text
stringlengths
14
1.76M
# On the automorphism group of a toral variety Anton Shafarevich and Anton Trushin Moscow Center of Fundamental and Applied Mathematics, Moscow, Russia; and Faculty of Computer Science, HSE University, Pokrovsky Boulevard 11, Moscow, 109028 Russia<EMAIL_ADDRESS>Moscow Center of Fundamental and Applied Mathematics, Moscow, Russia; and Faculty of Computer Science, HSE University, Pokrovsky Boulevard 11, Moscow, 109028 Russia<EMAIL_ADDRESS> ###### Abstract. Let ${\mathbb{K}}$ be an algebraically closed field of characteristic zero. An affine algebraic variety $X$ over ${\mathbb{K}}$ is toral if it is isomorphic to a closed subvariety of a torus $({\mathbb{K}}^{*})^{d}$. We study the group $\operatorname{Aut}(X)$ of regular automorpshims of a toral variety $X$. We prove that if $T$ is a maximal torus in $\operatorname{Aut}(X)$, then $X$ is a direct product $Y\times T$, where $Y$ is a toral variety with a trivial maximal torus in the automorphism group. We show that knowing $\operatorname{Aut}(Y)$, one can compute $\operatorname{Aut}(X)$. In the case when the rank of the group ${\mathbb{K}}[Y]^{*}/{\mathbb{K}}^{*}$ is $\dim Y+1$, the group $\operatorname{Aut}(Y)$ can be described explicitly. ###### Key words and phrases: Affine variety, invertible function, algebraic torus, automorphism, rigid variety ###### 2010 Mathematics Subject Classification: Primary 14M25, 14L30; Secondary 14R20, 13N15, 14M17 The work was supported by the Foundation for the Advancement of Theoretical Physics and Mathematics “BASIS” ## 1\. Introduction Let ${\mathbb{K}}$ be an algebraically closed field of characteristic zero. The set of solutions of a system of polynomial equations in affine space has been studied for a very long time. But some interesting properties may appear when we consider the set of solutions inside a torus $({\mathbb{K}}^{*})^{d}$. In other words we consider only solutions with non-zero coordinates. One of examples of this approach is the Bernstein-Kushnirenko Theorem; see [2, 6]. Vladimir Popov in [9] proposed the following definition. ###### Definition 1. An irreducible affine algebraic variety $X$ is called _toral_ if it is isomorphic to a closed subvariety of a torus $({\mathbb{K}}^{*})^{d}$. Some authors also use the term a ”very affine variety”; see [11, 3]. It can be seen that $X$ is toral if and only if the algebra of regular functions on $X$ is generated by invertible functions; see [9, Lemma 1.14]. One of the reasons why toral varieties are interesting is that they are rigid varieties; see [9, Lemma 1.14]. ###### Definition 2. An affine algebraic variety $X$ is called _rigid_ if there is no non-trivial action of the additive group $({\mathbb{K}},+)$ on $X.$ Despite the fact that the automorphism group of an affine algebraic variety has a complicated structure sometimes it is possible to describe it for rigid varieties. It was proven in [1] that the group of regular automorphisms $\operatorname{Aut}(X)$ of a rigid variety $X$ contains a unique maximal torus $T$. Moreover if we denote by $C(T)$ the centralizer of $T$ in $\operatorname{Aut}(X)$ then the quotient group $\operatorname{Aut}(X)/C(T)$ is a discrete group. One can find examples of computation of $\operatorname{Aut}(X)$ for rigid varieties in [1, 7, 8]. In this paper we study the automorphism group $\operatorname{Aut}(X)$ of a toral variety $X$. We denote by ${\mathbb{K}}[X]$ the algebra of regular functions on $X$ and by ${\mathbb{K}}[X]^{*}$ the multiplicative group of invertible regular functions on $X$. Let $E(X)$ be the quotient group ${\mathbb{K}}[X]^{*}/{\mathbb{K}}^{*}.$ By [10], the group $E(X)$ is a free finitely generated abelian group. For a toral variety $X$ the rank of $E(X)$ is not less than $\dim X.$ Any automorphism of $X$ induces an automorphism of $E(X).$ So we obtain a homomorphism from $\operatorname{Aut}(X)$ to $E(X).$ We denote by $H(X)$ the kernel of this homomorphism. Suppose that $X$ is a closed subvariety of a torus $T_{d}=({\mathbb{K}}^{*})^{d}$. In Proposition 1 we show that the group $H(X)$ is naturally isomorphic to a subgroup in $T_{d}$ which consists of elements that preserve $X$ under the action by multiplication. In Proposition 2 we propose a way to compute the subgroup $H(X)$. In Theorem 1 we show that if $T$ is a maximal torus in $\operatorname{Aut}(X)$ then $X$ is isomorphic to a direct product $T\times Y$ where $Y$ is a toral variety with a discrete automorphism group. Theorem 3 gives a way to find $\operatorname{Aut}(X)$ knowing $\operatorname{Aut}(Y).$ If the rank of $E(Y)$ is $\dim Y+1$ then one can describe $\operatorname{Aut}(Y)$ (Theorem 3). Also we consider the case when the rank of $E(X)$ is equal to $\dim X.$ By Proposition 3 in this case $X$ is a torus. Moreover, it is the only case when $\operatorname{Aut}(X)$ acts on $X$ with an open orbit. We use the following notation. If $\varphi$ is a regular automorphism of an affine variety $X$ then by $\varphi^{*}$ we mean an automorphism of ${\mathbb{K}}[X]$ dual to $\varphi$. If $A$ is a group and $B$ is a normal subgroup in $A$ then by $[a]$ we denote the image of an element $a\in A$ in the quotient group $A/B.$ If $X$ is a closed subvariety of an affine variety $Z$ then by $I(X)$ we mean the ideal of regular functions on $Z$ which are equal to zero on $X$. The authors are grateful to Segrey Gaifullin for useful discussions. Also we would like to thank Ivan Arzhantsev for helpful remarks and comments. ## 2\. General facts on toral varieties Here we prove some initial properties of toral varieties and propose a way to compute the group $H(X)$ for a toral variety $X$ Let $T_{d}$ be a torus of dimension $d$. We recall that the group $\operatorname{Aut}(T_{d})$ is isomorphic to $T_{d}\rtimes\mathrm{GL}_{d}({\mathbb{Z}})$; see[1, Example 2.3]. Here the left factor $T_{d}$ acts on itself by multiplications and a matrix $(a_{ij})\in\mathrm{GL}_{d}({\mathbb{Z}})$ defines an automorphism of $T_{d}$ which is given by the formula $t_{i}\to t_{1}^{a_{i1}}\ldots t_{d}^{a_{id}},$ where $t_{1},\ldots,t_{d}$ are coordinate functions on $T_{d}.$ ###### Proposition 1. 1. (1) Let $X$ be a closed irreducible subvariety of the torus $T_{d}.$ Then $H(X)$ is a quasitorus isomorphic to a closed subgroup in $T_{d}$ which consists of all elements that preserve $X$. 2. (2) Let $X$ be a toral variety. Suppose that the rank of the group $E(X)$ is equal to $r$. Then $X$ is isomorphic to a closed subvariety of $T_{r}=({\mathbb{K}}^{*})^{r}$ and the group $\operatorname{Aut}(X)$ is isomorphic to a subgroup of $\operatorname{Aut}(T_{r})=T_{r}\rtimes\mathrm{GL}_{r}({\mathbb{Z}})$ which consists of all automorphisms of $T_{r}$ that preserve $X.$ ###### Proof. 1) Let $H$ be a subgroup in $T_{d}$ which consists of all elements which preserve $X$. Any element of $H$ defines an automorphism of $X$. So we obtain a homomorphism $\psi:H\to\operatorname{Aut}(X).$ We will show that $\psi$ is injective and the image of $\psi$ is $H(X)$. Let $t_{1},\ldots,t_{d}$ be coordinate functions on $T_{d}.$ Then their restrictions to $X$ define invertible functions $f_{1},\ldots,f_{d}$ which generate ${\mathbb{K}}[X].$ Let $\varphi$ be an automorphism in $H(X)$. Then the automorphism $\varphi^{*}$ defines a trivial automorphism of $E(X).$ So $\varphi^{*}(f_{i})=\lambda_{i}f_{i},$ for some $\lambda_{1},\ldots,\lambda_{d}\in{\mathbb{K}}^{*}.$ Then the element $(\lambda_{1},\ldots,\lambda_{d})\in T_{d}$ preserves $X$ and its restriction to $X$ is $\varphi.$ So the image of $\psi$ contains $H(X).$ Conversely, consider an element $t=(\lambda_{1},\ldots,\lambda_{d})\in H$. Then $(\psi(t))^{*}$ maps $f_{i}$ to $\lambda_{i}f_{i}.$ So it acts trivially on $E(X).$ Therefore, the image of $\psi$ is contained in $H(X).$ The group $T_{d}$ acts on itself freely. So only the identity can act on $X$ trivially. Hence $\psi$ is injective. 2) Since rank $E(X)$ is equal to $r$ one can choose invertible functions $f_{1},\ldots,f_{r}\in{\mathbb{K}}[X]^{*}$ such that $[f_{1}],\ldots,[f_{r}]$ form a basis of the group $E(X)$. Then $f_{1},\ldots,f_{r}$ generate the algebra ${\mathbb{K}}[X]$ and there is a closed embedding of $X$ into $T_{r}.$ We denote by $t_{1},\ldots,t_{r}$ coordinate functions on $T_{r}$ which corresponds to $f_{1},\ldots,f_{r}.$ Let $P$ be the subgroup in $T_{r}\rtimes\mathrm{GL}_{r}({\mathbb{Z}})$ of all automorphisms that preserve $X$. Again, there is a homomorphism $\psi:P\to\operatorname{Aut}(X)$ and we will show that $\psi$ is a bijection. Each automorphism $\varphi\in\operatorname{Aut}(X)$ defines an automorphism of the group $E(X).$ Since $E(X)$ is a free abelian group of the rank $r$ the group $\operatorname{Aut}E(X)$ is isomorphic to $\mathrm{GL}_{r}({\mathbb{Z}}).$ So the automorphism $\varphi^{*}$ has a form $\varphi^{*}(f_{i})=\lambda_{i}f_{1}^{a_{i1}}\ldots f_{r}^{a_{ir}},$ where $(a_{ij})\in\mathrm{GL}_{r}({\mathbb{Z}})$ and $\lambda_{i}\in{\mathbb{K}}^{*}.$ But then the automorphism $\varphi$ is the restriction of an automorphism $\widetilde{\varphi}\in\operatorname{Aut}(T_{r})$ which is given by the formula $\widetilde{\varphi}(t_{i})=\lambda_{i}t_{1}^{a_{i1}}\ldots t_{r}^{a_{ir}}.$ So $\psi$ is surjective. Now let $\phi$ be an automorphism of $T_{r}$ which acts trivially on $X$. Then $\phi$ is given by the formula $\phi(t_{i})=\lambda_{i}t_{1}^{a_{i1}}\ldots t_{r}^{a_{ir}}$ where $\lambda_{i}\in{\mathbb{K}}^{*}$ and the matrix $(a_{ij})$ belongs to $\mathrm{GL}_{r}({\mathbb{Z}}).$ The corresponding automorphism of ${\mathbb{K}}[X]$ is given by the formula $f_{i}\to\lambda_{i}f_{1}^{a_{i1}}\ldots f_{r}^{a_{ir}}$ but it is a trivial automorphism so we have $f_{i}=\lambda_{i}f_{1}^{a_{i1}}\ldots f_{r}^{a_{ir}}.$ But the elements $[f_{1}],\ldots[f_{r}]\in E(X)$ are linearly independent. The matrix $(a_{ij})$ is the identity matrix. So $\phi(t_{i})=\lambda_{i}t_{i}$ and $\phi$ belongs to $T_{r}\subseteq\operatorname{Aut}(T_{r})$. But the group $T_{r}$ acts freely on itself. Hence $\phi$ is a trivial automorphism. So $\psi$ is injective. ∎ ###### Remark 1. It follows from Proposition 1 that a toral variety $X$ can be embedded in a torus $T_{r}$ in such a way that any automorphism $X$ can be extended to an automorphism of $T_{r}.$ If $X$ is a subvariety of $Z$ it is always natural to ask whether an automorphism of $X$ can be extended to an automorphism of $Z$. Some results concerning this problem can be found in [4, 5]. Now let $X$ be a closed irreducible subvariety in $T_{d}.$ Then one can describe the subgroup $H(X).$ Let $M\simeq{\mathbb{Z}}^{d}$ be the lattice of characters of $T_{d}$. For $m=(m_{1},\ldots,m_{d})\in M$ by $\chi^{m}$ we mean the character $t\to t_{1}^{m_{1}}\ldots t_{d}^{m_{d}}.$ Then each function in ${\mathbb{K}}[t_{1}^{\pm 1},\ldots,t_{d}^{\pm 1}]$ is a linear combinations of characters. For a function $f=\sum_{i}\alpha_{m_{i}}\chi^{m_{i}}\in{\mathbb{K}}[t_{1}^{\pm 1},\ldots,t_{d}^{\pm 1}]$ by _support of $f$_ we mean the subset $\operatorname{Supp}f=\\{m_{i}\in M|\ \alpha_{m_{i}}\neq 0\\}\subseteq M.$ Let $I(X)$ be the ideal of functions in ${\mathbb{K}}[t_{1},\ldots,t_{d}]$ which are equal to zero on $X$. We say that $f\in I(X)$ is _minimal_ if there is no non-zero $g\in I(X)$ such that $\operatorname{Supp}g\subsetneq\operatorname{Supp}f.$ ###### Lemma 1. The ideal $I(X)$ is generated by minimal polynomials. ###### Proof. If $f\in I(X)$ is not minimal then there is a $g\in I(X)$ with $\operatorname{Supp}g\subsetneq\operatorname{Supp}f$. One can choose a constant $\alpha$ such that $\operatorname{Supp}(f-\alpha g)\subsetneq\operatorname{Supp}f.$ Applying induction by cardinality of $\operatorname{Supp}f$ we see that $g$ and $f-\alpha g$ can be represented as a sum of minimal polynomials. Then $f$ is also a sum of minimal polynomials. ∎ ###### Remark 2. We denote by $M(X)$ the subgroup of $M$ which is generated by all elements of the form $m_{a}-m_{b}$ such that there is a minimal polynomial $f=\sum\alpha_{m_{i}}\chi^{m_{i}}\in I(X)$ with non-zero $\alpha_{m_{a}}$ and $\alpha_{m_{b}}$. ###### Proposition 2. The subgroup $H(X)$ is isomorphic to a subgroup in $T_{d}$ which is given by equations $\chi^{m}(t)=1$ for all $m\in M(X).$ ###### Proof. Denote by $H$ the subgroup in $T_{d}$ consisting of all elements which preserve $X$. By Proposition 1 the group $H$ is isomorphic to $H(X).$ We will show that $H$ is given in $T_{d}$ by equations $\chi^{m}(t)=1$ for all $m\in M(X).$ Let $h\in H$ and $f=\sum_{i}\alpha_{m_{i}}\chi^{m_{i}}$ be a minimal polynomial in $I(X).$ Then $h\circ f=\sum_{i}\alpha_{i}\chi^{m_{i}}(h)\chi^{m_{i}}.$ The ideal $I(X)$ is invariant under the action of $H$. So $h\circ f\in I(X)$. Suppose that there are $a,b$ such that $\alpha_{a},\alpha_{b}\neq 0$ and $\chi^{m_{a}}(h)\neq\chi^{m_{b}}(h).$ Then $g=\chi^{m_{a}}(h)f-h\circ f$ is a non-zero function in $I(X)$ and $\operatorname{Supp}g\subsetneq\operatorname{Supp}f.$ But $f$ is minimal. So $\chi^{m_{a}}(h)=\chi^{m_{b}}(h)$. Therefore, $\chi^{m_{a}-m_{b}}(h)=1$ and this implies that $\chi^{m}(h)=1$ for all $m\in M(X).$ Now consider the subgroup $H^{\prime}=\\{t\in T_{d}|\chi^{m}(t)=1,\ \forall\ m\in M(X)\\}.$ Then every minimal polynomial in $I(X)$ is semi-invariant with respect to $H^{\prime}$. But $I(X)$ is generated by minimal polynomials. So $I(X)$ is invariant under the action of $H^{\prime}$. Therefore, $H^{\prime}\subseteq H.$ ∎ At the end of this section, we note that toral varieties satisfy the following conjecture formulated by Alexander Perepechko and Mikhail Zaidenberg. ###### Conjecture 1 (Conjecture 1.0.1 in [8]). If $Y$ is a rigid affine algebraic variety over ${\mathbb{K}}$, then the connected component $\operatorname{Aut}^{0}(Y)$ is an algebraic torus of rank not greater than $\dim Y$. Indeed, if $Y$ is a toral variety then the group $\operatorname{Aut}(Y)/H(Y)$ is isomorphic to a subgroup in $\operatorname{Aut}(E(Y))\simeq\mathrm{GL}_{r}({\mathbb{Z}})$, where $r$ is a rank of $E(Y).$ Then $\operatorname{Aut}(Y)/H(Y)$ is a discrete group. So $\operatorname{Aut}^{0}(Y)$ is contained in $H(Y).$ But $H(Y)$ is a quasitorus. Therefore, $\operatorname{Aut}^{0}(Y)$ is a torus. ## 3\. The structure of the automorphism group It follows from [9, Lemma 1.14] that toral varieties are rigid. By [1, Theorem 2.1], there is a unique maximal torus in the automorphism group of an irreducible rigid variety. ###### Theorem 1. Let $X$ be a toral variety and $T$ be the maximal torus in $\operatorname{Aut}(X).$ Then $X\simeq Y\times T$ where $Y$ is a toral variety with a discrete automorphism group. ###### Proof. Let $r$ by the rank of the group $E(X).$ By Proposition 1, the variety $X$ can be embedded into the torus $T_{r}.$ We denote by $M$ the lattice of characters of $T_{r}$ and by $M(X)$ the sublattice in $M$ which corresponds to $X$; see Remark 2. One can choose a basis $e_{1},\ldots,e_{r}\in M$ such that $b_{1}e_{1},\ldots b_{l}e_{l}$ is a basis of $M(X)$ for some $b_{1},\ldots,b_{l}\in{\mathbb{N}}$ and $l\leq r$. Denote by ${t_{1}},\ldots{t_{r}}$ coordinates on $T_{r}$ corresponding to $e_{1},\ldots,e_{r}$. Then the equations $\chi^{m}(t)=1$ for all $m\in M(X)$ define a subgroup $H$ in $T_{r}$ which consists of elements of the form $(\epsilon_{1},\ldots,\epsilon_{l},t_{l+1},\ldots,t_{r}),$ where $\epsilon_{1},\ldots,\epsilon_{l}$ are the roots of unity of degrees $b_{1},\ldots,b_{l}$ respectively and $t_{l+1},\ldots t_{r}\in{\mathbb{K}}^{*}.$ Then all minimal polynomials in $I(X)$ are semi-invariant with respect to $H$. This means that minimal polynomials in $I(X)$ are homogeneous with respect to variables ${t}_{l+1},\ldots,{t}_{r}.$ Since functions ${t}_{i}$ are invertible one can choose a set of minimal generators of $I(X)$ which do not depend on ${t}_{l+1},\ldots,{t}_{r}.$ It implies that $X\simeq Y\times T_{r-l}$ where $Y$ is a subvariety of $T_{l}.$ We see that $T_{r-l}$ coincides with the maximal torus in $H\simeq H(X)$. The variety $Y$ is also a toral variety which is given by the ideal $I(X)\cap{\mathbb{K}}[t_{1},\ldots t_{l}].$ Since the maximal torus in $\operatorname{Aut}(X)$ is isomorphic to $T_{l}$ the maximal torus in $\operatorname{Aut}(Y)$ is trivial. It follows from Proposition 1 that in this case $\operatorname{Aut}(Y)$ is a discrete group. ∎ Let $X$ be a toral variety and suppose that $X\simeq T_{s}\times Y$ where $Y$ is a toral variety with a discrete automorphism group and $T_{s}$ is the torus $({\mathbb{K}}^{*})^{s}.$ One can see that $\operatorname{Aut}(X)$ contains the following subgroups. There is a subgroup which is isomorphic to $\operatorname{Aut}(Y)$. This subgroup acts in a natural way on $Y$ and acts trivially on $T_{s}.$ The subgroup $\mathrm{GL_{s}}({\mathbb{Z}})$ acts naturally on $T_{s}$ and trivially on $Y$. Moreover there is a subgroup which is isomorphic to $({\mathbb{K}}[Y]^{*})^{s}\simeq(E(Y)\times{\mathbb{K}}^{*})^{s}.$ This subgroup acts in the following way. If $f_{1},\ldots,f_{s}\in{\mathbb{K}}[Y]^{*}$ then we can define an automorphism of $T_{s}\times Y$ as follows $(t_{1},\ldots,t_{s},y)\to(f_{1}(y)t_{1},\ldots,f_{s}(y)t_{s},y).$ The following theorem was proposed to the authors by Sergey Gaifullin. ###### Theorem 2. Let $X\simeq T_{s}\times Y$ be a toral variety, where $Y$ is a toral variety with a discrete automorphism. Then $\operatorname{Aut}(X)\cong\operatorname{Aut}(Y)\ltimes(\mathrm{GL}_{s}({\mathbb{Z}})\ltimes(E(Y)\times{\mathbb{K}}^{*})^{s}).$ ###### Proof. There is a natural action of $T_{s}$ on $X$. We see that ${\mathbb{K}}[Y]$ is the algebra of invariants of this action. Since $T_{s}$ is a maximal torus in $\operatorname{Aut}(X)$ each automorphism of $T_{s}\times Y$ preserves ${\mathbb{K}}[Y].$ So we obtain a homomorphism $\Phi:\operatorname{Aut}(X)\to\operatorname{Aut}(Y).$ Let $B$ be the kernel of $\Phi$. The group $\operatorname{Aut}(Y)$ is naturally embedded into $\operatorname{Aut}(T_{s}\times Y)$ and intersects trivially with $B$. At the same time $\operatorname{Aut}(Y)$ maps isomorphically to the image of $\Phi$. It implies that $\operatorname{Aut}(T_{s}\times Y)=\operatorname{Aut}(Y)\ltimes B.$ We denote by $t_{1},\ldots,t_{s}$ coordinate functions on $T_{s}.$ Then ${\mathbb{K}}[T_{s}\times Y]\simeq{\mathbb{K}}[T_{s}]\otimes{\mathbb{K}}[Y]={\mathbb{K}}[Y][t_{1}^{\pm 1},\ldots t_{s}^{\pm 1}].$ Let $\phi\in B.$ The algebra ${\mathbb{K}}[Y]$ is invariant with respect to $\phi^{*}$. So for all $t\in T_{s}$ and $y\in Y$ we have $\phi((t,y))=(t^{\prime},y),$ for some $t^{\prime}\in T_{s}.$ Therefore, for each $y\in Y$ the automorphism $\phi$ defines an automorphism $\phi_{y}:T_{s}\to T_{s}.$ Hence for each $y\in Y$ we have $\phi^{*}(t_{i})(t,y)=t_{i}((\phi(t),\phi(y)))=t_{i}((\phi_{y}(t),y))=f_{i}(y)t_{1}^{a_{i1}(y)}\ldots t_{s}^{a_{is}(y)}$ for some non-zero constant $f_{i}(y)$ and a matrix $A(y)=(a_{ij}(y))\in\mathrm{GL}_{s}({\mathbb{Z}}).$ In reasons of continuity the matrix $A(y)$ is the same for all $y\in Y$ and $f_{i}:Y\to Y$ are regular functions on $Y$. Since $f_{i}(y)\neq 0$ for all $y\in Y$ the functions $f_{i}$ are invertible. So we have $\phi^{*}(t_{i})=f_{i}t_{1}^{a_{i1}}\ldots t_{s}^{a_{is}}$ for some $f_{i}\in{\mathbb{K}}[Y]^{*}$ and $A\in\mathrm{GL}_{s}({\mathbb{Z}}).$ Then we have a homomorphism $\overline{\Phi}:B\to\mathrm{GL}_{s}({\mathbb{Z}}),\ \phi\to A.$ Again, it is easy to see that $\mathrm{GL}_{s}({\mathbb{Z}})$ is naturally embedded into $B$ and maps isomorphically to $\mathrm{GL}_{s}({\mathbb{Z}})$ with respect to $\overline{\Phi}.$ So $B=\mathrm{GL}_{s}({\mathbb{Z}})\ltimes\mathrm{Ker}\ \overline{\Phi}.$ The kernel of $\overline{\Phi}$ consists of automorphisms $\varphi\in\operatorname{Aut}(T_{s}\times Y)$ which have a form $\varphi(t_{1},\ldots,t_{s},y)=(f_{1}(y)t_{1},\ldots,f_{s}(y)t_{s},y).$ for some $f_{1},\ldots,f_{s}\in{\mathbb{K}}[Y]^{*}.$ We see that for all $f_{1},\ldots,f_{s}\in{\mathbb{K}}[Y]^{*}$ this formula defines an automorphism of $T_{s}\times Y,$ so $\mathrm{{\rm Ker}\,}\ \overline{\Phi}=({\mathbb{K}}[Y]^{*})^{s}=(E(Y)\times{\mathbb{K}}^{*})^{s}.$ ∎ ## 4\. The case $\mathrm{rk}\ E(X)=\dim X$ Let $X$ be a toral variety. Then $\mathrm{rk}\ E(X)\geq\dim X.$ Indeed, suppose that $f_{1},\ldots,f_{r}$ are invertible functions and $[f_{1}],\ldots,[f_{r}]$ is a basis in $E(X).$ Then $f_{1},\ldots,f_{r}$ generate ${\mathbb{K}}[X].$ So $r\geq\dim X$. The following result shows that if $\mathrm{rk}\ E(X)=\dim X$ then $X$ is a torus. Moreover this is the only case when $\operatorname{Aut}(X)$ acts with an open orbit on $X$. ###### Proposition 3. Let $X$ be a toral variety. Then the following conditions are equivalent. 1. (1) $X$ is a torus; 2. (2) $\mathrm{rk}\ E(X)=\dim X.$ Suppose also that the field ${\mathbb{K}}$ is uncountable. Then $X$ is a torus if and only if $\operatorname{Aut}(X)$ acts on $X$ with an open orbit. ###### Proof. The implication $1)\Rightarrow 2)$ is trivial. Suppose that $\mathrm{rk}\ E(X)=\dim X.$ Then one can choose invertible functions $f_{1},\ldots,f_{n}$ such that $[f_{1}],\ldots,[f_{n}]$ is a basis of $E(X)$. Then ${\mathbb{K}}[X]$ is generated by $f_{1},f_{1}^{-1},\ldots,f_{n},f_{n}^{-1}$. But $f_{1},\ldots,f_{n}$ are algebraically independent, otherwise $\dim X<\mathrm{rk}\ E(X)$. So ${\mathbb{K}}[X]$ is isomoprhic to the algebra of Laurent polynomials. So we obtain implication $2)\Rightarrow 1).$ Now we prove the second statement. If $X$ is a torus then $\operatorname{Aut}(X)$ acts on $X$ transitively. So suppose $X$ is a toral variety over an uncountable field ${\mathbb{K}}$ and $\operatorname{Aut}(X)$ acts on $X$ with an open orbit $U$. Let $T$ be the maximal torus in $\operatorname{Aut}(X).$ Sine the quotient group $\operatorname{Aut}(X)/T$ is a discrete group the set $U$ is a countable union of orbits of $T.$ Since ${\mathbb{K}}$ is uncountable it implies that one of the orbits of $T$ is open in $X.$ Then $\dim X=\dim T.$ By Theorem 1 we have $X\simeq T\times Y$ for some toral variety $Y$. But since $\dim T=\dim T\times Y$ we obtain that $Y$ is a point and $X\simeq T.$ ∎ ## 5\. The case $\mathrm{rk}\ E(X)=\dim X+1$ By Theorem 1 any toral variety is a direct product $T\times Y$ where $T$ is a torus and $Y$ is a toral variety with a discrete automorphism group. By Theorem 2 one can find $\operatorname{Aut}(X)$ knowing $\operatorname{Aut}(Y).$ In this section we provide a way to find $\operatorname{Aut}(Y)$ when $\mathrm{rk}\ E(Y)=\dim\ Y+1.$ Let $Y$ be a toral variety with a trivial maximal torus in $\operatorname{Aut}(Y)$. Let $r$ be the rank of $E(Y)$. We suppose that $r=\dim(Y)+1.$ By Proposition 1 the variety $Y$ can be embedded into $T_{r}$ as a hypersurface. Then there is an irreducible polynomial $h\in{\mathbb{K}}[t_{1}^{\pm 1},\ldots,t_{r}^{\pm 1}]$ such that $I(Y)=(h).$ The polynomial $h$ has a form $h=\sum_{m\ \in\ \mathrm{Supp}\ h}\alpha_{m}\chi^{m}.$ Let $M$ be the lattice of characters of $T_{r}$ and $M(Y)$ be a sublattice in $M$ which corresponds to $Y$; see Definition 2. Since the maximal torus in $\operatorname{Aut}(Y)$ is trivial, the rank of the lattice $M(Y)$ is equal to $r$. It means that the elements $m_{a}-m_{b}$ with $m_{a},m_{b}\in\mathrm{Supp}\ h$ generate a sublattice of full rank in $M$. We denote by $\mathrm{GAff}(M,h)$ the group of all degenerate integer affine transformations $\varphi$ of $M$ which preserve $\mathrm{Supp}\ h$ and for any linear combination $\sum_{i}{a_{i}}(m_{i,1}-m_{i,2})=0,\ a_{i}\in\mathbb{Z},\ m_{i,1},m_{i,2}\in\mathrm{Supp}\ h$ $\varphi$ satisfies $\prod_{i}\left(\frac{\alpha_{m_{i,1}}}{\alpha_{m_{i,2}}}\right)^{a_{i}}=\prod_{i}\left(\frac{\alpha_{\varphi(m_{i,1})}}{\alpha_{\varphi(m_{i,2})}}\right)^{a_{i}}.$ (1) ###### Remark 3. Let us fix a point $m_{0}\in\mathrm{Supp}\ h.$ Then any equation $\sum_{i}{a_{i}}(m_{i,1}-m_{i,2})=0,\ a_{i}\in\mathbb{Z},\ m_{i,1},m_{i,2}\in\mathrm{Supp}\ h$ can be written as $\sum_{i}{b_{i}}(m_{i}-m_{0})=0,\ b_{i}\in{\mathbb{Z}},\ m_{i}\in\mathrm{Supp}\ h.$ (2) Therefore, it is enough to check equations (1) for linear combinations of the form (2). Let $N$ be the number of points in $\mathrm{Supp}\ h.$ Then the set $\\{(b_{1},\ldots,b_{N-1})\in\mathbb{Z}^{N-1}|\sum_{i}{b_{i}}(m_{i}-m_{0})=0\\}$ is a sublattice in ${\mathbb{Z}}^{N-1}$ and it is enough to check the equations in (1) for some basis in this sublattice. ###### Theorem 3. Let $Y$ be a toral variety with trivial maximal torus in $\operatorname{Aut}(Y).$ Suppose that $\mathrm{rk}\ E(Y)=\dim Y+1.$ Then $\operatorname{Aut}(Y)/H(Y)\simeq\mathrm{GAff}(M,h).$ ###### Proof. Let $\psi$ be an automorphism of $Y$. By Proposition 1 the automorphism $\psi$ can be extended to an automorphism of $T_{r}$. We denote by $\psi^{*}$ the respective automorphism of ${\mathbb{K}}[t^{\pm 1}_{1},\ldots,t^{\pm 1}_{r}]$. Then $\psi^{*}$ has the form $\psi^{*}(t_{i})=\lambda_{i}t_{1}^{a_{i1}}\ldots t_{r}^{a_{ir}},$ where $\lambda_{i}\in{\mathbb{K}}^{*}$ and $(a_{ij})\in\mathrm{GL}_{r}({\mathbb{Z}}).$ We denote by $\lambda$ the element $(\lambda_{1},\ldots,\lambda_{r})\in T_{r}$ and by $\overline{\psi}$ the automorphism of $M$ which respects to the matrix $(a_{ij}).$ Then $\psi^{*}(\chi^{m})=\chi^{m}(\lambda)\chi^{\overline{\psi}(m)}$ for all $m\in M.$ The polynomial $\psi^{*}(h)$ also generates $I(Y).$ So it differs from $h$ by an invertible element of ${\mathbb{K}}[t^{\pm 1}_{1},\ldots,t^{\pm 1}_{r}].$ Then $\psi^{*}(h)=\alpha\chi^{v}h$ for some $\alpha\in{\mathbb{K}}^{*}$ and $v\in M.$ Therefore, we have an equation $\psi^{*}(h)=\sum_{m\ \in\ \mathrm{Supp}\ h}\alpha_{m}\chi^{m}(\lambda)\chi^{\overline{\psi}(m)}=\alpha\sum_{m\ \in\ \mathrm{Supp}\ h}\alpha_{m}\chi^{m+v}.$ (3) We define the map $\varphi:M\to M$ by the following formula: $\varphi(m)=\overline{\psi}(m)-v.$ Then $\psi$ is an affine transformation of $M$ which preserves $\mathrm{Supp}\ h.$ We consider a linear combination $\sum_{i}{a_{i}}(m_{i,1}-m_{i,2})=0,\ a_{i}\in\mathbb{Z},\ m_{i,1},m_{i,2}\in\mathrm{Supp}\ h.$ Equation (3) can be written as $\sum_{m\ \in\ \mathrm{Supp}\ h}\alpha_{m}\chi^{m}(\lambda)\chi^{\varphi(m)}=\alpha\sum_{m\ \in\ \mathrm{Supp}\ h}\alpha_{m}\chi^{m}=\alpha\sum_{m\ \in\ \mathrm{Supp}\ h}\alpha_{\varphi(m)}\chi^{\varphi(m)}.$ and it implies $\frac{\alpha_{m_{i,1}}\chi^{m_{i,1}}(\lambda)}{\alpha_{m_{i,2}}\chi^{m_{i,2}}(\lambda)}=\frac{\alpha_{m_{i,1}}}{\alpha_{m_{i,2}}}\chi^{m_{i,1}-m_{i,2}}(\lambda)=\frac{\alpha_{\varphi(m_{i,1})}}{\alpha_{\varphi(m_{i,2})}}$ for all $i.$ Then we have $\prod_{i}\left(\frac{\alpha_{m_{i,1}}}{\alpha_{m_{i,2}}}\chi^{m_{i,1}-m_{i,2}}(\lambda)\right)^{a_{i}}=\prod_{i}\left(\frac{\alpha_{m_{i,1}}}{\alpha_{m_{i,2}}}\right)^{a_{i}}\left(\chi^{\sum_{i}{a_{i}}(m_{i,1}-m_{i,2})}(\lambda)\right)=\prod_{i}\left(\frac{\alpha_{m_{i,1}}}{\alpha_{m_{i,2}}}\right)^{a_{i}}$ $=\prod_{i}\left(\frac{\alpha_{\varphi(m_{i,1})}}{\alpha_{\varphi(m_{i,2})}}\right)^{a_{i}}.$ So $\varphi\in\mathrm{Gaff}(M,h).$ Then we obtain a homomorphism $\eta:\operatorname{Aut}(Y)\to\mathrm{Gaff}(M,h).$ Moreover we see that the kernel of $\eta$ is $H(Y).$ Now we will show that $\eta$ is surjective. Let $\varphi\in\mathrm{Gaff}(M,h)$ and $f_{1},\ldots,f_{r}$ be a basis in $M(Y).$ Then there are $m_{i,1},m_{i,2}\in\ \mathrm{Supp}(h),$ and $a_{ij}\in{\mathbb{Z}}$ such that $f_{j}=\sum_{i}a_{ij}(m_{i,1}-m_{i,2}).$ There is a $\lambda\in T_{r}$ such that $\chi^{f_{j}}(\lambda)=\prod_{i}\left(\frac{\alpha_{m_{i,1}}}{\alpha_{m_{i,2}}}\right)^{-a_{ij}}\prod_{i}\left(\frac{\alpha_{\varphi(m_{i,1})}}{\alpha_{\varphi(m_{i,2})}}\right)^{a_{ij}}.$ Let $d\varphi$ be the differential of $\varphi.$ We define an automorphism $\psi^{*}$ of ${\mathbb{K}}[t_{1}^{\pm 1}\ldots t_{r}^{\pm 1}]$ by the following rule: $\psi^{*}(\chi^{m})=\chi^{m}(\lambda)\chi^{d\varphi(m)}.$ Let us check that $\psi^{*}$ preserves $I(Y).$ We have $\psi^{*}(h)=\sum\alpha_{m}\chi^{m}(\lambda)\chi^{d\varphi(m)}.$ There is $v\in M$ such that $\varphi(m)=d\varphi(m)+v.$ Then $\chi^{v}\psi^{*}(h)=\sum\alpha_{m}\chi^{m}(\lambda)\chi^{\varphi(m)}.$ We see that $\mathrm{Supp}\ \chi^{v}\psi^{*}(h)=\mathrm{Supp}\ h.$ We will show that there is $\alpha\in{\mathbb{K}}$ such that $\chi^{v}\psi^{*}(h)=\alpha h.$ For any $b,c\in\mathrm{Supp}\ h$ there are numbers $d_{j}\in{\mathbb{Z}}$ such that $b-c=\sum_{j}d_{j}f_{j}.$ So $\frac{\alpha_{b}\chi^{b}(\lambda)}{\alpha_{c}\chi^{c}(\lambda)}=\frac{\alpha_{c}}{\alpha_{b}}\chi^{b-c}(\lambda)=\frac{\alpha_{b}}{\alpha_{c}}\chi^{\sum_{j}d_{j}f_{j}}(\lambda)=$ $\frac{\alpha_{b}}{\alpha_{c}}(\prod_{j}\chi^{f_{j}}(\lambda))^{d_{j}}=\frac{\alpha_{b}}{\alpha_{c}}\prod_{i,j}\left(\frac{\alpha_{m_{i,1}}}{\alpha_{m_{i,2}}}\right)^{-d_{j}a_{ij}}\prod_{i,j}\left(\frac{\alpha_{\varphi(m_{i,1})}}{\alpha_{\varphi(m_{i,2})}}\right)^{d_{j}a_{ij}}.$ We have a combination $b-c-\sum_{j}d_{j}f_{j}=b-c-\sum_{j,i}d_{j}a_{ij}(m_{i,1}-m_{i,2})=0.$ Since $\varphi\in\mathrm{Gaff}(M,h)$ we obtain $\frac{\alpha_{b}}{\alpha_{c}}\prod_{i,j}\left(\frac{\alpha_{m_{i,1}}}{\alpha_{m_{i,2}}}\right)^{-d_{j}a_{ij}}=\frac{\alpha_{\varphi(b)}}{\alpha_{\varphi(c)}}\prod_{i,j}\left(\frac{\alpha_{\varphi(m_{i,1})}}{\alpha_{\varphi(m_{i,2})}}\right)^{-d_{j}a_{ij}}.$ It implies that $\frac{\alpha_{b}\chi^{b}(\lambda)}{\alpha_{c}\chi^{c}(\lambda)}=\frac{\alpha_{\varphi(b)}}{\alpha_{\varphi(c)}}.$ So the coefficients of the polynomials $\chi^{v}\psi^{*}(h)$ and $h$ are proportional. Then there is an $\alpha\in{\mathbb{K}}$ such that $\chi^{v}\psi^{*}(h)=\alpha h.$ Hence $\psi^{*}(h)=\alpha\chi^{-v}h\in I(Y)$. Therefore, $\psi^{*}$ preserves $I(Y)$ and defines an automorphism $\psi$. It is a direct check that $\eta(\psi)=\varphi$. So $\varphi$ is surjective. ∎ ###### Corollary 1. Let $Y$ be a toral variety with trivial maximal torus in $\operatorname{Aut}(Y).$ Suppose that $\mathrm{{\mathrm{rk}}}\ E(Y)=\dim Y+1.$ Then $\operatorname{Aut}(Y)$ is a finite group. It is natural to formulate the following question. ###### Question 1. Let $Y$ be a toral variety with trivial maximal torus in $\operatorname{Aut}(Y).$ Is $\operatorname{Aut}(Y)$ a finite group? Note that this is not true for a general rigid variety. One can find a counterexample in [7]. At the end, we give two examples illustrating Theorem 3. ###### Example 1. Let $Y$ be the affine line $\mathbb{A}^{1}$ without two points. The variety $Y$ can be given in $({\mathbb{K}}^{*})^{2}$ as the set of solutions of the equation $h=x+y-1=0,\ (x,y)\in({\mathbb{K}}^{*})^{2}.$ So $Y$ is a toral variety. Let $M\simeq{\mathbb{Z}}^{2}$ be the lattice of characetrs of $({\mathbb{K}}^{*})^{2}$. We see that the lattice $M(Y)$ contains elements $(1,0),(0,1),(1,-1)$, so $M(Y)=M.$ Therefore, $H(Y)$ is a trivial group. The group $E(Y)$ is generated by $[x]$ and $[y]$. So we can apply Theorem 3. Now we find the group $\mathrm{Gaff}(M,h)$. The set $\mathrm{Supp}\ h$ consists of points $m_{0}=(0,0),\ m_{1}=(1,0),m_{2}=(0,1).$ $m_{2}$$m_{0}$$m_{1}$$M$ Figure 1. $\mathrm{Supp}(x+y-1)$ A linear combination $a_{1}(m_{1}-m_{0})+a_{2}(m_{2}-m_{0})+a_{3}(m_{2}-m_{1})=a_{1}(1,0)+a_{2}(0,1)+a_{3}(-1,1)$ is equal to zero if and only if $a_{1}=a_{3}=-a_{2}.$ But then equations (1) are trivial. By affine transformations of $M$ we can permute all points in $\mathrm{Supp}\ h.$ Therefore, $\operatorname{Aut}(Y)\simeq\mathrm{GAff}(M,h)\simeq S_{3}.$ The answer looks natural since the affine line without two points is the projective line without three points. ###### Example 2. Let $a,b,c\in{\mathbb{K}}^{*}$ be non-zero constants and $Y$ be the set of solutions of the equation $h=1+ax+by+cxy=0,\ (x,y)\in({\mathbb{K}}^{*})^{2}.$ Let $M\simeq{\mathbb{Z}}^{2}$ be the lattice of characters of $({\mathbb{K}}^{*})^{2}$. The lattice $M(Y)$ contains $(1,0)$ and $(0,1)$ so $M(Y)=M.$ Then $H(Y)$ is a trivial group. The group $E(Y)$ is generated by $[x]$ and $[y]$. Again, we can apply Theorem 3. The set $\mathrm{Supp}\ h$ consists of points $m_{0}=(0,0),\ m_{1}=(1,0),m_{2}=(0,1),m_{3}=(1,1).$ $m_{3}$$m_{2}$$m_{0}$$m_{1}$$M$ Figure 2. $\mathrm{Supp}(1+ax+by+cxy)$ The set of affine transformation which preserve $\mathrm{Supp}\ h$ is the dihedral group $D_{4}.$ To understand what elements of $D_{4}$ belong to $\mathrm{GAff}(M,h)$ we need to check the equations (3). To simplify notations we will denote $m_{0}$ by $1,$ $m_{1}$ by $x,$ $m_{2}$ by $y$ and $m_{3}$ by $xy$. By $\alpha_{m}$ we denote the coefficient in $h$ of monomial $\chi^{m}$. So we have $\alpha_{1}=1,\ \alpha_{x}=a,\ \alpha_{y}=b,\ \alpha_{xy}=c.$ By Remark 3 it is enough to check the equations (3) only for linear combinations of the form $a_{1}(m_{1}-m_{0})+a_{2}(m_{2}-m_{0})+a_{3}(m_{3}-m_{0})=a_{1}(1,0)+a_{2}(0,1)+a_{3}(1,1)=0,$ where $a_{1},a_{2},a_{3}\in{\mathbb{Z}}.$ So we have $a_{1}=a_{2}=-a_{3}.$ Again, by Remark 3 it is enough to check the equation $\left(\frac{\alpha_{x}}{\alpha_{1}}\right)^{1}\left(\frac{\alpha_{y}}{\alpha_{1}}\right)^{1}\left(\frac{\alpha_{xy}}{\alpha_{1}}\right)^{-1}=\left(\frac{\alpha_{\varphi(x)}}{\alpha_{\varphi(1)}}\right)^{1}\left(\frac{\alpha_{\varphi(y)}}{\alpha_{\varphi(1)}}\right)^{1}\left(\frac{\alpha_{\varphi(xy)}}{\alpha_{\varphi(1)}}\right)^{-1}.$ So an element $\varphi\in D_{4}$ belongs to $\mathrm{GAff}(M,h)$ if and only if $\frac{ab}{c}=\frac{\alpha_{\varphi(x)}\alpha_{\varphi(y)}}{\alpha_{\varphi(1)}\alpha_{\varphi(xy)}}.$ Substituting various elements of the dihedral group into this equality, we obtain that the rotation by $\frac{\pi}{2}$ and the symmetries with respect to the lines passing through the midpoints of the opposite sides belong to $\mathrm{GAff}(M,h)$ if and only if $\frac{ab}{c}=\frac{c}{ab}\iff(ab)^{2}=c^{2}.$ The rotation by $\pi$ and the symmetries with respect to the diagonals belong to $\mathrm{GAff}(M,h)$ for any $a,b,c.$ So we obtain that if $(ab)^{2}=c^{2}$ then $\operatorname{Aut}(Y)\simeq D_{4}$ and if $(ab)^{2}\neq c^{2}$ then $\operatorname{Aut}(Y)\simeq({\mathbb{Z}}/2{\mathbb{Z}})\oplus({\mathbb{Z}}/2{\mathbb{Z}}).$ ## References * [1] Ivan Arzhantsev and Sergey Gaifullin. The automorphism group of a rigid affine variety. Math. Nachr., 290:5, 662–671, 2017 * [2] David Bernstein. The number of roots of a system of equations. Funct. Anal. Appl., 9:3, 183–185, 1975 * [3] June Huh. The maximum likelihood degree of a very affine variety. Compos. Math., 149:8, 1245–1266, 2013 * [4] Shulim Kaliman. Extensions of isomorphisms between affine algebraic subvarieties of $k^{n}$ to automorphisms of $k^{n}$. Proc. Amer. Math. Soc. 113:2, 325–334, 1991 * [5] Shulim Kaliman and David Udumyan. On automorphisms of flexible varieties. Adv. Math., 396, Article no. 108112, 2022 * [6] Anatoly Koushnirenko. Newton polytopes and the Bezout theorem. Funct. Anal. Appl., 10:3, 82–83, 1976 * [7] Alexander Perepechko. Automorphisms of surfaces of Markov type. Math. Notes, 110:5, 732–737, 2021 * [8] Alexander Perepechko and Michael Zaidenberg. Automorphism group of affine rigid surfaces: the identity component. arXiv:2208.09738, 2022 * [9] Vladimir Popov. On the Makar-Limanov, Derksen invariants, and finite automorphism groups of algebraic varieties. CRM Proc. Lect. Notes, 54:17, 289–311, 2011 * [10] Maxwell Rosenlicht. Some rationality questions on algebraic groups. Ann. Mat. Pura Appl., 43, 25–50, 1957 * [11] Jenia Tevelev. Compactifications of subvarieties of tori. Amer. J. Math., 129:4, 1087–1104, 2007
Omm$#2\text{\,}#\mathrm{3}$ HIKE, High Intensity Kaon Experiments at the CERN SPS Letter of Intent The HIKE Collaboration***Address all correspondence to<EMAIL_ADDRESS> ###### Abstract A timely and long-term programme of kaon decay measurements at a new level of precision is presented, leveraging the capabilities of the CERN Super Proton Synchrotron (SPS). The proposed programme is firmly anchored on the experience built up studying kaon decays at the SPS over the past four decades, and includes rare processes, CP violation, dark sectors, symmetry tests and other tests of the Standard Model. The experimental programme is based on a staged approach involving experiments with charged and neutral kaon beams, as well as operation in beam-dump mode. The various phases will rely on a common infrastructure and set of detectors. The HIKE collaboration E. Cortina Gil1, J. Jerhot1, N. Lurkin1, T. Numao2, B. Velghe2, V. W. S. Wong2, D. Bryman3, L. Bician4, Z. Hives4, T. Husek4, K. Kampf4, M. Koval4, A. T. Akmete5, R. Aliberti5, V. Büscher5, L. Di Lella5, N. Doble5, L. Peruzzo5, M. Schott5, H. Wahl5, R. Wanke5, B. Döbrich6, L. Montalto7, D. Rinaldi7, F. Dettori8,9, A. Cardini9, A. Lai9, L. Bomben10, S. Carsi10, M. Prest10, A. Selmi10, G. Lezzani11, P. Monti-Guarnieri11, L. Perna11, P. Dalpiaz12,13, V. Guidi,12,13, A. Mazzolari12,13, I. Neri12,13, F. Petrucci12,13, M. Soldani12,13, L. Bandiera13, A. Cotta Ramusino13, A. Gianoli13, M. Romagnoni13, A. Sytov13, M. Lenti14,15, I. Panichi14,15, G. Ruggiero14,15, A. Bizzeti15, F. Bucci15, A. Antonelli16, E. Di Meco16,30, G. Lanfranchi16, S. Martellotti16, M. Martini16,31, M. Moulson16, D. Paesani16,30, I. Sarra16, T. Spadaro16, G. Tinti16, E. Vallazza17, F. Ambrosino18,20, R. Giordano18,20, P. Massarotti18,20, M. Napolitano18,20, G. Saracino18,20, C. Di Donato19,20, G. D’Ambrosio20, M. D’Errico20, M. Mirra20, S. Neshatpour20, R. Fiorenza21, I. Rosa21, D. De Salvador22, F. Sgarbossa22, G. Anzivino23,24, S. Germani23,24, R. Volpe23,24, P. Cenci24, S. Cutini24, V. Duk24, P. Lubrano24, M. Pepe24, M. Piccini24, F. Costantini25,26, S. Donati25,26, M. Giorgi25,26, S. Giudici25,26, G. Lamanna25,26, E. Pedreschi25,26, J. Pinzino25,26, M. Sozzi25,26, R. Fantechi26, V. Giusti26, F. Spinella26, I. Mannelli27, M. Raggi28,29, A. Biagioni29, P. Cretaro29, O. Frezza29, F. Lo Cicero29, A. Lonardo29, M. Turisini29, P. Vicini29, R. Ammendola30, V. Bonaiuto30, A. Fucci30, A. Salamon30, F. Sargeni30, R. Arcidiacono32,33, B. Bloch-Devaux32, E. Menichetti32,33, E. Migliore32,33, C. Biino33, F. Marchetto33, D. Baigarashev34, Y. Kambar34, D. Kereibay34, Y. Mukhamejanov34, S. Sakhiyev34, A. Briano Olvera35, J. Engelfried35, N. Estrada-Tristan35, R. Piandani35, M. A. Reyes Santos35, K. A. Rodriguez Rivera35, P. C. Boboc36, A. M. Bragadireanu36, S. A. Ghinescu36, O. E. Hutanu36, T. Blazek37, V. Cerny37, A. Kleimenova37, Z. Kucerova37, D. Martinez Santos38, C. Prouve39, M. Boretto40, F. Brizioli40, A. Ceccucci40, M. Corvino40, H. Danielsson40, F. Duval40, E. Gamberini40, R. Guida40, E. B. Holzer40, B. Jenninger40, G. Lehmann Miotto40, P. Lichard40, K. Massri40, E. Minucci40, M. Perrin-Terrin40, V. Ryjov40, J. Swallow40, M. Van Dijk40, M. Zamkovsky40, R. Marchevski41, A. Gerbershagen42, J. R. Fry43, F. Gonnella43, E. Goudzovski43, J. Henshaw43, C. Kenworthy43, C. Lazzeroni43, C. Parkinson43, A. Romano43, J. Sanders43, A. Shaikhiev43, A. Tomczak43, H. Heath44, D. Britton45, A. Norton45, D. Protopopescu45, J. B. Dainton46, R. W. L. Jones46, A. De Santo47, F. Salvatore47, P. Cooper48, D. Coward48, P. Rubin48 1 Université Catholique de Louvain, B-1348 Louvain-La-Neuve, Belgium 2 TRIUMF, Vancouver, British Columbia, V6T 2A3, Canada 3 University of British Columbia, Vancouver, British Columbia, V6T 1Z4, Canada 4 Charles University, 116 36 Prague 1, Czech Republic 5 Johannes Gutenberg Universität Mainz, D-55099 Mainz, Germany 6 Max-Planck-Institut für Physik (Werner-Heisenberg-Institut), D-80805 München, Germany 7 Dipartimento di Scienze e Ingegneria della Materia, dell’Ambiente ed Urbanistica, Università Politecnica delle Marche, I-60131 Ancona, Italy 8 Dipartimento di Fisica dell’Università, I-09042 Cagliari, Italy 9 INFN, Sezione di Cagliari, I-09042 Cagliari, Italy 10 Università degli Studi dell’Insubria, 22100, Como and INFN Sezione Milano Bicocca, I-20126 Milano, Italy 11 Dipartimento di Scienza e Alta Tecnologia, Università degli Studi dell’Insubria, I-22100, Como, Italy 12 Dipartimento di Fisica e Scienze della Terra dell’Università, I-44122 Ferrara, Italy 13 INFN, Sezione di Ferrara, I-44122 Ferrara, Italy 14 Dipartimento di Fisica e Astronomia dell’Università, I-50019 Sesto Fiorentino, Italy 15 INFN, Sezione di Firenze, I-50019 Sesto Fiorentino, Italy 16 INFN Laboratori Nazionali di Frascati, I-00044 Frascati, Italy 17 INFN Sezione di Milano Bicocca, I-20126 Milano, Italy 18 Dipartimento di Fisica “Ettore Pancini” dell’Università degli Studi di Napoli Federico II, I-80126 Napoli, Italy 19 Dipartimento di Ingegneria, Università degli Studi di Napoli Parthenope, I-80143 Napoli 20 INFN, Sezione di Napoli, I-80126 Napoli, Italy 21 Scuola Superiore Meridionale e INFN, Sezione di Napoli, I-80138 Napoli, Italy 22 Dipartimento di Fisica e Astronomia dell’ Università degli Studi di Padova, I-35131 Padova e INFN Laboratori Nazionali di Legnaro, I-35020 Legnaro, Italy 23 Dipartimento di Fisica e Geologia dell’Università, I-06100 Perugia, Italy 24 INFN, Sezione di Perugia, I-06100 Perugia, Italy 25 Dipartimento di Fisica dell’Università, I-56100 Pisa, Italy 26 INFN, Sezione di Pisa, I-56100 Pisa, Italy 27 Scuola Normale Superiore e INFN, Sezione di Pisa, I-56100 Pisa, Italy 28 Dipartimento di Fisica, Sapienza Università di Roma, I-00185 Roma, Italy 29 INFN, Sezione di Roma I, I-00185 Roma, Italy 30 Dipartimento di Fisica, Università degli Studi di Roma Tor Vergata, I-00133 Roma, Italy 31 Dipartimento di Scienze Ingegneristiche, Università degli Studi Guglielmo Marconi, I-00193 Roma, Italy 32 Dipartimento di Fisica dell’Università, I-10125 Torino, Italy 33 INFN, Sezione di Torino, I-10125 Torino, Italy 34 Institute of Nuclear Physics, 050032 Almaty, Kazakhstan 35 Instituto de Física, Universidad Autónoma de San Luis Potosí, 78240 San Luis Potosí, Mexico 36 Horia Hulubei National Institute for R&D in Physics and Nuclear Engineering, 077125 Bucharest-Magurele, Romania 37 Faculty of Mathematics, Physics and Informatics, Comenius University, 842 48, Bratislava, Slovakia 38 Axencia Galega de Innovacion, Conselleria de Economia e Industria, Xunta de Galicia, 15704 Santiago de Compostela, Spain 39 Instituto Galego de Física de Altas Enerxías (IGFAE), Universidade de Santiago de Compostela, 15782 Santiago de Compostela, Spain 40 CERN, European Organization for Nuclear Research, CH-1211 Geneva 23, Switzerland 41 Ecole polytechnique fédérale de Lausanne (EPFL), CH-1015 Lausanne, Switzerland 42 PARTREC, UMCG, University of Groningen, 9747 AA Groningen, The Netherlands 43 School of Physics and Astronomy, University of Birmingham, Edgbaston, Birmingham, B15 2TT, UK 44 School of Physics, University of Bristol, Bristol, BS8 1TH, UK 45 School of Physics and Astronomy, University of Glasgow, Glasgow, G12 8QQ, UK 46 Faculty of Science and Technology, University of Lancaster, Lancaster, LA1 4YW, UK 47 Department of Physics and Astronomy, University of Sussex, Brighton, BN1 4GE, UK 48 Physics and Astronomy Department, George Mason University, Fairfax, VA 22030, USA ###### Contents 1. 1 Overview of HIKE 1. 1.1 The scientific goals of HIKE 2. 2 Scientific context 1. 2.1 Experimental context for high-scale BSM physics 2. 2.2 Experimental context for feebly-interacting particles 3. 2.3 The role of HIKE 3. 3 Physics with high-intensity kaon beams 1. 3.1 Flavour-changing neutral currents 2. 3.2 Lepton flavour universality tests 3. 3.3 Lepton flavour and lepton number violation 4. 3.4 Tests of low-energy QCD 5. 3.5 Test of first-row CKM unitarity 6. 3.6 Production of feebly-interacting particles in kaon decays 4. 4 Physics in the beam-dump mode 5. 5 High-intensity beams 1. 5.1 Beam delivery to the kaon production target 2. 5.2 Charged kaon beamline 3. 5.3 Neutral kaon beams 6. 6 Phase 1: a multi-purpose $K^{+}$ decay experiment 1. 6.1 Experimental layout 2. 6.2 Physics sensitivity 7. 7 Phase 2: a multi-purpose $K_{L}$ decay experiment 1. 7.1 Experimental layout 2. 7.2 Physics sensitivity 8. 8 Phase 3 (KLEVER): measurement of the $K_{L}\to\pi^{0}\nu\bar{\nu}$ decay 1. 8.1 Experimental layout 2. 8.2 Rates and timing performance 3. 8.3 Expected performance for $K_{L}\to\pi^{0}\nu\bar{\nu}$ 9. 9 Operation in beam-dump mode 1. 9.1 Experimental layout 2. 9.2 Physics sensitivity 10. 10 The HIKE detectors 1. 10.1 Detectors upstream of the fiducial decay volume 2. 10.2 Fiducial decay volume and its detectors 3. 10.3 Detectors downstream of the fiducial decay volume 4. 10.4 Detectors specific to KLEVER 11. 11 Data acquisition and high-level trigger 1. 11.1 Readout boards 2. 11.2 Streaming readout 3. 11.3 Triggered readout 4. 11.4 Event-building farm and event filter 12. 12 Online and offline computing 1. 12.1 Online monitoring 2. 12.2 Calibration and data quality 3. 12.3 Data processing model 4. 12.4 Distributed computing model 5. 12.5 Data reduction model 13. 13 Infrastructure and safety 14. 14 Summary and conclusions ## 1 Overview of HIKE Over the past 75 years, experimental studies of kaon decays have played a singular role in propelling the development of the Standard Model. As in other branches of flavour physics, the continuing experimental interest in the kaon sector derives from the possibility of conducting precision measurements, particularly of suppressed rare processes, that may reveal the effects of new physics with mass-scale sensitivity exceeding that which can be explored directly, e.g., at the LHC or next-generation colliders. Because of the relatively small number of kaon decay modes and the relatively simple final states, combined with the relative ease of producing intense kaon beams, kaon decay experiments are in many ways the quintessential intensity-frontier experiments. Over the past four decades, the CERN North Area has been the host to a successful series of precision kaon decay experiments. Among the many results obtained by these experiments is the discovery of direct CP violation, widely quoted to be among the major discoveries made at CERN. Continuation of high-intensity kaon experiments at CERN, including future upgrades (providing a unique probe into BSM physics complementary to the $B$-sector) has been identified as an essential scientific activity in the 2020 Update of the European Strategy for particle physics, and is strongly supported in the national roadmaps across Europe. The High Intensity Kaon Experiments (HIKE) project represents a broad, long- term programme at CERN after LS3, based in the North Area ECN3 experimental hall, covering all the main aspects of rare kaon decays and searches accessible via kaon physics, from ultra-rare kaon decays to precision measurements and searches for new phenomena. HIKE is intended to continue the very successful tradition of kaon experiments at CERN in ECN3, the latest of which is the currently operating NA62 experiment. HIKE will profit from a beam intensity increase by a factor between four and six with respect to NA62, and cutting-edge detector technologies. This will allow HIKE to play a pivotal role in the quest for New Physics (NP) at the sensitivity required by the present experimental limits and theoretical models, in a wide range of possible masses and interaction couplings. The programme will consist of multiple phases, first with charged and then neutral kaon beams, as well as periods in beam dump mode. The long decay volume and detector characteristics needed for kaon physics make HIKE suitable to search for new feebly- interacting, long-lived particles, providing unique sensitivity to forward processes. The detector design is challenging but at least one technological solution exists for each subsystem, thanks also to the synergy with HL-LHC. The various phases allow for insertion, repositioning or removal of specific detector elements depending on the physics requirements, while the overall setup remains broadly the same. ### 1.1 The scientific goals of HIKE The HIKE programme consists of several phases using shared detectors and infrastructure: a charged kaon phase between LS3 and LS4, and neutral kaon experiments commencing after LS4 including a phase with tracking, and a phase optimised specifically for the $K_{L}\to\pi^{0}\nu\bar{\nu}$ measurement (KLEVER). Phase 1: a multi-purpose $K^{+}$ decay experiment The physics of $K^{+}$ decays will be scrutinised with the highest precision, expanding and improving the physics programme of the NA62 experiment presently running in ECN3. New detectors will replace those of NA62 with the goal of improving the performance and sustaining higher rates. The principal goals of this phase are: * • Measurement of the $K^{+}\to\pi^{+}\nu\bar{\nu}$ branching ratio to 5% relative precision, matching the SM theoretical uncertainty. * • Precision measurements of $K^{+}\to\pi^{+}e^{+}e^{-}$ and $K^{+}\to\pi^{+}\mu^{+}\mu^{-}$ decays, and a precision lepton universality test. * • Searches for lepton flavour/number violating decays $K^{+}\to\pi(\pi^{0})\mu e$, $K^{+}\to\pi^{-}(\pi^{0})\ell^{+}\ell^{+}$, $K^{+}\to\ell_{1}^{-}\nu\ell_{2}^{+}\ell_{2}^{+}$ and $\pi^{0}\to\mu^{\pm}e^{\mp}$ at the ${\cal O}(10^{-12})$ sensitivity. * • Measurement of the quantity $R_{K}=\Gamma(K^{+}\to e^{+}\nu)/\Gamma(K^{+}\to\mu^{+}\nu)$ to 0.1% precision. * • Measurement of the ratios of the branching ratios of the main decay modes $K^{+}\to\pi^{0}\ell^{+}\nu$, $K^{+}\to\pi^{+}\pi^{0}$, $K^{+}\to\pi^{+}\pi^{+}\pi^{-}$ to a relative precision of a few per mille. * • Improvements of existing measurements of rare decays, including $K^{+}\to\pi^{+}\gamma\gamma$, $K^{+}\to\pi^{+}\gamma\ell^{+}\ell^{-}$, $K^{+}\to\pi^{+}\pi^{0}\gamma$, $K^{+}\to\pi^{+}\pi^{0}e^{+}e^{-}$. * • Searches for production of feebly-interacting particles in $K^{+}$ decays. * • Collection of a dataset in the beam-dump mode (with appropriate time sharing with the kaon mode), aiming for a factor 10 improvement in sensitivity to decays of feebly-interacting particles with respect to what NA62 can achieve by LS3. Phase 2: a multi-purpose $K_{L}$ decay experiment The physics of the $K_{L}$ decays will be addressed, including $K_{L}$ decays into charged particles in the final state. This mode of operation will require some modifications of the detector layout with respect to the charged mode. The main objectives of this phase are: * • Observation of the ultra-rare decays $K_{L}\to\pi^{0}e^{+}e^{-}$ and $K_{L}\to\pi^{0}\mu^{+}\mu^{-}$, or establishment of stringent upper limits on the branching ratios of these decays at the ${\cal O}(10^{-11})$ level. * • Measurement of the $K_{L}\to\mu^{+}\mu^{-}$ decay branching ratio to 1% relative precision. * • Searches for lepton flavour violating decays including $K_{L}\to(\pi^{0})(\pi^{0})\mu^{\pm}e^{\mp}$ and $K_{L}\to e^{\pm}e^{\pm}\mu^{\mp}\mu^{\mp}$ with ${\cal O}(10^{-12})$ sensitivity. * • Measurement of the ratios of the branching ratios of the main decay modes $K_{L}\to\pi^{\pm}\ell^{\mp}\nu$, $K_{L}\to\pi^{+}\pi^{-}(\pi^{0})$, $K_{L}\to\pi^{0}\pi^{0}(\pi^{0})$ to a relative precision of a few per mille. * • Collection of a further dataset in the beam-dump mode (with appropriate time sharing with the kaon mode), and characterisation of the neutral beam necessary to proceed to the third phase of HIKE. Phase 3 (KLEVER): measurement of the $K_{L}\to\pi^{0}\nu\bar{\nu}$ decay The experimental apparatus will be modified to specifically address the $K_{L}\to\pi^{0}\nu\bar{\nu}$ decay, and to measure its branching ratio to a 20% relative precision. To optimise the measurement, the tracking systems will be removed from the fiducial volume and additional photon veto detectors installed. Although this phase will focus on $K_{L}\to\pi^{0}\nu\bar{\nu}$, the setup will allow study of additional topics: * • Searches for additional flavour-changing neutral current decays, such as $K_{L}\to\pi^{0}\pi^{0}\nu\bar{\nu}$. * • Limits on the production of a long-lived, feebly interacting particle $X$ in $K_{L}\to\pi^{0}X$ decays, as a by-product of the measurement of $K_{L}\to\pi^{0}\nu\bar{\nu}$. * • Searches for $K_{L}$ decays to feebly-interacting particles $X$ such as $K_{L}\to XX$, with $X\to\gamma\gamma$. * • Searches for forbidden decays such as $K_{L}\to\pi^{0}\gamma$. ## 2 Scientific context The Standard Model (SM) of particle physics describes the experimental results obtained so far by experiments with exceptional precision. The culmination of this success was the discovery of the Higgs boson by the ATLAS and CMS collaborations at the LHC in 2012 [1, 2]. Thus the SM provides a solid foundation for the present understanding of elementary particles and their interactions. Despite this success, cosmological observations, experimental tensions, and theoretical motivations strongly suggest that the SM is an approximation of a more fundamental theory. One paradigm assumes that this theory lies above the electroweak (EW) mass scale and manifests itself in terms of new particles with masses well above the Higgs boson mass and having sizeable interactions with the SM fields. Another paradigm accounts for extensions of the SM that predict particles with masses below the EW scale, feebly interacting with the SM fields. Such a low-energy NP paradigm is usually referred to as a Feebly Interacting Particles (FIPs) scenario. Both high-scale BSM and FIPs models provide explanations for the main open questions in modern physics: the nature of dark matter, the baryon asymmetry of the universe, cosmological inflation, the origin of neutrino masses and oscillations, the strong CP problem, the hierarchical structure of the Yukawa couplings, the hierarchy problem, and the cosmological constant. HIKE intends to study NP in both high-scale BSM and FIPs directions. In this respect, HIKE is perfectly aligned with the recommendation of the 2020 Update of the European Strategy for Particle Physics [3], where a whole chapter is devoted to flavour physics research with the statement: “Experimental hints for deviations from SM predictions in flavour processes are one of our best hopes to direct research towards the right energy scale where NP can be found”. ### 2.1 Experimental context for high-scale BSM physics The experimental techniques to investigate the presence of BSM physics at high mass scales are direct searches for massive particles or processes forbidden by the SM, and precision measurements of observables precisely known within the SM. The first method is usually referred to as direct searches; the second as indirect searches. Direct searches allow unambiguous identification of NP already in single processes. After the observation and the study of the Higgs boson, direct searches for BSM particles were a primary goal of the LHC experiments ATLAS and CMS. Still, the LHC centre-of-mass energy limits the sensitivity of this technique to the TeV scale, or slightly above. Besides, a direct observation alone gives little information on the BSM structure. The absence of any significant direct observations of NP so far sets strong bounds on several types of BSM models up to the TeV scale. The bounds are still statistically limited in the TeV region, and the LHC high-luminosity programme will allow them to be extended to the kinematic limits, if no observations emerge in the meantime. Another type of direct searches focuses on the study of processes forbidden by the SM due to accidental SM symmetries like such as charged lepton flavour (CLF) or lepton number (LN) conservation. These searches typically involve the study of hadron or lepton decays forbidden by the SM, for example, $\tau\to\mu\mu\mu$ and $K_{L}\to\mu^{\pm}e^{\mp}$. Presently, these searches are statistically limited, and experiments at the high- intensity frontier are planned with increased sensitivity. Indirect searches exploit the possibility that NP particles affect low-energy observables virtually, via loop-level corrections to the SM prediction. The golden modes of indirect searches are theoretically clean observables of rare processes in the SM, i.e. occurring at loop-level at the lowest order, because NP here may favorably compete with SM. Indirect searches allow sensitivities to NP up to mass scales well above the TeV, but may require measurements of several observables to provide conclusive evidence of NP and to determine its nature. Given the limits to NP at TeV-scale already set by LHC through direct searches, indirect searches represent a promising tool to further boost BSM searches. Experimental flavour physics, measurements of Higgs couplings, and measurements in the gauge sectors of the SM ($W$, $Z$, top masses and gauge couplings) are typical examples of indirect searches. The existing measurements are broadly consistent with the SM and already suggest that NP may occur well beyond the TeV scale. Nevertheless, tensions between experiments and the SM exist mainly in flavour physics observables, and statistical limitations or the lack of measurements of rare processes demonstrate the need for additional efforts and motivate future directions in experimental particle physics. Presently, the most significant experimental tensions with the SM are: * • the violation of lepton flavour universality in $b\to s$ and $b\to c$ transitions reported by LHCb and $B$-factories [4, 5, 6, 7, 8, 9, 10], with a global significance of about 3–4$\sigma$ [11, 12, 13, 14]; * • the violation of CKM unitarity in the first row, known as the “Cabibbo angle anomaly”, with a significance of about 2–3$\sigma$ [15, 16, 17]; * • the inconsistency between the exclusive and inclusive measurements of the CKM matrix element $|V_{cb}|$ at the level above $3\sigma$ [18]; * • the inconsistency between the experiment and the SM prediction for the anomalous magnetic dipole moment of the muon at the level of 4.2$\sigma$ [19]. The sensitivity of the indirect searches to BSM physics scales with the size of the datasets, which is particularly important in case of rare processes. Moving to higher intensities is the way to increase statistics, however this requires improvements of the detector performance to cope with harsher experimental environments. Upgraded detectors also allow the improvement of the accuracy of already precise measurements, overcoming systematic uncertainties whenever they are the main limiting factor. The indirect search approach is at the heart of the LHC intensity frontier programme, the motivation for the upgrade of the LHCb experiment and for the continuation of the $B$-factory programme with the Belle II experiment. ### 2.2 Experimental context for feebly-interacting particles Direct searches for new low-mass particles can provide experimental insight into the models involving dark sectors and predicting the existence of FIPs [20, 21]. FIPs can be produced in meson decays or hadronic interactions of protons, and can decay to SM particles. They are produced in rare processes and are expected to have long lifetimes with respect to typical experimental scales as a result of their feeble couplings to the SM fields. Experiments may search for FIP production or decays. The first method focuses on the associated production of FIPs with SM particles in meson decays; see Ref. [22] for a review of possible production processes in kaon decays. The meson mass ($\pi$, $K$, $D$, $B$) determines the kinematic region that can be explored. Some examples are production of heavy neutral leptons in $K^{+}\to\ell^{+}N$, dark photons in $\pi^{0}\to\gamma X$, and dark scalars in $B\to KS$ or $K\to\pi S$. These kinds of searches can be pursued by flavour experiments, like LHCb and NA62, but also at ATLAS and CMS. The second method is based on FIP production, mainly in proton-nucleon interactions, and the detection of the subsequent FIP decays. Forward configurations and long experimental setups are best suited for this type of search, as those of fixed-target experiments, but also LHCb. The solid angle covered by the detector defines the kinematic range of the FIPs that can be studied. Examples of FIP decays include dark scalars or dark photon decays $S/A^{\prime}\to\mu^{+}\mu^{-}$ and heavy neutral lepton decays $N\to\pi^{\pm}\mu^{\mp}$. Searches for FIPs performed so far have not reported any hints of NP, and only limits in the plane of the coupling versus FIP mass have been set within the framework of various NP models. Nevertheless, it is worth continuing to explore the parameter space for FIP models. High beam intensities and experiments with appropriate geometries are necessary for the pursuit of these studies. ### 2.3 The role of HIKE The HIKE experiment is primarily a flavour physics experiment, with a fixed target configuration appropriate for the study of decays of kaons at high energies, on the order of 100 GeV. The study of the kaon sector makes HIKE complementary to LHCb and Belle II: kaons provide different, and in some cases higher, sensitivity to NP than $B$ and $D$ mesons. Measurements of kaon observables and correlations among them can help to find NP through the comparison with corresponding SM expectations and investigation of the present experimental inconsistencies with SM. More generally, the comparison between the flavour picture emerging from kaons with that from $B$ mesons is a powerful tool to investigate the indirect effects of NP: to shed light on existing experimental anomalies, push the sensitivity to mass scales beyond those attainable by $B$ and $D$ only, and provide insights to the flavour structure of possible NP models. Several models of BSM physics predict effects on kaon observables, ranging from supersymmetric to non-supersymmetric models, NP involving heavy $Z^{\prime}$ bosons, vector-like couplings, leptoquarks, and extra dimensions, in both minimally flavour violating (MFV) scenarios or with new sources of flavour violation. A common feature is that bounds on the parameter space of these models coming from direct searches at LHC and $B$ and $D$ physics have only a marginal impact on the possible effects on kaon observables, even more so if the models are non-MFV and non-supersymmetric. In contrast, the main constraints on the effects of NP on kaon observables come from the precisely measured parameters $\varepsilon_{K}$ and $\Delta M_{K}$. Table 1 gives a selected list of the most relevant NP models that may affect significantly kaon observables [23]. Table 1: Compendium of new-physics models relevant for kaon processes [23]. NP scenarios | Process | References ---|---|--- Z-FCNC | $K^{+}\rightarrow\pi^{+}\nu\bar{\nu}$, $K_{L}\to\pi^{0}\nu\bar{\nu}$, $\varepsilon^{\prime}/\varepsilon$ | [24, 25, 26, 27, 28] Z′ | $K^{+}\to\pi^{+}\nu\bar{\nu}$, $K_{L}\to\pi^{0}\nu\bar{\nu}$, $\varepsilon^{\prime}/\varepsilon$, $\Delta M_{K}$ | [29, 24, 25, 30, 31] Simplified models | $K_{L}\to\pi^{0}\nu\bar{\nu}$, $\varepsilon^{\prime}/\varepsilon$ | [32] LHT | All $K$ decays | [33, 34, 35] 331 models | Small effects in $K\to\pi\nu\bar{\nu}$ | [30] Vector-like quarks | $K^{+}\to\pi^{+}\nu\bar{\nu}$, $K_{L}\to\pi^{0}\nu\bar{\nu}$, $\Delta M_{K}$ | [36] Supersymmetry | $K^{+}\to\pi^{+}\nu\bar{\nu}$, $K_{L}\to\pi^{0}\nu\bar{\nu}$ | [37, 38, 39, 40, 41, 42, 43, 44, 45] 2HDM | $K^{+}\to\pi^{+}\nu\bar{\nu}$, $K_{L}\to\pi^{0}\nu\bar{\nu}$ | [46, 47] Universal extra dimensions | $K^{+}\to\pi^{+}\nu\bar{\nu}$, $K_{L}\to\pi^{0}\nu\bar{\nu}$ | [48, 49] Randall-Sundrum models | All rare $K$ decays | [50, 51, 52, 53, 54] Leptoquarks | All rare $K$ decays | [55, 56, 57] SMEFT | Several processes in $K$ system | [25, 58] SU(8) | $K^{+}\to\pi^{+}\nu\bar{\nu}$, $K_{L}\to\pi^{0}\nu\bar{\nu}$ | [59] Diquarks | $K^{+}\to\pi^{+}\nu\bar{\nu}$, $K_{L}\to\pi^{0}\nu\bar{\nu}$, $\varepsilon_{K}$ | [60, 61] Vector-like compositeness | $K^{+}\to\pi^{+}\nu\bar{\nu}$, $K_{L}\to\pi^{0}\nu\bar{\nu}$, $\varepsilon_{K}$ | [62] Presently, the main limitation to the investigation of the above models, and more generally to the quest for NP with kaons, comes from the experimental precision of the measurements of the kaon observables. To a lesser extent, the limitation is due to theoretical uncertainties and the precision of the measurements of the SM parameters, which will eventually be improved in the upcoming years. The primary goal of HIKE is to improve the accuracy of the measurements, wherever they exist, in order to match and possibly challenge the theory precision, to study and measure for the first time channels not yet observed, and to search with unprecedented sensitivity for kaon decays forbidden by the SM. A natural by-product of HIKE is the possibility to search for kaon decays to FIPs, a line of inquiry that is already being pursued with success by NA62. In addition, the fixed-target configuration and the high beam intensity requirement make HIKE suitable to search for decays of FIPs, exploring regions mainly below the $D$ mass, but with unprecedented sensitivity. ## 3 Physics with high-intensity kaon beams Kaons offer a unique opportunity to investigate BSM physics in a way complementary to $B$ or $D$ mesons, lepton physics, and EDM searches. The strategy is to perform multiple experimental studies ranging from measurements of branching ratios and form factors of kaon decays to searches for processes forbidden or not existing within the SM. A global fit to the experimental inputs provides the best sensitivity to NP, although single measurements also put significant constraints. Evidence for NP can also show up as inconsistencies between the observed flavour pattern and that expected from the SM, violation of accidental SM symmetries like lepton flavour universality (LFU), lepton flavour or number (LFV or LNV), and direct production of new low-mass particles in kaon decays. Because the BSM paradigm assumes NP at high masses, NP is expected to affect kaon processes at short-distance (SD) scales. The link between kaon-physics observables and BSM physics relies on the SM theoretical knowledge of these observables. The development of the theoretical understanding of the kaon sector is focused exclusively on calculations of observables relevant to pin down SD contributions from experimental results. Examples are: the SM calculation of the $K^{+}\to\pi^{+}\nu\bar{\nu}$ and $K_{L}\to\pi^{0}\nu\bar{\nu}$ decay branching ratios at NNLO leading to a few percent precision on these quantities; the recent calculation of the CP violation parameter $\varepsilon^{\prime}/\varepsilon$; computations of the branching ratios for the rare decays $K^{+}\to\pi^{+}\ell^{+}\ell^{-}$, $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$, $K_{L,S}\to\mu^{+}\mu^{-}$. It is convenient to group kaon decays into the following classes. * • Flavour-changing neutral currents (Section 3.1). The $K\to\pi\nu\bar{\nu}$, $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$, $K^{+}\to\pi^{+}\ell^{+}\ell^{-}$ and $K_{L}\to\mu^{+}\mu^{-}$ decays are suppressed in the SM, occurring at loop- level at the lowest order. The corresponding SD contributions are related to CKM matrix elements and are generally calculated with high precision within the SM. Together with $\varepsilon_{K}$, rare decays allow the unitary triangle to be over-constrained, independently from $B$ processes. With the exception of $K\to\pi\nu\bar{\nu}$ decays, the above decays receive important contributions from long distance (LD) physics. The theoretical framework used to study these processes involves low-energy effective theories, like Chiral Perturbation Theory (ChPT), which require experimental inputs (chiral parameters) to provide and validate the predictions, as well as lattice QCD. * • Lepton flavour universality tests (Section 3.2). The ratios ${\cal B}(K^{+}\to\pi^{+}e^{+}e^{-})/{\cal B}(K^{+}\to\pi^{+}\mu^{+}\mu^{-})$ and ${\cal B}(K^{+}\to e^{+}\nu)/{\cal B}(K^{+}\to\mu^{+}\nu)$ are extremely sensitive probes of lepton flavour universality. Another sensitive test is based on the ratio ${\cal B}(K\to\pi e\nu)/{\cal B}(K\to\pi\mu\nu)$, where both neutral and charged kaon decays ($K_{L}\to\pi^{\pm}\ell^{\mp}\nu$ and $K^{+}\to\pi^{0}\ell^{+}\nu$) can be used. * • Lepton flavour and number violating decays (Section 3.3). Decay modes forbidden by the SM by either lepton flavour or number conservation include $K^{+}\to\pi^{-}(\pi^{0})\ell_{1}^{+}\ell_{2}^{+}$, $K^{+}\to\pi^{+}\mu^{\pm}e^{\mp}$, $K^{+}\to\ell_{1}^{-}\nu\ell_{2}^{+}\ell_{2}^{+}$, $K_{L}\to(\pi^{0})(\pi^{0})\mu^{\pm}e^{\mp}$ and $\pi^{0}\to\mu^{\pm}e^{\mp}$. Any experimental signal of these processes would constitute direct evidence for NP. It should be noted that kaon beams represent sources of tagged, Lorentz-boosted $\pi^{0}$ mesons. * • Tests of low-energy QCD (Section 3.4). The abundant kaon decays $K\to\pi\pi$ and $K\to\pi\pi\pi$ occur at the tree level and are dominated by LD physics. The decays are described by ChPT which is the low-energy effective field theory of QCD. Global fits to the measured branching ratios and kaon lifetimes of the $K\to\pi\pi$ and $K\to\pi\pi\pi$ decays play a crucial role in the extraction of the ChPT low-energy constants. Rare and radiative decays including $K\to\pi\gamma\gamma$, $K\to\ell\nu\gamma$, $K\to\pi\ell\nu\gamma$, $K\to\pi\pi\gamma$, $K\to\pi\pi\pi\gamma$, $K\to\pi\pi e^{+}e^{-}$, and $K\to\pi\pi\ell\nu$ are also dominated by LD physics and similarly provide access to the ChPT parameters. * • CKM first-row unitarity tests (Section 3.5). Branching ratios and form factors of the tree-level $K^{+}\to(\pi^{0})\ell^{+}\nu$ and $K_{L}\to\pi^{\pm}\ell^{\mp}\nu$ decays represent the principal experimental inputs for the determination of CKM matrix element $V_{us}$, with input from lattice QCD. * • Production of feebly-interacting particles in kaon decays (Section 3.6) is predicted by FIP models involving new particles below the kaon mass scale. Examples include production of a heavy neutral lepton ($K^{+}\to\ell^{+}N$), or a dark scalar ($K^{+}\to\pi^{+}S$). HIKE intends to address experimentally all these processes at a new level of sensitivity, improving the existing experimental picture to fully exploit the complementarity of kaon with respect to heavy-quark physics in the quest for NP. ### 3.1 Flavour-changing neutral currents Kaon decays proceeding at loop level in the SM are denoted “rare”, and typically have branching ratios below $10^{-7}$. They provide a tool independent from $B$ and $D$ physics to test the flavour structure of the SM by constraining the unitary triangle via loop-level observables. The link between rare kaon decays and flavour physics in the $\rho-\eta$ plane is schematically shown in Fig. 1. CP-violating short distance (SD) physics contributes to the $K_{L}\to\pi^{0}\nu\bar{\nu}$, $K_{L}\to\pi^{0}e^{+}e^{-}$, $K_{L}\to\pi^{0}\mu^{+}\mu^{-}$, and $K_{S}\to\mu^{+}\mu^{-}$ decays. The amplitudes of these decays thus depend on the height, $\eta$, of the unitary triangle. The SD physics contributing to the decay $K_{L}\to\mu^{+}\mu^{-}$ is CP-conserving, and its amplitude depends on the base, $\rho$, of the unitary triangle. The amplitude of the rare decay $K^{+}\to\pi^{+}\nu\bar{\nu}$ contains terms depending on both $\eta$ and $\rho$. Figure 1: Relation between kaon rare decay modes and the parameters $\rho$ and $\eta$ of the unitary triangle (UT). The direct link between decay modes and the UT indicates short distance terms dependent on $\rho$ or $\eta$ contributing to the corresponding decay amplitudes. Decays not directly connected to the UT are relevant to interpret the experimental results of the decay modes to which they are related. The $K_{L}\to\pi^{0}e^{+}e^{-}$, $K_{L}\to\pi^{0}\mu^{+}\mu^{-}$ and $K_{L}\to\mu^{+}\mu^{-}$ decays are LD dominated. The extraction of their SD components from experimental data proceeds through the study of ancillary decay modes, listed in Fig. 1 in association with the rare decays mentioned above. In contrast, LD contributions are sub-dominant to the amplitudes of the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay, and negligible for the $K_{L}\to\pi^{0}\nu\bar{\nu}$ decay. As a matter of fact, these two decay modes belong to the theoretically cleanest probes of the flavour structure of the SM among all the kaon and $B$ meson decays. #### 3.1.1 $K\to\pi\nu\bar{\nu}$ decays The $K\to\pi\nu\bar{\nu}$ decays involve $s\to d$ quark transitions. They proceed through box and penguin diagrams, which are SD dominated because of the leading contribution of the virtual $t$ quark exchange. The quadratic GIM- mechanism and the $t\to d$ Cabibbo suppression make place these processes among the rarest meson decays in the SM. Their decay amplitudes can be parameterised in terms of the precisely measured $K^{+}\to\pi^{0}e^{+}\nu$ decay, allowing theoretical computation free from hadronic uncertainties. The SM predictions of the $K^{+}\to\pi^{+}\nu\bar{\nu}$ and $K_{L}\to\pi^{0}\nu\bar{\nu}$ branching ratios can be written in terms of CKM parameters as [63] $\displaystyle{\cal B}(K^{+}\to\pi^{+}\nu\bar{\nu})$ $\displaystyle=(8.39\pm 0.30)\times 10^{-11}\left(\frac{|V_{cb}|}{0.0407}\right)^{2.8}\left(\frac{\gamma}{73.2^{\circ}}\right)^{0.74},$ (1) $\displaystyle{\cal B}(K_{L}\to\pi^{0}\nu\bar{\nu})$ $\displaystyle=(3.36\pm 0.05)\times 10^{-11}\left(\frac{|V_{ub}|}{3.88\times 10^{-3}}\right)^{2}\left(\frac{|V_{cb}|}{0.0407}\right)^{2}\left(\frac{\sin{\gamma}}{\sin{73.2^{\circ}}}\right)^{2}.$ (2) Here $V_{cb}$ and $V_{ub}$ are elements of the CKM matrix, and $\gamma$ is the angle of the unitary triangle defined as $\arg[(-V_{ud}V_{ub}^{*})/(V_{cd}V_{cb}^{*})]$. The theoretical uncertainty depends on the NNLO approximation of the computation of the Feynmann diagrams, on the radiative corrections, and on the estimates of the long distance contributions due to the exchange of the $u,d,c$ quarks. The prediction for the neutral mode prediction is more precise than that for the charged mode because the neutral amplitude is purely imaginary, allowing cancellation of the long distance corrections. In either case, the main contribution to the uncertainties for the branching ratio calculation comes from the precision of the measurements of the CKM matrix elements, known as the parametric uncertainty, and could be as large as 9% if the current measurements of $V_{cb}$, $V_{ub}$ and $\gamma$ from the CKM fit are considered. A suitable combination of external parameters nearly independent of new physics contributions allows the reduction of the parametric uncertainty, reducing the overall relative uncertainties in the branching ratios to 5% [64]. Other recent calculations reach a similar quantitative conclusion [65]. A global fit to the CKM unitary triangle, instead, allows the parametric uncertainty to be factored out, and offers a powerful tool to compare the SM flavour structure arising from kaon and $B$ physics. Equations 1 and 2 also indicate that precise measurements of $K\to\pi\nu\bar{\nu}$ decays would shed light on CKM parameters historically measured with $B$ mesons and suffering from long- standing experimental tensions like $|V_{cb}|$. The extreme SM suppression makes these decays particularly sensitive to NP. From a model-independent point of view, the $K\to\pi\nu\bar{\nu}$ decays probe NP at the highest mass scales [63], of the order of hundreds of TeV. Existing experimental constraints on NP affect the $K\to\pi\nu\bar{\nu}$ branching ratio weakly. Model-dependent scenarios predict sizeable deviations of the branching ratios from the SM, as well as correlations between the branching ratios for the charged and neutral modes, depending on the model [46, 55, 36, 28, 45, 44, 35, 66] (see also Table 1). The NA62 experiment at CERN has observed $20$ candidate events for the decay $K^{+}\to\pi^{+}\nu\bar{\nu}$ with 7 expected background events and 10 expected SM signal events [67], leading to the measurement $\mathcal{B}(K^{+}\rightarrow\pi^{+}\nu\bar{\nu})=(10.6^{+4.0}_{-3.4}|_{\text{stat}}\pm 0.9_{\text{syst}})\times 10^{-11}$ at $68\%$ confidence level. This represents the most precise measurement to date of this process, providing first evidence for its existence and falsifying the background-only hypothesis with $3.4\sigma$ significance. The experiment is currently taking data in Run 2 (2021–LS3) with the aim of reaching a ${\cal O}(10\%)$ uncertainty measurement, and has demonstrated its ability to sustain nominal beam intensity. The NA62 experiment has therefore shown that the decay-in-flight technique works well and is scalable to larger data samples. Beyond the $\mathcal{B}(K^{+}\to\pi^{+}\nu\bar{\nu})$ measurement, it is important to establish if the decay has a purely vector nature as expected within the SM. In BSM scenarios, an additional scalar contribution to the decay is predicted [68, 69], leading to possibly LNV/LFV $K^{+}\to\pi^{+}\nu\nu$ contributions, and since the neutrinos are not detected, a $\nu\bar{\nu}$ pair cannot be directly distinguished from a $\nu\nu$ pair, and the measurement of the branching ratio alone is insufficient to understand the nature of the process. Because of the suppression from the small neutrino mass, there is a negligible interference between vector and scalar contributions, and the experimentally measured branching ratio is given by [68, 23]: $\mathcal{B}(K^{+}\rightarrow\pi^{+}\nu\overset{\textbf{(---)}}{\nu})=\mathcal{B}(K^{+}\rightarrow\pi^{+}\nu\bar{\nu})_{\rm SM}+\sum_{i\leq j}^{3}\mathcal{B}(K^{+}\rightarrow\pi^{+}\nu_{i}\nu_{j})_{\rm LNV}.$ (3) To fully test the SM prediction of a purely vector, lepton-number-conserving $K^{+}\rightarrow\pi^{+}\nu\bar{\nu}$ decay, the momentum-transfer spectrum must be studied to rule out the additional scalar BSM $K^{+}\rightarrow\pi^{+}\nu\nu$ contribution to the overall experimentally measured rate. Such a measurement requires a much larger dataset, as offered within the HIKE programme. The current upper limit on the branching ratio of the $K_{L}\to\pi^{0}\nu\bar{\nu}$ decay is $3\times 10^{-9}$ at 90% CL, set by the KOTO experiment at J-PARC [70]. The KOTO experiment is currently taking data with the goal of reaching ${\cal O}(10^{-11})$ sensitivity in the next 5 years. In the longer term, an ambitious upgrade, KOTO Step-2, is planned to begin construction after 2025 in a proposed extension of the Hadron Experimental Facility, with the goal of measuring the branching ratio to within $\sim$20% in about three years of data taking [71]. #### 3.1.2 $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ decays The ultra-rare $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ decays represent another set of theoretically clean “golden modes” in kaon physics, second only to $K\to\pi\nu\bar{\nu}$ in terms of their importance, and allowing for the direct exploration of new physics contributions in $s\to d\ell\ell$ transitions (to be compared to $b\to s\ell\ell$ transitions). In the SM, they are dominated by the indirect (mixing-induced) CP violation and its interference with the direct CP-violating contribution. The decay rates can be enhanced significantly in the presence of large new CP-violating phases, in a correlated way with the effects in $K_{L}\to\pi^{0}\nu\bar{\nu}$ and $\varepsilon^{\prime}/\varepsilon$ [23]. The SM expectation for the decay branching ratios is [72, 73, 74] $\displaystyle{\cal B}(K_{L}\to\pi^{0}e^{+}e^{-})$ $\displaystyle=$ $\displaystyle 3.54^{+0.98}_{-0.85}~{}\left(1.56^{+0.62}_{-0.49}\right)\times 10^{-11},$ $\displaystyle{\cal B}(K_{L}\to\pi^{0}\mu^{+}\mu^{-})$ $\displaystyle=$ $\displaystyle 1.41^{+0.28}_{-0.26}~{}\left(0.95^{+0.22}_{-0.21}\right)\times 10^{-11},$ with the two sets of values corresponding to constructive (destructive) interference between direct and indirect CP-violating contributions. Experimentally, the most stringent upper limits (at 90% CL) of the decay branching ratios have been obtained by the KTeV experiment [75, 76]: ${\cal B}(K_{L}\to\pi^{0}e^{+}e^{-})<28\times 10^{-11},\quad{\cal B}(K_{L}\to\pi^{0}\mu^{+}\mu^{-})<38\times 10^{-11}.$ The irreducible $K^{+}\to\gamma\gamma\ell^{+}\ell^{-}$ background, whose rate vastly exceeds that of the signal decays $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$, represents a significant problem in achieving experimental sensitivity at the SM level, especially for $K_{L}\to\pi^{0}e^{+}e^{-}$ [77]. #### 3.1.3 $K^{+}\to\pi^{+}\ell^{+}\ell^{-}$ decays The amplitudes of the $K^{+}\to\pi^{+}\ell^{+}\ell^{-}$ decays ($\ell=e,\mu$) are LD-dominated and their branching ratios are $\mathcal{O}(10^{-7})$. The differential decay amplitude of these decays with respect to the di-lepton invariant mass depends on two form factor parameters, denoted $a_{+}$ and $b_{+}$, that lepton universality predicts to be independent from the flavour of the leptons. Differences between $a_{+}$ and $b_{+}$ between the electron and muon channels can be correlated to LUV effects that can explain the possible observed lepton anomalies in $B$ physics [78, 79]. The NA62 experiment has recently reported a new measurement of the form factors of the $K^{+}\to\pi^{+}\mu^{+}\mu^{-}$ decay using data collected in 2017–2018 [80]. The excellent kinematic resolution allowed a nearly background-free selection of $3\times 10^{4}$ decays. The decay $K^{+}\to\pi^{+}\pi^{+}\pi^{-}$ was used for normalisation, and the parameter values $a_{+}=-0.575\pm 0.013$ and $b_{+}=-0.722\pm 0.043$ were extracted from a fit to the di-lepton mass spectrum. This improves significantly on the previous measurements in the muon channel while also consistent within uncertainty with measurements of the same parameters performed both in the muon and electron channels [81, 82, 83, 84]. The NA62 experiment is expected to improve the precision of the LFU test in the next years, addressing also the measurement of the $K^{+}\to\pi^{+}e^{+}e^{-}$ form factors. However a significant increase in precision is required to perform a LFU test with comparable sensitivity to NP models as in current results from $B$ physics [78, 85]. HIKE will offer a significantly larger data sample, further reducing statistical uncertainties, which remain a large fraction of the total uncertainty in the most recent NA62 measurement [80]. #### 3.1.4 $K_{L}\to\mu^{+}\mu^{-}$ decays The $K_{L}\to\mu^{+}\mu^{-}$ is a rare and helicity suppressed decay. Its branching ratio is driven by LD contributions, with a sub-dominant CP- conserving SD component [86]. The SM prediction exhibits large uncertainties due to the sign ambiguity of the interference between the LD and SD terms contributing to the decay amplitude. The sign ambiguity leads to two SM predictions: $\text{[LD$+$]:}~{}~{}\left(6.82^{+0.77}_{-0.24}\pm 0.04\right)\times 10^{-9},~{}~{}~{}\text{[LD$-$]:}~{}~{}\left(8.04^{+1.66}_{-0.97}\pm 0.04\right)\times 10^{-9}.$ (4) The first uncertainty comes from the calculation of the LD terms, dominates the overall precision, and makes the uncertainty quite asymmetrical. This asymmetry is reflected in the computation of the $K_{L}\to\mu^{+}\mu^{-}$ rate with the inclusion of NP. This decay has been studied experimentally and the measured branching ratio is $(6.84\pm 0.11)\times 10^{-9}$ [87]. The sign ambiguity and the LD uncertainties prevent a clean theoretical interpretation of this result. Theoretical efforts are foreseen in the next few years to resolve the sign ambiguity through an appropriate matching between LD and SD contributions and to reduce the LD uncertainty by at least a factor of two. This could allow a sizeable reduction of the uncertainty of the SM prediction, opening the possibility to exploit the high sensitivity of this decay to BSM physics models. The $K_{S}\to\mu^{+}\mu^{-}$ decay provides a complementary, theoretically clean observable [88, 89]. While the total $K\to\mu^{+}\mu^{-}$ rate is dominated by long-distance (LD) physics, the measurement (in a dedicated high- intensity $K_{S}$ experiment) of CP violation in the interference of mixing and decay via the time-dependent rate enables the extraction of the purely CP odd short-distance amplitude, which is predicted within the SM with an ${\cal O}(1\%)$ uncertainty. The study of the $K_{S}\to\mu^{+}\mu^{-}$ decay is the main goal of the kaon physics programme of the LHCb experiment. The high production rate of kaons in proton-proton collisions at 13 TeV partly compensates the small LHCb acceptance for kaon decays due to the long lifetime. LHCb has recently reported an upper limit on the $K_{S}\to\mu^{+}\mu^{-}$ branching fraction using the combined Run 1+2 dataset, ${\cal B}(K_{S}\to\mu^{+}\mu^{-})<2.1\times 10^{-10}$ 90% CL [90], to be compared to the SM prediction, ${\cal B}_{\rm SM}(K_{S}\to\mu^{+}\mu^{-})=(5.18\pm 1.50)\times 10^{-12}$ [91, 88]. The ultimate expected LHCb sensitivity for this decay is close to the SM branching ratio and is expected to be limited by the statistical uncertainty on the background subtraction and the signal yield [92]. #### 3.1.5 Global sensitivity to New Physics The rare decays discussed above offer the possibility to look for NP using a global fit technique. As an example, Fig. 2 shows the result of this technique using a low energy effective theory approach [85]. In this approximation, the physics is described in terms of an effective Hamiltonian, which is an expansion series of fermion current operators. The coefficient of this expansion is the Wilson coefficient $C_{k}$, where $k$ is an index running over the expansion terms. The assumption is that NP modifies $C_{k}$ with respect to the SM values. The Wilson coefficients are then fitted using experimental values of observables from rare kaon decays as input variables. The results shown in Fig. 2 assume NP scenarios compatible with the LFUV anomalies observed in $B$ decays. Figure 2: Global fit to current data (purple, labelled as “current fit” in the legend) and the results of two projections discussed in Ref. [85]: A, light- red, where central values of observables currently only constrained with upper bounds are set to the SM predictions; B, dark-red, where central values for all of the observables are best-fit projections from existing data. The SM and best-fit point with the current data are indicated with black and purple crosses, respectively. The two figures represent the two possible signs of the long-distance contribution to the $K_{L}\to\mu^{+}\mu^{-}$ process; the sign ambiguity may be resolved with future data. ### 3.2 Lepton flavour universality tests Lepton flavour universality (LFU) is a cornerstone of the SM postulating that the lepton coupling to gauge bosons is independent of lepton type (“flavour”), in contrast with the flavour-dependence of quark interactions. The origin of the observed LFU, which is not associated with any known symmetry and is thus not a fundamental conservation law, is a major question in modern physics. Recently observed tensions between the SM predictions and experimental results such as the measurement of the anomalous magnetic moment of the muon [93], the Cabibbo angle anomaly [94], and the $B$-meson anomalies [8, 95, 6], suggest possible LFU violation and motivate further the quest for improved experimental LFU measurements. The kaon sector offers opportunities for precision lepton flavour universality tests by measuring the ratio ${\cal B}(K^{+}\to\pi^{+}e^{+}e^{-})/{\cal B}(K^{+}\to\pi^{+}\mu^{+}\mu^{-})$ as discussed in the framework of a global fit in a low-energy effective theory approach in Sections 3.1.3 and 3.1.5, as well as ${\cal B}(K^{+}\to e^{+}\nu)/{\cal B}(K^{+}\to\mu^{+}\nu)$ and ${\cal B}(K\to\pi e\nu)/{\cal B}(K\to\pi\mu\nu)$, exploiting cancellation of hadronic effects [17, 78]. The ratio of purely leptonic decay rates of the charged kaon $R_{K}=\Gamma(K^{+}\to e^{+}\nu)/\Gamma(K^{+}\to\mu^{+}\nu)$ is strongly suppressed in the SM by conservation of angular momentum and an extremely sensitive probe of LFU. Its SM expectation $R^{\rm SM}_{K}=(2.477\pm 0.001)\times 10^{-5}$ [96] is known to excellent (0.4‰) precision and the currently most precise (4‰) experimental result is $R_{K}=(2.488\pm 0.007_{\rm stat}\pm 0.007_{\rm syst})\times 10^{-5}$ [97]. The ratio $R_{K}$ is highly sensitive to the possible violation of LFU naturally arising in new physics scenarios involving sterile neutrinos [98, 99], leptoquarks [100], massive gauge bosons [101], or an extended Higgs sector [102, 103]. Variations of $R_{K}$ up to a few per mille from its SM expectation are predicted by these models, without contradicting any present experimental constraints. An analogous LFU test is based on the ratio of semileptonic kaon decay rates $R_{K}^{(\pi)}=\Gamma(K\to\pi e\nu)/\Gamma(K\to\pi\mu\nu)$, where both neutral and charged kaon decays ($K_{L}\to\pi^{\pm}\ell^{\mp}\nu$ and $K^{+}\to\pi^{0}\ell^{+}\nu$) can be used. For a given neutral or charged initial state kaon, the Fermi constant, $V_{us}$, short-distance radiative corrections, and the hadronic form factor at zero momentum transfer cancel out when taking the ratio $R_{K}^{(\pi)}$ [17]. Therefore, in the SM this ratio is entirely determined by phase space factors and long-distance radiative corrections [104, 105, 106, 107, 16]. ### 3.3 Lepton flavour and lepton number violation In the SM, lepton number and lepton flavour are conserved due to accidental global symmetries. Additionally, the SM requires neutrinos to be strictly massless due to the absence of right-handed chiral states. However the discovery of neutrino oscillations has demonstrated non-zero neutrino mass and non-conservation of lepton flavour in the neutrino sector. On the other hand, no evidence has yet been found for lepton flavour violation (LFV) in the charged lepton sector, or lepton number violating (LNV) processes. Any such observation would be a clear indication of NP. Mixing of active neutrinos can, in principle, mediate LFV decays in the SM. However the branching ratios are vanishingly small, and therefore any observation of LFV kaon decays would be clear evidence of new physics. Such LFV kaon decays are predicted in BSM scenarios with light pseudoscalar bosons (ALPs) [108], a $Z^{\prime}$ boson [109, 110] or leptoquarks [111, 112]. As discussed in Section 2.1, several hints of violations of lepton flavour universality (LFUV) have been reported. Models which predict LFUV and explain these potential anomalies in general naturally predict LFV processes [113, 114, 115]. Therefore search for LFV processes are of strong interest. In BSM scenarios, LNV kaon decays are predicted for example in models with Majorana neutrinos. In the minimal Type-I seesaw model [116, 117, 118], the neutrino is a Majorana fermion with a mass term that violates lepton number by two units. Non-zero neutrino masses make it possible, in principle, to distinguish experimentally between the possible Dirac and Majorana natures of the neutrino. Strong evidence for the Majorana nature of the neutrino would be provided by the observation of lepton number violating (LNV) processes, including kaon decays. The recent results of LFV/LNV searches with the NA62 Run 1 (2016–2018) dataset [119, 120, 121] include upper limits on the $K^{+}\to\pi^{-}\ell^{+}_{1}\ell^{+}_{2}$ decay branching ratios at the ${\cal O}(10^{-11})$ level, already leading to stringent constraints on active- sterile mixing angles between Majorana neutrinos. Below the kaon mass, these constraints are competitive with those obtained from neutrinoless double beta decay [122, 123, 124, 125]. The NA62 searches are not limited by backgrounds, and indicate that HIKE Phase 1 would make significant improvements in sensitivity to LFV and LNV $K^{+}$ decays, extending the limits to the ${\cal O}(10^{-12})$ level. On the other hand, the existing limit on the LFV $K_{L}\to\mu^{\pm}e^{\mp}$ decay [126], which represents a stringent constraint on several BSM scenarios, would be improved significantly by the HIKE Phase 2; further details are provided in Section 7.2. ### 3.4 Tests of low-energy QCD Most kaon decays are governed by physics at long distances. These decays can be described by chiral perturbation theory (ChPT), which is the low-energy effective field theory of QCD. The ChPT framework determines the kaon decay amplitudes in terms of the so-called low-energy constants, which are determined using experimental data. Therefore, measurements of kaon decay rates and form factor parameters of various decay channels represent essential tests of the ChPT predictions and crucial inputs to the theory at the same time. A comprehensive overview of kaon decays and their relation to the ChPT can be found in Ref. [127]. The HIKE dataset will provide a unique opportunity to perform precision measurements of rare and radiative decays including $K\to\pi\gamma\gamma$, $K\to\ell\nu\gamma$, $K\to\pi\ell\nu\gamma$, $K\to\pi\pi\gamma$, $K\to\pi\pi\pi\gamma$, $K\to\pi\pi e^{+}e^{-}$, and $K\to\pi\pi\ell\nu$ of both $K^{+}$ and $K_{L}$ mesons. Possible studies of the abundant decay modes $K\to\pi\pi$ and $K\to\pi\pi\pi$ will also provide important inputs to future ChPT parameter fits. ### 3.5 Test of first-row CKM unitarity Measurements of semileptonic kaon decays $K\to\pi\ell\nu$ provide the principal input for the extraction of the CKM parameter $V_{us}$, while the ratios of (semi)leptonic $K^{+}$ and $\pi^{+}$ decay rates are used to extract the ratio $V_{us}/V_{ud}$, with inputs provided from lattice QCD [128]. Determination of $V_{us}$ from kaon, pion, and $\tau$ decays, combined with $V_{ud}$ measurement from super-allowed beta decays [129] and neutron decays [130, 131], gives rise to a $3\sigma$ deficit in first-row CKM unitarity relation known as the Cabibbo angle anomaly; a tension of similar significance is observed between $K\to\ell\nu$ and $K\to\pi\ell\nu$ rates [15, 17]. The uncertainty in $V_{us}$ comes in equal parts from the experimental errors and theoretical uncertainties in the ratio of decay constants, $f_{K}/f_{\pi}$, and the $K\to\pi\ell\nu$ form factor, $f_{+}(0)$. Substantial improvements in the lattice QCD calculations of these hadronic factors are expected in the next five years, thanks to decreased lattice spacing and accurate evaluation of electromagnetic effects [132, 133, 134, 135]. Significant progress on the calculation of radiative corrections has been achieved recently, reducing the uncertainties on the long-distance electromagnetic corrections to a negligible level [136, 107, 137, 16]. In order to shed light on the Cabibbo angle anomaly, improved measurements of the principal $K^{+}$ and $K_{L,S}$ branching ratios are essential (and it should be noted that no $K_{L}$ decay measurements have been made in the past decade). In particular a precision measurement of the ratio of $K^{+}\to\pi^{0}\mu^{+}\nu$ and $K^{+}\to\mu^{+}\nu$ rates is well-motivated [15]. In case the Cabibbo angle anomaly persists, HIKE would be able to make precision measurements of (semi)leptonic $K^{+}$ and $K_{L}$ decays based on special datasets collected with minimum-bias triggers at low beam intensity. It should be noted that the Belle II experiment is planning to extract $V_{us}$ at an improved precision from a suite of inclusive and exclusive measurements of $\tau$ decays, including ${\cal B}(\tau^{-}\to K^{-}\nu)/{\cal B}(\tau^{-}\to\pi^{-}\nu)$, combined with theory improvements [138]. This would provide a possibility to cross-check the kaon results. ### 3.6 Production of feebly-interacting particles in kaon decays An alternative paradigm to the searches for NP at high energy is the search for feebly-interacting particles (FIPs), which would be the manifestations of weakly-interacting NP potentially at much lower energy scales. Interactions of FIPs with SM fields are possible through BSM extensions involving low- dimension operators, the so-called portals [20, 21], the simplest of which are summarised in Table 2. For further discussion of the benchmark scenarios, see also Section 9.2. Table 2: Summary of generic portal models with FIPs [20, 21]. Portal | FIPs ---|--- Vector | Dark photon, $A^{\prime}$ Scalar | Dark Higgs/scalar, $S$ Fermion | Heavy neutral lepton (HNL), $N$ Pseudoscalar | Axion/axion-like particle (ALP), $a$ Due to the availability of large datasets and the suppression of the total kaon decay width, kaon decays represent uniquely sensitive probes of light hidden sectors via FIP production searches. The possible search strategies have been reviewed recently [22], and the following have been identified as the most promising. Searches for the $K\to\pi X_{\rm inv}$ decay by extension of the $K\to\pi\nu\bar{\nu}$ measurements, where $X_{\rm inv}$ represents an invisible particle, represent a unique probe into the dark-scalar and ALP parameter space. Searches for heavy neutral lepton ($N$) production in $K^{+}\to\ell^{+}N$ decays are approaching the seesaw neutrino mass models with ${\cal O}(100~{}{\rm MeV})$ sterile neutrinos [139]. Searches for resonances in the $K\to\pi\ell^{+}\ell^{-}$ and $K\to\pi\gamma\gamma$ decay spectra are complementary to searches at beam-dump experiments for a significant ALP mass range. Finally, searches for a leptonic force mediator ($X$) in $K^{+}\to\mu^{+}\nu X$ decays can probe a region of parameter space providing an explanation for the muon $g-2$ anomaly [140]. Experimentally, the NA62 Run 1 dataset has been used to establish the analysis methods for searches for dark-scalar and HNL production in $K^{+}$ decays, producing world-leading limits on the dark-scalar coupling [67, 141, 142] and HNL mixing parameters [143, 144] below the kaon mass. Further details on the experimental aspects and HIKE sensitivity projections are provided in Sections 6.2.3 and 6.2.4. ## 4 Physics in the beam-dump mode Thanks to the high-intensity beam and high-performance detectors (redundant particle identification capability, highly efficient veto systems and high resolution measurements of momentum, time, and energy), HIKE can achieve the sensitivities required for competitive searches for production and decay of long-lived light mediators in a variety of new-physics scenarios. A number of possible Standard Model (SM) extensions aimed at explaining the abundance of dark matter in our universe predict a new “hidden sector” with mediator fields or matter fields in the MeV–GeV mass range. The realisations of such scenarios are usually classified in terms of the Lagrangian operators (the so-called “portals” listed in Table 2) connecting SM particles to the new mediators. The related coupling constants are small, thus evading the existing exclusion bounds and justifying the label of “feebly interacting” given to these new-physics models [20, 21]. The mass of the mediator and hidden-sector matter fields and the couplings are free parameters of the models. The relevant features of the phenomenology of such models are: * • The mediators can be produced in proton-nucleus interactions through a number of mechanisms, the most important being direct Primakoff production and meson- mediated tertiary production. The specific production mechanism for each mediator type differs in terms of production cross section and momentum-angle spectra of the mediator emitted. * • At the 400-GeV energy of SPS protons, mediators with momenta above 10 GeV have decay lengths ranging from tens of meters to tens of kilometres in a suitable, interesting interval of the feeble coupling. * • Due to the feeble interaction with the SM particles, the emitted mediators can reach the decay volume after punching through tens of meters of traversed material before decaying. As demonstrated by the NA62 Run 2 operation [145], the intense P42 400 GeV proton beam extracted from the CERN SPS and the HIKE setup can be exploited to search for production and decay of the emitted mediators. For this purpose, HIKE can be operated in the so-called beam-dump mode. Considering that HIKE will have the same type of forward geometry as NA62, the HIKE beam-dump operation will be particularly sensitive to specific physics cases in which the hidden sector mediator is produced by Primakoff scattering or light-meson mediated decays. Prominent examples are the vector and some axion-like particle scenarios. We address a number of new physics scenarios in Section 9, tentatively assuming a total integrated statistics of $10^{19}$ POT for HIKE Phase 1, and $5\times 10^{19}$ POT for the complete HIKE programme. ## 5 High-intensity beams As the acronym HIKE suggests, the beam line must provide significantly higher beam intensities than those that are now delivered to the NA62 experiment, on the order of 4 to 6 times of that currently provided, which would correspond to between 1.2 and $2\times 10^{13}$ protons on the T10 target per SPS extraction. Here, we assume that a typical running year has 200 days with 3000 spills each, including an SPS uptime of about 80%, and a flat top length of 4.8 seconds. The proposed HIKE program includes both a charged beam ($K^{+}$) phase and a neutral beam ($K_{L}$) phase. For both parts of the program, protons at much higher intensity have to be transported via the P42 beam line from the T4 target to the kaon production target T10. Conceptually, the $K^{+}$ beam is proposed to be the same as the present beam line, but consolidation and upgrades are needed to handle the higher beam intensities. Improved beam instrumentation will keep the beam losses better under control. A more precise calibration of the primary beam intensity measurements by the secondary emission monitors at T4 and T10 is important for beam optimisation, and for normalisation purposes in dump mode. Efforts in this respect have already been started by the BI and EA groups at CERN. The neutral beamline is an entirely new design and requires minor local modifications to the last section of the P42 line. ### 5.1 Beam delivery to the kaon production target The P42 beam line started operation in 1980 and hardly has been modified since then [146, Chapter 5]. Most of the equipment dates from then and is showing strong effects of ageing. Many issues and suggestions for mitigation have been described in the report from the PBC Conventional Beams working group [147], which has recently been updated. The first phase of a vast consolidation program for the CERN North Area, titled NACONS, has recently been approved and funded, and the second phase has been prepared [148, 149, 150]. This will allow restoration of the equipment and beam infrastructure to a reliable state. Some upgrades are required to cope with the up to 6 times higher beam intensity, including operational requirements from the rest of the North Area complex and with stricter safety and radiation protection requirements compared to the original installation. Many studies have been performed by the SPS Losses and Activation working group (SLAWG) [151, 152, 153], the Physics Beyond Collider (PBC) Proton Delivery working group [154], the PBC Conventional Beams working group [147], and more recently, by the PBC ECN3 beam delivery task force [155], focusing on the possibility to extract and deliver such high intensities towards TCC8 and ECN3. The present optics of the P42 beam is shown in Fig. 4. In order to reduce the attenuation of the primary beam in the T4 target, several options are considered. One is to increase the vertical beam size so that most of the beam bypasses the target without interacting. Another option is to run dedicated cycles for ECN3 in which the beam is bumped to pass above or below the target plate. This option could allow in principle even higher intensities than $2\times 10^{13}$ ppp on the T10 target. At present, the beam is provided in spills of 4.8 seconds, typically twice per super-cycle of about 40 to 45 seconds during day time, and two equally-spaced spills during 33.6 seconds during nights. One spill every 14.4 seconds is also technically possible and will be the standard cycle adopted during the night from 2023 (Fig. 3). Figure 3: The spill configuration used on 22 October 2022, as seen from SPS Page 1. In recent years, on average 3000 spills per day are delivered, including the typical availability of SPS beams of about 80%. The figure of merit is the time on flat top per year, which is proportional to the duty cycle and the running time. It is also important that the beam intensity is delivered as uniformly as possible. At the time of the NA62 proposal one assumed an effective spill of 3.3 seconds, but thanks to recent improvements we anticipate that it can be significantly better, ideally better than 4 seconds for a 4.8 second spill. The spill length can be modified to some extent, but it is important to maintain the duty cycle as well as the instantaneous intensity, respecting both the RMS power limitation of the SPS and transfer lines as well as the maximum intensity that the SPS is able to accelerate. The latter implies that the intensity per spill shall be kept proportional to the spill length. Studies by the SY-STI group in the PBC context suggest that the TAXs in P42 and K12 are already running close to the maximum intensity allowed [147]. An upgrade is being studied at the technical level in the framework of the ECN3 task force. This implies a substantial upgrade of the T10 target-TAX complex to withstand the higher beam power and to be compatible with RP and maintenance requirements. A new target design could be inspired by the concepts of the CNGS target. The target material might be changed from beryllium to carbon for better handling. However, also more dense target materials can be considered as an option outside of the present baseline, as they could be advantageous in particular for the neutral beamline. The new K12 TAX will be based on the concept that is developed for the P42 TAX. TAX holes will be defined to ensure compatibility with the charged beam as well as the neutral beam. It will probably require a different choice of material in the TAX blocks and cooling closer to the beam impact point. The present interlocks, P0-Survey and Dump Control, protect both the beam line equipment and the NA62 experiment against incorrect magnet currents, T10 target and K12 TAX cooling problems and closed vacuum valves in the K12 beam line. For the higher intensities in the future, this system must be sped up. This is planned within the NACONS program baseline, independently of the proposed high intensity upgrade for ECN3, profiting from a new generation of power converters. On the shorter time scale, some improvements are already under consideration, compatible with the presented HIKE timeline. Many studies are on-going concerning radiation dose levels to equipment and on surface, which must remain under control also at the higher intensities. All these issues will be addressed in the forthcoming report from the ECN3 beam delivery task force. For a $K_{L}\to\pi^{0}\nu\bar{\nu}$ experiment, the production angle of the kaon beam has to be increased to 8 mrad instead of the zero production angle used for the charged mode. This can be achieved by a minor modification of the P42 beam line, i.e. realigning the three last dipoles as documented in the Conventional Beams Working Group Report [147], shown in Fig. 5. The high- intensity proton beam will ultimately impinge on the kaon production target, followed by the TAX beam dumps. ### 5.2 Charged kaon beamline The HIKE charged kaon beamline will remain conceptually as it is now, with essentially the same beam properties apart from the beam intensity (Fig. 6). Consolidation of some equipment is required in the framework of NACONS, such as the K12 machine protection system in synergy with the P42 interlock upgrade. The main upgrades concern the T10 production target and the K12 TAX, including the needed shielding. A pre-study was presented in the Conventional Beams Working Group report [147] and is being followed up further in the working group. The K12 beam is a mixed beam containing 6% of kaons at 75 GeV/$c$. A study has been performed to evaluate if the beam can be enriched by RF separation. RF separation requires a much longer beamline, implying that more kaons decay before the fiducial volume. This decreases not only the kaon flux but also the initial kaon fraction in the beam before enrichment. An RF-separated beam cannot fulfil the HIKE requirements, even with state-of-the-art RF systems [156]. Figure 4: Beam optics for the P42 primary beam line, delivering protons from the T4 production target to the T10 kaon production target. The red line depicts the local values of the R12 (horizontal) and R34 (vertical) beam transport matrix terms. The dotted blue line shows the local dispersion terms R16 (horizontal) and R36 (vertical). Figure 5: Modifications of the T10 target steering dipoles for increasing the production angle to 8 mrad, as required for the KLEVER phase of the HIKE programme. Figure 6: Beam optics for the K12 beamline. The red line depicts the local values of the R12 (horizontal) and R34 (vertical) beam transport matrix terms. The dotted black line shows the local dispersion terms R16 (horizontal) and R36 (vertical). ### 5.3 Neutral kaon beams The original design and simulations for the KLEVER beamline were performed by the Conventional Beams Working Group [147]. The design is based on the experience with the $K_{L}$ beam lines for NA31 and NA48, following guidelines laid out in Ref. [157], and is shown schematically in Fig. 7. It forms the basis for the beamline to be used also for HIKE Phase 2, and will be extended in length for the precise measurement of $K_{L}\to\pi^{0}\nu\bar{\nu}$ (KLEVER). Figure 7: Baseline design of neutral beamline for HIKE Phases 2 and 3. The primary proton beam impinges on the T10 target with a downward vertical angle of 8 mrad for KLEVER, and possibly 2.4 mrad for the $K_{L}$ experiment with tracking. These choices balance several factors for the measurements of $K_{L}\to\pi^{0}\nu\bar{\nu}$ and $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$, respectively. As seen from Figs. 8 and 9, a smaller production angle increases $K_{L}$ flux and hardens the $K_{L}$ spectrum, facilitating the task of vetoing background channels with extra photons, such as $K_{L}\to\pi^{0}\pi^{0}$. On the other hand, with harder input momentum spectra, a smaller $K_{L}$ fraction and a larger $\Lambda$ faction decay inside the fiducial volume. For the $K_{L}\to\pi^{0}\nu\bar{\nu}$ measurement, $\Lambda\to n\pi^{0}$ decays constitute a potentially dangerous background. Neutron interactions on residual gas in the vacuum tank and from halo neutrons striking the detector constitute another dangerous background, which is reduced by going to larger angle. The production angle of 8 mrad softens the $\Lambda$ spectrum, so that almost all $\Lambda$ hyperons decay before the start of the fiducial volume. For the $K_{L}\to\pi^{0}\nu\bar{\nu}$ measurement, the reduced $K_{L}$ flux is compensated to some extent by an increase in the fraction of $K_{L}$ mesons that decay in the fiducial volume. However, this is not true for $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$, and moreover, background from $\Lambda$ decays and neutron interactions is not a concern in this case. A production angle as small as possible is therefore preferred for HIKE Phase 2, subject to limitations from the neutron flux in the beam at very small angles. A production angle of 2.4 mrad (the same as for the $K_{L}$ beam in NA48) is a likely choice. The $K_{L}$ and $\Lambda$ momentum spectra for production angles of 2.4 and 8 mrad are shown in Fig. 10, and the fluxes and other parameters are tabulated in Table 3. Figure 8: Mean (left) and peak (right) momentum for $K_{L}$ (red) and $\Lambda$ (blue) components vs beam production angle, from FLUKA (squares) and Geant4 (circles), with solid (dashed) curves from fits to the form in Ref. [158]. Figure 9: Left: neutral beam fluxes per POT per $\mu$sr of acceptance for $K_{L}$ (blue) and $\Lambda$ (red) vs beam production angle. Right: ratios of $\gamma$ (green), $n$ (red), and $\Lambda$ (blue) to $K_{L}$ fluxes vs beam production angle. FLUKA (Geant4) results are shown with squares (circles), with solid (dashed) curves from fits. Figure 10: $K_{L}$ (blue) and $\Lambda$ (red) beam momentum spectrum (number of particles in beam per POT per $\mu$sr of neutral beam acceptance) for production angles of 2.4 mrad (left) and 8 mrad (right). The solid curves show a parameterisation [158] of FLUKA results at production; dashed curves show the momentum spectra for particles decaying in an indicative fiducial volume of the experiment extending from 135 m to 195 m downstream of the target. Table 3: Parameters of $K_{L}$, $n$, $\Lambda$ production at angles of 2.4 and 8.0 mrad, obtained from the FLUKA simulation of the beamline. An indicative fiducial volume (FV) extending from 135 m to 195 m from the production target is considered. | 2.4 mrad | 8.0 mrad ---|---|--- Mean $K_{L}$ momentum (GeV/$c$) | | at production | 79 | 39 in FV | 46 | 26.4 Mean $\Lambda$ momentum (GeV/$c$) | | at production | 151 | 66 in FV | 285 | 251 $K_{L}$ rate ($10^{-6}$/pot/$\mu$sr) | | at production | 187 | 73 in FV | 6.8 | 4.1 $\Lambda$ rate ($10^{-6}$/pot/$\mu$sr) | | at production | 310 | 37 in FV | 0.053 | 0.000132 Table 4: Particle fluxes in the neutral beam, obtained from the FLUKA simulation. The rates in MHz/GHz assume a primary beam intensity of $6.7\times 10^{12}$ pot/s. Beam component | Before converter | After converter | After final collimator ---|---|---|--- | pot-1 | GHz | pot-1 | MHz | pot-1 | MHz $\gamma$ | | | | | | 2em $E>1$ GeV | $4.2\times 10^{-3}$ | 27.9 | $9.1\times 10^{-4}$ | 6100 | $2.97\times 10^{-5}$ | 198 2em $E>5$ GeV | $2.19\times 10^{-3}$ | 14.6 | $7.1\times 10^{-5}$ | 470 | $7.9\times 10^{-6}$ | 53 2em $E>10$ GeV | $1.55\times 10^{-3}$ | 10.3 | $1.81\times 10^{-5}$ | 121 | $3.15\times 10^{-6}$ | 21 2em $E>30$ GeV | $6.3\times 10^{-4}$ | 4.2 | $2.06\times 10^{-6}$ | 13.7 | $6.2\times 10^{-7}$ | 4.1 $n$ | | | | | | 2em $E>1$ GeV | $4.3\times 10^{-4}$ | 2.88 | $4.2\times 10^{-4}$ | 2820 | $6.7\times 10^{-5}$ | 440 $K_{L}$ | | | | | | 2em $E>1$ GeV | $1.37\times 10^{-4}$ | 0.91 | $1.29\times 10^{-4}$ | 870 | $2.11\times 10^{-5}$ | 140 The target is immediately followed by a first collimator that stops hadrons outside the beam acceptance before they decay into muons. The non-interacting protons are swept further downward by a strong dipole magnet. Protons and charged secondaries are dumped in the TAX, which has apertures that allow the beam to pass. A photon converter consisting of 9$X_{0}$ of high-$Z$ material is positioned at the centre of the TAX between the two modules and reduces the flux of high-energy photons ($E>5$ GeV) in the neutral beam by two orders of magnitude. A thinner oriented crystal converter is an optional possibility under study. Shower products emerging from the TAX are swept horizontally by a dipole magnet. Three collimators define the beam acceptance in a clean way. The defining collimator is located at $1/3$ of the distance to the final collimator and defines the beam angular acceptance to not more than $\pm$0.4 mrad, which is necessary to measure the transverse momentum of the $\pi^{0}$ to sufficient precision in order to reject background from $K_{L}\to\pi^{0}\pi^{0}$ decays. A cleaning collimator stops debris from scatterings in the jaws of the defining collimator, and a final collimator stops scattering products from the cleaning collimator. The defining and cleaning collimator are each followed by strong sweeping magnets. The final collimator will be active, and will define the start of the fiducial volume. Sufficient space is available to add more collimation stages, depending on future design iterations within the Conventional Beams Working Group. #### 5.3.1 Simulation of the $K_{L}$ beamline A detailed FLUKA simulation of the entire beamline has been developed by the Conventional Beams Working Group. Fig. 11 shows the momentum distributions for $K_{L}$, photons, and neutrons in the beam at various stages of the beamline simulation: at generation (at the exit from the target), after the converter and each of the collimators, and at the end of the beamline, at the front face of the small-angle calorimeter (SAC). The normalisation of these distributions provides the particle fluxes in the neutral beam per incident proton. Of particular interest are the photon and neutron fluxes after the final collimator. The fluxes of photons, neutrons and $K_{L}$ in the beam are summarised in Table 4. For the purposes of the sensitivity estimates for the KLEVER phase, $2.1\times 10^{-5}$ $K_{L}$ mesons enter the detector per proton incident on the target. The beam rates in the table assume a primary beam intensity of $2\times 10^{13}$ ppp with a (pessimistic) effective spill length of 3 seconds. In addition to the 140 MHz of $K_{L}$ mesons entering the detector, there are 440 MHz of neutrons: the $n/K_{L}$ ratio of about 3 observed for particle production at 8 mrad is not significantly changed during transport of the neutral beam. In addition, there are about 50 MHz of photons with $E>5$ GeV entering the detector, most of which are incident on the SAC at the downstream end of the detector (the condition $E<5$ GeV corresponds to the SAC threshold for KLEVER). The FLUKA simulation also contains an idealised representation of the KLEVER experimental setup, for the purposes of evaluating rates on the detectors from beam halo. These rates, with specific reference to the KLEVER experimental geometry, are discussed in Section 8.2. As seen from Fig. 11 and Table 4, the photon converter in the TAX dramatically reduces the photon flux in the beam, especially for high energy photons. This is critical to avoid blinding the SAC at the downstream end of the beamline. In the baseline design, the converter is a tungsten prism of 32.9 mm thickness, corresponding to 9.4$X_{0}$, or 7.3 photon conversion lengths. This thickness is chosen to keep the rate of photons with $E>5$ GeV below 40 MHz at the entrance to the SAC. On the other hand, this thickness corresponds to 58% of a nuclear collision length and 33% of an interaction length, so that about 35% of $K_{L}$ mesons and 40% of neutrons interact in or are scattered out of the beam by the converter. Figure 11: Momentum distributions for $K_{L}$ mesons (top), photons (bottom left), and neutrons (bottom right) in the neutral beam at various points along the neutral beamline. Figure 12: Photon fluxes (for $E>5$ GeV) at the SAC as functions of photon converter thickness for beryllium, copper, and lead targets. The dotted line indicates the flux corresponding to a rate of 40 MHz, considered to be the maximum tolerable rate for high-energy photons on the SAC. The use of a high-$Z$ material for the target is an interesting alternative to the current baseline. Most of the prompt photons come from $\pi^{0}$ decays within the target, and a high-$Z$ material causes many of these photons to be converted immediately, while the overall thickness of the target in nuclear interaction lengths is unchanged. No large differences are observed in terms of hadronic production for targets of identical thickness in nuclear interaction lengths [159]. Photon rates at the SAC for beryllium, copper and tungsten targets (each with a nuclear interaction length equivalent to 400 mm Be), determined with the FLUKA simulation, are shown in Fig. 12. The required converter thicknesses for beryllium, copper, and tungsten targets of equivalent thickness are 7.3, 5.2, and 3.8 photon conversion lengths, respectively, resulting an effective increase in the $K_{L}$ flux into the detector of 15% (28%) for a copper (tungsten) target relative to a beryllium target. Changing the target material strongly impacts the target design, as the energy deposit per unit of volume is much higher in a high-$Z$ target than in beryllium, adding to cost and complexity of the new target design. Another promising technique to reduce the photon content in the beam is to use a crystal metal photon converter. If the crystal axis is aligned with the incoming photon direction, the coherent effect of the crystal lattice promotes pair production, leading to an effective decrease in the photon conversion length [160, 161, 162]. The effects of coherent interactions increase with photon energy and for decreasing angle of incidence. A series of exploratory tests was performed with a set of tungsten crystals at the CERN SPS in summer 2018, together with the AXIAL collaboration [163]. In particular, a commercial quality tungsten crystal of 10-mm thickness was targeted with a tagged photon beam. When the <111> crystal axis was aligned with the beam to within 2.5 mrad, the multiplicity of charged particles was found to be enhanced by a factor 1.6–2.3 for photon energies over the range of 30–100 GeV. Ref. [163] also reports simulations validated by this result that suggest that a crystal of this type could be used to reduce the thickness of the photon converter by 15–20% at no cost in effectiveness. Such a solution appears to be relatively easy to implement. #### 5.3.2 Beam for Phase 2 (a multi-purpose $K_{L}$ experiment) The proposed HIKE Phase 2 is based on the $K_{L}$ beam, but a detector very similar to the Phase 1 experiment, including a tracking system. This phase of the programme will make use of the neutral beamline described above with a production angle chosen to optimise the statistics for the measurement of $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ decays, likely 2.4 mrad. Alternative beamline options can be studied. The full FLUKA simulation will be performed when the choice of production angle is finalised. For the sensitivity estimates in Section 7, the entries in Table 4 are scaled to account for production angle of 2.4 mrad (Table 3). #### 5.3.3 Beam for Phase 3 (KLEVER: measurement of $K_{L}\to\pi^{0}\nu\bar{\nu}$ decay) After the initial neutral beam studies for the KLEVER phase of HIKE were completed, it was discovered that the background from $\Lambda\to\pi^{0}n$ decays is more difficult to suppress than previously thought. Therefore, the amount of $\Lambda$ decays in the fiducial region has to be reduced further by several orders of magnitude, whilst preserving as much as possible the kaon flux. Various options were considered, such as lengthening the distance between the target and the start of the fiducial volume, reducing the primary proton beam energy (for example, from 400 to 300 GeV), and increasing the production angle (for example, from 8 mrad to 20 mrad). Lengthening the beam line by 150 m seems to be the best option, and several possibilities are under consideration. Moving the T10 target further upstream is one of them, but this may have complex radiation protection implications and might require the whole experiment to be installed on a slope, depending on the exact target location. The more straightforward option might be to prolong the ECN3 cavern by 150 m, implying civil engineering works. Simulations show that this brings the background from $\Lambda$ decays to an acceptable level (Section 8.3). The technical details and costs of the different options are under discussion with the Civil Engineering group of CERN. An extension of the ECN3 cavern implies that the detector will be installed 150 m downstream of the present NA62 location. At the moment, we do not contemplate changing the transverse dimensions of the setup. The positions of the beamline elements would change as noted in Table 5. The beam solid angle and hence the $K_{L}$ flux decreases by a factor of $(0.256/0.400)^{2}\approx 0.41$. Since the $p_{\perp}$ reconstruction at the downstream end of the FV depends on the beam spot size, it is not trivial to recover the sensitivity of the experiment by enlarging its transverse dimensions, and we favour the other approaches discussed in Section 8.3. The full FLUKA simulation of the extended configuration will be performed as soon as the design is finalised. For the sensitivity estimates, the rates in Table 4 can be scaled to account for the reduction in solid angle from the extension. In case resources allow to advance the construction of the ECN3 extension, HIKE Phase 2 can be performed with the extended beamline and the detector in the new location. This would eliminate the need for infrastructural work to change beamline configurations between Phase 2 and KLEVER and allow precise information on the beam composition and halo to be obtained during Phase 2, thereby improving the understanding of beam-related backgrounds in KLEVER. Table 5: Collimator positions and other parameters for standard and extended KLEVER beamlines. Configuration | Standard | Extended ---|---|--- Target position [m] | 0 (T10) | 0 (T10) Defining collimator position [m] | 40 | 90 Cleaning collimator position [m] | 80 | 180 Final collimator position [m] | 120 | 270 Beam opening angle [mrad] | 0.400 | 0.256 ## 6 Phase 1: a multi-purpose $K^{+}$ decay experiment ### 6.1 Experimental layout While the success of the NA62 experiment has proven that its layout is suitable for a precision ${\cal B}(K^{+}\to\pi^{+}\nu\bar{\nu})$ measurement, new or upgraded detectors will replace those of NA62 with the goal of sustaining secondary-beam rates four times higher than those of NA62, to substantially boost the statistical sensitivity. To this end, new technologies will be used, in synergy with the upgrades of the LHC experiments. While some detectors will be renovated for the start of the HIKE $K^{+}$ phase (Phase 1), others are already intrinsically fast in terms of detector technologies but will need readout upgrades. Therefore a staged approach will be used, where renovated or new detectors are inserted as soon as they are needed and ready, while maintaining the general principle that changes must serve the remaining phases of the programme once they are applied. In the baseline configuration, the $K^{+}$ beam is produced by $1.2\times 10^{13}$ $400~{}{\rm GeV}/c$ protons/spill at zero angle on target. The mean $K^{+}$ momentum at the entrance to the decay volume is 75 GeV/$c$. The layout (Fig. 13) includes a beam tracker, a kaon-identification detector, and a veto counter upstream of the decay volume; a main tracking system with the MNP33 magnet in the decay volume; calorimeters, vetoes and particle identification detectors downstream of the decay volume; and a large-angle veto system surrounding the decay volume and part of the region downstream. HIKE Phase 1 ($K^{+}$) Figure 13: HIKE Phase 1 layout, with an aspect ratio of 1:10. ### 6.2 Physics sensitivity #### 6.2.1 $K^{+}\to\pi^{+}\nu\bar{\nu}$ measurement The first phase of HIKE will be dedicated to a detailed measurement of the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay, with a target of reaching a branching ratio measurement with $\mathcal{O}(5\%)$ precision, competitive with the theoretical uncertainty. The approach towards this goal will continue, and build upon, the successful strategy of the NA62 experiment. The keystones of the measurement, required to suppress backgrounds at the $\mathcal{O}(10^{11})$ level, are the following: * • High efficiency and high-precision tracking of both the $K^{+}$ upstream and $\pi^{+}$ downstream. This, coupled with a careful choice of signal regions, will allow kinematic suppression of backgrounds by a factor of $\mathcal{O}(10^{3})$. * • High precision time measurements, allowing time-matching between upstream and downstream detectors with $\mathcal{O}(20~{}\text{ps})$ precision. Simulations show that with this time-matching performance, the effect of the higher intensity can be totally compensated in the matching between upstream and downstream tracks, without losing efficiency for the signal selection and without increasing the relative contamination of background coming from upstream decays and interactions with respect to NA62. * • Comprehensive and hermetic veto systems: * – Photon veto detectors with hermetic coverage of the 0–50 mrad range for photons from $\pi^{0}\to\gamma\gamma$ decays, allowing photon suppression by a factor of $\mathcal{O}(10^{8})$. * – Veto detectors to cover downstream regions not covered by the principal detectors. * – Rejection of upstream decays and interactions, for suppressing possible upstream backgrounds by an additional factor of $\mathcal{O}(10)$, with changes to the beam line setup and new systems dedicated to the detection of the upstream background events. * • High-performance particle identification system (for $\pi$/$\mu$ discrimination), suppressing backgrounds with muons by a factor $\mathcal{O}(10^{7})$. Crucial to the study of the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay at HIKE will be the exploitation of the high-intensity environment, taking advantage of the high $K^{+}$ flux while mitigating detector pileup effects. A critical performance indicator is the “random veto efficiency”, i.e., the fraction of events passing signal selection criteria that are sensitive to accidental activity in the detector systems. The random veto efficiency measured within the NA62 selection for the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay at the current intensity working-point is $\varepsilon_{\rm RV}\approx 65\%$. This is dependent on selection criteria that are limited by the timing precision of the detectors. The random veto efficiency is approximately linear as a function of the instantaneous beam intensity. At HIKE, this random veto efficiency must be maintained or improved, requiring an improvement in time resolution by the same factor as the intensity increase. With this performance, a projection for the random veto efficiency as a function of intensity is shown in Fig. 14. Individual contributions from photon veto subsystems (SAV, LAV, ECAL) are indicated which in total give the ‘photon rejection’ curve. The curve labelled ‘multiplicity rejection’ relates to the selection criteria required to reject additional activity in an event (outside the photon veto detectors). Finally, the total random veto efficiency, combining photon and multiplicity rejection, is shown. At a secondary beam intensity 4 times higher than NA62, with the detector updates and corresponding selection changes shrinking veto windows by a factor of 4, the random veto efficiency is maintained at the same level as currently achievable at NA62. Figure 14: Expected HIKE Phase 1 random veto efficiency $\varepsilon_{\rm RV}$ for the $K^{+}\to\pi^{+}\nu\bar{\nu}$ analysis as a function of instantaneous beam intensity: the total effect (grey band) and the contributions from the individual photon/multiplicity veto conditions. The NA62 graph is also shown for comparison: the lower $\varepsilon_{\rm RV}$ values with respect to HIKE Phase 1 are due to the worse time resolution. The HIKE sensitivity for the ${\cal B}(K^{+}\to\pi^{+}\nu\bar{\nu})$ measurement is estimated as follows. * • The effective number of collected $K^{+}$ decays per year is $N_{K}/{\rm year}\simeq 2\times 10^{13}$. This can be extrapolated from the data already collected by the NA62 experiment, scaling for an intensity four times larger than the NA62 nominal intensity and 200 days of data-taking per year. This automatically includes effects already present in the current setup, such as duty cycle, SPS beam availability, secondary beamline downtime, and detector and DAQ efficiencies. * • Acceptance of the signal selection: $\varepsilon_{A}\simeq 0.1$. The improvements on the time-matching performance and on the resolution of the charged-particle identification system will lead to an increase of about 50% relative to that at the current NA62 working point. * • Random veto efficiency ($\varepsilon_{\rm RV}$) and trigger efficiency ($\varepsilon_{\rm trig}$), which according to the projections already described should remain the same as for the current NA62 working point even at higher intensity: $\varepsilon_{\rm RV}\simeq 0.7$, $\varepsilon_{\rm trig}\simeq 0.9$. Under these conditions, the single event sensitivity attainable in one year of data-taking is found to be ${\cal B}_{\rm SES}(1\;{\rm year})=(N_{K}/{\rm year}\cdot\varepsilon_{A}\cdot\varepsilon_{\rm RV}\cdot\varepsilon_{\rm trig})^{-1}\simeq 8\times 10^{-13}$, leading to the number of expected SM signal events per year: $N_{\pi\nu\bar{\nu}}/{\rm year}\simeq 100.$ With an $O(10\%)$ relative background contamination in the signal sample and systematic uncertainty sources under control to $O(1\%)$ or better, a measurement of the branching ratio $\mathcal{B}(K^{+}\to\pi^{+}\nu\bar{\nu})$ to $\mathcal{O}(5\%)$ precision can be reached by HIKE in 4 years of data- taking. Assuming that the number of observed events is equal to the SM expectation, values of ${\cal B}(K^{+}\to\pi^{+}\nu\bar{\nu})$ higher than 10% with respect to the SM prediction will be excluded at 95% CL. #### 6.2.2 Test for scalar amplitudes in the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay As introduced in Section 3.1.1, in BSM scenarios the measurable branching ratio of $K^{+}\to\pi^{+}\nu\bar{\nu}$ is formed from two components: the SM process $K^{+}\to\pi^{+}\nu\bar{\nu}$ with a purely vector nature and a possible BSM process $K^{+}\to\pi^{+}\nu\nu$ with a scalar nature [68, 69]. Therefore the experimentally measured branching ratio is given by [23, 68]: $\mathcal{B}(K^{+}\to\pi^{+}\nu\bar{\nu})=\mathcal{B}_{\rm SM}(K^{+}\to\pi^{+}\nu\bar{\nu})+\sum_{i\leq j}^{3}\mathcal{B}_{\rm LNV}(K^{+}\to\pi^{+}\nu_{i}\nu_{j}).$ (5) The SM and BSM contributions are predicted to have different kinematic distributions (Fig. 15). To identify the nature of the decay, an investigation of the shape of the distribution of selected signal candidates as a function of kinematic variables will be performed. In absolute terms, within the signal regions (as defined in the current NA62 analysis), the acceptance for the BSM mode is approximately 75% of the acceptance for the SM mode. With a sample of several hundred $K^{+}\to\pi^{+}\nu\bar{\nu}$ candidates, HIKE will be able to study the shape of the kinematic distributions as well as compare sets of sub- categories of phase-space to distinguish between, or constrain the relative contributions of, the possible vector and scalar natures of the process. Figure 15: Simulated distributions of the squared missing mass $m_{\rm miss}^{2}=(P_{K^{+}}-P_{\pi^{+}})^{2}$ vs pion momentum in the laboratory frame for the SM $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay with a vector nature (left) and the LNV $K^{+}\to\pi^{+}\nu\nu$ decay with a scalar nature (right). Black boxes indicate the signal regions for which experimental sensitivity is highest. #### 6.2.3 Feebly-interacting particle production in $K^{+}\to\pi^{+}X_{\rm inv}$ decays As described in Section 3.6, searches for the $K^{+}\to\pi^{+}X_{\rm inv}$ decay, where $X_{\rm inv}$ is a FIP (either a dark scalar or ALP), can be performed within the analysis framework developed for the study of the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay. This is possible because the signature of these two decays is identical except that the $K^{+}\to\pi^{+}X_{\rm inv}$ decay produces a peak in the $m_{\rm miss}^{2}=(P_{K^{+}}-P_{\pi^{+}})^{2}$ variable at $m_{X}^{2}$, where $m_{X}$ is the mass of the FIP. Searches are therefore performed by looking for evidence of a peaking signal on top of the background dominated by the SM $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay. The search strategy is described in Ref. [141], which furthermore presents results based on the analysis of 2017 data, which has been updated for the full Run 1 (2016–2018) dataset in Ref. [67]. Based on this experience, a sensitivity projection for HIKE Phase 1 has been performed assuming a 40-fold increase in the size of the data sample with respect to NA62 Run 1. The principal background comes from the SM $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay, and other backgrounds are negligible in comparison. The selection acceptance and efficiency are assumed to be consistent with the results obtained for the NA62 2018 data [67]. In this way realistic projections are performed, accounting for all first-order effects. The results are shown in Fig. 16 for the excluded regions in the interpretations of $X_{\rm inv}$ as either a dark scalar mixing with the Higgs boson or an ALP with fermionic couplings. Additional specialisation of the selection used for the $K^{+}\to\pi^{+}X_{\rm inv}$ decays might be possible, further enhancing sensitivity. HIKE Phase 3 sensitivity projections for the search for the $K_{L}\to\pi^{0}X_{\rm inv}$ decay, refined with respect to Ref. [20], are also shown in Fig. 16. HIKE Phases 1 and 3 provide complementary sensitivities, allowing for exclusion of, or discovery within, new phase-space regions which are otherwise difficult to cover. Figure 16: Left: excluded regions at 90% CL of the $(m_{S},\sin^{2}\theta)$ parameter space for a dark scalar, $S$, of the BC4 model [20]. Right: excluded regions of the parameter space $(m_{a},g_{Y})$ for an ALP, $a$, of the BC10 model [20]. The exclusion bound from analysis of NA62 Run 1 data [141, 142, 67] is shown in purple; the projected exclusion for HIKE Phase 1 and Phase 3 (KLEVER, assuming invisible $S$ or $a$ and updated from Ref. [20]) is shown in red and light blue, respectively. Other exclusion bounds are shown in grey, including results from E949 [164], CHARM [165, 166], NA48/2 [167], LHCb [168, 169], Belle [170], $K_{\mu 2}$ experiment [171], CLEO [172], KTEV [76], MicroBooNE [173]. #### 6.2.4 Heavy neutral lepton production in $K^{+}\to\ell^{+}N$ decays Searches for heavy neutral lepton (HNL) production in $K^{+}\to\ell^{+}N$ decays have been well established by the NA62 experiment using the main $K^{+}\to\pi^{+}\nu\bar{\nu}$ trigger chain for the $K^{+}\to e^{+}N$ case, and a downscaled control trigger chain for the $K^{+}\to\mu^{+}N$ case. NA62 has published world-leading exclusion on the HNL mixing parameters $|U_{\ell 4}|^{2}$ over much of the accessible mass range of 144–462 MeV/$c^{2}$ with the Run 1 dataset [143, 144]. Both searches are limited by background. In particular, the $K^{+}\to\mu^{+}\nu$ decay followed by $\mu^{+}\to e^{+}\nu\bar{\nu}$ decay in flight, and the $\pi^{+}\to e^{+}\nu$ decay of the pions in the unseparated beam, represent irreducible backgrounds to the $K^{+}\to e^{+}N$ process. The peaking nature of the $K^{+}\to\ell^{+}N$ signal in terms of the reconstructed missing mass allows for data-driven background evaluation, reducing the systematic uncertainties in the background estimates. HIKE Phase 1 sensitivity projections for HNL production searches in $K^{+}\to\ell^{+}N$ decays, obtained from a detailed analysis assuming NA62-like trigger chains and conditions, are shown in Fig. 17. HIKE Phase 1 offers world-leading sensitivity to $|U_{e4}|^{2}$ in the region $m_{N}>140~{}{\rm MeV}/c^{2}$, approaching the seesaw bound and complementing future long-baseline neutrino experiments [139], and is expected to improve on the state-of-the-art for lower HNL masses via searches for the $K^{+}\to\pi^{0}e^{+}N$ [174] and $\pi^{+}\to e^{+}N$ decays. HIKE Phase 1 also provides competitive sensitivity to $|U_{\mu 4}|^{2}$. The projection for $|U_{\mu 4}|^{2}$ assumes data collection with a highly-downscaled control trigger, and may improve by an order of magnitude in case a software trigger based on the streaming readout (Section 11.2) is employed. Figure 17: Summary of the upper limits at 90% CL of the HNL mixing parameters $|U_{e4}|^{2}$ and $|U_{\mu 4}|^{2}$ obtained from production searches, and HIKE Phase 1 sensitivity with $K^{+}\to\ell^{+}N$ decays. HIKE projection for $|U_{\mu 4}|^{2}$ assumes data collection with a highly-downscaled control trigger, and may improve by an order of magnitude in case a software trigger is employed. #### 6.2.5 Other measurements HIKE Phase 1 is expected to increase the world samples of many rare $K^{+}$ decays by an order of magnitude, developing upon the trigger strategies developed for the NA62 experiment. This would bring many rare decay measurements to a new level of precision. In particular, we expect to collect background-free samples of several times $10^{5}$ events of both $K^{+}\to\pi^{+}e^{+}e^{-}$ and $K^{+}\to\pi^{+}\mu^{+}\mu^{-}$ decays, leading to a powerful test of lepton universality (LFU) by comparing the form- factor parameters ($a_{+}$, $b_{+}$) of the two decay modes. This is of particular interest in the framework of the global NP fit [85] to the kaon decay data (Section 3.1.5, Fig. 2). Considering that the precision of the recent NA62 measurement in the muon channel [80] is limited by the size of the dataset, HIKE Phase 1 is expected to measure the difference of the form-factor parameter $a_{+}$ between the two decay modes, $\Delta a_{+}^{e\mu}$, to a precision of $\pm 0.007$, and the corresponding difference $\Delta b_{+}^{e\mu}$ to a precision of $\pm 0.015$. Exploitation of the correlations between the measured $a_{+}$ and $b_{+}$ parameters in each mode will provide additional power in terms of LFU tests. In terms of searches for $K^{+}$ and $\pi^{0}$ decays violating lepton number or flavour conservation, including $K^{+}\to\pi^{-}(\pi^{0})\ell_{1}^{+}\ell_{2}^{+}$, $K^{+}\to\pi^{+}\mu^{\pm}e^{\mp}$, $K^{+}\to\ell_{1}^{-}\nu\ell_{2}^{+}\ell_{2}^{+}$ and $\pi^{0}\to\mu^{\pm}e^{\mp}$, the experimental technique based on dedicated di-lepton trigger chains has been firmly established by the NA62 experiment, leading to world-leading upper limits of ${\cal O}(10^{-11})$ on the branching ratios of a number of processes with the NA62 Run 1 dataset [119, 120, 121]. These searches are not limited by background, and HIKE Phase 1 sensitivity is expected to improve in the future almost linearly with the size of the dataset to the ${\cal O}(10^{-12})$ level. ## 7 Phase 2: a multi-purpose $K_{L}$ decay experiment ### 7.1 Experimental layout The baseline design of the multi-purpose $K_{L}$ decay experiment includes a 120 m long neutral beamline with the secondary beam opening angle of 0.4 mrad, proposed and studied extensively for the KLEVER phase of the HIKE programme. The beamline involves four stages of collimation, and provides adequate suppression of the short-lived $K_{S}$ and $\Lambda$ components. A detailed description of the beamline, including the expected $K_{L}$ yields and momentum spectra, is provided in Section 5.3 Unlike the KLEVER phase, HIKE Phase 2 detector will be equipped with a tracker, which offers the opportunity to perform characterisation of the KLEVER beam. For the forthcoming proposal, alternative beamline configurations can be studied. The proposed $K_{L}$ production angle is 2.4 mrad (to be compared to 8 mrad for the KLEVER phase), which improves the $K_{L}$ yield and increases the mean $K_{L}$ momentum (Figs. 8, 10) therefore improving the acceptances for $K_{L}\to(\pi^{0})\ell^{+}\ell^{-}$ decays. Thanks to the relatively compact detector, HIKE Phase 2 allows for a 90 m long fiducial decay volume to be accommodated in the present ECN3 experimental hall, and no major civil engineering work is required in preparation for this phase of the programme. It is proposed to use the HIKE Phase 1 experimental setup with minimal modifications (Fig. 18). The GTK, KTAG, RICH and SAC detectors will be removed, the STRAW spectrometer will be shortened to a total length of 25 m, and central holes of the STRAW chambers will be realigned on the neutral beam axis (Section 10.2.3). Reduction of the magnetic field of the spectrometer dipole magnet by about 20%, leading to a momentum kick of 210 MeV/$c$, is possible without significant degradation of the mass resolution. HIKE Phase 2 ($K_{L}$) Figure 18: HIKE Phase 2 layout, with an aspect ratio of 1:10. ### 7.2 Physics sensitivity The expected $K_{L}$ yield in the beam is $5.4\times 10^{-5}$ per proton on target, and the assumed integrated proton flux of $1.2\times 10^{19}$ pot/year leads to the number of $K_{L}$ decays in the decay volume of $3.8\times 10^{13}$/year. The mean momentum of $K_{L}$ mesons entering the decay volume is 79 GeV/$c$, while the mean momentum of decaying $K_{L}$ mesons is 46 GeV/$c$ (Table 3, Fig. 8). The expected signal yields for several principal rare and forbidden $K_{L}$ decays have been evaluated using a full Geant4-based simulation, reconstruction and analysis chain within the flexible HIKE offline software platform being developed on the basis of the NA62 software (Section 12). The results are summarised in Table 6, and acceptances in bins of the longitudinal coordinate of the decay vertex and $K_{L}$ momentum are shown in Fig. 19. Table 6: HIKE Phase 2 sensitivity estimate for the principal rare and forbidden $K_{L}$ decays: assumed branching ratios (see Section 3.1.2 for details), acceptances and expected signal yields in five years of data taking. Mode | Assumed branching ratio | Acceptance | Signal yield in five years ---|---|---|--- $K_{L}\to\pi^{0}e^{+}e^{-}$ | $3.5\times 10^{-11}$ | 2.1% | 140 $K_{L}\to\pi^{0}\mu^{+}\mu^{-}$ | $1.4\times 10^{-11}$ | 6.0% | 160 $K_{L}\to\mu^{+}\mu^{-}$ | $7\times 10^{-9}$ | 17% | $2.3\times 10^{5}$ $K_{L}\to\mu^{\pm}e^{\mp}$ | – | 16% | – Figure 19: Acceptances for rare $K_{L}$ decays for HIKE Phase 2, in bins of the decay vertex longitudinal coordinate (left) and $K_{L}$ momentum (right). Momentum distribution of the $K_{L}$ mesons decaying in the decay volume is shown by a continuous curve in the right panel. For the $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ decays, the expected SM Dalitz plot distributions [72] are used in the simulation, and a selection condition $m_{ee}>140~{}{\rm MeV}/c^{2}$ is applied in the $K_{L}\to\pi^{0}e^{+}e^{-}$ case as required to suppress backgrounds from major $K_{L}$ decays into neutral pions followed by $\pi^{0}\to\gamma e^{+}e^{-}$ decays. The resulting single-event sensitivities for $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ decays (Table 6) improve by more than two orders of magnitude on the previous searches at the kTeV experiment [75, 76]. Suppression of the $K_{L}\to\gamma\gamma\ell^{+}\ell^{-}$ background, which represents a significant limitation especially in the $K_{L}\to\pi^{0}e^{+}e^{-}$ case [77], relies on the good photon energy resolution provided by the EM calorimeter. The current NA62 LKr calorimeter [175] would provide a 2.2 MeV di-photon mass resolution for the proposed HIKE Phase 2 setup. Further potential backgrounds to $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ decays come from major $K_{L}$ decays such as $K_{L}\to\pi^{+}\pi^{-}\pi^{0}$ with $\pi^{\pm}$ misidentification or decay in flight, and from pileup effects. Background estimation for the $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ decays is currently in progress, however the extensive experience of the NA62 lepton flavour/number violation programme in terms of background reduction [119, 120, 121] suggests that the proposed experimental technique allows for background suppression to the level of ${\cal O}(10^{-11})$ for rare and forbidden decays into multiple tracks and photons, for a well-collimated beam. Considering the expected signal yields in excess of 100 events (Table 6), we conclude that the first observation of the $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ decays is likely, and a sensitivity at the SM branching ratio level of ${\cal O}(10^{-11})$ can certainly be reached. Detailed HIKE Phase 2 sensitivity studies are in progress for the forthcoming proposal. The expected $K_{L}\to\mu^{+}\mu^{-}$ signal yield suggests a 0.2% statistical precision on the measurement of the decay branching ratio. We expect that overall a precision better than 1% is attainable when normalising to the $K_{L}\to\pi^{+}\pi^{-}$ decay, which improves on the previous measurement by the BNL-E871 experiment [87]. HIKE Phase 2 offers sensitivities of ${\cal O}(10^{-12})$ for branching ratios of a broad range of rare and forbidden $K_{L}$ decay modes. As an example, the acceptance for the lepton flavour violating $K_{L}\to\mu^{\pm}e^{\mp}$ decays (Table 6) leads to expected upper limits of both decay branching ratios of $8\times 10^{-14}$ at 90% CL in five years of operation (the null hypothesis and assuming low background), improving the limit by a factor of 60 on the previous search by the BNL-E871 experiment [126]. ## 8 Phase 3 (KLEVER): measurement of the $K_{L}\to\pi^{0}\nu\bar{\nu}$ decay Exploratory work on an experiment to measure ${\cal B}(K_{L}\to\pi^{0}\nu\bar{\nu})$ in a future stage of the ECN3 high-intensity kaon physics programme began even before the start of NA62 data taking. The basic design work for the KLEVER experiment ($K_{L}$ Experiment for VEry Rare events) was carried out as a part of the first Physics Beyond Colliders initiative at CERN, in preparation for the 2020 update of the European Strategy for Particle Physics [176]. As in the other phases of the programme, KLEVER makes use of the 400-GeV primary proton beam from the SPS, slow extracted and transported to the T10 target (Section 5.1) at an intensity of $2\times 10^{13}$ protons per pulse (ppp), corresponding to about six times the nominal NA62 intensity. The neutral secondary beam for KLEVER is derived at an angle of 8 mrad; the beamline design is described in Section 5.3. In practice, the choice of production angle has been optimised together with the limits of the fiducial volume (FV) to maximise the signal sensitivity and minimise $K_{L}\to\pi^{0}\pi^{0}$ and $\Lambda\to n\pi^{0}$ backgrounds. For a production angle of 8 mrad, the neutral beam has a mean $K_{L}$ momentum of 40 GeV, so that 5% of $K_{L}$ decay inside an FV extending from 130 m to 190 m downstream of the target. The $K_{L}$ mesons decaying inside the FV have a mean momentum of 27 GeV/$c$. The boost from the high-energy beam facilitates the rejection of background channels such as $K_{L}\to\pi^{0}\pi^{0}$ by detection of the additional photons in the final state. On the other hand, the layout poses particular challenges for the design of the small-angle vetoes, which must reject photons from $K_{L}$ decays escaping through the beam exit amidst an intense background from soft photons and neutrons in the beam. Background from $\Lambda\to n\pi^{0}$ decays in the beam must also be kept under control. The KLEVER goal is to achieve a sensitivity of about 60 events for the $K_{L}\to\pi^{0}\nu\bar{\nu}$ decay at the SM BR with an $S/B=1$. At the SM BR, this would correspond to a relative uncertainty of about 20%. We would expect to be able to observe a discrepancy with SM predictions with $5\sigma$ significance if the true BR is a bit more than twice or less than one-quarter of the SM BR, or with $3\sigma$ significance if the true BR is less than half of the SM rate. As noted in Section 5.3, with an opening angle of 0.4 mrad, the $K_{L}$ yield in the beam is $2.1\times 10^{-5}$ per proton on target (pot). With a fiducial-volume acceptance of 5% and a selection efficiency of 3%, collection of 60 SM events would require a total primary flux of $6\times 10^{19}$ pot, corresponding to five years of running at an intensity of $2\times 10^{13}$ ppp under NA62-like slow-extraction conditions with 3000 spills per day for 200 days per year. The scheduling of the KLEVER phase is influenced by two considerations. Firstly, to provide maximum protection from $\Lambda\to n\pi^{0}$ background, the beamline would need to be lengthened by 150 m (Section 5.3.3). Secondly, despite the efforts to benchmark the KLEVER beam simulation by comparison with existing data on inclusive particle production by protons on lightweight targets at SPS energies (Section 5.3.1), confidence in the simulations and background estimates would be greatly increased by the possibility to acquire particle production data and measure inclusive rates in the KLEVER beamline setup with a tracking experiment in place. This would be possible in HIKE Phase 2 (Section 7). Therefore, although in principle KLEVER could aim to start data taking after LS4 (for which injector physics is currently foreseen to begin in early 2034), the most natural timescale might envisage KLEVER as running after Phase 2. We are studying scheduling options to allow KLEVER data taking to begin before LS5 (foreseen to start in 2038). ### 8.1 Experimental layout The KLEVER setup largely consists of a collection of high-efficiency photon detectors arranged around a 160-m-long vacuum volume to guarantee hermetic coverage for photons from $K_{L}$ decays emitted at polar angles out to 100 mrad and to provide a nearly free path through vacuum up to the main electromagnetic calorimeter (MEC) for photons emitted into a cone of at least 7.5 mrad. The fiducial volume (FV) spans about 60 m just downstream of the active final collimator (AFC), but the photon veto coverage extends along the entire length up to the MEC. The layout of the detector elements is schematically illustrated in Fig. 20. HIKE Phase 3 (KLEVER) Figure 20: HIKE Phase 3 (KLEVER) layout, with an aspect ratio of 1:10. The baseline configuration, without the extension of the beamline by 150 m is shown. The largest elements are about 3 m in diameter. The beginning of the vacuum volume is immediately downstream of the cleaning collimator at $z=80$ m, i.e., 80 m downstream of the T10 target. A 40-m vacuum decay region allows the upstream veto calorimeter (UV) surrounding the AFC to have an unobstructed view for the rejection of $K_{L}$ decays occurring upstream of the detector volume. The UV and AFC at $z=120$ m define the start of the detector volume, which is lined with 25 large-angle photon and charged-particle veto stations (LAV) in five different sizes, placed at intervals of 4 to 6 m to guarantee hermeticity for decay particles with polar angles out to 100 mrad. The MEC, at the downstream end of the vacuum volume, replaces the NA48 LKr calorimeter used in NA62: it reconstructs the $\pi^{0}$ vertex for signal events and helps to reject events with extra photons. A charged-particle veto detector (CPV) in front of the MEC rejects $K_{L}\to\pi^{\pm}e^{\mp}\nu$ and $K_{L}\to\pi^{+}\pi^{-}\pi^{0}$ backgrounds, and a preshower detector (PSD) allows reconstruction of the angles of incidence for photons, providing additional constraints on signal candidates even if only one photon converts. The small-angle vetoes on the downstream side of the MEC intercept photons from $K_{L}$ decays that pass through the beam pipe. The small-angle calorimeter (SAC) intercepts the neutral beam; its angular coverage as seen from the downstream end of the FV extends to $\pm 2$ mrad. Because of the high neutron and photon rate in the beam, the SAC design is one of the most challenging aspects of the experiment. The intermediate-ring calorimeter (IRC) is a ring-shaped detector between the SAC and MEC and intercepts photons from downstream decays that make it through the calorimeter bore at slightly larger angles. In addition to the photon vetoes, the experiment makes use of hadronic calorimeters downstream of the MEC, to help reject background from hadron interactions and the copious $K_{L}$ decays into charged particles such as muons and pions. Because of the experimental challenges involved in the $K_{L}\to\pi^{0}\nu\bar{\nu}$ measurement, and in particular, the very high efficiency required for the photon veto systems, the performance requirements for KLEVER drive the specifications for many of the new detectors for HIKE. In this case, the LAVs, MEC, IRC, SAC, and hadronic calorimeters built for Phases 1 and 2 would all be able to be used for KLEVER in Phase 3. The UV, CPV, and PSD would have to be built new for KLEVER, possibly with recycled components from the analogous detectors for the earlier phases. In the baseline design with the original 120 m beamline, the FV covers the region $130~{}{\rm m}<z<190$ m. The positioning of the FV significantly upstream of the calorimeter, together with the relatively high $K_{L}$ momentum, is key to obtaining sufficient $K_{L}\to\pi^{0}\pi^{0}$ background rejection with the ring-shaped LAV geometry covering polar angles out to 100 mrad. This comes at a cost in acceptance for signal decays, which increases significantly as the FV is moved closer to the calorimeter. Extending the downstream limit of the FV would increase the sensitivity of the experiment but will require improvements in background rejection; further optimisation along these lines is under study. Fig. 20 shows the experimental configuration without a lengthened beamline to suppress background from $\Lambda$ decays. Lengthening the beamline would change the distance from the cleaning collimator to the AFC, so that the cleaning collimator would be upstream of the start of the vacuum tank instead of right at its entrance. The detector configuration would be otherwise unchanged. ### 8.2 Rates and timing performance Table 7: Rates for events with hits on KLEVER detectors, by detector system (left) and for certain event classes (right), evaluated for the 120 m beamline configuration. Detector | Event rate (MHz) | Event class | Rate (MHz) ---|---|---|--- AFC | 2.3 | Exactly 2 hits on MEC | 4.8 UV | 7.1 | Exactly 2 photons on MEC | 1.0 LAV | 14 | 2 hits on MEC with UV, LAV veto | 3.1 MEC | 18 | 2 hits on MEC, no other hits | 0.007 IRC | 22 | | SAC | 95 | | The FLUKA simulation of the beamline performed by the Conventional Beams Working Group (Section 5.3) contains an idealised representation of the experimental setup for the purposes of evaluating rates on the detectors, both from the decays of $K_{L}$ mesons and from the beam halo. The estimated rates by detector system are listed in Table 7, assuming a primary intensity of $2\times 10^{13}$ ppp and a (pessimistic) 3 s effective spill. The simulation was performed for the standard, 120 m beamline (without extension). The rates are obtained with a geometrical representation of the detector volumes and assumed probabilities for hadrons to register signals. For the AFC, UV, LAV, and MEC, the probability for registering signals is assumed to be 25% for neutrons and $K_{L}$ mesons, and 100% for all other particles. For the IRC and SAC, the assumed probability is 10% for neutral hadrons, and 100% for all other particles. The particle tracking threshold is 1 GeV. Since a particle detected in any of the detectors would effectively constitute a veto (the rates of signal candidates being negligible for these purposes), this provides an estimate of the total veto rate: 134 MHz, of which 104 MHz from events with hits on a single detector and 30 MHz from events with hits on multiple detectors. The total rate is dominated by the rate of interactions of beam particles in the SAC, consisting of 13 MHz of hits from $K_{L}$ mesons, 44 MHz of hits from neutrons, and 40 MHz of hits from beamline photons with $E>5$ GeV. Hadronic interactions can be efficiently recognised offline and we assume only 10% of the corresponding rate contributes to the inefficiency from accidental coincidence. The remaining veto rate from accidentals is then at most 100 MHz, which we take as a figure of merit. The event time is obtained from the $\pi^{0}$ candidate reconstructed in the MEC, and the accidental coincidence rate is dominated by events on the SAC. Assuming a $\pm 5\sigma_{t}$ coincidence window, to limit the inefficiency from accidental coincidence to below 25%, the time resolution $\sigma_{t}$ on the coincidence must be better than 250 ps, and the detector design must minimise any contributions from non-gaussian fluctuations to the timing distribution. These considerations lead to the specification that the nominal time resolution for the most critical detectors (the MEC and the SAC) must be better than 100 ps. As noted above, the full FLUKA simulation was performed with the original, 120 m beamline configuration. Many of the rates in Table 7 would be expected to decrease by the ratio of beam solid angle, i.e., by a factor of 0.41, when the beamline is extended. This scaling may not be exact for the upstream detectors (AFC, UV, LAV), as the rates on those detectors are determined in part by the details of the collimation and sweeping scheme, but is is certainly true for the downstream detectors, and in particular, for the SAC, which drives the random veto rate for the experiment. Thus, the downstream extension not only provides a significant general decrease in the rates on the detectors, but would lower the estimated random veto rate from 25% to 10%. The rates for certain classes of events to be acquired are also listed in Table 7: for events with exactly two hits and with exactly two photons on the calorimeter, independently of the other detectors; for events with two hits on the calorimeter without hits on the UV or LAV (presumed to be in online veto); and for events with two hits on the calorimeter and no hits on any of the other detectors. These are estimates of the rates for the physical events—there is no simulation of the detector response apart from the efficiency assumptions outlined above—but these results do suggest that the total rates for event classes that would approximately correspond to level-0 trigger conditions in NA62 are in fact of about the same order of magnitude as in NA62. Figure 21: Radial distribution of particle fluxes at the MEC for KLEVER, from FLUKA beamline simulation. Fig. 21 shows the radial distribution of the particle fluxes at the front face of the MEC, as estimated with the FLUKA simulation of the neutral beam. The beam halo from photons and neutral hadrons drops off rapidly for $10<r<20$ cm, with $r$ the radial distance from the beam axis. For neutrons, the halo-to- core ratio decreases from $2.5\times 10^{-2}$ to $8.8\times 10^{-5}$ over this interval, corresponding to a decrease in the total neutron rate in the halo from 12 MHz to 43 kHz. For $r>20$ cm, the neutral hadron fluxes are on the order of a few Hz/cm2. At $r=20$ cm, there are 1 kHz/cm2 of photons and charged pions and a few hundred Hz/cm2 of muons, dropping off more or less linearly with radius. Once the extended beamline layout is finalised, it will be necessary to re-run the FLUKA-based simulation to obtain definitive estimates of the beam flux, halo, and veto rates in the detectors. ### 8.3 Expected performance for $K_{L}\to\pi^{0}\nu\bar{\nu}$ Simulations of the experiment in the standard configuration carried out with fast-simulation techniques (idealised geometry, parameterised detector response, etc.) suggest that the target sensitivity (60 SM events with $S/B=1$) is achievable. However, due to the the reduction in the beam solid angle, the $K_{L}$ flux is decreased by 60% when the beamline is extended by 150 m. We are exploring options to increase the acceptance for signal decays. The acceptance for $K_{L}\to\pi^{0}\nu\bar{\nu}$ decays in the fiducial volume is only a few percent due to the cuts used to reject background, primarily from $K_{L}\to\pi^{0}\pi^{0}$ decays, so increasing the signal acceptance is closely related to optimising the background rejection. #### 8.3.1 Signal selection and rejection of $K_{L}$ background The basic criteria for signal selection are as follows: * • Events with exactly two photon clusters on the MEC and no other activity in the detector, including in any of the photon vetoes, are selected. * • Assuming that the two photons have invariant mass equal to $m_{\pi^{0}}$ and that the $K_{L}$ decays on the nominal beam axis ($r=0$), the position of the decay vertex $z_{\rm rec}$ is calculated from the distance between the photon clusters on the MEC. The vertex is required to lie inside the FV. * • The requirement that both photons must have radius $r>r_{\rm min}=35$ cm from the beam axis helps to increase the rejection for events with overlapping clusters, provides a safeguard against background from $K_{L}\to\pi^{0}\pi^{0}$ decays upstream of the AFC, and helps to reject residual $\Lambda\to n\pi^{0}$ decays. * • A specific background topology with two photons at large angle from an odd- paired $K_{L}\to\pi^{0}\pi^{0}$ decaying just upstream of the MEC but falsely reconstructing in the FV is rejected by the requirement that the photon of lower energy have $E_{\rm min}>2~{}{\rm GeV}/r_{E_{\rm\min}}$, where $r_{E_{\rm min}}$ refers to the photon with $E=E_{\rm min}$. * • Signal events are required to have $p_{\perp\,{\rm min}}>140$ MeV, where $p_{\perp}$ refers to the $\pi^{0}$ as reconstructed by the MEC. * • At least one photon is required to convert in the PSD, allowing reconstruction of the vertex position and transverse momentum with relaxed assumptions on the radial position of the decaying $K_{L}$. The vertex position $z_{\rm PSD}$ must be upstream of the downstream edge of the FV and the transverse momentum $p_{\perp\,{\rm PSD}}$ must be greater than $p_{\perp\,{\rm min}}$. Table 8 shows the signal acceptance from the fast simulation. The numbers of events at each stage of selection when the above cuts are sequentially applied are listed, both for the original 120 m and extended 270 m beamline configurations, assuming ${\cal B}(K_{L}\to\pi^{0}\nu\bar{\nu})=3\times 10^{-11}$. Note that for the extended beamline configuration, the FV extends 10 m further downstream than it does for the original configuration. Plots of the distributions of the events satisfying these minimal criteria in the plane of $p_{\perp}$ vs $z_{\rm rec}$ are shown in Fig. 22, for the extended beamline configuration only. The two panels show the distributions for all events with two photons on the MEC, and with the addition of the $r_{\rm min}$, $E_{\rm min}$, and PSD requirements. Two features are apparent. Firstly, the acceptance for signal events increases dramatically in proximity to the calorimeter: the majority of the events with two photons on the MEC actually lie downstream of the FV. Secondly, the $p_{\perp}$ resolution degrades rapidly downstream of the FV, due to the error in $p_{\perp}$ reconstruction resulting from ignoring the radial displacement of the decaying $K_{L}$. Note finally that the PSD cuts partially enforce the definition of the signal box, which explains the shape of the distribution in Fig. 22 (right). For events passing the FV and cluster reconstruction cuts, the $p_{\perp}<140$ MeV selection retains 60% of signal events. Table 8: Estimated sensitivity for $K_{L}\to\pi^{0}\nu\bar{\nu}$. | Stage | Acceptance | Cumulative acceptance | Total events ---|---|---|---|--- 120 m beamline, 60 m fiducial volume (130–190 m) | Produced $K_{L}\to\pi^{0}\nu\bar{\nu}$ | 1 | 1 | 37800 | Decay in FV | 5.93% | 5.93% | 2240 | $2\gamma$ in MEC | 2.34% | 2.34% | 884 | \+ reconstructed in FV | 0.205 | $7.13\times 10^{-3}$ | 269 | \+ $r_{\rm min}>35$ cm | 0.504 | $3.60\times 10^{-3}$ | 136 | \+ $E_{\rm min}$ cut | 0.911 | $3.28\times 10^{-3}$ | 124 | \+ $p_{\perp}>140$ MeV | 0.568 | $1.86\times 10^{-3}$ | $70.4$ | \+ $(z,p_{\perp})_{\rm PSD}$ cuts with $0.5X_{0}$ | 0.469 | $0.874\times 10^{-3}$ | $33.0$ | or with $1.0X_{0}$ | 0.684 | $1.28\times 10^{-3}$ | $48.2$ 270 m beamline, 70 m fiducial volume (280–350 m) | Produced $K_{L}\to\pi^{0}\nu\bar{\nu}$ | 1 | 1 | 15483 | Decay in FV | 5.07% | 5.07% | 785 | $2\gamma$ in MEC | 1.95% | 1.95% | 302 | \+ reconstructed in FV | 0.429 | $8.38\times 10^{-3}$ | 130 | \+ $r_{\rm min}>35$ cm | 0.474 | $3.97\times 10^{-3}$ | 61.5 | \+ $E_{\rm min}$ cut | 0.867 | $3.44\times 10^{-3}$ | 53.3 | \+ $p_{\perp}>140$ MeV | 0.595 | $2.05\times 10^{-3}$ | 31.7 | \+ $(z,p_{\perp})_{\rm PSD}$ cuts with $0.5X_{0}$ | 0.479 | $0.982\times 10^{-3}$ | 15.2 | or with $1.0X_{0}$ | 0.699 | $1.43\times 10^{-3}$ | 22.2 Figure 22: Distribution of $K_{L}\to\pi^{0}\nu\bar{\nu}$ events in the $(z_{\rm rec},p_{\perp})$ plane, from fast MC simulation of the configuration with beamline extension, for all events with two photons on the MEC (left), and after $r_{\rm min}$, $E_{\rm min}$, and preshower cuts (right). In analogy to Table 8, Table 9 shows the number of $K_{L}\to\pi^{0}\pi^{0}$ events surviving the various signal selection cuts. Here, it is useful to distinguish between events with the two detected photons from the same $\pi^{0}$ (even pairings), with the two photons from different $\pi^{0}$’s (odd pairings), and with at least one cluster containing overlapping photons (fused clusters). As is readily seen from the left panels of Fig. 23, the background from $\pi^{0}\pi^{0}$ events from all three categories is concentrated in the proximity of the MEC, even more so than for signal events. This arises from veto inefficiencies for high-angle, low-energy photons from $K_{L}$ decays close to the MEC, in addition to the geometrical effect also present for signal events. The background for odd events extends further upstream into the FV than for the other categories, because of the incorrect assumption that the two photon clusters come from a single $\pi^{0}$. These events are rejected effectively by the PSD. Differently to Fig. 22, the right panels of Fig. 23 show the distributions without the application of the preshower cuts. This is because, in the case of the background, the cuts on $(z,p_{\perp})_{\rm PSD}$ are almost as effective as those on $(z_{\rm rec},p_{\perp})$ from the calorimeter. Table 9: Estimated background from $K_{L}\to\pi^{0}\pi^{0}$. | Stage | Total | Even | Odd | Fused ---|---|---|---|---|--- 120 m beamline, 60 m fiducial volume (130–190 m) | Produced $K_{L}\to\pi^{0}\pi^{0}$ | $1.089\times 10^{12}$ | Decay in FV | $5.93\%=64.5\times 10^{9}$ | $2\gamma$ in MEC | $3.83\times 10^{8}$ | $0.94\times 10^{8}$ | $2.64\times 10^{8}$ | $0.26\times 10^{8}$ | \+ reconstructed in FV | $8.40\times 10^{6}$ | $0.70\times 10^{6}$ | $6.70\times 10^{6}$ | $1.00\times 10^{6}$ | \+ $r_{\rm min},E_{\rm min}$ cuts | $4.34\times 10^{6}$ | $0.33\times 10^{6}$ | $3.84\times 10^{6}$ | $0.18\times 10^{6}$ | \+ $p_{\perp}>140$ MeV | $261\pm 16$ | $154\pm 12$ | $86\pm 9$ | $21\pm 5$ | \+ $(z,p_{\perp})_{\rm PSD}$ cuts with $0.5X_{0}$ | $80\pm 9$ | $61\pm 8$ | $11\pm 3$ | $8\pm 3$ | or with $1.0X_{0}$ | $117\pm 13$ | $89\pm 11$ | $16\pm 5$ | $12\pm 4$ 270 m beamline, 70 m fiducial volume (280–350 m) | Produced $K_{L}\to\pi^{0}\pi^{0}$ | $0.446\times 10^{12}$ | Decay in FV | $5.07\%=22.6\times 10^{9}$ | $2\gamma$ in MEC | $9.45\times 10^{7}$ | $2.02\times 10^{7}$ | $6.71\times 10^{7}$ | $0.72\times 10^{7}$ | \+ reconstructed in FV | $5.12\times 10^{6}$ | $0.48\times 10^{6}$ | $4.07\times 10^{6}$ | $0.58\times 10^{6}$ | \+ $r_{\rm min},E_{\rm min}$ cuts | $21.3\times 10^{5}$ | $2.31\times 10^{5}$ | $18.3\times 10^{5}$ | $0.71\times 10^{5}$ | \+ $p_{\perp}>140$ MeV | $234\pm 15$ | $175\pm 13$ | $48\pm 7$ | $11\pm 3$ | \+ $(z,p_{\perp})_{\rm PSD}$ cuts with $0.5X_{0}$ | $70\pm 8$ | $59\pm 8$ | $7\pm 3$ | $4\pm 2$ | or with $1.0X_{0}$ | $102\pm 12$ | $86\pm 11$ | $10\pm 4$ | $6\pm 3$ Figure 23: Distribution of $K_{L}\to\pi^{0}\pi^{0}$ events in the $(z_{\rm rec},p_{\perp})$ plane, from fast MC simulation of the configuration with beamline extension, for all events with two photons on the MEC (left column), and after $r_{\rm min}$ and $E_{\rm min}$ cuts (right column). The preshower cuts are not applied. From top to bottom, rows show events with the two photons from the same $\pi^{0}$ (even), with the two photos from different $\pi^{0}$’s (odd), and with at least one cluster containing overlapping photons (fused). To ensure that it does not interfere with the detection of low-energy photons by the MEC, the PSD radiator is thin, and the PSD cuts, although effective at suppressing background, penalize the signal acceptance because of the requirement for at least one photon to convert. With a $1X_{0}$ radiator, the acceptance is increased by about 50% with respect to that obtained with a $0.5X_{0}$ radiator, but the optimum thickness will require more detailed simulations to determine. Tables 8 and 9 thus give the final expected numbers of signal and background events in both hypotheses. For the original 120 m beamline configuration, assuming a $1X_{0}$ radiator, Tables 8 and 9 give 48.2 expected signal events and $117\pm 13$ background events from $K_{L}\to\pi^{0}\pi^{0}$. For the case of the extended beamline, 22.2 signal events are expected after accounting for the reduction of the $K_{L}$ flux, and $102\pm 12$ background events, where the increased background results mainly from extending the fiducial volume by an additional 10 m in the downstream direction to partially make up for the reduced $K_{L}$ flux. This scheme is far from final. We are continuing to investigate potential improvements to both the experiment and the analysis to increase the sensitivity, reduce the background, and increase redundancy, including the following: * • Recovering sensitivity for signal events by further extending the FV in the downstream direction. As noted above, this will require improved background rejection. * • Making better use of information from the PSD. In particular, a kinematic fit to combine information from calorimeter and PSD shows promise for improving the $(z,p_{\perp})$ reconstruction in the downstream region, where the angles of incidence are greatest and the $p_{\perp}$ reconstruction is poorest, but where the potential for recovery of signal events is highest. * • Continuing to improve the effectiveness of the photon veto systems, especially in the downstream region. At the moment, we assume that the MEC gives no response for photons with $E<100$ MeV (Section 10.3.3.2). If this threshold can be lowered, which is likely with the design proposed for the MEC, the populations of events in the downstream regions seen in Fig. 23 will be significantly decreased. * • Use of multivariate analysis techniques to efficiently combine the event selection criteria. While efforts have been made in this direction, current attention is on development of the full simulation, which is prerequisite to establishing a definitive search strategy. #### 8.3.2 $\Lambda\to n\pi^{0}$ background and beamline extension As noted in Section 5.3.3, additional measures must be taken to ensure sufficient suppression of background from $\Lambda\to n\pi^{0}$ decays. The most attractive option is to maintain the 8 mrad production angle and increase the length of the beamline from target to AFC by 150 m. Table 10: Estimated background from $\Lambda\to n\pi^{0}$. | Stage | Acceptance | Cumulative acceptance | Total events ---|---|---|---|--- 120 m beamline, 60 m fiducial volume (130–190 m) | Produced $\Lambda\to\ n\pi^{0}$ | 1 | 1 | $2.00\times 10^{14}$ | Decay in FV | $4.70\times 10^{-6}$ | $4.70\times 10^{-6}$ | $9.42\times 10^{8}$ | $2\gamma$ in MEC | $7.86\times 10^{-6}$ | $7.86\times 10^{-6}$ | $1.58\times 10^{9}$ | \+ reconstructed in FV | 0.351 | $2.76\times 10^{-6}$ | $5.53\times 10^{8}$ | \+ $r_{\rm min},E_{\rm min}$ cuts | 0.316 | $8.71\times 10^{-7}$ | $1.75\times 10^{8}$ | \+ $p_{\perp}>140$ MeV | $6.80\times 10^{-3}$ | $5.92\times 10^{-9}$ | ($1.19\pm 0.04)\times 10^{6}$ | \+ $(z,p_{\perp})_{\rm PSD}$ cuts with $0.5X_{0}$ | 0.132 | $7.81\times 10^{-10}$ | $(1.57\pm 0.14)\times 10^{5}$ | or with $1.0X_{0}$ | 0.193 | $1.14\times 10^{-9}$ | $(2.29\pm 0.21)\times 10^{5}$ 270 m beamline, 70 m fiducial volume (280–350 m) | Produced $\Lambda\to n\pi^{0}$ | 1 | 1 | $8.22\times 10^{13}$ | Decay in FV | $2.07\times 10^{-9}$ | $2.07\times 10^{-9}$ | $1.70\times 10^{6}$ | $2\gamma$ in MEC | $2.79\times 10^{-9}$ | $2.79\times 10^{-9}$ | $2.29\times 10^{5}$ | \+ reconstructed in FV | 0.445 | $1.24\times 10^{-9}$ | $1.02\times 10^{5}$ | \+ $r_{\rm min},E_{\min}$ cuts | 0.222 | $2.75\times 10^{-10}$ | $2.26\times 10^{4}$ | \+ $p_{\perp}>140$ MeV | 0.0207 | $5.70\times 10^{-12}$ | $468\pm 54$ | \+ $(z,p_{\perp})_{\rm PSD}$ cuts with $0.5X_{0}$ | 0.0760 | $4.33\times 10^{-13}$ | $35.6\pm 14.0$ | or with $1.0X_{0}$ | 0.111 | $6.32\times 10^{-13}$ | $52.0\pm 20.4$ Figure 24: Distribution of $\Lambda\to n\pi^{0}$ events in the $(z_{\rm rec},p_{\perp})$ plane from fast MC simulation of the configuration with beamline extension, for all events with two photons on the MEC (left), and after $r_{\rm min}$, $E_{\rm min}$, and preshower cuts (right). The effectiveness of this solution is demonstrated by Table 10, from which it is seen that the beamline extension decreases the number of $\Lambda$ decays with two photons on the MEC by a factor of nearly 7000 and the number passing all cuts by a factor of more than 4000, from 229 000 to $52\pm 14$. The centre-of-mass decay momentum of the $\pi^{0}$ in $\Lambda\to n\pi^{0}$ decay is only 104 MeV, and even with the contribution from the neutral beam divergence, the true value of $p_{\perp}$ is always less than 120 MeV, so the $p_{\perp}$ and preshower cuts are particularly effective. The distributions in $(z_{\rm rec},p_{\perp})$ are shown in Fig. 24. Before the $r_{\rm min}$, $E_{\rm min}$, and preshower cuts, most of the $\Lambda$ decays occur at the most upstream end of the FV: the 70 m FV spans more than 3 average decay lengths. The distribution for the 52 surviving events is seen in the right panel. Most of these events can be eliminated using a hadronic calorimeter at small angle backing up the SAC to veto the high-energy neutron from the $\Lambda\to n\pi^{0}$ decay. Because of the mass asymmetry in the decay, the neutrons from decays in the FV that pass analysis cuts are always emitted into the SAC acceptance and have an energy distribution with a mean of 220 GeV; 98% of these neutrons have $E>150$ GeV. Relatively few of the beam neutrons have energies this high (Fig. 11); a hadronic calorimeter with a threshold at $E=150$ GeV would add less than 10 MHz to the total ${\cal O}$(100 MHz) veto rate for the extended configuration. The concept requires full simulation, but it seems reasonable to assume that the residual $\Lambda$ background could be reduced by an order of magnitude or more. #### 8.3.3 Other backgrounds In addition to the backgrounds from $K_{L}\to\pi^{0}\pi^{0}$ and $\Lambda\to n\pi^{0}$, there are a number other background channels that require further advances in the simulation to properly study, and in particular, the full incorporation of the detailed FLUKA beamline simulation into the HIKE Monte Carlo. While mitigation of potential background contributions from these sources might ultimately require specific modifications to the experimental setup, we expect this task to be less complicated than dealing with the primary challenges from $K_{L}\to\pi^{0}\pi^{0}$ and $\Lambda\to n\pi^{0}$. ##### $K_{L}\to 3\pi^{0}$ Background from $K_{L}\to 3\pi^{0}$ is not expected to be significant due to the relative improbability of losing four photons. A preliminary study was performed with the fast simulation for an earlier configuration of the experiment. Essentially all of the selected events have the $K_{L}$ vertex very near the calorimeter; the detected photons have low energy and are concentrated at very small radius. For combinatoric reasons, there are significantly more events reconstructed in the FV with one or more photons lost to cluster fusion than in the case of $K_{L}\to\pi^{0}\pi^{0}$. The fiducial volume cut and $r_{\rm min}$ cuts are highly effective at eliminating these events, and the residual background has very low $p_{\perp}$. This study will be repeated with the updated experimental configuration at a level of statistics sufficient to rule out or evaluate residual contributions to the background. ##### $K_{L}\to\gamma\gamma$ This channel has also been studied by fast simulation. Because the transverse momentum of the $\gamma\gamma$ pair is zero in the rest frame, the $p_{\perp}$ cut used to reject $K_{L}\to\pi^{0}\pi^{0}$ background is extremely effective, and no events are expected from $K_{L}\to\gamma\gamma$ decays in the core of the neutral beam. Any background from this channel must come from the decays of $K_{L}$’s scattering from the collimator surfaces. The KOTO experiment has reported background from $K_{L}\to\gamma\gamma$ at a level corresponding to about 6 events at SM sensitivity [177]. Cluster shape cuts to constrain the angles of photon incidence were helpful in reducing this background. KLEVER has numerous advantages, including a long beamline with four stages of collimation, resulting in beam with an excellent core-to-halo ratio, and reconstruction constraints from the PSD. ##### Decays to charged particles The role of the CPV in rejecting decays with charged particles in the final state is discussed in Section 10.4. An important decay against which the CPV not completely effective is $K_{L}\to\pi^{+}\pi^{-}\pi^{0}$: because of the low decay momentum, it is possible for the $\pi^{+}$ and/or $\pi^{-}$ to escape down the beam pipe. KOTO has resolved this problem by installing additional charged particle vetoes along the walls of the downstream beam pipe, and it is straightforward to adopt the same solution for KLEVER. Radiative semileptonic decays and $K_{e4}$ decays will also require simulation with the full HIKE MC. ##### Beam interactions on collimator surfaces There are a number of potential backgrounds from the inelastic scattering of beam particles from the collimator surfaces. Hyperons and $K_{S}$ mesons can produced in these interactions. Compared to the $\Lambda$ and $K_{S}$ components of the neutral beam, these secondaries have much lower momentum and should decay away if produced on the upstream collimators. Scattering in the AFC is more dangerous, so further beamline optimisation, including the possibility of introducing additional stages of collimation, may be necessary. Based on KOTO’s experience, $K^{0}p\to K^{+}n$ charge exchange is possibly the most important process [177], because the $K^{+}$ can undergo $K_{e3}$ decay inside the FV with the $\pi^{0}$ reconstructed and the $e^{+}$ emitted at very low energy and at very high angle. KOTO is exploring the use of very thin upstream charged particle detectors to reduce this background. In KLEVER, the first three stages of collimation are followed by magnetic sweepers, so this background can only arise from scattering in the AFC, which for this process leaves a signal to veto. ##### Neutron interactions in the MEC Neutron interactions in the calorimeter can create background due to the possibility for hadronic shower fragments to travel a substantial distance in the calorimeter before reinteracting, creating two clusters. The principal strategy for abatement of this background is careful beamline design to reduce the neutral beam halo. This is also the logic behind widening the calorimeter bore so that the beam penumbra intercepts the IRC downstream of the MEC. In addition, the particle-identification capabilties of the calorimeter play an important role. KOTO, which uses a CsI calorimeter, identifies neutron interactions via cluster shape analysis and double-sided light readout. The MEC design permits optimisation of the transverse segmentation to allow cluster shape analysis, while longitudinal shower information is obtained from the use of spy tiles (Section 10.3.3.2). Since the halo is concentrated on the inner regions of the calorimeter, the $r_{\rm min}$ cut is expected to be effective. KLEVER will also make use of information from the HIKE hadron calorimeters to reject hadronic showers. ##### Beam-gas interactions The background from single $\pi^{0}$ production in interactions of beam neutrons on residual gas has been estimated with the FLUKA-based simulation (Section 5.3) to be at most a few percent of the expected signal for a residual gas pressure of $10^{-7}$ mbar. #### 8.3.4 Outlook As discussed in the previous sections, the current simulations give 22.2 expected signal events, $102\pm 12$ expected background events from $K_{L}\to\pi^{0}\pi^{0}$, and $52\pm 20$ expected background events from $\Lambda\to n\pi^{0}$, which can potentially be reduced to the low single digits or eliminated by the introduction of a hadronic calorimeter to back up the SAC. We have identified in-principle solutions for the mitigation of the remaining backgrounds that require further study. The designs of the beamline and experiment may require relatively minor modifications, such as the introduction of additional collimation stages or of small detectors to accomplish specific tasks such as upstream or downstream charged-particle veto, but these are well-defined problems amenable to solution. We also consider the following factors in interpreting the signal and background estimates: * • Random veto will reduce both signal and background by about 10% (Section 8.2). * • The use of an oriented crystal metal photon converter in the TAX collimator will reduce the scattering of the neutral beam by about 20%, leading to an effective increase in $K_{L}$ flux of 10% (Section 5.3.1). Since this involves no significant technical difficulty, we assume that this will cancel the loss from random veto. * • Alternatively, the use of a high-$Z$ target could increase the effective $K_{L}$ flux by up to 30% due to the dramatic reduction in the thickness required for the photon converter. However, we do not take this into account for the moment, as the associated costs and technical difficulties have not been evaluated. With these factors taken into consideration, the signal and background estimates from the simulation remain unchanged. Relative to the sensitivity target of 60 SM events, the apparent shortfall arises principally because of the reduction of the the $K_{L}$ flux a factor of 2.4 with the extended beamline. With the performance gains expected from reoptimisation of the analysis, with particular attention to extension of the FV, better use of the information from the PSD, and continued improvements to the effectiveness of the photon veto systems with emphasis on the calorimeter design, this loss can largely be recovered, so that the original sensitivity goal remains within reach. An effort is underway to develop a comprehensive simulation and to use it to validate the results obtained so far. This work is being carried out within the framework of the flexible HIKE Monte Carlo platform, allowing new configurations of the existing and proposed detectors to be simulated on the basis of the NA62 software. ## 9 Operation in beam-dump mode ### 9.1 Experimental layout The HIKE beam dump operation will build upon the experience accumulated in NA62 with beam-dump data taking. In 2021, $1.4\times 10^{17}$ protons were collected in 10 days of data taking at NA62 in beam-dump mode, with the T10 target used to generate the standard NA62 secondary hadron beam removed from the beamline. Proton beam was made to interact in the NA62 movable collimators, called TAX, located 23 m downstream of T10 within two pair of dipoles, and 80 m upstream of the decay volume. An ad-hoc setting of the dipoles allows a substantial reduction of the rate of muons emitted by pion decays in the proton-induced hadronic showers in the TAX (the so called “halo” muons). Secondary interactions of halo muons can induce background to the searches proposed, and optimisation of the beamline reduces this background by at least one order of magnitude [147]. On-going and completed studies have shown that the residual background is negligible, in particular when searching for two-body decays of new-physics mediators. In this case, the background rejection is improved by requiring that the trajectory of the mediator points to the beam dump. Based on the NA62 experience, it is assumed that the background will remain negligible at HIKE after an increase of 10 to 50 times in statistics with respect to the NA62 beam-dump operation. Therefore sensitivity projections are performed assuming no background limitation. Sensitivity projections are produced assuming $10^{19}$ and $5\times 10^{19}$ POT accumulated in beam-dump mode. Operation at $4\times 10^{13}$ POT for 4.8 s spills is assumed. When compared to the NA62 2021 beam-dump period, this corresponds to an intensity increase by a factor of 8. Assuming the same efficiencies as for the 2021 NA62 data taking, HIKE will be able to collect $10^{19}$ POT in three to four months. The trigger rate foreseen during beam- dump operation is modest, amounting to less than 200 kHz overall. Using a movable dump allows for a quick switch between a kaon-beam and a beam- dump operation, making it possible to perform specific online calibration procedures in kaon-beam mode during beam-dump operation periods. The importance of this aspect was demonstrated in 2021, significantly improving the overall data quality. A careful consideration concerning the cooling of the dump is needed: the intensity foreseen demands a cooling power exceeding the maximum presently available for the NA62 TAX collimators by a factor of four. Cooling and radio-protection considerations are discussed in Section 5. As demonstrated by the 2021 NA62 beam-dump data, frequent and accurate calibration of the beam secondary-emission intensity monitors (BSI) is necessary, to ensure an overall uncertainty in the integrated proton flux at the level of 5%. Monitors based on the present BSI-technology are adequate to withstand the projected beam intensity. Exploratory experimental studies by the CERN beam-group experts show that the goal of achieving a 5% uncertainty in the proton flux measurement is within reach. ### 9.2 Physics sensitivity The projected sensitivity curves of the HIKE experiments in beam-dump mode are evaluated as exclusion plots at 90% CL, obtained in the assumption of zero observed events and negligible expected background. Note that some of the scenarios are also accessible by HIKE operating in kaon mode (Section 3.6). With reference to the portal classification discussed in Ref. [20], the models relevant to the HIKE sensitivity projections are: * BC1 A new single vector state $A^{\prime}$, that interacts with the electromagnetic current through a small coupling constant, $\varepsilon$. In the NA62 beam-dump setup, such dark photon can be produced by proton strahlung and through the decays of secondary mesons ($\pi^{0}$, $\eta^{(\prime)}$, etc.). The dark photon decays to SM particles, as final states with dark matter fields are assumed to be kinematically forbidden. Below 700 MeV, di- lepton decays dominate the $A^{\prime}$ width. The free parameters of the model are $M_{A^{\prime}}$ and $\varepsilon$. The expected sensitivity of HIKE to the dark photon production and decay is shown in Fig. 25. * BC4 A new scalar state $S$, that interacts with the SM Higgs through a small mixing constant, $\theta$. In the beam-dump setup, such dark scalar is mostly produced through flavor-changing neutral-current decays of $B$ mesons ($B\to K^{(\ast)}S$). The effective coupling constant is proportional to the fermion mass. Hence, when scanning the $S$ mass from MeV to GeV and above, each new decay mode kinematically accessible ($e^{+}e^{-}$, $\mu^{+}\mu^{-}$, $\pi^{+}\pi^{-}$, etc.) tends to either saturate or significantly increase the $S$ width. The free parameters of the model are $m_{S}$ and $\theta$. The expected HIKE sensitivity is shown in Fig. 26 (left), where we have used the scalar hadronic decay widths from [165]. * BC6, 7, 8 New fermions $N$, that interact with the SM lepton doublets. These heavy- neutral leptons (HNL) mix with the SM neutrinos through a matrix called $U$. Models BC6, BC7 and BC8 assume the presence of a single HNL of Majorana type and the dominance of electron, $\mu$, or $\tau$ neutrino couplings, respectively. In the beam-dump setup, the production of HNL is dominated by leptonic and semileptonic decays of charmed mesons or $\tau$ leptons, each model favouring the corresponding lepton flavour. The decay width is dominated by meson-lepton or meson-neutrino two-body final states. The free parameters of the BC6,7,8 models are $m_{N}$ and $U_{e,\mu,\tau}$, respectively. The expected HIKE sensitivity curves are obtained with the same assumptions of [178] and are shown in Fig. 27. * BC9, 10, 11 A new pseudoscalar state $a$ (axion-like particle, ALP), that interacts with the SM fields. Models BC9, BC10 and BC11 assume a dominant coupling of an ALP with SM photons, fermions, and gluons, respectively. In BC9, the production in a proton beam dump is dominated by the interaction of photons from the decays of secondary mesons ($\pi^{0}$, $\eta^{(\prime)}$) with the EM fields of the TAX nuclei. The decay width is dominated by the ALP decay to two photons in the whole $m_{a}$ range. The free parameters of the BC9 model are $m_{a}$ and the coupling $C_{\gamma\gamma}$. The expected HIKE sensitivity is shown in Fig. 26 (right). In BC10, the production is dominated by decays of secondary $B$ mesons ($B\to K^{(\ast)}a$). For $m_{a}$ below 700 MeV, the decay width is dominated by decays to di-lepton final states. The free parameters of the BC10 model are $m_{a}$ and the Yukawa coupling $g_{Y}$ to the SM fermionic fields. The expected HIKE sensitivity is shown in Fig. 28 (left). The broad phenomenology of BC11 can reproduce either of the two former models, depending on the ALP mass: one-loop corrections lead to an effective coupling to photons, and an effective coupling to quarks is generated. The former case leads to the ALP di-photon decay; the latter to hadronic decays, e.g. $a\to\pi^{+}\pi^{-}\gamma$, but not to leptonic decays. The expected HIKE sensitivity is shown in Fig. 28 (right). Figure 25: Expected HIKE 90% CL exclusion limits for the BC1 scenario ($A^{\prime}\to\ell^{+}\ell^{-}$). The currently excluded areas of parameter space are shown in grey. Figure 26: 90% CL exclusion limits obtained in simulations using the ALPINIST framework with the selection criteria described in [179]. Current exclusion limits are shown with filled and future projections with empty contours. Left: Search for a dark scalar $S$ in the BC4 scenario. For HIKE, production from $B\to K^{(\ast)}S$ decays and sensitivity to $S\to ee,\mu\mu,\pi\pi,KK$ decays are considered. The $S$ hadronic decay widths derived in [165] have been used. Exclusions from LHCb [168, 169], NA62 Run 1 [67, 142], E137, and CHARM measurements are separately shown. The E137 exclusion is obtained using the ALPINIST framework with data provided in [180]. Right: search for an axion- like particle $a$ in the BC9 scenario. HIKE sensitivity to $a\to\gamma\gamma$ decays is considered. Figure 27: HIKE 90% CL exclusion limits for BC6 (top-left), BC7 (top-centre) and BC8 (bottom) scenarios in the beam-dump mode, assuming sensitivity to two- track final states and NA62 geometrical acceptance. The currently excluded areas of parameter space are shown in grey. Figure 28: Exclusion limits at 90% CL obtained with the ALPINIST framework for an axion-like particle $a$. Current exclusion limits are shown with filled and future projections with empty contours. In the BC10 scenario (left), we assume HIKE sensitivity to ALP production from $B$ meson decays and to the ALP decays $a\to ee$ and $a\to\mu\mu$. In the BC11 scenario (right), we assume HIKE sensitivity to $a\to\gamma\gamma$ decays (dashed contour) and also to the hadronic decays $a\to\pi^{+}\pi^{-}\gamma$, $3\pi$, $2\pi\eta$, etc. (full contour). ## 10 The HIKE detectors The aim is to achieve the same detector performances of NA62 at a factor 4–5 higher beam intensity, with suitable time resolution and rate capability. Describe which detector modifications and readout we need. ### 10.1 Detectors upstream of the fiducial decay volume #### 10.1.1 Cherenkov kaon tagger (KTAG) A differential ring-focusing Cherenkov detector (KTAG), used for $K^{+}$ tagging in the NA62 experiment at a kaon rate of about 40 MHzprovides tagging efficiency above 95$\%$ [175, 181]. The KTAG uses nitrogen radiator gas (to be replaced with hydrogen before LS3) and ring-imaging optics to focus Cherenkov photons onto eight spherical mirrors reflecting them onto eight photodetector arrays. A diaphragm is used so that only Cherenkov light emitted by kaons reaches the photodetectors. Each array is equipped with a matrix of 48 Hamamatsu R7400-U03 and R9880-210 single-anode phototubes (PMTs) with peak quantum efficiency (QE) of 20–40$\%$. The average rate of detected photons per PMT depends on the QE and PMT position and varies between 3–5 MHz, for an effective area of 2.5 cm2. A single-photon time resolution of 300 ps and the mean number of detected photons per kaon of about 20 lead to a 70 ps overall kaon time resolution. For HIKE Phase 1, the KTAG is expected to provide a time resolution of 15–20 ps and a tagging efficiency above 95%. Simulations of the KTAG with hydrogen radiator gas show that a mean number of detected photons per kaon of about 30 is expected; the improvement in the light yield with respect to nitrogen radiator is due to newly built ring-imaging optical components optimised for hydrogen. The expected $K^{+}$ rate in the HIKE beam is about 200 MHz, corresponding to a 10 MHz/cm2 maximum rate of detected photons (including a safety margin) assuming the NA62 values of the QE and acceptance of the photodetection system. The KTAG should be upgraded with new photodetection, front-end and readout systems to satisfy these requirements. Photodetection devices under consideration for the HIKE KTAG detector are micro-channel plate photomultipliers (MCP-PMTs) which are compact devices providing a single- photon time resolution of 50–70 ps with a low dark count rate (below 1 kHz/cm2), able to operate at the expected maximum photon rate of 10 MHz/cm2. Assuming a gain of $10^{6}$, 300 days of runtime and a 50% duty cycle, the above photon rate leads to an integrated anode charge (IAC) of 16 C/cm2. Figure 29: Schematics of a Photonis Planacon square-shaped MCP-PMT array under investigation. Figure 30: Sketches of a single MCP-PMT array (left) and a matrix of four MCP- PMTs (right) to be used to detect photons in each of the eight octants of the HIKE KTAG detector. Figure 31: QE of the Photonis Planacon MCP-PMTs as functions of wavelength for the two photocathode types considered (standard bialkali and Hi-QE Blue). A cutoff at 180 nm corresponds to the minimum wavelength of the Cherenkov photons reaching the photodetection system. A limitation of standard MCP-PMTs is their limited lifetime, caused by a rapid decrease in the QE of the photocathode (PC) with increasing IAC. This ageing of the PC is caused by heavy feedback ions from the residual gas. For standard MCP-PMTs the QE was observed to drop by 50% for IAC of 0.2 C/cm2 [182]. MCP- PMT treatment with atomic layer deposition (ALD) coating increases the lifetime dramatically [183]. The Planacon XP85112 from Photonis, a square- shaped MCP-PMT treated with an improved ALD technique involving two layers, has been demonstrated to reach a lifetime of 33 C/cm2 without any sign of ageing [184]. This represents an improvement in lifetime by a factor of more than 150 with respect to standard MCP-PMTs, and a factor of at least 10 compared to earlier lifetime optimisation techniques. A Photonis Planacon MCP-PMT of 2”$\times$2” size consisting of $8\times 8$ pixel sensors (Fig. 29), treated with two-layer ALD coating, is a viable solution for the photodetection system of the KTAG in high intensity applications. The Cherenkov photons at each photodetector plane can be detected with a matrix of four MCP-PMTs (Fig. 30). The MCP-PMT is available with standard bialkali and Hi-QE Blue photocathode types; the two QE parameterisations provided by the manufacturer are shown in Fig. 31. Simulation results obtained with a geometrical filling factor of 75% and a collection efficiency of 60% show that the 15–20 ps kaon time resolution is achievable. The linearity of MCP-PMTs with rate at high gain is under investigation. However the rate stability can be adjusted by modifying the resistance and capacitance of the MCP coating inside the MCP-PMT. The above results for the kaon time resolution do not include contributions from front-end and readout systems. The fastIC [185] and picoTDC [186] electronics (Section 11.1) being currently developed for other applications are viable options. The FastIC technology provides a 20 ps time resolution and better linearity for signals from SiPM and MCP-PMTs. PicoTDC chips with 64 channels and 12 ps binning can provide a 4 ps time resolution. Both time contributions will only have a marginal effect on the KTAG timing capability. #### 10.1.2 Beam tracker The GigaTracker [187] is the beam tracker used by the NA62 experiment to measure the momentum and time of the beam particles with high momentum, angular and time resolutions. The detector is made by planes of hybrid pixel silicon detectors, with a dipole magnets between the planes to allow the momentum measurement. For NA62 Run 1 (2016–2018) three silicon planes were used, then for NA62 Run 2 (2021–LS3) a fourth plane was added. Each plane is made of a n-in-p planar sensor, with a thickness of $200\text{\,}\mathrm{\SIUnitSymbolMicro m}$ and a sensitive area of $60\times 27~{}\rm{mm^{2}}$, bump-bonded to 10 readout ASICs (TDCPix). The ASICs need to operate with a cooling system: the NA62 GigaTracker used a silicon microchannel cooling plate that represented the first use of this technology in high-energy experiments. In order to minimise the probability of interactions between the beam particles and the detector, the material budget must be kept as low as possible: the total thickness of a single plane is around $500\text{\,}\mathrm{\SIUnitSymbolMicro m}$, namely $0.5\%X_{0}$. The NA62 GigaTracker has achieved very good performances, namely a track time resolution of ${\cal O}(100~{}{\rm ps})$, an angular resolution of 16 $\mu$rad and a momentum resolution of 0.2%. Among all the above characteristics of the detector, the time resolution represents a limiting parameter for its operations in the context of HIKE. With more than four-fold instantaneous intensity increase, the time resolution should scale accordingly, namely from less than 200 ps to below 50 ps for a single hit. The radiation hardness of the new detector needs to be improved by at least a factor four if a replacement approach is to be preserved, or more in case of a single installation for the entire HIKE data taking campaign. Table 11 summarises the comparison between the current NA62 GigaTracker and the new detector needed for HIKE. Table 11: Comparison between the NA62 Gigatracker main characteristics and the requirements for the HIKE beam tracker. One year corresponds to 200 days of beam. | NA62 GigaTracker | New beam tracker ---|---|--- Single hit time resolution | < 200 ps | < 50 ps Track time resolution | < 100 ps | < 25 ps Peak hit rate | $2~{}\rm{MHz/mm^{2}}$ | $8~{}\rm{MHz/mm^{2}}$ Pixel efficiency | > 99 % | > 99 % Peak fluence / 1 year [$10^{14}~{}1~{}\rm{MeV~{}n_{eq}/cm^{2}}$] | 4 | 16 The interest for silicon detectors with fast timing information capable to operate in a high-radiation environment is shared among different particle- physics collaborations, including the LHC experiments for the high luminosity phase of the collider. Following the 2020 Update to the European Strategy for Particle Physics, 4D tracking at high fluences was identified as one of the most urgently needed technologies for future detectors (2021 European Collaboration for Future Accelerators; Detector Research and Development Roadmap DRDT 3.2 and DRDT 3.3). Therefore several promising R&D projects are ongoing and their outcome could be adopted for HIKE. A strong option is the timeSPOT project [188, 189]: which develops a technology called hybrid 3D-trenched pixels in which the pixel electrode geometry is optimised for timing performance. A representation of one cell of a 3D-trench sensor is reported in Fig. 32. As for standard 3D-pixels, the sensors are able to withstand very large irradiation. The project has experimentally demonstrated that sensors with a pitch of $55\text{\,}\mathrm{\SIUnitSymbolMicro m}$ provide a time resolution of $10\text{\,}\mathrm{p}\mathrm{s}$ up to fluence of $2.5\text{\times}{10}^{16}\text{\,}\mathrm{M}\mathrm{e}\mathrm{V}~{}\mathrm{n}_{eq}\mathrm{/}\mathrm{c}\mathrm{m}^{2}$ [190]. The timeSPOT collaboration is planning to extend these tests up to fluences of $1\text{\times}{10}^{17}\text{\,}\mathrm{M}\mathrm{e}\mathrm{V}~{}\mathrm{n}_{eq}\mathrm{/}\mathrm{c}\mathrm{m}^{2}$, that has never been achieved so far. Excellent detection efficiencies were also measured [190] by operating the sensor inclined by an angle of $20\text{\,}\mathrm{\SIUnitSymbolDegree}$ with respect to the beam incidence. As the electrode itself is not sensitive, the inclination ensures that particle always cross some active pixel region. Sensor with a size of $2\times 2$ cm2 can be produced and technical solution like stitching are being explored to produce larger devices. A first prototype of the ASIC, TIMESPOT1, has been realised [191]. The chip contains a pixel matrix of $32\times 32$ elements, with a pixel pitch of $55~{}\rm{\mu m}$. Each pixel in the matrix contains an analogue and a digital circuit, with its TDC to measure the time of the arrival of the signals from the sensor. Digitised signals are then sent to the periphery of the matrix where 8 multiplexed output links send the data out at 1.28 Gb/s speed. Efforts are on-going to scale the size of the ASIC. The above considerations make timeSPOT a viable option for the HIKE beam tracker. Figure 32: Elementary cell of a 3D-trench silicon detector, as designed by the TIMESPOT project [192]. Doping profiles are shown (n++ in red, p– in green, p++ in blue). A) 3D rendering; B) pixel section; C) pixel layout. Other projects involving different technologies, such as monolithic pixel sensors [193] or LGADs [194, 195], are also taken in consideration. Monolithic detectors can be made with very small thickness and excellent time resolutions, however they have a lower radiation tolerance compared to the 3D-trench sensors. #### 10.1.3 Veto counter The introduction of the veto counter detector (VC) surrounding the beamline has helped reduce the upstream background for the $K^{+}\to\pi^{+}\nu\bar{\nu}$ analysis. However, a part of the background remain undetected, passing at a very small angle inside the holes accommodating the beam pipe. The random veto due to the usage of the veto counter is less than 2% at NA62 nominal intensity, but is expected to increase with the beam intensity. When more than one particle is present in the event, the granularity of the current setup does not allow to reliably separate them and distinguish between accidental activity due to halo muons and genuine background candidates. An improved design based on scintillating fibre (SciFi) technology, developed for the SciFi tracking detector at LHCb [196], is proposed to tackle these limitations. The detector will consist of three detection stations: two located in front of the main collimator, separated by a layer of lead acting as a photon converter, and a third one located immediately behind the collimator, see figure 33. Each station needs to extend at least up to $10\text{\,}\mathrm{cm}$ above the beamline and $30\text{\,}\mathrm{cm}$ below the beamline to cover the full range where upstream decays are expected to cross the detection planes. For the same reason, the horizontal coverage needs to extend up to $6\text{\,}\mathrm{cm}$ on either side of the beam centre. A central hole will accommodate the reduced-size beam pipe, allowing it to reach only $2\text{\,}\mathrm{cm}$ from the beam centre. The thin SciFi modules would then be attached to the beam pipe allowing for an increased detection efficiency for interactions of photons and charged particles with the beam pipe. Following these specifications, the active area of the new veto counter will cover a rectangular surface area of $26\times$$78\text{\,}\mathrm{c}\mathrm{m}^{2}$ with a hole at the centre to accommodate the passing beam. With the present state of the technology as used in LHCb, each station is made of scintillating fibres of diameter $250\text{\,}\mathrm{\SIUnitSymbolMicro m}$, read out by silicon photo- multipliers (SiPMs). The fibres are arranged in mats of multiple fibre layers forming a detection plane. Two orthogonal SciFi planes form a single veto counter station. The described detector scheme will provide X-Y reading of charged particles crossing a detection station with a spatial resolution $\sigma_{x,y}\sim$200\text{\,}\mathrm{\SIUnitSymbolMicro m}$$, a time resolution below $200\text{\,}\mathrm{ps}$, and detection efficiency larger than 99$\%$. The number of readout channels for the above design will be about 4000 channels per station. If needed, several fibres could be coupled together into a single SiPM to reduce the number of channels at the expense of some spatial resolution. The thickness of the detection planes can also be modified to increase detection efficiency and improve time resolution. The small granularity and high spatial resolution of each detection plane will provide tracking capabilities by combining information from consecutive stations, further improving the time resolution. Figure 33: Zoomed view of upstream region of HIKE Phase 1 (see Fig. 13), highlighting the location of the veto counter stations. The expected hit rate in the detector is extrapolated from the rates at the nominal NA62 beam intensity, assuming it grows linearly with the beam intensity. The hit rate at four times the NA62 beam intensity is expected to be 30 MHz per detection plane, which is too high for the existing readout electronics of the SciFi detector but can be addressed with the TDC-Felix electronics commissioned in NA62 (Section 11). The front-end electronics must also be adapted for HIKE. The present front-end ASIC used for the LHCb SciFi receives signals from the SiPMs, which are then shaped and digitised by a 2-bit ADC with a sampling rate of 40 MHz. Due to the time resolution requirements for HIKE, new front-end electronics will be used, based on ToT or FADC, similar to the solutions for other detector systems (see for example the KTAG and LAV sections). A different technology option is an update on the design currently used in NA62 using tiles of plastic scintillator. In this case, the use of tiles with smaller height (4 cm in NA62) and the usage of a double plane for each station (a plane with tiles placed horizontally and a plane with tiles placed vertically) will allow to achieve a smaller granularity and separate signals left by multiple particles. An increased acceptance coverage will be achieved by adding tiles also on the left and right of the beam. A rectangular hole of $42\times 78$ mm2 size will allow the beam pipe to pass through the detector. In addition, some tiles could be placed along the beam pipe to increase the detection of particles escaping the beam pipe between stations. Further improvements in upstream background rejection power could be achieved by rearranging the beamline elements in the veto counter region. Combined with the proposed hardware improvements, the new veto counter detector will reduce the upstream background at least three times more than the present system. #### 10.1.4 ANTI-0 The baseline option for the charged particle veto upstream of the decay volume (very upstream ANTI station called ANTI-0) is a cell structure hodoscope covering area of $\oslash$2.2 m around the beampipe at the entrance of the fiducial decay volume. The NA62 ANTI-0 hodoscope design is based on rectangular scintillating tiles with transverse size of $120\times 120$ mm2 read by SiPMs through short lightguides arranged in a chessboard style at both sides of the central foil. The general design of ANTI-0 for HIKE follows that of the NA62 ANTI-0 hodoscope [197] with the two major changes: finer granularity to sustain up to x6 intensity, and two active layers to ensure higher efficiency. An additional improvement, currently under investigation, is the possibility to use reabsorption-free fast nanocrystal-based plastic scintillators [198]. The new scintillator already now can sustain a higher radiation dose, having an emission peak shifted to $\lambda>550$ nm, tolerating transparency losses in UV and blue regions. Some perovskite-based scintillators show very fast decay time, with the first time component of the order of $\sim 0.3$ ns [199], which could be crucial for building fast future detectors. These technological possibilities will be further investigated for the HIKE Proposal. More details can be found in the description of the MEC. ### 10.2 Fiducial decay volume and its detectors #### 10.2.1 Charged anti-coincidence detector The Charged Anti-coincidence detector (CHANTI) provides veto for events with particles scattered inelastically off the last station of the beam tracker and halo particles originating from upstream decays entering the fiducial decay region close to the beam. The NA62 CHANTI comprises six stations, each station covering an area of $30\times 30$ cm2 with a $9\times$5\text{\,}\mathrm{c}\mathrm{m}^{2}$$ inner opening for the beam. The CHANTI covers hermetically the angular region between $34\text{\,}\mathrm{m}\mathrm{r}\mathrm{a}\mathrm{d}$ and $1.38\text{\,}\mathrm{r}\mathrm{a}\mathrm{d}$ with respect to the beam axis. The CHANTI, with its time resolution of about $800\text{\,}\mathrm{p}\mathrm{s}$, induces a signal loss due to accidental activity of about 5$\%$ at the nominal intensity but is expected to become more significant with increasing intensity. The design of the HIKE CHANTI stations is based on the usage of thick plastic scintillating fibres (from $\oslash 2$ up to $\oslash 3$ mm) read by SiPMs from both ends. The expected particle rate has been estimated using the NA62 beamline simulation, assuming an hadron beam intensity six times higher than the nominal 750MHz of NA62. The expected rate per channel in kHz is shown in Fig. 34 for the option with $\oslash 2$ mm thick fibres overlapping by 0.5 mm. The option with $\oslash 2$ mm fibres, that is the maximum diameter currently widely available on the market, assumes 464 fibres per station (208 forvertically oriented X station, and 256 for horizontally oriented Y station) and about 1000 SiPMs/station. The required number of electronic channels could be reduced by connecting signals from the peripheral fibres into analogue OR before digitization. ##### 10.2.1.1 Alternative detector design using SciFi technology A more ambitious option would be to pursue the SciFi detector technology rather than scintillator technology. The new detector would consists of six stations with $39\times$$39\text{\,}\mathrm{c}\mathrm{m}^{2}$ transverse dimensions, positioned to hermetically cover the required angular region with respect to the beam axis. Each station is made of scintillating fibres ($\oslash$250\text{\,}\mathrm{\SIUnitSymbolMicro m}$$) arranged in two planes that provide X-Y information about the incoming charged particles. Each plane is made of mats of several fibre layers, which are read out by SiPMs. The layout allows a time resolution below $200\text{\,}\mathrm{p}\mathrm{s}$ per station and a spatial resolution $\sigma_{x,y}\sim$200\text{\,}\mathrm{\SIUnitSymbolMicro m}$$. The time resolution would be at least a four-fold improvement, and the exceptional spatial resolution would provide tracking capabilities. The stations would be thicker than the ones currently used at LHCb [196], bringing the detection efficiency to above 99$\%$. The new CHANTI detector could also be used in conjunction with the new VetoCounter (Section 10.1.3), both made of SciFi technology, to separate halo muons from charged particles produced in interactions in the beam-tracker. Owing to their identical spatial and time resolutions and largely overlapping sensitive regions, a combination of the information between the two detectors can reduce significantly the accidental veto due to halo-muon activity. Using a SciFi detector for the new CHANTI presents a technological challenge. At present detectors using SciFi technology are operated in air and a dedicated R$\&$D program is required to enable their operation in vacuum. A solution to this challenge will allow cooling of the SiPM to liquid Nitrogen temperatures, eliminating the noise completely and improving radiation tolerance. This is an important improvement over the present CHANTI, which uses SiPMs without cooling. The SiPMs currently in operation at NA62 are heavily impacted by the effect of accumulated radiation dose; such issue would not be present for a SciFi detector with cooled SiPMs in vacuum. The number of readout channels for the above design would be about 3000 channels per station, but it can be reduced if several fibres are coupled together and fed into a single SiPM. This would be at the expense of some spatial resolution, and would not be an issue. The readout is planned to be based on the TDC-Felix electronics 11, that is able to sustain the hit rate presented in Fig. 34. Figure 34: CHANTI expected rate per fibre for the vertical and horizontal planes. #### 10.2.2 Large angle vetos As demonstrated by the successful $K^{+}\to\pi^{+}\nu\bar{\nu}$ measurements at NA62, the efficiency of the existing photon veto systems, including the existing Large Angle Vetoes (LAV) [175] based on the OPAL lead glass [200], is adequate for the HIKE $K^{+}\to\pi^{+}\nu\bar{\nu}$ measurement. In the analysis of NA62 Run 1 data, the overall $\pi^{0}$ detection inefficiency was about $1\times 10^{-8}$ [201], and the expected number of background events from $K^{+}\to\pi^{+}\pi^{0}$ (the only significant background channel with photons) was about 7% of the expected number of signal events [67]. The $\pi^{0}$ veto efficiency estimate was validated by studies of the single- photon detection efficiency for each veto subsystem [142]. The inefficiency of the NA62 LAVs for single photons was found to rapidly decrease with photon energy to a value of about $3\times 10^{-3}$ for photons with $E=300$ MeV, and thereafter to slowly decrease with energy to values of a few $10^{-4}$ for photons of $4$–$6$ GeV.111This refers to the LAV system in operation, including the effects of the non-hermeticity of the retrofitted arrangement of lead-glass blocks. The actual detection efficiency for the lead-glass detectors themselves was found with electrons to be an order of magnitude higher [202]. In the decay-in-flight technique for the $K^{+}\to\pi^{+}\nu\bar{\nu}$ measurement, the $K^{+}$/$\pi^{+}$ vertex is accurately reconstructed and required to be in the fiducial volume, and the fiducial momentum cuts on the secondary $\pi^{+}$ guarantee that the $\pi^{0}$ in $K^{+}\to\pi^{+}\pi^{0}$ decays has at least 40 GeV of energy. As a result, the LAV system only has to cover out to 50 mrad in the polar angle as seen from the fiducial volume. Moreover, the two photons from $K^{+}\to\pi^{+}\pi^{0}$ decays have an anticorrelation such that for events with one low-energy photon heading into the LAVs, the second photon has high energy and is seen in the LKr. These conditions allow the very low detection inefficiency for $\pi^{0}$’s to be obtained even with LAVs with single-photon inefficiencies on the order of $10^{-3}$. On the other hand, the time resolution obtained with the existing LAVs is problematic for HIKE. The Cherenkov light produced in the large lead-glass blocks is highly directional, and the light propagates to the PMT photocathodes via complicated paths with multiple reflections. There is considerable time spread arising from the angle of particle incidence, and perhaps for different particle species, the Cherenkov characteristics of mips and electromagnetic showers being quite different. Above all, the light yield of the lead-glass blocks for electromagnetic showers is just 0.2 p.e./MeV. As a result, the LAVs have a time resolution of about 1 ns with significant extra-gaussian tails [175]. In order to maintain the random veto rate constant in HIKE, the time resolution needs to be improved by at least a factor of 4, and the tails substantially reduced or eliminated. The LAVs are potentially the most resource-intensive subsystem to build for HIKE, so it is essential that they be reusable for all phases of the program. The design of the LAVs for HIKE is therefore driven by the efficiency requirements for the $K_{L}$ program, and in particular, for the measurement of $K_{L}\to\pi^{0}\nu\bar{\nu}$. The time resolution requirements for the $K_{L}$ phase are similar to those for the $K^{+}$ phase. The preliminary estimate of the total hit rate on the LAVs in KLEVER is 14 MHz (Table 7), implying a 3.5% random veto rate for a $\pm 5\sigma$ coincidence window of 2.5 ns. The efficiency requirements for the $K_{L}$ phase, on the other hand, are much more stringent than for the $K^{+}$ phase. Unlike for the case of $K^{+}\to\pi^{+}\nu\bar{\nu}$, for $K_{L}\to\pi^{0}\nu\bar{\nu}$ decays, the reconstruction of the $K_{L}$ decay vertex is uncertain, so the photon vetoes need to cover the entire length of the experiment from the upstream end of the fiducial volume to the calorimeter. Moreover, the kinematic anticorrelation between the energies (or angles) of missed photons from $K_{L}\to\pi^{0}\pi^{0}$ decays is largely washed out. Nevertheless, because of the boost from the high-energy beam, the KLEVER simulations show that it is sufficient for the large-angle photon vetoes (LAVs) to cover polar angles out to 100 mrad, as long as the fiducial volume is located well upstream of the calorimeter and the LAV detectors themselves satisfy the efficiency requirements discussed below: indicatively, the photon detection inefficiency must be less than 5% at 10 MeV, less than $2.5\times 10^{-4}$ at 100 MeV, and less than $2.5\times 10^{-6}$ for energies above 2.5 GeV. ##### 10.2.2.1 Layout for the $K^{+}$ phase Table 12: Parameters of the existing NA62 Large Angle Vetoes. Stations | Diameter [mm] | Block radius [mm] | Layers | Blocks ---|---|---|---|--- | Outer wall | Inner | Outer | | LAV1–LAV5 | 2168 | 537 | 907 | 5 | 160 LAV6–LAV8 | 2662 | 767 | 1137 | 5 | 240 LAV9–LAV11 | 3060 | 980 | 1350 | 4 | 240 LAV12 | 3320 | 1070 | 1440 | 4 | 256 The positions and radii of the LAV stations for the HIKE $K^{+}$ phase are expected to be quite similar to those for NA62. In NA62, there are 12 LAV stations: 11 incorporated in the vacuum tank, and one in air just upstream of the calorimeter. Parameters of the NA62 LAV stations are listed in Table 12. ##### 10.2.2.2 Layout for KLEVER For KLEVER, a total of 25 LAV stations in five different sizes, operated in vacuum and placed at intervals of 4 to 6 m, are needed to guarantee coverage out to $\theta=100$ mrad. The most downstream LAV station (LAV25) leaves as much of the front face of the Main Electromagnetic Calorimeter (MEC) uncovered as possible, and the dimensions and positions of the other LAVs have been chosen so that LAV25 defines the most restrictive aperture for photons from the fiducial volume. The dimensions and segmentation of the 25 LAV stations are specified in Table 13. Table 13: Dimensions, segmentation, number of readout channels, and total quantity of plastic scintillator for the five different types of LAV stations. LAVs | $r_{\rm int}$ (m) | $r_{\rm ext}$ (m) | Sectors | Total channels | Tot. scint. (kg) ---|---|---|---|---|--- 1–11 | 0.44 | 0.85 | 40 | 3520 | 9690 12–15 | 0.58 | 0.99 | 48 | 1536 | 4290 16–18 | 0.72 | 1.23 | 56 | 1344 | 4970 19–21 | 0.86 | 1.37 | 64 | 1536 | 5680 22–25 | 1.00 | 1.51 | 72 | 2304 | 8525 As noted above, 11 of the LAV stations will be constructed for the HIKE Phase 1. The 12th station, operated in air rather than in vacuum, might be used for Phase 1 only. The remaining 14 stations in vacuum will be constructed for the HIKE Phase 2, as needed, and for KLEVER. The sizes and positions of the LAVs were implemented in the KLEVER simulation before the HIKE $K^{+}$ phase was proposed. Because the polar angle coverage was extended by increasing the number of LAV stations rather than their dimensions, the dimensional parameters of the KLEVER LAVs, though not the same as for NA62, are quite similar. Although the LAVs must be positioned carefully in KLEVER to guarantee the needed coverage, with some careful engineering work, it should be possible to reconcile the LAV designs for the HIKE $K^{+}$ and $K_{L}$ phases. For example, from the comparison of Tables 12 and 13, it seems likely that LAVs 1–5 for the $K^{+}$ program could be recycled as LAVs 11–15 for $K_{L}$, LAVs 6–8 for $K^{+}$ would become LAVs 16–18 for $K_{L}$, and LAVs 9–11 for $K^{+}$ would become LAVs 19–21 for $K_{L}$. Various sections of the NA62 vacuum tank would have to be modified or remade. ##### 10.2.2.3 Design and expected performance As a reference for what low-energy photon detection efficiencies can be achieved for the new LAVs, Fig. 35 shows the inefficiency parameterization used for the KOPIO proposal [203]. Figure 35: Photon detection inefficiency parameterization from KOPIO [203], broken down by source. Measured inefficiencies for the E949 barrel veto [204] are also plotted. For the energy range $50~{}{\rm MeV}<E<170~{}{\rm MeV}$, the overall parameterization (black circles) is based on detection inefficiencies measured for the E787/949 barrel photon veto [204] using $K^{+}\to\pi^{+}\pi^{0}$ events; these data are also shown in the figure. Outside of this range, the parameterization is guided by FLUKA simulations with different detector designs, and the overall result is (slightly) adjusted to reflect the segmentation of the KOPIO shashlyk calorimeter, but for most of the interval $E<200$ MeV, the results do not differ much from the E949 measurements. The contributions to the inefficiency from photonuclear interactions, sampling fluctuations, and punch through were estimated from known cross sections, statistical considerations, and mass attenuation coefficients. One possible design for the HIKE LAVs would be similar to the Vacuum Veto System (VVS) detectors planned for the CKM experiment at Fermilab [205]. The CKM VVS is a lead/scintillator-tile detector with a segmentation of 1 mm Pb + 5 mm scintillator, for an electromagnetic sampling fraction of 36%. This segmentation is the same as for the E787/949 barrel photon veto, so the same low-energy efficiencies might be expected. The wedge-shaped tiles are stacked into modules and arranged to form a ring-shaped detector. The scintillation light is collected and transported by 1-mm-diameter WLS fibers in radial grooves, as seen in Fig. 36. Figure 36: Prototype tile with WLS fiber readout for the CKM VVS detector [205]. In the approximate geometry for HIKE, the LAV modules would consist of 96 layers, for a total thickness of $\sim$60 cm, corresponding to $\sim$18 $X_{0}$. There are 40–72 modules per detector, each with 20 fibers per tile. In the original VVS design, the fibers brought the light to optical windows for readout by PMTs outside of the vacuum. Readout by SiPMs inside the vacuum would make for shorter fibers and would facilitate the mechanical design—the availability of economical SiPM arrays with large effective area makes this an attractive option. In the HIKE geometry, the fibers in a module would be bundled for readout by eight SiPM arrays. As in the VVS design, alternating fibers from each tile would be read by different SiPMs to provide redundancy, and there would be four readout layers in depth. Figure 37: Measurements of detection inefficiency for tagged electrons with 203, 350, 483 MeV energies for three veto prototypes, made at the Frascati BTF in 2007. The red squares are for the CKM VVS prototype. In the original NA62 proposal [206], before the OPAL lead glass became available, very similar detectors were the baseline solution for the existing LAVs, and in 2007, the efficiency of the CKM VVS prototype in the energy range 200–500 MeV was measured using a tagged electron beam at the Frascati Beam- Test Facility (BTF) [202]. The results are shown in Fig. 37. Earlier, the efficiency of the same prototype was measured at the Jefferson National Laboratory with tagged electrons in the interval 500–1200 MeV. An inefficiency of $3\times 10^{-6}$ was found at 1200 MeV, even with a high threshold (80 MeV, or 1 mip) [207]. The efficiency parameterization for the LAVs used in our simulations is based on the KOPIO efficiencies up to the point at 129 MeV, and then extrapolated through the three points measured at the BTF, to $2.5\times 10^{-6}$ for photons with $E>2.5$ GeV. This is not unreasonable, considering that the Jefferson Lab measurement shows that nearly this inefficiency is already obtained at 1.2 GeV. The tests of this prototype at the BTF were not optimized for the measurement of the time resolution, but indicated a time resolution of better than 250 ps for 500 MeV electrons, which should be sufficient for HIKE, including KLEVER. ##### 10.2.2.4 Readout A leading contribution to the random veto inefficiency for the LAVs in NA62 is from halo muons. If the interactions of halo muons could be reliably distinguished from photon showers, the random veto inefficiency for the LAVs would be decreased, potentially relaxing the requirements on the LAV time resolution. In NA62, the LAV signals are discriminated against two thresholds, and signal amplitudes are obtained using the time over threshold technique. While some use of this information has been made in NA62 to partially recover the LAV random veto inefficiency and efforts are continuing in this direction, in practice, the separation between mips and low-energy photon showers has not so far proved to be reliable. To improve upon this situation, we are investigating the gains to be had with a fully digitizing FADC readout for the LAVs, as is foreseen for all of the photon veto detectors in KLEVER. #### 10.2.3 Spectrometer A spectrometer similar to the NA62 straw tracker [175], comprising four straw chambers and a dipole magnet, is planned to reconstruct the momentum and direction of charged particles in the final state. The straw tubes are expected to remain in the vacuum tank containing the decay region, profiting from the successful technology developed for NA62. Straws with diameter less than 5 mm are necessary to handle the expected particle rates at the higher beam intensity. Straw diameter reduction by a factor $\sim 2$ with respect to the NA62 straws will lead to shorter drift times and an improvement in the resolution of the trailing edge time from the current ${\sim}30$ ns to ${\sim}6$ ns. A smaller diameter of the straw also requires a change in the geometric placement of the straws in a single view. Design work based on Monte Carlo simulations was performed, and the straw layout was optimised taking into account realistic spacing and dimension requirements, resulting in a choice of eight straw layers per view, shown in Fig. 38. The material used to make new straws using the ultrasonic welding technique will be the same as in the current spectrometer, namely Mylar coated with 50 nm of copper and 20 nm of gold on the inside. To reduce the detector material budget, the thickness of the mylar will be reduced from $36~{}\mu$m to either $12~{}\mu$m or $19~{}\mu$m. The diameter of the gold-plated tungsten anode wires might be reduced from $30~{}\mu$m to $20~{}\mu$m. The final decision on the mylar thickness and the wire diameter will be made based on mechanical stability tests. The development of small-diameter thin-walled straws has synergies with R&D work for COMET phase II at J-PARC [208], and is included in the ECFA detector R&D roadmap [209]. Based on the results of the design study, a Geant4-based simulation of the new spectrometer was developed using the same dimensions and positions of the straw chambers, the number and orientation of views in the chamber, the gas composition (Ar+CO2 with 70:30 ratio) and the properties of the dipole magnets as in the current NA62 layout. A comparison between the two straw detectors is given in Table 14, and a Geant4 visualisation of the new spectrometer is shown in Fig. 39. The new spectrometer is planned to have the capability of aligning the central holes of the straw chambers on the beam axis for both $K^{+}$ (Section 6) and $K_{L}$ (Section 7) modes of operation. The NA62 track reconstruction algorithm was adapted for the new detector, and a preliminary resolution comparison indicates that the new spectrometer could improve the resolution for the reconstructed track angles and momenta by 10–20% with respect to the existing NA62 spectrometer while maintaining the high track reconstruction efficiency. Investigations of possible technological solutions for the straw connectivity, design of a new high-voltage board, and a pre-production of straw tubes with a diameter of 4.82 mm have already been started (Figs. 42, 43, 44). Figure 38: Optimised layout of straw tubes with a diameter of 4.82 mm in a view. Table 14: Comparison of the NA62 spectrometer and the new straw spectrometer. | Current NA62 spectrometer | New straw spectrometer ---|---|--- Straw diameter | 9.82 mm | 4.82 mm Straw length | 2100 mm | 2100 mm Planes per view | 4 | 8 Straws per plane | 112 | ${\sim}160$ Straws per chamber | 1792 | ${\sim}5200$ Mylar thickness | $36~{}\upmu$m | (12 or 19) $\upmu$m Anode wire diameter | 30 $\upmu$m | (20 or 30) $\upmu$m Total material budget | 1.7% $X_{0}$ | (1.0 – 1.5)% $X_{0}$ Maximum drift time | ${\sim}150~{}$ns | ${\sim}80~{}$ns Hit leading time resolution | (3 – 4) ns | (1 – 4) ns Hit trailing time resolution | ${\sim}30~{}$ns | ${\sim}6~{}$ns Average number of hits hits per view | 2.2 | 3.1 Figure 39: Geant4 visualisation of the new straw spectrometer. Figure 40: Geant4 visualisation of a new straw chamber: (left) front view; (right) tilted back view. Figure 41: Preliminary comparison of resolutions of track momentum (left) and track $\theta_{X}$ angle (right) between the existing NA62 spectrometer (blue) and the new spectrometer with $12~{}\mu$m mylar thickness (red). A similar improvement is observed in the reconstructed $\theta_{Y}$ angle. Figure 42: Left: detail of the connectivity elements (coaxial) in order to minimise the signal path to the front-end electronics. Right: cross-section of the prototype to validate connectivity and basic performance of a 4.82 mm diameter straw using new front-end electronics with the capability to measure the trailing edge. Figure 43: Left: pre-production of straws with a diameter of 4.82 mm and a wall thickness of 19 $\mu$m. Right: test of signal connectivity and high- voltage stability of individual components. Figure 44: Prototype of a HV board for the new straw spectrometer. ### 10.3 Detectors downstream of the fiducial decay volume #### 10.3.1 The RICH detector The NA62 RICH detector, which uses neon at atmospheric pressure as the radiator, is well suited for operation within the HIKE programme in terms of the mechanical structure (vessel, mirror support, end-caps). Major changes should only concern the Cherenkov light sensors and the two flanges hosting them. Improvement of the geometrical acceptance for negative particles is also being considered. ##### 10.3.1.1 New photodetectors The present NA62 RICH is equipped with Hamamatsu R7400-U03 phototubes with a time resolution of 240 ps for single photons, quantum efficiency (QE) of $\sim$20$\%$ and with a distance between the centres of adjacent sensors of 18 mm [210]. This distance, constrained by the sensor size, gives the main contribution to the resolution on the single hit position ($\sim$4.7 mm) and consequently to the overall resolution on the ring radius ($\sim$1.5 mm), the main parameter driving the performance on the particle identification (PID) of the RICH detector [211]. The single hit time resolution together with the number of hits associated to each ring (which depends on the QE) determines the overall time resolution of the RICH for positive tracks of 80 ps on average. The main requirement for RICH operation at HIKE Phase 1 is a ring time resolution of 20–30 ps, in order to reduce the coincidence window with the upstream detectors (beam tracker and KTAG). The RICH is the only downstream detector that could reach such resolution for the kaon decay products. Besides, a reduction of the sensor size would improve the particle identification performance. Silicon photomultiplier (SiPMs) meet these requirements: SiPMs with 100 ps time resolution and QE of 40% or above are already available from the main manufacturers. Considering a similar active area geometrical filling factor (80%) and a similar average number of hits per ring to the NA62 case, and taking into account a factor 2 improvement in the QE, we have evaluated the number of hits per ring and the track time resolution with the new configuration. The time resolution for pion momenta of 15 and 45 GeV/$c$ (the limits of the RICH working region) for the NA62 RICH, and those expected for the future RICH instrumented with SiPMs, are listed in Table 15. The latter meets the HIKE requirements. Table 15: Comparison of the time resolution of the NA62 RICH and the new RICH. | NA62 RICH | Future RICH ---|---|--- Sensor type | PMT | SiPM Sensor time resolution | 240 ps | 100 ps Sensor quantum efficiency | 20% | 40% Number of hit for $\pi^{+}$ at 15 GeV/$c$ | 7 | 14 Number of hit for $\pi^{+}$ at 45 GeV/$c$ | 12 | 24 Time resolution for $\pi^{+}$ at 15 GeV/$c$ | 90 ps | 27 ps Time resolution for $\pi^{+}$ at 45 GeV/$c$ | 70 ps | 20 ps Replacement of the light sensors, necessary to improve the time resolution, also represents an opportunity to improve the RICH performance in terms of particle identification. A smaller sensor size and an improved QE will allow to establish a more optimal RICH working point, improving both the muon rejection and pion identification efficiency in the $K^{+}\to\pi^{+}\nu\bar{\nu}$ analysis. Fig. 45 illustrates the improvement in the ring radius resolution $\sigma_{\rm Radius}$ achieved by sensor size reduction and QE improvement. For a small SiPM size, the contributions to the single hit resolution coming from misalignment of the reflecting mirrors and neon (radiator) dispersion, 0.6 mm and 2.1 mm respectively, become dominant. Figure 45: Layout of the present RICH NA62 sensors compared to SiPM candidates (in scale) and the corresponding performance in terms of the ring radius resolution. In considering SiPMs as photodetector candidates for the HIKE RICH, the effects of the main drawbacks related to these sensors must be evaluated: * - Dark count: for a 800 ps coincidence window, and an annulus area of $7\times 10^{4}~{}{\rm mm}^{2}$ considered for evaluation of the possible particle identification hypotheses, a dark count rate of several kHz/mm2 would produce a non-negligible number of spurious hits. To lower the contamination to the level of few percent, the SiPM should be operated at low temperature. A possible layout of the flanges housing the SiPM and the cooling system is discussed in Section 10.3.1.2. * - Cross-talk is strongly dependent on the SiPM type, can be reduced by cooling the SiPM, and will be considered for the choice of the final sensor. Cross- talk leads to an extra contribution (for a small fraction of events) to the hit space resolution, and consequently to the ring radius resolution. * - Ageing/radiation hardness: the RICH flanges are located at a distance of 1.5 m from the beam pipe and are not traversed by the bulk of the particle flux. Nevertheless the radiation level in the high intensity environment should be investigated. Cooling of the SiPMs would reduce the effects of radiation damage. Alternative photodetectors are the MCP-PMTs considered for the HIKE KTAG detector and described in Section 10.1.1. ##### 10.3.1.2 New mechanics The NA62 RICH vessel, end-caps and the mirror support panel will not be changed in the HIKE configuration, apart for minor modifications. Concerning the mirrors, a replacement or a re-aluminisation should be considered to recover from the deterioration of the coating layer already observed in 2014 during their installation. The new reflecting surface must match, in terms of reflectivity as a function of light wavelength, the quantum efficiency of the photodetectors, so both the reflective and coating layers will be chosen accordingly. The change of sensors will request modification of the two flanges hosting them. This represents an opportunity to increase the RICH acceptance for negatively-charged particles (note that the NA62 RICH is optimised for the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay). Simulations show that increasing the instrumented area from 5700 cm2 to 7600 cm2 would lead to a good acceptance for negative tracks. The end-cap region that can be instrumented without compromising the vacuum-proof mechanics is shown in Fig. 46. In the redesign of the new flanges, a photosensor cooling system will be introduced, which would avoid inducing a temperature gradient in the RICH radiator gas. An adequate system to guarantee thermal insulation between the sensor flanges and the vessel must be considered. Figure 46: The upstream RICH end-cap after its construction, in 2013. The circular hole hosts the sensor flange in the NA62 RICH. The red contour delimits the region that can be instrumented with new photo-sensors. #### 10.3.2 The timing detector The timing detector will allow to have a redundant and complementary measurement of the transit time of charged particles in the downstream detector region with respect to the RICH. The information of the timing plane can be used for efficiency studies, trigger purposes and to veto accidental activity. The timing detector can also be used to have more than one time measurement for pions and muons with momentum below the Cherenkov thresholds in the RICH of 12 and 8 GeV/$c$, respectively. The requested time resolution is about 100 ps. For the construction of this detector HIKE will profit from the expertise gained by NA62 with the NA48-CHOD and NA62-CHOD detectors. The NA48-CHOD, consisting of two planes of scintillator slabs of approximately $100\times 6$ cm2 size, cannot be used in the HIKE environment; the high particle rate (and the high probability of more than one track hitting the same counter) will not allow to correct the measured time for the light propagation time in the scintillator to the photo-sensors depending on the track impact point. On the contrary, the NA62-CHOD layout (Fig. 47), consisting of tiles of approximately $10\times 13$ cm2 (in the central region) will be suitable for HIKE. The expected maximum rate per tile of 2.8 MHz (inferred from experience with NA62 data taking) is affordable and can be lowered further by reducing the dimension of the tiles near the beam pipe, where the particle rate is higher. Figure 47: Layout of the present NA62 CHOD in the $xy$ and $xz$ planes. The NA62-CHOD time resolution (500 ps) is limited by the light transportation outside the detector acceptance via wavelength shifting fibers. In the HIKE setup, the photosensors will be attached directly to the tiles to avoid the degradation. In this layout the sensors will be irradiated by the downstream particle flux. The photosensor type will be chosen profiting from the studies foreseen for other detectors (KTAG, RICH, HCAL, MUV). To further improve the time resolution, an additional detector plane can be considered. #### 10.3.3 The HIKE electromagnetic calorimeter The forward electromagnetic calorimeter is crucial for any kaon program. In particular, in the HIKE $K^{+}$ phase this detector serves as the principal photon veto for the measurement of $K^{+}\to\pi^{+}\nu\bar{\nu}$ and is essential for the reconstruction of the final states for decays like $K_{\ell 2}$, $K^{+}\to\pi^{+}\ell^{+}\ell^{-}$, and $K^{+}\to\pi^{+}\gamma\gamma$. For the neutral beam configurations, the electromagnetic calorimeter reconstructs the $\pi^{0}$ vertex for $K_{L}\to\pi^{0}\nu\bar{\nu}$ events and helps to reject events with extra photons. Many of the same issues arise in the design of the electromagnetic calorimeter for the $K^{+}$ and $K_{L}$ phases. We therefore seek a design for a fast calorimeter with excellent photon detection efficiency and energy resolution to be used in all phases of the HIKE program. While the new calorimeter is investigated and built, the existing NA62 LKr calorimeter is presented as a possible solution for the $K^{+}$ program since it could meet the performance requirements. ##### Performance requirements The principal performance requirements for the electromagnetic calorimeter are excellent energy resolution and intrinsic detection efficiency for high-energy photons, good two-cluster separation for photons, and excellent time resolution. It is natural to inquire as to whether the liquid-krypton calorimeter (LKr) [212] currently used in NA62 can be reused for HIKE. The energy, position, and time resolution of the LKr calorimeter were measured in NA48 to be $\displaystyle\frac{\sigma_{E}}{E}$ $\displaystyle=0.0042\oplus\frac{0.032}{\sqrt{E{\rm(GeV)}}}\oplus\frac{0.09}{E{\rm(GeV)}},$ (6) $\displaystyle\sigma_{x,y}$ $\displaystyle=0.06~{}{\rm cm}\oplus\frac{0.42~{}{\rm cm}}{\sqrt{E{\rm(GeV)}}},$ (7) $\displaystyle\sigma_{t}$ $\displaystyle=\frac{2.5~{}{\rm ns}}{\sqrt{E{\rm(GeV)}}}.$ (8) Indeed, the efficiency and energy resolution of the LKr appear to be satisfactory for all phases of HIKE. Studies of $K^{+}\to\pi^{+}\pi^{0}$ decays in NA48 data and tests conducted in 2006 with tagged photons from an electron beam confirmed that the LKr has an inefficiency of less than $10^{-5}$ for photons with $E>10$ GeV, providing the needed rejection for forward photons [213]. These studies were fully confirmed in NA62. Notwithstanding the presence of a much larger amount of material upstream of the LKr calorimeter in NA62 than in NA48, a study of single-photon efficiency underpinning the NA62 measurement of ${\rm BR}(K^{+}\to\pi^{+}\nu\bar{\nu})$ (and used to obtain a limit on ${\rm BR}(\pi^{0}\to{\rm invisible})$) found an inefficiency of about $10^{-5}$ at $E=20$ GeV, slightly decreasing at higher energies [142]. The LKr time resolution, however, is a significant issue. The necessary upgrade for the LKr calorimeter to be used in the $K^{+}$ phase is discussed below (Section 10.3.3.1). For the $K_{L}$ phases, and for KLEVER in particular, the calorimeter provides the measurement of the event time, and must have a time resolution of 100 ps or better for the reconstruction of $\pi^{0}$’s with energies of a few GeV. Additionally, the size of the LKr inner bore would limit the beam solid angle and hence the kaon flux during the $K_{L}$ phases. A new calorimeter that would meet these requirements is described in Section 10.3.3.2. ##### 10.3.3.1 The NA62 LKr calorimeter The LKr energy resolution meets the HIKE requirements, while the time resolution must be substantially improved. This can be achieved by performing a major upgrade of the LKr readout electronics, as discussed below. ##### Major upgrade of the LKr readout electronics The readout of the NA62 liquid krypton calorimeter is based on the measurement of the initial current induced by the charge generated by the showers, to have a fast response and to separate pile-up pulses. The electronics chain for each cell is composed of three elements: * • a preamplifier, sitting in the cold liquid, to collect the charge with an integration time of about 150 ns; * • a so-called transceiver module, mounted immediately after the cryostat feed- throughs, whose function is to restore the fast signal, removing the pole of the integration in the preamplifier, and to prepare a differential signal to drive long cables to the next stage; * • a shaper, to have a narrow signal (70 ns), followed by an FADC to digitise the signal at a rate of 40 MHz. The increase in intensity by a factor 4 will produce more pile-up events, whose detection could be difficult with the existing chain. To cope with this, the baseline for a new readout structure will be the reduction of the shaping time to the minimum possible of about 28 ns (the transceiver output has a risetime of about 20 ns which adds to the time constants of the shaper) with a reduction of the amplitude of about 40%, and subsequent digitization at 160 MHz, which is above the Nyquist limit, but which could help in identifying superposition of pulses looking at the width of the pulse. ##### Space charge The increase of intensity will also worsen the effect of the space charge: already in the current operation we observe a loss of response in the hottest cells, due to the build up of the space charge during the burst. This effect is inversely proportional to the square of the value of the high voltage across the cell and directly proportional to the energy density created by ions. The effect of the space charge is, at low densities, to reduce the electric field at the anode and to increase it at the cathode. There is then a threshold value after which the field at the anode is zero and above this value an increasing fraction of the cell is unresponsive. With the current NA62 beam intensity we see already that in the hottest cells we are near or just above the critical value, clearly seen both from the decrease in the response across the burst and from a modulation of the response across the $x$ coordinate of the cells. An increase in intensity will worsen this behaviour for the hottest cells, but it will also extend the area affected by this problem. To mitigate this issue, an increase of the high voltage is possible: going from the actual 3.5 kV to 5 kV will reduce by two the value of the critical parameter and the performance will be the same as now, which is not a problem for veto operations. An additional improvement for all analyses that do not use the calorimeter as a veto will be to increase the size of the actual Intermediate Ring Counter (IRC, Section 10.3.4), to cover the area most affected by the space charge. ##### 10.3.3.2 The new HIKE electromagnetic calorimeter We are investigating the possibility of replacing the LKr with a shashlyk calorimeter patterned on the PANDA FS calorimeter, in turn based on the calorimeter designed for the KOPIO experiment [214]. This design featured modules $110\times 110$ mm2 in cross section made of alternating layers of 0.275-mm-thick lead absorber and 1.5-mm-thick injection-moulded polystyrene scintillator. This composition has a radiation length of 3.80 cm and a sampling fraction of 39%. The scintillator layers were optically divided into four $55\times 55$ mm2 segments; the scintillation light was collected by WLS fibres traversing the stack longitudinally and read out at the back by avalanche photodiodes (APDs). KOPIO was able to obtain an energy and time resolution of 3.3% and 73 ps at 1 GeV with this design, establishing that it is capable of providing the same energy resolution as the LKr while meeting the time resolution requirements for HIKE. For HIKE, the design would be updated to use silicon photomultipliers (SiPMs) instead of APDs. The final choice of module size and readout granularity has yet to be determined, but on the basis of KLEVER simulations that assume that clusters are resolved if more than 6 cm apart, readout cells of $5\times 5$ cm2 seem reasonable. The Molière radius for the fine-sampling shashlyk design described above is 3.39 cm, so smaller cells can be used if needed. For comparison, the Molière radius of liquid krypton is 5.86 cm, and the NA48 LKr calorimeter features $2\times 2$ cm2 cells. Figure 48: Wireframe drawing of the MEC as implemented in Geant4 for the NA62 Monte Carlo. The new calorimeter, referred to as the main electromagnetic calorimeter (MEC) in KLEVER, would have an inner bore of at least 12 cm in radius to allow the passage of the neutral beam. The bore could be widened to as much as 15 cm to allow the penumbra of beam photons and neutral hadrons to pass through and be intercepted by the small-angle vetoes. For the $K^{+}$ phase, a smaller inner aperture would be required; in NA62, this angular region is covered by the IRC (Section 10.3.4), and a new IRC based on the same shashlyk technology would be used together with the MEC during the $K^{+}$ phase. The sensitive area has an outer radius of 125 cm. Figure 48 shows a wireframe drawing of the calorimeter as implemented in the NA62 Geant4-based Monte Carlo for HIKE. Figure 49: Geant4 simulation of a 5 GeV photon showering in a MEC module. Figure 50: Detection efficiency as a function of incident particle energy, for various thresholds on visible energy, from the Geant4 simulation. For the KLEVER sensitivity studies of Section 8.3, the inefficiency for photon detection assumed for the MEC was assumed to be 1 for energies below 100 MeV, falling exponentially to $10^{-3}$ at 1 GeV, $10^{-4}$ at 5 GeV, and thence to $10^{-5}$ at 15 GeV, in accordance with experience with the LKr as discussed above. The model of the shashlyk calorimeter developed for the HIKE Monte Carlo does not include the light and signal readout, which are still being defined, but does contain a detailed representation of the module structure, allowing studies of the efficiency as a function of visible energy deposition. Figure 49 shows the simulated interaction of a 5 GeV photon in a module of the calorimeter. Figure 50 shows results obtained from the simulation for the photon detection inefficiency as a function of incident energy for various thresholds on visible energy. The simulation, which confirms the visible energy fraction of 39%, demonstrates that the fine-samplng shashlyk design satisfies the efficiency requirements for KLEVER, as expected. The radiation resistance of the scintillator is a potential concern, especially for the $K_{L}$ phase. In KLEVER, the dose rate is dominated by photons from $K_{L}$ decays. The rate is most intense on the innermost layers of the calorimeter, for which it is about 2 kHz/cm2. Precise dose rate calculations have yet to be performed, but an estimate suggests a dose of 4 kGy/yr to the scintillator for the innermost layers. This estimate would suggest that radiation damage, while a concern, is likely manageable. Radiation robustness may be a factor in the final choice of scintillator. Figure 51: Shashlyk tiles made of perovskite nanocomposite scintillator tested as an alternative to conventional scintillator, in ambient light (left) and under ultraviolet light (right). Although current information suggests that optimized formulations of conventional polystyrene scintillator are sufficiently luminous, fast, and radiation resistant, in synergy with the AIDAinnova project NanoCal, we are evaluating the advantages that can be obtained with less conventional choices for the light emitter (e.g., perovskite [198, 215] or chalcogenide [216] quantum dots) or matrix material (e.g., polysiloxane [217]). In particular, we have participated in a recent head-to-head experimental comparison of small shashlyk prototypes made from conventional scintillator, specifically, the extrusion-moulded polystyrene scintillator formulated at IHEP Protvino for KOPIO [214], with 1.5% PTP and 0.04% POPOP, and a nanocomposite scintillator consisting of 0.2% lead halide bromide (CsPbBr3) nanocrystals in PMMA222Glass To Power SpA, Rovereto TV, Italy.. The latter, which emits at 520 nm (Figure 51), is expected to be a very fast and bright alternative to conventional scintillators; its comparatively long wavelength emission and use of PMMA as a matrix material is expected to confer good radiation hardness. Figure 52: Small shashlyk prototypes tested in fall 2022, during construction (left) and on the H2 beam line for testing (right). Identical prototypes consisting of 12 layers of 0.6-mm lead and 3-mm scintillator were constructed out of components from the PANDA/KOPIO calorimeter (Figure 52). For the conventional prototype, the original scintillating tiles were used with Kuraray Y-11(200) green-emitting WLS fibre. The nanocomposite prototype used Kuraray O-2(100) orange-emitting WLS fibre. Each was read out with a single Hamamatsu 13360-6050 SiPM ($6\times 6$ mm2, 50 $\mu$m pixel size) and fast amplifier with a gain of 4. Both protoypes worked well; the conventional prototype showed significantly more light output than the nanocomposite prototype, but we have reason to believe that the luminosity and attenuation length of the orange fibres plays a significant role. A series of tests are planned to explore different fibre-scintillator pairings (including the use of custom-produced fibres) as well as successively optimised nanocomposite scintillator formulations. In addition to the basic criteria on energy resolution, efficiency, time resolution, and two-cluster separation, ideally, the MEC would provide information useful for particle identification. For example, in KLEVER, identification of pion interactions would provide additional suppression of background for decays with charged particles in the final state, and, as the experience with KOTO suggests, it is crucial to have as much information as possible to assist with $\gamma/n$ discrimination. The fine transverse segmentation of the MEC will play an important role: a simple cut on cluster RMS with the existing LKr can suppress up to 95% of pion interactions in NA62 data. Fast digitisation of the signals from the MEC is expected to provide additional $\gamma/n$ discrimination. For both the $K^{+}$ and $K_{L}$ phases, the MEC will be backed up with hadronic veto calorimeters as discussed in Section 10.3.5. Figure 53: Top: Geant4 model of small prototype for romashka calorimeter, featuring spy tiles placed at key points in the shashlyk stack, together with a photograph of the module constructed at Protvino. Bottom: Fiber routing scheme for independent readout of spy tiles, giving rise to the name romashka (chamomile). We are also experimenting with concepts to obtain information on the longitudinal shower development from the shashlyk design. One possible concept makes use of “spy tiles”, 10-mm thick scintillator bricks incorporated into the shashlyk stack but optically isolated from it and read out by separate WLS fibres. The spy tiles are located at key points in the longitudinal shower development: near the front of the stack, near shower maximum, and in the shower tail, as illustrated in Fig. 53. This provides longitudinal sampling of the shower development, resulting in additional information for $\gamma/n$ separation. The prototype shown in Fig. 53 was constructed at Protvino in early 2018. Its basic functionality was tested in the OKA beamline in April 2018, and more comprehensive tests were carried out in September 2019 at DESY in collaboration with LHCb. Simulations suggest (and preliminary test beam data validate, to a certain extent) that the romashka design with spy tiles can give at least an order of magnitude of additional neutron rejection relative to what can be obtained from the transverse segmentation of the calorimeter alone, providing an overall suppression of up to 99.9% for neutron interactions. The small prototype has a cross sectional area of only $55\times 55$ mm2 (one readout cell) and a depth of $14X_{0}$ (60% of the design depth), and both transverse and longitudinal leakage significantly complicated attempts to measure the time resolution. For electrons with $1<E<5$ GeV, the measured time resolution was about 200 ps and virtually constant; we expect that this can be significantly improved. The time resolution for hits on the spy tiles (independently of information from the shashlyk stack) was on the order of 500–600 ps, which may be difficult to improve. This is not expected to be a problem, however: the main shashlyk signal establishes the event time and the association of the PID information from the shashlyk tiles is based on the segmentation, with occupancies per cell of at most a few tens of kHz on the innermost layers. Although promising, the romashka design is not the only solution under investigation for obtaining information on the longitudinal shower development. Other concepts being explored, besides variations on the spy tile readout such as on-board SiPMs, include alternatives such as two-sided front/back readout and explicit segmentation into two or more modules in depth. We plan to continue R&D work on improvements to the shashlyk design such as from the use of innovative scintillators and longitudinal segmentation schemes through 2023, while at the same time, we are designing a full-scale prototype of the baseline solution with conventional scintillator and a uniform shashlyk stack for beam testing next year. #### 10.3.4 Intermediate-ring calorimeter (IRC) In NA62, the IRC is a small, ring-shaped calorimeter that sits just upstream of the LKr to provide photon veto coverage for the angular interval corresponding to the inactive region near the inner surface of the LKr cryostat. The NA62 IRC has an inner radius of 60 mm and an outer radius of 145 mm. The inner and outer radii are not concentric—the inner bore is offset in the horizontal direction by 10 mm, so that it is centered on the beam axis. In NA62, the IRC is a shashlyk calorimeter with layers consisting of 1.5 mm of lead and 1.5 mm of scintillator. It is segmented into quadrants and read out with PMTs. In HIKE, as discussed in Section 10.3.3, the IRC will be used to extend the coverage of the calorimeter to small radii, and performance specifications similar to those for the calorimeter will be necessary. If the LKr used is during the $K^{+}$ phase, the IRC will cover the region of highest rates on the LKr, where space-charge effects are important. The new MEC, not needing a cryostat, will have minimum dead space at its inner radius. However, it could be convenient to have a bore of 15 cm or more to allow passage of the neutral beam halo during the $K_{L}$ phase of the experiment. In that case, during the $K^{+}$ phase, the IRC would be used in the same way with the MEC as it would with the LKr. For the HIKE $K_{L}$ phases, the IRC will then be moved downstream of the calorimeter and used to veto photons at the small radii occupied by the penumbra of the neutral beam. Ideally, the same instrument would be used for both phases, but due to the horizontal displacement of the $K^{+}$ beam at the location of the IRC, changes to the geometry may be necessary. If two different IRCs are needed, they can certainly be built with identical technology, namely, as shashlyk calorimeters with geometry similar to that for the NA62 IRC and sampling and readout granularity as for the HIKE MEC, providing more light for better time resolution and higher readout granularity for better rate resistance. #### 10.3.5 Hadron Calorimeter (HCAL) and Muon Veto Detector (MUV) ##### 10.3.5.1 Hadron calorimeter requirements As in NA62, the hadron calorimeter (HCAL) is the main detector for $\pi/\mu$ identification and separation. While minimum ionizing muons can be distinguished from pion showers without difficulty, so-called catastrophically interacting muons, which deposit all or a large fraction of their energy in the calorimeter, are of more concern. The $\pi/\mu$ separation is therefore achieved both by longitudinal tracking through the whole calorimeter and by examining the transversal shower shapes of electromagnetic muon versus hadronic pion showers. Similar to the NA62 MUV1-3 detectors, an average muon misidentification probability of ${\cal O}(10^{-6})$ over the momentum range from 10 to about 50 GeV is necessary (see Section 6). The HCAL therefore needs segmentation in all three dimensions. The 4–6 times higher intensity in HIKE with respect to NA62 leads to a total rate of charged tracks from $K^{+}$ decays of ${\cal O}(10~{}\text{MHz})$ on the calorimeters. This converts into a sub-nanosecond time resolution which is required to avoid dead-time from random veto. Such a time resolution cannot be achieved with scintillating strips as in the NA62 MUV1/MUV2, since the long light paths cause a resolution of 1 ns or worse. In addition, a strip read-out would suffer from double hits on the same strip at high rates. The HIKE HCAL therefore needs to have a cellular layout to reduce both the rate on each channel and the time resolution. With an iron absorber with a Molière radius of 1.72 cm, a cell size of $6\times 6$ cm2 allows a sufficient distinction between hadronic pion showers and electromagnetic showers from catastrophically interacting muons. A corresponding layout of an HCAL with an octagonal shape with an inner radius of 126 cm is shown in Fig. 54. It consists of a grid of $42\times 42$ cells. The subtraction of the beam pipe region and the corners to obtain an octagon results in 1440 cells in the transverse plane. Figure 54: Transverse HCAL layout with an inner radius of 126 cm and readout cells of $6\times 6$ cm2 cross section. The plane comprises in total 1440 cells. We consider two options for an iron-scintillator based calorimeter, which are described in the following. The decision between the two options will be taken in the near future, after more detailed simulation of the particle ID capabilities of the two designs. ##### 10.3.5.2 High-granularity hadron calorimeter For the ILD detector at the proposed International Linear Collider (ILC) the CALICE Collaboration has developed a highly granular analogue hadron calorimeter (AHCAL) [218] and a similar detector is being constructed for the CMS High Granularity Calorimeter Upgrade (HGCAL) [219]. In both these cases the high granularity is needed to improve jet energy resolution with particle- flow methods. While jet reconstruction is not necessary in HIKE, high granularity is nevertheless required for optimum $\pi/\mu$ separation and sub- nanosecond time resolution in a high- rate environment as described above. The basic design of a high-granularity HCAL is an iron-scintillator sampling calorimeter with 3 cm iron absorber plates interleaved with 40 layers of scintillating tiles of $6\times 6$ cm2 cross section and 1 cm thickness, which gives enough light yield to detect minimum ionizing particles. The total HCAL therefore comprises 7.2 nuclear interaction lengths and has a length of about 2 m, when considering additional space for electronics and air. Each scintillating tile is readout separately by a single SiPM, which is connected to a common readout board (see Fig. 55 for the Calice AHCAL layout). The SiPM needs a relatively large dynamic range, i.e. a large number of pixels, for a linear energy measurement for both minimum ionizing particles and electromagnetic and hadronic showers. Candidates for such SiPMs are e.g. the existing Hamamatsu types S13360 or S14160 with more than 14000 pixels and a sensitive area of $3\times 3$ or $6\times 6$ mm2. Figure 55: Left: $3\times 3$ cm2 scintillator tiles of the CALICE AHCAL, wrapped and unwrapped, mounted on a common readout board with SiPMs [218]. Centre and right: back and front sides of a readout board with $6\times 6$ cm2 scintillator tiles [220]. Each readout board of $36\times 36$ cm2 size carries $6\times 6=36$ scintillating tiles, wrapped in reflective foil and glued to the PCB. The SiPMs are surface-mounted on the PCB and are housed in small cavities on the bottom side of the tiles (Fig. 55). The front-end electronics together with a digitizing ASIC occupy the back side of the readout board. The digitization therefore takes place directly next to the photo detectors, ensuring an excellent time resolution. Assuming the general design shown in Fig. 54 with 1440 cells in each layer and in total 40 scintillating layers, the number of readout boards is 1800 (not all fully equipped) and the total number of channels is 57600. The possibility of a reduced channel count by e.g. using larger tiles in the outer region or less active layers is being studied at the moment. As described above, the digitization of the SiPM signals is done with an on- board ASIC. A suitable such chip is the HGCROC [221], which has been developed for the high-rate environment of the CMS HGCAL. ##### 10.3.5.3 Shashlik hadron calorimeter As second option we consider a shashlik calorimeter with the same $6\times 6$ cm2 cell size as described above. Each module (cell) consists of 80 absorber layers of 1.5 cm iron and 80 active layers of 0.5 cm plastic scintillator (Fig. 56). The scintillator light is read out with wavelength-shifting (WLS) fibres to either photomultiplier tubes or large SiPMs or SiPM arrays. For calibration, the scintillators can be activated by light input through an additional clear fibre. In the simplest set-up the number of channels is just 1440, greatly reduced with respect to the Calice-like design. However, such a design would not provide any information about the longitudinal shower development. In NA62, the two separate modules MUV1 and MUV2 are of great importance for $\pi/\mu$ separation, therefore a shashlik design should also at least contain two modules with separate readout or, alternatively, having light readout both on the front and the back side of the detector. Figure 56: A shashlik HCAL module. The beam comes from the left. The central fibre serves for injection of monitoring light. ##### 10.3.5.4 Muon veto detector (MUV) The muon veto detector (MUV) vetoes muons and contributes to the muon identification and suppression at a event-building-farm or the analysis level. It is placed following the HCAL behind a 80 cm thick iron wall (corresponding to about 5 additional nuclear interaction lengths). The MUV needs a time resolution of the order of 100–150 ps, to keep the loss of signal due to random activity at few percent level, as it is presently achieved by NA62 with a time resolution of about 500 ps. The MUV will be built from scintillating tiles with direct photo detector readout. Two feasible options exist. The first scenario is applicable with a high-granularity HCAL, where the MUV detector would be just another scintillating layer, but behind the iron wall and with scintillating tiles of 5 cm thickness to ensure high light yields for an optimum time resolution. The second MUV option is a copy of the NA62 MUV3 detector, also consisting of 5 cm thick scintillating tiles, but read out by PMTs from the back side. In contrast to NA62, where the tile front faces measure $22\times 22$ cm2, the tile size would need to be reduced to about $8\times 8$ cm2 to be able to cope with the higher rate in HIKE. #### 10.3.6 Small-angle electromagnetic calorimeter (SAC) For the measurements of $K^{+}\to\pi^{+}\nu\bar{\nu}$ and $K_{L}\to\pi^{0}\nu\bar{\nu}$, as well as other measurements of rare decays requiring hermetic photon vetoes, the coverage of the veto system must extend down to zero in the polar angle, to intercept photons that would otherwise escape the detector through the downstream beam pipe. Vetoing photons at small angle is easier for the $K^{+}$ phase than for the $K_{L}$ phases, because the $K^{+}$ beam can be diverted and dumped outside the acceptance of the small- angle calorimeter. For the $K_{L}$ phase, instead, the SAC sits directly in the neutral beam; it must reject photons from $K_{L}$ decays that would otherwise escape via the downstream beam exit while remaining relatively insensitive to the very high flux of neutral hadrons in the beam, so that the experiment is not blinded by the random vetoes from these hadrons. The design and construction of the SAC for the $K_{L}$ phase is thus a unique challenge. To the extent that R&D for the KLEVER SAC is well underway, its construction should be complete before LS3, allowing it to be validated for the $K^{+}$ phase before the end of the present run. The operational experience gained during the $K^{+}$ phase will then be used to validate the SAC with respect to the more stringent requirements for KLEVER, which form the basis for the following discussion. Note that for much of HIKE Phase 2 (specifically, for $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ measurements) a production angle of 2–4 mrad is preferred. For running at the smallest production angles, the SAC might not be installed because of the extremely high total neutral beam rate. Figure 57: Maximum tolerable inefficiency of the KLEVER SAC vs incident photon energy, for photons from $K_{L}\to\pi^{0}\pi^{0}$ events, from simulation. The maximum tolerable inefficiency for photon detection for the KLEVER SAC is illustrated as a function of incident photon energy in Fig. 57. This plot was obtained from the fast simulation described in Section 8.3 by cumulating the energy distribution in the SAC for $K_{L}\to\pi^{0}\pi^{0}$ events that evade all other vetoes and pass all other cuts used to select $K_{L}\to\pi^{0}\nu\bar{\nu}$ events. The following, indicative efficiency requirements can be identified: * • For $E<5~{}{\rm GeV}$, the SAC can be blind. For the $K\to\pi\nu\bar{\nu}$ studies in both the $K^{+}$ and $K_{L}$ experiments, background processes that are not otherwise efficiently vetoed do not have photons on the SAC. * • For photons with $5~{}{\rm GeV}<E<30~{}{\rm GeV}$, the SAC inefficiency must be less than 1%. Most photons on the SAC with energies in this range from $K_{L}$ decays that are otherwise accepted as signal candidates are from events in which there are two photons on the SAC, so that the efficiency requirement is relatively relaxed. * • Only for photons with $E>30~{}{\rm GeV}$ must the inefficiency be very low ($<10^{-4}$). For $K_{L}$ decays passing analysis cuts with a photon on the SAC in this energy range, the other photon is emitted at large angle and has low energy, so the SAC veto is important. Although these SAC efficiency requirements are not intrinsically challenging, from the simulations of the beamline without extension (Section 5.3), there are about 130 MHz of $K_{L}$ mesons, 440 MHz of neutrons, and 40 MHz of high- energy ($E>5$ GeV) beam photons incident on the SAC, and the required efficiencies must be attained while maintaining insensitivity to the nearly 600 MHz of neutral hadrons in the beam. In order to keep the false veto rate to an acceptable level, the hadronic component must be reduced to at most a few tens of MHz, so that the total accidental rate is dominated by the beam photons and in any case significantly less than the 100 MHz target cited in Section 8.2. These requirements lead to the following considerations: * • The SAC must be as transparent as possible to the interactions of neutral hadrons. In practice, this means that the nuclear interaction length $\lambda_{\rm int}$ of the SAC must be much greater than its radiation length $X_{0}$. * • The SAC must have good transverse segmentation to provide $\gamma/n$ discrimination. * • It would be desirable for the SAC to provide additional information useful for offline $\gamma/n$ discrimination, for example, from longitudinal segmentation, from pulse-shape analysis, or both. * • The SAC must have a time resolution of 100 ps or less. * • The SAC must have double-pulse resolution capability at the level of a few ns. In addition, in five years of operation, the SAC will be exposed to a neutral hadron fluence of about $10^{14}$ $n$/cm2, as well as a dose of up to several MGy from photons. Figure 58: Dimensional sketch of a SAC for KLEVER based on dense, high-$Z$ crystals with both transverse and longitudinal segmentation. One possible design that is well-matched to the KLEVER requirements would be to use a highly segmented, homogeneous calorimeter with dense, high-$Z$ crystals providing very fast light output. As an example, the small-angle calorimeter for the PADME experiment used an array of 25 lead fluoride (PbF2) crystals of $30\times 30\times 140~{}{\rm mm}^{3}$ dimensions. PbF2 is a Cherenkov radiator and provides very fast signals. For single crystals read out with PMTs, a time resolution of 81 ps and double-pulse separation of 1.8 ns were obtained for 100–400 MeV electrons [222], satisfying the KLEVER timing performance requirements. The PADME performance was obtained with fully digitising waveform readout at 5 GS/s; waveform digitisation would also be required for KLEVER (Section 11). At the doses expected at KLEVER, radiation-induced loss of transparency to Cherenkov light could be significant for PbF2, as suggested by existing studies with ionising doses of up to ${\cal O}(10~{}{\rm kGy})$ [223, 224, 225]. However, these studies also found significant annealing and dose-rate effects in PbF2, as well as the effectiveness of bleaching with UV light. If the effects of radiation damage to PbF2 prove to be a significant problem, a good, radiation-hard alternative could be recently developed, optimised lead tungstate (PbWO4, PWO) [226, 227, 228]. In particular, ultrafast PWO (PWO-UF) with a decay time constant of 640 ps, good light yield, and high radiation tolerance has recently been developed [229]. For the KLEVER SAC, a design with longitudinal segmentation is under study, as shown in Fig. 58. This design would feature four layers of $20\times 20\times 40$ mm3 PbF2 or PWO-UF crystals (each $\sim$4$X_{0}$ in depth). To minimise leakage, the gaps between layers would be kept as small as possible. Compact PMTs such as Hamamatsu’s R14755U-100 could fit into a gap of as little as 12 mm. Readout with SiPMs would facilitate a compact SAC design even further, but may require advances in SiPM radiation resistance and timing performance. Figure 59: Geant4 simulation of the KLEVER SAC showing a 1-GeV photon incident from the right. White tracks show photons, cyan tracks show Cherenkov photons, and the red track shows a negatively charged particle. A Geant4 simulation of the KLEVER SAC has been developed for the HIKE Monte Carlo. This simulation can be used to study the performance with different crystals and geometries (the readout response is not yet simulated). Fig. 59 shows the simulation of an interaction of a 1-GeV photon in the KLEVER SAC with PbF2 crystals and a 5-mm gap width between layers for readout with SiPMs. The Cherenkov photon yield is 85,000 per GeV of incident photon energy. There is a $\sim$10% loss of shower containment at the highest photon energies, which is not seen to depend on the size of the gaps between layers over the interval of 0–20 mm. This is considered to be an acceptable trade-off for the purpose of maintaining the detector relatively transparent to hadrons. The response to the main components of the neutral beam has been studied. Figure 60: SAC inefficiency for photons (top) and efficiency for neutrons and $K_{L}$ mesons (bottom left and right) vs. incident particle energy, for various thresholds on observed incident particle energy as measured from the number of Cherenkov photons produced. Fig. 60 shows the inefficiency of the SAC for photon detection and the efficiency of the SAC for neutron and $K_{L}$ detection as a function of incident particle energy, for various thresholds set on the measured incident particle energy from the number of Cherenkov photons produced. It is seen that with a threshold set at the number of Cherenkov photons produced by a photon of 4–5 GeV, the SAC efficiency requirements are satisfied, while about 30% of the neutral hadrons leave signals above threshold, contributing to the random veto rate. R&D work on the KLEVER SAC is currently being carried out in synergy with other collaborations, facilitated in part by participation in the AIDAinnova research network. CRILIN, an electromagnetic calorimeter under development for the International Muon Collider Collaboration, is an independently proposed, highly granular, longitudinally segmented, fast crystal calorimeter with SiPM readout and performance requirements similar to those for the KLEVER SAC [230], and much development work for KLEVER is being carried out in collaboration with CRILIN, with particular emphasis on the SiPMs, front-end electronics, and signal readout, as well as on solutions for detector mechanics and SiPM cooling. The first test beam measurements with individual PbF2 and PWO crystals were performed in summer 2021 at the Frascati BTF and the SPS North Area, followed by additional tests in fall 2022 in which some of the first commercially available samples of PWO-UF were also tested. These tests were focused on understanding the best possible time resolution that can be obtained, studying the systematics of light collection in the small crystals, and validating the CRILIN choices of SiPMs and the design of the readout amplifier. In autumn 2022, single $10\times 10\times 40$ mm3 crystals of PbF2 (4.3$X_{0}$) and PWO-UF (4.5$X_{0}$) were exposed to high-energy (60–120 GeV) electron beams at the SPS H2 beamline. Each crystal was viewed by a matrix of four Hamamatsu 14160-4010 SiPMs ($4\times 4$ mm2, 10 $\mu$m pixel size), which were read out in pairs, providing independent readout channels for the left and right sides of the crystal (Fig. 61). Figure 61: Left: Single crystal read out with $2\times 2$ matrix of four $4\times 4$ mm2 SiPMs. Right: Installation of crystal on H2 beam line during fall 2022 KLEVER/CRILIN test beam. The SiPM was chosen to represent a plausible choice in the current state of the art for high-speed response, short pulse width, and good radiation resistance. The signals from the SiPMs were amplified with the CRILIN electronics and digitized at 5 GS/s. The data collected are currently being analyzed, but some preliminary conclusions are already apparent: * • The time resolution obtainable from the combination of either crystal, PbF2 or PWO-UF, with the chosen SiPM and CRILIN electronics is excellent, with the final time resolution expected to be significantly better than required for the KLEVER SAC. * • The light yield for PWO-UF is approximately twice that for PbF2; the time resolution for PbF2 is slightly better. * • For either crystal, the light produced is highly localized on the rear face of the short crystal, requiring some care with the segmented readout. * • The CRILIN electronics performed very well; KLEVER should evaluate the possibility of faster shaping to obtain better double pulse discrimination. Figure 62: Two-layer $3\times 3$ crystal CRILIN prototype: mechancial drawing (left), assembled front module (center), and PCB with SiPMs for one layer (right). The next step of CRILIN development is to test a two-layer, $3\times 3$ crystal array. The mechanical prototype (Fig. 62) features a scheme for SiPM cooling. The light is read out in the same manner as for the single crystals tested in 2021–2022, with two channels per crystal, for a total of 36 channels. Modular electronics for power distribution, SiPM signal processing, and control have also been developed. This prototype will be populated with both PbF2 and PWO-UF crystals and tested in the coming months to validate the segmentation scheme, test the time resolution, study various optimisations such as the crystal surface preparation, and evaluate the performance of the engineering solutions adopted. As noted above, the full SAC design with 4 layers of 4-cm crystals suffers slightly from shower leakage, while the interaction probability for hadrons in the beam is about 30%. A possibility to make the SAC more hermetic for photon showers and less sensitive to hadrons is to exploit the effects of the coherent interactions of high-energy photons in oriented crystals to induce prompt electromagnetic showering. This is the same technique to be used for optimisation of the photon converter in the neutral beam (Section 5.3.1). Because of the relative ease in producing high-$Z$ optical crystals of high quality, there are good prospects for using this technique to construct a compact calorimeter with a very small radiation length, referred to the primary interaction [231]. A decrease by a factor of 5 in the effective radiation length for a 4-mm thick PWO crystal has been observed for 120 GeV electrons incident to within 1 mrad of the [001] axis [232]. Considering that the SAC acceptance for photons from $K_{L}\to\pi^{0}\pi^{0}$ decays in the fiducial volume extends at most to $\pm 2$ mrad, it should be possible to orient the crystals to enhance the probability for photon conversion. This would allow the thickness of the SAC layers to be reduced, reducing the rate of neutral hadron interactions while maintaining shower containment and photon detection efficiency. The potential gains from aligning the crystals, as well as the procedures and mechanics needed, are under study. In summer 2021, KLEVER and CRILIN, together with the STORM collaboration, used a tagged photon beam to measure the enhancement of shower processes in thicker PbF2 and PWO crystals as a function of angular alignment, correlating variables such as the energy dispersed in the crystal and multiplicity of charged particles produced in the shower with the production of Cherenkov and scintillation light in the crystals. STORM collected similar data with electron beams in summer 2021 and 2022. The data are still under analysis, but as a preliminary generalisation, the effective radiation length for crystals of 2–4 cm thickness is reduced by 20–30% by alignment, with an angular acceptance of about 1 mrad [233]. Further studies are planned with the CRILIN prototypes, and the OREO collaboration, continuing the work of STORM, has plans to build a prototype PWO-UF calorimeter in which the [100] axes of the crystals on the front layer are aligned. In addition to allowing studies of the variation of the energy deposition, radiation length and Molière radius with beam energy and crystal alignment, the OREO prototype will serve as a platform for developing the procedures needed to build a calorimeter with aligned crystals. #### 10.3.7 The hadron sampling calorimeter (HASC) The original NA62 setup was improved by the addition of a hadron sampling calorimeter (HASC-1) adjacent to the beampipe, downstream of the muon detector. The primary purpose of this detector is to reduce the $K\to\pi^{+}\pi^{+}\pi^{-}$ background to the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay, by vetoing the topology in which the $\pi^{-}$ undergoes hadronic interaction in the first STRAW chamber, while an energetic $\pi^{+}$ travels in the beampipe undetected by the IRC and emerges downstream. Analysis of the NA62 Run 1 dataset has revealed that HASC-1 is also efficient as a photon veto, providing a 30% reduction of the $K^{+}\to\pi^{+}\pi^{0}$ background to the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay. Consequently, a second calorimeter (HASC-2) was installed in 2021 at the HASC-1 longitudinal position, symmetrically with respect to the beam axis, which improves further the $K^{+}\to\pi^{+}\pi^{0}$ rejection. Each of the HASC-1 and HASC-2 stations consists of 9 identical modules. Each module is a sandwich of 120 lead/scintillator alternating tiles, with a total volume of $10\times 10\times 120~{}\mathrm{cm}^{3}$ (W x H x L). The sampling ratio is 4:1, the scintillator tiles having a dimension of $100\times 100\times 4~{}{\rm mm}^{3}$ while the lead thickness is 16 mm. Each module is organised in 10 longitudinal readout sections, each scintillator tile of every section being optically coupled with a wavelength shifting (WLS) optical fiber of 1 mm2 cross-section. At the rear side of each module there are 10 optical connectors, originally designed to be coupled with 3$\times$3 mm2 green- sensitive micro-pixel avalanche photodiodes (MAPD) (currently the S12572-015C Hamamatsu SiPM sensors are used). The FE electronics and SiPMs installed on the HASC-2 station are cooled down to 21ºC with a custom-made system consisting of three Peltier thermoelectric coolers / module and a water-air heat exchanger used to blow cold air in the modules end-cap cases. The temperature is maintained constant by a temperature controller – MCU based – which regulates the Peltier supply voltage via a PID routine. For HIKE, this scheme will be extended to HASC-1, improving the quality of SiPM signals and reducing ageing due to radiation to an acceptable level. Signal rates observed in the HASC during the NA62 data taking in 2022 at nominal beam intensity are summarised in Table 16. The rates fall rapidly with the distance from the beampipe. The modules next to the beampipe contain $35\%$ and $25\%$ of the total activity in the HASC-1 and HASC-2 stations, respectively. Table 16: Signal rates observed in the HASC during the NA62 data taking in 2022. Detector part | Rates in NA62 setup at nominal intensity ---|--- HASC-1 | 1 MHz HASC-1, most active module | 0.35 MHz HASC-1, most active channel | 0.05 MHz HASC-2 | 0.21 MHz HASC-2, most active module | 0.05 MHz HASC-2, most active channel | 6 kHz Total (HASC-1 and HASC-2) | 1.21 MHz The NA62 HASC provides near-optimal geometric coverage for the relevant topologies of the $K^{+}\to\pi^{+}\pi^{+}\pi^{-}$ and $K^{+}\to\pi^{+}\pi^{0}$ backgrounds to the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay, and is suitable also for HIKE Phase 1. Assuming the HIKE beam profile to be similar to the NA62 one, we expect at most a rate of 200 kHz in the most active channel for the HIKE HASC. The main issue in the operation of the current HASC at the HIKE beam intensity is the random veto. It is determined chiefly by the time resolution of the detector, presently at the level of 370 ps FWHM for HASC-2 and 430 ps for HASC-1, in agreement with HAMAMATSU specifications [234]. For HIKE, the resolution should be improved to $\mathcal{O}(100~{}\mathrm{ps})$. The HAMAMATSU S13362-3050DG SiPMs achieving this value are available on the market [235], and represent a promising candidate for the upgrade. Table 17 shows a comparison between the current HASC SiPMs and the candidate for the upgrade. Another appealing feature of the new SiPMs is the embedded two-stage thermoelectric cooling which greatly simplifies the current cooling scheme. Table 17: Summary of SIPM characteristics for the model currently employed in the NA62 HASC and the one proposed for the HIKE upgrade. SIPM model | S12572-015C (current) | S13362-3050DG (upgrade) ---|---|--- Photon detection efficiency | 25% | 40% Typical dark count | 1000 kcps | 25 kcps Gain | $2.3\times 10^{5}$ | $1.7\times 10^{6}$ Time resolution FWHM | 400 ps | 110 ps [236] An alternative option for the HASC upgrade, which is currently under study, involves PMTs. A particularly interesting example is HAMAMATSU H14220A, which could be used to read out individual scintillator plates (instead of groups of six, as done presently), eliminating pileup due to detecting a convoluted signal hence offering lower rise time ($\tau_{\mathrm{r}}$) and fall time ($\tau_{\mathrm{f}}$). The main drawback of reading single scintillators is the lower number of collected photons ($N_{\mathrm{ph}}$), compared with SiPM readout, which could have a negative impact on the time resolution which is known to be proportional with $\sqrt{\tau_{\mathrm{r}}\tau_{\mathrm{f}}/N_{\mathrm{ph}}}$ [237]. ### 10.4 Detectors specific to KLEVER #### 10.4.1 Upstream veto (UV) and active final collimator (AFC) The upstream veto (UV) rejects $K_{L}\to\pi^{0}\pi^{0}$ decays in the 40 m upstream of the fiducial volume where there are no large-angle photon vetoes. The UV is a shashlyk calorimeter with the same basic structure as the MEC (without the spy tiles). Because the UV does not participate in event reconstruction, the readout granularity can be somewhat coarser than for the MEC. The sensitive area of the UV has inner and outer radii of 10 cm and 100 cm. The active final collimator (AFC) is inserted into the hole in centre of the UV. The AFC is a LYSO collar counter with angled inner surfaces to provide the last stage of beam collimation while vetoing photons from $K_{L}$’s that decay in transit through the collimator itself. The collar is made of 24 crystals of trapezoidal cross section, forming a detector with an inner radius of 60 mm and an outer radius of 100 mm. The UV and AFC are both 800 mm in depth. The maximum crystal length for a practical AFC design is about 250 mm, so the detector consists of 3 or 4 longitudinal segments. Each crystal is read out on the downstream side with two avalanche photodiodes (APDs). These devices couple well with LYSO and offer high quantum efficiency and simple signal and HV management. Studies indicate that a light yield in excess of 4000 p.e./MeV should be easy to achieve. #### 10.4.2 Charged-particle rejection For the rejection of charged particles, $K_{L}\to\pi^{+}e^{-}\nu$ is a benchmark channel because of its large branching ratio and because the final state electron can be mistaken for a photon. Simulations indicate that the needed rejection can be achieved with two planes of charged-particle veto (CPV) each providing 99.5% detection efficiency, supplemented by the $\mu^{\pm}$ and $\pi^{\pm}$ recognition capabilities of the MEC (assumed in this case to be equal to those of the LKr) and the HIKE hadronic calorimeters, which could be reused in KLEVER. The CPVs are positioned $\sim$3 m upstream of the MEC and are assumed to be constructed out of thin scintillator tiles. In thicker scintillation hodoscopes, the detection inefficiency arises mainly from the gaps between scintillators. For KLEVER, the scintillators will be only $\sim$5 mm thick ($1.2\%X_{0}$), and the design will be carefully optimized to avoid insensitive gaps. #### 10.4.3 Preshower detector The PSD measures the directions for photons incident on the MEC. Without the PSD, the $z$-position of the $\pi^{0}$ decay vertex can only be reconstructed by assuming that two clusters on the MEC are indeed photons from the decay of a single $\pi^{0}$. With the PSD, a vertex can be reconstructed by projecting the photon trajectories to the beamline. The invariant mass is then an independent quantity, and $K_{L}\to\pi^{0}\pi^{0}$ decays with mispaired photons can be efficiently rejected. The vertex can be reconstructed using a single photon and the constraint from the nominal beam axis. Simulations show that with $0.5X_{0}$ of converter (corresponding to a probability of at least one conversion of 50%) and two tracking planes with a spatial resolution of 100 $\mu$m, placed 50 cm apart, the mass resolution is about 20 MeV and the vertex position resolution is about 10 m. The tracking detectors must cover a surface of about 5 m2 with minimal material. Micropattern gas detector (MPGD) technology seems perfectly suited for the PSD. Information from the PSD will be used for bifurcation studies of the background and for the selection of control samples, as well as in signal selection. ## 11 Data acquisition and high-level trigger Particle rate in HIKE Phase 1 is 3000 MHz in the beam tracker and 200 MHz for $K^{+}$ identification. Single-particle rates in the downstream detectors is about 50 MHz, including both $K^{+}$ decay products and the charged beam’s muon halo. The neutral beam of HIKE Phases 2 and 3 is 50% more intense that the Phase 1 charged beam. The small-angle calorimeter (SAC) in the neutral beam will encounter rates of about 100 MHz, and the other detectors encounter rates of about 20 MHz. The high particle rates will give rise to a harsh radiation environment in the HIKE experimental cavern. To limit the impact of radiation on the DAQ system, we aim to minimise the amount of electronics located in the cavern, particularly electronics related to data processing and storage, and use radiation-tolerant optical links and error-correcting data formats such as (lp)GBT. ### 11.1 Readout boards The readout boards serve two purposes: to host devices that timestamp digital signals produced by the detectors, and to encode and transmit these data away. The readout boards are located in the high-radiation environment close to the detector, so radiation-hardened technologies will be used. The time of signals from the detectors will be recorded either with time-digital converters (TDC), analogue-to-digital converters (ADC), or dedicated ASIC devices. The 64-channel PicoTDC is under development at CERN and is anticipated to be ready before HIKE Phase 1. The PicoTDC can timestamp signals at 12 ps precision [238] and outputs data in the GBT format, allowing simple readout boards to be used that only convert electric signals from the PicoTDC into optical signals that are sent to the rest of the DAQ system. Possible commercial solutions for fast FADC-based readout boards, to be adapted to the requirements of the DAQ in synergy with the manufacturer, are under investigation. The HIKE Phase 1 beam tracker data will be handled by dedicated ASICs. The number of detector channels read by TDC or ADC in HIKE Phase 1 is estimated to be about 50000. Approximately half of the detector channels will be read by TDC devices, the other half by ADCs. Therefore an estimated 400 TDC devices with 64 channels each, and 1600 ADC devices with 16 channels each, will be required for HIKE Phase 1. For HIKE Phase 2, the 10240 channels of the LAV, $\mathcal{O}(2000)$ channels of the MEC, and 400 channels of the SAC will be read using ADCs, requiring about 800 ADC devices with 16 channels each. Spy tiles in the MEC would contribute an additional 8000 channels that can be read using TDCs, in which case about 125 TDC devices with 64 channels each will be needed. ### 11.2 Streaming readout Moving the bulk of the DAQ hardware away from the high-radiation environment suggests implementing a streaming readout system, where detector signals are collected by radiation-tolerant readout boards and then transferred immediately, without an external trigger signal and without complicated processing or storage. The transferred data are received and buffered in the RAM of a set of servers, then transmitted to a computing farm for selection and filtering. The first stage of data readout may be implemented using a FELIX-based system, developed at CERN and exploited in several HEP use cases. The current version of the FELIX PCIe card has 24 optical links with a 9.6 Gb/s line rate and supports several data formats. When operating with the radiation-tolerant GBT format the effective data bandwidth is 3.2 Gb/s per link. New versions of FELIX, currently under development and anticipated on the timescale of HIKE, can provide faster links (e.g. 25 Gb/s) or a larger number of links (e.g. 48). The FELIX system also takes care of distributing clock, timing and commands to the front-ends, as well as receiving monitoring data for detector control purposes. FELIX cards are hosted in servers, and all data received by the FELIX cards are transferred to the host and buffered in memory. The servers then provide data to the computing farm as needed, over a high-speed ethernet network. As modern servers can be equipped with 1 TB of RAM, the data recorded during a whole SPS spill (4.8 s) may be buffered, greatly relaxing any latency requirement on the data processing as the inter-spill period (at least 9.6 s) can be exploited. Other options for interfacing the detector data with the computing farm are presently available, such as the PCI40 board developed for the LHCb and ALICE experiments, and the NaNet board, used by the KM3Net experiment. The main challenges of a streaming readout system are the transfer and processing of large quantities of data. The highest data rates are expected during the KLEVER phase, in particular from the SAC, MEC, and LAV detectors. The raw data rates are summarised in Table 18 assuming the use of 14-bit ADCs. The largest rate is 72 Tb/s from the LAV. This can be considerably reduced by only reading channels that contain non-zero energy, by implementing a zero- suppression (ZS) algorithm in the readout boards. Assuming that reading data covering 16 ns is enough to contain the data from one event, the zero- suppressed data rates are given in Table 19. In this case, the SAC dominates the data rate. Taking the FELIX system with GBT data format as the benchmark solution, the zero-suppressed SAC data could be collected by 9 FELIX cards. Reading the non-suppressed data is unfeasible, and the streaming readout design depends on the successful implementation of zero-suppression in the ADC-based readout boards. Table 18: Raw data rates expected from the MEC, SAC, and LAV in HIKE Phase 3. Detector | Channels | Sampling rate (GHz) | Raw data rate (Tb/s) ---|---|---|--- LAV | 10240 | 0.5 | 72 MEC | 2000 | 1 | 26 SAC | 400 | 1 | 6 Table 19: Zero-suppressed data rates expected from the MEC, SAC, and LAV in HIKE Phase 3. Detector | Event rate (MHz) | Average N channels hit per event | ZS data rate (Gb/s) ---|---|---|--- LAV | 14 | 5 | 8 MEC | 18 | 10 | 40 SAC | 95 | 30 | 640 ### 11.3 Triggered readout In case zero suppression of the calorimeter data proves unfeasible, or data transmission at the required speed cannot be achieved on the calorimeter readout boards, a trigger may still be necessary to reduce the required data throughput. The key parameters of the trigger are latency, which dictates how much data will have to be stored on the readout boards, and selectivity, which dictates how much data will have to be transferred to the computing farm. The level of data reduction that the trigger can achieve is a compromise between the sophistication of the algorithms, with more sophisticated algorithms providing better selectivity, and the time taken to form the trigger decision. We envisage a software trigger, with the trigger decision prepared using dedicated trigger algorithms running on the computing farm, most likely on a subset of the event data. Recalling the anticipated LAV data rate is 72 Tb/s (9 TB/s), assuming a software-trigger latency of 100 ms and 100 LAV readout boards, each LAV readout board must be able to cache about 9 GB of data. If the trigger can achieve a reduction factor of 100 – the same as the combined hardware and software trigger of NA62 – the transfer rate would be 720 Gb/s, which can be handled by 10 FELIX cards. ### 11.4 Event-building farm and event filter Regardless of whether a streaming or triggered DAQ is implemented, an event filter will be used to select candidate events, reducing the amount of data sent to permanent storage. The first stage of the event-filter is to combine fragments of events coming from the different detectors and reconstruct basic objects in each detector to identify physics events. Taking $K^{+}\to\pi^{+}\nu\bar{\nu}$ decays as an example, the event-building code will identify coincident signals in the RICH, STRAW, and KTAG as a potential $K^{+}\to\pi^{+}\nu\bar{\nu}$ event. This stage will be performed in an event- building computing farm located in the surface building of the experimental cavern. The backbone of the event-building farm will be standard server computers running C++ algorithms. However, the event-building process centres around combinatorial algorithms that are ideally suited for hardware acceleration using FPGA or GPU devices. We envisage the use of heterogeneous computing architectures to handle the demanding processing task given to the event-building farm. Subsequently, the event filter will examine signals in the events to identify the small fraction that should be kept for detailed analysis offline. This stage will be performed using dedicated servers in the Computing Centre on the CERN Prévessin site. A number of machine-learning- based algorithms have been developed at NA62 for particle identification and other tasks, and these can be the basis for evaluating events in the HIKE event filter. ## 12 Online and offline computing The baseline computing model is a scaled version of the current NA62 system. A flexible HIKE software platform is currently under development, on the basis of the NA62 software. HIKE will profit from the use of dedicated machines in Prévessin for many, if not all, of the described functions below, provided enough band capability is present between the experiment site and the computing centre. ### 12.1 Online monitoring The HIKE online monitoring system will be based on an updated model of the current one used in NA62. At present in NA62, a subset of the data are sent from farm nodes running the event-building processes to dedicated four monitoring computers; each event is then processed by the standard reconstruction software. The reconstructed events are then forwarded to a last machine where the different streams are recombined for displaying by dedicated processes. A set of low-level monitoring histograms produced by the reconstruction processes are also combined and displayed, for the purpose of quick data quality checks. For HIKE, a multi-level analysis pipeline, where each level performs more complex and time consuming tasks, should be implemented to shorten the latency between the acquisition of the data and the first feedback. Low level (data corruption) and high level (data quality) algorithms will perform error detection and alert the shift crew and the run control software. As often as possible automatic recovery procedures will be implemented to minimise data losses. A set of low-level calibration constants will also be computed during this first reconstruction and used both for monitoring purposes and to be fed back into the acquisition system. Depending on the details of the trigger scheme, the computing farm could also perform some of data filtering tasks. Reconstructed events will be processed with data quality algorithms, also used during full offline data processing (see next section), yielding output information collected over a whole run (about 1500 bursts) to be reviewed by the detector experts. ### 12.2 Calibration and data quality Calibration and data quality monitoring strategy will be built on the NA62 systems. Calibration will be performed in several stages. First, geometrical detector alignments will be initially obtained from special “muon” runs, where TAXes are closed and the spectrometer magnet is switched off to obtain straight penetrating tracks. Later, these geometrical alignments will be refined and monitored using kaon decays in standard data-taking conditions. Raw timing alignments within individual detectors, like slewing corrections, will be obtained from data and assumed to be constant over a certain data taking period, although they are permanently monitored. A similar procedure to that of NA62 for relative timing alignment between different detectors (“T0s”) will be developed, to achieve an average precision at the picosecond level, with a width corresponding to the individual timing resolutions of each detector. This will be done on a per-burst basis. During processing information on detector-specific calibrations will also be obtained, on a per- burst basis, like energy conversions, dead/noisy channels etc, which will be fed back to the analysis software. Data quality will be additionally monitored during full offline processing by software specific to each detector, but also including more common properties like efficiencies for different (online or offline) trigger algorithms. Graphical and numerical outputs will be produced for every detector, together with a database containing the monitored (and calibrated) quantities and alarms raised if certain thresholds are passed for one burst. ### 12.3 Data processing model Based on experience, the NA62 processing model, with some improvements and changes, can be adapted for HIKE. In NA62, raw data processing and user analysis is performed only on CERN computing clusters, while Monte Carlo production is performed using the international GRID, including CERN Grid resources. Processing of raw data for HIKE will be performed at CERN and using GRID resources (with a split still to be defined depending on the details of available computing and data sizes at the time of data taking), due to the increased amount of data. The splitting of the processed data according to trigger and/or physics analysis interests (“filtering”) must be refined to obtain stream sizes that are smaller and easier to handle. Stronger selection criteria will be applied, and the amount of information written for every selected event (slimming) will be reduced. ### 12.4 Distributed computing model Large-scale generation of Monte Carlo simulations will be performed exclusively on the Grid. The existing NA62 Grid framework is a one-of-its-kind system, created specifically for NA62 and providing an easy-to-use and fully automated production system [239]. With an uptime of 99.99% over the last ten years, it is a proven system that we envisage to extend to HIKE. Besides its simulation and reconstruction capabilities, our Grid framework can be extended to handle user analysis, in case this becomes a necessity for HIKE. ### 12.5 Data reduction model Due to the expected increase in the amount of data (due to the increased intensity) and in the size of data (due to the increase of signal channels), both filtering and slimming are foreseen for HIKE. For filtering, overlap between different filtering streams must be minimised, while several filter outputs may be combined at analysis level. For slimming, after reconstruction no low-level hit information will be stored, discarding as much as possible any information that can be: 1. 1. Recomputed from stored information. 2. 2. Likely not needed at analysis level. In specific cases of detector studies, or when an analysis requires access to low-level information, a selected list of events/bursts will be reconstructed again. The low-level information will be kept for those small dedicated samples for the duration of the study. Generally, most analyses make use of standard high-level objects (downstream tracks, vertices) without the need to access most of the information stored in the underlying objects from which they were built (clusters, tracks, candidates). Those high level object could therefore contain only the subset of information commonly used (time and position at associated sub-detectors) and stored in high-level analysis files which would be the basis for most user analyses. Again, some dedicated studies occasionally will require access to the full reconstructed objects, but these are generally aiming at deeper understanding of detector effects and to develop procedures to deal with them. In most cases the final procedure will not rely on the full reconstruction information, but only on that already available in the high-level objects. In addition, such procedures can generally be standardised and applied to several analyses without further study. It will therefore be sufficient to reconstruct a subset of data with full information for the duration of the study. ## 13 Infrastructure and safety HIKE will be housed in TCC8 and ECN3 where NA62 is at present. An evaluation of the necessary modifications to the experimental area has been carried out in collaboration with the ECN3 Task Force [155], the North Area Consolidation Project (NA-CONS) [148, 149, 150] and the Conventional Beams Working Group [147], where topics such as requirements for beam infrastructure, vacuum, cooling and ventilation, electrical distribution, handling and transport of detectors, access to the cavern, IT infrastructure, gas distribution, cryogenic systems, civil engineering and radiation protection have been addressed. The full list of requirements is provided in Ref. [240]. In order to ensure full compatibility with the North Area consolidation and as a central access point to all CERN service groups, discussion for the implementation of all requirements is steered through the NA-CONS Technical Coordination Committee (TCC) meetings, while integration aspects are handled in the Integration Committee for Experimental Areas (ICEA) as part of NA-CONS. The best match between the HIKE requirements and the infrastructure studies will be further investigated for the Proposal. We would like to highlight that the experimental layout of HIKE does not require any civil engineering work before at least LS4. The minimal changes and updates in the infrastructure will have a positive impact on the foreseen costs. Furthermore, the requirements for the electrical distribution and the detector cooling will be similar to that of NA62. The vacuum system with seven cryo-pumps went through thorough maintenance in 2021 and will be suitable to provide a good vacuum at the $10^{-6}$ mbar level for HIKE. To keep the NA62 LKr calorimeter operational for HIKE Phase 1, continuous maintenance of the cryogenics is needed. In addition, besides the installation of a new readout back end, the legacy components of the readout chain should be reviewed. Actions in this direction will be started already in the next YETS and procedures for regular review intervention will be defined. The human, technical and financial resources needed for HIKE are being evaluated. Table 20 shows a tentative indication of the cost of the main detectors for HIKE. The total for Phase 1 is 22.3 MCHF. The Phase 2 will see the addition of the Main Electromagnetic Calorimeter (for an extra 5 MCHF). The remaining costs for Phase 3 (KLEVER) will be mainly the increase in the number of Large Angle Vetoes from 12 to 25 units (for an extra 9 MCHF). The numbers are only intended to give an idea of the financial extent of the project. A refined estimate will be prepared for the forthcoming proposal. Table 20: Tentative indication of detector costs for HIKE. Detector | Cost (MCHF) | Comments ---|---|--- Kaon ID (KTAG) | 0.5 | Using MCP-PMTs Beam Tracker | 2.5 | Development, and production of 4 planes Charged particle veto (CHANTI) | 0.4 | 6 stations, with SiPMs VetoCounter (VC) | 0.2 | 3 stations (SciFi technology) Anti0 | 0.1 | 1 plane, same technology as TimingPlanes Large Angle Vetos (LAV) | 8 | 12 modules (Phase 1) STRAW | 3.5 | 4 Straw chambers and associated readout LKr upgrade OR MEC | 2.5 OR 5 | Readout upgrade OR new MEC Small Angle Calorimeter (SAC) | 2 | High-Z crystals Pion ID (RICH) | 0.8 | Using SiPMs Timing Detector | 0.2 | 2 planes, scint. tiles and SiPMs Hadronic Calorimeter + Muon plane | 1.5 | Shashlyk technology and SiPMs HASC | 0.1 | Using SiPMs MEC (if not already in Phase 1) | 5 | Shashlyk technology and SiPMs (Phase 2) Large Angle Vetoes (LAV) | 9 | Additional 13 modules (Phase 3) ## 14 Summary and conclusions The HIKE project presented in this document constitutes a multi-phase experimental programme taking full advantage of the ECN3 experimental hall to provide a comprehensive study of flavour physics in the kaon sector. This physics programme is complementary to, and out of reach of, the LHC experiments. Importantly, HIKE will be sensitive to new physics up to the highest mass scales. The HIKE facility will, in addition, search for feebly- interacting particles with unprecedented sensitivity. A summary of the HIKE phases and related intensity requirements is presented in Table 21. The experimental apparatus changes over time with a staged approach, which allows HIKE to evolve and adapt the physics scope in its three phases, an important feature for a project that embraces a time scale of more than decade during which the physics landscape could change. Thanks to the successful experience of NA62 and its predecessor NA48, the experimental technique is well established and robust expectations of sensitivity can be provided driven by the analysis of the existing data. In this way the expected sensitivity for benchmark studies have been presented above. The HIKE facility allows other users to run in parallel, even in the same area. This facilitates a high degree of diversity in the North-Area physics programme which, we believe, is crucial for the future of particle physics. Table 21: Summary of intensity requirements for the HIKE Phases. The number of kaon decays refers to those inside a fiducial decay volume (see Sections about physics sensitivity). For the $K^{+}$ phase, the maximum required intensity is specified; possibilities of reducing the intensity and increasing the momentum bite will be studied for the full proposal. An intensity higher than $2\times 10^{13}$ can be used by HIKE when operating in dump mode. Phase | Protons on target/spill | $K$ decays/year | Protons on TAX ---|---|---|--- Phase 1 ($K^{+}$) | $1.2\times 10^{13}$ | $2\times 10^{13}$ | - Phase 2 ($K_{L}$+tracking) | $2\times 10^{13}$ | $3.8\times 10^{13}$ | - Dump Mode | - | - | (2–4)$\times 10^{13}$ Phase 3 ($K_{L}$, KLEVER) | $2\times 10^{13}$ | $1.3\times 10^{13}$ | - ## References * [1] Georges Aad “Observation of a new particle in the search for the Standard Model Higgs boson with the ATLAS detector at the LHC” In _Phys. Lett. B_ 716, 2012, pp. 1–29 DOI: 10.1016/j.physletb.2012.08.020 * [2] Serguei Chatrchyan “Observation of a New Boson at a Mass of 125 GeV with the CMS Experiment at the LHC” In _Phys. Lett. B_ 716, 2012, pp. 30–61 DOI: 10.1016/j.physletb.2012.08.021 * [3] “2020 Update of the European Strategy for Particle Physics” Geneva: CERN Council, 2020 DOI: 10.17181/ESU2020 * [4] R. Aaij “Differential branching fractions and isospin asymmetries of $B\to K^{(*)}\mu^{+}\mu^{-}$ decays” In _JHEP_ 06, 2014, pp. 133 DOI: 10.1007/JHEP06(2014)133 * [5] Roel Aaij “Angular analysis and differential branching fraction of the decay $B^{0}_{s}\to\phi\mu^{+}\mu^{-}$” In _JHEP_ 09, 2015, pp. 179 DOI: 10.1007/JHEP09(2015)179 * [6] Roel Aaij “Test of lepton universality in beauty-quark decays” In _Nature Phys._ 18.3, 2022, pp. 277–282 DOI: 10.1038/s41567-021-01478-8 * [7] Roel Aaij “Angular analysis of the $B^{0}\to K^{*0}\mu^{+}\mu^{-}$ decay using 3 fb-1 of integrated luminosity” In _JHEP_ 02, 2016, pp. 104 DOI: 10.1007/JHEP02(2016)104 * [8] R. Aaij “Test of lepton universality with $B^{0}\rightarrow K^{*0}\ell^{+}\ell^{-}$ decays” In _JHEP_ 08, 2017, pp. 055 DOI: 10.1007/JHEP08(2017)055 * [9] A. Abdesselam “Test of Lepton-Flavor Universality in ${B\to K^{\ast}\ell^{+}\ell^{-}}$ Decays at Belle” In _Phys. Rev. Lett._ 126.16, 2021, pp. 161801 DOI: 10.1103/PhysRevLett.126.161801 * [10] S. Choudhury “Test of lepton flavor universality and search for lepton flavor violation in $B\rightarrow K\ell\ell$ decays” In _JHEP_ 03, 2021, pp. 105 DOI: 10.1007/JHEP03(2021)105 * [11] Jason Aebischer et al. “$B$-decay discrepancies after Moriond 2019” In _Eur. Phys. J. C_ 80.3, 2020, pp. 252 DOI: 10.1140/epjc/s10052-020-7817-x * [12] Wolfgang Altmannshofer and Peter Stangl “New physics in rare B decays after Moriond 2021” In _Eur. Phys. J. C_ 81.10, 2021, pp. 952 DOI: 10.1140/epjc/s10052-021-09725-1 * [13] Claudia Cornella et al. “Reading the footprints of the B-meson flavor anomalies” In _JHEP_ 08, 2021, pp. 050 DOI: 10.1007/JHEP08(2021)050 * [14] Gino Isidori, Davide Lancierini, Patrick Owen and Nicola Serra “On the significance of new physics in b → s$\ell$+$\ell$$-$ decays” In _Phys. Lett. B_ 822, 2021, pp. 136644 DOI: 10.1016/j.physletb.2021.136644 * [15] Vincenzo Cirigliano, Andreas Crivellin, Martin Hoferichter and Matthew Moulson “Scrutinizing CKM unitarity with a new measurement of the $K_{\mu 3}/K_{\mu 2}$ branching fraction”, 2022 arXiv:2208.11707 [hep-ph] * [16] Chien-Yeah Seng, Daniel Galviz, Mikhail Gorchtein and Ulf-G. Meißner “Complete theory of radiative corrections to Kℓ3 decays and the Vus update” In _JHEP_ 07, 2022, pp. 071 DOI: 10.1007/JHEP07(2022)071 * [17] Douglas Bryman, Vincenzo Cirigliano, Andreas Crivellin and Gianluca Inguglia “Testing Lepton Flavor Universality with Pion, Kaon, Tau, and Beta Decays”, 2021 arXiv:2111.05338 [hep-ph] * [18] R.. Workman “Review of Particle Physics” In _PTEP_ 2022, 2022, pp. 083C01 DOI: 10.1093/ptep/ptac097 * [19] Alex Keshavarzi, Kim Siang Khaw and Tamaki Yoshioka “Muon g $-$ 2: A review” In _Nucl. Phys. B_ 975, 2022, pp. 115675 DOI: 10.1016/j.nuclphysb.2022.115675 * [20] J. Beacham “Physics Beyond Colliders at CERN: Beyond the Standard Model Working Group Report” In _J. Phys. G_ 47.1, 2020, pp. 010501 DOI: 10.1088/1361-6471/ab4cd2 * [21] Prateek Agrawal “Feebly-interacting particles: FIPs 2020 workshop report” In _Eur. Phys. J. C_ 81.11, 2021, pp. 1015 DOI: 10.1140/epjc/s10052-021-09703-7 * [22] Evgueni Goudzovski “New Physics Searches at Kaon and Hyperon Factories (accepted for publication in Reports on Progress in Physics)”, 2022 arXiv:2201.07805 [hep-ph] * [23] Jason Aebischer, Andrzej J. Buras and Jacky Kumar “On the Importance of Rare Kaon Decays: A Snowmass 2021 White Paper” In _2022 Snowmass Summer Study_ , 2022 arXiv:2203.09524 [hep-ph] * [24] Andrzej J. Buras “New physics patterns in $\varepsilon^{\prime}/\varepsilon$ and $\varepsilon_{K}$ with implications for rare kaon decays and $\Delta M_{K}$” In _JHEP_ 04, 2016, pp. 071 DOI: 10.1007/JHEP04(2016)071 * [25] Jason Aebischer, Andrzej J. Buras and Jacky Kumar “Another SMEFT story: $Z^{\prime}$ facing new results on $\epsilon^{\prime}/\epsilon$, $\Delta M_{K}$ and $K\to\pi\nu\overline{\nu}$” In _JHEP_ 12, 2020, pp. 097 DOI: 10.1007/JHEP12(2020)097 * [26] Christoph Bobeth, Andrzej J. Buras, Alejandro Celis and Martin Jung “Yukawa enhancement of $Z$-mediated new physics in $\Delta S=2$ and $\Delta B=2$ processes” In _JHEP_ 07, 2017, pp. 124 DOI: 10.1007/JHEP07(2017)124 * [27] Motoi Endo, Teppei Kitahara and Daiki Ueda “SMEFT top-quark effects on $\Delta F=2$ observables” In _JHEP_ 07, 2019, pp. 182 DOI: 10.1007/JHEP07(2019)182 * [28] Motoi Endo, Teppei Kitahara, Satoshi Mishima and Kei Yamamoto “Revisiting Kaon Physics in General $Z$ Scenario” In _Phys. Lett. B_ 771, 2017, pp. 37–44 DOI: 10.1016/j.physletb.2017.05.026 * [29] Andrzej J. Buras, Fulvia De Fazio and Jennifer Girrbach “The Anatomy of Z’ and Z with Flavour Changing Neutral Currents in the Flavour Precision Era” In _JHEP_ 02, 2013, pp. 116 DOI: 10.1007/JHEP02(2013)116 * [30] Andrzej J. Buras, Fulvia De Fazio, Jennifer Girrbach and Maria V. Carlucci “The Anatomy of Quark Flavour Observables in 331 Models in the Flavour Precision Era” In _JHEP_ 02, 2013, pp. 023 DOI: 10.1007/JHEP02(2013)023 * [31] Jason Aebischer, Andrzej J. Buras, Maria Cerdà-Sevilla and Fulvia De Fazio “Quark-lepton connections in Z’ mediated FCNC processes: gauge anomaly cancellations at work” In _JHEP_ 02, 2020, pp. 183 DOI: 10.1007/JHEP02(2020)183 * [32] Andrzej J. Buras, Dario Buttazzo and Robert Knegjens “$K\to\pi\nu\overline{\nu}$ and $\varepsilon$’/$\varepsilon$ in simplified new physics models” In _JHEP_ 11, 2015, pp. 166 DOI: 10.1007/JHEP11(2015)166 * [33] Monika Blanke et al. “Rare and CP-Violating $K$ and $B$ Decays in the Littlest Higgs Model with $T^{-}$ Parity” In _JHEP_ 01, 2007, pp. 066 DOI: 10.1088/1126-6708/2007/01/066 * [34] Andrzej J. Buras, Anton Poschenrieder, Selma Uhlig and William A. Bardeen “Rare $K$ and $B$ Decays in the Littlest Higgs Model without $T^{-}$ Parity” In _JHEP_ 11, 2006, pp. 062 DOI: 10.1088/1126-6708/2006/11/062 * [35] Monika Blanke, Andrzej J. Buras and Stefan Recksiegel “Quark flavour observables in the Littlest Higgs model with T-parity after LHC Run 1” In _Eur. Phys. J. C_ 76.4, 2016, pp. 182 DOI: 10.1140/epjc/s10052-016-4019-7 * [36] Christoph Bobeth, Andrzej J. Buras, Alejandro Celis and Martin Jung “Patterns of Flavour Violation in Models with Vector-Like Quarks” In _JHEP_ 04, 2017, pp. 079 DOI: 10.1007/JHEP04(2017)079 * [37] Andrzej J. Buras, Thorsten Ewerth, Sebastian Jager and Janusz Rosiek “K+ —$>$ pi+ nu anti-nu and K(L) —$>$ pi0 nu anti-nu decays in the general MSSM” In _Nucl. Phys. B_ 714, 2005, pp. 103–136 DOI: 10.1016/j.nuclphysb.2005.02.014 * [38] Gino Isidori et al. “Exploring the flavour structure of the MSSM with rare K decays” In _JHEP_ 08, 2006, pp. 064 DOI: 10.1088/1126-6708/2006/08/064 * [39] Tomáš Blažek and Peter Maták “Left–left squark mixing, $K^{+}\rightarrow\pi^{+}\nu\bar{\nu}$ and minimal supersymmetry with large tan$\beta$” In _Int. J. Mod. Phys. A_ 29.27, 2014, pp. 1450162 DOI: 10.1142/S0217751X14501620 * [40] Wolfgang Altmannshofer et al. “Anatomy and Phenomenology of FCNC and CPV Effects in SUSY Theories” In _Nucl. Phys. B_ 830, 2010, pp. 17–94 DOI: 10.1016/j.nuclphysb.2009.12.019 * [41] Morimitsu Tanimoto and Kei Yamamoto “Probing SUSY with 10 TeV stop mass in rare decays and CP violation of kaon” In _PTEP_ 2016.12, 2016, pp. 123B02 DOI: 10.1093/ptep/ptw160 * [42] Teppei Kitahara, Ulrich Nierste and Paul Tremper “Supersymmetric Explanation of CP Violation in $K\to\pi\pi$ Decays” In _Phys. Rev. Lett._ 117.9, 2016, pp. 091802 DOI: 10.1103/PhysRevLett.117.091802 * [43] Motoi Endo, Satoshi Mishima, Daiki Ueda and Kei Yamamoto “Chargino contributions in light of recent $\epsilon^{\prime}/\epsilon$” In _Phys. Lett. B_ 762, 2016, pp. 493–497 DOI: 10.1016/j.physletb.2016.10.009 * [44] Andreas Crivellin, Giancarlo D’Ambrosio, Teppei Kitahara and Ulrich Nierste “$K\to\pi\nu\overline{\nu}$ in the MSSM in light of the $\epsilon^{\prime}_{K}/\epsilon_{K}$ anomaly” In _Phys. Rev. D_ 96.1, 2017, pp. 015023 DOI: 10.1103/PhysRevD.96.015023 * [45] Motoi Endo et al. “Gluino-mediated electroweak penguin with flavor-violating trilinear couplings” In _JHEP_ 04, 2018, pp. 019 DOI: 10.1007/JHEP04(2018)019 * [46] Chuan-Hung Chen and Takaaki Nomura “$Re(\epsilon^{\prime}_{K}/\epsilon_{K}$) and $K\to\pi\nu\bar{\nu}$ in a two-Higgs doublet model” In _JHEP_ 08, 2018, pp. 145 DOI: 10.1007/JHEP08(2018)145 * [47] Chuan-Hung Chen and Takaaki Nomura “$\epsilon^{\prime}/\epsilon$ from charged-Higgs-induced gluonic dipole operators” In _Phys. Lett. B_ 787, 2018, pp. 182–187 DOI: 10.1016/j.physletb.2018.11.006 * [48] Andrzej J. Buras, Michael Spranger and Andreas Weiler “The Impact of universal extra dimensions on the unitarity triangle and rare K and B decays” In _Nucl. Phys. B_ 660, 2003, pp. 225–268 DOI: 10.1016/S0550-3213(03)00250-5 * [49] Andrzej J. Buras, Anton Poschenrieder, Michael Spranger and Andreas Weiler “The Impact of universal extra dimensions on B —$>$ X(s) gamma, B —$>$ X(s) gluon, B —$>$ X(s) mu+ mu-, K(L) —$>$ pi0 e+ e- and epsilon-prime / epsilon” In _Nucl. Phys. B_ 678, 2004, pp. 455–490 DOI: 10.1016/j.nuclphysb.2003.11.010 * [50] Monika Blanke et al. “Rare K and B Decays in a Warped Extra Dimension with Custodial Protection” In _JHEP_ 03, 2009, pp. 108 DOI: 10.1088/1126-6708/2009/03/108 * [51] Bjoern Duling “$K$ and $B$ meson mixing in warped extra dimensions with custodial protection” In _J. Phys. Conf. Ser._ 171, 2009, pp. 012061 DOI: 10.1088/1742-6596/171/1/012061 * [52] Michaela E. Albrecht et al. “Electroweak and Flavour Structure of a Warped Extra Dimension with Custodial Protection” In _JHEP_ 09, 2009, pp. 064 DOI: 10.1088/1126-6708/2009/09/064 * [53] S. Casagrande et al. “Flavor Physics in the Randall-Sundrum Model: I. Theoretical Setup and Electroweak Precision Tests” In _JHEP_ 10, 2008, pp. 094 DOI: 10.1088/1126-6708/2008/10/094 * [54] M. Bauer, S. Casagrande, U. Haisch and M. Neubert “Flavor Physics in the Randall-Sundrum Model: II. Tree-Level Weak-Interaction Processes” In _JHEP_ 09, 2010, pp. 017 DOI: 10.1007/JHEP09(2010)017 * [55] Christoph Bobeth and Andrzej J. Buras “Leptoquarks meet $\varepsilon^{\prime}/\varepsilon$ and rare Kaon processes” In _JHEP_ 02, 2018, pp. 101 DOI: 10.1007/JHEP02(2018)101 * [56] S. Fajfer, N. Košnik and L. Vale Silva “Footprints of leptoquarks: from $R_{K^{(*)}}$ to $K\rightarrow\pi\nu\bar{\nu}$” In _Eur. Phys. J. C_ 78.4, 2018, pp. 275 DOI: 10.1140/epjc/s10052-018-5757-5 * [57] David Marzocca, Sokratis Trifinopoulos and Elena Venturini “From B-meson anomalies to Kaon physics with scalar leptoquarks” In _Eur. Phys. J. C_ 82.4, 2022, pp. 320 DOI: 10.1140/epjc/s10052-022-10271-7 * [58] Jason Aebischer, Christoph Bobeth, Andrzej J. Buras and David M. Straub “Anatomy of $\varepsilon^{\prime}/\varepsilon$ beyond the standard model” In _Eur. Phys. J. C_ 79.3, 2019, pp. 219 DOI: 10.1140/epjc/s10052-019-6715-6 * [59] Shinya Matsuzaki, Kenji Nishiwaki and Kei Yamamoto “Simultaneous interpretation of $K$ and $B$ anomalies in terms of chiral-flavorful vectors” In _JHEP_ 11, 2018, pp. 164 DOI: 10.1007/JHEP11(2018)164 * [60] Chuan-Hung Chen and Takaaki Nomura “$\epsilon_{K}$ and $\epsilon^{\prime}/\epsilon$ in a diquark model” In _JHEP_ 03, 2019, pp. 009 DOI: 10.1007/JHEP03(2019)009 * [61] Chuan-Hung Chen and Takaaki Nomura “Left-handed color-sextet diquark in the Kaon system” In _Phys. Rev. D_ 99.11, 2019, pp. 115006 DOI: 10.1103/PhysRevD.99.115006 * [62] Shinya Matsuzaki, Kenji Nishiwaki and Kei Yamamoto “Simultaneous explanation of $K$ and $B$ anomalies in vectorlike compositeness” In _PoS_ CORFU2018, 2019, pp. 031 DOI: 10.22323/1.347.0031 * [63] Andrzej J. Buras, Dario Buttazzo, Jennifer Girrbach-Noe and Robert Knegjens “${K}^{+}\to{\pi}^{+}\nu\overline{\nu}$ and ${K}_{L}\to{\pi}^{0}\nu\overline{\nu}$ in the Standard Model: status and perspectives” In _JHEP_ 11, 2015, pp. 033 DOI: 10.1007/JHEP11(2015)033 * [64] Andrzej J. Buras and Elena Venturini “Searching for New Physics in Rare $K$ and $B$ Decays without $|V_{cb}|$ and $|V_{ub}|$ Uncertainties” In _Acta Phys. Polon. B_ 53.6, 2021, pp. A1 * [65] Joachim Brod, Martin Gorbahn and Emmanuel Stamou “Updated Standard Model Prediction for $K\to\pi\nu\bar{\nu}$ and $\epsilon_{K}$” In _PoS_ BEAUTY2020, 2021, pp. 056 DOI: 10.22323/1.391.0056 * [66] Marzia Bordone, Dario Buttazzo, Gino Isidori and Joachim Monnard “Probing Lepton Flavour Universality with $K\to\pi\nu\bar{\nu}$ decays” In _Eur. Phys. J. C_ 77.9, 2017, pp. 618 DOI: 10.1140/epjc/s10052-017-5202-1 * [67] Eduardo Cortina Gil “Measurement of the very rare K+→${\pi}^{+}\nu\overline{\nu}$ decay” In _JHEP_ 06, 2021, pp. 093 DOI: 10.1007/JHEP06(2021)093 * [68] Frank F. Deppisch, Kåre Fridell and Julia Harz “Constraining lepton number violating interactions in rare kaon decays” In _JHEP_ 12, 2020, pp. 186 DOI: 10.1007/JHEP12(2020)186 * [69] Òscar L. Crosas et al. “Flavor Non-universal Vector Leptoquark Imprints in $K\to\pi\nu\bar{\nu}$ and $\Delta F=2$ Transitions”, 2022 arXiv:2207.00018 [hep-ph] * [70] J.. Ahn “Search for the $K_{L}\\!\to\\!\pi^{0}\nu\overline{\nu}$ and $K_{L}\\!\to\\!\pi^{0}X^{0}$ decays at the J-PARC KOTO experiment” In _Phys. Rev. Lett._ 122.2, 2019, pp. 021802 DOI: 10.1103/PhysRevLett.122.021802 * [71] “Searches for new physics with high-intensity kaon beams” In _2022 Snowmass Summer Study_ , 2022 arXiv:2204.13394 [hep-ex] * [72] Gino Isidori, Christopher Smith and Rene Unterdorfer “The Rare decay $K_{L}\to\pi^{0}\mu^{+}\mu^{-}$ within the SM” In _Eur. Phys. J. C_ 36, 2004, pp. 57–66 DOI: 10.1140/epjc/s2004-01879-0 * [73] Samuel Friot, David Greynat and Eduardo De Rafael “Rare kaon decays revisited” In _Phys. Lett. B_ 595, 2004, pp. 301–308 DOI: 10.1016/j.physletb.2004.05.069 * [74] Federico Mescia, Christopher Smith and Stephanie Trine “K(L) —$>$ pi0 e+ e- and K(L) —$>$ pi0 mu+ mu-: A Binary star on the stage of flavor physics” In _JHEP_ 08, 2006, pp. 088 DOI: 10.1088/1126-6708/2006/08/088 * [75] A. Alavi-Harati “Search for the rare decay K(L) —$>$ pi0 e+ e-” In _Phys. Rev. Lett._ 93, 2004, pp. 021805 DOI: 10.1103/PhysRevLett.93.021805 * [76] A. Alavi-Harati “Search for the Decay $K_{L}\to\pi^{0}\mu^{+}\mu^{-}$” In _Phys. Rev. Lett._ 84, 2000, pp. 5279–5282 DOI: 10.1103/PhysRevLett.84.5279 * [77] H.. Greenlee “Background to $K^{0}_{L}\to\pi^{0}ee$ From $K^{0}_{L}\to\gamma\gamma ee$” In _Phys. Rev. D_ 42, 1990, pp. 3724–3731 DOI: 10.1103/PhysRevD.42.3724 * [78] Andreas Crivellin, Giancarlo D’Ambrosio, Martin Hoferichter and Lewis C. Tunstall “Violation of lepton flavor and lepton flavor universality in rare kaon decays” In _Phys. Rev. D_ 93.7, 2016, pp. 074038 DOI: 10.1103/PhysRevD.93.074038 * [79] Giancarlo D’Ambrosio, David Greynat and Marc Knecht “On the amplitudes for the CP-conserving $K^{\pm}(K_{S})\to\pi^{\pm}(\pi^{0})\ell^{+}\ell^{-}$ rare decay modes” In _JHEP_ 02, 2019, pp. 049 DOI: 10.1007/JHEP02(2019)049 * [80] “A measurement of the $K^{+}\to\pi^{+}\mu^{+}\mu^{-}$ decay”, 2022 arXiv:2209.05076 [hep-ex] * [81] R. Appel “A New measurement of the properties of the rare decay K+ —$>$ pi+ e+ e-” In _Phys. Rev. Lett._ 83, 1999, pp. 4482–4485 DOI: 10.1103/PhysRevLett.83.4482 * [82] J.. Batley “Precise measurement of the K+- —$>$ pi+-e+e- decay” In _Phys. Lett. B_ 677, 2009, pp. 246–254 DOI: 10.1016/j.physletb.2009.05.040 * [83] H. Ma “A New measurement of the rare decay K+ —$>$ pi+ mu+ mu-” In _Phys. Rev. Lett._ 84, 2000, pp. 2580–2583 DOI: 10.1103/PhysRevLett.84.2580 * [84] J.. Batley “New measurement of the K+- –$>$ pi+-mu+mu- decay” In _Phys. Lett. B_ 697, 2011, pp. 107–115 DOI: 10.1016/j.physletb.2011.01.042 * [85] G. D’Ambrosio, A.. Iyer, F. Mahmoudi and S. Neshatpour “Anatomy of kaon decays and prospects for lepton flavour universality violation”, 2022 arXiv:2206.14748 [hep-ph] * [86] Andrzej J. Buras and Robert Fleischer “Quark mixing, CP violation and rare decays after the top quark discovery” In _Adv. Ser. Direct. High Energy Phys._ 15, 1998, pp. 65–238 DOI: 10.1142/9789812812667_0002 * [87] D. Ambrose “Improved branching ratio measurement for the decay K0(L) –$>$ mu+ mu-” In _Phys. Rev. Lett._ 84, 2000, pp. 1389–1392 DOI: 10.1103/PhysRevLett.84.1389 * [88] Giancarlo D’Ambrosio and Teppei Kitahara “Direct $CP$ Violation in $K\to\mu^{+}\mu^{-}$” In _Phys. Rev. Lett._ 119.20, 2017, pp. 201802 DOI: 10.1103/PhysRevLett.119.201802 * [89] Avital Dery, Mitrajyoti Ghosh, Yuval Grossman and Stefan Schacht “K → $\mu$+$\mu$- as a clean probe of short-distance physics” In _JHEP_ 07, 2021, pp. 103 DOI: 10.1007/JHEP07(2021)103 * [90] Roel Aaij “Constraints on the $K^{0}_{S}\rightarrow\mu^{+}\mu^{-}$ Branching Fraction” In _Phys. Rev. Lett._ 125.23, 2020, pp. 231801 DOI: 10.1103/PhysRevLett.125.231801 * [91] G. Ecker and A. Pich “The Longitudinal muon polarization in K(L) —$>$ mu+ mu-” In _Nucl. Phys. B_ 366, 1991, pp. 189–205 DOI: 10.1016/0550-3213(91)90056-4 * [92] A.. Alves Junior “Prospects for Measurements with Strange Hadrons at LHCb” In _JHEP_ 05, 2019, pp. 048 DOI: 10.1007/JHEP05(2019)048 * [93] B. Abi “Measurement of the Positive Muon Anomalous Magnetic Moment to 0.46 ppm” In _Phys. Rev. Lett._ 126.14, 2021, pp. 141801 DOI: 10.1103/PhysRevLett.126.141801 * [94] Antonio M. Coutinho, Andreas Crivellin and Claudio Andrea Manzari “Global Fit to Modified Neutrino Couplings and the Cabibbo-Angle Anomaly” In _Phys. Rev. Lett._ 125.7, 2020, pp. 071802 DOI: 10.1103/PhysRevLett.125.071802 * [95] Roel Aaij “Search for lepton-universality violation in $B^{+}\to K^{+}\ell^{+}\ell^{-}$ decays” In _Phys. Rev. Lett._ 122.19, 2019, pp. 191801 DOI: 10.1103/PhysRevLett.122.191801 * [96] Vincenzo Cirigliano and Ignasi Rosell “Two-loop effective theory analysis of pi (K) —$>$ e anti-nu/e [gamma] branching ratios” In _Phys. Rev. Lett._ 99, 2007, pp. 231801 DOI: 10.1103/PhysRevLett.99.231801 * [97] C. Lazzeroni “Precision Measurement of the Ratio of the Charged Kaon Leptonic Decay Rates” In _Phys. Lett. B_ 719, 2013, pp. 326–336 DOI: 10.1016/j.physletb.2013.01.037 * [98] A. Abada et al. “Tree-level lepton universality violation in the presence of sterile neutrinos: impact for $R_{K}$ and $R_{\pi}$” In _JHEP_ 02, 2013, pp. 048 DOI: 10.1007/JHEP02(2013)048 * [99] D.. Bryman and R. Shrock “Constraints on Sterile Neutrinos in the MeV to GeV Mass Range” In _Phys. Rev. D_ 100, 2019, pp. 073011 DOI: 10.1103/PhysRevD.100.073011 * [100] Andreas Crivellin, Dario Müller and Luc Schnell “Combined constraints on first generation leptoquarks” In _Phys. Rev. D_ 103.11, 2021, pp. 115023 DOI: 10.1103/PhysRevD.103.115023 * [101] Bernat Capdevila, Andreas Crivellin, Claudio Andrea Manzari and Marc Montull “Explaining $b\to s\ell^{+}\ell^{-}$ and the Cabibbo angle anomaly with a vector triplet” In _Phys. Rev. D_ 103.1, 2021, pp. 015032 DOI: 10.1103/PhysRevD.103.015032 * [102] Jennifer Girrbach and Ulrich Nierste “Gamma(K -$>$ e nu)/Gamma(K -$>$ mu nu) in the Minimal Supersymmetric Standard Model”, 2012 arXiv:1202.4906 [hep-ph] * [103] R.. Fonseca, J.. Romao and A.. Teixeira “Revisiting the $\Gamma(K\to e\nu)/\Gamma(K\to\mu\nu)$ Ratio in Supersymmetric Unified Models” In _Eur. Phys. J. C_ 72, 2012, pp. 2228 DOI: 10.1140/epjc/s10052-012-2228-2 * [104] V. Cirigliano et al. “Radiative corrections to K(l3) decays” In _Eur. Phys. J. C_ 23, 2002, pp. 121–133 DOI: 10.1007/s100520100825 * [105] V. Cirigliano, H. Neufeld and H. Pichl “K(e3) decays and CKM unitarity” In _Eur. Phys. J. C_ 35, 2004, pp. 53–65 DOI: 10.1140/epjc/s2004-01745-1 * [106] Vincenzo Cirigliano, Maurizio Giannotti and Helmut Neufeld “Electromagnetic effects in K(l3) decays” In _JHEP_ 11, 2008, pp. 006 DOI: 10.1088/1126-6708/2008/11/006 * [107] Chien-Yeah Seng, Daniel Galviz, Mikhail Gorchtein and Ulf G. Meißner “High-precision determination of the Ke3 radiative corrections” In _Phys. Lett. B_ 820, 2021, pp. 136522 DOI: 10.1016/j.physletb.2021.136522 * [108] Claudia Cornella, Paride Paradisi and Olcyr Sumensari “Hunting for ALPs with Lepton Flavor Violation” In _JHEP_ 01, 2020, pp. 158 DOI: 10.1007/JHEP01(2020)158 * [109] L.. Landsberg “Is it still worth searching for lepton flavor violation in rare kaon decays?” In _Phys. Atom. Nucl._ 68, 2005, pp. 1190–1210 DOI: 10.1134/1.1992575 * [110] Paul Langacker “The Physics of Heavy $Z^{\prime}$ Gauge Bosons” In _Rev. Mod. Phys._ 81, 2009, pp. 1199–1228 DOI: 10.1103/RevModPhys.81.1199 * [111] Jogesh C. Pati and Abdus Salam “Lepton Number as the Fourth Color” [Erratum: Phys.Rev.D 11, 703–703 (1975)] In _Phys. Rev. D_ 10, 1974, pp. 275–289 DOI: 10.1103/PhysRevD.10.275 * [112] Marzia Bordone, Claudia Cornella, Javier Fuentes-Martı́n and Gino Isidori “Low-energy signatures of the $\mathrm{PS}^{3}$ model: from $B$-physics anomalies to LFV” In _JHEP_ 10, 2018, pp. 148 DOI: 10.1007/JHEP10(2018)148 * [113] Sheldon L. Glashow, Diego Guadagnoli and Kenneth Lane “Lepton Flavor Violation in $B$ Decays?” In _Phys. Rev. Lett._ 114, 2015, pp. 091801 DOI: 10.1103/PhysRevLett.114.091801 * [114] Lorenzo Calibbi, Andreas Crivellin and Toshihiko Ota “Effective Field Theory Approach to $b\to s\ell\ell^{(^{\prime})}$, $B\to K^{(*)}\nu\overline{\nu}$ and $B\to D^{(*)}\tau\nu$ with Third Generation Couplings” In _Phys. Rev. Lett._ 115, 2015, pp. 181801 DOI: 10.1103/PhysRevLett.115.181801 * [115] Admir Greljo, Gino Isidori and David Marzocca “On the breaking of Lepton Flavor Universality in B decays” In _JHEP_ 07, 2015, pp. 142 DOI: 10.1007/JHEP07(2015)142 * [116] Peter Minkowski “$\mu\to e\gamma$ at a Rate of One Out of $10^{9}$ Muon Decays?” In _Phys. Lett. B_ 67, 1977, pp. 421–428 DOI: 10.1016/0370-2693(77)90435-X * [117] Murray Gell-Mann, Pierre Ramond and Richard Slansky “Complex Spinors and Unified Theories” In _Conf. Proc. C_ 790927, 1979, pp. 315–321 arXiv:1306.4669 [hep-th] * [118] Rabindra N. Mohapatra and Goran Senjanovic “Neutrino Mass and Spontaneous Parity Nonconservation” In _Phys. Rev. Lett._ 44, 1980, pp. 912 DOI: 10.1103/PhysRevLett.44.912 * [119] Eduardo Cortina Gil “Searches for lepton number violating $K^{+}$ decays” In _Phys. Lett. B_ 797, 2019, pp. 134794 DOI: 10.1016/j.physletb.2019.07.041 * [120] Eduardo Cortina Gil “Search for Lepton Number and Flavor Violation in $K^{+}$ and $\pi^{0}$ Decays” In _Phys. Rev. Lett._ 127.13, 2021, pp. 131802 DOI: 10.1103/PhysRevLett.127.131802 * [121] Eduardo Cortina Gil “Searches for lepton number violating K+ → $\pi$$-$($\pi$0)e+e+ decays” In _Phys. Lett. B_ 830, 2022, pp. 137172 DOI: 10.1016/j.physletb.2022.137172 * [122] Laurence S. Littenberg and Robert Shrock “Implications of improved upper bounds on |Delta L| = 2 processes” In _Phys. Lett. B_ 491, 2000, pp. 285–290 DOI: 10.1016/S0370-2693(00)01041-8 * [123] Anupama Atre, Vernon Barger and Tao Han “Upper bounds on lepton-number violating processes” In _Phys. Rev. D_ 71, 2005, pp. 113014 DOI: 10.1103/PhysRevD.71.113014 * [124] Anupama Atre, Tao Han, Silvia Pascoli and Bin Zhang “The Search for Heavy Majorana Neutrinos” In _JHEP_ 05, 2009, pp. 030 DOI: 10.1088/1126-6708/2009/05/030 * [125] Asmaa Abada et al. “Effective Majorana mass matrix from tau and pseudoscalar meson lepton number violating decays” In _JHEP_ 02, 2018, pp. 169 DOI: 10.1007/JHEP02(2018)169 * [126] D. Ambrose “New limit on muon and electron lepton number violation from K0(L) —$>$ mu+- e-+ decay” In _Phys. Rev. Lett._ 81, 1998, pp. 5734–5737 DOI: 10.1103/PhysRevLett.81.5734 * [127] Vincenzo Cirigliano et al. “Kaon Decays in the Standard Model” In _Rev. Mod. Phys._ 84, 2012, pp. 399 DOI: 10.1103/RevModPhys.84.399 * [128] Y. Aoki “FLAG Review 2021”, 2021 arXiv:2111.09849 [hep-lat] * [129] J.. Hardy and I.. Towner “Superallowed $0^{+}\to 0^{+}$ nuclear $\beta$ decays: 2020 critical survey, with implications for Vud and CKM unitarity” In _Phys. Rev. C_ 102.4, 2020, pp. 045501 DOI: 10.1103/PhysRevC.102.045501 * [130] Andrzej Czarnecki, William J. Marciano and Alberto Sirlin “Neutron Lifetime and Axial Coupling Connection” In _Phys. Rev. Lett._ 120.20, 2018, pp. 202002 DOI: 10.1103/PhysRevLett.120.202002 * [131] Chien-Yeah Seng, Xu Feng, Mikhail Gorchtein and Lu-Chang Jin “Joint lattice QCD–dispersion theory analysis confirms the quark-mixing top-row unitarity deficit” In _Phys. Rev. D_ 101.11, 2020, pp. 111301 DOI: 10.1103/PhysRevD.101.111301 * [132] N. Carrasco et al. “QED Corrections to Hadronic Processes in Lattice QCD” In _Phys. Rev. D_ 91.7, 2015, pp. 074506 DOI: 10.1103/PhysRevD.91.074506 * [133] P.. Boyle et al. “QED corrections to leptonic decay rates” In _PoS_ LATTICE2018, 2019, pp. 267 DOI: 10.22323/1.334.0267 * [134] Xu Feng, Luchang Jin and Michael Joseph Riberdy “Lattice QCD Calculation of the Pion Mass Splitting” In _Phys. Rev. Lett._ 128.5, 2022, pp. 052003 DOI: 10.1103/PhysRevLett.128.052003 * [135] Chien-Yeah Seng et al. “New method for calculating electromagnetic effects in semileptonic beta-decays of mesons” In _JHEP_ 10, 2020, pp. 179 DOI: 10.1007/JHEP10(2020)179 * [136] Chien-Yeah Seng, Daniel Galviz and Ulf-G Meißner “A New Theory Framework for the Electroweak Radiative Corrections in $K_{l3}$ Decays” In _JHEP_ 02, 2020, pp. 069 DOI: 10.1007/JHEP02(2020)069 * [137] Chien-Yeah Seng, Daniel Galviz, Mikhail Gorchtein and Ulf-G. Meißner “Improved Ke3 radiative corrections sharpen the Kμ2–Kl3 discrepancy” In _JHEP_ 11, 2021, pp. 172 DOI: 10.1007/JHEP11(2021)172 * [138] Latika Aggarwal “Snowmass White Paper: Belle II physics reach and plans for the next decade and beyond”, 2022 arXiv:2207.06307 [hep-ex] * [139] Asli M. Abdullahi “The Present and Future Status of Heavy Neutral Leptons” In _2022 Snowmass Summer Study_ , 2022 arXiv:2203.08039 [hep-ph] * [140] Gordan Krnjaic, Gustavo Marques-Tavares, Diego Redigolo and Kohsaku Tobioka “Probing Muonphilic Force Carriers and Dark Matter at Kaon Factories” In _Phys. Rev. Lett._ 124.4, 2020, pp. 041802 DOI: 10.1103/PhysRevLett.124.041802 * [141] Eduardo Cortina Gil “Search for a feebly interacting particle $X$ in the decay $K^{+}\rightarrow\pi^{+}X$” In _JHEP_ 03, 2021, pp. 058 DOI: 10.1007/JHEP03(2021)058 * [142] Eduardo Cortina Gil “Search for $\pi^{0}$ decays to invisible particles” In _JHEP_ 02, 2021, pp. 201 DOI: 10.1007/JHEP02(2021)201 * [143] Eduardo Cortina Gil “Search for heavy neutral lepton production in $K^{+}$ decays to positrons” In _Phys. Lett. B_ 807, 2020, pp. 135599 DOI: 10.1016/j.physletb.2020.135599 * [144] Eduardo Cortina Gil “Search for $K^{+}$ decays to a muon and invisible particles” In _Phys. Lett. B_ 816, 2021, pp. 136259 DOI: 10.1016/j.physletb.2021.136259 * [145] E. Minucci “First Results for Searches of Exotic Decays with NA62 in Beam-dump Mode” Contribution to ICHEP 2022 URL: https://agenda.infn.it/event/28874/contributions/169469/ * [146] Dipanwita Banerjee et al. “The North Experimental Area at the Cern Super Proton Synchrotron”, 2021 URL: https://cds.cern.ch/record/2774716 * [147] Lau Gatignon, Johannes Bernhard, Markus Brugger and Niels Doble “Report from the Conventional Beams Working Group to the Physics Beyond Collider Study and to the European Strategy for Particle Physics”, 2018 URL: https://cds.cern.ch/record/2650989 * [148] Yacine Kadi “Strategy for the Consolidation and Upgrade of the North Experimental Area”, 2018 URL: https://edms.cern.ch/document/2042932 * [149] Yacine Kadi “Addendum to the NA Consolidation Study Report”, 2019 URL: https://edms.cern.ch/document/2144442 * [150] Yacine Kadi “North Area Consolidation Project Roadmap”, 2021 URL: https://edms.cern.ch/document/2458866 * [151] Bruno Balhan et al. “Improvements to the SPS Slow Extraction for High Intensity Operation”, 2019 URL: https://cds.cern.ch/record/2668989 * [152] M.. Fraser “SPS Slow Extraction Losses and Activation: Challenges and Possibilities for Improvement” In _Proc. IPAC’17_ JACoW Publishing, Geneva, Switzerland, pp. 611–614 DOI: 10.18429/JACoW-IPAC2017-MOPIK045 * [153] M.. Fraser “SPS Slow Extraction Losses and Activation: Update on Recent Improvements” In _Proc. IPAC’19_ JACoW Publishing, Geneva, Switzerland, pp. 2391–2394 DOI: 10.18429/JACoW-IPAC2019-WEPMP031 * [154] Eirini Koukovini Platia et al. “Protons Beyond the LHC Injectors Upgrade Project”, 2019 URL: https://cds.cern.ch/record/2675225 * [155] Markus Brugger “Mandate of the PBC ECN3 Task Force”, 2022 URL: https://edms.cern.ch/document/2790130 * [156] Robert Barrie Appleby et al. “Considerations for a RF separated K+ beam for NA62”, 2021 URL: http://cds.cern.ch/record/2765940 * [157] Lau Gatignon “Design and Tuning of Secondary Beamlines in the CERN North and East Areas”, 2020 URL: https://cds.cern.ch/record/2730780 * [158] A.. Malensek “Empirical Formula for Thick Target Particle Production”, 1981 * [159] M.W.U Dijk and M. Rosenthal “Target studies for the proposed KLEVER experiment” CERN-ACC-NOTE-2018-0066, CERN-PBC-Notes-2018-002, KLEVER-PUB-18-01, 2018 * [160] J.. Bak “$e^{+}e^{-}$ Pair Creation by 40-GeV to 150-GeV Photons Incident Near the $\left<110\right>$ Axis in a Germanium Crystal” In _Phys. Lett._ B202, 1988, pp. 615–619 DOI: 10.1016/0370-2693(88)91874-6 * [161] J.. Kimball and N. Cue “QUANTUM ELECTRODYNAMICS AND CHANNELING IN CRYSTALS” In _Phys. Rept._ 125, 1985, pp. 69–101 DOI: 10.1016/0370-1573(85)90021-3 * [162] V.. Baryshevsky and V.. Tikhomirov “Synchrotron type radiation processes in crystals and polarization phenomena accompanying them” [Usp. Fiz. Nauk159,529(1989)] In _Sov. Phys. Usp._ 32, 1989, pp. 1013–1032 DOI: 10.1070/PU1989v032n11ABEH002778 * [163] M. Soldani “Experimental evidence of strong electromagnetic shower enhancement induced by high-energy photons in a thick oriented tungsten crystal”, 2022 arXiv:2203.07163 [hep-ex] * [164] A.. Artamonov “Study of the decay $K^{+}\to\pi^{+}\nu\bar{\nu}$ in the momentum region $140<P_{\pi}<199$ MeV/c” In _Phys. Rev. D_ 79, 2009, pp. 092004 DOI: 10.1103/PhysRevD.79.092004 * [165] Martin Wolfgang Winkler “Decay and detection of a light scalar boson mixing with the Higgs boson” In _Phys. Rev. D_ 99.1, 2019, pp. 015018 DOI: 10.1103/PhysRevD.99.015018 * [166] F. Bergsma “Search for Axion Like Particle Production in 400-GeV Proton - Copper Interactions” In _Phys. Lett. B_ 157, 1985, pp. 458–462 DOI: 10.1016/0370-2693(85)90400-9 * [167] J.. Batley “Searches for lepton number violation and resonances in $K^{\pm}\to\pi\mu\mu$ decays” In _Phys. Lett. B_ 769, 2017, pp. 67–76 DOI: 10.1016/j.physletb.2017.03.029 * [168] R. Aaij “Search for long-lived scalar particles in $B^{+}\to K^{+}\chi(\mu^{+}\mu^{-})$ decays” In _Phys. Rev. D_ 95.7, 2017, pp. 071101 DOI: 10.1103/PhysRevD.95.071101 * [169] Roel Aaij “Search for hidden-sector bosons in $B^{0}\\!\to K^{*0}\mu^{+}\mu^{-}$ decays” In _Phys. Rev. Lett._ 115.16, 2015, pp. 161802 DOI: 10.1103/PhysRevLett.115.161802 * [170] J.-T. Wei “Measurement of the Differential Branching Fraction and Forward-Backward Asymmetry for $B\to K^{(*)}\ell^{+}\ell^{-}$” In _Phys. Rev. Lett._ 103, 2009, pp. 171801 DOI: 10.1103/PhysRevLett.103.171801 * [171] T. Yamazaki “SEARCH FOR A NEUTRAL BOSON IN A TWO-BODY DECAY OF K+ —$>$ PI+ X0” In _Phys. Rev. Lett._ 52, 1984, pp. 1089–1091 DOI: 10.1103/PhysRevLett.52.1089 * [172] R. Ammar “Search for the familon via B+- —$>$ pi+- X0, B+- —$>$ K+- X0, and B0 —$>$ K0(S)X0 decays” In _Phys. Rev. Lett._ 87, 2001, pp. 271801 DOI: 10.1103/PhysRevLett.87.271801 * [173] P. Abratenko “Search for a Higgs Portal Scalar Decaying to Electron-Positron Pairs in the MicroBooNE Detector” In _Phys. Rev. Lett._ 127.15, 2021, pp. 151803 DOI: 10.1103/PhysRevLett.127.151803 * [174] Jean-Loup Tastet, Evgueni Goudzovski, Inar Timiryasov and Oleg Ruchayskiy “Projected NA62 sensitivity to heavy neutral lepton production in K+→$\pi$0e+N decays” In _Phys. Rev. D_ 104.5, 2021, pp. 055005 DOI: 10.1103/PhysRevD.104.055005 * [175] Eduardo Cortina Gil “The Beam and detector of the NA62 experiment at CERN” In _JINST_ 12.05, 2017, pp. P05025 DOI: 10.1088/1748-0221/12/05/P05025 * [176] F. Ambrosino “KLEVER: An experiment to measure BR($K_{L}\to\pi^{0}\nu\bar{\nu}$) at the CERN SPS”, 2019 arXiv:1901.03099 [hep-ex] * [177] J.. Ahn “Study of the $K_{L}\to\pi^{0}\nu\bar{\nu}$ Decay at the J-PARC KOTO Experiment” In _Phys. Rev. Lett._ 126.12, 2021, pp. 121801 DOI: 10.1103/PhysRevLett.126.121801 * [178] Marco Drewes, Jan Hajer, Juraj Klaric and Gaia Lanfranchi “NA62 sensitivity to heavy neutral leptons in the low scale seesaw model” In _JHEP_ 07, 2018, pp. 105 DOI: 10.1007/JHEP07(2018)105 * [179] Jan Jerhot et al. “ALPINIST: Axion-Like Particles In Numerous Interactions Simulated and Tabulated” In _JHEP_ 07, 2022, pp. 094 DOI: 10.1007/JHEP07(2022)094 * [180] Matthew J. Dolan et al. “Revised constraints and Belle II sensitivity for visible and invisible axion-like particles” [Erratum: JHEP 03, 190 (2021)] In _JHEP_ 12, 2017, pp. 094 DOI: 10.1007/JHEP12(2017)094 * [181] Evgueni Goudzovski “Development of the kaon tagging system for the NA62 experiment at CERN” In _Nucl. Instrum. Meth. A_ 801, 2015, pp. 86–94 DOI: 10.1016/j.nima.2015.08.015 * [182] A. Britting “Lifetime-issues of MCP-PMTs” In _JINST_ 6.10, 2011, pp. C10001 DOI: 10.1088/1748-0221/6/10/C10001 * [183] M. Pfaffinger “Recent results with lifetime enhanced microchannel-plate photomultipliers” In _Nucl. Instrum. Meth. A_ 912, 2018, pp. 123–127 DOI: 10.1016/j.nima.2017.10.084 * [184] Albert Lehmann “Status and Perspectives of Vacuum-based Photon Detectors”, Talk at 11th International Workshop on Ring Imaging Cherenkov Detectors (RICH2022), 2022 URL: https://indico.cern.ch/event/1094055/contributions/4974129/attachments/2508971/4311863/Lehmann_Talk_RICH2022_final.pdf * [185] S. Gómez “FastIC: a fast integrated circuit for the readout of high performance detectors” In _JINST_ 17.05, 2022, pp. C05027 DOI: 10.1088/1748-0221/17/05/C05027 * [186] Christiansen Jorgen, Horstmann Moritz, Perktold Lukas and Prinzie Jeffrey “picoTDC:Pico-second TDC for HEP”, Talk at Picosecond Workshop, Prague 2015, 2015 URL: https://indico.cern.ch/event/396441/contributions/1836631/attachments/794758/1089390/picoTDC_Prague.pdf * [187] G. Rinella et al. “The NA62 GigaTracKer: a low mass high intensity beam 4D tracker with 65 ps time resolution on tracks” In _Journal of Instrumentation_ 14.07, 2019, pp. P07010 DOI: 10.1088/1748-0221/14/07/P07010 * [188] A. Lai “A System Approach towards Future Trackers at High Luminosity Colliders: the TIMESPOT Project” In _2018 IEEE Nuclear Science Symposium and Medical Imaging Conference Proceedings (NSS/MIC)_ Sydney, NSW, Australia: IEEE, 2018, pp. 1–3 DOI: 10.1109/NSSMIC.2018.8824310 * [189] Adriano Lai “Timespot project testbeam results and simulations” talk given at "1st MONOLITH Workshop on silicon sensors for timing and their applications", Geneve, September 2022, https://indico.cern.ch/event/1179742/contributions/4961955/attachments/2502408/4298914/Timing_ALai_UniGe.pdf, 2022 * [190] Andrea Lampis “10ps timing with 3D trench silicon pixel sensors” talk given at "23rd IWORID", Riva del Garda, July 2022., https://indi.to/KPGtS, 2022 * [191] Sandro Cadeddu “Timespot1: A 28nm CMOS Pixel Read-Out ASIC for 4D tracking a High Rates”, 2022 DOI: 10.48550/arXiv.2209.13242. * [192] L. Anderlini et al. “Intrinsic time resolution of 3D-trench silicon pixels for charged particle detection” In _Journal of Instrumentation_ 15.09, 2020, pp. P09029 DOI: 10.1088/1748-0221/15/09/P09029 * [193] G. Iacobucci “Efficiency and time resolution of monolithic silicon pixel detectors in SiGe BiCMOS technology” In _JINST_ 17.02, 2022, pp. P02019 DOI: 10.1088/1748-0221/17/02/P02019 * [194] G. Pellegrini “Recent Technological Developments on LGAD and iLGAD Detectors for Tracking and Timing Applications” In _Nucl. Instrum. Meth. A_ 831, 2016, pp. 24–28 DOI: 10.1016/j.nima.2016.05.066 * [195] G. Paternoster et al. “Trench-Isolated Low Gain Avalanche Diodes (TI-LGADs)” In _IEEE Electron Device Letters_ 41.6, 2020, pp. 884–887 DOI: 10.1109/LED.2020.2991351 * [196] Christian Joram et al. “LHCb Scintillating Fibre Tracker Engineering Design Review Report: Fibres, Mats and Modules”, 2015 URL: https://cds.cern.ch/record/2004811 * [197] H. Danielsson et al. “New veto hodoscope ANTI-0 for the NA62 experiment at CERN” In _JINST_ 15.07, 2020, pp. C07007 DOI: 10.1088/1748-0221/15/07/C07007 * [198] M. Gandini “Efficient, fast and reabsorption-free perovskite nanocrystal-based sensitized plastic scintillators” In _Nat. Nanotechnol._ 15, 2020, pp. 462–468 DOI: 10.1038/s41565-020-0683-8 * [199] Atanu Jana et al. “Perovskite: Scintillators, direct detectors, and X-ray imagers” In _Materials Today_ 55, 2022, pp. 110–136 DOI: https://doi.org/10.1016/j.mattod.2022.04.009 * [200] K. Ahmet “The OPAL detector at LEP” In _Nucl. Instrum. Meth. A_ 305, 1991, pp. 275–319 DOI: 10.1016/0168-9002(91)90547-4 * [201] Eduardo Cortina Gil “An investigation of the very rare ${K}^{+}\to{\pi}^{+}\nu\overline{\nu}$ decay” In _JHEP_ 11, 2020, pp. 042 DOI: 10.1007/JHEP11(2020)042 * [202] F. Ambrosino “A Prototype large-angle photon veto detector for the P326 experiment at CERN” In _Proceedings, 2007 IEEE Nuclear Science Symposium and Medical Imaging Conference (NSS/MIC 2007): Honolulu, Hawaii, October 28-November 3, 2007_ 1, 2007, pp. 57–64 DOI: 10.1109/NSSMIC.2007.4436288 * [203] J.R. Comfort “KOPIO Project: Conceptual Design Report”, 2005 * [204] M.. Atiya “A detector to search for K+ —> pi+ neutrino anti-neutrino” In _Nucl. Instrum. Meth._ A321, 1992, pp. 129–151 DOI: 10.1016/0168-9002(92)90382-E * [205] J. Frank “Charged Kaons at the Main injector (CKM): A Proposal for a Precision Measurement of the Decay $K^{+}\to\pi^{+}\nu\bar{\nu}$ and other rare $K^{+}$ processes at Fermilab using the Main Injector”, 2001 DOI: 10.2172/878912 * [206] G. Anelli “Proposal to Measure the Rare Decay $K^{+}\to\pi^{+}\nu\bar{nu}$ at the CERN SPS” CERN-SPSC-2005-013/SPSC-P-326, 2005 URL: https://cds.cern.ch/record/832885 * [207] E. Ramberg, P. Cooper and R. Tschirhart “A photon veto detector for the CKM experiment” In _Real-time. Proceedings, 13th Conference on Computing Applications in Nuclear and Plasma Sciences, RT2003, Montreal, Canada, May 18-23, 2003_ 51, 2004, pp. 2201–2204 DOI: 10.1109/TNS.2004.836738 * [208] H. Nishiguchi “Development of an extremely thin-wall straw tracker operational in vacuum – The COMET straw tracker system” In _Nucl. Instrum. Meth. A_ 845, 2017, pp. 269–272 DOI: 10.1016/j.nima.2016.06.082 * [209] ECFA “The 2021 ECFA detector research and development roadmap”, 2021 DOI: 10.17181/CERN.XDPL.W2EX * [210] G. Anzivino et al. “Light detection system and time resolution of the NA62 RICH” In _Journal of Instrumentation_ 15.10, 2020, pp. P10025 DOI: 10.1088/1748-0221/15/10/P10025 * [211] G. Anzivino et al. “Precise mirror alignment and basic performance of the RICH detector of the NA62 experiment at CERN” In _Journal of Instrumentation_ 13.07, 2018, pp. P07012 DOI: 10.1088/1748-0221/13/07/P07012 * [212] V. Fanti “The Beam and detector for the NA48 neutral kaon CP violations experiment at CERN” In _Nucl. Instrum. Meth._ A574, 2007, pp. 433–471 DOI: 10.1016/j.nima.2007.01.178 * [213] NA62 Collaboration, 2007 * [214] G.. Atoian “An Improved Shashlyk Calorimeter” In _Nucl. Instrum. Meth._ A584, 2008, pp. 291–303 DOI: 10.1016/j.nima.2007.10.022 * [215] Katerina Decka “Timing performance of lead halide perovskite nanoscintillators embedded in a polystyrene matrix” In _J. Mater. Chem. C_ 10, 2022, pp. 12836 DOI: 10.1039/d2tc02060b * [216] Chao Liu “Transparent Ultra-High-Loading Quantum Dot/Polymer Nanocomposite Monolith for Gamma Scintillation” In _ACS Nano_ 11, 2017, pp. 6422–6430 DOI: 10.1021/acsnano.7b02923 * [217] F. Acerbi “Polysiloxane-based scintillators for shashlik calorimeters” In _Nucl. Instrum. Meth. A_ 956, 2020, pp. 163379 DOI: 10.1016/j.nima.2019.163379 * [218] Felix Sefkow and Frank Simon “A highly granular SiPM-on-tile calorimeter prototype” In _J. Phys. Conf. Ser._ 1162.1, 2019, pp. 012012 DOI: 10.1088/1742-6596/1162/1/012012 * [219] “The Phase-2 Upgrade of the CMS Endcap Calorimeter”, 2017 * [220] Naoki Tsuji et al. “Study on Granularity Optimization for ILD Hadron Calorimeter” In _JPS Conf. Proc._ 27, 2019, pp. 012015 DOI: 10.7566/JPSCP.27.012015 * [221] F. Bouyjou “HGCROC3: the front-end readout ASIC for the CMS High Granularity Calorimeter” In _JINST_ 17.03, 2022, pp. C03015 DOI: 10.1088/1748-0221/17/03/C03015 * [222] A. Frankenthal “Characterization and performance of PADME’s Cherenkov-based small-angle calorimeter” In _Nucl. Instrum. Meth. A_ 919, 2019, pp. 89–97 DOI: 10.1016/j.nima.2018.12.035 * [223] A. Cemmi “Radiation study of Lead Fluoride crystals”, 2021 arXiv:2107.12307 [physics.ins-det] * [224] Peter Kozma, Petr Kozma Jr. and R. Bajgar “Radiation resistivity of PbF-2 crystals” In _Nucl. Instrum. Meth. A_ 484, 2002, pp. 149–152 DOI: 10.1016/S0168-9002(01)02011-3 * [225] D.. Anderson, M. Kobayashi, Y. Yoshimura and C.. Woody “Lead Fluoride: An Ultracompact Cherenkov Radiator for EM Calorimetry” In _Nucl. Instrum. Meth. A_ 290, 1990, pp. 385–389 DOI: 10.1016/0168-9002(90)90553-I * [226] M. Kavatsyuk “Performance of the prototype of the electromagnetic calorimeter for PANDA” In _Nucl. Instrum. Meth. A_ 648, 2011, pp. 77–91 DOI: 10.1016/j.nima.2011.06.044 * [227] E. Auffray “Luminescence rise time in self-activated PbWO 4 and Ce-doped Gd 3 Al 2 Ga 3 O 12 scintillation crystals” In _J. Lumin._ 178, 2016, pp. 54–60 DOI: 10.1016/j.jlumin.2016.05.015 * [228] M. Follin et al. “Scintillating properties of today available lead tungstate crystals” In _JINST_ 16.08, 2021, pp. P08040 DOI: 10.1088/1748-0221/16/08/P08040 * [229] M. Korzhik “Ultrafast PWO scintillator for future high energy physics instrumentation” In _Nucl. Instrum. Meth. A_ 1034, 2022, pp. 166781 DOI: 10.1016/j.nima.2022.166781 * [230] S. Ceravolo “Crilin: A CRystal calorImeter with Longitudinal InformatioN for a future Muon Collider” In _JINST_ 17.09, 2022, pp. P09033 DOI: 10.1088/1748-0221/17/09/P09033 * [231] L. Bandiera, V. Haurylavets and V. Tikhomirov “Compact electromagnetic calorimeters based on oriented scintillator crystals” In _Nucl. Instrum. Meth. A_ 936, 2019, pp. 124–126 DOI: 10.1016/j.nima.2018.07.085 * [232] L. Bandiera “Strong reduction of the effective radiation length in an axially oriented scintillator crystal” In _Phys. Rev. Lett._ 121.2, 2018, pp. 021603 DOI: 10.1103/PhysRevLett.121.021603 * [233] Pietro Monti-Guarnieri “Beam test characterization of oriented crystals in strong field conditions” In _PoS_ ICHEP2022, 2022, pp. 342 DOI: 10.22323/1.414.0342 * [234] , https://www.alldatasheet.com/datasheet-pdf/pdf/575353/HAMAMATSU/S12572-015C.html * [235] , https://www.hamamatsu.com/content/dam/hamamatsu-photonics/sites/documents/99_SALES_LIBRARY/ssd/s13362_series_kapd1059e.pdf * [236] E.A. Kravchenko, V.V. Porosev and G.A. Savinov “Measurement of the time resolution of small SiPM-based scintillation counters” In _Journal of Instrumentation_ 12.12 IOP Publishing, 2017, pp. P12020–P12020 DOI: 10.1088/1748-0221/12/12/p12020 * [237] S Gundacker et al. “Time of flight positron emission tomography towards 100ps resolution with L(Y)SO: an experimental and theoretical analysis” In _Journal of Instrumentation_ 8.07 IOP Publishing, 2013, pp. P07014–P07014 DOI: 10.1088/1748-0221/8/07/p07014 * [238] “PicoTDC data sheet”, https://kt.cern/sites/default/files/technology/picotdc/tech-brief/picotdc_0.pdf * [239] D. Protopopescu “NA62 Grid Monte Carlo Production Tools” NA62-13-12, 2013 URL: https://na62.web.cern.ch/restricted/NotesDoc/NA62-13-12.pdf * [240] Alexios Charalambous “User Requirements for a High Intensity Beam Facility at TCC8/ECN3”, 2022 URL: https://edms.cern.ch/document/2791543
# Rigorous Assessment of Model Inference Accuracy using Language Cardinality Donato Clun 0000-0001-5190-8957 Imperial College LondonUK <EMAIL_ADDRESS>, Donghwan Shin 0000-0002-0840-6449 University of SheffieldUK<EMAIL_ADDRESS>, Antonio Filieri 0000-0001-9646-646X Imperial College LondonUK<EMAIL_ADDRESS>and Domenico Bianculli 0000-0002-4854-685X University of LuxembourgLuxembourg <EMAIL_ADDRESS> ###### Abstract. Models such as finite state automata are widely used to abstract the behavior of software systems by capturing the sequences of events observable during their execution. Nevertheless, models rarely exist in practice and, when they do, get easily outdated; moreover, manually building and maintaining models is costly and error-prone. As a result, a variety of model inference methods that automatically construct models from execution traces have been proposed to address these issues. However, performing a systematic and reliable accuracy assessment of inferred models remains an open problem. Even when a reference model is given, most existing model accuracy assessment methods may return misleading and biased results. This is mainly due to their reliance on statistical estimators over a finite number of randomly generated traces, introducing avoidable uncertainty about the estimation and being sensitive to the parameters of the random trace generative process. This paper addresses this problem by developing a systematic approach based on analytic combinatorics that minimizes bias and uncertainty in model accuracy assessment by replacing statistical estimation with deterministic accuracy measures. We experimentally demonstrate the consistency and applicability of our approach by assessing the accuracy of models inferred by state-of-the-art inference tools against reference models from established specification mining benchmarks. ## 1\. Introduction Software system models, typically in the form of Finite State Automata (FSA), have been widely used to abstract the behavior of software components and their interactions. Such models are important in many applications, including test data generation (Fraser and Walkinshaw, 2012), model checking (Clarke Jr et al., 2018), and program comprehension (Cook and Wolf, 1998). Nevertheless, they are rarely available during software development and, if they do, get easily outdated. This is mainly because manually building and maintaining such models is both time-consuming and error-prone. To address this problem, a variety of _model inference_ algorithms have been proposed (Biermann and Feldman, 1972; Beschastnikh et al., 2011; Walkinshaw et al., 2016; Emam and Miller, 2018; Mariani et al., 2017). These algorithms automatically extract a system’s behavior model by capturing visible events generated during the system’s execution. _Model assessment_ , i.e., assessing the accuracy of inferred models, is an essential task to evaluate and compare different model inference algorithms. When reference models are given as ground truth, model assessment seems straightforward; for example, one can compute the similarity between the languages defined by the inferred and reference models. However, the languages are often infinite, making it complex to measure the degree of their similarity. To address this issue, it is common to rely on statistical accuracy estimation. For example, the idea behind a popular method, called _trace similarity_ (Lo and Khoo, 2006a), is to generate sampled traces111A _trace_ is the sequence of events generated during the system’s execution. Also known as _run_ , _sequence_ , _execution trace._ by randomly traversing both reference and inferred models, and estimate the accuracy by checking how many of the traces generated using one model are accepted by the other. The idea is that a high number of accepted sampled traces is an indicator of a high degree of similarity. Although this idea is very intuitive, the random traverse can introduce a large degree of uncertainty about the estimation, depending on the parameters of the random trace generative process (e.g., the probability of ending the traverse in an accepting state rather than continuing through one of its outgoing transitions), and the topology of the model (e.g., a model may have features that are hard to exercise through random exploration). While there is no intrinsic probability distribution over the traces accepted by a finite state model, the random trace generation does induce such a distribution, therefore introducing an evaluation bias: the measured accuracy will reflect how well the inferred model can classify traces drawn according to that distribution. In an attempt to reduce the uncertainty due to the random traversal, an alternative consists in using deterministic trace generation methods (Walkinshaw et al., 2008). These generate sets of traces that are guaranteed to cover all the models’ behaviors without favoring any specific one. Although this increases the reproducibility and the chances of revealing discrepancies, there are still remaining issues. First, each discrepancy (a trace upon which the reference and the inferred model disagree) may be representative of a larger or smaller class of traces, thus making it difficult to quantify its impact. Second, these methods often generate sets of traces containing a disproportionately large amount of traces not accepted by the reference model, which may lead to skewed accuracy results. Third, the generated set of traces can become exceedingly large if the difference between the number of states in the reference and the inferred model is large, which is a common occurrence in model inference, hindering its practical usability. In this paper we propose a method to rigorously measure the accuracy of an inferred model against a reference one (both in the form of FSA) by considering all the possible traces up to an arbitrary finite maximum trace length. The maximum trace length considered is a parameter set by the user, which can be set to a value relevant for the application under analysis, to a value large enough to guarantee that all the behaviors of the reference and inferred models are exercised, or to an arbitrarily large value, effectively computing the asymptotic accuracy. Our method is deterministic (i.e., for a given pair of reference and inferred models it always returns the same result), does not depend on the model structure (i.e., assessing against the same reference model different models accepting the same language will generate the same accuracy measurement), and does not introduce any evaluation bias other than the maximum trace length (i.e., the computed metrics accurately describe how well the inferred model can classify traces drawn with uniform distribution from the set of all the traces up to the maximum length). Within this paper we also highlight how an inferred model can show a variable level of accuracy depending on the length of the traces used in the assessment. This is an aspect that is generally not considered by current assessment methodologies. Following this observation, we propose an additional assessment method that computes a pair of precision and recall values for each trace length, within a range specified by the user. Each value considers all the possible traces of a given length. Central to our solution is the use of analytic combinatorics to count the number of traces in a (possibly infinite) regular language, up to an arbitrary maximum trace length, without explicitly enumerating them. In this regard, to increase the practical applicability of our approach, we improved the scalability of the analytic combinatorics approaches currently used to compute the cardinality of regular languages (Flajolet and Sedgewick, 2009). We have implemented the proposed method in a prototype and experimentally evaluated it using reference models previously used in the model inference literature, and inferred models generated using well known model inference methods. The experimental results obtained when assessing the applicability of our method on the real-world model show that it is scalable enough to be used in practice. We have also compared the generated accuracy measurements with the results obtained using other popular assessment methods. The results indicate differences caused by the evaluation bias introduced by current assessment methods, and show that our proposed approach can be used to address this issue. In summary, this paper makes the following contributions: * • A novel model assessment method that measures the precision and recall values of an inferred model, with respect to a reference ground truth model, considering all the traces up to a finite maximum trace length. This method is: * – Deterministic: repeated executions return the same result. * – Comprehensive: it considers all the traces up to the maximum length. * – Unbiased: all the traces considered in the evaluation have the same weight on the result. * – Model-independent: assessing different models accepting the same language against the same reference model generates the same result. * • A further development of the assessment method, measuring the model accuracy separately for each trace length, over a given range. * • An empirical comparison of the accuracy measurement obtained using our methods with measurements generated using other popular assessment methods. * • An experimental evaluation of the applicability of our method. * • An improvement of currently used analytic combinatorics approaches to compute the cardinality of regular languages, up to a finite maximum trace length. The rest of the paper is organized as follows. In Section 2, we present a brief background on the theory of formal languages, model assessment, and analytic combinatorics as a tool to compute the cardinality of regular languages. In Section 3, we discuss two prominent classes of model assessment methods from the literature: statistical estimation (the most commonly used accuracy assessment method) and model-based (developed to address shortcomings of the first), reflecting on the open issues that motivated this work. Section 4 presents our main contribution: a novel assessment method, based on analytical measures for the cardinality of languages, that addresses the issues highlighted in the preceding section, and a further development that allows to evaluate how the model accuracy changes over a range of trace lengths. To increase the practical applicability of our method, in Section 5, we present an improvement of currently used analytic combinatorics approaches to compute the cardinality of regular languages. In Section 6 we evaluate our method experimentally, focusing on whether it is suitable to replace the assessment methods discussed in Section 3, and how the different methods’ assessment results compare. Section 7 presents relevant related work, and Section 8 concludes the paper with future work directions. ## 2\. Background ### 2.1. Models and Languages In this paper, we consider models in the form of Deterministic Finite-state Automata (DFA). A DFA is a tuple $\mathcal{A}=(\Sigma,Q,q_{0},F,\delta)$, where $\Sigma$ is a finite alphabet, $Q$ is the set of states, $q_{0}\in Q$ is the initial state, $F\subseteq Q$ is the set of accepting states, and $\delta:Q\times\Sigma\rightarrow Q$ is the transition function. A trace is a finite sequence $t=\langle\sigma_{1}\sigma_{2}\dots\sigma_{n}\rangle$ of elements $\sigma_{i}\in\Sigma$ for $i=1,\dots,n$. A trace $t=\langle\sigma_{1}\sigma_{2}\dots\sigma_{n}\rangle$ is _accepted_ by $\mathcal{A}$ if there exists a sequence of states $\langle q_{0},q_{1},\dots,q_{n}\rangle$ such that (1) $q_{i}\in Q$ for $i=1,\dots,n$, (2) $\delta(q_{i-1},\sigma_{i})=q_{i}$ for $i=1,\dots,n$, (3) $q_{0}$ is the initial state, and (4) $q_{n}\in F$. A state $q\in Q$ is an _error state_ if no accepting state is reachable from $q$. Let $\Sigma^{*}$ be the set of all possible traces over $\Sigma$ (including the empty trace). The language accepted by $\mathcal{A}$, denoted by $\mathfrak{L}(\mathcal{A})\subseteq\Sigma^{*}$, is the set of all traces accepted by $\mathcal{A}$. Two DFAs are equivalent if they accept the same language. DFAs accept regular languages, which are closed under union, intersection, and complement. These operations on regular languages correspond to analogous operations on the automata accepting the languages. Therefore, with abuse of notation, we will use the union ($\cdot\cup\cdot$), intersection ($\cdot\cap\cdot$), and complement ($\overline{\cdot}$) operators on both automata and the languages they accept; for example, $\mathcal{A}\cup\overline{\mathcal{B}}$ represents the DFA accepting any trace that is accepted by $\mathcal{A}$ or not accepted by $\mathcal{B}$. ### 2.2. Model Assessment as Language Comparison Given a reference model $\mathcal{R}$ and an inferred model $\mathcal{H}$ over the same alphabet $\Sigma$, to assess the accuracy of the inferred model we need a measure of how well the language accepted by $\mathcal{H}$ approximates the language accepted by $\mathcal{R}$. Drawing from established theory on classification assessment methods (Tharwat, 2020), $\Sigma^{*}$ can be partitioned into four subsets of traces: True Positives ($\Sigma^{*}_{\mathit{TP}}=\mathfrak{L}(\mathcal{R}\cap\mathcal{H})$), True Negatives ($\Sigma^{*}_{\mathit{TN}}=\mathfrak{L}(\overline{\mathcal{R}}\cap\overline{\mathcal{H}})$), False Positives ($\Sigma^{*}_{\mathit{FP}}=\mathfrak{L}(\overline{\mathcal{R}}\cap\mathcal{H})$), and False Negatives ($\Sigma^{*}_{\mathit{FN}}=\mathfrak{L}(\mathcal{R}\cap\overline{\mathcal{H}})$). By generalization, any set of traces $E\subseteq\Sigma^{*}$ (which we call evaluation set) can be partitioned into four subsets $E_{\mathit{TP}}$, $E_{\mathit{TN}}$, $E_{\mathit{FP}}$, and $E_{\mathit{FN}}$, containing true positives, true negatives, false positives and false negatives, respectively. If the evaluation set is finite, several accuracy metrics can be defined based on the _relative cardinality_ ($|\cdot|$) of these four languages. In this paper, we will focus on precision and recall: (1) $\mathit{precision}=\frac{|E_{\mathit{TP}}|}{|E_{\mathit{TP}}|+|E_{\mathit{FP}}|};\quad\textit{recall}=\frac{|E_{\mathit{TP}}|}{|E_{\mathit{TP}}|+|E_{\mathit{FN}}|}$ which are widely used in model inference literature (Krka et al., 2014; Lo et al., 2012; Lo and Khoo, 2006a). In the rest of this paper we will continue to use $\mathcal{R}$ and $\mathcal{H}$ to denote the reference and the inferred model respectively, both assumed to be defined over the same alphabet $\Sigma$. ### 2.3. Analytic Combinatorics and Cardinality of Regular Languages Analytic combinatorics (Flajolet and Sedgewick, 2009) is a theory used to build, manipulate and analyze exact enumerative descriptions of combinatorial structures, typically focusing on structures whose realizations are too many for explicit enumeration. _Generating functions_ are a core tool in analytic combinatorics, as they enable to concisely define and operate on certain infinite sequences. Let us consider a discrete, possibly infinite sequence of real values $a_{n}$, with $n=0,1,2,\cdots$. An _ordinary generating function_ (OGF) is a mathematical object encoding the sequence as the coefficients of a formal power series: $f(z)=a_{0}+a_{1}z+a_{2}z^{2}+a_{3}z^{3}+\cdots=\sum_{n=0}^{\infty}a_{n}z^{n}$. Given an OGF $f(z)$, the $n$-th coefficient $a_{n}$ can be retrieved as the $n$-th Taylor coefficient of $f(z)$, i.e., $a_{n}=\frac{1}{n!}\frac{\partial^{n}}{\partial z^{n}}f(z)\big{|}_{z=0}$. A regular language $L$, accepted by a DFA $\mathcal{A}$, is a combinatorial structure representing all the traces belonging to the language. The _sequence of cardinalities_ of $L$ is the sequence of number of traces $a_{n}$ of length $n=0,1,2,\cdots$ accepted by $\mathcal{A}$, i.e., the cardinality of the language accepted by $\mathcal{A}$ restricted to traces of length $n$, $\mathfrak{L}(\mathcal{A})\cap\Sigma^{n}$. For our purpose, we aim at constructing an OGF $\mathfrak{G}_{L}(z)$ encoding in a compact way the sequence of cardinalities of $L$. The OGF for the sequence of cardinalities of a regular language is always a rational function (Flajolet and Sedgewick, 2009). For example, let us consider the language $L=\Sigma^{*}$, with $\Sigma=\\{0,1\\}$. The empty trace, with length 0, is accepted ($a_{0}=1$); two traces of length 1 are accepted ($a_{1}=2$); four of length 2 ($a_{2}=4$); and so on. The infinite sequence $a_{n}=\\{1,2,4,8,16,\cdots\\}$ can be compactly encoded by the OGF $\mathfrak{G}_{L}(z)=\frac{1}{1-2z}$. Given a regular language $L$, the construction of the OGF $\mathfrak{G}_{L}(z)$ is typically reduced to the problem of counting the number of accepting paths of length $n$ of the minimal automaton accepting $L$, via the established _transfer matrix method_ (Flajolet and Sedgewick, 2009; Stanley, 2011). This algorithm relies on linear algebraic operations on the matrix representing the transition relation of the automaton (Flajolet and Sedgewick, 2009, Proposition I.3). While the transfer matrix method works in general, in Section 5 we will present an alternative, equivalent algorithm to obtain the OGF of a regular language that significantly outperformed the transfer matrix method in our tests. Once the OGF $\mathfrak{G}_{L}(z)$ is known, the values of the sequence $a_{i},i\geq 0$ can be recovered as the Taylor coefficients of $\mathfrak{G}_{L}(z)$. Since the OGF for the sequence of cardinalities of a regular language is a rational function, however, it is possible (and faster) to derive a recurrence relation that allows us to compute the values of the sequence without performing symbolic differentiation, using the following method. Let the OGF $A(z)$ be defined as $A(z)=N(z)/D(z)$, where $N(z)$ and $D(z)$ are polynomials of maximum degree $m$. Then (2) $\frac{N(z)}{D(z)}=\frac{b_{0}+b_{1}z+\cdots+b_{m}z^{m}}{c_{0}+c_{1}z+\cdots+c_{m}z^{m}}=\sum_{i=0}^{\infty}a_{i}z^{i}$ By multiplying by $D(z)$ and expanding the sum, one obtains: (3) $b_{0}+b_{1}z+\cdots+b_{m}z^{m}=(c_{0}+c_{1}z+\cdots+c_{m}z^{m})(a_{0}+a_{1}z+a_{2}z^{2}+\cdots)$ and by equating the coefficients one obtains a set of equations that recursively defines the sequence (4) $a_{n}=\begin{cases}\frac{b_{n}-\sum_{i=1}^{n}c_{i}a_{n-i}}{c_{0}}&\text{if $n\leq m$}\\\ \frac{-\sum_{i=1}^{m}c_{i}a_{n-i}}{c_{0}}&\text{if $n>m$}\end{cases}$ ## 3\. Existing methods for Model Assessment Following Equation (1), when a finite evaluation set of traces $E$ is available, measuring the accuracy of $\mathcal{H}$ against $\mathcal{R}$ using precision and recall can be achieved by comparing the acceptance of each trace in $E$. However, obtaining a finite and representative evaluation set over $\Sigma^{*}$ is a challenging problem. In this section, we discuss the two prominent classes of methods from the literature to generate evaluation sets for model assessment. Methods in the first class — _Statistical Estimation_ — generate the evaluation set of traces by means of a random walk over the automata’s paths. Methods in the second class — _Model-Based_ — use a deterministic traversal of the automata’s transition relation to generate an evaluation set that is likely to reveal divergences between the reference model and the inferred one. ### 3.1. Statistical Estimation using Random Walks Statistical estimation (Busany et al., 2019; Lo et al., 2012, 2009; Walkinshaw and Bogdanov, 2013; Walkinshaw et al., 2013; Mariani et al., 2010) represents a class of model assessment methods based on finite evaluation multisets (denoted by $E$) of traces randomly generated (sampled) over $\Sigma^{*}$. The random process used to generate $E$ is the characterizing element of each method. The most popular statistical estimation method is _trace similarity_ , proposed by Lo and Khoo (2006a), which generates $E$ by performing _random walks_ on the reference and inferred models, rather than generating random sequences from $\Sigma^{*}$. The core random walk algorithm is defined as follows, starting from the initial state. If the current state is accepting, the procedure randomly decides whether to terminate the walk or to continue; if the walk is not terminated, it randomly selects one among the outgoing transitions from the current state, adds the transition symbol to the trace, updates the current state to the target of the transition, and then the random walk continues from the new current state. If an error state222As defined in Section 2.1, an error state is a state from which no accepting state can be reached. is reached, the current trace is discarded and the random walk restarts from the initial state. To compute the _precision_ using trace similarity, an evaluation multiset $E$ is generated by performing random walks on the inferred model $\mathcal{H}$, therefore generating traces that are either true positives or false positives. Precision is then defined as the ratio of traces in $E$ accepted also by the reference model $\mathcal{R}$, i.e. the proportion of true positives333The rationale for this process stems from the observation that $\frac{|TP|}{|TP|+|FP|}=\frac{|\mathfrak{L}(\mathcal{R}\cap\mathcal{H})|}{|\mathfrak{L}(\mathcal{R}\cap\mathcal{H})|+|\mathfrak{L}(\overline{\mathcal{R}}\cap\mathcal{H})|}=\frac{|\mathfrak{L}(\mathcal{R}\cap\mathcal{H})|}{|\mathfrak{L}(\mathcal{H})|}$, which on the surface reminds of the conditional probability of a trace being accepted by $\mathcal{R}$ given that the trace is accepted by $\mathcal{H}$. We shall demonstrate in the remaining of this section that this conditional probability interpretation is itself conditional on the random walk process, which may render the computed precision meaningless.. Similarly, to compute the _recall_ , the evaluation multiset $E$ is generated by performing random walks on the reference model $\mathcal{R}$, generating traces that are either true positives or false negatives. Recall is then defined as the proportion of traces in $E$ accepted also by the inferred model $\mathcal{H}$. In both cases, model coverage can be achieved by repeated, independent random walks, e.g., until each transition is traversed a minimum number of times. It is important to note that, when Lo and Khoo (2006a) originally proposed the trace similarity method, it was assumed either (a) that the models are probabilistic and the random walk is performed according to the transition probabilities of the models, or (b) that the models are deterministic and the user specifies how the random choices necessary to generate traces are made (defaulting to a uniform distribution over the outgoing transitions if not otherwise specified). The transition probability distribution is an important aspect of the assessment because, as we will soon discuss, it affects the results. In the model inference literature, the use of trace similarity on deterministic models, i.e., whose transition probabilities are not specified by the model itself, is predominant. This is because the inference methods that are most well-known (e.g., K-tail (Chow, 1978)) or considered state-of- the-art (e.g., MINT (Walkinshaw et al., 2016)) generate deterministic models. Moreover, the reference models used in the assessment are also most often deterministic, because they are either manually created (for example, from API documentation and reference books (Pradel et al., 2010)) or they come from a formal specification of the target language, thus with no information about the probability distribution of the traces. On the other hand, a random walk algorithm does require a probability distribution to select which transition to traverse at each step, and to decide whether to continue or to terminate the random walk when an accepting state is reached. To decide which transition to follow, it is common practice to select one of the possible alternatives with uniform probability distribution (Lo and Khoo, 2006a). For the decision of whether to continue or terminate the random walk once a final state is reached, a number of strategies are commonly used; e.g., Walkinshaw et al. (2013) define a termination probability that is inversely proportional to the number of outgoing transitions. A factor that is often overlooked in the statistical estimation of the model accuracy is how the (arbitrary) randomness of the trace generation impacts the model assessment. Inferred and reference models merely represent languages, i.e., (possibly infinite) sets of traces over a finite alphabet. There is no intrinsic probability distribution over the traces of a finite state model. On the other hand, the random trace generation process used to produce the evaluation multiset (e.g., a random walk, in the case of trace similarity) induces a probability distribution over the accepted language. This distribution may be non-uniform (i.e., different traces may be generated with different probabilities) therefore introducing a _sampling bias_ in the model assessment: the precision and recall values obtained using this random sampling will reflect how well the inferred model can classify traces drawn according to that distribution. Unless this distribution reflects domain knowledge about the application being analyzed (i.e., if the models are purposefully probabilistic, and they describe the probability that different traces have of being generated by the system under analysis), this sampling bias is not desirable. For example, let us consider a random walk using a fixed termination probability $p_{a}$ to decide the trace termination, and uniform sampling among the available alternatives to decide which transition to follow. Intuitively, the termination probability has an impact on the length of the sampled traces: increasing $p_{a}$ makes shorter traces more likely to be sampled (correspondingly, longer traces less likely). However, it is hard to predict the actual length distribution without taking into account the specific topology of the model from which the traces are sampled. Moreover, the topology of the model also induces a non-uniform distribution among accepted traces of the same length: the likelihood of a trace depends on the probability of selecting each of its transitions in a given order, with each of these selections generally depending on the number of successors of the source state. We have thus identified two sampling biases in the random walk: * • shorter traces are more likely to be sampled (but exactly how likely is topology-dependent); * • for a given trace length, some traces are more likely to be sampled than others. When a random walk is used to generate the evaluation multiset for trace similarity, both sampling biases become systematic faults of the method, and their impact on the model assessment is difficult to quantify a priori since both depend on specific random walk parameters and the topology of the model to which the random walk is applied. Furthermore, since for trace similarity methods precision and recall are measured using random walks on different models (i.e., $\mathcal{H}$ for precision, $\mathcal{R}$ for recall), the assessment mixes two _different sampling biases_ , making the two measures incomparable and generally impossible to aggregate. (a) Reference model $\mathcal{R}$. (b) Inferred model $\mathcal{H}$. Figure 1. Example of reference and inferred models Let us show with an example how sensitive the model’s accuracy calculated by trace similarity is to the choice of random walk parameters. We take a concrete example based on the reference model Signature, adapted from the benchmark models provided in Krka et al. (2014). For readability, we used single symbols on transition labels and omitted transitions to error states. The reference model $\mathcal{R}$ is shown in Figure 1(a), while Figure 1(b) shows the inferred model $\mathcal{H}$, obtained with MINT (Walkinshaw et al., 2016) — a state-of-the-art model inference tool — with its default configuration, on a set of 100 traces accepted by $\mathcal{R}$. To see the impact of different parameters of random walks on the precision of $\mathcal{H}$, we vary the termination probability (denoted with $p_{a}$) between $0.01$ and $1$, while the choice of the outgoing transition is always done uniformly among the available alternatives. For each value of $p_{a}$, 50,000 random traces are generated to assess the precision of $\mathcal{H}$. Figure 2 shows the result. When $p_{a}=1$, the random walk always generates a trace of length zero (since the empty string is accepted by $\mathcal{H}$), which is a true positive, and therefore the precision value converges to 1. On the contrary, as $p_{a}$ decreases, longer traces are generated and the precision value converges to 0.2. In this simple example, this range of values could have been predicted analytically because, for traces of length $l$ ($l\geq 2$), the number of true positives is $5^{(l-2)}$ (the set of true positives is the language determined by the regular expression $a\cdot b\cdot\\{b,c,d,e,f\\}^{*}$) while the number of false positives is $4\cdot 5^{(l-2)}$ (the set of false positives is the language determined by the regular expression $a\cdot\\{c,d,e,f\\}\cdot\\{b,c,d,e,f\\}^{*}$), and therefore $\lim_{l\to\infty}\frac{5^{(l-2)}}{5^{(l-2)}+4\cdot 5^{(l-2)}}=0.2$. As a result, we can see that the precision value varies between 0.2 and 1.0 depending on the value of $p_{a}$, implying that the result of trace similarity is extremely sensitive to the choice of the random walk parametrization. Unlike in this simple example, however, the effect of the random walk on the assessment result is in general difficult to predict analytically. Figure 2. Sensitivity of statistical estimation to changes in the random walk for the models in Figure 1 #### 3.1.1. Statistical Estimation by Sampling $\Sigma^{*}$ The sampling bias caused by the random trace generation is fundamentally unavoidable, because the sampled languages are in general infinite and there exists no uniform distribution over a discrete infinite set. There are, however, alternative methods to generate a random evaluation multiset $E$ (and then compute precision and recall using equation (1)) that allow for more control over the sampling bias. The simplest method consists in generating traces by combining symbols randomly selected from the alphabet $\Sigma$, without considering the models under analysis, thus making the sampling bias model-independent. This can be done, for example, by firstly selecting a random trace length $l$, and then concatenating $l$ symbols, each one randomly selected from $\Sigma$. The generated trace is then added to the evaluation multiset $E$, and the process is repeated until a stopping condition is met (e.g., a certain number of traces has been generated). Note that, when the maximum length of a sample trace $n$ is fixed, it is possible to perform this trace generation in a way that ensures uniform probability distribution over $\Sigma^{\leq n}$: each trace length $l\leq n$ must be chosen with probability proportional to $|\Sigma^{l}|$, and each symbol must be chosen with uniform distribution. Despite its simplicity and its ability to reduce the sampling bias, compared to methods based on random walks like trace similarity, sampling from $\Sigma^{n}$ is rarely used in practice due to its computational cost: for most practical problems and large values of $n$, the proportion of $\Sigma^{n}$ accepted by either $\mathcal{H}$ or $\mathcal{R}$ tends to decrease with $n$, leading to most samples being true negatives, whose number does not contribute to the computation of either precision or recall. We will show instances of this problem experimentally in Section 6. ### 3.2. Model-Based Assessment Walkinshaw et al. (Walkinshaw et al., 2008) proposed a new method based on methodologies from model-based testing (MBT), aimed at mitigating the sampling bias introduced by the random sampling and increasing the reproducibility of the results, as it is deterministic and thus not affected by the statistical uncertainty that comes from using a finite randomly generated evaluation set. The intuition is that MBT methods can deterministically generate a set of traces that is _comprehensive_ (in the sense that any erroneous behavior of the inferred model would be detected) that can be used as evaluation set. There are multiple MBT methods for DFAs (Chow, 1978; Bonifácio et al., 2008) that, given a reference model $\mathcal{R}$ and an upper bound on the number of states the inferred model ($\mathcal{H}$ in our case) is allowed to contain, can check the equivalence of $\mathcal{R}$ and $\mathcal{H}$ by testing them on a finite set of traces, i.e., it can generate a set of traces $T$ such that, ideally, $\mathfrak{L}(\mathcal{R})=\mathfrak{L}(\mathcal{H})\Leftrightarrow\forall t\in T:\mathcal{R}(t)=\mathcal{H}(t)$. The idea proposed by Walkinshaw et al. (Walkinshaw et al., 2008) is to compute precision and recall using the set of traces $T$ (generated using the MBT method of choice) as evaluation set $E$. As an example, let us describe how the W-method (Chow, 1978), a widely-used MBT method for DFA, generates a set of tests $T$. Given the reference model $\mathcal{R}=(\Sigma,Q_{\mathcal{R}},q_{i_{\mathcal{R}}},F_{\mathcal{R}},\delta_{\mathcal{R}})$ and an upper bound to the number of states of $\mathcal{H}$ denoted with $m$, it constructs two sets of traces: a state cover (denoted with $C_{\mathcal{R}}$) and a characterization set (denoted with $D_{\mathcal{R}}$). The state cover $C_{\mathcal{R}}$ is a prefix-closed subset of $\Sigma^{*}$ containing traces reaching each state of $\mathcal{R}$, i.e., $\forall q\in Q_{\mathcal{R}}\;\exists t\in C_{\mathcal{R}}:\delta_{\mathcal{R}}(q_{i_{\mathcal{R}}},t)=q$. The characterization set $D_{\mathcal{R}}$ is a subset of $\Sigma^{*}$ such that, for any pair of states $q_{a},q_{b}\in Q_{\mathcal{R}}$ with $q_{a}\neq q_{b}$, it contains a distinguishing trace $t$ such that $\delta_{\mathcal{R}}(q_{a},t)$ is an accepting state but $\delta_{\mathcal{R}}(q_{b},t)$ is not, or vice versa. The test set $T$ generated by the W-method is then defined as $T=C_{\mathcal{R}}(\\{\epsilon\\}\cup\Sigma\cup\Sigma^{2}\cup\dots\cup\Sigma^{k+1})D_{\mathcal{R}}$, where $k=m-|Q_{\mathcal{R}}|$ and $T_{a}T_{b}=\\{t_{a}t_{b}\mid t_{a}\in T_{a},t_{b}\in T_{b}\\}$ for two sets of traces $T_{a}$ and $T_{b}$. The comprehensiveness of $T$ ensures that any erroneous behavior in the inferred model will affect the accuracy metrics, and the way in which $T$ is generated does not favor specific parts of the model under analysis, leading to less skewed accuracy results when compared to statistical estimation. Nevertheless, using $T$ to measure the accuracy of an inferred model has three major shortcomings. First, there are multiple approaches in the area of MBT for DFA with different aims. For example, the Wp-method (Fujiwara et al., 1991) is a further development of the W-method aimed at reducing the size of $T$ while providing the same guarantees. Since different approaches will generate different sets of tests $T$, usually with different number of tests accepted or rejected by either of the models, they will also lead to different accuracy results. Second, as already noted by Walkinshaw et al. (2008), MBT methods often generate a set $T$ containing a disproportionately large amount of traces not accepted by the reference model, which may lead to skewed accuracy results. Furthermore, $T$ can become exceedingly large if the difference between the number of states of $\mathcal{R}$ and the number of states of $\mathcal{H}$ is large. For example, in the W-method described above, the cardinality of $T$ grows with $|\Sigma^{k+1}|$, where $k$ is the difference between the number of states. This is a significant issue especially in the context of model inference from positive examples only, where it is not unusual to obtain inferred models having a number of states that is one order of magnitude larger than the number of states of the reference model, making this type of assessment infeasible. This will also be shown experimentally in Section 6. Third, although MBT guarantees that the generated $T$ will certainly contain a counterexample trace highlighting any incorrect behavior of the inferred model, each counterexample trace has the same “weight” on the computed accuracy metrics, even though it could represent a smaller or larger class of errors. Let us consider the example shown in Figure 3; it contains a reference model $\mathcal{R}$ (Figure 3(a)) and two incorrect models $\mathcal{H}_{1}$ (Figure 3(b)) and $\mathcal{H}_{2}$ (figure 3(c)), which were manually created introducing the erroneous transitions highlighted in red. Both $\mathcal{H}_{1}$ and $\mathcal{H}_{2}$ have only false negative errors. (a) Reference model $\mathcal{R}$ (b) An incorrect model $\mathcal{H}_{1}$ (c) Another incorrect model $\mathcal{H}_{2}$ Figure 3. A example case where the assessment based on the W-method would give misleading results. By enumerating all traces we verified that, for any trace length, $\mathcal{H}_{1}$ has fewer false negatives than $\mathcal{H}_{2}$ (e.g., for traces of length less than or equal to 10, $\mathcal{H}_{1}$ has 11 false negative traces, while $\mathcal{H}_{2}$ has 64 false negative traces). However, when the model accuracy is evaluated using the set of traces $T$ generated by the W-method, $\mathcal{H}_{1}$ has 96% recall while $\mathcal{H}_{2}$ has 98% recall (precision is 100% in both cases). This happens because individual false negative traces in $T$ do not represent an equal number of false negative traces in the inferred languages. ## 4\. Measuring Accuracy with Trace Counting To overcome the limitations of statistical and model-based accuracy evaluations, we propose to compute precision and recall, using analytical measures for the cardinality of the languages of true positives, false positives and false negatives — normally, restricted up to a finite maximum trace length to handle languages with unbounded traces. In particular, the proposed cardinality measures should be: 1) _deterministic_ , making the process repeatable and avoiding the convergence limitations of statistical methods; 2) _comprehensive_ , accounting for every trace; 3) _unbiased_ , giving each trace the same weight on the result; 4) _model- independent_ , generating the same accuracy measurement on all models accepting the same language. Notably, a statistical estimation in which all the traces (up to the prescribed maximum length) are sampled uniformly with the same probability (such as via the model-independent random sampling from $\Sigma^{*}$ discussed in Section 3.1.1) would converge to the same values we propose to compute analytically. This will be shown experimentally in Section 6. ### 4.1. Language Cardinality Measures We propose to compute two classes of cardinality-based measures to evaluate the accuracy of an inferred model (the hypothesis $\mathcal{H}$) against a reference model $\mathcal{R}$. The first step is the construction of the languages corresponding to the definitions of true positive, false positive, and false negative, as required for the computation of precision and recall. These languages will be accepted, respectively, by the automata $\mathcal{A}_{TP}$, $\mathcal{A}_{FP}$, and $\mathcal{A}_{FN}$ defined as: (5) $\mathcal{A}_{TP}=\mathcal{R}\cap\mathcal{H};\quad\mathcal{A}_{FP}=\overline{\mathcal{R}}\cap\mathcal{H};\quad\mathcal{A}_{FN}=\mathcal{R}\cap\overline{\mathcal{H}}$ We can then use analytic combinatorics methods, as described in Section 2.3, to obtain the sequences $\mathit{tp}_{n}$, $\mathit{fp}_{n}$, $\mathit{fn}_{n}$ of the number of traces of length $n=0,1,2,\cdots$ accepted by $\mathcal{A}_{TP}$, $\mathcal{A}_{FP}$, and $\mathcal{A}_{FN}$ respectively — the _cardinality of the languages_ accepted by the three automata intersected with the evaluation set $E=\Sigma^{n}$. Finally, we compose these elementary cardinality measures to compute the derived precision and recall metrics for assessing an inferred hypothesis model: _cumulative-length_ and _single-length_. Cumulative-Length. This type of assessment considers all the traces of length _up to a provided value_ $n$ (i.e., the evaluation set $E$ is $\Sigma^{\leq n}$). We compute the cardinalities $C_{TP}=\sum_{i=0}^{n}\mathit{tp}_{i};\quad C_{FP}=\sum_{i=0}^{n}\mathit{fp}_{i};\quad C_{FN}=\sum_{i=0}^{n}\mathit{fn}_{i}$ using the sequences of cardinalities of the languages accepted by the automata in Equation 5, and then compute precision and recall: (6) $\textit{precision}_{\leq n}=\frac{C_{TP}}{C_{TP}+C_{FP}};\quad\textit{recall}_{\leq n}=\frac{C_{TP}}{C_{TP}+C_{FN}}$ Our analysis does not mandate for a specific value of the maximum trace length $n$. If any domain knowledge for the application under analysis is available and suggests the use of a particular value, this should be used. In the absence of this information, the user should consider that the choice of the maximum trace length may affect the assessment result — as we will investigate experimentally in Section 6.4. Resolving to using very large values of $n$ so to approximate the _asymptotic_ values of $\textit{precision}_{\leq n}$ and $\textit{recall}_{\leq n}$ (i.e., the limit of such measures for $n\to\infty$) is appealing, but should be done with caution, since models with the same asymptotic values can exhibit different accuracy for shorter trace lengths. Nonetheless, the asymptotic values of precision and recall carry useful information. If the precision eventually converges to zero, the hypothesis language contains more false positives than true positives (in order of magnitude); if it converges to one, there are more true positives than false positives, while a value in between these extremes is achievable if the numbers of true and false positive have the same order of magnitude (including when they are both finite). Analogous considerations can be formulated for the asymptotic convergence of the recall measure in Equation (6) by comparing the orders of magnitude of the true positives and false negatives languages. We will further investigate experimentally the behavior of our cardinality-based accuracy metrics for large values of $n$ in Section 6.4444The analytical computation of such asymptotic values is usually non-trivial and will not be considered in this work. The interested reader may refer to “Part B: Complex Asymptotics” of (Flajolet and Sedgewick, 2009) for an extensive treatment of the subject.. Single-length Assessment. A second pair of cardinality-based assessment metrics can be obtained by computing precision and recall on the sublanguages of $\mathcal{A}_{TP}$, $\mathcal{A}_{FP}$, and $\mathcal{A}_{FN}$ restricted to traces of exactly length $n$, i.e., the traces at the intersection of $\Sigma^{n}$ and the corresponding automaton. Given a trace length, computing precision and recall is straightforward, using directly values from the sequences of cardinalities previously defined: (7) $\textit{precision}_{=n}=\frac{\mathit{tp}_{n}}{\mathit{tp}_{n}+\mathit{fp}_{n}};\quad\textit{recall}_{=n}=\frac{\mathit{tp}_{n}}{\mathit{tp}_{n}+\mathit{fn}_{n}}$ This assessment type can be used to build a more comprehensive picture of the model accuracy by repeating the assessment for every trace length within a range specified by the user. The result has multiple desirable features. First, is not sensitive to the parameter selection: the choice of the range of trace lengths over which the single-length assessment is repeated does not affect the result, but just defines the scope of the assessment. Second, it makes clear how the model accuracy changes across the trace lengths considered — a characteristic that current popular assessment methods do not highlight despite the fact that, in general, models _do_ have variable accuracy depending on the trace length, as it will be observed in our experimental evaluation (Section 6). Third, it allows computing derived accuracy metrics that consider different trace lengths with different weight. This may be used, for example, to weight precision and recall on the frequency of trace lengths observed in a specific deployment of the system. Although our method does not mandate for a specific range, it should be wide enough to cover all the possible behaviors of the models under analysis, for example ensuring that the upper bound of the range is at least equal to the number of states of the largest automaton under analysis. ## 5\. Fast Computation of OGFs for Model Assessment In our model assessment method we compute the values of precision and recall using the counts of true positives, false positives and false negatives, up to a finite maximum trace length. To obtain these counts without explicitly enumerating all the possible traces, we use analytic combinatorics (see Section 2.3), to count the number of different accepting paths on the automata accepting these languages. To do so, the first step is to obtain the _ordinary generating function_ (OGF) for the cardinality sequence of the language under analysis. This is generally done using the _transfer matrix method_ (Flajolet and Sedgewick, 2009; Stanley, 2011), which relies on linear algebraic operations on the matrix representing the transition relation of the automaton (Flajolet and Sedgewick, 2009), resulting in its worst-case complexity being cubic in the state count. This section describes an alternative state elimination algorithm that, despite having the same worst case complexity, in practice in our preliminary evaluation performed significantly better than the transfer matrix method, by exploiting the sparsity of the transition matrix. Our method is analogous to the state elimination algorithm by Brzozowski and McCluskey (Brzozowski and McCluskey, 1963) to generate a regular expression given a finite state automaton. While in the Brzozowski and McCluskey’s algorithm each transition is labeled with a regular expression indicating the language that causes traversing the transition, in our method it is labeled with the OGF of the cardinality sequence of that same language. The reduction rules used when a state is eliminated then allow us to progressively build the OGF through operations between rational functions. Our method is presented in algorithm 1 Data: DFA $\mathcal{A}=(\Sigma,Q,q_{i},F,\delta)$ Result: The OGF of the sequence of cardinalities of the language accepted by $\mathcal{A}$ 1 $(\mathcal{G}=\langle N,E\rangle,\mathit{initial},\mathit{final})\leftarrow\texttt{digraphConstruction}(\mathcal{A})$ 2 while _$\texttt{hasNonInitialOrFinalNode}(\mathcal{G},\mathit{initial},\mathit{final})$_ do 3 $n\leftarrow\texttt{chooseNonInitialOrFinalNode}(\mathcal{G},\mathit{initial},\mathit{final})$ 4 $\texttt{eliminateNode}(\mathcal{G},n)$ 5 6 end while return $E(\mathit{initial},\mathit{final}\,)$ Algorithm 1 Fast computation of the OGF of the sequence of cardinalities of the language accepted by a DFA. and works as follows. Given an input DFA, we construct (line 1) a directed graph $\mathcal{G}$, using digraphConstruction (algorithm 2): Data: $\mathcal{A}=(\Sigma,Q,q_{i},F,\delta)$ deterministic finite-state automaton over the alphabet $\Sigma$. Result: $\mathcal{G}=\langle N,E\rangle$ digraph with edges labeled with rational functions, where $N$ is the set of nodes, and $E:N\times N\rightarrow RF$ are the edges labeled with rational functions (0 if no edge is present). 1 $N\leftarrow Q\cup\\{\mathit{initial},\mathit{final}\\}$ 2 foreach _$(a,b)\in N\times N$_ do 3 $E(a,b)\leftarrow 0$ 4 5 end foreach 6foreach _$q\in Q$_ do 7 foreach _$s\in\Sigma$_ do 8 $t\leftarrow\delta(q,s)$ 9 $E(q,t)\leftarrow E(q,t)+z$ 10 11 end foreach 12 13 end foreach 14$E(\mathit{initial},q_{i})\leftarrow 1$ 15 foreach _$q\in F$_ do 16 $E(q,\mathit{final})\leftarrow 1$ 17 18 end foreach 19$\mathcal{G}=\langle N,E\rangle$ return $(\mathcal{G},\mathit{initial},\mathit{final}\,)$ Algorithm 2 digraphConstruction. Generates the digraph corresponding to the given DFA. each state of the DFA corresponds to one node of $\mathcal{G}$ (distinct states correspond to distinct nodes), and the edge between any ordered pair of nodes of $\mathcal{G}$ is labeled with the generating function of the sequence of cardinalities of the language of words of length one, containing the symbols causing the transition between the two corresponding states of the DFA. If the cardinality of this language is $n$ (i.e., there are $n$ symbols causing the transition from the source state to the destination state), the cardinality sequence is $\\{0,n,0,0,0,\dots\\}$, thus the generating function on the corresponding edge of $\mathcal{G}$ is $\mathfrak{G}(z)=nz$. In addition, we add to $\mathcal{G}$ a node called initial, an edge from initial to the node corresponding to the initial state of the DFA, a node called final, and an edge from each node corresponding to an accepting state of the DFA to final. All these additional edges are labeled with the generating function $\mathfrak{G}(z)=1$, which is the generating function of the cardinality sequence $\\{1,0,0,\dots\\}$ (i.e., of the language containing only the empty string). Adding these edges does not change the final generating function, but it is a simple way of dealing with multiple final states and transitions back to the initial state. Figure 4(a) (a) Initial DFA (b) Corresponding digraph $\mathcal{G}$ Figure 4. Example of digraph construction shows the example DFA taken from (Aydin et al., 2015) (which computes the OGF using the transfer matrix method), whereas Figure 4(b) depicts the graph $\mathcal{G}$ constructed with these rules described above. Then, at lines 1–1 of algorithm 1, we eliminate, one by one and using eliminateNode (Algorithm 3), all the nodes of $\mathcal{G}$ except initial and final, as follows. Data: $\mathcal{G}=\langle N,E\rangle$ labeled directed graph, where $N$ is the set of nodes, and $E:N\times N\rightarrow RF$ are the edges labeled with rational functions (0 if no edge is present); $n\in N$ the node to be eliminated. Result: The node $n$ is removed from $\mathcal{G}$ and the OGFs on the remaining edges are updated. 1 2$\mathfrak{G}_{\mathit{loop}}\leftarrow\frac{1}{1-E(n,n)}$ $E(n,n)\leftarrow 0$ /* remove self loop if present */ 3 $P\leftarrow\\{p\in N:E(p,n)\neq 0\\}$ /* predecessors of $n$ */ $S\leftarrow\\{s\in N:E(n,s)\neq 0\\}$ /* successors of $n$ */ 4 5foreach _$p\in P$_ do 6 foreach _$s\in S$_ do 7 $E(p,s)\leftarrow E(p,s)+E(p,n)\,E(n,s)\,\mathfrak{G}_{\mathit{loop}}$ 8 9 end foreach 10 $E(p,n)\leftarrow 0$ 11 12 end foreach 13foreach _$s\in S$_ do 14 $E(n,s)\leftarrow 0$ 15 16 end foreach $N\leftarrow N\setminus n$ Algorithm 3 eliminateNode. Eliminates a node while maintaining the correct OGF on the remaining digraph edges. Let $n$ be the node to be eliminated; algorithm eliminateNode works by replacing every pair composed by one edge entering $n$ from a predecessor node $p$ and one edge exiting $n$ to a successor node $s$, with a single edge from $p$ to $s$, keeping into account the loop edge from $n$ to $n$ if present. The procedure computes the OGF on the new edges using the fact that union, concatenation, and Kleene star of regular languages translate to algebraic operations between the OGF of the cardinality sequences of the operand languages. Specifically: if $A$ and $B$ are regular languages, their concatenation $C=A\cdot B$ is a regular language with $\mathfrak{G}_{C}(z)=\mathfrak{G}_{A}(z)\mathfrak{G}_{B}(z)$; if $A$ and $B$ are regular languages and $A\cap B=\emptyset$ (which in our case is a consequence of the automaton determinism), then their union $C=A\cup B$ is a regular language with $\mathfrak{G}_{C}(z)=\mathfrak{G}_{A}(z)+\mathfrak{G}_{B}(z)$; if $A$ is a regular language with $\mathfrak{G}_{A}(z)=A(z)$ then $A^{*}$ (Kleene star) is a regular language with $\mathfrak{G}_{A^{*}}(z)=\frac{1}{1-A(z)}$. As a result, Algorithm 3 computes the OGF of the Kleene star of the language that would be on the loop of the node to be eliminated (line 3), eliminates the loop (line 2) — otherwise the set of predecessors and successors subsequently defined would contain $n$ — and then iterates through the edges entering $n$ and exiting $n$, updating the OGF on the edge from the predecessor node to the successor node as illustrated in Figure 5. (a) Before eliminating $q_{2}$ (b) After eliminating $q_{2}$ Figure 5. Node elimination When initial and final are the only nodes left in $\mathcal{G}$, the OGF on the edge from initial to final is the OGF of the cardinality sequence of the language accepted by the DFA. Figure 6 shows one possible sequence of state eliminations to obtain the generating functions of the DFA shown in Figure 4(a). (a) After eliminating $q_{3}$ (b) After eliminating $q_{2}$ (c) After eliminating $q_{1}$ Figure 6. Example sequence of state eliminations for the DFA shown in Figure 4(a) We remark the order with which the nodes are eliminated is not relevant for the correctness of the result. However, it does significantly affect the performance due to the cost of the operations between rational functions required to compute the generating functions on the edges involved in each elimination. A simple heuristic that proved to be beneficial in our experiments is choosing first the node with the lowest OGF degree555The OGF degree is defined as the maximum of the degrees of its constituent numerator and denominator polynomials. on the self loop (thus the nodes with no self loop are eliminated first). ## 6\. Experimental Evaluation In this section, we report on the experimental evaluation of our model assessment method in terms of the following aspects. First, since the cumulative-length assessment requires, as an input parameter, a maximum trace length, we investigate the impact of changing its value on the model assessment results. We also analyze whether the asymptotic values (i.e., the results obtained when the maximum trace length go to infinity) are useful to characterize the model accuracy. Second, to better understand to what extent the model assessment results differ depending on the methods used, we compare the results obtained using our method with the results obtained using existing methods, specifically the statistical estimation and MBT-based methods. Third, in Section 4, we claimed that a statistical estimation in which traces having the same length have the same probability of being sampled would converge to the same results obtained by our single-length assessment method. We will verify this experimentally by comparing the single-length assessment and a model-independent sampling of $\Sigma^{*}$ in terms of the precision and recall values. Finally, we evaluate the applicability of our method on models inferred using well known model inference approaches. To summarize, we address the following research questions: 1. RQ1: How does the choice of the maximum trace length affect the model assessment results? 2. RQ2: How do the values of precision and recall obtained using our method compare with the results obtained using other evaluation methods? 3. RQ3: Is the single-length assessment over a range of lengths a usable alternative to the statistical evaluation of model accuracy using model-independent sampling of $\Sigma^{*}$? 4. RQ4: Is our method applicable to the assessment of inferred models representing aspects of actual software systems? ### 6.1. Evaluation Subjects To evaluate model assessment methods, we need various pairs of reference and inferred models. We describe how we select reference models and generate inferred models from them below. #### Reference Models We selected 41 publicly available reference models, taken from existing studies (Pradel et al., 2010; Krka et al., 2014); all of them have a well- documented origin and were previously used in the model inference literature. Pradel et al. (2010) selected 32 commonly used classes from the Java SDK API and identified their method ordering constraints using the API documentation and well-known reference books. These constraints were then translated into reference models representing all possible valid traces of method calls. The resulting reference models are publicly available on the authors’ website666http://mp.binaervarianz.de/icsm2010/index.html. These models were used also in previous work on model inference (Busany et al., 2019). Krka et al. (2014) selected 9 open-source libraries, found the corresponding reference models (manually specified in previous work), and checked them manually for inconsistencies. These models were checked against execution traces collected from actual executions of software using those libraries, and the transitions on methods that were never invoked in the collected traces were eliminated. The resulting reference models are publicly available on the authors’ website777https://softarch.usc.edu/wiki/doku.php?id=inference:start. These models have been used also in previous work (Walkinshaw et al., 2016; Le et al., 2015) #### Inferred Models Ideally, the inferred models should be generated using model inference engines on traces produced by executions of actual software systems. Unfortunately, we were not able to find and execute the exact versions of the software systems represented by the 41 reference models, to collect their execution traces. As an alternative, we used the same type of random walk described in Section 3.1 to generate a set of random traces for each reference model, and then we processed each set of traces using different model inference engines. As for the random walk parameters, we used the termination probability $p_{a}=0.1$ and the uniform probability for selecting an outgoing transition among available transitions following Walkinshaw et al. (2013). For each reference model, we repeated the random walk until at least 100 traces were generated _and_ each state of the model had been visited at least four times, as suggested by Busany et al. (2019). From each set of traces we inferred two models, using two different model inference algorithms: K-tail (Chow, 1978) (the most well-known model inference algorithm) and MINT (Walkinshaw et al., 2016) (a state-of-the-art model inference technique). It is worth noting that, in the area of model inference, it is common practice to use randomly generated traces when traces coming from actual software executions are not available (e.g., (Busany et al., 2019; Lo et al., 2012; Walkinshaw and Bogdanov, 2013; Walkinshaw et al., 2013)). Moreover, model assessment is independent of how inferred models are generated, thus is not a major issue in our evaluation. Nevertheless, we will discuss this as a potential threat to validity in Section 6.8. To summarize, our experimental evaluation is based on 82 test subjects (pairs of reference and inferred models): 41 reference models, with two inferred models each. Table 1 shows the size of the reference and inferred models in terms of the minimum, the average, and the maximum number of states and transitions. Table 1. Characteristics of reference and inferred models | State count | | Transition count ---|---|---|--- | Min | Avg | Max | | Min | Avg | Max Reference | 2 | 9.2 | 41 | | 7 | 77.8 | 465 Inferred | 2 | 349 | 2060 | | 5 | 752 | 4994 ### 6.2. Evaluation Settings We performed the assessment of all the 82 test subjects using our implementation of trace similarity, MBT-based assessment, and our method with trace counting, all developed in Java and publicly available (see Section 6.3). All the experiments were executed on an AMD EPYC 7401P (24 cores, 48 threads) with 448 GB of RAM. Since the implementation of our evaluation method is single-thread, 24 evaluations were executed concurrently. The evaluation runtime of every model assessment performed in our evaluation was subject to a timeout of two days. It is worth pointing out that all the single-length and cumulative-length assessments on the same test subject require the same evaluation runtime. The reason is that this evaluation runtime is dominated by the time required to compute the generating functions, which are the same for all assessments of the same test subject. ### 6.3. Data availability To support open science and enhance the reproducibility of our evaluation, we provide a replication package for the reviewers, including all the artifacts: the source code, the reference models and the inferred models used. Upon acceptance, the replication package will be released to the public. ### 6.4. RQ1: Input Parameter Sensitivity ##### Methodology As described in Section 4, our method can be applied in two ways: _cumulative- length_ and _single-length_ assessments. The single-length assessment repeated over a range of trace lengths is not sensitive to parameter changes: the _range_ changes just the _scope_ of the assessment, i.e., for which lengths the values of precision and recall are computed, without affecting the output values themselves. To evaluate the parameter sensitivity of the cumulative- length assessment, we computed the precision and recall values varying the only parameter (i.e., the maximum trace length) from 0 to 200 in steps of 1. ##### Results All the assessments of 61 of the 82 test subjects terminated within the 2-day timeout. Figure 7 shows how the precision and recall values vary depending on the maximum trace length parameter; each line is a result for one test subject. (a) (b) Figure 7. Values of $\textit{precision}_{\leq n}$ and $\textit{recall}_{\leq n}$ measured using the cumulative-length assessment, varying the maximum trace length parameter $n$ in the range [0, 200]. One plot per test subject. Five test subjects are highlighted, and their characteristics are summarized in Table 2. The foremost takeaway of the evaluation results is that the choice of the maximum trace length parameter does indeed affect the computed accuracy values, and it does so differently, depending on the inferred model. The results highlighted the prevalence of cases in which precision and/or recall values cover a large part of the range $[0,1]$, depending on the parameter value. Specifically, among the 61 test subjects that finished within the timeout, 20 have a precision value (and 54 have a recall value) that covers the entire range $[0,1]$, depending on the maximum trace length parameter. This emphasizes how — for any choice of a single value of the maximum trace length — some relevant information about the model accuracy is inevitably lost, leading to possibly misleading results. Figure 7 also shows the variety of trends that the precision and recall values can have, as the maximum trace length is increased: e.g., convergence to 0 with and without a peak, constant equal to 1, and convergence to $0<k<1$. In the figure we have highlighted (with colors and special markers) five test subjects that, together, cover the most common trends we observed in the results of this experiment. Appendix A gives further details on how we selected these test subjects, and Appendix B describes the selected test subjects in more detail. Table 2 summarizes the characteristics of these representative test subjects. Table 2. Characteristics of the selected test subjects | Reference model | Inference approach | Inferred model ---|---|---|--- | Name | States | Trans. | States | Trans. Subject 1 (S1) | java.net.URL (Pradel et al., 2010) | 5 | 57 | MINT | 718 | 993 Subject 2 (S2) | SMTPProtocol (Krka et al., 2014) | 3 | 14 | MINT | 24 | 110 Subject 3 (S3) | StringTokenizer (Pradel et al., 2010) | 4 | 14 | 2-Tails | 83 | 175 Subject 4 (S4) | Signature (Krka et al., 2014) | 3 | 7 | MINT | 15 | 63 Subject 5 (S5) | StringTokenizer (Krka et al., 2014) | 4 | 7 | 2-Tails | 30 | 61 Furthermore, Figure 7 shows that the asymptotic value888 We determined the asymptotic value by manually checking to which value the accuracy stabilizes. There were some cases (seven for the precision value, four for recall value) in which we were not able to determine the asymptotic value because, within the observed range, it was not clear around which value it stabilized. can be reached through different paths, with fast or slow convergence, and with or without local minima or maxima. This confirms that judging or comparing model accuracies according to the asymptotic value of precision and recall can be misleading, since models with the same asymptotic accuracy generally can have different accuracies for shorter traces. Although investigating the asymptotic accuracy of models obtained with current inference methods is outside the scope of this paper, an interesting observation is that the asymptotic values of the accuracy metrics are generally either zero or one. This is because the precision and recall values are quotients between language cardinalities that tend to infinity as the maximum trace length considered is increased, therefore the quotient often goes to zero or one depending on the relative growth rate of the languages in Equation 1. In our evaluation, the most common asymptotic value is zero: 45 subjects have a precision value converging to zero, while 55 have a recall value converging to zero. The number of subjects having accuracy converging to a value greater than zero and smaller than one is only three for precision and two for recall. This is an observation that is outside the scope of this paper, and may not generalize to models obtained with different inference methods; it will be further investigated in future work. #### Answer to RQ1 The maximum trace length parameter does affect the computed cumulative-length precision and recall values, and it does so differently, depending on the inferred model. This parameter should be tuned, case by case, to a value that is relevant for the application under analysis. If no domain knowledge is available, it is preferable to consider how precision and recall values change considering different trace lengths, using the single-length metrics. ### 6.5. RQ2: Comparison of Model Assessment Methods This section discusses how the results generated by the assessment methods described and discussed in the paper compare with each other. In particular, we will examine trace similarity, MBT-based assessment, cumulative-length assessment, and single-length assessment. Note that the model assessment methods discussed in this paper can be classified into two types depending on their output types: trace similarity, MBT-based assessment, and the cumulative-length assessment are all methods that return one pair of precision and recall values, while the single-length assessment instead returns one pair of precision and recall values _per trace length_ in the given range (which is a parameter of the method). These two types cannot be directly compared. For this reason the comparison proposed in this section is divided in two parts. In the first part, we will compare trace similarity, MBT-based, and cumulative-length, with the goal of understanding the consistency and the differences between the computed precision and recall values across the different assessment methods and different parameter values. In the second part, we will discuss the single-length assessment results in relation with the results generated with the other currently used methods. In this case a direct comparison is neither possible nor meaningful, due to the different output type. To make them comparable, we will _condition the other assessment methods on the trace length_ : the evaluation sets generated by these methods will be partitioned according to the trace length, and the subsets will be used to compute one pair of precision and recall values per trace length, which can then be directly compared with the results from the single-length assessment. #### Comparison of Trace Similarity, MBT-based, and Cumulative-Length Assessments As discussed above, we first compare trace similarity, MBT-based, and cumulative-length. ##### Methodology To understand the consistency and the differences between the computed precision and recall values across the different assessment methods, we run the methods on the 81 test subjects, with the following parameters settings. Trace similarity requires defining how to make the nondeterministic choices needed to conduct the random walk (i.e., whether to terminate the walk, and which available transition should be traversed — see Section 3.1). This has a direct effect on the sampling bias, thus affecting the assessment results. In our setup, the selection of the outgoing transition to be traversed was done uniformly among the available alternatives, as it is often done in practice (e.g., (Walkinshaw et al., 2013; Lo and Khoo, 2006a)). Regarding the termination of the trace when a final state is reached, we found different approaches in the model inference literature. Sometimes the termination probability is a function of the outdegree of the accepting state (e.g., as in reference (Walkinshaw et al., 2013)), sometimes is not (e.g., as in reference (Lo et al., 2012)). To highlight how this choice may affect the assessment result, we used a fixed termination probability $p_{a}$, and we executed the trace similarity assessment setting $p_{a}$ to three values (0.02, 0.1 and 0.5), thus obtaining three pairs of precision and recall values for each test subject. Trace similarity also requires specifying the target size for the evaluation set $E$, which in our experiments was set to 100,000 traces. To further ensure adequate model coverage, we continued adding traces to $E$ beyond the target size, until each transition of the model was followed at least 10 times (as proposed in reference (Lo and Khoo, 2006a)), or a 30-minute time limit was reached. As a result, in some cases we generated more than 100,000 traces. The MBT-based assessment requires a model-based testing method for finite state automata, to generate the evaluation set $E$. We use the W-method, as done by the original proponents of the method. The cumulative-length assessment requires a parameter indicating the maximum length of the traces to be considered. Since there is no “correct” value to choose in the absence of domain knowledge for a specific test subject, we evaluate the results when varying the parameter from 0 to 200 in steps of 1. ##### Results All the model assessments performed using trace similarity returned a result within the timeout. The cumulative-length assessments terminated within the timeout for 61 of the 82 test subjects (for all parameter values), while for the remaining 21 subjects a timeout occurred during the computation of the generating function. The MBT-based assessment terminated within the timeout only in two of the 82 cases under analysis. The reason is that the cardinality of the evaluation set generated with the W-method grows with $|\Sigma^{d+1}|$, where $d$ is the difference in the number of states between the inferred and reference model, therefore the method is usable only when the reference and inferred models have a similar number of states. This is rarely the case in our set of test subjects: the difference $d$ is on average 247 ($\sigma=265$). Due to the lack of a sufficient number of results from the MBT-based assessment, we omit this method in the rest of the comparison. Before we compare the trace similarity and cumulative-length assessment results, it is worth focusing on how trace similarity is affected by the termination probability $p_{a}$ of the random walk. Table 3 shows the effect of $p_{a}$ on the average length of the traces generated in our experimental evaluation. These are consistent with an exponential distribution, however we remind that the relationship between termination probability and distribution of trace lengths is model dependent, and our experiments lead to this distribution only because most of the non-error states of the models under analysis are accepting states. Table 3. Mean trace length generated by the random walk, depending on the $p_{a}$ parameter. $p_{a}$ | Mean trace length ---|--- 0.5 | 1.99 0.1 | 11.28 0.02 | 49.45 Table 4 summarizes the absolute difference between the model accuracy evaluated using trace similarity with different random walk termination probability $p_{a}$. For example the mean absolute difference between the precision value computed using trace similarity with a random walk with $p_{a}=0.5$, and the precision value computed using trace similarity with a random walk with $p_{a}=0.02$, is $36\text{\,}\mathrm{\textup{pp}}$ ($\sigma=$29\text{\,}\mathrm{\textup{pp}}$$), with $\mathrm{\textup{pp}}$ indicating percentage points. This difference indicates that the choice of the random walk parameters values can affect the trace similarity result in a way that does not allow us to reliably determine the model accuracy. Note that a larger difference in the $p_{a}$ parameter leads to a larger difference in the measured accuracy (e.g., line 1 vs line 3 of Table 4). This is explained intuitively in terms of the sampling bias discussed in Section 3.1: a random walk with higher termination probability $p_{a}$ will generally produce shorter traces, and we know, from the RQ1 results in Section 6.4, that in most cases the model accuracy is higher on shorter traces. Table 4. Mean and standard deviation of the absolute difference between the model accuracy evaluated using trace similarity with different random walk termination probability $p_{a}$. Values in percentage points ($\mathrm{\textup{pp}}$). | Mean | $\sigma$ ---|---|--- $|\textit{Precision}_{\textit{TS}_{0.5}}-\textit{Precision}_{\textit{TS}_{0.1}}|$ | $19\text{\,}\mathrm{\textup{pp}}$ | $14\text{\,}\mathrm{\textup{pp}}$ $|\textit{Precision}_{\textit{TS}_{0.1}}-\textit{Precision}_{\textit{TS}_{0.02}}|$ | $19\text{\,}\mathrm{\textup{pp}}$ | $17\text{\,}\mathrm{\textup{pp}}$ $|\textit{Precision}_{\textit{TS}_{0.5}}-\textit{Precision}_{\textit{TS}_{0.02}}|$ | $36\text{\,}\mathrm{\textup{pp}}$ | $29\text{\,}\mathrm{\textup{pp}}$ $|\textit{Recall}_{\textit{TS}_{0.5}}-\textit{Recall}_{\textit{TS}_{0.1}}|$ | $37\text{\,}\mathrm{\textup{pp}}$ | $20\text{\,}\mathrm{\textup{pp}}$ $|\textit{Recall}_{\textit{TS}_{0.1}}-\textit{Recall}_{\textit{TS}_{0.02}}|$ | $15\text{\,}\mathrm{\textup{pp}}$ | $10\text{\,}\mathrm{\textup{pp}}$ $|\textit{Recall}_{\textit{TS}_{0.5}}-\textit{Recall}_{\textit{TS}_{0.02}}|$ | $52\text{\,}\mathrm{\textup{pp}}$ | $27\text{\,}\mathrm{\textup{pp}}$ Figure 8. Comparison of trace similarity with three different random walks ($\textit{TS}_{\textit{0.5}}$, $\textit{TS}_{\textit{0.1}}$ and $\textit{TS}_{\textit{0.02}}$), and the cumulative evaluation proposed in Section 4 (black plot) over five representative test subjects. Figure 8 shows the comparison of the precision and recall values computed using trace similarity and our cumulative-length assessment, for the five representative test subjects selected in RQ1. To improve the readability, the range of the trace length parameter for the cumulative-length assessment shown in the figure is from 0 to 100. The same information is available for all test subjects in the supporting material (see Section 6.3). Note the three different trace similarity assessments per test subject (using different values of the $p_{a}$ parameter of the random walk). The most important observation is that both trace similarity and cumulative- length results values are different in most cases, and vary over a large range of values depending on the methods’ parameters. A closer look reveals a consistency between the trend of the accuracy results computed using trace similarity as $p_{a}$ is decreased, and the trend of the accuracy results computed by the cumulative-length assessment for longer traces: in all subjects for which the cumulative-length assessment shows decreasing model accuracy for longer traces, also the trace similarity assessment shows decreasing accuracy as $p_{a}$ is decreased from 0.5 to 0.02. Analogously, the cumulative-length assessment for subject S1 shows an (initially) increasing precision, and the corresponding trace similarity result is also consistent with this trend. The reason behind this consistency is the effect of the random walk termination probability $p_{a}$ on the length of the traces generated by the random walk, as discussed before. However, it is important to stress that the distribution of trace lengths generated by the random walk is not enough to fully capture how the sampling bias affects the trace similarity result, because also different traces of the same length may have different probabilities of being generated. The nondeterministic choices required to conduct the random walk can be performed in a variety of ways, and their effect on the result is model-dependent. On the other hand, the cumulative-length assessment has only the maximum trace length parameter, whose effect can be intuitively understood, making it easier to tune. Section 6.5 will include a scenario in which trace similarity and cumulative-length assessment diverge due to this effect. Furthermore, in Figure 8 we can observe that for some test subjects all the model assessment methods compute the same value (100%) for precision. We manually verified that in these cases the inferred language is a subset of the reference language, hence the evaluation set cannot contain false positive traces, regardless how it is generated, and thus the computed precision value is always 100%. #### Comparison of Statistical Estimation and Single-Length Assessment We now turn our attention to the results of the single-length assessment over a range of trace lengths, and how it compares with the results of trace similarity. ##### Methodology As discussed above, a direct comparison is not possible, since the single- length assessment returns one pair of precision and recall values _per trace length_ , while trace similarity returns one pair of precision and recall values overall. To obtain comparable results, we will _condition the trace similarity on the trace length_ , by partitioning each evaluation set $E$ according to the trace length. This will allow us to use the resulting subsets to compute one pair of precision and recall values per trace length, and it will also allow us to look at the distribution of the traces in $E$ across different trace lengths. In fact, we will initially focus on this distribution, which is the first effect of the sampling bias induced by the random walk mentioned in Section 3.1. Then, we will compare the values of precision and recall for different trace lengths (from 1 to 100 in steps of 1). Since our single-length assessment has no sampling bias, any difference in precision or recall value for a specific trace length implies statistical noise and/or the non- uniformity of the sampling induced by the random walk (i.e., the second effect of the sampling bias as mentioned in Section 3.1). ##### Results The results for the five representative test subjects previously used in RQ1 are shown in Figure 9. In each plot, the red line and the black line with markers represent how the accuracy metric values (left y-axis) computed by trace similarity conditioned on the trace length and the single-length assessment, respectively, vary depending on the trace length (x-axis). In addition, the solid area filled in blue indicates the number of samples (right y-axis) in the evaluation set $E$ for each trace length (x-axis). Note that, in some of the plots, the red line and the blue area do not exist after a certain trace length, meaning that the random walk generated an empty $E$ and no precision and recall values were computed by trace similarity. (a) Trace similarity using a random walk with $p_{a}=0.5$ (b) Trace similarity using a random walk with $p_{a}=0.1$ (c) Trace similarity using a random walk with $p_{a}=0.02$ Figure 9. Comparison of the single-length assessment and trace similarity conditioned on the trace length, with three different random walk parameters Based on the results shown in Figure 9, we can make the following observations. * • The distribution of samples highlights how shorter traces are more likely to be sampled, and how the distribution of samples depends both on the random walk parameters and on the model upon which the random walk is performed. For example with $p_{a}=0.5$ (Figure 9(a)) the distribution of samples for subject S1 contains longer traces, compared to the other test subjects. Conversely, with $p_{a}=0.02$ (Figure 9(c)) the distribution of samples for the precision assessment of subject S1 contains shorter traces, compared to the other test subjects. * • The random walk can introduce a sampling bias between traces having the same length. This is particularly visible in Figure 9(c), in the recall plots of subject S5, which show a difference between trace similarity conditioned on the length (red line, which converges to 0.8) and the single-length evaluation (black line, which converges to 0.3). This observation explains why in Figure 8, in the case of subject S5, the recall values obtained using standard trace similarity consistently exceed the recall values obtained using our cumulative-length assessment, despite the fact that the inferred model consistently shows a 0.3 recall value for traces having a length greater than 20. Note that in the other test subjects, the plots of trace similarity conditioned on the trace length, and those of the single-length assessment show the same trend, highlighting how this effect is model-dependent. #### Answer to RQ2 The sampling bias affects the trace similarity result in a way that is hard to predict a priori, because it depends on both the topology of the model and the parameters of the random walk. When trace similarity is compared to the cumulative-length assessment, the conclusion is that the choice of the parameter values of both assessment methods affect the result to an extent that does not allow us to reliably determine the model accuracy, thus preventing a meaningful comparison among different models in terms of accuracy, or among the accuracy values of the same model measured with different methods. On the other hand, the single-length assessment is not affected by sampling bias because it measures the model accuracy for each trace length in the given range, considering (for each length) all the possible traces. Leveraging this characteristic, we compared the single-length assessment with trace similarity conditioned on the trace length, analyzing the effect of the non-uniform sampling among traces of the same length: this effect is model-dependent. We also examined the distribution of trace length generated by the various random walk in trace similarity, noticing how shorter traces are more likely to be sampled, and how this effect depends on the model topology and the parameters of the random walk. ### 6.6. RQ3: Single-length assessment and model-independent sampling of $\Sigma^{*}$ The goal of RQ3 is to compare, in terms of precision and recall, the single- length assessment we proposed and a model-independent sampling of $\Sigma^{*}$ inspired by the observation made in Sections 3.1.1 and 4. It is possible to statistically assess precision and recall using a sampling method that generates traces by combining symbols randomly chosen from the alphabet, rather than through random walks on models. In practice, however, using this method is feasible only for short trace lengths, because to evaluate precision (respectively, recall) it is necessary to generate traces that are accepted by the inferred (respectively, reference) model. This generation process becomes infeasible with real-world models and long traces, due to the gap between the exponential growth of $|\Sigma^{n}|$ and the slower growth of the size of the accepted language of length $n$, as the length $n$ is increased. Nonetheless, this approach has the advantage of controlling the sampling bias (which becomes model independent), enabling sampling with uniform distribution over a finite subset of $\Sigma^{*}$. Answering this research question will let us check experimentally whether a statistical estimation of the model accuracy (in which all the traces having the same length have the same probability of being sampled) generates the same results as our single-length assessment. If the results show consensus between the two assessment methods, they will confirm that our method is a usable alternative to the statistical evaluation of model accuracy using model- independent sampling of $\Sigma^{*}$. ##### Methodology For each test subject, we computed precision and recall for each trace length starting at 0 and increasing it until a one-hour timeout was reached. This was done using Algorithm 4, which takes as input the reference model $\mathcal{R}$, the inferred model $\mathcal{H}$, the accuracy metric $m$ to be computed (precision or recall) and the trace length $n$ for which to compute it. The algorithm generates traces without considering the reference or the inferred model, by concatenating $n$ symbols randomly selected from the alphabet with uniform distribution. Each generated trace is then tested against the models: if $m$ is _precision_ (respectively, _recall_), the trace is tested against $\mathcal{H}$ (respectively, $\mathcal{R}$), to determine whether it is part of the inferred (respectively, reference) language. If this is the case, the trace is further tested against $\mathcal{R}$ (respectively, $\mathcal{H}$) to determine whether it is a true positive. This is repeated until 1000 “useful” traces are generated, i.e., traces that are accepted by $\mathcal{H}$ (respectively $\mathcal{R}$) and therefore contribute to computing the metric. The target number of useful traces is a compromise between the scalability and the noise of the evaluation. It was chosen empirically, and deemed appropriate because it gives a 99% chance that the real precision/recall value is within ±4.08% of the measured value. Note that the maximum trace length reached with this model assessment method is model- dependent, since it depends on the proportion of traces in $\Sigma^{*}$ accepted by the models under analysis. In fact, it may differ even between precision and recall on the same test subject, since the proportion may be different between reference and inferred model. Finally, we compared the resulting precision and recall values with the single-length assessment results. Data: Reference model $\mathcal{R}$, inferred model $\mathcal{H}$, alphabet $\Sigma$, trace length $l$, target number of samples $n$, metric $m$ (either precision or recall) Result: Accuracy value 1 $\mathit{truePositives}\leftarrow 0$ 2 $\mathit{acceptedTraces}\leftarrow 0$ 3 while _$\mathit{acceptedTraces} <n$_ do $t\leftarrow\text{randomTrace}(\Sigma,l)$ // $l$ symbols chosen from $\Sigma$ with uniform distribution 4 if _m = precision $\wedge$ $t\in\mathfrak{L}(\mathcal{H})$_ then 5 $\mathit{acceptedTraces}\leftarrow\mathit{acceptedTraces}+1$ 6 else if _m = recall $\wedge$ $t\in\mathfrak{L}(\mathcal{R})$_ then 7 $\mathit{acceptedTraces}\leftarrow\mathit{acceptedTraces}+1$ 8 end if 9 if _$t\in\mathfrak{L}(\mathcal{R})\wedge t\in\mathfrak{L}(\mathcal{H})$_ then 10 $\mathit{truePositives}\leftarrow\mathit{truePositives}+1$ 11 end if 12 13 end while return _$\frac{\mathit{truePositives}}{\mathit{acceptedTraces}}$_ Algorithm 4 Statistical assessment of the model accuracy for traces of length $l$, using a model-independent sampling of $\Sigma^{l}$ ##### Results The differences between the model accuracies obtained using the single-length assessment and the model-independent sampling of $\Sigma^{*}$ (computed using all the trace lengths for which both results are available) are summarized in Table 5. Table 5. Mean and standard deviation of the absolute difference in precision and recall evaluated using the single-length assessment and the model-independent sampling of $\Sigma^{*}$ (values in percentage points). | $\textit{precision}_{=n}$ | $\textit{recall}_{=n}$ ---|---|--- Mean | $0.44\text{\,}\mathrm{\textup{pp}}$ | $0.22\text{\,}\mathrm{\textup{pp}}$ $\sigma$ | $0.70\text{\,}\mathrm{\textup{pp}}$ | $0.55\text{\,}\mathrm{\textup{pp}}$ The small differences between the methods’ results are caused by the sampling error intrinsic in any accuracy measure performed using randomly generated samples, and are within the margin of error expected for the used number of samples. This confirms that a statistical estimation in which all the traces of the same length have the same probability of being sampled generates the same results as our single-length assessment, in line with your discussion in Section 4. To have a closer look at the results for individual test subjects, Figure 10 shows the results of the same five representative test subjects previously used in RQ1 and RQ2. The results obtained using the model-independent sampling of $\Sigma^{*}$ are represented by red markers (note the different ranges of trace lengths for which the results are available, due to the timeout), while the single-length assessment is represented by the black line. We remark the two methods exhibit a similar trend also on test subject 5, for which instead the trace similarity method returned different results due to the non-uniform probability distribution of traces having the same length, induced by the random walk. Figure 10. Partitioned comparison of sampling $\Sigma^{*}$ #### Answer to RQ3 The precision and recall values obtained using the single-length assessment we proposed are consistent with those obtained using a model-independent sampling of $\Sigma^{*}$. This implies that our method is preferable, because it is not limited to short trace lengths, and because, being deterministic, is not affected by statistical noise. ### 6.7. RQ4: Practical applicability on inferred models This section discusses the practical applicability of our approach by evaluating its scalability with respect to the complexity of the models under analysis, measuring the execution time of each assessment on each test subject. ##### Methodology Model complexity is widely studied, and can be measured in a variety of ways, with different complexity metrics appropriate in different scenarios. We will evaluate the scalability of our method with respect to two different model complexity metrics: _deterministic state complexity_ (Yu, 2001) and _star height_ (Eggan, 1963). Specifically, given a regular language $L$, these metrics are defined as follows. The _deterministic state complexity_ is defined as the number of states of the minimal DFA accepting $L$. We will refer to this measure simply as the “number of states”, since all the automata used in this work are in their minimal form. The number of states is a commonly used model complexity metric in the field of model inference (Walkinshaw et al., 2016; Lo and Khoo, 2006b). The _star height_ of $L$ is the minimum star height among all the regular expressions representing $L$, where the star height of a regular expression is the maximum nesting depth of Kleene star operators in the expression. Though the start height is less common, it is a better predictor of the execution time of our assessment method since this is dominated by the computation of the OGFs. Referring to the OGF computation algorithm described in Section 5, it can be seen that an automaton without closed cycles leads to intermediate OGFs that are easy to compute since they are polynomials (instead of rational functions) generated using only sums and products of polynomials. Conversely, all cycles present in the automaton under analysis must eventually be eliminated during the OGF computation by composing the intermediate OGFs using sums, products, and quotients of rational functions. Intuitively, languages with higher star heights are likely to have more closed cycles in the corresponding DFA representation, increasing the complexity of the OGF computation, whereas a star height of zero indicates a finite language (thus with a corresponding DFA with no loops) of which the OGF is a polynomial of finite degree that is easy to compute. We should note that the star height is by no means a complete metric; the cost of each state elimination operation in the OGF computation of Section 5 is hard to predict a priori, because it depends on the complexity of the rational functions involved. Nonetheless, developing a new ad-hoc structural complexity metric for the OGF computation is outside the scope of this work. Finding the minimum star height of a language is a notoriously hard problem (Kirsten, 2005). Therefore, in our evaluation, we use an approximated value obtained by transforming the inferred model in a regular expression using the Brzozowski and McCluskey algorithm (Brzozowski and McCluskey, 1963) and then counting the maximum nesting depth of Kleene star operators in this expression. Note that both reference and inferred models influence the execution time. However, in our experiments the size of the inferred model is one or two orders of magnitude larger than the size of the reference model: therefore we consider the effect of the latter negligible. Since the differences in execution times between different assessments (single-length and cumulative-length) of the same subject are negligible, we give one execution time _per test subject_ rather than _per assessment_. This is because the execution time is dominated by the computation of the OGFs, which is the same for all the assessments on the same test subject. ##### Results Figure 11(a) shows the number of subject assessments completed over time. On one hand, 21 test subjects assessments did not finish within the timeout, and when the timeout occurred all of them were in the OGF computation phase. This confirms that the execution time of our method is dominated by the computation of the OGFs. Among the subject assessments that did not finish within the timeout, the minimum state count was 68 and the maximum was 2061, while the minimum approximated star height was 13 and the maximum was 304. On the other hand, 41 of 82 subjects had an assessment execution time of less than one minute, showing the practical suitability of the method. Additionally, a benefit of our method is that the OGFs, once generated, can be reused to perform further analyses with different parameters with negligible overhead. Figure 11(b) shows the correlation between number of states of the inferred model, and execution time. The largest model successfully assessed had 1465 states, and was assessed in $9935\text{\,}\mathrm{s}$. The Pearson correlation coefficient between execution time and number of states was 0.07, highlighting how inadequate is this complexity metric to capture the computational complexity of the assessment. Figure 11(c) shows the correlation between star height and execution time. The most complex inferred model successfully assessed had an approximated star height of 38 (i.e., the maximum nesting depth of the Kleene star operator in a regular expression accepting the inferred language was 38), and was assessed in $151\,000\text{\,}\mathrm{s}$ — the longest execution time encountered among the test subjects that finished within the timeout. The Pearson correlation coefficient between star height and execution time was 0.51, implying that the start height can be a rough indicator of the execution time of our method. $10^{-2}$$10^{-1}$$10^{0}$$10^{1}$$10^{2}$$10^{3}$$10^{4}$$10^{5}$$0$$20$$40$$60$$80$Elapsed time (s)Completed subject assessments (a) Number of subject assessments completed over time. $10^{1}$$10^{2}$$10^{3}$$10^{-2}$$10^{0}$$10^{2}$$10^{4}$Number of statesExecution time (s) (b) Assessment time, depending on the number of states. $10^{0}$$10^{1}$$10^{-2}$$10^{0}$$10^{2}$$10^{4}$Star heightExecution time (s) (c) Assessment time, depending on the star height of the inferred model. Figure 11. Scalability of our method Although the experiments were executed on a multicore system, our current implementation is still sequential, thus the performance metrics in this section are to be considered for a single core. #### Answer to RQ4 Despite the inadequacy of widely adopted model complexity metrics, we report that our approach was able to assess inferred models representing aspects of actual software systems, using commodity hardware and despite being at a prototype stage. ### 6.8. Threats to validity There are a number of factors that could impact the validity of our experimental results. _Choice of metrics._ The choice of the accuracy metrics used in this evaluation could threaten the validity of our results if the metrics fail to capture adequately the model accuracy. To mitigate this threat we used precision and recall (Tharwat, 2020), which are well known metrics and have often been used in the area of model inference (Lo and Khoo, 2006a; Pradel et al., 2010; Walkinshaw et al., 2016; Krka et al., 2014; Polyvyanyy et al., 2020; Walkinshaw et al., 2008). Moreover, this risk is also partially mitigated by the fact that the proposed assessment method can be adapted to compute additional metrics (e.g., specificity or F-measure) without affecting the scalability. The choice of the metric used to capture the model complexity in the evaluation of our method scalability could also pose a construct validity threat. To minimize this risk, we provided results using two different complexity metrics: model state count and language star height. _Reference models used._ Using certain reference models could limit the generalizability of our results. To mitigate this issue, we considered a variety of reference models (41 in total), obtained from two different publicly available sources we found in the model inference literature (Pradel et al., 2010; Krka et al., 2014). _Training set generation and model inference setup._ Due to the lack of a suitable and widely accepted model inference benchmark suite containing traces of real software executions, we used random walks on the reference models to generate the traces fed to a model inference method. This could introduce a bias in the training set, which may become a bias in the corresponding inferred models. Similarly, also the choice of the inference method (and its parameter values) could introduce a bias in the inferred model. Nevertheless, the model assessment method proposed in this paper and considered in our evaluation is independent of the method used for generating inferred models. Furthermore, we used the same evaluation subjects (i.e., pairs of reference and inferred models) for the model assessment methods used in our comparative evaluation, yielding fair comparison results. Moreover, testing inference methods using traces generated from random walks is common practice (Busany et al., 2019; Lo et al., 2012; Walkinshaw and Bogdanov, 2013; Walkinshaw et al., 2013). _Input parameters of other assessment methods._ The comparison discussed in RQ2 is affected by the parameters of the considered assessment methods, e.g., the characteristics of the random walk used for trace similarity. To obtain results that are representative of how the assessment methods are is used in practice, our choice of parameter values was guided by what is frequently used in the literature. Nonetheless, different method parameters would in general lead to different results. _Statistical noise_. Although single-length, cumulative-length, and MBT-based assessment methods are deterministic, trace similarity involves randomness, which could affect the evaluation results. However, due to the low computational cost for generating traces through random walks and checking them against a model, we were able to perform the trace similarity assessment using at least $100\,000$ samples, ensuring low statistical noise. _Errors in the implementation._ We used a prototype implementation of our method, which may contain faults. Moreover, in order to develop a self- contained solution that does not rely on external algebra systems such as Wolfram Mathematica (Inc., 2022) or Maple (Waterloo Maple, Inc.., 2022), the algebraic system used to perform operations on rational functions was implemented by us from scratch: it may also contain faults. To mitigate the issue, we performed systematic sanity checks by comparing the language cardinalities obtained using different counting methods. ## 7\. Related Work Our method is related to the work done in the area of _model assessment_ and _trace (model) counting_ 999This area of research is generally called _model counting_. In our context, this name could be confusing, since we have already used the term _model_ to indicate a representation of an aspect of a system. To avoid any confusion, hereafter we refer to it as _trace (model) counting_.. ### 7.1. Model Assessment As discussed in Section 3.1, Lo and Khoo (2006a) proposed a method called _trace similarity_ for empirically assessing the accuracy of inferred models against reference models, both in the form of probabilistic or non- probabilistic finite-state automata. The method measures the accuracy in terms of _precision_ and _recall_ using traces randomly generated from the models. Specifically, for a reference model $R$ and an inferred model $H$, they defined _recall_ as the percentage of traces generated by $R$ that are accepted by $H$ and _precision_ as the percentage of traces generated by $H$ that are accepted by $R$. They also proposed an algorithm called TraceGen that generates a set of random traces from a model with the aim of covering every transition in the model at least $n$ times, where $n$ is a coverage parameter. Lo et al. (2012) later extended the trace similarity method to additionally assess the accuracy of inferred models in terms of _specificity_ , which is a metric indicating the ability to correctly reject illegal behaviors. One common issue of these methods, as discussed in Section 3.1, is that the accuracy of inferred models depends on the traces drawn according to a certain probability distribution imposed by the random trace generation. Our method resolves this issue by computing precision and recall values that consider all the traces, up to a user-defined arbitrarily large maximum length. As discussed in Section 3.2, the aforementioned issue was also acknowledged by Walkinshaw et al. (2008), which addressed it by adapting the W-Method (Chow, 1978) (originally developed for model-based software testing to generate a set of traces covering all distinguishable runs of the model under test) to generate a “representative” trace set that covers all the model behaviors without privileging any specific one. Although this solution guarantees to cover all behaviors, it could still return misleading results (e.g., an inferred model having fewer counterexamples could obtain a lower accuracy value) since it does not consider the number of accepted traces affected by each behavior, which is something that our method does — up to a finite maximum trace length set by the user. Furthermore, as shown in our evaluation results, the MBT-based method is not scalable due to the exponential growth of the cardinality of the representative set with the number of different states between reference and inferred models. Recently, Polyvyanyy et al. (2020) have proposed a framework that compares reference and inferred models in terms of their languages (i.e., the languages accepted by the models). While they defined precision and recall similarly to ours (i.e., Equation 6), they overlooked the usage of language cardinality, arguing that it is useful only for finite languages. Instead, they suggested framework instantiations using _topological entropy_ as a language measure which, intuitively, measures the increase in the variability of the words of the language as their length goes to infinity. It is not clear whether, using this measure, an inferred model with fewer counterexamples would always obtain a higher accuracy value, even when the counterexample languages of the two inferred models are not in a subset-of relationship. In our method, all the traces up to the maximum trace length specified by the user are considered, and they all have the same weight on the result. Moreover, the maximum trace length parameter can be set either to a value that is relevant for the application under analysis, or to a very large value, resulting in effectively computing the asymptotic accuracy. Other than the _language_ perspective, where traces (and thus languages) generated by models matter in assessing the accuracy of inferred models, model assessment can also take a _structural_ perspective, where the model structure, i.e., how states and transitions are arranged, is considered. Walkinshaw and Bogdanov (2013) proposed an algorithm called LTSDiff that compares the structure of two models, identifying missing or superfluous states and transitions, and computing precision and recall based on these. Pradel et al. (2010) noticed that structural differences can have a variable impact on the size of the language difference. To address the issue, they proposed an approach that counts the number of transitions that are in common between reference and inferred models after abstracting the models using a variant of the k-tail algorithm. This allows one to identify states that are _similar_ , even when they are not exactly equivalent, with the parameter $k$ controlling how much imprecision and incompleteness the metrics accept. We remark that the language perspective and the structural one are complementary: a structural approach is unable to consider the full extent of the impact (i.e., number of traces affected) of each missing or erroneous state or transition, while a language approach cannot provide insights about the structural similarity of the models. ### 7.2. Trace (Model) Counting A number of approaches perform trace (model) counting using the _transfer matrix method_ (Flajolet and Sedgewick, 2009; Stanley, 2011), which relies on linear algebraic operations on the matrix representing the transition relation of the automaton. For example, Aydin et al. (2015) used it to perform quantitative information flow and probabilistic analysis of software, and Aydin et al. (2018) later extended the work including also parametric constraints, still using the same trace (model) counting approach. While the transfer matrix method generates the same OGF as our method, as discussed in Section 5, our method is more scalable, enabling the assessment of larger models. Luu et al. (2014) proposed an approach to count the number of strings satisfying a given set of constraints, using _analytic combinatorics_. Constraints are individually translated to OGF, and later composed. However, both constraint translation and composition are not precise, returning an upper and lower bound instead. Trinh et al. (2017) developed a trace (model) counting approach for a class of string constraints based on S3P (an SMT solver for strings constraints) (Trinh et al., 2016). While S3P works by building a reduction tree, reducing the original formula into simpler formulas until a satisfying assignment or a contradiction is found, trace (model) counting is performed by exhaustively building the entire reduction tree, in which each node is associated with the OGF representing the count for that node of the tree, and the counts are propagated bottom-up, from the leaf nodes. ## 8\. Conclusions In the evaluation of newly developed model inference systems, rigorously assessing the accuracy of the generated inferred models against ground truth reference models is crucial. In this paper we have highlighted how commonly used assessment methods provide results that may not reflect the actual accuracy of the inferred model under analysis: statistical methods are often affected by a systematic bias caused by overlooking the impact of the random trace generation, while deterministic methods that guarantee to discover any error in the model may not correctly capture the number of traces it affects. To tackle these shortcomings, we have proposed a method to rigorously measure the accuracy of an inferred model against a ground truth reference model. First, our method is comprehensive: it considers all the possible traces up to an arbitrary finite maximum trace length defined by the user. This is in contrast with other language-based methods that perform the assessment based on a much smaller subset of traces. Second, it is deterministic: our method does not rely on random sampling, and gives exact results. While current statistical methods may provide a more flexible time/accuracy trade-off, our approach avoids the convergence limitations of statistical methods, Third, it is unbiased: it considers all the traces up to the maximum length, and each trace has the same weight on the results. This is an advantage over assessment methods using a random trace generation process to generate a finite set of traces on which the assessment result is based: the random sampling induces a probability distribution over the accepted language introducing a bias in the result, which is not desirable unless it reflects the domain knowledge about the application being analyzed. Fourth, it is model-independent: assessing against the same reference model different models accepting the same language will generate the same accuracy measurement. This is in contrast with trace similarity, where the topology of the model affects the sampling, and thus the result. We have also highlighted how characterizing model accuracy using only a pair of precision and recall values may be inadequate because an inferred model can have a variable level of accuracy depending on the _length of the traces_ used in the assessment. As a result, we have proposed an additional assessment method, measuring the accuracy separately for each trace length, over a given range. We have evaluated our assessment methods experimentally and compared ours with the currently popular assessment methods, on reference models previously used in the model inference literature and inferred models generated using well known model inference methods. The results highlighted the shortcomings of current assessment methods that are addressed by our solution. The results also show that our approach is scalable enough to be used in practice. As part of future work, we plan to further evaluate our approach on test subjects from industry, investigate how the accuracy provided by modern model inference tools changes with the length of the traces considered, and further discuss the asymptotic analysis of precision and recall and its usefulness in assessing the model accuracy. ###### Acknowledgements. The support of the EPSRC Centre for Doctoral Training in High Performance Embedded and Distributed Systems (HiPEDS, Grant Reference EP/L016796/1) is gratefully acknowledged. ## References * (1) * Aydin et al. (2015) Abdulbaki Aydin, Lucas Bang, and Tevfik Bultan. 2015\. Automata-Based Model Counting for String Constraints. In _Computer Aided Verification_ , Daniel Kroening and Corina S. Păsăreanu (Eds.). Springer International Publishing, Cham, 255–272. https://doi.org/10.1007/978-3-319-21690-4_15 * Aydin et al. (2018) Abdulbaki Aydin, William Eiers, Lucas Bang, Tegan Brennan, Miroslav Gavrilov, Tevfik Bultan, and Fang Yu. 2018. Parameterized model counting for string and numeric constraints. In _Proceedings of the 2018 26th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering - ESEC/FSE 2018_ , Daniel Kroening and Corina S. Pasareanu (Eds.). ACM Press, New York, New York, USA, 400–410. https://doi.org/10.1145/3236024.3236064 * Beschastnikh et al. (2011) Ivan Beschastnikh, Yuriy Brun, Sigurd Schneider, Michael Sloan, and Michael D. Ernst. 2011. Leveraging Existing Instrumentation to Automatically Infer Invariant-constrained Models. In _Proceedings of the 19th ACM SIGSOFT Symposium and the 13th European Conference on Foundations of Software Engineering (ESEC/FSE 2011)_. ACM, New York, NY, USA, 267–277. https://doi.org/10.1145/2025113.2025151 * Biermann and Feldman (1972) Alan W. Biermann and J. A. Feldman. 1972. On the Synthesis of Finite-State Machines from Samples of Their Behavior. _IEEE Trans. Comput._ C-21, 6 (June 1972), 592–597. https://doi.org/10.1109/TC.1972.5009015 * Bonifácio et al. (2008) Adilson Luiz Bonifácio, Arnaldo Vieira Moura, and Adenilso da Silva Simão. 2008. A Generalized Model-Based Test Generation Method. In _2008 Sixth IEEE International Conference on Software Engineering and Formal Methods_. 139–148. https://doi.org/10.1109/SEFM.2008.17 * Brzozowski and McCluskey (1963) Janusz A. Brzozowski and Edward J. McCluskey. 1963. Signal Flow Graph Techniques for Sequential Circuit State Diagrams. _IEEE Transactions on Electronic Computers_ EC-12, 2 (1963), 67–76. https://doi.org/10.1109/PGEC.1963.263416 * Busany et al. (2019) Nimrod Busany, Shahar Maoz, and Yehonatan Yulazari. 2019\. Size and Accuracy in Model Inference. In _2019 34th IEEE/ACM International Conference on Automated Software Engineering (ASE)_. 887–898. https://doi.org/10.1109/ASE.2019.00087 * Chow (1978) T.S. Chow. 1978\. Testing Software Design Modeled by Finite-State Machines. _IEEE Transactions on Software Engineering_ SE-4, 3 (1978), 178–187. https://doi.org/10.1109/TSE.1978.231496 * Clarke Jr et al. (2018) Edmund M Clarke Jr, Orna Grumberg, Daniel Kroening, Doron Peled, and Helmut Veith. 2018. _Model checking_. MIT press, Cambridge, MA, USA. * Cook and Wolf (1998) Jonathan E. Cook and Alexander L. Wolf. 1998. Discovering Models of Software Processes from Event-Based Data. _ACM Transactions on Software Engineering and Methodology_ 7, 3 (July 1998), 215–249. https://doi.org/10.1145/287000.287001 * Eggan (1963) Lawrence C. Eggan. 1963\. Transition graphs and the star-height of regular events. _Michigan Mathematical Journal_ 10, 4 (1963), 385–397. * Emam and Miller (2018) Seyedeh Sepideh Emam and James Miller. 2018. Inferring Extended Probabilistic Finite-State Automaton Models from Software Executions. _ACM Trans. Softw. Eng. Methodol._ 27, 1, Article 4 (June 2018), 39 pages. https://doi.org/10.1145/3196883 * Flajolet and Sedgewick (2009) Philippe Flajolet and Robert Sedgewick. 2009. _Analytic Combinatorics_. Cambridge University Press, Cambridge. https://doi.org/10.1017/CBO9780511801655 * Fraser and Walkinshaw (2012) Gordon Fraser and Neil Walkinshaw. 2012. Behaviourally Adequate Software Testing. In _2012 IEEE Fifth International Conference on Software Testing, Verification and Validation_. IEEE Press, Piscataway, NJ, USA, 300–309. https://doi.org/10.1109/ICST.2012.110 * Fujiwara et al. (1991) S. Fujiwara, G. v. Bochmann, F. Khendek, M. Amalou, and A. Ghedamsi. 1991. Test selection based on finite state models. _IEEE Transactions on Software Engineering_ 17, 6 (1991), 591–603. https://doi.org/10.1109/32.87284 * Inc. (2022) Wolfram Research, Inc. 2022\. _Mathematica_. https://www.wolfram.com/mathematica * Kirsten (2005) Daniel Kirsten. 2005\. Distance desert automata and the star height problem. _RAIRO - Theoretical Informatics and Applications_ 39, 3 (2005), 455–509. https://doi.org/10.1051/ita:2005027 * Krka et al. (2014) Ivo Krka, Yuriy Brun, and Nenad Medvidovic. 2014. Automatic mining of specifications from invocation traces and method invariants. In _Proceedings of the 22nd ACM SIGSOFT International Symposium on Foundations of Software Engineering - FSE 2014_. ACM Press, New York, New York, USA, 178–189. https://doi.org/10.1145/2635868.2635890 * Le et al. (2015) Tien-Duy B. Le, Xuan-Bach D. Le, David Lo, and Ivan Beschastnikh. 2015. Synergizing Specification Miners through Model Fissions and Fusions. In _2015 30th IEEE/ACM International Conference on Automated Software Engineering (ASE)_. IEEE, 115–125. https://doi.org/10.1109/ASE.2015.83 * Lo and Khoo (2006a) David Lo and Siau-cheng Khoo. 2006a. QUARK: Empirical Assessment of Automaton-based Specification Miners. In _2006 13th Working Conference on Reverse Engineering_. IEEE, 51–60. https://doi.org/10.1109/WCRE.2006.47 * Lo and Khoo (2006b) David Lo and Siau-Cheng Khoo. 2006b. SMArTIC: Towards Building an Accurate, Robust and Scalable Specification Miner. In _Proceedings of the 14th ACM SIGSOFT International Symposium on Foundations of Software Engineering_ (Portland, Oregon, USA) _(SIGSOFT ’06/FSE-14)_. Association for Computing Machinery, New York, NY, USA, 265–275. https://doi.org/10.1145/1181775.1181808 * Lo et al. (2009) David Lo, Leonardo Mariani, and Mauro Pezzè. 2009\. Automatic steering of behavioral model inference. In _Proceedings of the 7th joint meeting of the European software engineering conference and the ACM SIGSOFT symposium on The foundations of software engineering on European software engineering conference and foundations of software engineering symposium - E_. ACM Press, New York, New York, USA, 345. https://doi.org/10.1145/1595696.1595761 * Lo et al. (2012) David Lo, Leonardo Mariani, and Mauro Santoro. 2012\. Learning extended FSA from software: An empirical assessment. _Journal of Systems and Software_ 85, 9 (2012), 2063–2076. https://doi.org/10.1016/j.jss.2012.04.001 * Luu et al. (2014) Loi Luu, Shweta Shinde, Prateek Saxena, and Brian Demsky. 2014\. A Model Counter for Constraints over Unbounded Strings. In _Proceedings of the 35th ACM SIGPLAN Conference on Programming Language Design and Implementation_ (Edinburgh, United Kingdom) _(PLDI ’14)_. Association for Computing Machinery, New York, NY, USA, 565–576. https://doi.org/10.1145/2594291.2594331 * Mariani et al. (2010) Leonardo Mariani, Mauro Pezzè, Oliviero Riganelli, and Mauro Santoro. 2010. SEIM: static extraction of interaction models. In _Proceedings of the 2nd International Workshop on Principles of Engineering Service-Oriented Systems - PESOS ’10_. ACM Press, New York, New York, USA, 22\. https://doi.org/10.1145/1808885.1808891 * Mariani et al. (2017) Leonardo Mariani, Mauro Pezzè, and Mauro Santoro. 2017\. GK-Tail+ An Efficient Approach to Learn Software Models. _IEEE Transactions on Software Engineering_ 43, 8 (Aug 2017), 715–738. https://doi.org/10.1109/TSE.2016.2623623 * Polyvyanyy et al. (2020) Artem Polyvyanyy, Andreas Solti, Matthias Weidlich, Claudio Di Ciccio, and Jan Mendling. 2020\. Monotone Precision and Recall Measures for Comparing Executions and Specifications of Dynamic Systems. _ACM Transactions on Software Engineering and Methodology_ 29, 3 (2020). https://doi.org/10.1145/3387909 arXiv:1812.07334 * Pradel et al. (2010) Michael Pradel, Philipp Bichsel, and Thomas R. Gross. 2010\. A framework for the evaluation of specification miners based on finite state machines. _IEEE International Conference on Software Maintenance, ICSM_ (2010), 1–10. https://doi.org/10.1109/ICSM.2010.5609576 * Stanley (2011) Richard P. Stanley. 2011\. _Enumerative Combinatorics_ (2 ed.). Cambridge Studies in Advanced Mathematics, Vol. 1. Cambridge University Press. https://doi.org/10.1017/CBO9781139058520 * Tharwat (2020) Alaa Tharwat. 2020\. Classification assessment methods. _Applied Computing and Informatics_ 17, 1 (2020), 168–192. https://doi.org/10.1016/j.aci.2018.08.003 * Trinh et al. (2016) Minh-Thai Trinh, Duc-Hiep Chu, and Joxan Jaffar. 2016. Progressive Reasoning over Recursively-Defined Strings. In _Computer Aided Verification_ , Swarat Chaudhuri and Azadeh Farzan (Eds.). Springer International Publishing, Cham, 218–240. https://doi.org/10.1007/978-3-319-41528-4_12 * Trinh et al. (2017) Minh-Thai Trinh, Duc-Hiep Chu, and Joxan Jaffar. 2017. Model Counting for Recursively-Defined Strings. In _Computer Aided Verification_ , Rupak Majumdar and Viktor Kunčak (Eds.). Springer International Publishing, Cham, 399–418. https://doi.org/10.1007/978-3-319-63390-9_21 * Walkinshaw and Bogdanov (2013) Neil Walkinshaw and Kirill Bogdanov. 2013. Automated comparison of state-based software models in terms of their language and structure. _ACM Transactions on Software Engineering and Methodology_ 22, 2 (2013). https://doi.org/10.1145/2430545.2430549 * Walkinshaw et al. (2008) Neil Walkinshaw, Kirill Bogdanov, and Ken Johnson. 2008\. Evaluation and comparison of inferred regular grammars. _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ 5278 LNAI (2008), 252–265. https://doi.org/10.1007/978-3-540-88009-7_20 * Walkinshaw et al. (2013) Neil Walkinshaw, Bernard Lambeau, Christophe Damas, Kirill Bogdanov, and Pierre Dupont. 2013\. _STAMINA: A competition to encourage the development and assessment of software model inference techniques_. Vol. 18. 791–824 pages. https://doi.org/10.1007/s10664-012-9210-3 * Walkinshaw et al. (2016) Neil Walkinshaw, Ramsay Taylor, and John Derrick. 2016\. Inferring extended finite state machine models from software executions. _Empirical Software Engineering_ 21, 3 (jun 2016), 811–853. https://doi.org/10.1007/s10664-015-9367-7 * Waterloo Maple, Inc.. (2022) Waterloo Maple, Inc.. 2022\. _Maple_. https://www.maplesoft.com/products/maple/ * Yu (2001) Sheng Yu. 2001. State Complexity of Regular Languages. _Journal of Automata, Languages and Combinatorics_ 6, 2 (2001), 221–234. https://doi.org/10.25596/jalc-2001-221 ## Appendix A Appendix: test subjects categorization and selection In this section we discuss how we categorized the results for the test subjects in the experimental evaluation, depending on the characteristics of the trends of the accuracy metric (precision and recall) values over the range of trace length, and how we used this categorization to choose the five test subjects that are given as examples in Section 6. To differentiate the trends, we looked at the value of each accuracy metric for the shortest and the longest available trace length, with the goal of obtaining a rough indication whether the metric is increasing, decreasing, or constant. The values were aggregated in three categories: greater than 0.999 (below represented with $\mathit{1}$), smaller than 0.001 (below represented with $\mathit{0}$), and any other value (below represented with $k$). This is motivated by the observation that the precision and recall values at the extremes of the trace length range are generally close to either zero or one, and when this is not the case it is worth considering the subject separately because it indicates that in Equation 6 numerator and denominator have the same order of magnitude, which is a relevant characteristic of the inferred language. Based on this categorization, we counted the number of test subjects in each category, obtaining Table 6 (e.g. $\mathit{1}\rightarrow k$ indicates that the value of the accuracy metric value for short traces is greater than 0.999 and for long traces is between 0.001 and 0.999). Table 6. Number of test subjects for each metric trend | | Recall ---|---|--- | | $\mathit{0}\rightarrow\mathit{0}$ | $\mathit{0}\rightarrow k$ | $\mathit{0}\rightarrow\mathit{1}$ | $k\rightarrow\mathit{0}$ | $k\rightarrow k$ | $k\rightarrow\mathit{1}$ | $\mathit{1}\rightarrow\mathit{0}$ | $\mathit{1}\rightarrow k$ | $\mathit{1}\rightarrow\mathit{1}$ Precision | $\mathit{0}\rightarrow\mathit{0}$ | 0 | 0 | 0 | 1 | 0 | 0 | 20 | 2 | 0 $\mathit{0}\rightarrow k$ | 0 | 0 | 0 | 0 | 0 | 0 | 3 | 0 | 0 $\mathit{0}\rightarrow\mathit{1}$ | 0 | 0 | 0 | 0 | 0 | 0 | 2 | 0 | 0 $k\rightarrow\mathit{0}$ | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 $k\rightarrow k$ | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 $k\rightarrow\mathit{1}$ | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 $\mathit{1}\rightarrow\mathit{0}$ | 0 | 0 | 0 | 0 | 0 | 0 | 14 | 2 | 2 $\mathit{1}\rightarrow k$ | 0 | 0 | 0 | 0 | 0 | 0 | 8 | 1 | 0 $\mathit{1}\rightarrow\mathit{1}$ | 0 | 0 | 0 | 0 | 0 | 0 | 5 | 1 | 0 Looking at table 6, we observed that all the most common accuracy trends have a recall value starting at 1 and ending at 0. Within these we selected four subjects, each one having one of the most common precision value trend. Specifically: * • Subject 1 was randomly selected among the 20 subjects having precision starting at 0 and ending at 0. It is worth noting that in all 20 subjects we observed that the precision increases to a maximum value, and then decreases. * • Subject 2 was randomly selected among the 14 subjects having both precision and recall starting at 1 and ending at 0. * • Subject 3 was randomly selected among the five subjects having precision starting at 1 and ending at 1. It is worth noting that in all 5 subjects we observed that the precision is actually constant throughout the range. * • Subject 4 was randomly selected among the eight subjects having precision starting at 1 and ending at a value between 0 and 1. These trend categories cover 47 of the 61 available test subjects. In addition, we selected one more subject (Subject 5) showing a recall value ending at a value greater than zero. This is motivated by the fact that although this trend is a rare occurrence, it highlights a relevant characteristic of the inferred model. ## Appendix B Appendix: description of the selected test subjects In test subject 1 the inferred model precision starts at zero, reaches a maximum for length 10 (47.5%), and then converges to zero. A possible motivation for this behavior is suggested by the large size of the inferred model: it is possible that the inference approach could not find any good generalization of the training data, and as a result the accuracy of the model follows the distribution of lengths of the traces contained in the training set. In fact, we verified that the most frequent trace length in the training data is indeed 10. In test subject 2 the precision and recall values converge to zero, indicating that the growth of the size of the language of true positives is slower than the growth of the size of the language of both inferred positives and reference positives. For test subject 3 the constant precision equal to 1 is explained by the fact that the language of false positives is empty (which we also manually verified), indicating that the inferred language is a (non-empty) subset of the reference language. In test subject 4 the precision converges to a value greater than zero, indicating that as the trace length is increased, the cardinalities of numerator and denominator of Equation 6 have the same order of growth. The same is true for the recall in test subject 5. Notably, test subject 5 has also an empty language of false positives, as suggested by the constant precision equal to 1. Consequently, the inferred language is a subset of the reference language, and thanks to the asymptotic value of the recall we can determine that it covers 30% of the reference language — an exceptionally accurate inference result.
# CMS RPC data taking during the LHC Run-2 and activities during Long Shutdown 2 Kevin Mota Amarilo on behalf of CMS Collaboration Rio de Janeiro State University, Institute of Physics, R. São Francisco Xavier, 524, Rio de Janeiro - RJ, 20550-013, Brasil<EMAIL_ADDRESS> ###### Abstract The CMS experiment collected around 150 fb-1 of proton-proton collision data at $\sqrt{s}$ = 13 TeV during the Run-2 data taking period of LHC. The CMS RPC system provided redundant information for robust muon triggering, reconstruction and identification. To ensure stable data taking, the CMS RPC collaboration has performed detector operation, calibration and performance studies. After the end of Run-2, it was started the second LHC long shutdown period (LS2), an important opportunity for maintenance and preparation for the next data taking period (Run-3) and the installation of services in preparation for the Phase-II upgrade. The activities included maintenance of power, gas and online systems. In this presentation, the overall performance of the CMS RPC system during the Run-2 period is summarized as well as all the activities done in preparation for future data taking periods. ## 1 Introduction One of the main actors of the CMS (Compact Muon Solenoid) experiment [1] is its Muon System [2], which uses gaseous detector technologies to accomplish muon triggering, identification, transverse momentum and charge measurement. Through the end of Run 2 in 2018, the components of the CMS Muon system were: drift tubes (DT) at barrel region ($|\eta|<1.2$), cathode strip chambers (CSC) at endcap region ($1.2<|\eta|<2.4$) and Resistive Plate Chambers (RPC), in barrel and endcap regions ($|\eta|<1.9$). During LS2 a new type of detector was installed, the gas electron multiplier (GEM), as part of the Muon System Phase-2 Upgrades to complement measurements in the higher $\eta$ region for Phase-II upgrade [3], which will also include new RPC chambers, the so called improved-RPC (iRPC). The CMS RPC system is composed of double gap phenolic resin (called bakelite) chambers, operating in avalanche mode with gas mixture 95.2% $C_{2}H_{2}F_{4}$, to enhance the ionization caused by incident particles, 4.5% $iC_{4}H_{10}$ as a quencher gas to reduce streamer formation and 0.3% $SF_{6}$ to control secondary ionization. The RPCs are designed and calibrated to have a time resolution of around 2 ns and a number of adjacent strips fired in a single muon hit (Cluster size) between 2 and 3. ## 2 CMS RPC Run-2 data taking Between 2015 to 2018 during Run-2, CMS recorded proton-proton collisions with $\sqrt{s}$ = 13 TeV from LHC, with total integrated luminosity of 150.26 fb-1, out of which only 0.15 % was lost due to RPC problems. The total accumulated charge in Run-1 and Run-2 of the RPC system was 2.3 mC/cm2 in barrel and 7.5 mC/cm2 in endcap. Studies at the CERN Gamma Irradiation Facility ++ (GIF++) have showed no drop in efficiency of RPCs up to an integrated charge of 153 mC/cm2 [8] ### 2.1 RPC efficiency and cluster size stability Hit efficiency and cluster size (CLS) are important parameters for RPC performance. An RPC hit’s geometrical coordinates are calculated as the geometrical center of the cluster formed by the fired adjacent strips. The CLS of RPC hits should be kept less than 4 strips to avoid fake muon triggers, therefore, proper calibration is very important to keep good performance. The calibration is done by analyzing the efficiency and CLS dependence on the effective high voltage (HV), which is the high voltage applied corrected by the environmental pressure and temperature variations. HV working point scans were performed once or twice a year in dedicated collision runs, effective voltage is applied in values between 8600 and 9800 V. The collected data is analyzed and the proper working points are selected. More information on the calibration can be found in [4]. Figure 1: RPC average efficiency vs integrated luminosity during Run-2 for barrel stations. Red vertical lines show the planned technical stops (TS) and the grey ones — Year-End-Technical stops (YETS). Figure 2: RPC average cluster size vs integrated luminosity during Run-2 for barrel stations. Red vertical lines show the planned technical stops (TS) and the grey ones — Year-End- Technical stops (YETS). The Figures 1 and 2 shows the Run-2 efficiency and CLS history. Each point corresponds to the average efficiency or CLS per barrel station. Each change in the trends is related to the deployment of a new working point. With exception to a drop in efficiency and CLS between in August 2018 caused by a configuration setting issue, it is possible to see that the RPC system had a stable performance with efficiency greater than 95 % and CLS between 2 and 3, within the CMS Muon Trigger requirements. ### 2.2 RPC contribution to the CMS Muon Trigger The main contribution of the RPC system is to the CMS Level-1 (L1) Muon Trigger [5], which is divided into 3 $\eta$-regions: the barrel muon track finder (BMTF) for $|\eta|<0.83$, the overlap muon track finder (OMTF) for $0.83<|\eta|<1.24$ and the endcap muon track finder (EMTF) for $|\eta|>1.24$. Each one of the track finders can access information of the detectors concurrently, building tracks and assigning pT and exploiting the redundancies of the muon system. In the BMTF, RPC timing information is used to improve DT trigger primitives and bunch crossing assignment. In the OMTF, RPC-only segments are build from the 8 chambers available, resulting in a substantial gain in in efficiency. Figure 3 demonstrates that using RPC information in the OMTF increases efficiency by 15 %. Finally, in the EMTF, RPC hit position is used in case of CSC trigger primitive absence. More information on this subject can be found in [6]. Figure 3: Trigger efficiency versus pT for the OMTF derived from algorithm emulation applied on real data. Using (Red) and not using (Blue) RPC information. Ref. [6]. ### 2.3 RPC ohmic current monitoring The ohmic current is defined as the current with no beam, up to around 7000 V, where the gas amplification contribution is negligible and the currents follow the Ohm’s law. Figure 4 shows the history of ohmic current measured in four RPC stations. An increase in the currents was observed in all stations, correlated to the background, the RE+4 and RE-4 background rate is about 40 Hz/cm2 and W+0 and RE-1 is less than 10 Hz/cm2. In November of 2018, at the start of the Heavy Ion period where the luminosity is very low and the background is effectively zero, the currents started to decrease. In RE-4 the decrease was faster with the increase of the gas flux. The increase was correlated with the production of flourine ions (F-) in the gas gap due to the electrical discharge inside the gap. The flourine can combine with the water in the gas and form HF, that can damage the gap. This explains the velocity of decrease of currents correlation with gas flux, as the gas can flush out the F-ions before the combination with water. More information on the fluorine formation inside RPC gas gap can be found in [7]. Figure 4: RPC ohmic current monitoring history for the stations W+0, RE+1, RE+4 and RE-4. ## 3 LS2 Activities During LS2, CMS is undergoing an intensive upgrade and maintenance program during the second two-year-long shutdown period (LS2). To ensure an excellent performance of the detector in the subsequent physics program, the RPC group pursued in the present shutdown a thorough detector consolidation program to repair most of the hardware problems. In preparation for future installation of new iRPC chambers, cooling and cable services for the new detectors were installed. This includes thousands of kilometres of high voltage and low voltage cables, stainless steel gas pipes between predistribution gas racks in the Service cavern (USC) and gas distribution racks in the experimental cavern (UXC), copper pipes between distribution racks and chambers, gas impedance boxes, support equipment, and optical fibers for the iRPCs. In order to keep the optimal performance of the system, an extensive HV and low voltage (LV) maintenance campaign was performed. The goal of HV maintenance was to identify the problematic parts of the HV power system and to fix it in the best possible way, recuperating the performance of the chambers. A total of 65 HV channels were repaired. The LV maintenance aim was to ensure a proper operation and configuration of the detector electronics and ensure a good functionality of the LV power boards and communication buses. A total of 12 LV problems were fixed. A very important activity was the extraction of the chambers from the two RE4 stations to allow the CSC ME4/1 chamber extraction for electronics refurbishment. The chambers were brought to the surface, and accommodated to a new laboratory with controlled environmental conditions and new gas lines to provide the standard RPC gas mixture. All needed reparation and re-validation was done in the laboratory before the re-installation. The activity with most priority during LS2, was gas system consolidation. The aim was to minimize the environmental impact of the RPC system, as the standard gas mixture is mainly composed of fluorine composed gases (F-gases) with high global warming potential (GWP). The actions taking place during LS2 are: * • Gas Leak identification and repair, where 49 out of 99 leaky chambers in barrel where repaired. * • Recuperation of the Exhaust, which was not working during Run-2 for the installation of the first $C_{2}H_{2}F_{4}$ recuperation system with efficiency of 80 %, which have been developed by CERN EP-DT Gas team. * • By end of 2021, CERN EP-DT Group is going to install automatic pressure regulation valves on the redistribution gas racks in USC to minimize pressure variations in the chambers, which can be a possible source of new leaks. * • Turn off the remaining leaky chambers which was not possible to repair (about 3.5 % of RPC system), to keep the amount of fresh gas added to the system at a minimum. ## 4 RPC commissioning with cosmic rays data during LS2 In Figure 5, the efficiency of the barrel chambers is compared using cosmic rays runs of 2018 (End of Run-2) and in 2021 (after all the repairs of LS2). The average efficiency of the chambers with efficiency greater than 70 % is compared and is about the same (gain of 0.6 %). The fraction of chambers with efficiency greater than 70 % increased in around 6 % as a result of all the repairs, which is very important for the muon system redundancy. Figure 5: RPC barrel efficiency comparison between cosmic runs of 2018 (Red) and 2021 (Blue). ## 5 Conclusion The CMS RPC system showed a stable performance after operation in Run-1 and Run-2 ($\sim$185 fb-1) with average efficiency greater than 95 % and average CLS between 2 and 3 strips/hit. We observed reversible ohmic current increase in high background regions correlated to the F- formation inside the gaps. During LS2 RPC did a massive maintenance campaign to repair gas leaks, HV and LV in order to keep the optimal performance and is planning to keep the greenhouse gases emissions at a minimum by keeping all the remaining leaky chambers off and using the new $C_{2}H_{2}F_{4}$ recuperation system. The performance of the RPC System is improved with respect to the end of Run-2 and the system is ready and commissioned for Run-3 data taking. ## References ## References * [1] CMS Collaboration 2008 The CMS experiment at the CERN LHC JINST 3 S08004 doi:10.1088/1748- 0221/3/08/S08004 * [2] CMS Collaboration 1997 The CMS muon project: Technical Design Report (Geneva: CERN) URL `https://cds.cern.ch/record/343814` * [3] CMS Collaboration 2017 The Phase-2 Upgrade of the CMS Muon Detectors (Geneva) URL `https://cds.cern.ch/record/2283189` * [4] Abbrescia M et al. 2005 Cosmic ray tests of double-gap resistive plate chambers for the CMS experiment, Nucl. Instrum. Meth. A 550 116 URL `https://www.sciencedirect.com/science/article/pii/S0168900205011915` * [5] CMS Collaboration 2013 CMS Technical Design Report for the Level-1 Trigger Upgrade (Geneva) URL `https://cds.cern.ch/record/1556311` * [6] A. Samalan et al 2020 RPC system in the CMS Level-1 Muon Trigger JINST 15 C10007 * [7] A. Gelmi, R. Guida and B. Mandelli on behalf of CMS Muon 2021 Gas mixture quality studies for the CMS RPC detectors during LHC Run 2 JINST 16 C04004 * [8] J. Eysermans and M. I. Pedraza Morales 2017 Aging studies for the CMS RPC system. Proceedings of Science. URL `https://cds.cern.ch/record/2286783/files/EPS-HEP2017_785.pdf`
ifaamas [AAMAS ’23]Proc. of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2023)May 29 – June 2, 2023 London, United KingdomA. Ricci, W. Yeoh, N. Agmon, B. An (eds.) 2023 2023 ??? # Artificial prediction markets present a novel opportunity for human-AI collaboration Anonymous submission ###### Abstract Despite high-profile successes in the field of Artificial Intelligence, machine-driven technologies still suffer important limitations, particularly for complex tasks where creativity, planning, common sense, intuition, or learning from limited data is required. These limitations motivate effective methods for human-machine collaboration. Our work makes two primary contributions. We thoroughly experiment with an artificial prediction market model to understand the effects of market parameters on model performance for benchmark classification tasks. We then demonstrate, through simulation, the impact of exogenous agents in the market, where these exogenous agents represent primitive human behaviors. This work lays the foundation for a novel set of hybrid human-AI machine learning algorithms. ###### Key Words.: prediction markets; machine learning; artificial intelligence; human-AI collaboration ## 1 Introduction A body of work on artificial prediction markets is emerging. These are numerically simulated markets, populated by artificial agents (bot-traders) for the purpose of supervised learning of probability estimators barbu2012introduction. While nascent, this literature has demonstrated the plausibility of using a trained market as a supervised learning algorithm, achieving comparable performance to standard approaches on simple classification tasks barbu2012introduction; barbu2013artificial; jahedpari2014artificial; nakshatri2021design. In fact, these results are sensible given the deep mathematical connections between prediction markets and learning chen2008complexity; chen2010new; abernethy2011optimization. Like other machine learning algorithms, functioning of an artificial prediction market depends on several researcher-determined parameters: number of agents; liquidity; initial cash; alongside parameters related to training processes. Scenarios in which performance is robust or brittle to these settings is yet unclear. Prior work has observed that artificial markets may suffer from lack of participation rajtmajer2022synthetic. That is, like their human counterparts in traditional prediction markets, agents may not invest in the market if they do not have sufficient information Arrow:2008; tetlock2008liquidity; rothschild2014extent; in practice, this occurs when an asset representing test data point is too dissimilar to training examples. In our view, the most promising opportunity afforded by artificial prediction markets is eventual human-AI collaboration – a market framework should theoretically support human traders participating alongside agents to evaluate outcomes. Whether and how artificial prediction markets might benefit from this hybrid scenario is an open question. The work we undertake here provides, through simulation, initial support for this opportunity in the context of a simple artificial market and primitive human behaviors. Our work is framed by two primary research questions. RQ1: How does performance of a simple artificial prediction market depend on hyper-parameter selection? RQ2: What impact does the inclusion of exogenous agents representing simple (human-like) behaviors have on market performance? Our findings support those of prior recent work indicating the promise of artificial prediction markets for classification tasks. We demonstrate the sensitivity of this approach to hyper-parameter selection and highlight, in particular, the role of liquidity in moderating performance. Finally, we demonstrate the exciting opportunity for _hybrid prediction markets_ to serve as a framework for human-AI collaboration. We suggest that this approach may be particularly valuable in contexts where machine learning falls short (e.g., lack of training data, complex tasks) and potential for human-only approaches is either undesirable or infeasible. ## 2 Related Work Our work builds upon and contributes to two primary literatures, namely, work on artificial prediction markets and work in collaborative human-AI technologies. ### 2.1 Artificial Prediction Markets Prediction markets are simple futures markets used to aggregate disperse information into efficient forecasts of uncertain future events wolfers2004prediction; hanson2006information; manski2006interpreting; wolfers2006interpreting. Specifically, market participants buy and sell contracts that pay out based on the outcomes of future events. Market prices generated from these contracts can be understood as a collective prediction among market participants. Prediction markets have been successfully used, e.g., for forecasting election outcomes berg2008prediction, sports betting spann2009sports, forecasting infectious disease activity polgreen2007use, and aggregating employee wisdom in corporate settings cowgill2009using; gillen2012information. Artificial prediction markets are a variation on this idea, wherein numerically simulated markets populated by trained agents (bot-traders) are used for the purpose of supervised learning of probability estimators barbu2012introduction; barbu2013artificial. In initial formulations by Barbu and Lay lay2010supervised; barbu2012introduction; lay2012artificial, each agent is represented as a budget and a simple betting function. During training, each agent’s budget is updated based on the accuracies of its predictions over a training dataset. Authors found that these markets outperformed standard approaches on benchmark classification and regression tasks. Later, Storkey and colleagues storkey2011machine; storkey2012isoelastic developed the so-called machine learning market, also for the purpose of classification. In their formulation, each agent purchases contracts in order to maximize a utility function. Most recently, Nakshatri et al. nakshatri2021design proposed an artificial prediction market wherein agent purchase logic is defined geometrically, in particular, by a convex semi- algebraic set in feature space. Time varying asset prices affect the structure of the semi-algebraic sets leading to time-varying agent purchase rules. Agent parameters are trained using an evolutionary algorithm. Authors show that their approach has desirable properties, e.g., the market satisfies certain universal approximation properties, and there exist sufficient conditions for convergence. Our work builds on this approach. Like their human-populated counterparts, artificial prediction markets have found a number of real-world applications barbu2013artificial; jahedpari2014artificial. Ongoing theoretical work has offered support for these promising experimental findings, highlighting the mathematical connections between artificial markets and machine learning chen2008complexity; ChenPennock:2010; abernethy2011optimization; hu2014multi. ### 2.2 Human-AI Collaboration Despite high-profile successes in the field of Artificial intelligence (AI) he2015delving; brown2019superhuman; kleinberg2018human; zhu2018human, machine- driven solutions still suffer important limitations particularly for complex tasks where creativity, common sense, intuition or learning from limited data is required jarrahi2018artificial; lai2019human; green2019principles; li2016crowdsourced; kamar2016directions; amershi2019guidelines; muller2017organic. Both the promises and challenges of AI have motivated work exploring frameworks for human-machine collaboration dellermann2021future; wang2019human; nunes2015survey; puig2020watch; wu2022survey. The hope is that we can eventually develop hybrid systems that bring together human intuition and machine rationality to effectively and efficiently tackle today’s grand challenges. Recent work in hybrid intelligence systems has demonstrated the feasibility and highlighted the potential of integrating human input into AI systems kamar2016directions, or even, of human-AI collaboration wang2020human. The spectrum of these efforts range from accounting for human factors in technology design bansal2019beyond; canonico2019wisdom; harper2019role to efficiently utilizing human inputs for training data amershi2014power in applications as diverse as business nagar2011making; sowa2021cobots, civic welfare fogliato2022case, criminal justice travaini2022machine, and healthcare tschandl2020human; lee2021human; rajpurkar2022ai. The work we describe here brings together the bodies of prior work on artificial prediction markets and hybrid intelligence, proposing hybrid prediction markets for direct integration of human wisdom into the deployment of a machine learning algorithm. ## 3 Data We consider three classification tasks. The first two are benchmark tasks used broadly to compare performance of machine learning algorithms. The third is the task of classifying scientific research outcomes as replicable or not replicable – a challenging, complex task on which both machine learning algorithms altmejd2019predicting; yang2020estimating; pawel2020probabilistic; wu2021predicting and human assessment dreber2015using; camerer2016evaluating; camerer2018evaluating; forsell2019predicting; gordon2020replication; gordon2021predicting have achieved respectable but not excellent performance. The replication prediction task, we suggest, is an example of the type of problem well-suited to hybrid human-AI approaches and, specifically, the hybrid prediction markets we propose. ### 3.1 Benchmark Machine Learning Datasets The Iris dataset misc_iris_53 is one of the best known datasets in statistics and machine learning, and one the earliest datasets used for evaluation of classification methodologies. The dataset contains three classes of 50 instances each, where each class refers to a type of iris plant. One class is linearly separable from the other; the latter are not linearly separable from each other. Approaches based on support vector classification mohan2020support, random forest classification mishina2015boosted; chicho2021machine, and logisitc regression pinto2018iris have been reported to work very well on this task, achieving $100\%$ or near-$100\%$ accuracy. The Heart Disease dataset misc_heart_disease_45 is also a multivariate dataset used for benchmark classification algorithms. Fourteen patient attributes are used to predict presence or absence of heart disease. Random forest singh2016heart, Xgboost rajadevi2021feature, and logistic regression desai2019back achieve performance just under $90\%$ accuracy. While, support vector classification achieves $86\%$ rajadevi2021feature. ### 3.2 Replication Studies Outcomes In the last decade, several large-scale replication projects have been undertaken across psychology, economics, political science, cancer biology and other domains open2015estimating; camerer2016evaluating; camerer2018evaluating; klein2014investigating; klein2018many; cova2021estimating; errington2014open. Amongst their important impacts, these studies have created small ground-truth datasets of replication studies outcomes that can be used for train and test of automated approaches for replication prediction. Specifically, we use the dataset and extracted features considered by rajtmajer2022synthetic for ease of comparison. The dataset containes 192 findings in the social and behavioral sciences, each labeled either Replicable or Not Replicable, and a set of 41 features extracted from each associated paper representing biblometric, venue-related, author-related, statistical and semantic information. See wu2021predicting for further detail on feature extraction processes. Of note, authors in rajtmajer2022synthetic achieve $89.4\%$ accuracy, remarkable for the task of replication prediction. However, accuracy is calculated based on the approximately one-third of the test data that gets evaluated by the market. Because agent participation is voluntary and agents do not participate if they do not have sufficient information about a test point, some (or much) of the data can be left unclassified. Our work uses the same data and market structure described in rajtmajer2022synthetic. This allows us to explore the effects of hyper-parameters (RQ1) and the inclusion of exogenous agents (RQ2) on these performance/participation trade-offs. ## 4 Prediction Market Model We use as a base model the artificial binary prediction market described in nakshatri2021design. The state of the prediction market is defined by a pair of integers $\mathbf{q}_{t}=(q^{0}_{t},q^{1}_{t})\in\mathbb{Z}_{+}^{2}$ giving the number of units of the two asset classes that have been sold. For simplicity we refer to the assets as $0$ and $1$. Traders are agents $\mathcal{A}=\\{a_{1},\dots,a_{n}\\}$ who buy assets $0$ and $1$ using policies $\\{\gamma_{1},\dots,\gamma_{n}\\}$. Also following nakshatri2021design, we assume for simplicity that agents cannot sell. If agent purchase policy $\gamma_{i}$ is conditioned on exogenous information $\mathbf{x}\in D\subseteq\mathbb{R}^{n}$ then, $\gamma_{i}:(\mathbf{q}_{t},\mathbf{x})\mapsto(r^{0},r^{1})$ and agent $i$ purchases $r^{0}$ units of $A_{0}$ and $r^{1}$ units of $A_{1}$, thus causing a state update. In what follows, we assume that agents specialize in the purchase of either Asset $0$ or Asset $1$ so that if $r^{0}>0$, then $r^{1}=0$. Asset prices are computed using a logarithmic market scoring rule (LMSR): $\displaystyle p^{0}_{t}=\frac{\exp{(\beta q^{0}_{t})}}{\exp{(\beta q^{0}_{t})}+\exp{(\beta q^{1}_{t})}}$ $\displaystyle p^{1}_{t}=\frac{\exp{(\beta q^{1}_{t})}}{\exp{(\beta q^{0}_{t})}+\exp{(\beta q^{1}_{t})}}.$ This is the softmax function of $(q^{0}_{t},q^{1}_{t})$. Liquidity $\beta$ adjusts the price change given a change in asset quantities lekwijit2018optimizing. The fact that prices vary as a function of $\mathbf{q}_{t}$ ensures that the policy need not take spot price into consideration explicitly. It is often more convenient to work in units of $1/\beta$ as $\beta$ can become arbitrarily close to zero. Experimental results are therefore reported for this quantity as the liquidity factor. To start the market, all agents may purchase assets at time $t=0$. After this, we assume that agents arrive at the market with arrival rate $\lambda$ and inter-arrival time governed by an exponential distribution. This allows us to avoid scenarios in the hybrid setting where the synthetic traders swamp the market. The LMSR imposes a market maker price, so that actual trade costs are given by: $\displaystyle\kappa_{t}^{0}(\Delta q^{0})=\frac{1}{\beta}\log\left\\{\frac{\exp[\beta(q^{0}_{t}+\Delta q^{0})]+\exp[\beta q^{1}_{t}]}{\exp[\beta q^{0}_{t}]+\exp[\beta q^{1}_{t}]}\right\\}$ $\displaystyle\kappa_{t}^{1}(\Delta q^{1})=\frac{1}{\beta}\log\left\\{\frac{\exp[\beta q^{0}_{t}]+\exp[\beta(q^{1}_{t}+\Delta q^{1})]}{\exp[\beta q^{1}_{t}]+\exp[\beta q^{0}_{t}]}\right\\}.$ Here $\kappa^{i}_{t}(\Delta q^{i})$ is the cost to a trader for purchasing $\Delta q^{i}$ units of Asset $i$ (with $i\in\\{0,1\\})$ at time $t$. For small values of $\beta$ (large values of $1/\beta$) the cost of purchase approaches the spot-price nakshatri2021design. Agent purchase logic is governed by a time-varying bank value $B_{i}$ and a characteristic function $\psi_{i}:\mathbb{R}^{n}\times\mathbb{R}^{2}\times\mathbb{R}^{m}\to\mathbb{R}$ to reason about information $\mathbf{x}$ and its decision to buy an asset in class $y_{i}$ is governed by: $\Delta q^{y_{i}}_{i}=H\left\\{\sigma[\psi_{i}(\mathbf{x},\mathbf{q};\bm{\theta})]-\kappa^{y_{i}}\right\\}\cdot H\left(B_{i}-\kappa^{y_{i}}\right).$ (1) Here $\sigma:\mathbb{R}\to[0,1]$ is a sigmoid function and $H(x)$ is the unit step function defined as $0$ at $x=0$. The expression $\sigma[\psi_{i}(\mathbf{x},\mathbf{q})]$ defines the value Agent $i$ places on Asset $y_{i}$ as a function of the market state (and hence spot-prices) and the information in the external information $\mathbf{x}$. If Agent $i$ places more value on Asset $y_{i}$ than its present price $\kappa^{y_{i}}$, then $H\left\\{\sigma[\psi_{i}(\mathbf{x},\mathbf{p})]-\kappa^{y_{i}}\right\\}=1$ and $\Delta q^{y_{i}}_{i}=1$ just in case the agent has sufficient funds given by $H\left[B_{i}-\kappa^{y_{i}}\right]$. That is, Agent $i$ purchases a share of Asset $i$. Notice we are assume that agents may buy one share of an asset at a time. This both simplifies the agent logic and also would prevent the agents from out-competing humans in the market in the hybrid scenario. The vector $\bm{\theta}$ is a set of parameters that define the specific outputs of $\psi_{i}$ and thus affect the agent purchase logic. Let $\bm{\Theta}$ be the (matrix) of all parameter vectors for the agents. After running for $T$ time units with input information $\mathbf{x}$, the spot price for Asset $1$ is $p^{1}_{T}(\mathbf{x};\bm{\Theta})$. If we are given input information $\\{\mathbf{x}_{1},\dots,\mathbf{x}_{N}\\}$ with class information $\\{y_{1},\dots,y_{N}\\}$, then training the market is the process of solving: $\min_{\bm{\Theta}}\;\;\frac{1}{N}\sum_{j=1}^{N}\left\lvert p^{1}_{T}(\mathbf{x}_{j};\bm{\Theta})-y_{j}\right\rVert^{2}.$ This problem is solved in nakshatri2021design using a genetic algorithm to obtain a market that can classify external information $\mathbf{x}\in D$. At the close of the market, the price of a each asset is taken as a proxy for the market’s confidence in the corresponding outcome. In our binary market model, there are two mutually exclusive possible outcomes and so the (normalized) prices should sum to $1$. In this way, the market can be used for regression or classification. In the three examples we consider here, the market is used for classification. A separate market is run for each point in the test set and the asset with the higher price is considered the market’s classification decision for that test point. We note, critically, that based on this model, agent participation is voluntary and decision to participate is driven by $\Delta q^{y_{i}}_{i}=1$ from Equation 1. If this condition is not met during the course of the market for any agents, there will be no market activity and thus no classification decision for that test point. Authors in rajtmajer2022synthetic have noted that this may occur frequently, particularly in cases where the training data set is small or points in the test set are significant different from training the data. Accordingly, we calculate accuracy and F1 based on the scored subset of the data, while also reporting participation as a performance metric. The artificial prediction market model includes five hyper-parameters that are not optimized by the genetic algorithm discussed in nakshatri2021design: 1. 1. Agent inter-arrival rate ($\lambda$); 2. 2. Agent initial bank value ($B_{i}(0)$); 3. 3. Market liquidity ($1/\beta$); 4. 4. Simulation running time ($T$) or duration; 5. 5. Number of generations in the genetic (training) algorithm. As such, these parameters are researcher-determined and warrant further study (RQ1). Our first set of experiments, described below, explore the specific roles of agent inter-arrival rate ($\lambda$), agent initial bank value ($B_{i}(0)$) (or, “cash”), and market liquidity ($1/\beta$) on performance. We explore the robustness of performance to selection of these hyper-parameters, highlighting accuracy and F1 score but also trade-offs with agent participation. In experiments that follow, the genetic algorithm is trained over five generation. The objective function of the genetic algorithm maximizes root mean square error of the estimated score. Agent performance is evaluated based on profit; nonprofitable agents are deleted from the pool. The ten most profitable agents are retained and, amongst them, the seven most profitable agents are selected for mutation and crossover. ## 5 Experimental Design The experiments we describe support the two primary research questions we have laid out, respectively. First, we capture the effects of different combinations of hyper-parameters on market performance (RQ1). Second, we explore the impact of exogenous agents not trained through the evolutionary training process, but rather who adopt one of a set of three simple purchasing rules meant to represent primitive human inputs (RQ2). ### 5.1 Market robustness to hyper-parameters We study the effects of inter-arrival rate $\lambda$, agent initial bank value $B_{i}(0)$ (or, “cash”), and market liquidity $1/\beta$ on artificial market performance. As mentioned, number of generations is fixed at five during training; while, market duration is fixed at 20. These parameters were fixed (vs. manipulated) to avoid combinatorial complexity during this initial study; however, they should be further studied in future work. In practice, we have found these values to be sufficient for market behavior to converge while also offering reasonable run time. Liquidity is tested for the set of values $\\{5,10,20,50,75,100,150,200,300\\}$. Initial cash is tested for $\\{1,2,3,4,5,10,20\\}$; $\lambda$ is tested for $\\{0.1,0.25,0.5,0.01,0.025,0.05,1\\}$. Our experiments consider all combinations of these hyper-parameter values, $441$ total, and measure corresponding performance in terms of accuracy, F1 score, and agent participation (i.e., total number of scored test points). Performance for each hyper-parameter set is determined based on 5-fold cross validation with 80/20 train/test splits. From these outcomes, we select best and worst-performing hyper-parameter sets to be used for downstream analyses. This process is outlined in Figure 10. ### 5.2 Market behavior with exogenous agents We introduce three classes of exogenous agents representing simple, fundamental behaviors which operate fully separate from the agent logic and feature-based training protocol used for the other agents in the market. These classes of behavior are intended to represent behavioral primitives that, in combination, would underlie the actions of human participants in a hybrid scenario. The first, _ground truth_ agents (GT) have perfect knowledge of the correct outcome and always buy contracts corresponding to the correct outcome whenever they have the opportunity to participate (which is moderated by their arrival rate, $\lambda$). The second are _ground truth inverse_ agents (GTinv). These agents also know the correct outcome but always buy contracts corresponding to the incorrect outcome whenever they have an opportunity to participate. This scenario is equivalent to the case where agents are simply certain but incorrect in their forecast. Finally, our third class of agents are _random_ agents which purchase contracts corresponding to one or the other outcome randomly. Understanding that the decisions of human participants in the hybrid prediction market would likely not exactly fall into any of these three categories, we consider our simulations of markets with agents drawn from these classes as drawing initial boundaries around the type of impact human participants might have on the performance of an artificial market depending on the complexity of the forecasting task, e.g., there may be a classification task which is very easy for humans but difficult for algorithms wherein we would expect near-perfect performance from human participants. Our experiments measure impact of exogenous agents on market performance measured, as before, by accuracy, F1 score, and agent participation. Because exogenous agents are not trained, they are not subject to the genetic algorithm. Rather, exogenous agents are added directly to the agent pool during test. Our experiments test the impact of adding varying number of agents from each class. We specify this number based on percentage of the total agent pool. Specifically, we test hybrid market performance with the inclusion of $\\{0.1\%,0.5\%,1\%\\}$ GT and GTinv agents. We test hybrid market performance with the inclusion of random agents accounting for $\\{1\%,5\%,10\%,50\%\\}$ of the total agent pool. Random agents are included at a higher rate given the comparatively lesser impact they have on asset prices. ## 6 Experimental Results This section discusses the results obtained from all the experiments executed on the three datasets mentioned above. We have analyzed the performance of the artificial prediction market using different metrics to capture the degree of coherence between observations and predictions. Owing to the size of these datasets, we chose to go for 5-fold cross validation to ensure that we make best use of the entire data and gather more accurate metrics in terms of the performance of our markets across various hyper-parameter combinations. Our market has been generalised to work like a binary classifier, therefore, accuracy and F1 score were extracted as the performance evaluation metrics. Accuracy is the fraction of correct predictions made by our model. For binary classification, accuracy can also be calculated in terms of two positives (True Positives(TP) and False Positives(FP)) and two negatives (True Negatives(TN) and False Negatives(FN))as follows. $\displaystyle Accuracy=\frac{TP+TN}{TP+TN+FP+FN}$ Apart from the accuracy of predictions, we also wanted to understand how well our model classifies across the two classes. Hence, we chose F1-Score as our evaluation metric since it combines the Precision and Recall of a classifier (which means that it accounts for both FPs and FNs) into a single metric by taking a harmonic mean as shown below. $\displaystyle F1-Score=2*\frac{\text{Precision}*\text{Recall}}{\text{Precision}+\text{Recall}}$ If a model produces no false positives then the precision of the model is 1. F1-score ranges between 0 and 1. Higher the F1, better the model. ### 6.1 Analysis using IRIS Data We started off our tests with the standard IRIS dataset which contains total 150 datapoints of 50 classes each as mentioned in Section 3.1. As shown in Figure 10, we capture the average accuracy and average f1 for all experimentally generated hyper-parameter combinations. Figure 1 highlights the average F1 score with different combinations of $\lambda$, initial cash and liquidity. From the colour datapoints in the figure, we can infer that the best F1 scores are obtained when initial cash ranges from 1-4 and when liquidity is greater than 100. We obtained a best case accuracy of 94.1% with a F1 score 0.91 for the combination (liquidity: 300, lambda: 1, initial cash:1). Hence, we a small subset was taken for deeper understanding to find clearer patterns across combinations by fixing a pair from this set and varying the third parameter. From Table 1, 2, 3, we can clearly observe that initial cash and liquidity tends to show a pattern in terms of the affect on performance of the market while it’s relatively difficult to assess the impact of $\lambda$ on the overall changes in captured metrics. The market performance increases as liquidity increases (Table 3) but decreases when initial cash increases(Table 1). For example, the accuracy is 94% when the initial cash is 1 but it decreases to 75% when the cash increased to 20. Figure 2 showcases the average accuracy vs F1 score for the IRIS data with different values of initial cash. In most of the cases, F1 scores are inline with the accuracy and the best numbers are seen with an initial cash of 1 (highlighted with blue markers). Table 1: Average F1 Score and Accuracy with different values of initial cash for best hyper-parameters of IRIS data Initial Cash | Liquidity | $\lambda$ | Accuracy | F1 Score ---|---|---|---|--- 1.0 | 300.0 | 1.0 | 0.94 | 0.91 2.0 | 300.0 | 1.0 | 0.81 | 0.58 3.0 | 300.0 | 1.0 | 0.81 | 0.64 4.0 | 300.0 | 1.0 | 0.87 | 0.76 5.0 | 300.0 | 1.0 | 0.76 | 0.42 10.0 | 300.0 | 1.0 | 0.75 | 0.35 20.0 | 300.0 | 1.0 | 0.75 | 0.37 | | | | Table 2: Average F1 Score and Accuracy with different values of $\lambda$ for best hyper-parameters of IRIS data $\lambda$ | Liquidity | Initial Cash | Accuracy | F1 Score ---|---|---|---|--- 0.100 | 300.0 | 1.0 | 0.87 | 0.79 0.250 | 300.0 | 1.0 | 0.87 | 0.80 0.500 | 300.0 | 1.0 | 0.87 | 0.78 0.010 | 300.0 | 1.0 | 0.87 | 0.79 0.025 | 300.0 | 1.0 | 0.91 | 0.86 0.050 | 300.0 | 1.0 | 0.88 | 0.82 1.000 | 300.0 | 1.0 | 0.94 | 0.91 | | | | Table 3: Average F1 Score and Accuracy with different values of Liquidity for best hyper-parameters of IRIS data Liquidity | $\lambda$ | Initial Cash | Accuracy | F1 Score ---|---|---|---|--- 5.0 | 1.0 | 1.0 | 0.67 | 0.00 10.0 | 1.0 | 1.0 | 0.67 | 0.00 20.0 | 1.0 | 1.0 | 0.67 | 0.03 50.0 | 1.0 | 1.0 | 0.75 | 0.33 75.0 | 1.0 | 1.0 | 0.85 | 0.72 100.0 | 1.0 | 1.0 | 0.86 | 0.74 150.0 | 1.0 | 1.0 | 0.86 | 0.76 200.0 | 1.0 | 1.0 | 0.88 | 0.81 300.0 | 1.0 | 1.0 | 0.94 | 0.91 | | | | Figure 1: Average F1 with different values of $\lambda$, liquidity and initial cash for Iris Data Figure 2: Average accuracy vs average F1 with different values of Initial Cash for Iris Data ### 6.2 Analysis using Heart Disease Data The complete set of 441 combinations were generated for heart disease dataset as mentioned in Section 3.1. Figure 3 shows the average F1 score for all 441 combinations of hyper-parameters where the yellow colored datapoints represent better F1 score than the magenda ones. These yellow colored datapoints are spreaded all over the combinations of hyper-parameters set. Figure 4 describe the obtained average accuracy with the average F1 score with different values of $\lambda$. There is no effect of lambda on market performance as we can see it is not possible to infer any pattern form the figure 4. The main intention to plot the figure is to show there is no proper best/worse value of $\lambda$ regarding market performance. The best performance was obtained with accuracy 66% and F1 score 0.71 when only one datapoint was not scored by the model. Now, to have more clear insight we took a small subset from the 441 combinations to explore more with these parameters. The best market performance was obtained when the liquidity is in the range (50-75), but there’s no defined pattern in the performance with change in values for $\lambda$ and initial cash in case of Heart data. From table 4 we can infer that with Only in case of increase in liquidity, effect the performance of the model. For heart data when we have liquidity 50 the obtained accuracy is 66% but it was reduced to 57% when liquidity is 200 which means higher values of liquidity are not good for this dataset. Table 4: Average F1 and Accuracy with different Liquidity for Heart Disease Data Liquidity | $\lambda$ | Initial cash | Accuracy | F1 score ---|---|---|---|--- 50.0 | 0.05 | 20.0 | 0.66 | 0.71 20.0 | 0.05 | 20.0 | 0.60 | 0.66 75.0 | 0.05 | 20.0 | 0.61 | 0.64 200.0 | 0.05 | 20.0 | 0.57 | 0.62 100.0 | 0.05 | 20.0 | 0.58 | 0.62 300.0 | 0.05 | 20.0 | 0.59 | 0.61 150.0 | 0.05 | 20.0 | 0.58 | 0.59 10.0 | 0.05 | 20.0 | 0.54 | 0.56 5.0 | 0.05 | 20.0 | 0.59 | 0.54 | | | | Figure 3: Average F1 with different values of $\lambda$, liquidity and initial cash for Heart Data Figure 4: Average accuracy vs average F1 with different values of $\lambda$ for Heart Data ### 6.3 Analysis using Replication Data We further repeated the phase 1 of the experiments on replication data to understand how the market parameters change across datasets of varying complexities and to capture similarities, dissimilarities and patterns across datasets. This generalization would be helpful for future research in this area. The average F1 score for all hyper-parameter combinations can be seen in Figure 5. Figure 6 is the density plot for the replication data across the accuracy and F1 score for all those 441 combinations. Figure 8 shows the average accuracy with average F1 score for the replication data with different values of initial cash. We observed that the market performs well with initial cash of 1 and 20 respectively in some of the cases. From these plots, we can further infer that there’s no clear best/worst performing range of initial cash and lambda for the replication data as well. Table 5: Agent participation with different values of Liquidity for Replication data Liquidity | Mean of agents participation ---|--- 5 | 24.49 10 | 26.15 20 | 29.13 50 | 35.43 75 | 37.48 100 | 38.88 150 | 40.57 200 | 41.51 300 | 42.45 | Table 6: Mean of Agent participation with different values of Initial Cash for Replication data Initial Cash | Mean of agents participation ---|--- 1 | 35.39 2 | 35.87 3 | 35.22 4 | 34.96 5 | 34.79 10 | 34.90 20 | 34.73 | Table 7: Mean of Agents participation with different values of $\lambda$ for Replication data $\lambda$ | Mean of agents participation ---|--- 0.01 | 35.08 0.025 | 34.93 0.05 | 35.06 0.1 | 35.13 0.25 | 35.18 0.5 | 35.06 1 | 35.41 | Table 8: First five best and worst Hyper-parameters for Replication Data Liquidity | $\lambda$ | Initial Cash | F1 | Accuracy | Scored% ---|---|---|---|---|--- 5 | 0.05 | 1 | 0.84 | 0.79 | 36 10 | 0.05 | 10 | 0.84 | 0.76 | 35 5 | 1 | 4 | 0.83 | 0.75 | 37 5 | 0.1 | 1 | 0.83 | 0.77 | 36 5 | 0.1 | 2 | 0.83 | 0.75 | 37 150 | 0.25 | 4 | 0.69 | 0.58 | 35 100 | 0.05 | 2 | 0.68 | 0.0.57 | 35 150 | 0.5 | 20 | 0.68 | 0.58 | 35 10 | 0.5 | 3 | 0.65 | 0.66 | 55 75 | 0.1 | 2 | 0.64 | 0.65 | 52 | | | | | Table 9: F1 score for different Exogenous Agents No-Ex | GT0.1% | GT0.5% | GT1% | GTinv0.1% | GTinv0.5% | GTinv1% | Random1% | Random5% | Random10% | Random50% ---|---|---|---|---|---|---|---|---|---|--- 0.84 | 0.93 | 1 | 1 | 0.34 | 0.23 | 0.09 | 0.79 | 0.81 | 0.80 | 0.82 0.84 | 0.91 | 0.96 | 0.97 | 0.34 | 0.31 | 0.28 | 0.74 | 0.76 | 0.75 | 0.75 0.83 | 0.94 | 1 | 1 | 0.32 | 0.25 | 0.06 | 0.79 | 0.81 | 0.83 | 0.78 0.83 | 0.90 | 0.95 | 0.99 | 0.33 | 0.28 | 0.24 | 0.76 | 0.74 | 0.82 | 0.77 0.83 | 0.88 | 0.91 | 0.94 | 0.33 | 0.32 | 0.29 | 0.75 | 0.77 | 0.79 | 0.73 0.69 | 0.89 | 0.94 | 0.96 | 0.34 | 0.31 | 0.28 | 0.76 | 0.77 | 0.81 | 0.77 0.69 | 0.91 | 0.97 | 0.96 | 0.34 | 0.29 | 0.29 | 0.78 | 0.77 | 0.77 | 0.78 0.68 | 0.90 | 0.95 | 0.96 | 0.33 | 0.29 | 0.29 | 0.78 | 0.78 | 0.79 | 0.76 0.66 | 0.89 | 0.90 | 0.94 | 0.33 | 0.33 | 0.29 | 0.76 | 0.76 | 0.79 | 0.74 0.65 | 0.90 | 0.91 | 0.94 | 0.34 | 0.32 | 0.30 | 0.77 | 0.75 | 0.81 | 0.74 | | | | | | | | | | Table 10: Accuracy score for different Exogenous Agents No-Ex | GT0.1% | GT0.5% | GT1% | GTinv0.1% | GTinv0.5% | GTinv1% | Random1% | Random5% | Random10% | Random50% ---|---|---|---|---|---|---|---|---|---|--- 0.79 | 0.92 | 1 | 1 | 0.27 | 0.14 | 0.05 | 0.76 | 0.77 | 0.77 | 0.78 0.76 | 0.90 | 0.95 | 0.97 | 0.27 | 0.21 | 0.17 | 0.67 | 0.68 | 0.67 | 0.70 0.75 | 0.94 | 1 | 1 | 0.26 | 0.14 | 0.03 | 0.76 | 0.77 | 0.80 | 0.77 0.77 | 0.89 | 0.95 | 0.99 | 0.26 | 0.18 | 0.14 | 0.70 | 0.66 | 0.77 | 0.73 0.75 | 0.87 | 0.90 | 0.94 | 0.27 | 0.23 | 0.18 | 0.67 | 0.70 | 0.74 | 0.68 0.58 | 0.88 | 0.93 | 0.95 | 0.26 | 0.21 | 0.17 | 0.70 | 0.70 | 0.76 | 0.71 0.57 | 0.90 | 0.96 | 0.95 | 0.25 | 0.18 | 0.18 | 0.73 | 0.70 | 0.71 | 0.72 0.58 | 0.89 | 0.94 | 0.95 | 0.25 | 0.19 | 0.17 | 0.73 | 0.71 | 0.73 | 0.74 0.65 | 0.88 | 0.89 | 0.93 | 0.25 | 0.26 | 0.18 | 0.71 | 0.68 | 0.74 | 0.69 0.64 | 0.89 | 0.90 | 0.94 | 0.27 | 0.23 | 0.19 | 0.72 | 0.68 | 0.76 | 0.70 | | | | | | | | | | Figure 5: Average F1 with different values of Lambda, liquidity and initial cash for Replication Data Figure 6: Density plot for replication data across Accuracy and F1 score Figure 7: Average accuracy vs average F1 with different values of Liquidity for Replication Data Figure 8: Average accuracy vs average F1 with different values of Initial Cash for Replication Data Figure 9: F1 with different percentage of added GT inverse agents for replication data Figure 10: Phase 1 Architecture Figure 11: Phase 2 Architecture Figure 12: F1 with different percentage of added Random agents for replication data Figure 13: F1 with different percentage of added GT agents for replication data Table 11: Agent participation with different percentages of GT agents No-Ex | GT0.1% | GT0.5% | GT1% ---|---|---|--- 4112 | 4543 | 4671 | 5352 3280 | 3766 | 4266 | 5339 3431 | 3325 | 4057 | 5132 4142 | 4442 | 5244 | 5496 3581 | 3989 | 4638 | 5023 5963 | 6462 | 6934 | 7544 5802 | 5784 | 6356 | 7028 6076 | 6100 | 6790 | 7475 3906 | 4063 | 4286 | 5180 6448 | 5436 | 5873 | 6766 | | | The most impactful hyper-parameter we found after doing the complete study is the liquidity for all three datasets. Figure 7 highlights the average accuracy vs average F1 score with different values of liquidity for the replication data. For all these three datasets the observation is same and it is evident that the performance of the market doesn’t depend on the lambda. The best performance was observed with accuracy 79% and F1 score 0.84 when the initial cash was 1 and liquidity was 5. The total number of papers were 145 out of which 93 were not scored and 52 were scored. Figure 14: Agent Participation with different percentage of added GT agents for replication data Table 12: Agent participation with different percentages of GTinv agents No-Ex | GTinv0.1% | GTinv0.5% | GTinv1% ---|---|---|--- 4112 | 4284 | 5374 | 6785 3280 | 3636 | 4512 | 5484 3431 | 3928 | 4664 | 5299 4142 | 4085 | 5169 | 7010 3581 | 4067 | 4883 | 6409 5963 | 6207 | 6414 | 7803 5802 | 5985 | 6174 | 7448 6076 | 6218 | 6529 | 7430 3906 | 4045 | 5042 | 6048 6448 | 5607 | 6120 | 7228 | | | Figure 15: Agent Participation with different percentage of added GT inverse agents for replication data Another very interesting observation is the mean of agent participation with different values of liquidity which is mentioned in Table 5. The agent participation clearly increases with increase in the value of liquidity. Liquidity constant directly impacts the shifts in the price.The agents tend to participate more as the price shifts tends to increase. Hence, this behavior is inline with our expectations. We captured similar data for different values of lambda and Initial cash as shown in Table 7 and Table 6 respectively. The mean of the agent participation tends to show no significant variation with change in lambda and initial cash and hence this supports our previous observation that these parameters don’t produce a consistent decipherable effect on the performance of the market. At the end of our experiments, with the hyper-parameter combinations we understood that there is no fixed best hyper-parameter set for all kinds of datasets. Every researcher needs to experiment and set the best hyper- parameter combination for their application to get the best performance from the prediction market. For the next set of experiments, we decided to take the best and worst 5 combinations of hyper-parameters for the replication dataset which is listed in Table 8. The prediction market classifies a subset of the input papers with good accuracy and F1 score. The last column shows the total percentage of papers scored for each of these combinations. The highest number of scored papers is seen with a (liquidity, $\lambda$, initial cash) combination of (10, 0.5, 3), when the market was able to get 55% scored papers with 65% accuracy and with an F1 of 0.66. #### 6.3.1 Discussion RQ1 Bring together. ### 6.4 Analysis using Exogenous Agents In the second phase of the study, we have introduced simulated exogenous agents to analyse their effect on the market performance. For this purpose, we chose to inject incremental percentages of simulated agents to track the shifts in accuracy, F1-score and eventually the overall agent participation. Our agents are divided into three categories - Ground truth (GT), Ground Truth Inverse(GT Inv) and Random agents respectively. The accuracy and F1-score for the top 5 best and worst combinations after the introduction of these agents has been listed down in Table 10 and Table 9 respectively. #### 6.4.1 Introduction of Ground Truth Agents in the Agent Pool Agents with Ground Truth information have the potential to causes sudden changes on how the market behaves. Hence, we selected very small percentages of such agents in our tests. $\%\>of\>GT\>agents=\\{0.1,0.5,1\\}$ In the set above, 0.1% implies that we would inject 1 GT agent for every 1000 agents currently in the pool. From Table 9 and Table 10, it is evident that even a small amount of GT agents can improve the market performance substantially. Figure 13 clearly charts down this incremental improvement. On injecting 0.1% of GT agents, we see a spike in F1 score from 0.84 to 0.93 and the accuracy shoots from 79% to 92% as shown in the tables above. Further gradual addition of such agents ensures 100% accuracy and F1 for the best combinations. The impact of such agents becomes more evident in the cases where the previously worst performing set of hyper-parameters now showed an improvement in accuracy from 64% to 89% with just 0.1% of GT agents. #### 6.4.2 Introduction of Ground Truth Inverse Agents in the Agent Pool Similar to GT agents, GTinv agents can intuitively cause sudden drops in performance. These drops were clearly seen in our tests and can be tracked from Table 9 and Table 10 respectively. Just like GT Agents, we chose the following set of incremental percentages for GTinv agents since they drive the market in a defined direction. $\%\>of\>GTInv\>Agents=\\{0.1,0.5,1\\}$ Figure 9 plots the F1 score for different percentage of GTinv agents. The black bar represents the case where there are no GTinv agents in the market. It is clearly visible from the bar plot that market performance deteriorates with the increase in amount of GTinv agents. Addition of just 0.1% of GTinv agents dropped the accuracy from 79% to 27%. The accuracy further drops down to 14% after adding 0.5% of agents and any further addition pushes the accuracy down to zero as per our tests. Similar behavior is seen in the F1 scores as well. The best F1 score is 0.84 prior to the introduction of GTinv agents but it drops down to 0.34 with just 0.1% of such agents. #### 6.4.3 Random Agents We further created a set of simulated agents which weren’t trained on the default feature set to simulate real-time human experiments. We found that it’s infeasible to assess a pattern in the market performance owing to the randomness introduced by these agents. Figure 12 shows how the F1 score changes when these Random agents are injected to the training pool. We have used a different set of percentages for these agents as shown below. $\%\>of\>Random\>Agents=\\{1,5,10,50\\}$ As expected, the F1 scores don’t showcase a clear shift in any direction as can be seen from Figure 12, but these agents have the potential to cause an overall change in the market performance. The impact of these agents varies across best to worst choices of markets. Adding 10% Random agents drops the F1-score from 0.84 to 0.80 for the best case but it also improves the score from 0.65 to 0.81 for the previously worst performing combination. Similar behavior is seen if we add 5% random agents as it changes F1-score from 0.84 to 0.81 for the best case but it improves from 0.65 to 0.75 for the worst case scenario. On increasing the percentage to 50%, we see a further dip in performance compared to the case with just 10% random agents But if you increase it to 50% then the overall performance will go down from the results obtained for 10% random agents. #### 6.4.4 Agent Participation varies with added Exogenous agents The addition of these agents provides an opportunity to derive various intuitions. One such analysis was how the agent participation changes with added GT and GTinv agents. Agent participation increases with the increase in percentage of these agents in the market as they have the potential to drive other agents by bidding in defined directions. Table 11 and Table 12 shows total agent participation for different percentages of GT agents and GTinv agents respectively. From Figure 14, we can observe that participation was highest for 1% of newly added GT agents and lowest for no exogenous agents for all 10 combinations. If we take the first combination for example, when there are no exogenous agents the participation was 4112, for 0.1% GT agents participation increases to 4543 and then it became 5352 for 1% added GT agents during market run. Similarly, Figure 15 shows total agent participation for different percentage of added GTinv agents. It can be seen that incremental addition of GTinv agents drives increased agent participation for all 10 combinations. For example, if we chose the last set of hyper-parameter combination the agent participation was 6448 for no exogenous agents but it increased to 7228 on adding 1% of GTinv agents. #### 6.4.5 Discussion RQ2 ## 7 Expected Contributions This study provides the research community with a new approach to the prediction market that has not been explored yet. The analysis of the simulated humans with trained AI agents and how they behave together in the field of research reproducibility is new. We have tried to analyze the scenario when there will be real human participants in the prediction market and will buy/sell alongside the trained AI agents and how their participation will affect the market performance as well as the agent participation. The comparative study of the artificial prediction market and the market with exogenous agents (also known as simulated humans) point out the need for human-AI collaboration in the prediction market which should motivate future researchers to explore more in the field of Human-AI study. The overall study with the hyper-parameters will be very helpful for future research because it will save time and effort. They don’t have to run multiple experiments to understand the behavior of these parameters. They will have a clear idea about the experiments needed for their dataset to choose the best set. The proposed research could have an impact on research society in broad ways because the prediction market has been massively used in almost every research domain. Scientific research reproducibility is one of the most sensitive areas for all researchers. The proposed research will produce a clear vision of the future hybrid prediction market. A lot of money and time will be saved. Previously it was observed that there was a lot of loss in the market due to replication failures. People can use that money and time for other productive research which can be used in different ways for the betterment of society. ## 8 Conclusions Previous research suggested that artificial prediction markets work wonders in various applications. In many cases, they perform better than human predictions. But some of the applications are very complex and they need human intelligence for better understanding. In these situations, the combined prediction of humans and AI agents can work wonders as per our hypothesis. Therefore, we have experimented with simulated humans in the artificial prediction market to explore how the participation of different exogenous agents will affect the market performance. The artificial prediction market with random agents, GT agents, and GT inverse agents was analyzed to understand the bounds of the behavior of the market in a hybrid setting. After adding a different percentage of exogenous agents, the performance of both markets (artificial prediction market and market with simulated humans) was compared using different evaluation matrices. Trained market agents will not participate if they don’t have enough similar training features. In those cases, human participation can drive agent participation in the market. The final hope of this research study is that it will encourage other researchers to explore more in Human-AI collaboration more. Also, we provided an important understanding of different hyper-parameters and how they differ for different datasets. Future researchers will have a complete understanding of how to use these parameters to improve the performance of the prediction market. For future work, we will introduce human participants in our prediction market with AI agents to investigate the real scenario.
# RPC based tracking system at CERN GIF++ facility K. Mota Amarilo<EMAIL_ADDRESS>A. Samalan M. Tytgat M. El Sawy G.A. Alves F. Marujo E.A. Coelho E.M. Da Costa H. Nogima A. Santoro S. Fonseca De Souza D. De Jesus Damiao M. Thiel M. Barroso Ferreira Filho A. Aleksandrov R. Hadjiiska P. Iaydjiev M. Rodozov M. Shopova G. Soultanov A. Dimitrov L. Litov B. Pavlov P. Petkov A. Petrov E. Shumka S.J. Qian H. Kou Z.-A. Liu J. Zhao J. Song Q. Hou W. Diao P. Cao C. Avila D. Barbosa A. Cabrera A. Florez J. Fraga J. Reyes Y. Assran M.A. Mahmoud Y. Mohammed I. Crotty I. Laktineh G. Grenier M. Gouzevitch L. Mirabito K. Shchablo I. Bagaturia I. Lomidze Z. Tsamalaidze V. Amoozegar B. Boghrati M. Ebraimi M. Mohammadi Najafabadi E. Zareian M. Abbrescia G. Iaselli G. Pugliese F. Loddo N. De Filippis R. Aly D. Ramos W. Elmetenawee S. Leszki I. Margjeka D. Paesani L. Benussi S. Bianco D. Piccolo S. Meola S. Buontempo F. Carnevali L. Lista P. Paolucci F. Fienga A. Braghieri P. Salvini P. Montagna C. Riccardi P. Vitulo E. Asilar J. Choi T.J. Kim S.Y. Choi B. Hong K.S. Lee H.Y. Oh J. Goh I. Yu C. Uribe Estrada I. Pedraza H. Castilla-Valdez A. Sanchez-Hernandez R. L. Fernandez M. Ramirez-Garcia E. Vazquez M. A. Shah N. Zaganidis A. Radi H. Hoorani S. Muhammad A. Ahmad I. Asghar W.A. Khan J. Eysermans F. Torres Da Silva De Araujo on behalf of the CMS Muon Group ###### Abstract With the HL-LHC upgrade of the LHC machine, an increase of the instantaneous luminosity by a factor of five is expected and the current detection systems need to be validated for such working conditions to ensure stable data taking. At the CERN Gamma Irradiation Facility (GIF++) many muon detectors undergo such studies, but the high gamma background can pose a challenge to the muon trigger system which is exposed to many fake hits from the gamma background. A tracking system using RPCs is implemented to clean the fake hits, taking profit of the high muon efficiency of these chambers. This work will present the tracking system configuration, used detector analysis algorithm and results. ###### keywords: Resistive plate chambers , CMS Experiment , gaseous detectors , HL-LHC ###### PACS: 0000 , 1111 ###### MSC: 0000 , 1111 ††journal: Nuclear Instruments and Methods in Physics Research A [a]Ghent University, Dept. of Physics and Astronomy, Proeftuinstraat 86, B-9000 Ghent, Belgium. [aa]Université Libre de Bruxelles, Avenue Franklin Roosevelt 50-1050 Bruxelles, Belgium. [b]Centro Brasileiro Pesquisas Fisicas, R. Dr. Xavier Sigaud, 150 - Urca, Rio de Janeiro - RJ, 22290-180, Brazil. [c]Dep. de Fisica Nuclear e Altas Energias, Instituto de Fisica, Universidade do Estado do Rio de Janeiro, Rua Sao Francisco Xavier, 524, BR - Rio de Janeiro 20559-900, RJ, Brazil. [d]Bulgarian Academy of Sciences, Inst. for Nucl. Res. and Nucl. Energy, Tzarigradsko shaussee Boulevard 72, BG-1784 Sofia, Bulgaria. [e]Faculty of Physics, University of Sofia,5 James Bourchier Boulevard, BG-1164 Sofia, Bulgaria. [f]School of Physics, Peking University, Beijing 100871, China. [g]State Key Laboratory of Particle Detection and Electronics, Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China. [gg]University of Chinese Academy of Sciences, No.19(A) Yuquan Road, Shijingshan District, Beijing 100049, China. [h]Universidad de Los Andes, Carrera 1, no. 18A - 12, Bogotá, Colombia [i]Egyptian Network for High Energy Physics, Academy of Scientific Research and Technology, 101 Kasr El-Einy St. Cairo Egypt. [ii]Suez University, Elsalam City, Suez - Cairo Road, Suez 43522, Egypt. [j]Center for High Energy Physics(CHEP-FU), Faculty of Science, Fayoum University, 63514 El-Fayoum, Egypt. [jj]Department of Physics, Faculty of Science, Ain Shams University, Cairo, Egypt. [jjj] Physics Department Faculty of Science Helwan University Ain Helwan 11795 Cairo Egypt. [k]Univ Lyon, Univ Claude Bernard Lyon 1, CNRS-IN2P3, IP2I Lyon, UMR 5822,F-69622, Villeurbanne, France. [l]Georgian Technical University, 77 Kostava Str., Tbilisi 0175, Georgia. [m]School of Particles and Accelerators, Institute for Research in Fundamental Sciences (IPM), P.O. Box 19395-5531, Tehran, Iran. [mm]School of Engineering, Damghan University, Damghan, 3671641167, Iran. [n]INFN, Sezione di Bari, Via Orabona 4, IT-70126 Bari, Italy. [o]INFN, Laboratori Nazionali di Frascati (LNF), Via Enrico Fermi 40, IT-00044 Frascati, Italy. [p]INFN, Sezione di Napoli, Complesso Univ. Monte S. Angelo, Via Cintia, IT-80126 Napoli, Italy. [pp]Dipartimento di Ingegneria Elettrica e delle Tecnologie dell’Informazione - Università Degli Studi di Napoli Federico II, IT-80126 Napoli, Italy. [q]INFN, Sezione di Pavia, Via Bassi 6, IT-Pavia, Italy. [qq]INFN, Sezione di Pavia and University of Pavia, Via Bassi 6, IT-Pavia, Italy. [r]Hanyang University, 222 Wangsimni-ro, Sageun- dong, Seongdong-gu, Seoul, Republic of Korea. [s]Korea University, Department of Physics, 145 Anam-ro, Seongbuk-gu, Seoul 02841, Republic of Korea. [t]Kyung Hee University, 26 Kyungheedae-ro, Dongdaemun-gu, Seoul 02447, Republic of Korea. [u]Sungkyunkwan University, 2066 Seobu-ro, Jangan-gu, Suwon, Gyeonggi- do 16419, Republic of Korea. [v]Benemerita Universidad Autonoma de Puebla, Puebla, Mexico. [w]Cinvestav, Av. Instituto Politécnico Nacional No. 2508, Colonia San Pedro Zacatenco, CP 07360, Ciudad de Mexico D.F., Mexico. [x]Universidad Iberoamericana, Mexico City, Mexico. [y]Sultan Qaboos University, Al Khoudh,Muscat 123, Oman. [z]National Centre for Physics, Quaid- i-Azam University, Islamabad, Pakistan. [za]Massachusetts Institute of Technology, 77 Massachusetts Ave, Cambridge, MA 02139, United States. [zc]III. Physikalisches Institut (A), RWTH Aachen University, Sommerfeldstrasse D-52056, Aachen, Germany. ## 1 Introduction The Resistive Plate Chambers (RPC) act as muon triggering detector in the Muon System of the Compact Muon Solenoid (CMS) experiment [1]. With the coming upgrade of the Large Hadron Collider (LHC) machine, the High Luminosity LHC (HL-LHC) [2], many detector systems are preparing to improve their capabilities. In particular, the CMS RPC project has two important planned activities – the replacement of the off-detector electronics (called link system) and the extension of the RPC coverage from $|\eta|=$ 1.9 to 2.4 with the new improved RPC (iRPC) detectors [3]. To study the performance and stability of RPCs in the HL-LHC conditions, tests are taking place in the CERN Gamma Irradiation Facility (GIF++), where a high energy particle beam (normally muons) and photons from a gamma source (13 TBq, 137Cs) are combined [4]. The maximum expected background rate for the RPCs in the current CMS Muon system is 600 Hz/cm2 and for the new iRPCs it goes up to 2 kHz/cm2 (including the safety factor of three) [3]. Such high rates can pose a challenge to the measurement, as the fake hits can bias the data-taking. All results shown in this work were recorded during GIF++ test beam in October 2021. ## 2 Experimental Setup Figure 1 shows a representation of the setup in GIF++. There are two trolleys where the RPC detectors are placed for irradiation. In addition to the RPCs, two scintillators are placed on each trolley to trigger on the muon beam. The data acquisition is performed using a CAEN Time-to-Digital Converter (TDC) module of type V1190A where the frontend electronics of the chambers are connected. A V1718 VME controller module is responsible for the communication between the TDC and the computer where the data are stored. To host the VME controller and the TDC a 6U VME 6021 crate is used. Figure 1: Experimental setup at GIF++. In the trolley farther from the gamma source, there are two tracking RPC chambers with their strip planes oriented perpendicular to each other to allow measurement in two directions. The chambers used are the following: * 1. Tracking chambers: Referred to as GT1 and GT2, these are double gap chambers made of 2 mm thick High Pressure Laminate (HPL, also known as bakelite). They are equipped with a strip plane with 32 strips of 1.45 cm pitch. The signals are read with the CMS FrontEnd electronics with threshold set to 150 fC. * 2. Test Chamber: Referred to as KODEL-C, this is also a double gap chamber made of HPL with 1.4 mm thickness. It is equipped with a strip plane with 32 strips of 1.95 cm pitch and custom FrontEnd electronics with threshold set to 75 fC. All the chambers are flushed with the CMS RPC standard gas mixture (95.2% freon ($C_{2}H_{2}F_{4}$), 4.5% isobutane ($iC_{4}H_{10}$), and 0.3% sulphur hexafluoride ($SF_{6}$)) with a relative humidity of $\approx$40%. The tracking chambers are always powered on with the high voltage (HV) working point (WP) applied to their electrodes. ## 3 Tracking algorithm To enable the tracking analysis, at least one hit is required in both tracking chambers inside the muon time window defined by a Gaussian fit, as indicated in Fig. 2. A 2D hit profile is made to check their alignment with the muon beam, as shown in Fig. 3. The hit profile is also determined for the test chamber and it is used for alignment between the chambers. Figure 2: Hits time profile of the tracking chamber GT2. The time window where the hits are accepted in the analysis is represented in the red hatched box. Figure 3: 2D hit profile of the strips signals for the hits in the muon time window. The tracking algorithm relies on the assumption that the beam is perpendicular to the strip plane, therefore it is only needed to extrapolate the position of the hit from the tracking to the test chamber and check for a matching hit. For every event, the following steps are taken: 1. 1. Perform the clusterization of the hits in the tracking chambers, where events with more than one cluster are rejected. The cluster barycentres are defined as the mean position in the cluster; 2. 2. Perform the clusterization for the test chambers and calculate the clusters’ barycentres; 3. 3. Form a perpendicular track starting from the tracking cluster barycentre; 4. 4. Check for a match in any cluster on the test chamber. Figure 4 shows an example of event where the analysis was performed and a matching hit was found. Figure 4: Example of one event in which the hit on tracking chamber was matched on the test chamber. The strips in red represent the hits and the point in black is the cluster barycentre. ## 4 Performance of the Tracking Algorithm Efficiency curves of the test chambers were used to evaluate the validity of the tracking in rejecting the fake hits caused by the gammas. On Fig. 5 the efficiency curves are compared for three gamma background conditions. There are three curves in each plot. The black curve is calculated without the tracking, taking into consideration all the hits in the muon time window. For the blue one, only the hits that passed the tracking criteria were considered. Finally, for the red curve, the HV was recalculated to remove the voltage drop caused by the resistance of the electrodes - this correction decouples the shift to the right of the curve on higher cluster rates, caused by the increase of the current. Therefore, HVgas is the effective HV applied to the gas volume. As expected, in Fig. 5(a), where the data were taken in a run with source OFF, the three curves are equivalent since the cluster rate (CLR) and the currents are low. in Fig. 5(b) with CLR $\approx$ 648 Hz/cm2, a shift to the right on the tracking corrected curve is observed, and it is much more evident in Fig. 5(c) with CLR $\approx$ 1.8 kHz/cm2. The increase of the maximum efficiency on the raw efficiency curve from the Fig. 5(b) to 5(c) indicates high gamma contamination. (a) (b) (c) Figure 5: Efficiencies and their sigmoid fits measured at three photon fluxes with measured at the working-point high voltages gamma cluster rates of 1.804 kHz (a), 0.645 kHz (b), and 1.48 Hz (c), respectively. The curve in black is the efficiency calculated with all hits inside the muon window, the blue one is with applied tracking correction and the one in red – with tracking and resistance correction. In Fig. 6(a), it is possible to see the curves with tracking and resistance correction. The shift to the right is caused by the reduction of the gas gain at higher rates, while the drop in the maximum efficiency is caused by the dead-time of the electronics. In Fig. 6(b), no resistance correction is applied, so the main reason for the shift is the voltage drop over the electrode’s resistance, coupled with the reasons mentioned before. (a) (b) Figure 6: Efficiency curves for various gamma background rates. In (a) the HV applied is corrected only by pressure and temperature (HVeff), while in (b) the HV is also corrected by the resistance of the gaps (HVgas) . ## 5 Conclusion The implementation of the tracking system was motivated by the necessity to test the CMS RPC chambers at the conditions of the HL-LHC. The results show that the tracking system performs very well to remove fake hits from the gamma background, even at rates as high as 2 kHz/cm2. The test chamber performed very well and showed increase on the working point of $\approx$650 V with efficiency loss of $\approx$ 7.5%, using the custom electronics with threshold of 75 fC. This system is currently being used for the ageing studies of the CMS RPC system. ## Acknowledgements We would like to thank our colleagues from CERN Gamma Irradiation facility. We would also like to acknowledge the enduring support for the Upgrade of the CMS detector and the supporting computing infrastructure provided by the following funding agencies: FWO (Belgium); CNPq, CAPES and FAPERJ (Brazil); MES and BNSF (Bulgaria); CERN; CAS, MoST, and NSFC (China); MINCIENCIAS (Colombia); CEA and CNRS/IN2P3 (France); SRNSFG (Georgia); DAE and DST (India); IPM (Iran); INFN (Italy); MSIP and NRF (Republic of Korea); BUAP, CINVESTAV, CONACYT, LNS, SEP, and UASLP-FAI (Mexico); PAEC (Pakistan); DOE and NSF (USA). ## References * [1] CMS Collaboration, The CMS Experiment at the CERN LHC, JINST 3 (2008) S08004. doi:10.1088/1748-0221/3/08/S08004. * [2] O. Aberle, et al., High-Luminosity Large Hadron Collider (HL-LHC): Technical design report, CERN Yellow Reports: Monographs, CERN, Geneva, 2020. doi:10.23731/CYRM-2020-0010. URL https://cds.cern.ch/record/2749422 * [3] CMS Collaboration, The Phase-2 Upgrade of the CMS Muon Detectors, Tech. rep., CERN, Geneva (2017). URL https://cds.cern.ch/record/2283189 * [4] R. Guida, GIF++: The new CERN irradiation facility to test large-area detectors for HL-LHC, in: 2015 IEEE Nuclear Science Symposium and Medical Imaging Conference (NSS/MIC), IEEE, 2015, pp. 1–4.
# On the number of subrings of $\mathbb{Z}^{n}$ of prime power index Hrishabh Mishra Chennai Mathematical Institute, H1, SIPCOT IT Park, Kelambakkam, Siruseri, Tamil Nadu 603103, India<EMAIL_ADDRESS>and Anwesh Ray Centre de recherches mathématiques, Université de Montréal, Pavillon André-Aisenstadt, 2920 Chemin de la tour, Montréal (Québec) H3T 1J4, Canada<EMAIL_ADDRESS> ###### Abstract. Let $n$ and $k$ be positive integers, and $f_{n}(k)$ (resp. $g_{n}(k)$) be the number of unital subrings (resp. unital irreducible subrings) of $\mathbb{Z}^{n}$ of index $k$. The numbers $f_{n}(k)$ are coefficients of certain zeta functions of natural interest. The function $k\mapsto f_{n}(k)$ is multiplicative, and the study of the numbers $f_{n}(k)$ reduces to computing the values at prime powers $k=p^{e}$. Given a composition $\alpha=(\alpha_{1},\dots,\alpha_{n-1})$ of $e$ into $n-1$ positive integers, let $g_{\alpha}(p)$ denote the number of irreducible subrings of $\mathbb{Z}^{n}$ for which the associated upper triangular matrix in Hermite normal form has diagonal $(p^{\alpha_{1}},\dots,p^{\alpha_{n-1}},1)$. Via combinatorial analysis, the computation of $f_{n}(p^{e})$ reduces to the computation of $g_{\alpha}(p)$ for all compositions of $i$ into $j$ parts, where $i\leq e$ and $j\leq n-1$. We extend results of Liu and Atanasov-Kaplan- Krakoff-Menzel, who explicitly compute $f_{n}(p^{e})$ for $e\leq 8$. The case $e=9$ proves to be significantly more involved. We evaluate $f_{n}(e^{9})$ explicitly in terms of a polynomial in n and p up to a single term which is conjecturally a polynomial. Our results provide further evidence for a conjecture, which states that for any fixed pair $(n,e)$, the function $p\mapsto f_{n}(p^{e})$ is a polynomial in $p$. A conjecture of Bhargava on the asymptotics for $f_{n}(k)$ as a function of $k$ motivates the study of the asymptotics for $g_{\alpha}(p)$ for certain infinite families of compositions $\alpha$, for which we are able to obtain general estimates using techniques from the geometry of numbers. ###### Key words and phrases: counting subrings of prescribed index, combinatorics of subrings of a fixed ring, zeta functions of groups and rings ###### 2020 Mathematics Subject Classification: 20E07, 11H06, 11A25, 11M41 ## 1\. Introduction Let $G$ be an infinite group. Given a finite index subgroup $H$, we set $[G\mathrel{\mathop{\mathchar 58\relax}}H]$ to denote the index of $H$ in $G$. For $k\in\mathbb{Z}_{\geq 1}$, let $a_{k}(G)$ be the number of finite-index subgroups $H$ of $G$ such that $[G\mathrel{\mathop{\mathchar 58\relax}}H]=k$. In [GSS88], Grunewald, Segal and Smith introduced the _zeta function_ of $G$, defined as follows $\zeta_{G}(s)\mathrel{\mathop{\mathchar 58\relax}}=\sum_{H}[G\mathrel{\mathop{\mathchar 58\relax}}H]^{-s}=\sum_{k=1}^{\infty}a_{k}(G)k^{-s},$ where the above sum runs over all finite index subgroups $H$ of $G$. The properties of such zeta functions are described in [DSWP08]. In this context, zeta functions measure _subgroup growth_. For a comprehensive account of this theme, we refer to [LS03]. Throughout, $n>1$ is an integer and $\mathbb{Z}^{n}$ is the ring consisting of integer $n$-tuples with componentwise addition and multiplication. It is well known that the zeta function $\zeta_{\mathbb{Z}^{n}}(s)$ is given by $\zeta_{\mathbb{Z}^{n}}(s)=\zeta(s)\zeta(s-1)\dots\zeta(s-(n-1))$ (cf. _loc. cit._ for five different proofs of the above result). We study distribution questions for the number of finite-index unital subrings of $\mathbb{Z}^{n}$ of prescribed index. This problem is closely related to the distributions of orders in a fixed number ring $\mathcal{O}_{K}$, a problem studied by Brackenhoff [Bra09], Kaplan Marcinek and Takloo-Bighash (cf. [KMTB15]). Given $k\in\mathbb{Z}_{\geq 1}$, following [Liu07], let $f_{n}(k)$ be the number of commutative unital subrings $S$ of $\mathbb{Z}^{n}$ such that $[\mathbb{Z}^{n}\mathrel{\mathop{\mathchar 58\relax}}S]=k$. The _subring zeta function_ is given by $\zeta_{\mathbb{Z}^{n}}^{R}(s)=\sum_{k=1}^{\infty}f_{n}(k)k^{-s},$ which decomposes into an Euler product $\zeta_{\mathbb{Z}^{n}}^{R}(s)=\prod_{p}\zeta_{\mathbb{Z}^{n},p}^{R}(s),$ where the product is over all primes $p$. The local Euler factor at $p$ is given by $\zeta_{\mathbb{Z}^{n},p}^{R}(s)\mathrel{\mathop{\mathchar 58\relax}}=\sum_{e=0}^{\infty}f_{n}(p^{e})p^{-es}.$ It is not hard to show that $\zeta_{\mathbb{Z}^{2}}^{R}(s)=\zeta(s)$, and that $\zeta_{\mathbb{Z}^{3}}^{R}(s)=\frac{\zeta(3s-1)\zeta(s)^{3}}{\zeta(2s)^{2}},$ cf. [Liu07, Proposition 4.1]. The formula for $n=3$ was first obtained by Datskovsky and Wright [DW88]. A formula for $n=4$ was obtained by Nakagawa cf. [Nak96]. Recently, there has been interest in the following question (cf. [Ish22b, Question 1.2]). ###### Question 1.1. Let $n,e$ be a pair of natural numbers. What can be said about $f_{n}(p^{e})$ as a function of $p$? For $e\leq 5$, Liu [Liu07] gives explicit formulae for $f_{n}(p^{e})$. Subsequently, these formulae were generalized by Atanasov, Kaplan, Krakoff and Menzel [AKKM21] for $e=6,7,8$. These computations indicate that for a fixed pair $(n,e)$, the function $p\mapsto f_{n}(p^{e})$ is a polynomial in $p$ (cf. [AKKM21, Question 1.13]), and this has been further studied by Isham in [Ish22a]. The zeta function $\zeta_{\mathbb{Z}^{n}}^{R}(s)$ is said to be _uniform_ if there is a rational function $W(X,Y)\in\mathbb{Q}(X,Y)$ such that for every prime $p$, $\zeta_{\mathbb{Z}^{n},p}^{R}(s)=W(p,p^{-s})$. It is not hard to show that the zeta function $\zeta_{\mathbb{Z}^{n}}(s)$ is uniform if for all $e$, $f_{n}(p^{e})$ is a polynomial in $p$. For further details, we refer to [AKKM21, section 5.1]. Below we summarize the known values of $f_{n}(p^{e})$ for $e\leq 5$. ###### Theorem 1.2 (Liu [Liu07]). With respect to notation above, $\begin{split}f_{n}(1)=&1,f_{n}(p)=\binom{n}{2},f_{n}(p^{2})=\binom{n}{2}+\binom{n}{3}+3\binom{n}{4},\\\ f_{n}(p^{3})=&\binom{n}{2}+(p+1)\binom{n}{3}+7\binom{n}{4}+10\binom{n}{5}+15\binom{n}{6},\\\ f_{n}(p^{4})=&\binom{n}{2}+(3p+1)\binom{n}{3}+(p^{2}+p+10)\binom{n}{4}+(10p+21)\binom{n}{5}+70\binom{n}{6}\\\ +&105\binom{n}{7}+105\binom{n}{8},\\\ f_{n}(p^{5})=&{n\choose 2}+(4p+1){n\choose 3}+(7p^{2}+p+13){n\choose 4}+(p^{3}+p^{2}+41p+31){n\choose 5}\\\ +&(15p^{2}+35p+141){n\choose 6}.\\\ \end{split}$ For the values for $e$ in the range $6\leq e\leq 8$, we refer to [KMTB15]. In this manuscript, we extend the above mentioned results to $e=9$. First, we introduce some further notation. A subring of prime power index decomposes into a direct product of irreducible subrings, which we now proceed to describe. ###### Definition 1.3. A subring $L$ of index $p^{e}$ is said to be _irreducible_ if for each vector $x=(x_{1},\dots,x_{n})^{t}\in L$, we have that $x_{1}\equiv x_{2}\equiv\dots\equiv x_{n}\left(\operatorname{mod}p\right).$ An irreducible subring matrix is a matrix $A$ in Hermite normal form (1.1) $A=\begin{pmatrix}p^{\alpha_{1}}&pa_{1,2}&pa_{1,3}&\cdots&pa_{1,n-1}&1\\\ &p^{\alpha_{2}}&pa_{2,3}&\cdots&pa_{2,n-1}&1\\\ &&p^{\alpha_{3}}&\cdots&pa_{3,n-1}&1\\\ &&&\ddots&\vdots&\vdots\\\ &&&&p^{\alpha_{n-1}}&1\\\ &&&&&1\end{pmatrix},$ with $\alpha_{i}\geq 1$ for all $i$. The subring $L$ of index $p^{e}$ is irreducible if and only if $A_{L}$ is an irreducible subring matrix. Let $g_{n}(p^{e})$ be the number of irreducible subrings of $\mathbb{Z}^{n}$ of index $p^{e}$. It is shown that every subring of $p$-power power index is a direct sum of irreducible subrings, and as a consequence, we have the following recursive formula (cf. Proposition 4.4 of [Liu07]) (1.2) $f_{n}(p^{e})=\sum_{i=0}^{e}\sum_{j=1}^{n}{n-1\choose j-1}f_{n-j}(p^{e-i})g_{j}(p^{i}).$ In order to compute $f_{n}(p^{9})$ it thus suffices to compute $g_{j}(p^{e})$ for all values $e\leq 9$, $1\leq j\leq n$. We remark that the convention for the definition of $g_{n}(p^{e})$ used here is that of [AKKM21], and differs from that in [Liu07]. A composition of $m$ into $r$ parts consists of a tuple $\alpha=(\alpha_{1},\dots,\alpha_{r})$ of positive integers such that $\sum_{i=1}^{r}\alpha_{i}=m$. Let $\mathcal{C}_{n,e}$ be set of all compositions of $e$ into $(n-1)$ parts. Given $\alpha\in\mathcal{C}_{n,e}$, let $g_{\alpha}(p)$ be the number of irreducible subring matrices (cf. Definition 2.3) of the form (1.1). We find that $g_{n}(p^{e})=\sum_{\alpha\in\mathcal{C}_{n,e}}g_{\alpha}(p)$, and thus to compute $g_{n}(p^{e})$, we need to compute $g_{\alpha}(p)$ for all $\alpha\in\mathcal{C}_{n,e}$. Therefore, the numbers $g_{\alpha}(p)$ can be thought of as the basic building blocks for computing the numbers $f_{n}(k)$. In practice, it is possible to compute $g_{\alpha}(p)$ in many cases via combinatorial case by case analysis. For larger values of $e$, the computation of $f_{n}(p^{e})$ becomes significantly more involved. The total number of new compositions that one must consider does grow exponentially, and even though one may compute the value of $g_{\alpha}(p)$ for various types of compositons that fit into a general framework, there are many exceptional compositions that do not fit into such a framework. Furthermore, the combinatorial (case by case) analysis does get increasing more challenging for exceptional compositions of longer length. With this in mind, we state the main result below. ###### Theorem 1.4. With respect to notation above, we have that $\displaystyle f_{n}(p^{9})-\gamma(n,p)={n\choose 2}+(p^{3}+4p^{2}+4p+1){n\choose 3}+(11p^{4}+30p^{3}+9p^{2}+p+25){n\choose 4}$ $\displaystyle+(11p^{6}+14p^{5}+137p^{4}-16p^{3}+81p^{2}+202p+73){n\choose 5}+(4p^{7}+76p^{6}+128p^{5}+71p^{4}$ $\displaystyle+464p^{3}+1088p^{2}+386p+571){n\choose 6}+(p^{9}+22p^{8}+23p^{7}+59p^{6}+31p^{5}+2032p^{4}$ $\displaystyle+2152p^{3}+2467p^{2}+4949p+2485){n\choose 7}+(p^{10}+p^{9}+2p^{8}+2p^{7}+507p^{6}+955p^{5}$ $\displaystyle+3293p^{4}+5980p^{3}+23410p^{2}+17011p+13707){n\choose 8}+(36p^{8}+37p^{7}+157p^{6}+1159p^{5}$ $\displaystyle+8545p^{4}+34997p^{3}+59371p^{2}+93649p+64019){n\choose 9}+(675p^{6}+795p^{5}+18960p^{4}$ $\displaystyle+48927p^{3}+250632p^{2}+330657p+297103){n\choose 10}+(990p^{5}+12540p^{4}+148830p^{3}$ $\displaystyle+497640p^{2}+1157145p+1245992){n\choose 11}+(13860p^{4}+97020p^{3}+1049895p^{2}+2961805p$ $\displaystyle+4727041){n\choose 12}+(135135p^{3}+1036035p^{2}+5900895p+15346045){n\choose 13}+(945945p^{2}$ $\displaystyle+7252245p+40500460){n\choose 14}+(4729725p+80615535){n\choose 15}+110810700{n\choose 16}$ $\displaystyle+91891800{n\choose 17}+34459425{n\choose 18},$ where $\gamma(n,p)\mathrel{\mathop{\mathchar 58\relax}}=\left(\sum_{k=2}^{n}{k-1\choose 5}\right)g_{(3,2,2,1,1)}(p)$. It proves to be difficult to compute the exact value of $g_{(3,2,2,1,1)}(p)$ since it reduces to the explicit computation of the number of solutions to the following system of polynomial equations over $\mathbb{F}_{p}=\mathbb{Z}/p\mathbb{Z}$ $\begin{split}&(x_{3}^{2}-x_{3})-x_{2}(x_{7}^{2}-x_{7})-x_{1}(x_{5}^{2}-x_{5})=0\\\ &(x_{4}^{2}-x_{4})-x_{2}(x_{8}^{2}-x_{8})-x_{1}(x_{6}^{2}-x_{6})=0\\\ &x_{3}x_{4}-x_{2}x_{7}x_{8}-x_{1}x_{5}x_{6}=0.\end{split}$ There are $3$ equations in $8$ variables. Computations done for primes $p\leq 19$ suggest that the number of solutions for the above equations is given by $N_{p}=p^{5}+12p^{4}-20p^{3}+30p^{2}-10p$. Our computations then suggest that $g_{(3,2,2,1,1)}(p)=p^{7}+24p^{6}-29p^{5}+21p^{4}-4p^{3}.$ We are unable to verify this claim for primes $p>19$, since it seems to be a difficult problem in general to compute the number of points on varieties over finite fields in a large number of variables. We note that in very specific cases, exact formulae are known. For instance, in the case of hypersurfaces (cf. [Dwo62]), or in the case when the system consists of diagonal equations (cf. [Spa79]) such questions are studied. One may however bound the size of $g_{(3,2,2,1,1)}(p)$, and it follows from Proposition 7.5 that $g_{(3,2,2,1,1)}(p)\geq p^{4}$. We describe the method of proof in more detail. It follows from results in [Liu07, AKKM21] that $g_{\alpha}(p)$ can be computed for compositions $\alpha$ of one of the following types * • $\alpha=(\beta,1,\dots,1)$ and $n\geq 2$, * • $\alpha=(2,1\dots,1,\beta,1,\dots,1)$ and $n\geq 3$. We refer to results in section 2 for further details. In section 3, we prove some general results which allow us to compute $g_{\alpha}(p)$, for compositions $\alpha$ of one of the following types * • $\alpha=(\beta,1,\dots,1,\gamma)$ is a composition of length $(n-1)\geq 3$ such that $\beta>2$ and $\gamma\geq\beta-1$, * • $\alpha$ is of the form $\alpha=(2,1,\dots,1,2,1,\dots,1,\beta)$, where $\beta>1$, * • $\alpha$ is of the form $\alpha=(2,1,\dots,1,2,1,\dots,1,\beta,1)$, where $\beta>1$, * • $\alpha$ is of the form $\alpha=(2,1,\dots,1,3,1,\dots,1,2)$. We refer to a composition that does not fit into any of the above families as an exceptional composition. In section 4 (resp. section 5), we compute $g_{\alpha}(p)$ for all relevant compositions beginning with $2$ (resp. $3$). In section 6, we compute $g_{\alpha}(p)$ for all relevant compositions beginning with $4,5$ or $6$. The proof of Theorem 1.4 is provided in section 7.1. All known computations indicate that for any composition $\alpha$, the function $g_{\alpha}(p)$ is a polynomial in $p$. The conditions for a matrix $A$ give rise to polynomial conditions on the entries $a_{i,j}$, and $g_{\alpha}(p)$ in many cases is the number of $\mathbb{F}_{p}$-points on a scheme defined by integral polynomial equations. It is certainly of interest to note that in all the examples considered, these schemes have polynomial point count in the sense of [HRV08, p.616, ll. -6 to -2]. We now come to describing the general asymptotic results proved in this manuscript. The analysis of the numbers $f_{n}(p^{e})$ can be translated into properties of the associated subring zeta functions. For instance, Isham in [Ish22b] proves lower bounds for $g_{n}(p^{e})$ and deduces that the subring zeta function $\zeta_{\mathbb{Z}^{n},p}^{R}(s)$ diverges for all $s$ such that $\operatorname{Re}s\leq c_{7}(n)$, where $c_{7}(n)\mathrel{\mathop{\mathchar 58\relax}}=\operatorname{max}_{0\leq d\leq n-1}\frac{d(n-1-d)}{(n-1+d)}$. This comes as a consequence of proving lower bounds for $f_{n}(p^{e})$. Some related results are also proven by Brackenhoff in [Bra09]. Liu attributes the following conjecture regarding the asymptotics for the numbers $f_{n}(k)$ to Bhargava (cf. [Liu07, p.298]). The conjecture was communicated to Liu via personal communication, cf. _loc. cit._ ###### Conjecture 1.5. For $n$ odd, $f_{n}(k)=O(k^{\frac{n-1}{6}+\epsilon})$, and for $n$ even, $f_{n}(k)=O(k^{\frac{n^{2}-2n}{6n-8}+\epsilon})$. In particular, for a fixed prime number $p$ and a fixed value of $n$, the conjecture predicts that $f_{n}(p^{e})=\begin{cases}&O(p^{\frac{e(n-1)}{6}})\text{ if }n\text{ is odd,}\\\ &O(p^{\frac{e(n^{2}-2n)}{6n-8}})\text{ if }n\text{ is odd.}\\\ \end{cases}$ Conjecture 1.5 motivates our results in section 7.2, where we investigate the asymptotics for $g_{\alpha}(p)$ for certain natural families of compositions $\alpha$. Let $n>1$ be a natural number and $t$ be in the range $1\leq t\leq n-1$. Let $\alpha=(\alpha_{1},\alpha_{2},\dots,\alpha_{n-1})$ be a composition of length $(n-1)$ and for $k\in\mathbb{Z}_{\geq 1}$, set $\alpha_{/k,t}\mathrel{\mathop{\mathchar 58\relax}}=(\alpha_{1},\alpha_{2},\dots k\alpha_{t}\dots,\alpha_{n-2},\alpha_{n-1})$, i.e., the composition obtained upon multiplying the $t$-th coordinate of $\alpha$ by $k$. Note that $g_{\alpha_{/k,t}}(p)$ contributes to $g_{n}(p^{e_{k}})$, where $e_{k}\mathrel{\mathop{\mathchar 58\relax}}=\sum_{i=1}^{n-1}\alpha_{i}+(k-1)\alpha_{t}$. Therefore, the asymptotic growth of $g_{\alpha_{/k,t}}(p)$ as a function of $k$ provides insight into that of $g_{n}(p^{e_{k}})$. The following result is proven via a combination of techniques from the geometry of numbers and the combinatorial analysis of large subring matrices. ###### Theorem 1.6. Let $n>1$ be a natural number and $\alpha=(\alpha_{1},\alpha_{2},\dots,\alpha_{n-1})$ a composition of length $(n-1)$ such that $\alpha_{i}>1$ are integers for each $i\in\\{1,2,\dots,n-1\\}$ and $1\leq t\leq n-1$. Then, $g_{\alpha_{/k,t}}(p)=O(p^{\gamma k(n-2)})\text{ as }k\to\infty,$ where $\gamma=\max\\{\alpha_{1},\dots,\alpha_{t}\\}$. The implied constant depends on $p$ and $n$, and not on $k$. Although the above result is weaker than the prediction of Conjecture 1.5, the authors expect that such arguments have the potential to lead to stronger results in the future. ### Acknowledgment The second author is supported by the CRM-Simons postdoctoral fellowship. ## 2\. Preliminaries Let $n\geq 1$ be an integer and $M_{n}(\mathbb{Z})$ denote the ring of $n\times n$-matrices with integer entries. By definition, a _lattice_ in $\mathbb{Z}^{n}$ is a subgroup $L$ of finite index in $\mathbb{Z}^{n}$. This index is denoted $[\mathbb{Z}^{n}\mathrel{\mathop{\mathchar 58\relax}}L]$. Set $u_{1},\dots,u_{n}$ to be the standard basis of $\mathbb{Z}^{n}$, with $u_{i}=(0,0,\dots,0,1,0,\dots)$, with $1$ in the $i$-th position and $0$s in all other positions. Let $v_{1},\dots,v_{n}$ be a basis of $L$, i.e., a set of vectors such that every element $v\in L$ is uniquely represented as an integral linear combination $v=\sum_{i=1}^{n}b_{i}v_{i}$. Expressing $v_{i}=\sum_{j=1}^{n}a_{i,j}u_{j}$, consider the integer matrix $A=\left(a_{i,j}\right)\in M_{n}(\mathbb{Z})$. We may choose $v_{1},\dots,v_{n}$ such that the associated matrix $A$ is in Hermite normal form, i.e., $A=\begin{pmatrix}a_{1,1}&a_{1,2}&a_{1,3}&\cdots&a_{1,n-1}&a_{1,n}\\\ &a_{2,2}&a_{2,3}&\cdots&a_{2,n-1}&a_{2,n}\\\ &&a_{3,3}&\cdots&a_{3,n-1}&a_{3,n}\\\ &&&\ddots&\vdots&\vdots\\\ &&&&a_{n-1,n-1}&a_{n-1,n}\\\ &&&&&a_{n,n}\end{pmatrix}$ with $0\leq a_{i,j}<a_{i,i}$ for all tuples $(i,j)$ such that $1\leq i<j\leq n$. The following result is used to reinterpret counting problems for subrings of $\mathbb{Z}^{n}$ in terms of intgral matrices satisfying precribed conditions. ###### Lemma 2.1. There is a bijection between lattices $L\subset\mathbb{Z}^{n}$ of index $k>0$ and integer $n\times n$ matrices $A$ in Hermite normal form with determinant $k$. ###### Proof. The reader is referred to the proof of [Liu07, Proposition 2.1]. ∎ Thus, to a lattice $L$ we associate the matrix $A_{L}$ in Hermite normal form, and to a matrix $A$ in Hermite normal form, we associate a unique lattice $L_{A}$. Given two integral vectors $u=(u_{1},\dots,u_{n})^{t}$ and $v=(v_{1},\dots,v_{n})^{t}$, denote the _composite_ by $u\circ v\mathrel{\mathop{\mathchar 58\relax}}=(u_{1}v_{1},\dots,u_{n}v_{n})^{t}$. A lattice $L$ is multiplicatively closed if $u\circ v\in L$ for all elements $u,v\in L$. A _subring_ of $\mathbb{Z}^{n}$ shall in this paper be taken to mean a multiplicatively closed lattice which contains the identity element $\mathbf{1}\mathrel{\mathop{\mathchar 58\relax}}=(1,1,\dots,1)^{t}$. In particular, the index of a subring is finite. Given a positive integer $k$, let $f_{n}(k)$ be the number of subrings of $\mathbb{Z}^{n}$ with index equal to $k$. Define the subring zeta function as follows $\zeta_{\mathbb{Z}^{n}}^{R}(s)\mathrel{\mathop{\mathchar 58\relax}}=\sum_{k=1}^{\infty}f_{n}(k)k^{-s}.$ The function $f_{n}(k)$ is multiplicative (cf. [Liu07, Proposition 2.7]), i.e., given two coprime integers $k_{1}>0$ and $k_{2}>0$, we have that $f_{n}(k_{1}k_{2})=f_{n}(k_{1})f_{n}(k_{2})$. The zeta function exhibits an Euler product $\zeta_{\mathbb{Z}^{n}}^{R}(s)=\prod_{p}\zeta_{\mathbb{Z}^{n},p}^{R}(s),$ where $\zeta_{\mathbb{Z}^{n},p}^{R}(s)=\sum_{e=0}^{\infty}f_{n}(p^{e})p^{-es}$. Thus, the study of the function $f_{n}(k)$ and the subring zeta function, comes down to the determination of $f_{n}(p^{e})$, where $n>0,e\geq 0$ and $p$ is a prime number. More precisely, given $n$ and $e$, we would like to determine $f_{n}(p^{e})$ as a function of $p$. We recall Liu’s bijection between subrings of $\mathbb{Z}^{n}$ and integral matrices in Hermite normal form. ###### Proposition 2.2. Let $n,k>0$ be integers. The association $L\mapsto A_{L}$ gives a bijection between subrings $L$ of $\mathbb{Z}^{n}$ of index $k$ and matrices $A\in\operatorname{M}_{n}(\mathbb{Z})$ in Hermite normal form with $\det(A)=k$ and columns $v_{1},\dots,v_{n}$ such that 1. (1) $\mathbf{1}$ is in the column span of $A$, 2. (2) for all $i,j$ in the range $1\leq i,j\leq n$, $v_{i}\circ v_{j}$ is in the column span of $A$. ###### Proof. The reader is referred to [Liu07, Proposition 2.1, 2.2]. ∎ ###### Definition 2.3. A matrix $A$ satisfying the conditions of Proposition 2.2 is referred to as a _subring_ matrix. The column span of $A$ is denoted $\operatorname{Col}(A)$. Let $\alpha=(\alpha_{1},\dots,\alpha_{n-1})$ be a composition of length $(n-1)$. Recall from the introduction that $g_{\alpha}(p)$ is the number of irreducible subring matrices with diagonal $(p^{\alpha_{1}},\dots,p^{\alpha_{n-1}},1)$. We recall some useful results from [AKKM21, Liu07] which shall allow us to deduce the values of $g_{\alpha}(p)$ for specific choices of $\alpha$. ###### Lemma 2.4. Let $n\geq 2$ and $\alpha=(\beta,1,\dots,1)$ be a composition of length $(n-1)$. Then, the following assertions hold. 1. (1) If $\beta=2$, then $g_{\alpha}(p)=p^{n-2}$. 2. (2) If $\beta\geq 3$, then $g_{\alpha}(p)=(n-1)p^{n-2}$. ###### Proof. The above result is [AKKM21, Lemma 3.5]. ∎ ###### Lemma 2.5. Let $n\geq 3$ and let $\alpha=(2,1\dots,1,\beta,1,\dots,1)$ be a composition of length $(n-1)$ where $\beta$ is in the $k$-th position. Let $r=n-1-k$, i.e., the number of $1$s after $\beta$. Then, the following assertions hold. 1. (1) If $\beta=2$, then, $g_{\alpha}(p)=p^{n-3+r}+(r+1)p^{n-3}(p-1).$ 2. (2) If $\beta\geq 3$, then $g_{\alpha}(p)=(r+1)\left(p^{n-3+r}+p^{n-3}(p-1)\right).$ ###### Proof. The above result is [AKKM21, Lemma 3.6]. ∎ ## 3\. General results for computing the value of $g_{\alpha}(p)$ In this section, we prove a number of general results computing $g_{\alpha}(p)$ for various choices of $\alpha$. Given an irreducible subring with associated matrix $A\in M_{n}(\mathbb{Z})$, recall that $v_{1},\dots,v_{n}$ shall denote the columns of $A$, where $v_{i}$ is the $i$-th column. ###### Lemma 3.1. Let $n\geq 4$, $\alpha=(\beta,1,\dots,1,\gamma)$ be a composition of length $(n-1)$ such that $\beta>2$ and $\gamma\geq\beta-1$. Then, we have $g_{\alpha}(p)=p^{n-3+\lfloor\frac{\beta}{2}\rfloor}+(n-3)p^{n-2}.$ ###### Proof. Consider a matrix $A\in M_{n}(\mathbb{Z})$ in Hermite normal form $A=\begin{pmatrix}p^{\beta}&pa_{1}&pa_{2}&\cdots&pa_{n-2}&1\\\ &p&0&\cdots&0&1\\\ &&p&\cdots&0&1\\\ &&&\ddots&\vdots&\vdots\\\ &&&&p^{\gamma}&1\\\ &&&&&1,\end{pmatrix}$ where the entries $a_{1},\dots,a_{n-2}$ satisfy $0\leq a_{i}\leq p^{\beta-1}-1$. First we write down conditions for $v_{j}^{2}\in\operatorname{Col}(A)$. Clearly, this condition is automatically satisfied for $j=1$ and $j=n$. Consider the values of $j$ that lie in the range $2\leq j\leq n-1$. We may write $j=i+1$, where $i$ lies in the range $1\leq i\leq n-2$. First, we consider the case when $i\leq n-3$. Observe that $v_{i+1}^{2}=(a_{i}^{2}p^{2},0,\dots,0,p^{2},0,\dots,0)^{t}.$ Note that $v_{i+1}^{2}$ is contained in $\operatorname{Col}(A)$ if and only if $v_{i+1}^{2}-pv_{i+1}$ is contained in $\operatorname{Col}(A)$. We find that $v_{i+1}^{2}-pv_{i+1}=\left((a_{i}^{2}-a_{i})p^{2},0,\dots,0\right)^{t}.$ Therefore, $v_{i+1}^{2}$ is contained in $\operatorname{Col}(A)$ if and only if (3.1) $a_{i}^{2}-a_{i}\equiv 0\mod{p^{\beta-2}}.$ We find that $v_{n-1}^{2}=(a_{n-2}^{2}p^{2},0,\dots,0,p^{2\gamma},0)^{t}$ is in $\operatorname{Col}(A)$ if and only if $v_{n-1}^{2}-p^{\gamma}v_{n-1}=\left(a_{n-2}^{2}p^{2}-a_{n-2}p^{\gamma+1},0,\dots,0\right)^{t}$ is contained in $\operatorname{Col}(A)$. Therefore, we deduce that $v_{n-1}^{2}$ is in $\operatorname{Col}(A)$ if and only if (3.2) $a_{n-2}^{2}p^{2}-a_{n-2}p^{\gamma+1}\equiv 0\mod{p^{\beta}}.$ It is assumed that $\gamma\geq\beta-1$ and hence the above condition is equivalent to $a_{n-2}^{2}\equiv 0\mod{p^{\beta-2}},$ i.e., (3.3) $a_{n-2}\equiv 0\mod{p^{\lceil\frac{\beta-2}{2}\rceil}}.$ For $1\leq i<j\leq n-2$, we find that $v_{i+1}\circ v_{j+1}$ is contained in $\operatorname{Col}(A)$ if and only if (3.4) $a_{i}a_{j}\equiv 0\mod{p^{\beta-2}}.$ Putting it all together, we find that $A$ is a subring matrix if and only if the following conditions are satisfied 1. (1) $a_{i}(a_{i}-1)\equiv 0\mod{p^{\beta-2}}$ for all $i$ in the range $1\leq i\leq n-3$, 2. (2) $a_{n-2}\equiv 0\mod{p^{\lceil\beta/2\rceil-1}}$, 3. (3) $a_{i}a_{j}\equiv 0\mod{p^{\beta-2}}$ for all $i,j$ in the range $1\leq i<j\leq n-2$. From condition (1) above, we find that at most one of $a_{i}\equiv 1\mod{p^{\beta-2}}$ for $i=1,\dots,n-3$. We thus are led to the following cases. Case 1: First, we consider the case when $a_{i}\equiv 0\mod{p^{\beta-2}}$ for all $i$ in the range $1\leq i\leq n-3$. Since $a_{i}$ is in the range $0\leq a_{i}<p^{\beta-1}$, we find that there are $p$ choices for each $a_{i}$ and $p^{\beta-\lceil\beta/2\rceil}=p^{\lfloor\beta/2\rfloor}$ choices for $a_{n-2}$. In total, we find that there are $p^{n-3+\lfloor\beta/2\rfloor}$ matrices in this case. Case 2: Consider the case when one of the $a_{i}$ satisfies $a_{i}\equiv 1\mod{p^{\beta-2}}$. Then, from (3) we find that $a_{n-2}\equiv a_{i}a_{n-2}\equiv 0\mod{p^{\beta-2}}.$ Thus, for each index $i$ in the range $1\leq i\leq n-2$, we have $p^{n-2}$ choices for which $a_{i}\equiv 1\mod{p^{\beta-2}}$. There are $(n-3)$ values taken by $i$ for which $a_{i}\equiv 1\mod{p^{\beta-2}}$. Thus, in this case, there are $(n-3)p^{n-2}$ choices. Putting together the calculations from cases 1 and 2, we find that $g_{\alpha}(p)=p^{n-3+\lfloor\beta/2\rfloor}+(n-3)p^{n-2}$. ∎ ###### Lemma 3.2. Let $\alpha$ be of the form $\alpha=(2,1,\dots,1,2,1,\dots,1,\beta)$, where $\beta>1$ and the second $2$ occurs at position $k\geq 2$, then $g_{\alpha}(p)=p^{n-3+r}+p^{n-3}(p-1)r,$ where $r=n-1-k$. ###### Proof. Let $A$ be an integral matrix in Hermite normal form $A=\begin{pmatrix}p^{2}&pa_{1}&pa_{2}&\cdots&pa_{k-1}&\cdots&\cdots&pa_{n-2}&1\\\ &p&0&\cdots&0&\cdots&\cdots&0&1\\\ &&p&\cdots&0&\cdots&\cdots&0&1\\\ &&&\ddots&\vdots&\vdots&\cdots&\vdots&\vdots\\\ &&&&p^{2}&pb_{1}&\cdots&pb_{r}&1\\\ &&&&&p&\cdots&0&1\\\ &&&&&&\ddots&\vdots&1\\\ &&&&&&&p^{\beta}&1\\\ &&&&&&&&1.\end{pmatrix}$ Since the first entry of $\alpha$ is equal to $2$, it is easy to see that $v_{i}^{2}$ is in $\operatorname{Col}(A)$ for $i=1,\dots,k-1$. We find that $v_{k}^{2}$ is in $\operatorname{Col}(A)$ if and only if $v_{k}^{2}-p^{2}v_{k}$ is in $\operatorname{Col}(A)$. Since $p^{2}$ divides $a_{k-1}^{2}p^{2}-p^{3}a_{k-1}$, this condition is seen to be satisfied. For $i=1,\dots,r$, we observe that $v_{k+i}^{2}=(a_{k-1+i}^{2}p^{2},\cdots,b_{i}^{2}p^{2},\cdots,p^{2})^{t}\text{ is in }\operatorname{Col}(A)$ if and only if $v_{k+i}^{2}-pv_{i+k}=\left((a_{k-1+i}^{2}-a_{k-1+i})p^{2},\dots,(b_{i}^{2}-b_{i})p^{2},0,\dots,0\right)^{t}$ is in $\operatorname{Col}(A)$. Subtracting $(b_{i}^{2}-b_{i})v_{k}$ from the above, we get $v_{k+i}^{2}-pv_{i+k}-(b_{i}^{2}-b_{i})v_{k}=\left((a_{k-1+i}^{2}-a_{k-1+i})p^{2}-(b_{i}^{2}-b_{i})a_{k-1}p,0,\dots,0\right)^{t},$ which is in $\operatorname{Col}(A)$ if and only if first entry is divisible by $p^{2}$. We have thus shown that $v_{k+i}^{2}$ is in $\operatorname{Col}(A)$ if and only if (3.5) $(b_{i}^{2}-b_{i})a_{k-1}\equiv 0\mod{p}.$ Now, $v_{n-1}^{2}=\left(a_{n-2}^{2}p^{2},\dots,b_{r}^{2}p^{2},0,\dots,0,p^{2\beta},0,\dots\right)^{t}$ is in $\operatorname{Col}(A)$ if and only if $v_{n-1}^{2}-p^{\beta}v_{n-1}$ is in $\operatorname{Col}(A)$. Note that therefore, $v_{n-1}^{2}$ is in $\operatorname{Col}(A)$ if and only if $b_{r}a_{k-1}\equiv 0\mod{p}$. It is clear that for all $1\leq i\leq k$ and all values of $j$, $v_{i}\circ v_{j}\in\operatorname{Col}(A)$. Now consider $v_{k+i}\circ v_{k+j}$ for $1\leq i<j\leq r$, it is equal to $\left(a_{k-1+i}a_{k-1+j}p^{2},0,\dots,0,b_{i}b_{j}p^{2},0,\dots,0\right)^{t}.$ We find that $v_{k+i}\circ v_{k+j}$ is contained in $\operatorname{Col}(A)$ if and only if $v_{k+i}\circ v_{k+j}-b_{i}b_{j}v_{k-1}\in\operatorname{Col}(A).$ Therefore, $v_{k+i}\circ v_{k+j}$ is contained in $\operatorname{Col}(A)$ if and only if $b_{i}b_{j}a_{k-1}\equiv 0\mod{p}$. Therefore, $A$ is a subring matrix if and only if 1. (1) $(b_{i}^{2}-b_{i})a_{k-1}\equiv 0\mod{p}$ for $i=1,\dots,r-1$, 2. (2) $b_{r}a_{k-1}\equiv 0\mod{p}$, 3. (3) $b_{i}b_{j}a_{k-1}\equiv 0\mod{p}$, for all $i,j$ such that $1\leq i<j\leq r$. We consider two cases. Case 1: Consider the case when $a_{k-1}=0$. In this case, the total number of choices is the total number of choices of $(a_{1},\dots,a_{k-2},a_{k},\dots,a_{n-2})$ and $(b_{1},\dots,b_{r})$. Thus, the total number of choices are $p^{n-3+r}$. Case 2: Consider the other case, i.e., when $a_{k-1}\neq 0$. Since $a_{k-1}<p$, we find that $p\nmid a_{k-1}$. Therefore, the conditions are as follows 1. (1) $b_{i}^{2}-b_{i}\equiv 0\mod{p}$, 2. (2) $b_{r}\equiv 0\mod{p}$, 3. (3) $b_{i}b_{j}\equiv 0\mod{p}$, $1\leq i<j\leq r$. Note that all elements $b_{i}$ satisfy the bounds $0\leq b_{i}<p$. Therefore, at most one of the $b_{i}$ satisfies $b_{i}=1$. Subdivide into two cases, first consider the case when all the $b_{i}$ are equal to $0$ and then consider the case when at most one of the $b_{i}$ is equal to $1$. The total number of choices is $p^{n-3}(p-1)r$. Upon adding up all the conclusions from the case decomposition above, we find that $g_{\alpha}(p)=p^{n-3+r}+p^{n-3}(p-1)r.$ ∎ ###### Lemma 3.3. Let $\alpha$ be of the form $\alpha=(2,1,\dots,1,2,1,\dots,1,\beta,1)$, where $\beta>1$, the second $2$ occurs at position $k\geq 2$ and set $r\mathrel{\mathop{\mathchar 58\relax}}=n-1-k$. Then, the following assertions hold. 1. (1) If $\beta=2$, then we find that $g_{\alpha}(p)=2p^{n-3+r}+p^{n-5+r}(p-2)+2rp^{n-3}(p-1)+p^{n-4}(p-1)(p-2)(p+r-2).$ 2. (2) If $\beta\geq 3$, then, $g_{\alpha}(p)=2p^{n-3+r}+2p^{n-5+r}(p-1)+2rp^{n-3}(p-1)+2p^{n-4}(p-1)^{2}(p+r-2).$ ###### Proof. Let $A$ be an integral matrix in Hermite normal form $A=\begin{pmatrix}p^{2}&pa_{1}&pa_{2}&\cdots&pa_{k-1}&\cdots&\cdots&\cdots&pa_{n-2}&1\\\ &p&0&\cdots&0&\cdots&\cdots&\cdots&0&1\\\ &&p&\cdots&0&\cdots&\cdots&\cdots&0&1\\\ &&&\ddots&\vdots&\vdots&\cdots&\cdots&\vdots&\vdots\\\ &&&&p^{2}&pb_{1}&\cdots&\cdots&pb_{r}&1\\\ &&&&&p&\cdots&\cdots&0&1\\\ &&&&&&\ddots&\cdots&\vdots&1\\\ &&&&&&&p^{\beta}&pc_{1}&1\\\ &&&&&&&&p&1\\\ &&&&&&&&&1.\end{pmatrix}$ Since the first entry is $p^{2}$, it is clear that $v_{i}^{2}$ is contained in $\operatorname{Col}(A)$ for $i=1,2,\dots,k$. For $i=1,2,\dots,r-2$, we find that $v_{k+i}^{2}$ is contained in $\operatorname{Col}(A)$ if and only if $v_{k+i}^{2}-pv_{k+i}$ is contained in $\operatorname{Col}(A)$. So, $(a_{k-1+i}^{2}p^{2}-a_{k-1}p^{2},\dots,b_{i}^{2}p^{2}-b_{i}p^{2},\dots)^{t}$ should be in $\operatorname{Col}(A)$. It is clear that this is true if and only if (3.6) $a_{k-1}(b_{i}^{2}-b_{i})\equiv 0\mod{p}$ Similarly we see that $v_{n-2}^{2}\in\operatorname{Col}(A)$ if and only if $v_{n-2}^{2}-p^{\beta}v_{n-2}\in\operatorname{Col}(A)$. Now, $(a_{n-3}^{2}p^{2}-a_{n-3}p^{\beta+1},\dots,b_{r-1}^{2}p^{2}-b_{r-1}p^{\beta+1},\dots)^{t}$ should be in $\operatorname{Col}(A)$. It is again clear that this is true if and only if (3.7) $a_{k-1}b_{r-1}\equiv 0\mod p$ as $\beta>1$. Again, $v_{n-1}^{2}\in\operatorname{Col}(A)$ if and only if $v_{n-1}^{2}-pv_{n-1}\in\operatorname{Col}(A)$, so $(a_{n-2}^{2}p^{2}-a_{n-2}p^{2},\dots,(b_{r}^{2}-b_{r})p^{2},\dots,(c_{1}^{2}-c_{1})^{2},0,0)^{t}$ should be in $\operatorname{Col}(A)$. Then we have $c_{1}^{2}-c_{1}\equiv 0\mod p^{\beta-2}$, and (3.8) $b_{r-1}(c_{1}^{2}-c_{1})\equiv 0\mod p^{\beta-1}$ (3.9) $a_{k-1}\left(b_{r}^{2}-b_{r}-\frac{b_{r-1}(c_{1}^{2}-c_{1})}{p^{\beta-1}}\right)+a_{n-3}\frac{c_{1}^{2}-c_{1}}{p^{\beta-2}}\equiv 0\mod p$ The requirement that $v_{i}v_{j}$ belongs to $\operatorname{Col}(A)$ for $i\neq j$ translates the following conditions on entries of the matrix (for all $1\leq i<j\leq r$, and $\\{i,j\\}\neq\\{r-1,r\\}$) $a_{k-1}b_{i}b_{j}\equiv 0\mod p$ First, we suppose that $\beta\geq 3$, then we have conditions, 1. (1) $a_{k-1}(b_{i}^{2}-b_{i})\equiv 0\mod{p}$, $i=1,\dots,r-2$, 2. (2) $b_{r-1}a_{k-1}\equiv 0\mod p$, 3. (3) $c_{1}^{2}-c_{1}\equiv 0\mod p^{\beta-2}$, 4. (4) $b_{r-1}(c_{1}^{2}-c_{1})\equiv 0\mod p^{\beta-1}$, 5. (5) $a_{k-1}\left(b_{r}^{2}-b_{r}-\frac{b_{r-1}(c_{1}^{2}-c_{1})}{p^{\beta-1}}\right)+a_{n-3}\frac{c_{1}^{2}-c_{1}}{p^{\beta-2}}\equiv 0\mod p$, 6. (6) $a_{k-1}b_{i}b_{j}\equiv 0\mod p$, $1\leq i<j\leq r$, and $\\{i,j\\}\neq\\{r-1,r\\}$. We consider two cases. Case 1 : First, we consider the case when $a_{k-1}=0$. Then, the above equations reduce to the following: 1. (1) $c_{1}^{2}-c_{1}\equiv 0\mod p^{\beta-2}$, 2. (2) $b_{r-1}(c_{1}^{2}-c_{1})\equiv 0\mod p^{\beta-1}$, 3. (3) $\frac{a_{n-3}(c_{1}^{2}-c_{1})}{p^{\beta-2}}\equiv 0\mod p$. If $c_{1}\in\\{0,1\\}$, then the number of such matrices is $2p^{n-3+r}$. On the other hand, if $c_{1}\notin\\{0,1\\}$, then, there are $2p-2$ choices for $c_{1}$. Since $b_{r-1}=0,a_{n-3}=0$, we deduce that there are $2(p-1)p^{n+r-5}$ more matrices. Case 2 : Suppose $a_{k-1}\neq 0$, then equations reduce to 1. (1) $b_{i}^{2}-b_{i}\equiv 0\mod p$, $i=1,\dots,r-2$, 2. (2) $b_{r-1}=0$, 3. (3) $c_{1}^{2}-c_{1}\equiv 0\mod p^{\beta-2}$, 4. (4) $a_{k-1}\left(b_{r}^{2}-b_{r}\right)+a_{n-3}\frac{c_{1}^{2}-c_{1}}{p^{\beta-2}}\equiv 0\mod p$, 5. (5) $b_{i}b_{j}\equiv 0\mod p$, $1\leq i<j\leq r$, and $\\{i,j\\}\neq\\{r-1,r\\}$ We divide into two sub-cases. * • Case 2A : Suppose that $c_{1}\in\\{0,1\\}$, then we get that $b_{r}\in\\{0,1\\}$ and as in Lemma 3.2, we get that at most one of the $b_{i}$ satisfies $b_{i}=1$. Subdivide into two cases, first consider the case when all the $b_{i}$ are equal to $0$ and then consider the case when at most one of the $b_{i}$ is equal to $1$. Therefore, the total number of choices is $2rp^{n-3}(p-1)$. * • Case 2B : Consider the other subcase, i.e., $c_{1}\notin\\{0,1\\}$. Then, there is a unique value of $a_{n-3}$ for each value of $b_{r},a_{k-1},c_{1}$, so we get $2p^{n-3}(p-1)^{2}(p+r-2)$ matrices Putting it all together, the result is proven. The case $\beta=2$ is similar, and the number of matrices change only in Case 1, with $c_{1}\notin\\{0,1\\}$ and Case 2B. ∎ ###### Lemma 3.4. Let $\alpha$ be of the form $\alpha=(2,1,\dots,1,3,1,\dots,1,2)$, where $3$ occurs at position $k\geq 2$. Then, we have that $g_{\alpha}(p)=rp^{n-3+r}+p^{n-3}(p-1)(p+r-1),$ where $r\mathrel{\mathop{\mathchar 58\relax}}=n-1-k$. ###### Proof. Let $A$ be an integral matrix in Hermite normal form $A=\begin{pmatrix}p^{2}&pa_{1}&pa_{2}&\cdots&pa_{k-1}&\cdots&\cdots&pa_{n-2}&1\\\ &p&0&\cdots&0&\cdots&\cdots&0&1\\\ &&p&\cdots&0&\cdots&\cdots&0&1\\\ &&&\ddots&\vdots&\vdots&\cdots&\vdots&\vdots\\\ &&&&p^{3}&pb_{1}&\cdots&pb_{r}&1\\\ &&&&&p&\cdots&0&1\\\ &&&&&&\ddots&\vdots&1\\\ &&&&&&&p^{2}&1\\\ &&&&&&&&1.\end{pmatrix}$ Since the first diagonal entry is $p^{2}$, we see that $v_{i}^{2}$ is contained in $\operatorname{Col}(A)$ for all $i=1,\dots,k$. Note that $v_{k+i}^{2}\in\operatorname{Col}(A)$ if and only if $v_{k+i}^{2}-pv_{k+i}\in\operatorname{Col}(A)$ for $i=1,2,\dots,r-1$. Hence, we find that $\left((a_{k-1+i}^{2}-a_{k-1+i})p^{2},0,\dots,0,(b_{i}^{2}-b_{i})p^{2},0,\dots,0\right)^{t}$ should be in $\operatorname{Col}(A)$. We deduce that that this is the case if and only if the following conditions are satisfied (3.10) $\begin{split}&b_{i}^{2}-b_{i}\equiv 0\mod p,\\\ &a_{k-1}(b_{i}^{2}-b_{i})\equiv 0\mod p^{2}.\end{split}$ It is also required that $v_{n-1}^{2}$ is contained in $\operatorname{Col}(A)$, which is the case if and only if $v_{n-1}^{2}-p^{2}v_{n-1}$ is contained in $\operatorname{Col}(A)$. Therefore, we find that $\left((a_{n-2}^{2}-a_{n-2}p)p^{2},0,\dots,0,(b_{r}^{2}-b_{r}p)p^{2},0,\dots,0\right)^{t}$ should be in $\operatorname{Col}(A)$. We get that $b_{r}^{2}-b_{r}p\equiv 0\mod p$ and $a_{k-1}(b_{r}^{2}-b_{r}p)\equiv 0\mod p^{2}$, therefore, we find that $p\mid b_{r}$. Using similar arguments we deduce that $v_{k+i}v_{k+j}\in\operatorname{Col}(A)$ for $1\leq i<j\leq r$ if and only if the following conditions are satisfied (3.11) $\begin{split}&b_{i}b_{j}\equiv 0\mod p,\\\ &a_{k-1}b_{i}b_{j}\equiv 0\mod p^{2}.\end{split}$ It is clear that $v_{i}v_{j}$ is contained in $\operatorname{Col}(A)$ for $1\leq i<j\leq k$. Therefore, we have following conditions on the entries of $A$ 1. (1) $b_{i}^{2}-b_{i}\equiv 0\mod p$ and $a_{k-1}(b-i^{2}-b_{i})\equiv 0\mod p^{2}$ for $i=1,\dots,r-1$, 2. (2) $b_{r}=b_{r}^{\prime}p$, where $0\leq b_{r}^{\prime}\leq p-1$, 3. (3) $b_{i}b_{j}\equiv 0\mod p$ and $a_{k-1}b_{i}b_{j}\equiv 0\mod p^{2}$ for all values $1\leq i<j\leq r$. We consider two cases as follows. Case 1 : First, we consider the case when $a_{k-1}=0$. In this case, the conditions on $A$ reduce to the following 1. (1) $b_{i}^{2}-b_{i}\equiv 0\mod p$ for all $i=1,\dots,r-1$, 2. (2) $b_{r}=b_{r}^{\prime}p$, where $0\leq b_{r}^{\prime}\leq p-1$, 3. (3) $b_{i}b_{j}\equiv 0\mod p$ for $1\leq i<j\leq r$. As in Lemma 3.3, either all the values of $b_{i}$ for $1\leq i\leq r-1$ are $0\mod p$, or at most one of these values is $1\mod p$. Further dividing into cases, it is easy to see that the number of matrices is $rp^{n-3+r}$. Case 2 : Next, we consider the other case, namely assume that $a_{k-1}\neq 0$. The conditions on $A$ then reduce to the following 1. (1) $b_{i}^{2}-b_{i}\equiv 0\mod p^{2}$ for all $i=1,\dots,r-1$, 2. (2) $b_{r}=b_{r}^{\prime}p$, where $0\leq b_{r}^{\prime}\leq p-1$, 3. (3) $b_{i}b_{j}\equiv 0\mod p^{2}$ for all $(i,j)$ satisfying $1\leq i<j\leq r$. Consequently, it follows that either $b_{i}=0$ for all $i=1,\dots,r-1$, or at most one of them is $1$. Therefore, we find that the number of matrices in this case is equal to $p^{n-3}(p-1)(p+r-1)$ matrices. Adding up our conclusions, we prove the assertion of the lemma. ∎ ## 4\. Calculating the values of $g_{\alpha}(p)$ for compositions beginning with $2$ ### 4.1. Compositions of length $4$ We consider the compositions $\alpha$ of length $4$. In all, there are $15$ of them, listed below $(2,5,1,1)$ | $(2,1,2,4)$ ---|--- $(2,1,5,1)$ | $(2,3,3,1)$ $(2,1,1,5)$ | $(2,3,1,3)$ $(2,4,2,1)$ | $(2,1,3,3)$ $(2,4,1,2)$ | $(2,3,2,2)$ $(2,2,4,1)$ | $(2,2,3,2)$ $(2,2,1,4)$ | $(2,2,2,3)$. $(2,1,4,2)$ | It follows directly from Lemmas 2.5, 3.2 and 3.3 that $g_{(2,5,1,1)}(p)$ | $3p^{4}+3p^{3}-3p^{2}$ ---|--- $g_{(2,1,5,1)}(p)$ | $4p^{3}-2p^{2}$ $g_{(2,1,1,5)}(p)$ | $p^{3}$ $g_{(2,2,4,1)}(p)$ | $4p^{4}+2p^{3}-4p^{2}$ $g_{(2,2,1,4)}(p)$ | $p^{4}+2p^{3}-2p^{2}$ $g_{(2,1,2,4)}(p)$ | $2p^{3}-p^{2}$. We compute $g_{\alpha}(p)$ for the rest of the compositions. We note that many arguments are similar, and we summarize the arguments that tend to repeat. ###### Lemma 4.1. With respect to notation above, we have that $g_{(2,1,4,2)}(p)=3p^{4}-2p^{3}$. ###### Proof. Let $A$ be the matrix, $\begin{pmatrix}p^{2}&a_{1}p&a_{2}p&a_{3}p&1\\\ 0&p&0&0&1\\\ 0&0&p^{4}&b_{1}p&1\\\ 0&0&0&p^{2}&1\\\ 0&0&0&0&1\end{pmatrix}$ where $0\leq a_{1}\leq p-1$ and $0\leq b_{1}\leq p^{3}-1$. We arrive at conditions for $A$ to be a subring matrix. It is easy to see that * • $v_{2}^{2}$,$v_{3}^{2}$,$v_{2}v_{3}$, $v_{2}v_{4}$, $v_{3}v_{4}$ are in $\operatorname{Col}(A)$. * • We find that $v_{4}^{2}$ is in $\operatorname{Col}(A)$ if and only if $v_{4}^{2}-p^{2}v_{4}$ is contained in $\operatorname{Col}(A)$. In other words, $\begin{split}&(a_{3}^{2}p^{2},0,b_{1}^{2}p^{2},p^{4},0)^{t}-p^{2}(a_{3}p,0,b_{1}p,p^{2},0)^{t}\\\ =&\left((a_{3}^{2}p^{2}-a_{3}p^{3}),0,(b_{1}^{2}p^{2}-b_{1}p^{3}),0,0\right)^{t}\end{split}$ is in $\operatorname{Col}(A)$. The expression $(b_{1}^{2}p^{2}-b_{1}p^{3})=b_{1}(b_{1}-p)p^{2}$ must be divisible by $p^{4}$ and moreover, we find that $v_{4}^{2}-p^{2}v_{4}$ is contained in $\operatorname{Col}(A)$ if and only if $v_{4}^{2}-p^{2}v_{4}-\frac{b_{1}(b_{1}-p)}{p^{2}}v_{3}$ is in $\operatorname{Col}(A)$. This holds if and only if in addition, we have that $b_{1}(b_{1}-p)a_{2}\equiv 0\mod p^{3}$. We deduce from above that the necessary conditions are as follows 1. (1) $b_{1}(b_{1}-p)\equiv 0\mod p^{2}$, 2. (2) $b_{1}(b_{1}-p)a_{2}\equiv 0\mod p^{3}$. From equation (1), we get that $b_{1}=b_{1}^{\prime}p$, where $0\leq b_{1}^{\prime}\leq p^{2}-1$. Equation (2) asserts that $b_{1}^{\prime}(b_{1}^{\prime}-1)a_{2}\equiv 0\mod p$. Consider the following case decomposition. * • Case 1 : Assume that $a_{2}=0$. In this case, the number of such matrices is $p^{4}$. * • Case 2 : Consider the other case, i.e., $a_{2}\neq 0$. Then, we have $b_{1}^{\prime}(b_{1}^{\prime}-1)\equiv 0\mod p$. Hence, the total number of such matrices is easily seen to be $2p^{3}(p-1)$. We conclude from the above that $g_{(2,1,4,2)}(p)=p^{4}+2p^{3}(p-1)=3p^{4}-2p^{3}$. ∎ ###### Lemma 4.2. With respect to notation above, we find that $g_{(2,4,1,2)}(p)=p^{5}+3p^{4}-p^{3}-p^{2}$. ###### Proof. Let $A$ be the matrix $\begin{pmatrix}p^{2}&a_{1}p&a_{2}p&a_{3}p&1\\\ 0&p^{4}&b_{1}p&b_{2}p&1\\\ 0&0&p&0&1\\\ 0&0&0&p^{2}&1\\\ 0&0&0&0&1\end{pmatrix}$ where $0\leq a_{i}\leq p-1$ and $0\leq b_{j}\leq p^{3}-1$. We obtain conditions for $A$ to be a subring matrix. By the same arguments as in Lemma 4.1, we find that * • $v_{2}^{2}$, $v_{2}v_{3}$, $v_{2}v_{4}$ are in $\operatorname{Col}(A)$. * • We find that $v_{3}^{2}$ is in $\operatorname{Col}(A)$ if and only if $b_{1}(b_{1}-1)\equiv 0\mod p^{2}$ and $b_{1}(b_{1}-1)a_{1}\equiv 0\mod p^{3}$. * • We find that $v_{4}^{2}$ is in $\operatorname{Col}(A)$ if and only if $b_{2}=b_{2}^{\prime}p$, $0\leq b_{2}^{\prime}\leq p^{2}-1$ and $b_{2}^{\prime}(b_{2}^{\prime}-1)a_{1}\equiv 0\mod p$. * • We find that $v_{3}v_{4}$ is in $\operatorname{Col}(A)$ if and only if $b_{1}b_{2}^{\prime}\equiv 0\mod p$ and $a_{1}b_{1}b_{2}^{\prime}\equiv 0\mod p^{2}$. Hence, $A$ is a subring matrix if and only if 1. (1) $b_{1}(b_{1}-1)\equiv 0\mod p^{2}$, 2. (2) $a_{1}b_{1}(b_{1}-1)\equiv 0\mod p^{3}$, 3. (3) $b_{2}=b_{2}^{\prime}p$ for $0\leq b_{2}^{\prime}\leq p^{2}-1$, 4. (4) $a_{1}b_{2}^{\prime}(b_{2}^{\prime}-1)\equiv 0\mod p$, 5. (5) $b_{1}b_{2}^{\prime}\equiv 0\mod p$, 6. (6) $a_{1}b_{1}b_{2}^{\prime}\equiv 0\mod p^{2}$. In order to compute the total number of matrices satisfying all of the above conditions, we consider the following cases. * • Case 1 : Assume that $a_{1}=0$. If $b_{1}\equiv 0\mod p^{2}$ then number of such matrices is $p^{5}$ otherwise $b_{1}\equiv 1\mod p^{2}$, number of such matrices is $p^{4}$. * • Case 2 : Consider the case when $a_{1}\neq 0$. We get $b_{1}=0$ or $b_{1}=1$. If $b_{1}=0$, number of such matrices is $2p^{3}(p-1)$. Otherwise $b_{1}=1$ and number of such matrices is $p^{2}(p-1)$ (as $b_{2}=0$ in this case). Therefore, we find that $g_{(2,4,1,2)}(p)=p^{5}+p^{4}+2p^{3}(p-1)+p^{2}(p-1)=p^{5}+3p^{4}-p^{3}-p^{2}$. ∎ ###### Lemma 4.3. We have that $\begin{split}&g_{(2,3,1,3)}(p)=3p^{4}-p^{2},\\\ &g_{(2,1,3,3)}(p)=p^{4},\\\ &g_{(2,2,3,2)}=p^{5}+p^{4}-p^{3}.\end{split}$ ###### Proof. The proof is similar to Lemma 4.2, and we omit it. ∎ ###### Lemma 4.4. We have that $g_{(2,3,2,2)}(p)=p^{5}+4p^{4}-9p^{3}+4p$. ###### Proof. Let $A$ be the matrix, $\begin{pmatrix}p^{2}&a_{1}p&a_{2}p&a_{3}p&1\\\ 0&p^{3}&b_{1}p&b_{2}p&1\\\ 0&0&p^{2}&c_{1}p&1\\\ 0&0&0&p^{2}&1\\\ 0&0&0&0&1\end{pmatrix}$ where $0\leq a_{i}\leq p-1$, $0\leq b_{j}\leq p^{2}-1$ and $0\leq c_{1}\leq p-1$. Below are the conditions for $A$ to be a subring matrix. * • It is easy to see that $v_{2}^{2}$ is in $\operatorname{Col}(A)$. * • We find that $v_{3}^{2}$ is contained in $\operatorname{Col}(A)$ if and only if $v_{3}^{2}-p^{2}v_{3}$ is contained in $\operatorname{Col}(A)$. Hence, we find that $v_{3}^{2}\in\operatorname{Col}(A)$ if and only if $b_{1}=b_{1}^{\prime}p$, where $0\leq b_{1}^{\prime}\leq p-1$. * • Similar reasoning shows that $v_{4}^{2}\in\operatorname{Col}(A)$ if and only if $b_{2}^{2}-(c_{1}^{2}-c_{1})b_{1}^{\prime}\equiv 0\mod p$ and $\left(\frac{b_{2}^{2}-(c_{1}^{2}-c_{1})b_{1}^{\prime}}{p}-b_{2}\right)a_{1}+c_{1}(c_{1}-1)a_{2}\equiv 0\mod p$. * • We next consider the requirement that $v_{i}v_{j}$ is in $\operatorname{Col}(A)$ for distinct values of $i$ and $j$. It is easy to see that $v_{1}v_{2},v_{1}v_{3},v_{1}v_{4},v_{2}v_{3}$ and $v_{2}v_{4}$ are in $\operatorname{Col}(A)$ and that $v_{3}v_{4}$ is in $\operatorname{Col}(A)$ if and only if $b_{1}^{\prime}a_{1}(b_{2}-c_{1})\equiv 0\mod p$. Summarizing the above, the conditions we arrive at are as follows 1. (1) $b_{1}=b_{1}^{\prime}p$, where $0\leq b_{1}^{\prime}\leq p-1$, 2. (2) $b_{2}^{2}-(c_{1}^{2}-c_{1})b_{1}^{\prime}\equiv 0\mod p$, 3. (3) $\left(\dfrac{b_{2}^{2}-(c_{1}^{2}-c_{1})b_{1}^{\prime}}{p}-b_{2}\right)a_{1}+c_{1}(c_{1}-1)a_{2}\equiv 0\mod p$, 4. (4) $b_{1}^{\prime}a_{1}(b_{2}-c_{1})\equiv 0\mod p$. Consider the following cases. * • Case 1 : First consider the case when $a_{1}=0$. We further subdivide our argument into the following subcases. * $\diamond$ Case 1A : Assume that $c_{1}\in\\{0,1\\}$. Then, we find that $b_{2}\equiv 0\mod p$. There are $2p^{4}$ such matrices. * $\diamond$ Case 1B : Assume that $c_{1}\notin\\{0,1\\}$. In this case, we get $a_{2}=0$, and thus, there are $p^{3}(p-2)$ such matrices. * • Case 2 : Next consider the case when $a_{1}\neq 0$. Then $b_{1}^{\prime}(b_{2}-c_{1})\equiv 0\mod p$. We have the following subcases. * $\diamond$ Case 2A : Assume that $b_{1}^{\prime}=0$, then we find that $b_{2}\equiv 0\mod p$. Now if $c_{1}\in\\{0,1\\}$ then we get $2p^{3}(p-1)$ such matrices. If $c_{1}\notin\\{0,1\\}$ we get $p^{2}(p-1)(p-2)$ such matrices. * $\diamond$ Case 2B : Assume that $b_{1}^{\prime}\neq 0$, then we get $b_{2}\equiv c_{1}\mod p$. If $c_{1}=0$ we get $p^{3}(p-1)^{2}$ such matrices. If $c_{1}=1$ we get no subring matrices as we get $b_{2}\equiv 1\mod p$ and $b_{2}\equiv 0\mod p$, but no such integer exists. If $c_{1}\notin\\{0,1\\}$, we get $p^{2}(p-1)(p-2)$ subring matrices. Adding up the numbers obtained in the above case decomposition, we find that $g_{(2,3,2,2)}(p)=p^{5}+4p^{4}-9p^{3}+4p$. ∎ ###### Lemma 4.5. We have that $\begin{split}&g_{(2,4,2,1)}(p)=2p^{5}+8p^{4}-15p^{3}+6p^{2},\\\ &g_{(2,3,3,1)}(p)=12p^{4}-10p^{3}+2p^{2}.\end{split}$ ###### Proof. The proof is similar to that of Lemma 4.4, and is omitted. ∎ ### 4.2. Compositions of length $5$ In this subsection we consider the compositions of length $5$ that begin with $2$. We need to compute $g_{\alpha}(p)$ for $20$ compositions of this form. They are listed below, $(2,4,1,1,1)$ | $(2,1,4,1,1)$ ---|--- $(2,1,1,4,1)$ | $(2,1,1,1,4)$ $(2,3,2,1,1)$ | $(2,3,1,2,1)$ $(2,3,1,1,2)$ | $(2,2,3,1,1)$ $(2,2,1,3,1)$ | $(2,2,1,1,3)$ $(2,1,3,1,2)$ | $(2,1,3,2,1)$ $(2,1,2,3,1)$ | $(2,1,2,1,3)$ $(2,1,1,3,2)$ | $(2,1,1,2,3)$ $(2,2,2,2,1)$ | $(2,2,2,1,2)$ $(2,2,1,2,2)$ | $(2,1,2,2,2)$ Table 1. Compositions of length $5$ that begin with $2$ In the next two lemmas we obtain values of $g_{\alpha}(p)$ for all the compositions listed above. ###### Lemma 4.6. We have following values $g_{(2,4,1,1,1)}(p)$ | $p^{6}+4p^{4}-4p^{3}$ ---|--- $g_{(2,1,4,1,1)}(p)$ | $p^{5}+3p^{4}-3p^{3}$ $g_{(2,1,1,4,1)}(p)$ | $3p^{4}-2p^{3}$ $g_{(2,1,1,1,4)}(p)$ | $p^{4}$ $g_{(2,3,1,1,2)}(p)$ | $3p^{6}+p^{5}+p^{4}-2p^{3}$ $g_{(2,1,3,1,2)}(p)$ | $3p^{5}-p^{3}$ $g_{(2,1,1,3,2)}(p)$ | $p^{5}$ $g_{(2,2,1,1,3)}(p)$ | $p^{6}+3p^{4}-3p^{3}$ $g_{(2,1,2,1,3)}(p)$ | $p^{5}+2p^{4}-2p^{3}$ $g_{(2,1,1,2,3)}(p)$ | $2p^{4}-p^{3}$ $g_{(2,2,1,3,1)}(p)$ | $2p^{6}+4p^{5}+2p^{4}-8p^{3}+2p^{2}$ $g_{(2,1,2,3,1)}(p)$ | $4p^{5}+4p^{4}-4p^{3}$ ###### Proof. The result follows immediately from Lemmas 2.5, 3.2, 3.3 and 3.4. ∎ Next, we obtain values of $g_{\alpha}(p)$ for remaining $8$ compositions. ###### Lemma 4.7. We have following values, $g_{(2,3,2,1,1)}(p)$ | $3p^{7}+9p^{6}+7p^{5}-18p^{4}-3p^{2}$ ---|--- $g_{(2,2,3,1,1)}(p)$ | $p^{7}+5p^{6}+8p^{5}-11p^{4}$ $g_{(2,1,3,2,1)}(p)$ | $10p^{5}-11p^{4}+4p^{3}$ $g_{(2,3,1,2,1)}(p)$ | $7p^{6}+2p^{5}-2p^{4}-5p^{3}+2p^{2}$ $g_{(2,2,2,2,1)}(p)$ | $9p^{6}+2p^{5}-18p^{4}+12p^{3}-4p^{2}$ $g_{(2,2,2,1,2)}(p)$ | $2p^{6}+9p^{5}-16p^{4}+8p^{3}-2p^{2}$ $g_{(2,2,1,2,2)}(p)$ | $p^{6}+3p^{5}-3p^{4}$ $g_{(2,1,2,2,2)}(p)$ | $2p^{5}-p^{3}$ ###### Proof. We will prove this for the composition $(2,1,2,2,2)$, expression for other compositions can be obtained in a similar way. Let $A$ be a matrix in Hermite normal form of following type, $\begin{pmatrix}p^{2}&a_{1}p&a_{2}p&a_{3}p&a_{4}p&1\\\ &p&0&0&0&1\\\ &&p^{2}&b_{1}p&b_{2}p&1\\\ &&&p&c_{1}p&1\\\ &&&&p^{2}&1\\\ &&&&&1\\\ \end{pmatrix}$ First, we determine the conditions on entries of the matrix which make $A$ a subring matrix. 1. (1) First, we note that $v_{2}^{2}\in\mathrm{Col}(A)$ and $v_{3}^{2}\in\mathrm{Col}(A)$. 2. (2) Next, we know that $v_{4}^{2}\in\mathrm{Col}(A)$ if and only $v_{4}^{2}-p^{2}v_{4}\in\mathrm{Col}(A)$. Now, this is the case if and only if $b_{1}a_{2}\equiv 0\mod p$. 3. (3) Similarly, $v_{5}^{2}\in\mathrm{Col}(A)$ if and only if $v_{5}^{2}-p^{2}v_{5}\in\mathrm{Col}(A)$ and this is true if and only if $b_{1}c_{1}\equiv 0\mod p$ and $b_{2}^{2}a_{2}-c_{1}^{2}a_{3}\equiv 0\mod p$. 4. (4) We also note that if above conditions are satisfied then $v_{2}v_{3},v_{2}v_{4},v_{2}v_{5},v_{3}v_{4},v_{3}v_{5}$ and $v_{4}v_{5}$ are in $\mathrm{Col}(A)$. The conditions we get are, * • $b_{1}a_{2}=0$, * • $b_{1}c_{1}=0$, * • $b_{2}^{2}a_{2}-c_{1}^{2}a_{3}\equiv 0\mod p$. First we assume that $b_{1}=0$. Now further if $c_{1}=0$ then we get $p^{3}(2p-1)$ matrices. If $c_{1}\neq 0$ then we get $p^{4}(p-1)$ matrices. Next, we assume that $b_{1}\neq 0$ then $a_{2}=0$ and $c_{1}=0$, therefore we get $p^{4}(p-1)$ matrices. Therefore, we get that $g_{(2,1,2,2,2)}(p)=2p^{5}-p^{3}$. ∎ ## 5\. Calculating the values of $g_{\alpha}(p)$ for compositions beginning with $3$ ### 5.1. Compositions of length $4$ In this section, we compute $g_{\alpha}(p)$ for $10$ compositions $\alpha$ beginning with $3$ and of length $4$. They are listed as follows $(3,4,1,1)$ | $(3,3,1,2)$ ---|--- $(3,1,4,1)$ | $(3,2,3,1)$ $(3,1,1,4)$ | $(3,1,2,3)$ $(3,2,1,3)$ | $(3,3,2,1)$ $(3,1,3,2)$ | $(3,2,2,2)$. We make note of some known computations. ###### Lemma 5.1. The following values of $g_{\alpha}(p)$ are known $g_{(3,4,1,1)}(p)$ | $12p^{4}-9p^{3}+p^{2}$ ---|--- $g_{(3,1,4,1)}(p)$ | $2p^{4}+6p^{3}-2p^{2}$ $g_{(3,1,1,4)}(p)$ | $3p^{3}$. ###### Proof. The above result follows from [Ish22a, Lemma 4.6]. ∎ ###### Lemma 5.2. We have that $g_{(3,2,1,3)}(p)=5p^{4}-4p^{3}+2p^{2}.$ ###### Proof. Let $A$ be any integer matrix in Hermite normal form of the type $\begin{pmatrix}p^{3}&a_{1}p&a_{2}p&a_{3}p&1\\\ &p^{2}&b_{1}p&b_{2}p&1\\\ &&p&0&1\\\ &&&p^{3}&1\\\ &&&&1.\\\ \end{pmatrix}$ We determine the conditions so that $A$ is a subring matrix. 1. (1) It is required that $v_{2}^{2}\in\mathrm{Col}(A)$ and this is the case if and only if $(a_{1}^{2}p^{2},p^{4},0,0,0)^{t}-p^{2}(a_{1}p,p^{2},0,0,0)^{t}$ is in $\mathrm{Col}(A)$. We deduce that $a_{1}\equiv 0\mod p$, hence $a_{1}=a_{1}^{\prime}p$ where $0\leq a_{1}^{\prime}\leq p-1$. 2. (2) It is required that $v_{3}^{2}\in\mathrm{Col}(A)$; clearly this is equivalent to the statement that $v_{3}^{2}-pv_{3}\in\mathrm{Col}(A)$. Now, it is easy to see that latter condition is true if and only if (5.1) $(a_{2}^{2}-a_{2})-a_{1}^{\prime}(b_{1}^{2}-b_{1})\equiv 0\mod p.$ 3. (3) Similarly, it is clear that $v_{4}^{2}\in\mathrm{Col}(A)$ if and only if $v_{4}^{2}-p^{3}v_{4}\in\mathrm{Col}(A)$ and latter condition is equivalent to (5.2) $a_{3}^{2}-b_{2}^{2}a_{1}^{\prime}\equiv 0\mod p.$ 4. (4) One can easily check that $v_{2}v_{3}$ and $v_{2}v_{4}$ are both in $\mathrm{Col}(A)$ and $v_{3}v_{4}\in\mathrm{Col}(A)$ if and only (5.3) $a_{2}a_{3}-a_{1}^{\prime}b_{1}b_{2}\equiv 0\mod p.$ Thus, to summarize, we arrive at the following conditions 1. (1) $a_{1}=a_{1}^{\prime}p$ where $0\leq a_{1}^{\prime}\leq p-1$, 2. (2) $(a_{2}^{2}-a_{2})-a_{1}^{\prime}(b_{1}^{2}-b_{1})\equiv 0\mod p$, 3. (3) $a_{3}^{2}-b_{2}^{2}a_{1}^{\prime}\equiv 0\mod p$, 4. (4) $a_{2}a_{3}-a_{1}^{\prime}b_{1}b_{2}\equiv 0\mod p.$ We count matrices $A$ satisfying these conditions by dividing into cases. Case 1 : First consider the case when $a_{3}\equiv\mod p$. From equation (5.2) we deduce that $a_{1}^{\prime}b_{2}\equiv 0\mod p$. We consider two subcases below. * • Case 1A : We consider the case when $b_{2}=0$. If $b_{1}\in\\{0,1\\}$, then, there are $4p^{3}$ matrices. If $b_{1}\notin\\{0,1\\}$, then, there are $p^{3}(p-2)$ matrices. Thus, there are $p^{3}(p+2)$ in this case in total. * • Case 1B : Next consider the case when $b_{2}\neq 0$. From equation (5.2) we find that $a_{1}^{\prime}=0$. From (5.1), we deduce $a_{2}^{2}-a_{2}\equiv 0\mod p$. Hence, there are $2p^{3}(p-1)$ matrices in this case. Case 2 : In this case we count matrices such that $a_{3}\not\equiv 0\mod p$. From (5.2), we deduce that $b_{2}\neq 0$. Consider following subcases. * • Case 2A : Suppose $b_{1}=0$, then from equation (5.3) we get that $a_{2}\equiv 0\mod p$. Hence, there are $p^{2}(p-1)^{2}$ matrices in this case. * • Case 2B : Assume that $b_{1}=1$, then from equation (5.1) we find that $a_{2}(a_{2}-1)\equiv 0\mod p$. However, from (5.2) we cannot have $a_{2}\equiv 0\mod p$, therefore $a_{2}\equiv 1\mod p$. Hence, we find that $a_{3}\equiv b_{2}\mod p$. As a result, there are $p^{2}(p-1)$ matrices in this case. * • Case 2C : In the last case we consider the case when $b_{1}\notin\\{0,1\\}$ then from equations (5.1), (5.2), (5.3) we get that $a_{2}\not\equiv 0\mod p$. Now, we get that from the three equations (5.1), (5.2), (5.3) that, $a_{1}^{\prime}\equiv\frac{a_{2}(a_{2}-1)}{b_{1}(b_{1}-1)}\mod p,$ $a_{1}^{\prime}\equiv\frac{a_{3}^{2}}{b_{2}^{2}}\mod p,$ $a_{1}^{\prime}\equiv\frac{a_{2}a_{3}}{b_{1}b_{2}}\mod p.$ From these we find that $a_{2}\equiv b_{1}\mod p$ and $a_{3}\equiv b_{2}\mod p$. Hence, there are $(p-2)(p-1)p^{2}$ matrices in this case. Adding up the values in each case, we prove the assertion of the lemma. ∎ ###### Lemma 5.3. We have that $g_{\alpha}(p)=2p^{4}$ where $\alpha=(3,1,3,2).$ ###### Proof. Let $A$ be any integer matrix in Hermite normal form of the type $\begin{pmatrix}p^{3}&a_{1}p&a_{2}p&a_{3}p&1\\\ &p&0&0&1\\\ &&p^{3}&c_{1}p&1\\\ &&&p^{2}&1\\\ &&&&1\\\ \end{pmatrix}$ We derive the conditions so that $A$ is a subring matrix and use them to evaluate $g_{(3,1,3,2)}(p)$. 1. (1) It is clear that $v_{2}^{2}\in\mathrm{Col}(A)$ if and only if $v_{2}^{2}-pv_{2}\in\mathrm{Col}(A)$, the second condition translates to (5.4) $a_{1}^{2}-a_{1}\equiv 0\mod p.$ 2. (2) Similarly, $v_{3}^{2}\in\mathrm{Col}(A)$ if and only if $v_{3}^{2}-p^{3}v_{3}\in\mathrm{Col}(A)$, and therefore, $a_{2}\equiv 0\mod p$. In other words, $a_{2}=a_{2}^{\prime}p$, where $0\leq a_{2}^{\prime}\leq p-1$. 3. (3) Arguing as in previous two cases, we see that $v_{4}^{2}\in\mathrm{Col}(A)$ if and only if $c_{1}=c_{1}^{\prime}p$, where $0\leq c_{1}^{\prime}\leq p-1$ and $a_{3}=a_{3}^{\prime}p$ with $0\leq a_{3}^{\prime}\leq p-1$. 4. (4) It is easy to see that if the above conditions are satisfied, then $v_{2}v_{3},v_{2}v_{4}$ and $v_{3}v_{4}$ are in $\mathrm{Col}(A)$. To summarize, the conditions are as follows 1. (1) $a_{1}^{2}-a_{1}\equiv 0\mod p$, 2. (2) $a_{2}=a_{2}^{\prime}p$ where $0\leq a_{2}^{\prime}\leq p-1$, 3. (3) $c_{1}=c_{1}^{\prime}p$ where $0\leq c_{1}^{\prime}\leq p-1$, 4. (4) $a_{3}=a_{3}^{\prime}p$ with $0\leq a_{3}^{\prime}\leq p-1$. From the above, it is clear that $g_{(3,1,3,2)}(p)=2p^{4}$. ∎ ###### Lemma 5.4. We have that (5.5) $g_{(3,3,1,2)}(p)=8p^{4}-6p^{3}+2p^{2}.$ ###### Proof. Let $A$ be an integer matrix in Hermite normal form $\begin{pmatrix}p^{3}&a_{1}p&a_{2}p&a_{3}p&1\\\ &p^{3}&b_{1}p&b_{2}p&1\\\ &&p&0&1\\\ &&&p^{2}&1\\\ &&&&1.\\\ \end{pmatrix}$ We determine the conditions so that $A$ is a subring matrix. 1. (1) Arguing as in Lemma 5.3, we find that the condition $v_{2}^{2}\in\mathrm{Col}(A)$ translates to the condition that $a_{1}=a_{1}^{\prime}p$, with $0\leq a_{1}^{\prime}\leq p-1$. 2. (2) Next, we see that $v_{3}^{2}\in\mathrm{Col}(A)$ if and only if $v_{3}^{2}-pv_{3}\in\mathrm{Col}(A)$. Therefore we deduce that $b_{1}^{2}-b_{1}\equiv 0\mod p$ and (5.6) $(a_{2}^{2}-a_{2})p-a_{1}^{\prime}(b_{1}^{2}-b_{1})\equiv 0\mod p.$ 3. (3) It is again clear that $v_{4}^{2}\in\mathrm{Col}(A)$ if and only if $v_{4}^{2}-p^{2}v_{4}\in\mathrm{Col}(A)$. We deduce that this is true if and only if $b_{2}=b_{2}^{\prime}p$, where $0\leq b_{2}^{\prime}\leq p-1$, and $a_{3}=a_{3}^{\prime}p$, where $0\leq a_{3}^{\prime}\leq p-1$. 4. (4) It is clear that $v_{2}v_{3}$ and $v_{2}v_{4}$ are in $\mathrm{Col}(A)$ whenever above conditions are satisfied. By similar arguments as above we see that $v_{3}v_{4}\in\mathrm{Col}(A)$ if and only if $a_{1}^{\prime}b_{1}b_{2}^{\prime}\equiv 0\mod p$. We have the following conditions on $A$. 1. (1) $a_{1}=a_{1}^{\prime}p$ with $0\leq a_{1}^{\prime}\leq p-1$, 2. (2) $b_{1}^{2}-b_{1}\equiv 0\mod p$ and $(a_{2}^{2}-a_{2})p-a_{1}^{\prime}(b_{1}^{2}-b_{1})\equiv 0\mod p$, 3. (3) $b_{2}=b_{2}^{\prime}p$ where $0\leq b_{2}^{\prime}\leq p-1$, 4. (4) $a_{3}=a_{3}^{\prime}p$ where $0\leq a_{3}^{\prime}\leq p-1$, 5. (5) $a_{1}^{\prime}b_{1}b_{2}^{\prime}\equiv 0\mod p$. We consider three cases depending upon whether $b_{1}$ is $0$, $1$ or not in $\\{0,1\\}$, and count number of matrices in each case. Case 1 : First consider the case when $b_{1}=0$. Clearly there are $2p^{4}$ matrices in this case. Case 2 : Consider the case in which $b_{1}=1$. It is easy to see that there are $2p^{2}(2p-1)$ matrices in this case. Case 3 : Assume that $b_{1}\notin\\{0,1\\}$. Then if $b_{2}^{\prime}=0$ we get $2p^{3}(p-1)$ matrices. On the other hand, if $b_{2}\neq 0$ then $a_{1}^{\prime}=0$ follows from $a_{1}^{\prime}b_{2}^{\prime}\equiv 0\mod p$. Therefore, we find that $a_{2}^{2}-a_{2}\equiv 0\mod p$, and there are $4p^{2}(p-1)^{2}$ matrices. Adding up the number of matrices from each case we arrive at the assertion. ∎ ###### Lemma 5.5. The following equality holds (5.7) $g_{(3,2,3,1)}(p)=12p^{4}-2p^{3}+4p^{2}.$ ###### Proof. Let $A$ be any integer matrix in Hermite normal form of the type $\begin{pmatrix}p^{3}&a_{1}p&a_{2}p&a_{3}p&1\\\ &p^{2}&b_{1}p&b_{2}p&1\\\ &&p^{3}&c_{1}p&1\\\ &&&p&1\\\ &&&&1.\\\ \end{pmatrix}$ We determine the conditions so that $A$ is a subring matrix. 1. (1) It is easy to see that $v_{2}^{2}\in\mathrm{Col}(A)$ if and only if $a_{1}=a_{1}^{\prime}p$ where $0\leq a_{1}^{\prime}\leq p-1$. 2. (2) We note that $v_{3}^{2}$ is in $\mathrm{Col}(A)$ if and only if $v_{3}^{3}-p^{3}v_{3}\in\mathrm{Col}(A)$. Therefore, we require that $(a_{2}^{2}p^{2},b_{1}^{2}p^{2},p^{6},0,0)^{t}-p^{3}(a_{2}p,b_{1}p,p^{3},0,0)^{t}$ is in $\mathrm{Col}(A)$. Clearly, this is equivalent to the congruence (5.8) $a_{2}^{2}-a_{1}^{\prime}b_{1}^{2}\equiv 0\mod p.$ 3. (3) It is required that $v_{4}^{2}\in\mathrm{Col}(A)$. It is easy to see that this condition is equivalent to the condition that $v_{4}^{2}-pv_{4}\in\mathrm{Col}(A)$. The last entry in the column $v_{4}^{2}-pv_{4}$ is $(c_{1}^{2}-c_{1})p^{2}$ therefore, first condition is $c_{1}^{2}-c_{1}\equiv 0\mod p$. As in previous lemmas we obtain following second condition, $b_{1}(c_{1}^{2}-c_{1})\equiv 0\mod p^{2}$ and from this we deduce that $b_{1}(c_{1}^{2}-c_{1})=0$ and the third condition is, (5.9) $(a_{3}^{2}-a_{3})p^{2}-a_{2}(c_{1}^{2}-c_{1})-a_{1}^{\prime}(b_{2}^{2}-b_{2})p^{2}\equiv 0\mod p^{3}.$ 4. (4) If entries of matrix $A$ satisfy above conditions then it is easy to see that $v_{2}v_{3}$, $v_{2}v_{4}$ are in $\mathrm{Col}(A)$ and $v_{3}v_{4}\in\mathrm{Col}(A)$ if and only if the following congruence holds (5.10) $a_{2}(a_{3}-c_{1})-a_{1}^{\prime}b_{1}(b_{2}-c_{1})\equiv 0\mod p.$ Thus, the conditions so that $A$ is a subring matrix are as follows 1. (1) $a_{1}=a_{1}^{\prime}p$ with $0\leq a_{1}^{\prime}\leq p-1$, 2. (2) $c_{1}^{2}-c_{1}\equiv 0\mod p$ and $b_{1}(c_{1}^{2}-c-1)=0$, 3. (3) $(a_{3}^{2}-a_{3})p^{2}-a_{2}(c_{1}^{2}-c_{1})-a_{1}^{\prime}(b_{2}^{2}-b_{2})p^{2}\equiv 0\mod p^{3}$, 4. (4) $a_{2}(a_{3}-c_{1})-a_{1}^{\prime}b_{1}(b_{2}-c_{1})\equiv 0\mod p$. There are three cases to consider, depending upon the value of $c_{1}$. Case 1 : First suppose that $c_{1}=0$. Then by symmetry, the set of matrices is in bijection with the set of irreducible subring matrices with diagonal $(3,2,1,3)$. Therefore, we deduce that $g_{(3,2,1,3)}(p)=5p^{4}-4p^{3}+2p^{2}$ (from the argument in Lemma 5.2). Case 2 : Next, consider the case when $c_{1}=1$. The equation (5.9) reduces to the following (5.11) $(a_{3}^{2}-a_{3})-a_{1}^{\prime}(b_{2}^{2}-b_{2})\equiv 0\mod p.$ First we suppose that $b_{1}=0$. Then we find that $a_{2}\equiv 0\mod p$. Again considering two cases when $b_{2}\in\\{0,1\\}$ or $\notin\\{0,1\\}$ it is clear that there are $p^{3}(p+2)$ matrices in this case with $b_{1}=0$. If, $b_{1}\neq 0$, then we further divide into cases such that $b_{2}=0$, $1$ or not in $\\{0,1\\}$. It is easy to see that $2p^{2}(p-1)(2p-1)$ in this case with $b_{1}\neq 0$. Case 3 : Finally, we consider the case when $c_{1}\notin\\{0,1\\}$. From conditions above we find that $b_{1}=0$; from equation 5.9 we deduce that $a_{2}=0$ and equation 5.9 gets reduced to the equation 5.11. We consider sub- cases depending upon whether $b_{2}\in\\{0,1\\}$ or not. There are $8p^{3}$ matrices in the case when $b_{2}\in\\{0,1\\}$. On the other hand, there are $2p^{3}(p-2)$ matrices for which $b_{2}\notin\\{0,1\\}$. Adding up the number of matrices in each case we prove the result. ∎ ###### Lemma 5.6. The following equality holds $\begin{split}&g_{(3,1,2,3)}(p)=p^{4}+2p^{3}-p^{2},\\\ &g_{(3,3,2,1)}(p)=p^{6}+2p^{5}-13p^{4}+9p^{3},\\\ &g_{(3,2,2,2)}(p)=p^{6}-2p^{5}+6p^{4}-4p^{3}+3p^{2}-5p+2.\end{split}$ ###### Proof. The proof is very similar to previous results, and we omit it. ∎ ### 5.2. Compositions of length $5$ We consider the composition of length $5$ that begin with $3$ we need to evaluate $g_{\alpha}(p)$ for $10$ compositions of this form. We list all compositions of this form in the table below $(3,3,1,1,1)$ | $(3,2,1,1,2)$ ---|--- $(3,1,3,1,1)$ | $(3,1,2,1,2)$ $(3,1,1,3,1)$ | $(3,1,2,2,1)$ $(3,1,1,1,3)$ | $(3,1,1,2,2)$ $(3,2,1,2,1)$ | $(3,2,2,1,1)$. Table 2. Compositions of length $5$ that begin with $3$. For some of these compositions, we can use results from [Ish22a] to deduce value of $g_{\alpha}(p)$. The next result follows from the results in _loc. cit._ ###### Lemma 5.7. We have following values $g_{(3,3,1,1,1)}(p)$ | $16p^{6}+12p^{5}-20p^{4}+8p^{3}$ ---|--- $g_{(3,1,3,1,1)}(p)$ | $18p^{5}-6p^{4}$ $g_{(3,1,1,3,1)}(p)$ | $2p^{5}+10p^{4}-4p^{3}$ $g_{(3,1,1,1,3)}(p)$ | $4p^{4}$. ###### Proof. The above results are direct consequence of [Ish22a, Lemma 4.6]. ∎ Next we evaluate the value of $g_{\alpha}(p)$ for the remaining compositions. ###### Lemma 5.8. The following table of relations holds $g_{(3,1,2,2,1)}(p)$ | $13p^{5}-8p^{4}+8p^{3}-2p^{2}$ ---|--- $g_{(3,1,2,1,2)}(p)$ | $5p^{5}-4p^{4}+2p^{3}$ $g_{(3,1,1,2,2)}(p)$ | $p^{5}+3p^{4}-2p^{3}$ $g_{(3,2,1,1,2)}(p)$ | $4p^{6}-8p^{5}+35p^{4}-31p^{3}+4p^{2}$ $g_{(3,2,1,2,1)}(p)$ | $9p^{6}-4p^{5}+28p^{4}-39p^{3}+20p^{2}-10p$. ###### Proof. We will prove this result only for the composition $(3,1,1,2,2)$, and the proof is similar for the other compositions listed above. Therefore, we will count the number of integer subring matrices $A$ in Hermite normal form of the type $\begin{pmatrix}p^{3}&a_{1}p&a_{2}p&a_{3}p&a_{4}p&1\\\ &p&0&0&0&1\\\ &&p&0&0&1\\\ &&&p^{2}&b_{1}p&1\\\ &&&&p^{2}&1\\\ &&&&&1.\\\ \end{pmatrix}$ We determine the conditions on entries of the matrix which make $A$ a subring matrix. 1. (1) We note that $v_{2}^{2}$ and $v_{3}^{2}\in\mathrm{Col}(A)$ if and only if $a_{1}^{2}-a_{1}\equiv 0\mod p$ and $a_{2}^{2}-a_{2}\equiv\mod p$. 2. (2) We also want $v_{4}^{2}$ to be in $\mathrm{Col}(A)$ and this is true if and only $(a_{3}^{2}p^{2},0,0,p^{4},0,0)^{t}-p^{2}(a_{3}p,0,0,p^{2},0,0)^{t}$ is in $\mathrm{Col}(A)$. Therefore, $v_{4}^{2}\in\operatorname{Col}A$ if and only if $a_{3}=a_{3}^{\prime}p$ with $0\leq a_{3}^{\prime}\leq p-1$. 3. (3) Similarly, $v_{5}^{2}\in\mathrm{Col}(A)$ if and only if $v_{5}^{2}-p^{2}v_{5}\in\mathrm{Col}(A)$. This condition is equivalent to (5.12) $(a_{4}^{2}-a_{4})-a_{3}^{\prime}b_{1}^{2}\equiv 0\mod p.$ 4. (4) Similar arguments show that $v_{2}v_{4},v_{3}v_{4},v_{4}v_{5}$ are in $\mathrm{Col}(A)$ if above conditions are satisfied. Furthermore, $v_{2}v_{3},v_{2}v_{5},v_{3}v_{5}\in\mathrm{Col}(A)$ if and only if $a_{1}a_{2},a_{1}a_{4},a_{2}a_{4}$ are all divisible by $p$. We count the number of matrices satisfying the above conditions. It benefits us to consider two cases. Case 1: First, we consider the case when $b_{1}=0$. Note that in this case, at most one of $a_{1},a_{2},a_{4}$ is $1$ $\mod p$, and the others need to be $0$ $\mod p$. Hence, our count leads to $4p^{4}$ matrices for which $b_{1}=0$. Case 2: Next, we consider the case when $b_{1}\neq 0$. Note that in this case there are $p^{4}(p-1)$ matrices for which $a_{1},a_{2}$ are both divisible by $p$. On the other hand, either $a_{1}$ or $a_{2}$ is not divisible by $p$, then, $a_{4}$ is divisible by $p$. The equation 5.12 implies that $a_{3}^{\prime}=0$. Thus, there are $2p^{3}(p-1)$ matrices for which either $a_{1}$ or $a_{2}$ is divisible by $p$. Thus, the total number of matrices for this case is $p^{3}(p+2)(p-1)$. Putting together the number of matrices from both cases we deduce that $g_{(3,1,1,2,2)}(p)$ $=p^{5}+3p^{4}-2p^{3}$. ∎ There is one composition $\alpha$ for which we are not able to compute $g_{\alpha}(p)$ using the above arguments. This is the composition $\alpha=(3,2,2,1,1)$. Let us consider this case in some further detail. Let $A$ be an integer subring matrix of the form $A=\begin{pmatrix}p^{3}&a_{1}p&a_{2}p&a_{3}p&a_{4}p&1\\\ &p^{2}&b_{1}p&b_{2}p&b_{3}p&1\\\ &&p^{2}&c_{1}p&c_{2}p&1\\\ &&&p&0&1\\\ &&&&p&1\\\ &&&&&1.\\\ \end{pmatrix}$ Since the above matrix is in Hermite normal form, $1\leq a_{i}\leq p^{2}-1$, and $1\leq b_{i},c_{j}\leq p-1$. Below, we list the conditions for $A$ to be a subring matrix * • $a_{2}^{2}-a_{1}^{\prime}b_{1}^{2}\equiv 0\mod p$, * • $b_{1}c_{1}(c_{1}-1)=0$, $b_{1}c_{2}(c_{2}-1)=0$, $b_{1}c_{1}c_{2}=0$, * • $(a_{3}^{2}-a_{3})-a_{2}(c_{1}^{2}-c_{1})/p-a_{1}^{\prime}(b_{2}^{2}-b_{2})\equiv 0\mod p$, * • $(a_{4}^{2}-a_{4})-a_{2}(c_{2}^{2}-c_{2})/p-a_{1}^{\prime}(b_{3}^{2}-b_{3})\equiv 0\mod p$, * • $a_{2}(a_{3}-c_{1})-a_{1}^{\prime}b_{1}(b_{2}-c_{1})\equiv 0\mod p$, * • and $a_{2}(a_{4}-c_{2})-a_{1}^{\prime}b_{1}(b_{3}-c_{2})\equiv 0\mod p$, * • $a_{3}a_{4}-a_{2}c_{1}c_{2}/p-a_{1}^{\prime}b_{2}b_{3}\equiv 0\mod p$. In the subcase when $b_{1}=0$ counting solutions to the equations is reduced to counting number of solutions in $\mathbb{F}_{p}^{8}$ to following system of polynomial equations: $(x_{3}^{2}-x_{3})-x_{2}(x_{7}^{2}-x_{7})-x_{1}(x_{5}^{2}-x_{5})=0$ $(x_{4}^{2}-x_{4})-x_{2}(x_{8}^{2}-x_{8})-x_{1}(x_{6}^{2}-x_{6})=0$ $x_{3}x_{4}-x_{2}x_{7}x_{8}-x_{1}x_{5}x_{6}=0$ Let $N_{p}$ denote the number of solutions to this system in $\mathbb{F}^{8}_{p}$. Using SageMath we calculated $N_{p}$ for $p=2,3,5,7,11,13,17,19$ and these computations suggest that $N_{p}=p^{5}+12p^{4}-20p^{3}+30p^{2}-10p$. Using this and arguments similar to those used in proving previous results we deduce that $g_{(3,2,2,1,1)}(p)=p^{7}+24p^{6}-29p^{5}+21p^{4}-4p^{3}$. In summary, our computations lead us to make the following conjecture. ###### Conjecture 5.9. We have that $g_{(3,2,2,1,1)}(p)=p^{7}+24p^{6}-29p^{5}+21p^{4}-4p^{3}$. ## 6\. Calculating the values of $g_{\alpha}(p)$ for compositions beginning with $4,5$ or $6$ ### 6.1. Values of $g_{\alpha}(p)$ for compositions beginning with $4$ #### 6.1.1. Compositions of length $4$ We will evaluate $g_{\alpha}(p)$ for $6$ compositions of length $4$ that begin with $4$. They are listed below $(4,2,2,1)$ | $(4,2,1,2)$ ---|--- $(4,1,2,2)$ | $(4,3,1,1)$ $(4,1,3,1)$ | $(4,1,1,3)$. Table 3. Compositions of length $4$ that begin with $4$. ###### Lemma 6.1. We have following values, $g_{(4,2,2,1)}(p)$ | $6p^{6}+12p^{5}-10p^{4}-7p^{3}+4p^{2}$ ---|--- $g_{(4,2,1,2)}(p)$ | $p^{5}+9p^{4}-5p^{3}-4p^{2}$ $g_{(4,1,2,2)}(p)$ | $p^{5}+2p^{4}-p^{2}$ $g_{(4,3,1,1)}(p)$ | $2p^{6}-2p^{5}+17p^{4}-6p^{3}-2p^{2}$ $g_{(4,1,3,1)}(p)$ | $2p^{4}+6p^{3}$ $g_{(4,1,1,3)}(p)$ | $p^{4}+2p^{3}$. ###### Proof. The proof is omitted. ∎ #### 6.1.2. Compositions of length $5$ We will evaluate $g_{\alpha}(p)$ for $4$ compositions of length $5$ that begin with $4$. We list them below $(4,2,1,1,1)$ | $(4,1,2,1,1)$ ---|--- $(4,1,1,2,1)$ | $(4,1,1,1,2)$. Table 4. Compositions of length $5$ that begin with $4$. ###### Lemma 6.2. We have following values, $g_{(4,2,1,1,1)}(p)$ | $5p^{6}+16p^{5}-17p^{4}-27p^{3}+12p^{2}$ ---|--- $g_{(4,1,2,1,1)}(p)$ | $p^{6}+13p^{5}-8p^{4}+3p^{3}+2p^{2}$ $g_{(4,1,1,2,1)}(p)$ | $p^{6}+8p^{4}-4p^{3}$ $g_{(4,1,1,1,2)}(p)$ | $p^{5}+3p^{4}$. ###### Proof. We will explicitly obtain the expression for $g_{(4,1,1,1,2)}(p)$. The other computations are similar, and thus omitted. Let $A$ be a matrix in Hermite normal form of following type $\begin{pmatrix}p^{4}&a_{1}p&a_{2}p&a_{3}p&a_{4}p&1\\\ &p&0&0&0&1\\\ &&p&0&0&1\\\ &&&p&0&1\\\ &&&&p^{2}&1\\\ &&&&&1.\\\ \end{pmatrix}$ First, we determine the conditions on entries of the matrix which make $A$ a subring matrix. Note that for $i=2,3,4$, $v_{i}^{2}\in\mathrm{Col}(A)$ if and only if $a_{i-1}(a_{i-1}-1)\equiv 0\mod p^{2}$ and $v_{5}^{2}\in\mathrm{Col}(A)$ if and only $a_{4}\equiv 0\mod p$. We also note that for $1<i<j$, $v_{i}v_{j}\in\mathrm{Col}(A)$ if and only if $a_{i-1}a_{j-1}\equiv 0\mod p^{2}$. First we suppose that for $i=2,3,4$, we have that $a_{i-1}\equiv 0\mod p^{2}$. There are $p^{5}$ matrices of this form. Next, suppose for exactly one of $a_{1},a_{2},a_{3}$ is congruent to $1$ modulo $p^{2}$ and others are divisible by $p^{2}$ then we get $3p^{4}$ matrices of this form. Therefore, $g_{(4,1,1,1,2)}(p)$ = $p^{5}+3p^{4}$. ∎ ### 6.2. Values of $g_{\alpha}(p)$ for compositions beginning with $5$ #### 6.2.1. Compositions of length $4$ We calculate $g_{\alpha}(p)$ for $3$ compositions of length $4$ that begin with $5$. The compositions are listed below $(5,2,1,1)$ --- $(5,1,2,1)$ $(5,1,1,2)$. Table 5. Compositions of length $4$ that begin with $5$. ###### Lemma 6.3. The following table of relations hold $g_{(5,2,1,1)}(p)$ | $p^{6}-3p^{5}+17p^{4}-8p^{3}-5p^{2}+2p$ ---|--- $g_{(5,1,2,1)}(p)$ | $7p^{4}-p^{3}-2p^{2}$ $g_{(5,1,1,2)}(p)$ | $p^{4}+2p^{3}$. ###### Proof. We will prove the result only for the composition $(5,1,2,1)$. The other cases are omitted since the arguments are similar to this case. Let $A$ be a matrix in Hermite normal form of the following type $\begin{pmatrix}p^{5}&a_{1}p&a_{2}p&a_{3}p&1\\\ &p&0&0&1\\\ &&p^{2}&b_{1}p&1\\\ &&&p&1\\\ &&&&1.\\\ \end{pmatrix}$ We obtain the conditions on the entries of $A$ below. 1. (1) First, we must have that $v_{2}^{2}\in\mathrm{Col}(A)$ and this is true if and only if $v_{2}^{2}-pv_{2}\in\mathrm{Col}(A)$. The latter condition is equivalent to the condition $a_{1}(a_{1}-1)\equiv 0\mod p$. 2. (2) We also want that $v_{3}^{2}\in\mathrm{Col}(A)$ and we see that this is equivalent to the condition $v_{3}^{2}-p^{2}v_{3}\in\mathrm{Col}(A)$, which is true if and only if $a_{2}(a_{2}-p)\equiv\mod p$. Therefore we get that $a_{2}=a_{2}^{\prime}p$ with $0\leq a_{2}^{\prime}\leq p^{3}-1$ and $a_{2}^{\prime}(a_{2}^{\prime}-1)\equiv 0\mod p$. 3. (3) Using similar arguments we obtain that $v_{4}^{2}\in\mathrm{Col}(A)$ if and only if $a_{3}(a_{3}-1)-b_{1}(b_{1}-1)a_{2}^{\prime}\equiv 0\mod p^{3}$. We also deduce that $v_{2}v_{3},v_{2}v_{4}$ and $v_{3}v_{4}$ are in $\mathrm{Col}(A)$ if and only if $a_{1}a_{2}\equiv 0\mod p$, $a_{1}a_{3}\equiv 0\mod p$ and $a_{2}(a_{3}-b_{1})\equiv 0\mod p$. The conditions we get are as follows * • $a_{1}(a_{1}-1)\equiv 0\mod p$, * • $a_{2}=a_{2}^{\prime}p$ with $0\leq a_{2}^{\prime}\leq p^{3}-1$ and $a_{2}^{\prime}(a_{2}^{\prime}-1)\equiv 0\mod p$, * • $a_{3}(a_{3}-1)-b_{1}(b_{1}-1)a_{2}^{\prime}\equiv 0\mod p^{3}$, * • $a_{1}a_{2}\equiv 0\mod p$, $a_{1}a_{3}\equiv 0\mod p$ and $a_{2}(a_{3}-b_{1})\equiv 0\mod p$. Next, we count the number of matrices by dividing in two cases depending upon if $a_{1}\equiv 0\mod p^{3}$ or $a_{1}\equiv 0\mod p^{3}$. Case 1 : First consider the case when $a_{1}\equiv 1\mod p^{3}$, as we must have $a_{1}a_{2}\mod p^{3}$ we deduce that $a_{2}=a_{2}^{\prime\prime\prime}p^{3}$ and $a_{3}\equiv 0\mod p$. Therefore, conditions reduce to $b_{1}(b_{1}-1)a_{2}^{\prime\prime\prime}\equiv 0\mod p$. Counting we see that there are $3p^{3}-2p^{2}$ matrices in this case. Case 2 : Now, we consider that case when $a_{1}\equiv 0\mod p^{3}$. We further consider two separate cases depending upon whether $a_{2}^{\prime}$ is divisible by $p$ or not. * • Case 2A : Consider the case when $a_{2}^{\prime}$ is divisible by $p$. we have two possibilities, either $p^{3}\nmid a_{2}$ or $p^{3}\mid a_{2}$, if $p^{3}\nmid a_{2}$ then we deduce that $a_{3}\equiv b_{1}\mod p$, therefore $b_{1}=0$ or $1$. We get $2p^{3}(p-1)$ matrices. if $p^{3}\mid a_{2}$, again considering two cases depending upon whether $b_{1}\in\\{0,1\\}$ or not and counting we see that we get $2p^{4}$ matrices. * • Case 2B : We consider the case when $a_{2}^{\prime}\equiv 1\mod p$, we deduce that $a_{3}\equiv b_{1}\mod p^{3}$. Conditions reduce to $b_{1}(b_{1}-1)(1-a_{2}^{\prime})\equiv 0\mod p^{3}$. Now, suppose $b_{1}\in\\{0,1\\}$, then we get $2p^{4}$ matrices. Otherwise, $a_{2}^{\prime}=1$ and we get $p^{2}(p-2)$ matrices. It follows from the above observations that the number of matrices in each case we deduce that $g_{(5,1,2,1)}(p)=7p^{4}-p^{3}-2p^{2}$. ∎ #### 6.2.2. Compositions of length $5$ In this section we evaluate $g_{\alpha}(p)$ for the composition $\alpha=(5,1,1,1,1)$. This is the only composition of length $5$ which starts with $5$. We have the following result. ###### Lemma 6.4. We have that $g_{\alpha}(p)=5p^{4}$, where $\alpha=(5,1,1,1,1)$. ###### Proof. The result follows directly from [AKKM21, Lemma 3.5]. ∎ ### 6.3. Values of $g_{\alpha}(p)$ for compositions beginning with $6$ We evaluate $g_{\alpha}(p)$ for the only composition $\alpha=(6,1,1,1)$ of length $4$ which begins with $6$. ###### Lemma 6.5. We have that $g_{(6,1,1,1)}(p)=4p^{3}$. ###### Proof. The above result is an immediate consequence [AKKM21, Lemma 3.5]. ∎ ## 7\. Main results ### 7.1. Proof of the main result In this section we prove Theorem 1.4. We note that for $n>1$ and $e\geq n-1$ the following relation holds (7.1) $g_{n}(p^{e})=g_{n-1}(p^{e-1})+\sum_{\alpha\in\mathcal{C}^{\prime}_{n,e}}g_{\alpha}(p),$ where $\mathcal{C}^{\prime}_{n,e}$ denotes the set of compositions in $\mathcal{C}_{n,e}$ whose first coordinate is greater than $1$. ###### Proposition 7.1. We have that, $g_{5}(p^{9})=11p^{6}+14p^{5}+137p^{4}-16p^{3}+p^{2}+2p+3,$ and $g_{6}(p^{9})=4p^{7}+76p^{6}+128p^{5}+56p^{4}-111p^{3}+43p^{2}-9p+1+g_{(3,2,2,1,1)}(p).$ ###### Proof. The result follows immediately from (7.1) and computations in previous sections. ∎ Liu computes the values of $g_{n}(p^{e})$ for $n=3,4$ (cf. [Liu07, Proposition 6.1 and 6.2]). We record these values for $e=9$ and refer to [AKKM21, p. 231] the values of $e$ which lie in the range $4\leq e\leq 8$. We find that $g_{3}(p^{9})=p^{3}+4p^{2}+4p+1,$ and that $g_{4}(p^{9})=11p^{4}+30p^{3}+9p^{2}+p+1.$ ###### Proposition 7.2. For $n>1$, the following relation holds $\displaystyle f_{n}(p^{9})-f_{n-1}(p^{9})$ $\displaystyle=$ $\displaystyle Q(n)(276480P_{10}(n)p^{10}+276480P_{9}(n)p^{9}+138240P_{8}(n)p^{8}$ $\displaystyle+34560P_{7}(n)p^{7}+34560P_{6}(n)p^{6}+11520P_{5}(n)p^{5}$ $\displaystyle+11520P_{4}(n)p^{4}+1440P_{3}(n)p^{3}+1440P_{2}(n)p^{2}$ $\displaystyle+240P_{1}(n)p+P_{0}(n))+{n-1\choose 5}g_{(3,2,2,1,1)}(p),$ where, $\displaystyle Q(n)$ $\displaystyle=$ $\displaystyle\frac{n-1}{1393459200},$ $\displaystyle P_{10}(n)$ $\displaystyle=$ $\displaystyle(n-2)(n-3)(n-4)(n-5)(n-6)(n-7),$ $\displaystyle P_{9}(n)$ $\displaystyle=$ $\displaystyle n(n-2)(n-3)(n-4)(n-5)(n-6),$ $\displaystyle P_{8}(n)$ $\displaystyle=$ $\displaystyle(n-2)(n-3)(n-4)(n-5)(n-6)(9n^{2}-131n+784),$ $\displaystyle P_{7}(n)$ $\displaystyle=$ $\displaystyle(n-2)(n-3)(n-4)(n-5)(37n^{3}-761n^{2}+6482n-18144),$ $\displaystyle P_{6}(n)$ $\displaystyle=$ $\displaystyle(n-2)(n-3)(n-4)(75n^{5}-2468n^{4}+36349n^{3}-279672n^{2}$ $\displaystyle+1101372n-1732080).$ $\displaystyle P_{5}(n)$ $\displaystyle=$ $\displaystyle(n-2)(n-3)(n-4)(33n^{6}-1220n^{5}+21757n^{4}-208732n^{3}$ $\displaystyle+1053542n^{2}-2414076n+1592640),$ $\displaystyle P_{4}(n)$ $\displaystyle=$ $\displaystyle(n-2)(n-3)(42n^{8}-2102n^{7}+51106n^{6}-749135n^{5}$ $\displaystyle+7010050n^{4}-41819711n^{3}+152633830n^{2}$ $\displaystyle-307862664n+260890560),$ $\displaystyle P_{3}(n)$ $\displaystyle=$ $\displaystyle(n-2)(273n^{10}-18123n^{9}+571282n^{8}-10999886n^{7}$ $\displaystyle+140978185n^{6}-1241293667n^{5}+7531038196n^{4}$ $\displaystyle-30849468932n^{3}+81151302432n^{2}-123180801984n$ $\displaystyle+81613163520),$ $\displaystyle P_{2}(n)$ $\displaystyle=$ $\displaystyle(n-2)(147n^{11}-10843n^{10}+377832n^{9}-8095642n^{8}$ $\displaystyle+117343063n^{7}-1198590955n^{6}+8745674590n^{5}$ $\displaystyle-45324367680n^{4}+162663439888n^{3}-383226514464n^{2}$ $\displaystyle+531138427776n-326776343040),$ $\displaystyle P_{1}(n)$ $\displaystyle=$ $\displaystyle(n-2)(315n^{12}-25368n^{11}+956025n^{10}-22147762n^{9}$ $\displaystyle+349611873n^{8}-3946086924n^{7}+32542876455n^{6}$ $\displaystyle-196940866290n^{5}+865325849028n^{4}-2683887407672n^{3}$ $\displaystyle+5560471449216n^{2}-6887555482752n+3844971970560),$ $\displaystyle P_{0}(n)$ $\displaystyle=$ $\displaystyle 135n^{16}-14400n^{15}+730980n^{14}-23334360n^{13}+522513250n^{12}$ $\displaystyle-8678453720n^{11}+110319301164n^{10}-1092302298312n^{9}$ $\displaystyle+8494128305343n^{8}-51930963880392n^{7}+248190356069720n^{6}$ $\displaystyle-915171074718208n^{5}+2545385435375472n^{4}$ $\displaystyle-5146605000021888n^{3}+7110255039457536n^{2}$ $\displaystyle-5973926003435520n+2289569122713600.$ ###### Proof. Recall the recurrence relation (1.2) $f_{n}(p^{9})=\sum_{i=0}^{9}\sum_{j=1}^{n}{n-1\choose j-1}f_{n-j}(p^{9-i})g_{j}(p^{i}).$ Note that since $i\leq 9$, we find that $j\leq 10$ in the above recurrence relation. We compute the values of $g_{j}(p^{i})$ and then recursively, use the above to compute the value of $f_{n}(p^{9})$. In greater detail, we use values from [AKKM21, p. 231], the Theorem 1.2, and the values computed in previous sections, to obtain the above recurrence relation. ∎ Using the recurrence relation above and noting that $f_{n}(p^{9})=\sum_{k=2}^{n}\left(f_{k}(p^{9})-f_{k-1}(p^{9})\right)$ we obtain Theorem 1.4. We performed all the computations in SageMath. ### 7.2. Some bounds on $g_{\alpha}(p)$ In this section we obtain upper bounds on $g_{\alpha}(p)$ for certain compositions $\alpha$. Let $u_{1},\dots,u_{n}$ be the standard basis of $\mathbb{Z}^{n}$. The vector $u_{i}=(0,0,\dots,0,1,0,\dots,0)$ consists of $0$ in all entries except for $1$ in the $i$-th entry. Let $L$ be an irreducible subring of $\mathbb{Z}^{n}$ and let $\mathfrak{m}_{L}$ be the ideal in $L$ consisting of vectors all of whose coordinates are divisible by $p$. Setting $\rho_{L}\mathrel{\mathop{\mathchar 58\relax}}=\operatorname{dim}_{\mathbb{F}_{p}}\mathfrak{m}_{L}/\mathfrak{m}_{L}^{2}$. Following [Liu07, p. 292, l.-1], an irreducible subring is _full_ if $\rho_{L}=n-1$. Assume that the matrix for $L$ is in Hermite normal form with respect to the basis $u_{1},\dots,u_{n}$. Let $\pi\mathrel{\mathop{\mathchar 58\relax}}\mathbb{Z}^{n}\rightarrow\mathbb{Z}^{n-1}$ be the projection onto the last $(n-1)$ coordinates $u_{2},\dots,u_{n}$. Suppose that the basis is chosen so that with respect to the ordered basis $(u_{1},\dots,u_{n})$, the matrix $A$ associated with $L$ is in the Hermite normal form. Let $v_{1},\dots,v_{n}$ be the columns of $A$, and let $\tau(L)$ be the lattice generated by $L$ and by $v_{1}^{\prime}\mathrel{\mathop{\mathchar 58\relax}}=\frac{1}{p}v_{1}$. It is easy to see that $\tau(L)$ is a subring (cf. [Liu07, p.291, l.-2]). ###### Proposition 7.3. Let $L$ be an irreducible subring of index $p^{e+n}$ such that $\pi(L)$ is full and $pu_{1}\notin L$. Then the map $L\mapsto\frac{1}{p}\mathfrak{m}_{\tau(L)}$ is a $p^{n-2}$-to-one surjection onto subrings of index $p^{e}$. ###### Proof. The above result is [Liu07, Proposition 5.6]. ∎ Let $\beta\mathrel{\mathop{\mathchar 58\relax}}=(\beta_{1},\beta_{2},\dots,\beta_{n-1})$ be a composition of length $n-1$ such that $\beta_{1}>1$, $\beta_{i}>0$ for $i\in\\{2,\dots,n-1\\}$. By abuse of notation, we say that an irreducible subring $L$ has diagonal $\beta$ if the entries on the diagonal of the matrix $A_{L}$ are $p^{\beta_{1}},\dots,p^{\beta_{n-1}},1$. We set $\mathcal{R_{\beta}}$ to denote the set of irreducible subrings with diagonal $\beta$. Let $\beta^{\prime}\mathrel{\mathop{\mathchar 58\relax}}=(\beta_{1}-2,\beta_{2}-1,\dots,\beta_{n-1}-1)$, let $\mathcal{S}_{\beta^{\prime}}$ denote the set of subring matrices with diagonal $\beta^{\prime}$. Let $\mathcal{R}^{\prime}_{\beta}$ be the subset of $\mathcal{R}_{\beta}$ consisting of irreducible subrings with diagonal $\beta$ such that $\pi(L)$ is full and $pu_{1}\notin L$, with respect to the basis $u_{1},\dots,u_{n}$ for which $A_{L}$ is in Hermite normal form. Given a finite set $S$, set $\\#S$ to denote the cardinality of $S$. ###### Lemma 7.4. Let $\beta\mathrel{\mathop{\mathchar 58\relax}}=(\beta_{1},\beta_{2},\dots,\beta_{n-1})$ be a composition of length $n-1$ such that the following conditions are satisfied 1. (1) $\beta_{1}>1$, 2. (2) $\beta_{i}>0$ for $i\in\\{2,\dots,n-1\\}$, 3. (3) $\sum_{i=1}^{n-1}\beta_{i}>n$. Then, there is a $p^{n-2}$-to-one surjection from $\mathcal{R}_{\beta}^{\prime}$ to $\mathcal{S}_{\beta^{\prime}}$. In particular, we have that $\\#\mathcal{R}_{\beta}\geq\\#\mathcal{R}^{\prime}_{\beta}\geq p^{n-2}\\#\mathcal{S}_{\beta^{\prime}}.$ ###### Proof. It follows from Proposition 7.3 that the map $L\mapsto\frac{1}{p}\mathfrak{m}_{\tau(L)}$ defines a $p^{n-2}$ to one surjective map $\Phi\mathrel{\mathop{\mathchar 58\relax}}\mathcal{R}_{\beta}^{\prime}\rightarrow\mathcal{S}_{\beta^{\prime}}$. The inequality follows immediately from this. ∎ The following result comes as a consequence of the above Lemma. ###### Proposition 7.5. Let $\beta\mathrel{\mathop{\mathchar 58\relax}}=(\beta_{1},\beta_{2},\dots,\beta_{n-1})$ be a composition of length $n-1$ such that $\beta_{1}>1$, $\beta_{i}>0$ for $i\in\\{2,\dots,n-1\\}$ and $\sum_{i=1}^{n-1}\beta_{i}>n$. Setting $m_{\beta}\mathrel{\mathop{\mathchar 58\relax}}=\min\\{\lfloor\frac{\beta_{1}}{2}\rfloor,\beta_{2},\dots,\beta_{n-1}\\}$, we have the following lower bound for $g_{\beta}(p)$ $g_{\beta}(p)\geq p^{(n-2)m_{\beta}}.$ ###### Proof. The result follows upon repeatedly applying the Lemma 7.4. ∎ In particular we see that if $\alpha\mathrel{\mathop{\mathchar 58\relax}}=(k,\ell,\dots,\ell)$ of length $n-1$, then $g_{\alpha}(p)\geq p^{{\lfloor\frac{k}{2}\rfloor}(n-2)}$. We note that this is a special case of [Ish22b, Corollary 4.6]. ###### Theorem 7.6. We have the following polynomial lower bound $g_{n}(p^{e})\geq\sum_{j=1}^{\lfloor e/n\rfloor}p^{(n-2)j}\left(\binom{e-1-nj}{n-2}-\binom{e-1-n(j+1)}{n-2}\right).$ ###### Proof. Let $\mathcal{C}_{n,e}^{j}$ be the subset of $\mathcal{C}_{n,e}$ consisting of all compositions $\beta=(\beta_{1},\dots,\beta_{n-1})$ such that $m_{\beta}=j$. It is easy to see that $\\#\mathcal{C}_{n,e}^{j}=\\#\mathcal{C}_{n,e-nj}-\\#\mathcal{C}_{n,e-n(j+1)}=\binom{e-1-nj}{n-2}-\binom{e-1-n(j+1)}{n-2}.$ It follows from Proposition 7.5 that $g_{n}(p^{e})\geq\sum_{j=1}^{\lfloor e/n\rfloor}p^{(n-2)j}\\#\mathcal{C}_{n,e}^{j},$ and the result follows. ∎ We now prove Theorem 1.6. ###### Proof of Theorem 1.6. Let us assume for simplicity that $t=n-1$. The proof in generality is identical to this case. Given an $n\times n$ matrix $A$, let $A_{[i,j]}$ be the $i\times j$ matrix obtained upon deleting the last $(n-i)$ rows and $(n-j)$ columns. Thus, $A_{[i,j]}$ is the upper left $i\times j$ submatrix of $A$. Let $S$ denote the set of subring matrices of the form $A=\begin{pmatrix}p^{\alpha_{1}}&\dots&\dots&\dots&a_{1}p&1\\\ &p^{\alpha_{2}}&\dots&\dots&a_{2}p&1\\\ &&\ddots&&\vdots&\vdots\\\ &&&&\vdots&\vdots\\\ &&&&p^{k\alpha_{n-1}}&1\\\ &&&&&1\end{pmatrix}$ that are in Hermite normal form. We note that the following conditions hold 1. (1) $0\leq a_{i}\leq p^{\gamma-1}-1$, 2. (2) setting $\hat{A}\mathrel{\mathop{\mathchar 58\relax}}=A_{[n-2,n-2]}$, we note that $(a_{1}^{2}p^{2}-a_{1}p^{k\alpha_{n-1}+1},\dots,a_{n-2}^{2}p^{2}-a_{n-2}p^{k\alpha_{n-1}+1})^{t}\in\operatorname{Col}\hat{A}=\hat{A}(\mathbb{Z}^{n-2}).$ We note in passing that $g_{\alpha_{/k,t}}(p)=\\#S$. We denote the set of all irreducible subring matrices of size $(n-1)$ and with diagonal $(p^{\alpha_{1}},p^{\alpha_{2}},\dots,p^{\alpha_{n-2}},1)$ by $\mathcal{A}$. We note that for each index $i$ in the range $1\leq i\leq n-2$ we have that $|a_{i}^{2}p^{2}-a_{i}p^{k\alpha_{i}+1}|\leq|a_{i}^{2}|p^{2}+|a_{i}p^{k\alpha_{i}+1}|\leq 2p^{(k+1)\gamma}$. In order to simplify notation, for $B\in\mathcal{A}$, set $B^{\prime}\mathrel{\mathop{\mathchar 58\relax}}=B_{[n-2,n-2]}$. Setting $\mathcal{C}\mathrel{\mathop{\mathchar 58\relax}}=[-2p^{(k+1)\gamma},2p^{(k+1)\gamma}]^{n-2}$, we find that (7.2) $\\#S\leq\sum_{B\in\mathcal{A}}\\#\left(B^{\prime}(\mathbb{Z}^{n-2})\cap\mathcal{C}\right).$ For $B\in\mathcal{A}$, the number of points in $v\in\mathbb{Z}^{n-2}$ such that $B^{\prime}v\in\mathcal{C}$ is same as the number of points $w\in\mathcal{C}$ such that $\left(B^{\prime}\right)^{-1}(w)\in\mathbb{Z}^{n-2}$. Therefore using inequality 7.2 we deduce that (7.3) $\\#S\leq\sum_{B\in\mathcal{A}}\\#(\left(B^{\prime}\right)^{-1}(\mathcal{C})\cap\mathbb{Z}^{n-2}).$ For $w=(w_{1},\dots,w_{n-2})\in\mathbb{Z}^{n-2}$, set $\lVert w\rVert$ to be the Euclidean norm $\sqrt{\sum_{i}w_{i}^{2}}$. For any $w\in\mathcal{C}$ we have that $\lVert(B^{\prime})^{-1}w\rVert\leq\lVert(B^{\prime})^{-1}\rVert\lVert w\rVert$. It is easy to see that $\lVert w\rVert\leq 2p^{(k+1)\gamma}\sqrt{n-2}$. On the other hand, since $\mathcal{A}$ is a finite set which is defined independent of $k$, we find that for any $B\in\mathcal{A}$ and $w\in\mathcal{C}$, $\lVert(B^{\prime})^{-1}w\rVert\leq M\sqrt{n-2}p^{(k+1)\gamma}$ where $M\in\mathbb{R}_{>0}$ is a suitably large constant (not depending on $k$). Using the inequality (7.3), we deduce that (7.4) $\\#S\leq\sum_{L\in\mathcal{A}}\\#([-M\sqrt{n-2}p^{(k+1)\gamma},M\sqrt{n-2}p^{(k+1)\gamma}]^{n-2}\cap\mathbb{Z}^{n-2})$ Fxing an integer $N>M\sqrt{n-2}$, we deduce that (7.5) $\\#S\leq\sum_{L\in\mathcal{A}}\\#([-Np^{(k+1)\gamma},Np^{(k+1)\gamma}]^{n-2}\cap\mathbb{Z}^{n-2})$ Therefore, $\\#S\leq\\#(\mathcal{A})(2N)^{n-2}p^{(k+1)\gamma{(n-2)}}$. We have proved that $g_{\alpha_{/k,t}}(p)=\\#S\leq\\#(\mathcal{A})(2N)^{n-2}p^{\gamma{(n-2)}}p^{\gamma k{(n-2)}}$ Therefore, $g_{\alpha_{/k,t}}(p)=O(p^{\gamma k(n-2)})\text{ as }k\to\infty.$ ∎ ## References * [AKKM21] Stanislav Atanasov, Nathan Kaplan, Benjamin Krakoff, and Julia H Menzel. Counting finite index subrings of $\mathbb{Z}^{n}$. Acta Arithmetica, 197(3), 2021. * [Bra09] Johannes Franciscus Brakenhoff. Counting problems for number rings. PhD thesis, Leiden University, 2009. * [DSWP08] Marcus Du Sautoy, Luke Woodward, and D Phil. Zeta functions of groups and rings, volume 1925. Springer, 2008. * [DW88] Boris Datskovsky and David J Wright. Density of discriminants of cubic extensions. 1988\. * [Dwo62] Bernard Dwork. On the zeta function of a hypersurface. Publications Mathématiques de l’IHÉS, 12:5–68, 1962. * [GSS88] Fritz J Grunewald, Dan Segal, and Geoff C Smith. Subgroups of finite index in nilpotent groups. Inventiones mathematicae, 93(1):185–223, 1988. * [HRV08] Tamás Hausel and Fernando Rodriguez-Villegas. Mixed hodge polynomials of character varieties. Inventiones mathematicae, 174(3):555–624, 2008. * [Ish22a] Kelly Isham. Is the number of subrings of index $p^{e}$ in $\mathbb{Z}^{n}$ polynomial in $p$? arXiv preprint arXiv:2203.00646, 2022. * [Ish22b] Kelly Isham. Lower bounds for the number of subrings in $\mathbb{Z}^{n}$. Journal of Number Theory, 234, 2022. * [KMTB15] Nathan Kaplan, Jake Marcinek, and Ramin Takloo-Bighash. Distribution of orders in number fields. Research in the Mathematical Sciences, 2(1):1–57, 2015. * [Liu07] Ricky Ini Liu. Counting subrings of $\mathbb{Z}^{n}$ of index k. Journal of Combinatorial Theory, Series A, 114(2):278–299, 2007\. * [LS03] Alexander Lubotzky and Dan Segal. Subgroup growth, volume 212. Springer, 2003. * [Nak96] Jin Nakagawa. Orders of a quartic field, volume 583. American Mathematical Soc., 1996. * [Spa79] Kenneth W Spackman. Simultaneous solutions to diagonal equations over finite fields. Journal of Number Theory, 11(1):100–115, 1979.
# A Component of the Smith High Velocity Cloud Now Crossing the Galactic Plane Felix J. Lockman Green Bank Observatory, Green Bank, WV 24944, USA Robert A. Benjamin Dept. of Physics, University of Wisconsin-Whitewater, Whitewater, WI 53190 Nicolas Pichette Dept. of Physics, Montana State University, Bozeman, MT Christopher Thibodeau Current Address: 236 Braeden Brooke Dr. San Marcos, TX 78666 Department of Physics, Astronomy & Geosciences, Towson University, Towson, MD. ###### Abstract We have identified a new structure in the Milky Way: a leading component of the Smith high velocity cloud that is now crossing the Galactic plane near longitude $25\arcdeg$. Using new 21cm Hi data from the Green Bank Telescope (GBT) we measured the properties of several dozen clouds that are part of this structure. Their kinematics is consistent with that of the Smith Cloud with a $V_{\mathrm{LSR}}$ exceeding that permitted by circular rotation in their direction. Most of the clouds in the Leading Component show evidence that they are interacting with disk gas allowing the location of the interaction to be estimated. The Leading Component crosses the Galactic plane at a distance from the Sun of 9.5 kpc, about 4.5 kpc from the Galactic Center. Its Hi mass may be as high as $10^{6}$ $M_{\mathrm{\odot}}$, comparable to the mass of the neutral component of the Smith Cloud, but only a fraction of this is contained in clouds that are resolved in the GBT data. Like the Smith Cloud, the Leading Component appears to be adding mass and angular momentum to the ISM in the inner Galaxy. We suggest that the Smith Cloud is not an isolated object, but rather part of a structure that stretches more than $40\arcdeg$ ($\sim 7$ kpc) across the sky, in two pieces separated by a gap of $\sim 1$ kpc. ISM: clouds – ISM: individual objects (Smith Cloud) – Galaxy: structure – Galaxy: evolution ††facilities: GBT ## 1 Introduction The old puzzle of the Milky Way’s high velocity clouds – those aggregations of neutral and ionized gas whose velocities deviate significantly from that allowed by Galactic rotation – continues to confound and delight. The origin of most high velocity clouds (HVCs) is uncertain. Because the category is defined solely on kinematics (e.g., Wakker & van Woerden, 1991), it can encompass everything from gas stripped from satellite galaxies, accretion from the intergalactic and circumgalactic medium, and material recycled from the Galactic disk (Oort, 1969; Wakker & van Woerden, 1997; Marinacci et al., 2010; Binney & Fraternali, 2012; Putman et al., 2012). HVCs cover a large fraction of the sky (Lockman et al., 2002; Lehner et al., 2022). The large HVC complexes lie in the Galactic halo and are typically found within 10 kpc of the Galactic plane (Putman, 2017). HVCs have now been detected in other galaxies, e.g., M31 (Braun & Thilker, 2004; Thilker et al., 2004; Westmeier et al., 2008; Wolfe et al., 2016), though not with the sensitivity and richness of detail available to studies of the Milky Way. HVCs are important in understanding the evolution of galaxies as they can be manifestations of both gas accretion and recycling from a galactic fountain (Putman, 2017; Richter, 2017; Fraternali, 2017; Li & Tonnesen, 2020). A number of HVCs have a cometary or head-tail morphology or other indications that they are interacting with an external medium (Putman et al., 2011; Barger et al., 2020). There is also evidence for the interaction of infalling HVCs with neutral gas in the Galactic plane (Lockman, 2003; McClure-Griffiths et al., 2008; Park et al., 2016). These clouds must have survived their passage through the halo to merge with the Milky Way disk. The Smith Cloud (catalog ) is one of the most prominent high-velocity clouds; it was discovered even before the identification of HVCs as a class of interstellar object (Smith, 1963; Muller et al., 1963). In neutral hydrogen emission the Smith Cloud extends over more than $10^{\circ}$ on the sky with a highly-organized cometary shape. It has a magnetic field draped around it (Hill et al., 2013; Betti et al., 2019). The Cloud contains at least $10^{6}$ $M_{\mathrm{\odot}}$ of Hi and probably an equal mass in ionized gas but has no prominent stellar counterpart (Lockman et al., 2008; Hill et al., 2009; Stark et al., 2015). Its metallicity has been determined through optical emission and UV absorption spectroscopy to be approximately half Solar (Hill et al., 2009; Fox et al., 2016), but this may reflect more the medium with which it has exchanged material than any initial property of the cloud (Gritton et al., 2014; Henley et al., 2017; Heitsch et al., 2022). The distance to the Smith Cloud derived from three different methods is 12.4 kpc, placing it 2.9 kpc below the Galactic Plane (Putman et al., 2003; Wakker et al., 2008; Lockman et al., 2008). The kinematics of the cloud has been analyzed by Lockman et al. (2008, hereafter L08) who suggest that it is on a course to collide with the Milky Way disk in $\approx 30$ Myr, adding mass and angular momentum to the Galaxy, and possibly triggering a burst of star formation (Alig et al., 2018). The Smith Cloud has been modelled as the gas remnant of a dwarf galaxy merging with the Milky Way (Bland-Hawthorn et al., 1998), the baryonic component of a dark matter subhalo (Nichols & Bland- Hawthorn, 2009; Nichols et al., 2014; Tepper-García & Bland-Hawthorn, 2018), and the product of supernove or the Galactic fountain (Sofue et al., 2004; Marasco & Fraternali, 2017). This is the first in a series of papers exploring aspects of the Smith Cloud. Future papers will report on studies of its structure, interaction with the circumgalactic medium, molecular content, and trajectory. Here we describe a major interstellar feature associated with the Smith Cloud: a leading component that is now passing through the Galactic disk. HVCs are easily detected at high Galactic latitude where their discrepant velocities stand out most clearly from Galactic disk gas. There is no reason to believe, however, that the phenomenon is not widespread, and that a significant fraction of the HVC population has been missed through confusion with gas at permitted velocities (Zheng et al., 2015). Indeed, distinctions based solely on velocity certainly give a biased view of phenomena in the Galactic halo (Bish et al., 2021; Marasco et al., 2022). In a search for potential HVCs at low Galactic latitude, we constructed a slice through the LAB Hi survey (Kalberla et al., 2005) in the first Galactic quadrant, integrating over a velocity interval 15-35 $\rm km\,s^{-1}$ greater than expected for circular Galactic rotation for a reasonable Galactic rotation curve. In the inner Galaxy, the maximum value of $|{V_{\mathrm{LSR}}}|$ from circular rotation occurs at the tangent point, where the distance from the Sun, r, is $R_{0}cos(\ell)$, and $R_{0}$ is the distance to the Galactic Center. By sampling emission from velocities slightly greater than allowed by circular rotation we are measuring gas in the line wing, which in the absence of non-circular motions should arise from regions close to the tangent point. In this instance we used the rotation curve function given by Brand & Blitz (1993) with coefficients from Burton (1988), although as we will show, the basic findings are not sensitive to details of the adopted rotation curve. The results over a $40^{\circ}\times 40^{\circ}$ region are shown in Figure 1. The Hi emission visible in this figure arises from three separate sources. First, because of temperature and turbulence there will always be some emission in the line wings at velocities extending beyond that allowed by Galactic rotation. The amount of emission in the line wings depends on the velocity dispersion of the gas, but also on $|d{V_{\mathrm{LSR}}}/dr|^{-1}$ – often called velocity crowding – which is the degree to which the velocity of gas at different distances from the Sun, r, projects to a difference in $V_{\mathrm{LSR}}$ (Burton, 1971; Celnik et al., 1979). The sense of this geometric effect is that $V_{\mathrm{LSR}}$ changes more slowly with distance at high longitudes than at low longitudes, so more emission appears in line wings at a forbidden velocity at $\ell=50\arcdeg$ than at $\ell=30\arcdeg$. The dashed yellow lines above and below latitude $0\arcdeg$ in Figure 1 mark a distance $\pm 0.25$ kpc from the plane at the tangent points and outline expectations for the contribution from line wings and velocity crowding. A second factor affecting the appearance of Figure 1 is large-scale streaming motions that arise from, for example, gas response to spiral arms (Burton, 1971). Although these are easily detected in the data (e.g., Fig. 8 in McClure-Griffiths & Dickey (2007) and Figs. 6 and 7 in Levine et al. (2008)), they are not included in rotation curves that describe only the symmetric component. Large-scale streaming contributes to the emission observed in Figure 1 around longitude $32\arcdeg$ and $50\arcdeg$. Note that both turbulence and streaming motions are expected to be approximately symmetric around the Galactic Plane, and that is what is observed. But Figure 1 also shows emission not related to Galactic rotation in any sense. The Smith HVC dominates latitudes $b\lesssim-10{{}^{\circ}}$ between longitudes $35^{\circ}\leq\ell\leq 50^{\circ}$, and there is a component in Hi emission almost as extensive as the Smith Cloud that lies along its axis but extends through the Galactic plane, crossing it around longitude $\ell=25\arcdeg$. A dashed cyan line connects these two features, which are clearly not symmetric about the Galactic Plane. The existence of a band of emission at forbidden velocity, aligned with the Smith Cloud, suggests that both may be part of a larger structure that shares a common kinematic anomaly. Because the morphology of the Smith Cloud indicates that its space motion is towards lower longitude and more positive latitude (L08), i.e., to the upper right in Figure 1, we call the component that extends through the Galactic plane the “Leading Component”. To understand its properties more fully we analyze new 21cm Hi observations from the Green Bank Telescope (GBT) made over a large region for another purpose, and here present those data and the results. Figure 1: Map of the integrated Hi brightness temperature along a slice through the LAB survey that includes only emission between 15 and 35 $\rm km\,s^{-1}$ in excess of that allowed by circular Galactic rotation. Dashed yellow lines mark a distance $z=\pm 0.25$ kpc from the plane at the tangent point distance of each longitude. Normal thermal and turbulent motions, along with velocity crowding, will produce forbidden-velocity emission in this range whose brightness is expected to increase to higher longitude over areas approximately bound by curves of constant $|z|$. The Smith Cloud is in the lower left between longitudes $40\arcdeg\lesssim\ell\lesssim 50\arcdeg$ centered around latitude $b\approx-15\arcdeg$. This Figure shows that there is also emission that extends through the Galactic plane along the axis of the Cloud, marked by the diagonal dashed cyan line. We call this the Leading Component. The grey scale is proportional to the square root of the integrated Hi brightness temperature. In section 2 the new GBT Hi survey is described, as well as an existing study that provides high resolution images of several clouds in the Leading Component. A catalog of Hi clouds found in the Leading Component in presented in section 3 along with a discussion of their general properties. Section 4 discusses evidence that some of these leading clouds are interacting with normally rotating disk gas, allowing derivation of a kinematic distance. Their overall space velocity is analyzed in section 5, which is followed by a discussion of the implications of the present work. ## 2 Observations ### 2.1 The GBT Hi Survey Figure 2: Longitude-latitude maps from the new GBT Hi survey. Coverage is almost complete over $13\arcdeg\leq\ell\leq 35\arcdeg$ and $-10\arcdeg\leq b\leq+10\arcdeg$. The new data show the tip of the Smith Cloud in the lower left of the left panel. The Leading Component stands out as a band of emission with the same position angle as the Smith Cloud, extending from $(\ell,b)=(32\arcdeg,-7\arcdeg)$ to $(22\arcdeg,+4\arcdeg)$. Black rectangles mark areas blanked in the GBT data. Left Panel: The integrated Hi emission over $+226\leq{V_{\mathrm{GSR}}}\ \leq+275\ {\rm km\,s^{-1}}$. Right Panel: Channel map at ${V_{\mathrm{GSR}}}=243.8\ {\rm km\,s^{-1}}$ over the central region of the GBT survey. Here we see that some of the Leading Component is resolved into individual clouds. Measurements of the 21cm line of Hi were made with the Green Bank Telescope (GBT) using the standard L band receiver and the GBT spectrometer (Prestage et al., 2009). Data were taken ”on-the-fly” using in-band frequency switching and were calibrated as described in Boothroyd et al. (2011). A low order polynomial was fit to emission-free regions of the spectra after they were gridded into a cube with a channel spacing of 0.80 $\rm km\,s^{-1}$ over a velocity range of $-200\leq{\rm V_{LSR}}\leq+275$ $\rm km\,s^{-1}$. The data cover $13^{\circ}\leq\ell\ \leq 35^{\circ}$ over a latitude range $\pm 10^{\circ}$ with a partial extension to latitude $-14^{\circ}$ at $\ell\geq 30^{\circ}$ to include the tip of the Smith Cloud. The effective angular resolution is $10\arcmin$. The median rms noise in brightness temperature for a 1 $\rm km\,s^{-1}$ channel is 0.12 K. The noise varies somewhat across the mapped area, being greatest within $2^{\circ}$ of the Galactic plane where there was often a significant increase in the system temperature from continuum emission. A small fraction of the mapped area had to be blanked over regions of strong continuum emission or because of occasional radio frequency interference. For study of the Leading Component, the new data have a distinct advantage over the premier existing Hi survey, HI4PI (HI4PI Collaboration et al., 2016). Emission from the leading component is not particularly faint, so although HI4PI has about one-third lower noise level than the new GBT data, noise is not a critical factor. The $10\arcmin$ angular resolution of the GBT data compared with the $16\farcm 2$ of HI4PI is, however, important in sorting out confusion between the leading component and other Hi emission. Virtually all of the Leading Component lies at J2000 declination $\delta<-5{{}^{\circ}}$, which is outside the declination limit of any survey using the Arecibo telescope. The new GBT survey is shown in Figure 2 as emission integrated over forbidden velocities (left), i.e., those velocities not permitted by circular Galactic rotation, and as a map of emission in a single velocity channel (right). The left panel shows the tip of the Smith Cloud at the lower left and the Leading Component in the center of the Figure. The velocity with respect to the Local Standard of Rest, $V_{\mathrm{LSR}}$, is not appropriate for the study of objects that cover a large angle on the sky, as the changing projection of the LSR across the structure becomes significant. A more suitable measure is the velocity with respect to the Galactic Standard of Rest defined as ${V_{\mathrm{GSR}}}={V_{\mathrm{LSR}}}+|V_{0}|sin(\ell)cos(b)$, where V0 is the circular velocity of Galactic Rotation at the Sun’s distance from the Galactic Center, ${\rm R_{0}}$. Throughout this paper we use ${\rm R_{0}}=8.1$ kpc and ${\rm V_{0}=230}\ {\rm km\,s^{-1}}$, consistent with the results of recent studies (Gravity Collaboration et al., 2018; Eilers et al., 2019). The right panel of Figure 2 is a channel map over the central part of the new GBT survey at a constant $V_{\mathrm{GSR}}$ of 243.8 $\rm km\,s^{-1}$, showing that the Leading Component contains discrete clouds. ### 2.2 High Angular Resolution Measurements Pidopryhora et al. (2015, hereafter P2015) combined GBT and VLA data to produce Hi maps of 10 interstellar clouds at $\sim 1\arcmin$ angular resolution, corresponding to a linear resolution of 3 pc at a distance of 10 kpc. The clouds cover a range of longitude ($16\fdg 0\leq\ell\ \leq\ 44\fdg 8$) and latitude ($-8\fdg 7\ \leq b\leq+4\fdg 5$). The sample was selected to have little confusion with other neutral Hi, so these clouds are somewhat outside of the Galactic plane (median $|b|=4\fdg 9)$ and have a velocity beyond the terminal velocity (${V_{\mathrm{LSR}}}\geq\ V_{t}$), where the terminal velocity is the maximum permitted by Galactic rotation in their direction. Three of the P2015 clouds turn out to be clouds in the Leading Component. Although the noise in the P2015 data is about three times worse than the GBT data (approximately 0.36 K in a 1 $\rm km\,s^{-1}$ channel) its much higher angular resolution reveals important structure within the clouds. ## 3 Leading Clouds Figure 3: GBT Hi spectra showing examples of the clouds detected in this survey associated with the Leading Component of the Smith Cloud identified by their $\ell,b$ coordinates. The leading cloud is always the highest velocity spectral feature, marked with the solid arrow. All these clouds appear to be interacting with a component of disk gas at a lower (permitted) velocity, marked with a dashed arrow. This is discussed in more detail in section 4. Figure 4: GBT Hi spectra of Leading Component clouds that are confused with lower velocity emission (arrows). The longitude and latitude is indicated in each panel. These spectral features are distinct clouds in channel maps and have a $V_{\mathrm{LSR}}$ exceeding that permitted by Galactic rotation, but cannot be separated from lower velocity emission and, like others with similar properties, are therefore not included in Table 1. ### 3.1 Cloud Properties At the angular resolution of the GBT data the Leading Component contains a number of individual clouds many of which are isolated in $\ell$, b, and velocity, so that their properties can be measured accurately. The right panel of Figure 2 shows a region of the survey centered near $(\ell,b)=(27\arcdeg,-2\arcdeg)$ in which a number of individual clouds can be identified. The clouds lie in a swath that begins a few degrees above the top of the Smith Cloud around $(\ell,b)=(35\arcdeg,-8\arcdeg)$ and extends at least to $b>+4{{}^{\circ}}$ at $\ell<24{{}^{\circ}}$. Discrete clouds associated with the leading component were identified using three criteria: 1) a location on or near the main region of the leading component; 2) a $V_{\mathrm{LSR}}$ that was close to or beyond that permitted by circular Galactic rotation in their direction; and, 3) for most of the clouds, evidence that they were interacting with gas at a lower, permitted velocity. Spectra towards 4 of the leading clouds are shown in Figure 3. They are detected well above the noise in the GBT data. There are also cloud-like structures visible in channel maps whose spectral components cannot be separated from lower velocity emission. This is especially common close to the Galactic plane. Examples are shown in Figure 4. These are not included in Table 1. Clouds were identified in channel maps and a Gaussian was fit to the spectrum at the location of the peak $N_{\mathrm{{\rm HI}}}$. A 2-d Gaussian model was used to derive an angular size and position angle. Hi masses were measured by integrating over an area of the channel map encompassing the measurable extent of the cloud at the velocity of the Hi peak. Background Hi emission was estimated from measurements around the edge of the cloud and subtracted. The result was then multiplied by the FWHM of the line at the brightest location to obtain a line power integrated over area and velocity. Some clouds were so confused with other emission that it was not possible to measure their size. Properties of 38 clouds are given in Table 1. We emphasize that the sample of clouds presented here does not completely account for all the clouds or all the Hi mass in the leading component; it represents only the more easily- identifiable and less-confused clouds in the GBT data. The longitude and latitude of the peak Hi brightness is in the first two columns, and quantities in cols. 3 through 6 refer to that location. Errors in line parameters were derived from the Gaussian fit to the spectrum. The median peak $T_{b}$ of the clouds is 2.5 K which is a detection at the $20\sigma$ level above the noise. The faintest cloud is detected at $6.5\sigma$. The angular size in col. 7 is not corrected for beam smearing. The position angle is measured counter- clockwise with reference to the Galactic North Pole such that a cloud elongated parallel to the Galactic plane would have PA $=90\arcdeg$. For the purposes of calculating general cloud properties we assume that they are all at a distance $d_{10}=10$ kpc. This is comparable to the distance of the Smith Cloud, 12.4 kpc (L08), and results can be scaled easily. The distance of the Leading Component is discussed further in section 4. Figure 5: Properties of the leading clouds. Top left: $N_{\mathrm{{\rm HI}}}$ measured at the position of peak line brightness assuming optically thin emission. Top right: Cloud diameter deconvolved for beam smearing assuming a 10 kpc distance. Center left: Hi mass of the leading clouds assuming a distance of 10 kpc. The median $M_{{\rm HI}}=1.6\times 10^{3}\ \ (d_{10}^{2})\ {M_{\mathrm{\odot}}}$. Center right: Estimate of the average volume density in the clouds from the ratio of the peak column density to the diameter. The diameter is corrected for beam convolution and scaled to a distance $d_{10}=10$ kpc. The median $n_{HI}=0.43\ {\rm cm^{-3}}$. Bottom: Distribution of position angles for the leading clouds. Each cloud is displayed twice, once at its PA and once at its PA+180∘. The two vertical arrows mark the PA of the Smith Cloud, $130^{\circ}$. The data suggest that many individual clouds share the same elongation as the Smith Cloud. Properties of the leading cloud population are summarized in Table 2. The measured angular size of the clouds, $\Theta_{obs}$, was corrected for beam convolution by subtracting the $10\arcmin$ resolution of the data cube in quadrature from the measured $\Theta_{maj}$ and $\Theta_{min}$. Cloud diameters were derived from the square root of the product of the major and minor deconvolved angular sizes, at a distance of 10 kpc. They are shown in the upper right panel of Figure 5. The cloud Hi mass distribution is shown in Figure 5 for an assumed distance of 10 kpc. The sum of the Hi mass of the tabulated clouds is $1.1\times 10^{5}$ $M_{\mathrm{\odot}}$. The Hi mass in the entire leading component was estimated by summing over a large region encompassing the leading component, and integrating over all ${V_{\mathrm{GSR}}}\geq 226$ $\rm km\,s^{-1}$. The result, $8\times 10^{5}$ $M_{\mathrm{\odot}}$ for a distance of 10 kpc, has a large uncertainly, but implies that the tabulated clouds contain only a small fraction ($\sim 15\%$) of the total Hi associated with the system. A lower limit to the volume density of the clouds can be derived from the ratio of the peak column density, $N_{\mathrm{{\rm HI}}}$, to the cloud diameter. This is a lower limit because the observed $N_{\mathrm{{\rm HI}}}$ is certainly reduced to some extent by beam convolution, while this effect was taken out, at least approximately, in calculating the diameter. The distribution of cloud position angles is shown in the bottom panel of Figure 5, where to avoid artificial discontinuities at $180^{\circ}$ each cloud is counted twice, once at its PA and once at its PA+180∘. The distribution peaks around the PA of the Smith Cloud, which is also the PA of the Leading Component, $130^{\circ}$, suggesting that some individual clouds are being shaped by the same interaction with their environment as the Smith Cloud itself. ### 3.2 Comparison with higher angular resolution observations Three of the leading clouds from Table 1 were observed in the P2015 survey and we can compare properties derived from GBT data with those from the higher angular resolution data. The results are given in Table 3, where P2015 diameters have been scaled to a distance of 10 kpc. Velocities are in good agreement, while the higher angular resolution of P2015 gives larger peak values of $N_{\mathrm{{\rm HI}}}$ by a factor of 2 - 3. The higher resolution data often reveal cloud cores with a diameter less than one-third that given by the GBT data. The increase in peak $N_{\mathrm{{\rm HI}}}$ and smaller cloud core size combine to increase the central cloud density by factors of more than six. It is interesting that these clouds are elongated along the direction of the Leading Component axis, with position angles clustered around the $130\arcdeg$ PA of the Smith Cloud. ### 3.3 Leading Cloud Kinematics Values of $V_{\mathrm{LSR}}$ and $V_{\mathrm{GSR}}$ for the clouds are shown in a position-velocity diagram in Figure 6. The solid line is at the median $V_{\mathrm{GSR}}$ of the sample, 237 $\rm km\,s^{-1}$, while the mean $V_{\mathrm{GSR}}$ is identical at $237.1\pm 6.8$ $\rm km\,s^{-1}$($1\sigma)$. The $V_{\mathrm{GSR}}$ of the central region of the Smith Cloud, marked by the dashed line, is essentially identical to that of the leading clouds at 241 $\rm km\,s^{-1}$. The dotted line at 226 $\rm km\,s^{-1}$ marks the lower velocity limit of the integrated emission displayed in the left panel of Figure 2, which was used to estimate the Hi mass of the entire Leading Component system. As we will discuss in section 5, the quantity $V_{\mathrm{GSR}}$ is directly related to the total space velocity of an object. The combination of similar $V_{\mathrm{GSR}}$ and an apparent association on the sky strongly supports a physical connection between the Leading Component and the Smith Cloud. From a rotation curve which gives $V_{\theta}(R)$, a terminal (maximum) velocity can be calculated for every longitude $|\ell|<90\arcdeg$: $Vt\equiv cos(b)\ [V_{\theta}(R_{0})sin(\ell)-V_{\theta}(R_{t})]$, where $R_{t}=R_{0}\ sin(\ell).$ Each cloud can then be assigned a ”forbidden” velocity, ${V_{\mathrm{for}}}\equiv{V_{\mathrm{LSR}}}-V_{t}$, which gives the amount by which the cloud’s $V_{\mathrm{LSR}}$ is in excess of that allowed by circular Galactic rotation at any location along that line of sight. We use two rotation curves: a flat rotation curve with $V_{\theta}(R)\equiv V_{0}=230\ {\rm km\,s^{-1}}$, which is very close to the curve measured by Eilers et al. (2019) at $R\geq 5.27$ kpc, and a ”Universal” rotation curve (Persic et al., 1996) with coefficients from Mróz et al. (2019), that we call the PM model. The PM model differs from that of Reid et al. (2014) by only a few percent over the region of interest. Figure 6: The leading clouds in position and velocity. The $V_{\mathrm{LSR}}$ (blue dots below) and $V_{\mathrm{GSR}}$ (red stars above) of the leading cloud sample is plotted vs. longitude. The solid horizontal line is the median $V_{\mathrm{GSR}}$ of the clouds, 237 $\rm km\,s^{-1}$, while the dashed line is the $V_{\mathrm{GSR}}$ of the Smith Cloud, 241 $\rm km\,s^{-1}$. There is no strong trend in $V_{\mathrm{GSR}}$ with longitude implying that the entire complex is consistent with having a single velocity vector projected to different values of $V_{\mathrm{LSR}}$ at different locations. The dotted line shows the lower limit to the $V_{\mathrm{GSR}}$ used in integrating over velocity to create the moment-zero map in the left panel of Figure 2. Figure 7 shows the ”forbidden velocities” $V_{\mathrm{for}}$ for the cloud sample plotted against longitude from the PM rotation curve on the left and the flat rotation curve on the right. The median value of $V_{\mathrm{for}}$ is 17.3 and 8.1 $\rm km\,s^{-1}$, respectively. Forbidden velocities are larger in the PM model because, unlike the flat model, the PM rotational velocities decrease toward the Galactic Center reducing the maximum permitted $V_{\mathrm{LSR}}$ at lower longitudes. When analyzed with a flat rotation curve there are five clouds with ${V_{\mathrm{for}}}\lesssim 0\ {\rm km\,s^{-1}}$, but these all show signs of interaction with lower-velocity gas implying that they have a high non-circular velocity component (discussed in section 4). In Figure 7 clouds with evidence of interaction are marked with a red star. Figure 7: The ”forbidden velocity” of the cloud sample, $V_{\mathrm{for}}$, defined as the difference between a cloud’s $V_{\mathrm{LSR}}$ and the maximum permitted by circular Galactic rotation in their direction (blue dots). This is a lower limit to the peculiar line-of-sight velocity of a cloud. A more realistic estimate of a cloud’s deviation from circular rotation is given by the quantity $V_{\mathrm{dif}}$, discussed in section 4. If the clouds show signs of interaction with lower velocity gas the points are overplotted with red stars. Left panel: Forbidden velocities evaluated for the Persic ”Universal” rotation curve (Persic et al., 1996) with coefficients from Mróz et al. (2019). Right Panel: Forbidden velocities for a flat rotation curve at a circular velocity of 230 $\rm km\,s^{-1}$. While there are 5 clouds with ${V_{\mathrm{for}}}\lesssim 0$ for a flat rotation curve, values for the more realistic Persic curve, and evidence for interaction, suggests that they have anomalous velocities compared to their location in the Galaxy and are properly included in the leading cloud sample. ## 4 Interaction with Disk Gas Most of the clouds appear to be interacting with gas at lower ${V_{\mathrm{LSR}}}$, i.e., with regularly rotating material in the Milky Way disc. In some cases the spectra show continuous emission between the velocity of the leading cloud and a lower-velocity component, overlapping in position with the cloud. The spectra in Figure 3 show four examples, where the lower velocity gas is marked with a dashed arrow. In many cases the connection is especially convincing, as illustrated in Figure 8. Here, in a display of latitude vs. $V_{\mathrm{LSR}}$, we can see one or more edges of leading clouds forming a continuous bridge of Hi emission to lower velocities, consistent with the outer regions of the clouds being stripped and decelerated. We compare Figure 8 to Figure 9 (from L08), which is a velocity- position cut through the minor axis of the Smith Cloud showing the same kinematic bridge between the anomalous velocity of the Smith Cloud and gas in the Milky Way disk. Figure 8: Examples of leading clouds showing their interaction with the Milky Way disk gas. In these latitude-$V_{\mathrm{LSR}}$ plots the leading cloud is always the feature at the highest velocity, marked with an arrow. There are ”bridges” to lower velocity Hi emission connecting these clouds to gas at a velocity consistent with normal Galactic rotation. This is often accompanied by an enhancement in the lower velocity emission. We interpret these figures as showing that these leading clouds are interacting with normal disk gas, with some cloud material being stripped away and decellerating to the lower, permitted velocity. Figure 9: From Lockman et al. (2008), a position-velocity cut through the minor axis of the Smith Cloud, showing the same transition at the cloud edge between anomalous velocity gas and normally-rotating disk gas as is seen at the edges of the leading clouds of Figure 8. The arrow marks a small cloud that we interpret as having been detached from the Smith Cloud and decellerated by its interaction with the circumgalactic medium of the Milky Way. The high-resolution P2015 data give additional information on the interaction for two leading clouds: G22.74+4.30 and G26.84-6.38 (we refer to these clouds by their GBT positions in Table 1, which deviate somewhat from the location of the brightest $N_{\mathrm{{\rm HI}}}$ in the P2015 data). A map of total $N_{\mathrm{{\rm HI}}}$ and a position-velocity cut through each cloud is shown in Figure 10. Unlike the clouds in Figure 8, these clouds do not exhibit kinematic bridges of material that join the cloud with lower-velocity gas, but in both cases there is a bright Hi emission component centered approximately at the position of the leading cloud, but at a much lower velocity. The interaction between cloud and disk gas for G25.20+4.42 is shown more clearly in the GBT data of Figure 8 than in the P2015 data because the P2015 data do not have the sensitivity to detect the extended bridge of emission that connects this cloud kinematically to the disk. It is interesting and perhaps significant that of the 10 clouds in the P2015 sample, only those three associated with the Leading Component show unambiguous evidence of interaction with gas at lower velocities. In Table 1 the $V_{\mathrm{LSR}}$ of gas interacting with leading clouds is given as $V_{MW}$. This quantity is blank for clouds without identifiable interaction. The quantity $V_{\mathrm{dif}}$, which is the difference in line- of-sight velocity between the cloud and the disk gas that it is encountering, ${V_{\mathrm{dif}}}\equiv{V_{\mathrm{LSR}}}-V_{MW}$, is shown in Figure 11. The evidence for interaction provides two important pieces of information: 1. 1. the leading clouds are in fact kinematically anomalous by a significant amount; 2. 2. the velocity of the interacting gas might be used to derive a distance to the interaction. Figure 10: Top panels: High angular resolution images of the total $N_{\mathrm{{\rm HI}}}$ from the P2015 data for two of the leading clouds observed in that survey: 22.74+4.30 (left) and 26.84-6.38 (right). Bottom: Cuts through the clouds in longitude-velocity (22.74+4.30, left) and latitude- velocity (26.84-6.38, right) along the directions given by dotted lines in the upper panels. The leading clouds are marked with a solid arrow and a dashed arrow identifies the Milky Way disk gas with which they are interacting. Unlike the examples in Figure 8, these data do not show kinematic bridges between the clouds and Milky Way gas. Instead, we derive evidence of interaction from the high degree of spatial coincidence between the gas at forbidden and permitted velocities, coincidence which is not seen in P2015 clouds except for those that are part of the Leading Component. The velocity of the interacting gas is used to derive a kinematic distance to the interaction. Figure 11: The line-of-sight velocity difference $V_{\mathrm{dif}}$ between the leading clouds and interacting gas, plotted against longitude. The median value is 44 $\rm km\,s^{-1}$, and this is the best estimate of the line-of-sight deviation of the leading clouds from circular rotation at their location. ### 4.1 Location of the Interaction Figure 12: Location of regions where the leading clouds appear to be interacting with disk gas (small blue circles). These were derived for the PM rotation curve with error bars showing the effect of a $\pm 5\ {\rm km\,s^{-1}}$ change in $V_{\mathrm{GSR}}$. Red star and crosses are components of the Smith Cloud (SC) assuming a constant distance from the Sun for this object of 12.4 kpc. The large blue circle is at the position of the Sun and the ”x” shows the Galactic center. (a) Projection of the cloud interaction locations on the x-z plane where x is in the direction of the Galactic Center (which is at 0,0), and z is distance from the Galactic plane. (b) Projection of cloud interaction locations on the z-y plane, where y is perpendicular to the line of sight from the Sun to the Galactic Center. To some extent the correlated structure in this panel simply reflects the alignment of the Leading Component with the Smith Cloud in longitude and latitude. (c) Cloud interaction locations in the x-y plane, which lies in the plane of the Milky Way. The black star shows the position of the Smith Cloud where it is projected to cross the Galactic plane in 27 Myr from the trajectory given in L08. The sense of Galactic rotation is indicated by the arrow extending upwards from the symbol for the Sun. We assume that the velocity of material apparently interacting with a leading cloud is in fact disk gas in circular rotation and thus can be used to identify the location of the interaction within the Galaxy. An identical line of argument was used for the Smith Cloud and produced a distance consistent with that derived from independent methods (L08). While there are many uncertainties in this approach, and the distance to any individual cloud may have a large error, this kinematic analysis has the potential to provide information on the location of the leading cloud population that is not otherwise available. In respect of the uncertainties in derived distances we do not discuss the location of individual clouds but rather limit our comments to properties of the ensemble. These do not vary greatly between the two rotation curves that were used. We analyze the cloud interactions with the two rotation curves described earlier: a flat rotation curve with $V_{\theta}(R)=230\ {\rm km\,s^{-1}}$, and the PM ”Universal” rotation curve (Persic et al., 1996; Mróz et al., 2019). For objects at the longitudes of the leading clouds, a given positive value of $V_{\mathrm{LSR}}$ occurs at a single value of R, the distance from the Galactic Center, but at two distances from the Sun. For the leading cloud sample the derived near distance is around 4 kpc, while the far distance is $>9$ kpc. Given that the Smith Cloud is at a distance of 12.2 kpc we always choose the far kinematic distance for the leading clouds. The velocity of the interacting gas is rarely determined with high precision: to estimate an effect of this uncertainly on our results we evaluated the kinematic distances at the $V_{\mathrm{LSR}}$ of the interaction $\pm 5\ {\rm km\,s^{-1}}$. In the few cases where a velocity is greater than permitted by the Galactic rotation model the tangent point distance is adopted as the most likely. Figure 12 shows derived distances to the interacting gas and by inference, to the leading clouds themselves, using the PM rotation curve where the Sun is at $x=-R_{0}=-8.1$ kpc. The interaction sites of the leading clouds cluster around median values of $x,y,z=0.2,4.5,-0.6$ kpc, at a median distance from the Sun of 9.5 kpc. This location is more than 8 kpc away from the spot where the Smith Cloud is expected to cross the Galactic disk according to the trajectory of L08, marked with a black star on panel (c) of Figure 12. Figure 13: Derived values of $V_{\theta}$ as a function of $cos(\ell-\theta)$ for $V_{R}=0$, assuming that so close to the Galactic plane $V_{z}sin(b)=0$. For very low values of $cos(\ell-\theta)$ any $V_{R}$ component will be projected across the line of sight and $V_{\mathrm{GSR}}$ will be a function of $V_{\theta}$ alone (Equation 1). The horizontal line shows the adopted $V_{\theta}=240$ $\rm km\,s^{-1}$. Figure 14: Derived values of $V_{R}$ for $V_{\theta}=240\ {\rm km\,s^{-1}}$ assuming that so close to the Galactic Plane $V_{z}sin(b)=0$. The horizontal line shows the median value, 65 $\rm km\,s^{-1}$. Error bars show the effect of changing $V_{\mathrm{\theta}}$ by $\pm 10$ $\rm km\,s^{-1}$, and the dashed lines show the median values of $V_{R}$ that result. ### 4.2 Infall or Outflow? Throughout this work we have assumed that the Smith Cloud and the Leading Component are manifestations of infall into the Milky Way. There have been suggestions, however, that the Smith Cloud results from a fountain process and is thus now moving away from the Galactic plane (Sofue et al., 2004; Marasco & Fraternali, 2017). We believe that the properties of clouds in the Leading Component resolve this issue. For one, the Leading Component is found both above and below the Galactic plane at the same angle with respect to the plane as the Smith Cloud. This seems a very unlikely geometry for a supernova-driven superbubble originating in the Milky Way disk, but is unremarkable for an object passing through the Galactic plane. Second, some leading clouds have components that connect them to lower velocity emission, emission that is allowed by Galactic rotation (Figure 8). Were these clouds being accelerated outwards by a fast wind, the presumably stripped gas would be at a higher velocity. These considerations lead to the same conclusion when applied to the Smith Cloud itself (Figure 9). This discussion has no bearing on the origin of these objects, but the case for their infall at the current time is persuasive. ## 5 Space Velocity In a right-handed cylindrical coordinate system centered on the Galactic center where the angle $\theta$ is measured from the Sun-center line such that $\ell\approx\theta$ at large distances, $V_{\mathrm{GSR}}$ can be written as the sum of three components, $V_{\theta},{V_{\mathrm{R}}},$ and $V_{\mathrm{z}}$ ${V_{\mathrm{GSR}}}=[-R_{0}\ sin(\ell)\ V_{\theta}/R\ \ +V_{R}\ cos(\ell-\theta)]cos(b)\ +V_{z}sin(b)$ (1) where in this system the circular velocity $V_{\theta}<0$. Figure 15: Derived values of $V_{z}sin(b)$ for $V_{\theta}=240\ {\rm km\,s^{-1}}$ and ${V_{\mathrm{R}}}=65\ {\rm km\,s^{-1}}$. The horizontal line shows the median value, 0.2 $\rm km\,s^{-1}$. Error bars show the effect of changing the adopted $V_{\mathrm{\theta}}$ by $\pm 10$ $\rm km\,s^{-1}$. The lines formed with dashes and plus signs show values expected for $V_{z}=-100$ $\rm km\,s^{-1}$ and $V_{z}=+100$ $\rm km\,s^{-1}$, respectively. The small values of the derived $V_{z}sin(b)$ imply that $V_{z}$ cannot be determined accurately from these data. There are two limiting cases of interest: first, near the Galactic Plane $sin(b)\approx 0$, so vertical velocities do not contribute significantly to $V_{\mathrm{GSR}}$. Second, at locations near the tangent points the quantity $cos(\ell-\theta)\approx 0$, and the radial component of motion, $V_{R}$, does not contribute to $V_{\mathrm{GSR}}$. At these locations $V_{\mathrm{GSR}}$ is a direct measure of $V_{\theta}$. Figure 13 shows $V_{\theta}$ derived from these assumptions plotted against $cos(\ell-\theta)$. The five clouds with the lowest values of $cos(\ell-\theta)$ give similar results for $V_{\theta}$ which we will adopt as its most likely value, 240 $\rm km\,s^{-1}$, indicated by a horizontal line in Figure 13. For a constant $V_{\theta}=240\ {\rm km\,s^{-1}}$, and, assuming that $V_{z}sin(b)$ is small close to the Galactic plane, Equation 1 can be solved for the $V_{R}$ of the leading clouds. The results are shown in Figure 14 where the median value, 65 $\rm km\,s^{-1}$, is indicated by a horizontal line. The error bars show the effect of a 10 $\rm km\,s^{-1}$ change in $V_{\mathrm{\theta}}$. In principle, having estimates for $V_{\mathrm{\theta}}$ and $V_{\mathrm{R}}$ should allow estimation of $V_{\mathrm{z}}$ for each cloud, but the determination is not robust, as sin(b) for these clouds is so small. Figure 15 shows $V_{\mathrm{z}}$sin(b) vs. latitude with expectations for $V_{z}=\pm 100$ $\rm km\,s^{-1}$. The small values and large uncertainties suggest that $V_{\mathrm{z}}$ cannot be determined reliably so close to the Galactic Plane. The morphology and kinematics of the Smith Cloud indicate that its motion is to higher Galactic latitude, which requires ${V_{\mathrm{z}}}>0$. If the leading clouds share this motion they should have ${V_{\mathrm{z}}}\ sin(b)<0$ at $b<0$ and lie along a curve like the one marked by the plus signs. The estimated values for the space velocity of the Leading Component clouds are compared with these for the Smith Cloud from L08 in Table 4. They are generally consistent. The total space velocity of the Leading Component clouds is $V_{tot}>250\ {\rm km\,s^{-1}}$ where the inequality indicates that we are unable to establish a reliable value for $V_{z}$. The Leading Clouds have a median line-of-sight velocity excess of 44 $\rm km\,s^{-1}$ with respect to the disk gas at their location. This is in the direction of Galactic rotation. As gas is stripped from these clouds (Figure 8) it adds not only mass but angular momentum to the disk. ## 6 What is the Leading Component? In this paper we have presented arguments for the existence of an organized interstellar structure whose orientation and kinematics are not consistent with Galactic rotation, but which has a connection to the Smith high velocity cloud. The principle evidence for these claims is 1. 1. A band of Hi emission crossing the Galactic plane a few degrees ahead of the Smith Cloud with an orientation identical to that of the Smith Cloud. 2. 2. The clouds in this structure have velocities larger than allowed by Galactic rotation (Figure 7) but the velocities are virtually identical to the $V_{\mathrm{GSR}}$ of the Smith Cloud, indicating that there is a kinematic as well as a spatial connection between the structures. 3. 3. Many of the leading clouds are interacting with material at permitted velocities, confirming their peculiar kinematics (Figure 8 and Figure 11). We thus believe that the Leading Component is an anomalous feature of the ISM related to the Smith Cloud, and despite its location in the Galactic plane this material had its origin as a high velocity cloud. Evidence of interaction with normal disk gas allows us to estimate that where the Leading Component crosses the Galactic plane near longitude $25\arcdeg$ it is 9.5 kpc from the Sun and 4.5 kpc from the Galactic Center. This is a region of the Galaxy with significant star formation activity as indicated by the presence of numerous HII regions (Anderson et al., 2012), some of which have a parallactic distance identical to the median kinematic distance to the Leading Component (Sato et al., 2014). The Leading Component has a neutral hydrogen mass comparable to that of the Smith Cloud, and it extends $\approx 17\arcdeg$ across the sky which, for our derived distance, makes its linear size $\sim 3$ kpc, identical to that of the Smith Cloud. From estimates of interaction locations we derive the velocity components for the leading clouds and find that they are similar to those of the Smith Cloud (Table 4). Because the two objects have a similar $V_{\mathrm{GSR}}$ (Figure 6) the similarity of the individual velocity components is not surprising, but is nonetheless reassuring that we are observing a single phenomenon. There is a gap between the Leading Component and the Smith Cloud of about $4\arcdeg$ in which there is no strong emission that can be associated with either object (see the left panel of Figure 2). This would correspond to a gap of 0.6 - 1.6 kpc depending on the projection and assumed distance of the clouds. The total space velocity of the Leading Component is $>250$ $\rm km\,s^{-1}$, where the limit reflects our inability to measure $V_{z}$ accurately. This is comparable to the L08 determination for the Smith Cloud of $V_{tot}=296\pm 20$ $\rm km\,s^{-1}$. There is no information to suggest that either object has a velocity capable of escaping the Galaxy. The estimates of cloud space velocity from section 5 required several approximations; a more complete study of the entire Smith Cloud and Leading Component taking into account the Galactic potential and dynamical effects is in progress. The Smith Cloud has been analyzed as an isolated system, perhaps the baryonic component of a dark matter subhalo (Nichols & Bland-Hawthorn, 2009; Nichols et al., 2014; Tepper-García & Bland-Hawthorn, 2018), and the question of the survival of its gaseous component given its strong interaction with the Milky Way’s circumgalactic medium has received considerable attention (Gritton et al., 2014; Galyardt & Shelton, 2016). The discovery of the Leading Component suggests that the Smith Cloud may not be an isolated object but instead the more prominent manifestation of a system that stretches over more than 7 kpc on the sky with a gap between its two parts. It is interesting that “chains” of infalling gas occur naturally in some simulations that include simultaneous infall and outflow – cloud chains that may survive to pass through a galaxy’s disk (Melso et al., 2019). Perhaps that is what we are observing here. If the metallicity of the Leading Component is similar to that of the Smith Cloud at $\approx 0.5$ solar, it will lower the average metallicity of disk gas over the region of interaction. In Lockman et al. (2008) and subsequent publications there was speculation about the future of the Smith Cloud, as its trajectory indicated that it would enter and merge with the Milky Way disk in a few tens of Myr perhaps triggering a burst of star formation (Alig et al., 2018). Given the existence of the Leading Component it appears that the merger is already underway. Acknowledgments The Green Bank Observatory is a facility of the National Science Foundation, operated by Associated Universities, Inc. The observations were made under GBT proposal codes 06C${\\_}$038, 08A${\\_}$014, and 09A${\\_}$007\. We thank E.C. Kornacki for assistance with the data reduction and the anonymous referee for useful comments. ECK, NP, and CT were supported by the NSF REU programs at the National Radio Astronomy Observatory and the Green Bank Observatory. FJL acknowledges the influence of the classic work by Toomre (1977) on the opening sentence of this paper. Table 1: Leading clouds detected in the GBT Observations. Columns as follow: (1) Galactic longitude; (2) Galactic latitude; (3) Maximum line brightness temperature; (4) LSR velocity; (5) Line width FWHM; (6) Maximum ${\rm HI}$ column density; (7) observed angular size FWHM; (8) position angle with respect to the Galactic pole; (9) ${\rm HI}$ mass assuming a distance of 10 kpc. Columns (3)-(5) are from a Gaussian fit. Column (7) is from a 2D Gaussian fit to the channel map at the velocity of peak emission and is not corrected for beam smearing. Column (9) is calculated assuming a distance from the Sun of 10 kpc. Uncertainties are propagated from the error associated with the Gaussian fits. If a cloud shows evidence that it is interacting with normal gas in the Milky Way disk the velocity of that gas is given as $V_{MW}$ in column (10). $\ell$ | $b$ | $T_{\mathrm{pk}}$ | ${V_{\mathrm{LSR}}}$ | FWHM | $N_{\mathrm{{\rm HI}}}$ | $\Theta_{obs}$ | PA | $M_{{\rm HI}}\ $ (10 kpc) | $V_{MW}$ ---|---|---|---|---|---|---|---|---|--- ($\arcdeg$) | ($\arcdeg$) | (K) | ($\rm km\,s^{-1}$) | ($\rm km\,s^{-1}$) | ($10^{19}\ {\rm cm^{-2}}$) | ($\arcmin$) | ($\arcdeg$) | ($M_{\mathrm{\odot}}$) | ($\rm km\,s^{-1}$) (1) | (2) | (3) | (4) | (5) | (6) | (7) | (8) | (9) | (10) 21.67 | +2.63 | 2.41$\pm$0.05 | 149.6$\pm$0.3 | 18.9$\pm$0.8 | 8.8 | — | — | 5.6e+03 | 119 22.74 | +4.30 | 3.48$\pm$0.04 | 137.3$\pm$0.1 | 11.0$\pm$0.2 | 7.5 | 17.9$\times$15.0 | 125$\pm$6 | 1.2e+03 | 78 23.45 | -2.00 | 1.84$\pm$0.04 | 146.2$\pm$0.1 | 12.3$\pm$0.3 | 4.4 | — | — | 2.4e+03 | 112 23.86 | -3.57 | 1.83$\pm$0.03 | 136.6$\pm$0.2 | 19.7$\pm$0.6 | 7.0 | 22.6$\times$16.0 | 84$\pm$3 | 1.3e+03 | 103 25.15 | -0.55 | 4.71$\pm$0.09 | 143.1$\pm$0.3 | 17.3$\pm$0.6 | 15.8 | 71.4$\times$57.0 | 27$\pm$6 | 1.0e+04 | — 25.20 | +4.42 | 2.76$\pm$0.04 | 147.7$\pm$0.1 | 12.8$\pm$0.2 | 6.9 | 16.7$\times$14.0 | 134$\pm$6 | 1.4e+03 | 87 25.36 | -3.41 | 2.63$\pm$0.05 | 132.2$\pm$0.2 | 16.9$\pm$0.5 | 8.6 | 72.0$\times$45.6 | 94$\pm$3 | 3.8e+03 | 91 25.60 | -2.71 | 1.72$\pm$0.05 | 145.0$\pm$0.2 | 13.5$\pm$0.6 | 4.5 | 37.0$\times$13.6 | 168$\pm$1 | 1.1e+03 | 126 25.72 | -1.42 | 1.70$\pm$0.10 | 146.7$\pm$0.3 | 9.1$\pm$0.9 | 3.0 | 17.8$\times$12.4 | 126$\pm$4 | 6.6e+02 | 105 25.74 | +4.78 | 1.06$\pm$0.03 | 142.6$\pm$0.3 | 24.8$\pm$0.8 | 5.1 | 17.2$\times$13.7 | 84$\pm$6 | 7.7e+02 | — 26.05 | -2.90 | 2.13$\pm$0.06 | 140.8$\pm$0.2 | 13.3$\pm$0.5 | 5.5 | 41.3$\times$25.0 | 126$\pm$2 | 1.3e+03 | 114 26.36 | +3.49 | 7.62$\pm$0.03 | 120.2$\pm$0.0 | 15.9$\pm$0.1 | 23.5 | 84.6$\times$45.6 | 1$\pm$2 | 1.1e+04 | 66 26.77 | -3.28 | 3.16$\pm$0.03 | 127.1$\pm$0.1 | 15.6$\pm$0.2 | 9.6 | — | — | 6.9e+03 | 81 26.84 | -2.24 | 1.44$\pm$0.06 | 136.8$\pm$0.4 | 12.4$\pm$1.0 | 3.4 | 36.3$\times$27.9 | 129$\pm$5 | 9.5e+02 | 116 26.84 | -6.38 | 3.78$\pm$0.04 | 123.0$\pm$0.1 | 10.0$\pm$0.1 | 7.4 | 29.2$\times$14.7 | 132$\pm$1 | 1.9e+03 | 94 27.94 | -4.11 | 2.90$\pm$0.04 | 132.7$\pm$0.1 | 12.5$\pm$0.2 | 7.0 | 47.9$\times$21.8 | 9$\pm$1 | 3.2e+03 | 87 28.02 | -2.75 | 4.03$\pm$0.05 | 124.6$\pm$0.2 | 18.4$\pm$0.5 | 14.4 | 77.4$\times$37.8 | 127$\pm$1 | 8.6e+03 | 110 28.76 | -1.83 | 11.35$\pm$0.08 | 121.8$\pm$0.1 | 12.8$\pm$0.1 | 28.2 | 37.4$\times$20.3 | 89$\pm$2 | 1.1e+04 | 96 29.29 | -4.34 | 3.46$\pm$0.04 | 135.8$\pm$0.1 | 9.0$\pm$0.1 | 6.1 | 19.9$\times$15.3 | 124$\pm$7 | 1.0e+03 | — 29.31 | -7.30 | 1.36$\pm$0.04 | 122.8$\pm$0.1 | 11.1$\pm$0.4 | 2.9 | 19.7$\times$11.0 | 51$\pm$2 | 4.3e+02 | — 29.36 | -5.03 | 1.71$\pm$0.05 | 130.9$\pm$0.1 | 6.8$\pm$0.2 | 2.3 | 15.3$\times$11.8 | 163$\pm$3 | 2.8e+02 | — 29.92 | -7.92 | 0.79$\pm$0.03 | 132.3$\pm$0.4 | 18.1$\pm$1.1 | 2.8 | 27.2$\times$18.0 | 16$\pm$4 | 1.1e+03 | 74 29.95 | -6.14 | 3.49$\pm$0.03 | 128.2$\pm$0.1 | 13.5$\pm$0.2 | 9.1 | 22.0$\times$15.8 | 149$\pm$3 | 1.9e+03 | 63 30.02 | -7.42 | 3.78$\pm$0.03 | 117.7$\pm$0.1 | 19.6$\pm$0.2 | 14.4 | 22.0$\times$16.8 | 98$\pm$3 | 1.3e+03 | 71 30.07 | -3.92 | 5.30$\pm$0.11 | 121.3$\pm$0.2 | 15.2$\pm$0.6 | 15.6 | — | — | 1.7e+03 | — 30.37 | -6.31 | 1.33$\pm$0.03 | 119.2$\pm$0.2 | 21.5$\pm$0.7 | 5.5 | 46.6$\times$38.0 | 91$\pm$7 | 2.0e+03 | — 30.64 | -4.16 | 5.35$\pm$0.04 | 128.2$\pm$0.0 | 12.5$\pm$0.1 | 13.0 | 28.3$\times$15.4 | 137$\pm$2 | 2.5e+03 | 82 30.98 | -6.85 | 1.01$\pm$0.02 | 119.2$\pm$0.5 | 29.6$\pm$1.2 | 5.8 | — | — | 2.3e+03 | 68 31.35 | -4.80 | 2.40$\pm$0.03 | 121.7$\pm$0.1 | 15.2$\pm$0.3 | 7.1 | 54.6$\times$20.2 | 129$\pm$1 | 6.2e+02 | 76 31.47 | -6.53 | 3.57$\pm$0.03 | 111.8$\pm$0.1 | 17.8$\pm$0.2 | 12.3 | 38.9$\times$22.1 | 162$\pm$1 | 3.1e+03 | 51 31.54 | -7.55 | 3.86$\pm$0.03 | 105.2$\pm$0.1 | 19.4$\pm$0.2 | 14.5 | 32.1$\times$18.2 | 48$\pm$2 | 1.5e+03 | 65 31.57 | -3.46 | 2.14$\pm$0.05 | 120.7$\pm$0.2 | 15.3$\pm$0.6 | 6.4 | — | — | 1.2e+03 | 95 31.64 | -5.73 | 4.39$\pm$0.05 | 112.2$\pm$0.0 | 8.0$\pm$0.1 | 6.8 | — | — | 1.1e+03 | 85 32.04 | -4.10 | 1.79$\pm$0.03 | 120.6$\pm$0.2 | 19.6$\pm$0.5 | 6.8 | — | — | 1.0e+03 | — 32.06 | -5.38 | 5.22$\pm$0.03 | 112.9$\pm$0.1 | 15.2$\pm$0.1 | 15.4 | — | — | 3.2e+03 | 76 32.15 | -7.35 | 2.73$\pm$0.03 | 110.2$\pm$0.2 | 18.8$\pm$0.4 | 10.0 | 30.0$\times$24.6 | 107$\pm$4 | 3.8e+03 | 50 32.28 | -4.97 | 3.15$\pm$0.04 | 121.1$\pm$0.1 | 13.7$\pm$0.2 | 8.3 | 21.8$\times$17.1 | 78$\pm$7 | 1.2e+03 | 56 32.46 | -6.55 | 2.46$\pm$0.03 | 123.4$\pm$0.1 | 14.0$\pm$0.2 | 6.7 | 44.1$\times$17.7 | 147$\pm$1 | 2.3e+03 | 43 Table 2: Leading cloud population properties. Distant-dependent quantities were calculated assuming a distance from the Sun, $d_{10}=10$ kpc. Spectral properties were derived from a Gaussian fit at the location of the peak 21cm brightness. The quantity $V_{\mathrm{GSR}}$ was calculated using a Galactic circular velocity of 230 $\rm km\,s^{-1}$ at the location of the Sun. $V_{\mathrm{dif}}$ is the difference in velocity between the cloud and the material it appears to be interacting with. Angular sizes were derived from a Gaussian fit at the velocity of peak emission corrected by subtracting in quadrature the $10\arcmin$ effective angular resolution of the survey. Column (4) gives the number of clouds used to establish each quantity. Property | Median | Range | n clouds ---|---|---|--- (1) | (2) | (3) | (4) $T_{pk}\ $ (K) | 2.7 | 0.79 – 11.35 | 38 FWHM ($\rm km\,s^{-1}$) | 15.2 | 6.8 – 29.6 | 38 Peak $N_{\mathrm{{\rm HI}}}$ $(10^{19}\ {\rm cm^{-2}})$ | 7.0 | 2.3 – 28.2 | 38 $V_{\mathrm{GSR}}$ ($\rm km\,s^{-1}$) | 237 | 222 – 248 | 38 $\Theta_{maj}\ (\arcmin)$ | 28.3 | 11.6 – 84 | 29 $\Theta_{min}\ (\arcmin)$ | 15.0 | 4.6 – 56 | 29 PA $(\arcdeg)$ | 124 | 1 – 168 | 29 Diam ($d_{10}$ pc) | 57 | 25 – 183 | 29 $n\ \ ({\rm cm^{-3}}\ \ d_{10}^{-1})$ | 0.43 | 0.13 – 1.24 | 29 $M_{{\rm HI}}\ (10^{3}\ d_{10}^{2}$ $M_{\mathrm{\odot}}$) | 1.6 | 0.28 – 11.0 | 38 ${V_{\mathrm{dif}}}\ $ ($\rm km\,s^{-1}$) | 44 | 15 – 81 | 30 Table 3: Properties of three leading component clouds observed by the GBT at $10\arcmin$ resolution and by P2015 at $1\arcmin$ resolution. The P2015 diameters refer to the cloud core and were scaled to a distance of 10 kpc. Property | G22.74+4.30 | G25.20+4.42 | G26.84-6.38 ---|---|---|--- | GBT, P2015 | GBT, P2015 | GBT, P2015 (1) | (2) | (3) | (4) Peak $N_{\mathrm{{\rm HI}}}$ $(10^{20}\ {\rm cm^{-2}})$ | 0.75, 1.7 | 0.69, 1.6 | 0.74, 2.2 FWHM ($\rm km\,s^{-1}$) aaG22.74+4.30 and G26.84-6.38 have two spectral components in the P2015 data, one broad and one narrow. | 11.0, 5.8/24.3 | 12.8, 13.4 | 10.0, 3.0/14.5 Diam ($d_{10}$ pc) | 37.5, 12.6 | 33.3, 13.3 | 50.0, 11.0 $n\ \ ({\rm cm^{-3}}\ \ d_{10}^{-1})$ | 0.65, 4.4 | 0.67, 3.9 | 0.48, 6.4 PA $(\arcdeg)$ | 125, 122 | 134, 156 | 132, 128 Table 4: Space velocities of the Smith Cloud and the Leading Component. Smith Cloud entries are from the model in L08. Leading cloud entries are from the discussion in section 5 where the the uncertainty in $V_{R}$ shows the effect of a 10 $\rm km\,s^{-1}$ change in the adopted $V_{\theta}$, and $V_{z}$ could not be determined. Object | $V_{\theta}$ | $V_{R}$ | $V_{z}$ ---|---|---|--- | $\rm km\,s^{-1}$ | $\rm km\,s^{-1}$ | $\rm km\,s^{-1}$ (1) | (2) | (3) | (4) Smith Cloud | $270\pm 21$ | $94\pm 18$ | $73\pm 26$ Leading Component | $240$ | $65\pm 35$ | — ## References * Alig et al. (2018) Alig, C., Hammer, S., Borodatchenkova, N., Dobbs, C. L., & Burkert, A. 2018, ApJ, 869, L2, doi: 10.3847/2041-8213/aaf1cb * Anderson et al. (2012) Anderson, L. D., Bania, T. M., Balser, D. S., & Rood, R. T. 2012, ApJ, 754, 62, doi: 10.1088/0004-637X/754/1/62 * Barger et al. (2020) Barger, K. A., Nidever, D. L., Huey-You, C., et al. 2020, ApJ, 902, 154, doi: 10.3847/1538-4357/abb376 * Betti et al. (2019) Betti, S. K., Hill, A. S., Mao, S. A., et al. 2019, ApJ, 871, 215, doi: 10.3847/1538-4357/aaf886 * Binney & Fraternali (2012) Binney, J., & Fraternali, F. 2012, in European Physical Journal Web of Conferences, Vol. 19, European Physical Journal Web of Conferences, 08001, doi: 10.1051/epjconf/20121908001 * Bish et al. (2021) Bish, H. V., Werk, J. K., Peek, J., Zheng, Y., & Putman, M. 2021, ApJ, 912, 8, doi: 10.3847/1538-4357/abeb6b * Bland-Hawthorn et al. (1998) Bland-Hawthorn, J., Veilleux, S., Cecil, G. N., et al. 1998, MNRAS, 299, 611, doi: 10.1046/j.1365-8711.1998.01902.x * Boothroyd et al. (2011) Boothroyd, A. I., Blagrave, K., Lockman, F. J., et al. 2011, A&A, 536, A81, doi: 10.1051/0004-6361/201117656 * Brand & Blitz (1993) Brand, J., & Blitz, L. 1993, A&A, 275, 67 * Braun & Thilker (2004) Braun, R., & Thilker, D. A. 2004, A&A, 417, 421, doi: 10.1051/0004-6361:20034423 * Burton (1971) Burton, W. B. 1971, A&A, 10, 76 * Burton (1988) —. 1988, in Galactic and Extragalactic Radio Astronomy, ed. K. I. Kellermann & G. L. Verschuur, 295–358 * Celnik et al. (1979) Celnik, W., Rohlfs, K., & Braunsfurth, E. 1979, A&A, 76, 24 * Eilers et al. (2019) Eilers, A.-C., Hogg, D. W., Rix, H.-W., & Ness, M. K. 2019, ApJ, 871, 120, doi: 10.3847/1538-4357/aaf648 * Fox et al. (2016) Fox, A. J., Lehner, N., Lockman, F. J., et al. 2016, ApJ, 816, L11, doi: 10.3847/2041-8205/816/1/L11 * Fraternali (2017) Fraternali, F. 2017, in Astrophysics and Space Science Library, Vol. 430, Gas Accretion onto Galaxies, ed. A. Fox & R. Davé, 323, doi: 10.1007/978-3-319-52512-9_14 * Galyardt & Shelton (2016) Galyardt, J., & Shelton, R. L. 2016, ApJ, 816, L18, doi: 10.3847/2041-8205/816/1/L18 * Gravity Collaboration et al. (2018) Gravity Collaboration, Abuter, R., Amorim, A., et al. 2018, A&A, 615, L15, doi: 10.1051/0004-6361/201833718 * Gritton et al. (2014) Gritton, J. A., Shelton, R. L., & Kwak, K. 2014, ApJ, 795, 99, doi: 10.1088/0004-637X/795/1/99 * Heitsch et al. (2022) Heitsch, F., Marchal, A., Miville-Deschênes, M. A., Shull, J. M., & Fox, A. J. 2022, MNRAS, 509, 4515, doi: 10.1093/mnras/stab3266 * Henley et al. (2017) Henley, D. B., Gritton, J. A., & Shelton, R. L. 2017, ApJ, 837, 82, doi: 10.3847/1538-4357/aa5df7 * HI4PI Collaboration et al. (2016) HI4PI Collaboration, Ben Bekhti, N., Flöer, L., et al. 2016, A&A, 594, A116, doi: 10.1051/0004-6361/201629178 * Hill et al. (2009) Hill, A. S., Haffner, L. M., & Reynolds, R. J. 2009, ApJ, 703, 1832, doi: 10.1088/0004-637X/703/2/1832 * Hill et al. (2013) Hill, A. S., Mao, S. A., Benjamin, R. A., Lockman, F. J., & McClure-Griffiths, N. M. 2013, ApJ, 777, 55, doi: 10.1088/0004-637X/777/1/55 * Kalberla et al. (2005) Kalberla, P. M. W., Burton, W. B., Hartmann, D., et al. 2005, A&A, 440, 775, doi: 10.1051/0004-6361:20041864 * Lehner et al. (2022) Lehner, N., Howk, J. C., Marasco, A., & Fraternali, F. 2022, MNRAS, 513, 3228, doi: 10.1093/mnras/stac987 * Levine et al. (2008) Levine, E. S., Heiles, C., & Blitz, L. 2008, ApJ, 679, 1288, doi: 10.1086/587444 * Li & Tonnesen (2020) Li, M., & Tonnesen, S. 2020, ApJ, 898, 148, doi: 10.3847/1538-4357/ab9f9f * Lockman (2003) Lockman, F. J. 2003, ApJ, 591, L33, doi: 10.1086/376961 * Lockman et al. (2008) Lockman, F. J., Benjamin, R. A., Heroux, A. J., & Langston, G. I. 2008, ApJ, 679, L21, doi: 10.1086/588838 * Lockman et al. (2002) Lockman, F. J., Murphy, E. M., Petty-Powell, S., & Urick, V. J. 2002, ApJS, 140, 331, doi: 10.1086/339371 * Marasco & Fraternali (2017) Marasco, A., & Fraternali, F. 2017, MNRAS, 464, L100, doi: 10.1093/mnrasl/slw195 * Marasco et al. (2022) Marasco, A., Fraternali, F., Lehner, N., & Howk, J. C. 2022, MNRAS, 515, 4176, doi: 10.1093/mnras/stac1172 * Marinacci et al. (2010) Marinacci, F., Binney, J., Fraternali, F., et al. 2010, MNRAS, 404, 1464, doi: 10.1111/j.1365-2966.2010.16352.x * McClure-Griffiths & Dickey (2007) McClure-Griffiths, N. M., & Dickey, J. M. 2007, ApJ, 671, 427, doi: 10.1086/522297 * McClure-Griffiths et al. (2008) McClure-Griffiths, N. M., Staveley-Smith, L., Lockman, F. J., et al. 2008, ApJ, 673, L143, doi: 10.1086/528683 * Melso et al. (2019) Melso, N., Bryan, G. L., & Li, M. 2019, ApJ, 872, 47, doi: 10.3847/1538-4357/aafaf5 * Mróz et al. (2019) Mróz, P., Udalski, A., Skowron, D. M., et al. 2019, ApJ, 870, L10, doi: 10.3847/2041-8213/aaf73f * Muller et al. (1963) Muller, C. A., Oort, J. H., & Raimond, E. 1963, Academie des Sciences Paris Comptes Rendus, 257, 1661 * Nichols & Bland-Hawthorn (2009) Nichols, M., & Bland-Hawthorn, J. 2009, ApJ, 707, 1642, doi: 10.1088/0004-637X/707/2/1642 * Nichols et al. (2014) Nichols, M., Mirabal, N., Agertz, O., Lockman, F. J., & Bland-Hawthorn, J. 2014, MNRAS, 442, 2883, doi: 10.1093/mnras/stu1028 * Oort (1969) Oort, J. H. 1969, Nature, 224, 1158, doi: 10.1038/2241158a0 * Park et al. (2016) Park, G., Koo, B.-C., Kang, J.-h., et al. 2016, ApJ, 827, L27, doi: 10.3847/2041-8205/827/2/L27 * Persic et al. (1996) Persic, M., Salucci, P., & Stel, F. 1996, MNRAS, 281, 27, doi: 10.1093/mnras/278.1.27 * Pidopryhora et al. (2015) Pidopryhora, Y., Lockman, F. J., Dickey, J. M., & Rupen, M. P. 2015, ApJS, 219, 16, doi: 10.1088/0067-0049/219/2/16 * Prestage et al. (2009) Prestage, R. M., Constantikes, K. T., Hunter, T. R., et al. 2009, IEEE Proceedings, 97, 1382, doi: 10.1109/JPROC.2009.2015467 * Putman (2017) Putman, M. E. 2017, in Astrophysics and Space Science Library, Vol. 430, Gas Accretion onto Galaxies, ed. A. Fox & R. Davé, 1, doi: 10.1007/978-3-319-52512-9_1 * Putman et al. (2003) Putman, M. E., Bland-Hawthorn, J., Veilleux, S., et al. 2003, ApJ, 597, 948, doi: 10.1086/378555 * Putman et al. (2012) Putman, M. E., Peek, J. E. G., & Joung, M. R. 2012, ARA&A, 50, 491, doi: 10.1146/annurev-astro-081811-125612 * Putman et al. (2011) Putman, M. E., Saul, D. R., & Mets, E. 2011, MNRAS, 418, 1575, doi: 10.1111/j.1365-2966.2011.19524.x * Reid et al. (2014) Reid, M. J., Menten, K. M., Brunthaler, A., et al. 2014, ApJ, 783, 130, doi: 10.1088/0004-637X/783/2/130 * Richter (2017) Richter, P. 2017, in Astrophysics and Space Science Library, Vol. 430, Gas Accretion onto Galaxies, ed. A. Fox & R. Davé, 15, doi: 10.1007/978-3-319-52512-9_2 * Sato et al. (2014) Sato, M., Wu, Y. W., Immer, K., et al. 2014, ApJ, 793, 72, doi: 10.1088/0004-637X/793/2/72 * Smith (1963) Smith, G. P. 1963, Bull. Astron. Inst. Netherlands, 17, 203 * Sofue et al. (2004) Sofue, Y., Kudoh, T., Kawamura, A., Shibata, K., & Fujimoto, M. 2004, PASJ, 56, 633, doi: 10.1093/pasj/56.4.633 * Stark et al. (2015) Stark, D. V., Baker, A. D., & Kannappan, S. J. 2015, MNRAS, 446, 1855, doi: 10.1093/mnras/stu2182 * Tepper-García & Bland-Hawthorn (2018) Tepper-García, T., & Bland-Hawthorn, J. 2018, MNRAS, 473, 5514, doi: 10.1093/mnras/stx2680 * Thilker et al. (2004) Thilker, D. A., Braun, R., Walterbos, R. A. M., et al. 2004, ApJ, 601, L39, doi: 10.1086/381703 * Toomre (1977) Toomre, A. 1977, ARA&A, 15, 437, doi: 10.1146/annurev.aa.15.090177.002253 * Wakker & van Woerden (1991) Wakker, B. P., & van Woerden, H. 1991, A&A, 250, 509 * Wakker & van Woerden (1997) —. 1997, ARA&A, 35, 217, doi: 10.1146/annurev.astro.35.1.217 * Wakker et al. (2008) Wakker, B. P., York, D. G., Wilhelm, R., et al. 2008, ApJ, 672, 298, doi: 10.1086/523845 * Westmeier et al. (2008) Westmeier, T., Brüns, C., & Kerp, J. 2008, MNRAS, 390, 1691, doi: 10.1111/j.1365-2966.2008.13858.x * Wolfe et al. (2016) Wolfe, S. A., Lockman, F. J., & Pisano, D. J. 2016, ApJ, 816, 81, doi: 10.3847/0004-637X/816/2/81 * Zheng et al. (2015) Zheng, Y., Putman, M. E., Peek, J. E. G., & Joung, M. R. 2015, ApJ, 807, 103, doi: 10.1088/0004-637X/807/1/103
# Structural evolution of the kagome superconductors $A$V3Sb5 ($A$ = K, Rb, and Cs) through charge density wave order Linus Kautzsch Materials Department, University of California Santa Barbara, Santa Barbara, CA, 93106, United States Brenden R. Ortiz Materials Department, University of California Santa Barbara, Santa Barbara, CA, 93106, United States Krishnanand Mallayya Department of Physics, Cornell University, Ithaca, NY, 14853, United States Jayden Plumb Materials Department, University of California Santa Barbara, Santa Barbara, CA, 93106, United States Ganesh Pokharel Materials Department, University of California Santa Barbara, Santa Barbara, CA, 93106, United States Jacob P. C. Ruff CHESS, Cornell University, Ithaca, NY, 14853, United States Zahirul Islam Advanced Photon Source, Argonne National Laboratory, Lemont, IL, 60439, United States Eun-Ah Kim Department of Physics, Cornell University, Ithaca, NY, 14853, United States Ram Seshadri Materials Department, University of California Santa Barbara, Santa Barbara, CA, 93106, United States Stephen D. Wilson<EMAIL_ADDRESS>Materials Department, University of California Santa Barbara, Santa Barbara, CA, 93106, United States ###### Abstract The kagome superconductors KV3Sb5, RbV3Sb5, and CsV3Sb5 are known to display charge density wave (CDW) order which impacts the topological characteristics of their electronic structure. Details of their structural ground states and how they evolve with temperature are revealed here using single crystal X-ray crystallographic refinements as a function of temperature, carried out with synchrotron radiation. The compounds KV3Sb5 and RbV3Sb5 present 2$\times$2$\times$2 superstructures in the $Fmmm$ space group with a staggered tri-hexagonal deformation of vanadium layers. CsV3Sb5 displays more complex structural evolution, whose details have been unravelled by applying machine learning methods to the scattering data. Upon cooling through the CDW transition, CsV3Sb5 displays a staged progression of ordering from a 2$\times$2$\times$1 supercell and a 2$\times$2$\times$2 supercell into a final 2$\times$2$\times$4 supercell that persists to $T$ = 11 K and exhibits an average structure where vanadium layers display both tri-hexagonal and Star of David patterns of deformations. Diffraction from CsV3Sb5 under pulsed magnetic fields up to $\mu_{0}H$ = 28 T suggest the real component of the CDW state is insensitive to external magnetic fields. ††preprint: APS/123-QED ## I Introduction The charge density wave (CDW) instability is central to many of the unconventional properties reported in the $A$V3Sb5 ($A=$ K, Rb, and Cs) family of kagome superconductors [1, 2, 3]. Within the kagome lattice of these materials, the CDW is hypothesized to harbor both real and imaginary components within the resulting order parameter, with the former corresponding to a real space charge inhomogeneity and the latter corresponding to a “chiral” flux phase that breaks time reversal symmetry [4]. The real component of the CDW results in a superstructure deformation of the crystal lattice. Precise details of the structural pattern of this deformation has been an active area of recent investigation. Understanding the pattern of the real space structural deformation stands to inform a number of theoretical models of the anomalous CDW transtion in AV3Sb5 compounds. For instance, models of the band folding through the transition and calculations of the resulting Berry curvature may help inform the origin of the large anomalous Hall effect reported in the CDW state [5, 6]. At high temperatures, the structure of all three variants assumes an ideal (undistorted) kagome network of vanadium ions in the space group $P6/mmm$ as shown in Fig. 1 [1]. The lattice deformation that results upon cooling through the CDW transition primarily arises from the displacement of the kagome network of vanadium atoms [7]. The most energetically favored possibilities are predicted to be one of two breathing modes: either the Star of David-type (SoD) deformation or its inverse, a tri-hexagonal (TrH) deformation [8, 9, 10]. Both are triple-q distortion modes, suggestive of the influence of nesting effects between the saddle-points ($M$-points) close to the Fermi level in AV3Sb5. Figure 1: (a) and (b) show the refined room temperature ($T=290$ K) structure of the $A$V3Sb5 ($A=$ K, Rb, and Cs) family in the hexagonal space group $P6/mmm$. An added complexity in solving the pattern of CDW order in AV3Sb5 is that interlayer interactions are strong enough to promote a fully three-dimensional CDW. This has been observed in scanning tunneling microscopy (STM) measurements as a half unit cell phase-shift between layers resolved across step-edges [11] as well as direct observation of superlattice reflections with a finite $q_{z}$ propagation wave vector in x-ray diffraction measurements [7, 12]. The out-of-plane $q_{z}$ periodicity varies between AV3Sb5 variants with the unit cell doubled along the c-axis in A= K, Rb variants [8, 12] and the unit cell at least partially quadrupled in A= Cs [7, 13, 14]. The increased interlayer spacing driven by the larger Cs interstitial layer seemingly modifies the interactions between kagome planes and drives the formation of a delicate three-dimensional CDW phase whose modulation between kagome planes potentially hosts phase coexistence and metastable states. Metastability in the CDW phase of CsV3Sb5 has been primarily observed with x-ray diffraction measurements. Initial reports found conflicting observations of $2\times 2\times 2$ [12] and $2\times 2\times 4$ CDW superlattices [7], suggesting sensitivity to growth conditions. Furthermore, recent x-ray diffraction studies have reported the coexistence of $2\times 2\times 2$ and $2\times 2\times 4$ structures with different onset temperatures and thermal quenching/annealing behaviors [13, 14]. Conclusions regarding which state is stabilized by annealing also vary, likely reflecting the added variable of disorder introduced via the growth process in the ground state selection. For instance, recent studies have shown that small amounts of kagome-plane dopants can quench the $2\times 2\times 4$ state, leaving only a quasi-two-dimensional $2\times 2\times 2$ phase remaining [15]. The above complexity of the CDW state in CsV3Sb5 is relevant as it is the most easily grown and widely studied AV3Sb5 variant. In our prior synchrotron study, an average structural solution for CsV3Sb5 was presented presuming a minimal model where the system maintains hexagonal symmetry ($P\bar{3}$). This average structure suggested a superstructure comprised of modulated TrH and SoD distortions in a $2\times 2\times 4$ cell [7]. This prior solution could also be indexed as an orthorhombic cell; however twinning/domain formation was not directly resolved in the diffraction data. Recently, however, STM [16] and optical measurements [17] have resolved three sets of orthorhombic twin domains in the CDW state of all three $A$V3Sb5 compounds. This provides an excellent basis for developing twinned orthorhombic structural models of these systems via synchrotron x-ray diffraction measurements. Here, we report the structural deformations resulting from CDW order in all three AV3Sb5 compounds determined via single crystal synchrotron x-ray diffraction measurements. Structural refinements using an orthorhombic, three domain, twinned model are presented for all three compounds. For KV3Sb5 and RbV3Sb5, a 2$\times$2$\times$2 superstructure is observed and structural refinement, combined with recent NMR analysis [18, 19, 20], allows for a staggered TrH deformation of the vanadium layers to be resolved and the atomic positions determined. We also revisit the average structure of CsV3Sb5 in an orthorhombic cell using the same three domain structure, where we identify a 2$\times$2$\times$4 superlattice in the ground state. The average low- temperature structural solution remains a modulation of SoD and TrH-type deformations, consistent with prior reports [7, 21, 22]. Curiously, we also resolve a cross-over regime below $T=94$ K where a 2$\times$2$\times$1 and a 2$\times$2$\times$2 lattice form and then diminish at lower temperatures. Finally, as a test for potential time reversal symmetry breaking in the charge density wave order parameter, we explore the field dependence of the superstructure in CsV3Sb5 under the application of a $\mu_{0}H=$ 28 T pulsed magnetic field. ## II Methods ### II.1 Structural solutions Synchrotron x-ray diffraction experiments were carried out at the QM2 beam line at CHESS. The incident x-ray wavelength of $\lambda=0.41328$ $\rm\AA$ was selected using a double-bounce diamond monochromator. A stream of cold flowing helium gas was used to cool the sample. The diffraction experiment was conducted in transmission geometry using a 6-megapixel photon-counting pixel- array detector with a silicon sensor layer. Data was collected in full 360∘ sample rotations with a step size of 0.1∘. Scattering planes in reciprocal space were visualized using the NeXpy software package. The diffraction data was indexed and integrated using the APEX3 software package including absorption and extinction correction. Crystallographic structural solutions were determined using the SHELX software package [23]. An unsupervised machine learning tool using Gaussian Mixture Models called X-ray Temperature Clustering (X-TEC) [24] was used to identify classes of peaks in reciprocal space with distinct temperature dependencies. ### II.2 Pulsed magnetic field experiments Pulsed magnetic field experiments were carried out at the beam line 6 ID-C of the Advanced Photon Source at Argonne National Laboratory using a dual cryostat, single solenoid pulsed-magnet system described in Ref. [25]. X-rays with a wavelength of $\lambda=0.56356$ $\rm\AA$ were used, and the crystallographic $c$-axis of CsV3Sb5 was aligned parallel to the direction of the incident x-ray beam and the direction of the magnetic field. A low-noise Lambda 750K detector was used with $55\times 55$ $\mu$ m2 CdTe pixels providing near 100% efficiency. The detector was run in 12-bit mode at a 1 kHz repetition rate with XRD pattern collected every 1 ms. Multiple frames were collected over the course of repeated 9 ms pulses with a maximum magnetic field of $H_{\mathrm{max}}=28$ T. Figure 2: Refined structural models for KV3Sb5 and RbV3Sb5. (a) shows the z = 0.5 layer comprised of vanadium and antimony atoms of the refined room temperature structure of KV3Sb5 and RbV3Sb5 in the hexagonal space group $P6/mmm$. (b) and (c) show the suggested TrH model for the 2$\times$2$\times$2 superstructure in the orthogonal space group $Fmmm$. The vanadium planes at $z=0$ and $z=0.5$ in the $Fmmm$ structure are displayed. ## III Experimental Results ### III.1 Structure of KV3Sb5 and RbV3Sb5 We begin by discussing diffraction experiments on KV3Sb5 at $T=10$ K. Scattering data indicate a 2$\times$2$\times$2 superstructure, and, similar to prior measurements [7], the area detector resolution does not allow for the momentum-space resolution of twinning in the distorted state. Recent scanning optical studies have shown the presence of three structural twins that form below the CDW transition [17]. These domains have their principle axes rotated by 120∘ in the ab-plane, consistent with an underlying orthorhombic lattice. This would imply a pseudohexagonal twinning with three orthorhombic twins that are rotated 120∘ with respect to each other in the $ab$-plane, which in turn raises the apparent symmetry of the diffraction pattern to hexagonal. Theoretical investigations identified the space groups $Cmcm$ [26] and $Fmmm$ [10] as possible starting orthorhombic solutions to the twinned pseudohexagonal structure. As reflection intensities indicate a Laue class of $mmm$, model refinements were performed in both the face-centered orthorhombic space group $Fmmm$ and the base-centered space group $Cmmm$. Refinement in both space groups yields identical deformations of the vanadium layers with the distinction being that $Cmmm$ allows for unique distortions within neighboring vanadium layers. Refinement in $Cmmm$ also implicitly allows for any intensity at positions forbidden in $Fmmm$ to be accounted for outside of twinning models. Analysis in these two possible groups gave comparable or slightly better $R_{1}$ values for the higher symmetry $Fmmm$ space group, and this group was used for the final low-temperature structural determination. KV3Sb5 data refined in $Fmmm$ is well-represented by two structural models (Tab. 1) consisting of either a staggered TrH ($R_{1}=6.22$, Fig. 2) or a staggered SoD distortion ($R_{1}=6.31$) of the vanadium layers. We note here that the assignments of TrH or SoD-type distortions in this paper are based on the dominant displacement type of V atoms, and that in all solutions there is detectable variance amongst V-V distances within a single plane (ie. there is not a single V-V distance in a given distorted kagome plane). This is further parameterized in Discussion section of this manuscript. The pattern of a given distortion type is shifted by half of a lattice constant ($0.5a$) with respect to neighboring layers, and the two solutions differ in the sign of the displacement vectors of the vanadium atoms in the $ab$-plane. X-ray refinement alone is unable to discriminate between the distortion types, which is a result of the previously discussed twinning. However, recent NMR studies identify a TrH-type distortion and allow us to break the degeneracy between these solutions [18, 19]. The resulting low-temperature structure generated from the staggered TrH solution of KV3Sb5 is plotted in Fig. 2 (b). This is consistent with recent studies [13, 27, 9], and crystallographic information files (CIF) for both TrH and SoD distortion types can be found in the supplemental information. After collecting data at 10 K, the sample was warmed to 290 K and the high temperature structure measured. This undistorted state was refined in $P6/mmm$ at 290 K with bond lengths illustrated in Fig. 2 (a). Similar analysis was employed for the x-ray diffraction data of RbV3Sb5, where RbV3Sb5 also forms a 2$\times$2$\times$2 superstructure in the CDW state. At $T=10$ K, the data is again best represented when indexed in the space group $Fmmm$, and our refinements show superior solutions using a staggered TrH model ($R_{1}=6.40$) versus a staggered SoD model ($R_{1}=7.41$). This is consistent with NMR results [18], and the TrH solution is selected as the correct structure for presentation and the resulting bond lengths in the vanadium network are shown in Fig. 2 (b). For reference, CIF files for both TrH and SoD distortion types can also be found in the supplemental information [28]. After measurement at 10 K, the 290 K structure was determined at 290 K in $P6/mmm$ with bond lengths illustrated in Fig. 2 (a). ### III.2 Structure of CsV3Sb5 Figure 3: Temperature dependence of the unit cell parameters of CsV3Sb5 upon traversing the CDW transition. The evolution of the $a$-, $b$-, and $c$-axes are plotted in panel (a). (b) shows $a$ and $b$ where $a_{\mathrm{hex}}=a/2$ for $T\leq 90$ K and $b_{\mathrm{hex}}=\sqrt{b^{2}+4a_{hex}^{2}}/4$ for $T\leq 90$ K and (c) shows $c$ where $c_{\mathrm{hex}}=c/2$ for $T=90$ K and $c_{\mathrm{hex}}=c/4$ for $T<90$ K. The subcell parameters undergo no significant changes. Figure 4: X-TEC analysis of CsV3Sb5 diffraction data. (a) X-TEC identifies three distinct temperature trajectories (blue, green, and yellow clusters) whose intensities exhibit a sharp onset near the CDW transition temperature. The lines denote the mean and shading denotes the standard deviation of the re-scaled (z-scored) intensities $\tilde{I}(T)$ within each cluster. The grey clusters identify the diffuse background scattering. (b) The average intensity of the blue, green, and yellow cluster reveals the different onset temperatures for the $2\times 2\times 4$ CDW ($T_{2\times 2\times 4}\approx 85K$) and $2\times 2\times 2$ CDW ($T_{2\times 2\times 2}\approx 90K$) state and the $2\times 2\times 1$ state emerging between 95 K and 80 K. $\Delta I(T)$ is the average intensity after subtracting its minimum value. (c) A section of the (H, 2.5, L) plane, in reciprocal lattice units (r.l.u.) and indexed in P6/mmm, showing the pixels colored by the cluster assignments of their intensities. (d,e,f) Distribution of the out-of-plane momentum $L$ of the peaks in the respective clusters. Now turning to analysis of the structure of CsV3Sb5, temperature dependent synchrotron x-ray diffraction experiments were carried out on a single crystal of CsV3Sb5. The final data set was collected at $T=290$ K where data were indexed in the hexagonal space group $P6/mmm$. The Bragg peaks of the CsV3Sb5 crystal studied are narrow with minimal stacking disorder, though the crystallinity is anisotropic. Quantifying this with Gaussian fits to representative Bragg peaks shows minimum correlation lengths in the $ab$-plane of $\xi_{min}=890$ $\rm\AA$ or $\xi_{min}=1000$ $\rm\AA$ (depending on the pixel-limited cut direction) and correlation lengths of $\xi=400$ $\rm\AA$ along the $c$-axis, indicating excellent crystallinity of the sample [28]. The lower correlation length along the $c$-axis (interlayer direction) is typical in layered materials and beyond the resolution of experiments using relaxed collimation. Figure 5: Refined structural model for CsV3Sb5. (a) shows the kagome network (z = 0.5) of the 290 K room temperature structure in the hexagonal space group $P6/mmm$. (b), (c), (d) and (e) represent a single-phase model of the 2$\times$2$\times$4 superstructure in the orthorhombic space group $Cmmm$. Three kagome layers form with TrH deformations that are capped by a layer with a predominant SoD deformation. Temperature-dependent diffraction data were collected on warming after using an initial cooling rate of approximately 10 K/min to cool to 10 K. Data were then collected by warming at $\approx 5$ K/min with pauses for measurement scans (12 min/scan) performed at select temperatures. Collecting a data series using this cooling/warming profile show that at temperatures below $T=94$ K, superstructure reflections appear as expected. Just below the CDW transition, a clear staging behavior is observed. At 90 K, superlattice reflections consistent with a 2$\times$2$\times$2 supercell are observed due to the appearance of half-integer $L$ peaks as well as half-integer peaks in the ($H$, $K$)-plane. Below this temperature, at 85 K, weak quarter-integer $L$ peaks also appear, signifying the onset of an enlarged 2$\times$2$\times$4 supercell. The cell parameters through the CDW transition are plotted in Fig. 3 Figure 6: Magnetic field dependence of x-ray diffraction of the Q=(0.5,0.5,0.25) CDW superlattice reflection in CsV3Sb5 in a pulsed magnetic field experiment with $H_{\mathrm{max}}=28$ T at $T=11$ K and $T=82.5$ K. (a) and (b) CDW reflection at $T=82.5$ K and $t=5$ ms from 1 sweep (1 ms exposure in (a)) and integrated over 61 sweeps (61 ms exposure in (b)). (c) Magnetic field $H$ as a function of the experiment time in milliseconds. (d) Sum of the integrated intensity of the (0.5,0.5,0.25) superlattice reflection from a total of 45 pulsed field sweeps at $T=11$ K and 61 sweeps at $T=82.5$ K. It is clear that the nature of the charge density wave and the potential coexistence/competition between different states involves subtle interactions and changes in the diffraction patterns. In addition, the large temperature- dependent and Q-dependent data sets contain substantial quantities of information that can be overlooked during standard data reduction and integration processes. Recently, an unsupervised machine learning tool called X-ray Temperature Clustering (X-TEC) was developed to enable analysis of parametric, complex data sets. X-TEC uses a Gaussian Mixture Model to identify peaks in reciprocal space with distinct temperature dependencies. In order to distinguish the functional form of the intensity-temperature trajectory rather than their magnitudes, the intensity of each momenta $\vec{q}$ is re-scaled (z-scored) as $\tilde{I}_{\vec{q}}(T)=\left(I_{\vec{q}}(T)-\mu_{\vec{q}}\right)/\sigma_{\vec{q}}$ where $\mu_{\vec{q}}$ is the mean over temperature $T$, and $\sigma_{\vec{q}}$ is the standard deviation in $T$ [24]. By applying X-TEC on the CsV3Sb5 data, $\sim 15,000$ peaks were identified with three distinct temperature trajectories color-coded as yellow, green and blue clusters. These clusters exhibit an onset behavior around the CDW transition temperature as shown in Figs. 4(a) and (b). By looking at the locations of these clusters in reciprocal space (Fig. 4(c)), we find that all three clusters have the same in-plane momenta at $(H,K)\equiv(0.5,0.5)$, but differ in their out-of-plane momentum $L$. While the green clusters are located at $L=1/4$ (Fig. 4(e)) corresponding to the $2\times 2\times 4$ structure, the blue clusters derive from the $L=1/2$ peaks (Fig. 4(f)) of the $2\times 2\times 2$ structure, and the yellow clusters to an additional class of $L=0$ peaks (Fig. 4 (d)) representing a $2\times 2\times 1$ structure. From their intensity-temperature trajectories (Fig. 4(b)), we find that the onset of CDW proceeds in two stages. First, the $2\times 2\times 1$ and $2\times 2\times 2$ cells appear below temperature $\approx 95$ K, followed by the onset of a $2\times 2\times 4$ state at $\approx 85$ K. Remarkably, with the onset of the $2\times 2\times 4$ peaks, the $2\times 2\times 1$ peaks are strongly suppressed and vanish below 80 K, and, additionally, the $2\times 2\times 2$ peaks initially decrease with the onset of $2\times 2\times 4$ order and then recover their intensity at lower temperatures. This suggests an exchange of scattering weight between order parameters. Table 1: Parameters of structural models for the room temperature and the low-temperature structures of KV3Sb5, RbV3Sb5, and CsV3Sb5. All .cif files from the refinements can be found in the supplemental information [28]. | KV3Sb5 | RbV3Sb5 | CsV3Sb5 ---|---|---|--- Wavelength ($\rm\AA$) | 0.41328 Temperature (K) | 290 | 10 | 290 | 10 | 290 | 90 | 11 Space group | $P6/mmm$ | $Fmmm$ | $P6/mmm$ | $Fmmm$ | $P6/mmm$ | $Fmmm$ | $Cmmm$ Type | | $2\times 2\times 2$ | | $2\times 2\times 2$ | | $2\times 2\times 2$ | $2\times 2\times 4$ Distortion | | TrH | SoD | | TrH | SoD | | TrH | SoD | TrH/SoD $N_{\mathrm{measured}}$ | 2751 | 17014 | 2239 | 16168 | 2888 | 6788 | 24048 $N_{\mathrm{independent}}$ | 232 | 2401 | 201 | 2444 | 238 | 2159 | 5002 $N_{\mathrm{sig.}}$ [$I>2\sigma(I)$] | 231 | 1617 | 201 | 1906 | 238 | 641 | 3098 $a$ ($\rm\AA$) | 5.4260(9) | 10.9576(9) | 5.4941(6) | 11.0065(9) | 5.5236(6) | 11.0275(9) | 11.005(2) $b$ ($\rm\AA$) | 5.4260(9) | 18.9813(16) | 5.4941(6) | 19.0575(16) | 5.5236(6) | 19.1082(16) | 19.046(4) $c$ ($\rm\AA$) | 8.845(2) | 17.8552(15) | 9.1071(14) | 18.1010(15) | 9.3623(15) | 18.6213(15) | 37.176(7) $\rho$ (g/cm3) | 5.90 | 5.73 | 5.91 | 5.93 | 6.00 | 6.06 | 7.22 $R_{1}>4\sigma$ (%) | 8.62 | 6.22 | 6.31 | 10.42 | 6.40 | 7.41 | 5.91 | 7.71 | 7.73 | 10.0 $R_{1}$ all (%) | 8.63 | 6.44 | 6.56 | 10.42 | 6.74 | 7.82 | 5.91 | 9.61 | 9.62 | 11.55 Twin fraction 1 | - | 0.35 | 0.36 | - | 0.32 | 0.32 | - | 0.31 | 0.30 | 0.31 Twin fraction 2 | - | 0.33 | 0.33 | - | 0.33 | 0.33 | - | 0.36 | 0.36 | 0.34 Twin fraction 3 | - | 0.32 | 0.31 | - | 0.34 | 0.34 | - | 0.33 | 0.34 | 0.35 The 2$\times$2$\times$4 superstructure persists down to temperatures of 11 K. We note here that the in-plane and out-of-plane correlation lengths of the CDW superlattice peaks here are identical to the primary structural peaks. This indicates that the 2$\times$2$\times$4 CDW state is long-range with correlation lengths constrained only by the sample’s underlying crystallinity. Notably, our diffraction data using this cooling/warming profile do not resolve anomalies at $T=60$ K as seen in magnetotransport experiments [29, 30] or a crossover to a 2$\times$2$\times$2 state similar to Ref. [13]. This difference potentially arises from subtle effects attributed to the different cooling profiles during measurement and potentially from sample dependent effects imparted during the crystal growth process. We now discuss the structural refinement in the 2$\times$2$\times$2 regime at $T=90$ K and the 2$\times$2$\times$4 superstructure at lower temperatures. Using the same twinning approach described earlier, the 2$\times$2$\times$2 state at 90 K is best fit in the $Fmmm$ group rendering either a staggered SoD or TrH distortion (CIFs provided in the supplemental material [28]). This is consistent with the low temperature state recently reported by Stahl et al. [13] where only a 2$\times$2$\times$2 superstructure was observed. Upon cooling into the 2$\times$2$\times$4 state, the data at 10 K are instead best refined in the lower symmetry, base-centered $Cmmm$ group. Fig. 5 shows the resulting model of the superstructure consisting of vanadium layers with both SoD and TrH deformations. The three TrH layers are staggered along the $c$-axis by $0.5a$ (Fig. 5 (d),(e),(f)) and are then capped by a SoD layer (Fig. 5 (c)). The model assumes a uniform phase at this temperature and represents the measured diffraction data well, especially considering the large number of significant reflections (3098 shown in Tab. 1). This mixed distortion-type result is consistent with previous refinements in a $P\overline{3}$ average structure containing both SoD and TrH layers [3] as well as recent angle-resolved photoemission data [21, 22]. Mixed reports from NMR measurements [19, 20] provide a picture interpreted as either SoD or TrH distortion types, and it is unclear whether a mixed distortion-type structure can also be modeled via the data. Moreover, our X-TEC analysis indicates that a 2$\times$2$\times$2 phase competes with and potentially coexists with the 2$\times$2$\times$4 phase at lowest temperatures and that the stacking sequence of the TrH- and SoD-like layers may be influenced by this coexistence. ### III.3 Pulsed magnetic field experiment Scanning tunneling microscopy measurements have reported that a magnetic field couples to the apparent Fourier weight attributed to CDW order at the surface in $A$V3Sb5 compounds [8], suggesting the presence of a chiral CDW state that breaks time reversal symmetry. As means of testing whether a magnetic field couples to the real component of the CDW order parameter in the bulk, the magnetic field dependence of the superlattice reflections was investigated. A pulsed magnetic field of $H_{\mathrm{pulse}}=28$ T was used, and the geometry of the inner bore of the magnet coil allows for the measurement of the Q$=(0.5,0.5,0.25)$ superlattice reflection in transmission geometry with the field applied parallel to the $c$-axis. Scattering data were collected at $T=11$ K and at $T=82.5$ K after cooling the sample into the CDW state under zero field conditions. Magnetic field pulses were applied that last for approximately 10 ms (half-sine, start to finish), and where the maximum magnetic field of 28 T is reached. Multiple pulses were used to improve measurement statistics. The Q$=(0.5,0.5,0.25)$ peak was measured at $T=11$ K for a total of 45 pulses and at $T=82.5$ K for a total of 61 pulses. For illustration, Fig. 6 (a) shows the result of 1 ms of data collection during a single field pulse on the Q$=(0.5,0.5,0.25)$ superlattice peak on a pixel array. Fig. 6 (b) then shows the same peak after a cumulative sum of 61 pulses. Each time bin corresponds to a field value shown in Fig. 6 (c). The integrated intensities of the Q$=(0.5,0.5,0.25)$ peak as a function of time were summed over the number of individual pulses to increase the signal- to-noise ratio with the results plotted in Fig. 6 (d). No significant change of the integrated intensity outside of the experimental error could be observed at either of the two temperatures explored. The pre-pulse fluctuations and sharp post-pulse decrease in the integrated intensity are attributed to mechanical vibrations. Within our experimental resolution, we do not observe any evidence of magnetic-field dependence of the superlattice peak intensity, consistent with a recent report [13] and signifying that the real component of the CDW order parameter is unchanged. ## IV Discussion and Conclusions Tab. 1 summarizes the parameters of the crystallographic refinements carried out on KV3Sb5, RbV3Sb5, and CsV3Sb5 crystals at key temperatures discussed in the preceding sections. All refined models represent the collected data well and the $R$-values are low considering the high number of significant reflections collected in synchrotron experiments. The CDW distortion results in an average V-V bond length that is nearly the same in all three AV3Sb5 variants with (V-V)avg=2.7022$\pm 0.022$ in A = K, (V-V)avg=2.6992$\pm 0.028$ in A = Rb, and (V-V)avg=2.6979$\pm 0.036$ in A = Cs. The standard deviations in these V-V distances highlight the fact that the assignments of TrH or SoD distortions are only effective classifications and that the V-V bond lengths have substantial variance within a given layer. The increase in the standard deviation of distorted bond lengths about a fixed average value likely indicates a progressively increasing instability of the TrH state as the A-site changes from K to Rb to Cs. Notably, the Cs system has the largest room temperature V-V distance, rendering its low-temperature reconstruction of the kagome network the most dramatic of the series. This is likely reflected in the larger heat capacity anomaly in the Cs system relative to the other two as well as its accommodation of a mixed TrH and SoD distortion. The mixed SoD and TrH state refined in the CDW state of CsV3Sb5 should also be viewed through the lens of the experimental parameters used in this study. Slower cooling or faster quenching may modify the resulting structure in the CDW state given the seeming competition with the $2\times 2\times 2$ TrH state and a newly observed $2\times 2\times 1$ state that both appear just below $T_{CDW}$. It is possible that materials with different levels of strain frozen during the growth process or dilute impurity content may modify this behavior as well. Subtle disorder along the interlayer direction in the form of variable crystallinity between samples does not seem to play a dominant role as we have screened many crystals with variable quality and they all exhibit the same $2\times 2\times 4$ low-temperature superlattice using this same cooling profile. In summary, we studied the structural deformations related to the CDW transitions in the systems CsV3Sb5, KV3Sb5, and RbV3Sb5 using synchrotron x-ray diffraction. Our diffraction data allow for twinned models of orthorhombic superstructures in the CDW state. We find 2$\times$2$\times$2 superstructures in KV3Sb5 and RbV3Sb5 and our crystallographic refinements allow us to parameterize the distortion of vanadium-based kagome layers in a staggered TrH state. Temperature dependent diffraction data from the CsV3Sb5 system indicate a complex evolution of the structural superlattice. Initially, a 2$\times$2$\times$1 state appears along with a 2$\times$2$\times$2 state at temperatures below 90 K, and this is followed by the onset of a 2$\times$2$\times$4 superlattice below 85 K. Structural refinement of this larger superstructure reveals a three-layer staggered TrH deformation capped by a layer with a SoD type distortion. Scattering data from CsV3Sb5 measured under a 28 T pulsed magnetic field fails to resolve a change in the CDW superlattice intensity, setting bounds on the response of the real component of the CDW state to the application of a magnetic field. ## V Acknowledgments SDW acknowledges helpful discussions with Turan Birol, Rafael Fernandes, and Binhai Yan. This work was supported by the National Science Foundation (NSF) through Enabling Quantum Leap: Convergent Accelerated Discovery Foundries for Quantum Materials Science, Engineering and Information (Q-AMASE-i): Quantum Foundry at UC Santa Barbara (DMR-1906325). The research made use of the shared facilities of the NSF Materials Research Science and Engineering Center at UC Santa Barbara (DMR- 1720256). The UC Santa Barbara MRSEC is a member of the Materials Research Facilities Network. (www.mrfn.org). This work is based upon research conducted at the Center for High Energy X-ray Sciences (CHEXS) which is supported by the National Science Foundation under award DMR-1829070. The high-field pulsed magnet and a choke coil were installed at the Advanced Photon Source through a partnership with International Collaboration Center at the Institute for Materials Research (ICC-IMR) and Global Institute for Materials Research Tohoku (GIMRT) at Tohoku University. The development and use of X-TEC by K.M. and E.-A.K. was supported by the U.S. Department of Energy, Office of Basic Energy Sciences, Division of Materials Sciences and Engineering. KM and E-AK are supported in part by the Gordon and Betty Moore Foundation’s EPiQS Initiative, Grant GBMF10436. ## References * Ortiz _et al._ [2019] B. R. Ortiz, L. C. Gomes, J. R. Morey, M. Winiarski, M. Bordelon, J. S. Mangum, I. W. Oswald, J. A. Rodriguez-Rivera, J. R. Neilson, S. D. Wilson, _et al._ , New kagome prototype materials: discovery of kv3sb5, rbv3sb5, and csv3sb5, Physical Review Materials 3, 094407 (2019). * Ortiz _et al._ [2020] B. R. Ortiz, S. M. Teicher, Y. Hu, J. L. Zuo, P. M. Sarte, E. C. Schueller, A. M. Abeykoon, M. J. Krogstad, S. Rosenkranz, R. Osborn, _et al._ , Csv3sb5: A $z$2 topological kagome metal with a superconducting ground state, Physical Review Letters 125, 247002 (2020). * Ortiz _et al._ [2021a] B. R. Ortiz, P. M. Sarte, E. M. Kenney, M. J. Graf, S. M. Teicher, R. Seshadri, and S. D. Wilson, Superconductivity in the z 2 kagome metal kv 3 sb 5, Physical Review Materials 5, 034801 (2021a). * Park _et al._ [2021] T. Park, M. Ye, and L. Balents, Electronic instabilities of kagome metals: Saddle points and landau theory, Phys. Rev. B 104, 035142 (2021). * Yang _et al._ [2020] S.-Y. Yang, Y. Wang, B. R. Ortiz, D. Liu, J. Gayles, E. Derunova, R. Gonzalez-Hernandez, L. Šmejkal, Y. Chen, S. S. Parkin, _et al._ , Giant, unconventional anomalous hall effect in the metallic frustrated magnet candidate, kv3sb5, Science advances 6, eabb6003 (2020). * Yu _et al._ [2021] F. Yu, T. Wu, Z. Wang, B. Lei, W. Zhuo, J. Ying, and X. Chen, Concurrence of anomalous hall effect and charge density wave in a superconducting topological kagome metal, Physical Review B 104, L041103 (2021). * Ortiz _et al._ [2021b] B. R. Ortiz, S. M. Teicher, L. Kautzsch, P. M. Sarte, N. Ratcliff, J. Harter, J. P. Ruff, R. Seshadri, and S. D. Wilson, Fermi surface mapping and the nature of charge-density-wave order in the kagome superconductor CsV3Sb5, Phys. Rev. X 11, 041030 (2021b). * Jiang _et al._ [2021] Y.-X. Jiang, J.-X. Yin, M. M. Denner, N. Shumiya, B. R. Ortiz, G. Xu, Z. Guguchia, J. He, M. S. Hossain, X. Liu, _et al._ , Unconventional chiral charge order in kagome superconductor kv3sb5, Nature Materials 20, 1353 (2021). * Tan _et al._ [2021] H. Tan, Y. Liu, Z. Wang, and B. Yan, Charge density waves and electronic properties of superconducting kagome metals, Physical review letters 127, 046401 (2021). * Christensen _et al._ [2021] M. H. Christensen, T. Birol, B. M. Andersen, and R. M. Fernandes, Theory of the charge density wave in a v 3 sb 5 kagome metals, Physical Review B 104, 214513 (2021). * Liang _et al._ [2021] Z. Liang, X. Hou, F. Zhang, W. Ma, P. Wu, Z. Zhang, F. Yu, J.-J. Ying, K. Jiang, L. Shan, Z. Wang, and X.-H. Chen, Three-dimensional charge density wave and surface-dependent vortex-core states in a kagome superconductor ${\mathrm{csv}}_{3}{\mathrm{sb}}_{5}$, Phys. Rev. X 11, 031026 (2021). * Li _et al._ [2021] H. Li, T. T. Zhang, T. Yilmaz, Y. Y. Pai, C. E. Marvinney, A. Said, Q. W. Yin, C. S. Gong, Z. J. Tu, E. Vescovo, C. S. Nelson, R. G. Moore, S. Murakami, H. C. Lei, H. N. Lee, B. J. Lawrie, and H. Miao, Observation of unconventional charge density wave without acoustic phonon anomaly in kagome superconductors ${A\mathrm{V}}_{3}{\mathrm{sb}}_{5}$ ($a=\mathrm{Rb}$, cs), Phys. Rev. X 11, 031050 (2021). * Stahl _et al._ [2022] Q. Stahl, D. Chen, T. Ritschel, C. Shekhar, E. Sadrollahi, M. Rahn, O. Ivashko, M. v. Zimmermann, C. Felser, and J. Geck, Temperature-driven reorganization of electronic order in csv3sb5, Physical Review B 105, 195136 (2022). * Xiao _et al._ [2022] Q. Xiao, Y. Lin, Q. Li, X. Zheng, W. Xia, S. Zhang, Y. Guo, J. Feng, and Y. Peng, Coexistence of multiple stacking charge density waves in kagome superconductor ${\mathrm{csv}}_{3}{\mathrm{sb}}_{5}$, arXiv preprint arXiv:2201.05211 (2022). * Kautzsch _et al._ [2022] L. Kautzsch, Y. M. Oey, H. Li, Z. Ren, B. R. Ortiz, R. Seshadri, J. Ruff, Z. Wang, I. Zeljkovic, and S. D. Wilson, Incommensurate charge-stripe correlations in the kagome superconductor csv3sb5-xsnx, arXiv preprint arXiv:2207.10608 (2022). * Li _et al._ [2022] H. Li, H. Zhao, B. R. Ortiz, T. Park, M. Ye, L. Balents, Z. Wang, S. D. Wilson, and I. Zeljkovic, Rotation symmetry breaking in the normal state of a kagome superconductor kv3sb5, Nature Physics 18, 265 (2022). * Xu _et al._ [2022] Y. Xu, Z. Ni, Y. Liu, B. R. Ortiz, Q. Deng, S. D. Wilson, B. Yan, L. Balents, and L. Wu, Three-state nematicity and magneto-optical kerr effect in the charge density waves in kagome superconductors, Nature Physics 10.1038/s41567-022-01805-7 (2022). * Frassineti _et al._ [2023] J. Frassineti, P. Bonfà, G. Allodi, E. Garcia, R. Cong, B. R. Ortiz, S. D. Wilson, R. De Renzi, V. F. Mitrović, and S. Sanna, Microscopic nature of the charge-density wave in the kagome superconductor rbv3sb5, Physical Review Research 5, L012017 (2023). * Wang _et al._ [2022] Y. Wang, T. Wu, Z. Li, K. Jiang, and J. Hu, The structure of kagome superconductors av3sb5 in the charge density wave states, arXiv preprint arXiv:2210.07585 (2022). * Luo _et al._ [2022] J. Luo, Z. Zhao, Y. Z. Zhou, J. Yang, A. F. Fang, H. T. Yang, H. J. Gao, R. Zhou, and G.-q. Zheng, Possible star-of-david pattern charge density wave with additional modulation in the kagome superconductor csv3sb5, npj Quantum Materials 7, 30 (2022). * Kang _et al._ [2022] M. Kang, S. Fang, J. Yoo, B. R. Ortiz, Y. Oey, S. H. Ryu, J. Kim, C. Jozwiak, A. Bostwick, E. Rotenberg, E. Kaxiras, J. Checkelsky, S. D. Wilson, J.-H. Park, and R. Comin, Microscopic structure of three-dimensional charge order in kagome superconductor av3sb5 and its tunability, arXiv preprint arXiv:2202.01902 (2022). * Hu _et al._ [2022] Y. Hu, X. Wu, B. R. Ortiz, X. Han, N. C. Plumb, S. D. Wilson, A. P. Schnyder, M. Shi, _et al._ , Coexistence of trihexagonal and star-of-david pattern in the charge density wave of the kagome superconductor a v 3 sb 5, Physical Review B 106, L241106 (2022). * Sheldrick [2008] G. M. Sheldrick, A short history of shelx, Acta Crystallographica Section A: Foundations of Crystallography 64, 112 (2008). * Venderley _et al._ [2022] J. Venderley, K. Mallayya, M. Matty, M. Krogstad, J. Ruff, G. Pleiss, V. Kishore, D. Mandrus, D. Phelan, L. Poudel, A. G. Wilson, K. Weinberger, P. Upreti, M. Norman, S. Rosenkranz, R. Osborn, and E.-A. Kim, Harnessing interpretable and unsupervised machine learning to address big data from modern X-ray diffraction, Proceedings of the National Academy of Sciences 119, e2109665119 (2022). * Islam _et al._ [2012] Z. Islam, D. Capatina, J. P. Ruff, R. K. Das, E. Trakhtenberg, H. Nojiri, Y. Narumi, U. Welp, and P. C. Canfield, A single-solenoid pulsed-magnet system for single-crystal scattering studies, Review of Scientific Instruments 83, 035101 (2012). * Subedi [2022] A. Subedi, Hexagonal-to-base-centered-orthorhombic 4 q charge density wave order in kagome metals kv 3 sb 5, rbv 3 sb 5, and csv 3 sb 5, Physical Review Materials 6, 015001 (2022). * Ratcliff _et al._ [2021] N. Ratcliff, L. Hallett, B. R. Ortiz, S. D. Wilson, and J. W. Harter, Coherent phonon spectroscopy and interlayer modulation of charge density wave order in the kagome metal csv 3 sb 5, Physical Review Materials 5, L111801 (2021). * [28] See Supplemental Information for further details. * Chen _et al._ [2021] H. Chen, H. Yang, B. Hu, Z. Zhao, J. Yuan, Y. Xing, G. Qian, Z. Huang, G. Li, Y. Ye, _et al._ , Roton pair density wave in a strong-coupling kagome superconductor, Nature 599, 222 (2021). * Xiang _et al._ [2021] Y. Xiang, Q. Li, Y. Li, W. Xie, H. Yang, Z. Wang, Y. Yao, and H.-H. Wen, Twofold symmetry of c-axis resistivity in topological kagome superconductor csv3sb5 with in-plane rotating magnetic field, Nature communications 12, 1 (2021).
# Orbits of linear series on the projective line Anand Deopurkar Anand Deopurkar, Mathematical Sciences Institute, Australian National University<EMAIL_ADDRESS>and Anand Patel Anand Patel, Department of Mathematics, Oklahoma State University<EMAIL_ADDRESS> ###### Abstract. We compute the equivariant fundamental class of the orbit closure of a linear series on the projective line. We also describe the boundary of the orbit closure and how the orbits specialise in one parameter families. ## 1\. Introduction Let $M$ be a variety with the action of an algebraic group $G$. Given a point $m\in M$, let $\operatorname{Orb}(m)\subset M$ be the closure of the $G$-orbit of $m$. The fundamental class of $\operatorname{Orb}(m)$ in the equivariant Chow ring of $M$ encodes fascinating geometric information. The class is particularly meaningful when the $G$-orbits have geometric significance. The main result of this paper (Theorem 1.1) is a complete and explicit expression for these classes for $M=\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ with the natural action of $G=\operatorname{GL}(r+1)\times\operatorname{GL}(2)$. Here $\mathbf{k}$ is an algebraically closed field in which $d!\neq 0$. The orbits of the $\operatorname{GL}(r+1)\times\operatorname{GL}(2)$ action on $\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ represent isomorphism classes of linear series of rank $r$ and degree $d$ on the projective line. Their fundamental class turns out to depend only on the ramification profile. For example, when $r=1$ and the linear series is base- point free, the class is a function of the multiplicities in the ramification divisor. Likewise, when $r=2$ and the linear series is base-point free, the class depends on the multiplicities of cusps and hyperflexes. The group $\operatorname{GL}(r+1)$ acts freely on the open subset of $\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ consisting of injective maps. The quotient is the Grassmannian ${\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$. So we have a surjective homomorphism (1) $A^{*}_{\operatorname{GL}(r+1)\times\operatorname{GL}(2)}(\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2}))\to A^{*}_{\operatorname{GL}(2)}({\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})).$ Taking images under this map gives classes of $\operatorname{GL}(2)$-orbits in the Grassmannian. Forgetting the group action gives a further homomorphism (2) $A^{*}_{\operatorname{GL}(2)}({\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2}))\to A^{*}({\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})).$ Taking images under this map gives non-equivariant classes of the $\operatorname{GL}(2)$-orbits. All of these, including the non-equivariant ones, were unknown. Table 1 shows the weighted (multiplied by the order of the stabiliser) orbit classes for all base-point free pencils of quartics. Ramification profile | Weighted orbit class in the Schubert basis ---|--- $2,2,2,2,2,2$ | $24\cdot\Omega_{3}+48\cdot\Omega_{2,1}$ $3,2,2,2,2$ | $16\cdot\Omega_{3}+40\cdot\Omega_{2,1}$ $4,2,2,2$ | $12\cdot\Omega_{3}+24\cdot\Omega_{2,1}$ $3,3,2,2$ | $8\cdot\Omega_{3}+32\cdot\Omega_{2,1}$ $3,3,3$ | $24\cdot\Omega_{2,1}$ $4,3,2$ | $4\cdot\Omega_{3}+16\cdot\Omega_{2,1}$ Table 1. The weighted orbit classes of base-point free pencils of quartics in ${\bf Gr}(2,\operatorname{Sym}^{4}\mathbf{k}^{2})$ as a function of the ramification profile. Many mathematicians have studied equivariant and non-equivariant orbit classes. We were inspired mainly by the work of Aluffi and Faber [3, 2, 1] on the (non-equivariant) classes of orbits of points on ${\bf P}^{1}$ and curves in ${\bf P}^{2}$. There has been exciting recent progress on extending their work to the equivariant setting; see [13] for points on ${\bf P}^{1}$ and quartic curves in ${\bf P}^{2}$, [12] for hyperplane arrangements in ${\bf P}^{n}$, and [6] for generic cubics in ${\bf P}^{3}$. There is also significant work in a more representation theoretic context. See, for example, [11, 5] for classes of orbits of quiver representations and [4] for orbits of matrices. ### 1.1. Main theorem To state the main result, we need some notation. We have an identification (3) $A^{*}_{\operatorname{GL}(r+1)\times\operatorname{GL}(2)}(\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2}))=\mathbf{Z}[\mu_{0},\dots,\mu_{r},\omega_{1},\omega_{2}]^{S_{r+1}\times S_{2}},$ where the $\mu_{i}$’s and the $\omega_{j}$’s are the “Chern roots” of the tautological representations of $\operatorname{GL}(r+1)$ and $\operatorname{GL}(2)$, respectively. Given integers $a_{0}<\dots<a_{r}$ in $\\{0,\dots,d\\}$, set $\phi_{a_{0},\dots,a_{r}}(\mu;\omega)=\prod_{j=0}^{r}\prod_{i}((d-i)\omega_{1}+i\omega_{2}-\mu_{j}),$ where the index $i$ ranges over $\\{0,\dots,d\\}-\\{a_{0},\dots,a_{r}\\}$. Set $\psi_{a_{0},\dots,a_{r}}(\mu;\omega)=\phi_{a_{0},\dots,a_{r}}(\mu;\omega_{1},\omega_{2})-\phi_{a_{0},\dots,a_{r}}(\mu;\omega_{2},\omega_{1}).$ Note that $\psi$ is symmetric in the $\mu$’s and anti-symmetric in the $\omega$’s. Given an injective $s\in\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$, let $S\subset\operatorname{Sym}^{d}\mathbf{k}^{2}$ be its image. Associated to $S$ and a point $p\in{\bf P}^{1}$, we have the vanishing sequence $a_{0}(p)<\dots<a_{r}(p)$ characterised by the property that for each $i$, the space $S$ contains a section vanishing to order exactly $a_{i}(p)$ at $p$. For all but finitely many points $p\in{\bf P}^{1}$, the vanishing sequence is given by $a_{i}(p)=i$. Let $B\subset{\bf P}^{1}$ be the points where it differs. The ramification profile of $S$ is the multi-set $\\{(a_{0}(b),\dots,a_{r}(b))\mid b\in B\\}.$ ###### Theorem 1.1 (Theorem 5.5). Let $\mathbf{k}$ be an algebraically closed field in which $d!\neq 0$. Let $s\in\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ be an injective map with image $S\subset\operatorname{Sym}^{d}\mathbf{k}^{2}$. Let $\\{(a_{i}(b))\mid b\in B\\}$ be the ramification profile of $S$. Let $\Gamma\subset\operatorname{PGL}(2)$ be the stabiliser of $S\subset\operatorname{Sym}^{d}\mathbf{k}^{2}$, and assume that it is finite. Let $\operatorname{Orb}(s)$ be the closure of the $\operatorname{GL}(r+1)\times\operatorname{GL}(2)$ orbit of $s$. Then the class $|\Gamma|\cdot[\operatorname{Orb}(s)]\in A^{(r+1)(d-r)-3}_{\operatorname{GL}(r+1)\times\operatorname{GL}(2)}(\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2}))$ is given by $|\Gamma|\cdot[\operatorname{Orb}(s)]=\frac{1}{(\omega_{1}-\omega_{2})^{3}}\sum_{b\in B}\psi_{a_{0}(b),\dots,a_{r}(b)}(\mu;\omega)+\frac{2-|B|}{(\omega_{1}-\omega_{2})^{3}}\cdot\psi_{0,\dots,r}(\mu;\omega).$ Despite the apparent denominators, the expression must be a polynomial (although this is not obvious). To get the $\operatorname{GL}(2)$-equivariant class of $\operatorname{Orb}(S)\subset{\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$, we apply the homomorphism (1), which sends the $\mu$’s to the Chern roots of the universal sub-bundle. To get the non-equivariant class of $\operatorname{Orb}(S)\subset{\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$, we further substitute $\omega_{1}=\omega_{2}=0$. ###### Remark 1.2. Although Theorem 1.1 gives the non-equivariant classes by the series of substitutions mentioned above, it does not give a simple expression for it. Indeed, the substitution $\omega_{1}=\omega_{2}=0$ can only be performed after dividing by the denominator in Theorem 1.1, and the result of the division is not obvious. ###### Remark 1.3. Given $[S]\in{\bf Gr}(r+1,\operatorname{Sym}^{d}W)$, we can consider $[S]$ as a point of the projective space ${\bf P}\left(\wedge^{r+1}\operatorname{Sym}^{d}W\right)$ by applying the Plucker embedding. The class of $\operatorname{Orb}([S])$ in ${\bf P}\left(\wedge^{r+1}\operatorname{Sym}^{d}W\right)$ depends only on the multiplicities in the ramification divisor of $S$. The class of $\operatorname{Orb}([S])$ in ${\bf Gr}(r+1,\operatorname{Sym}^{d}W)$ is a richer invariant, sensitive to the full ramification profile. ### 1.2. Orbit closures The method of proof of the main theorem also yields a description of the boundary points of the orbit closure. ###### Theorem 1.4 (Theorem 4.5). The orbit closure $\operatorname{Orb}(S)\subset{\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$ is a union of finitely many orbits. In addition to the orbit of $S$, these are the orbits of subspaces of the form $\langle w_{1}^{d-a_{0}}w_{2}^{a_{0}},\dots,w_{1}^{d-a_{r}}w_{2}^{a_{r}}\rangle,$ where $a_{0}<\dots<a_{r}$ is the vanishing sequence of $S$ at some point of ${\bf P}^{1}$ and $w_{1},w_{2}$ is a basis of $\mathbf{k}^{2}$. At a generic point of ${\bf P}^{1}$, the vanishing sequence is given by $a_{i}=i$. The orbit of the corresponding linear series $\langle w_{1}^{d}w_{2}^{0},\dots,w_{1}^{d-r}w_{2}^{r}\rangle$ is $1$-dimensional. At a ramification point, the vanishing sequence is different, and the orbit of the corresponding linear series $\langle w_{1}^{d-a_{0}}w_{2}^{a_{0}},\dots,w_{1}^{d-a_{r}}w_{2}^{a_{r}}\rangle$ is $2$-dimensional. ### 1.3. Orbit specialisation Consider a one-parameter family of linear series in which some ramification points collide. Then the flat limit of the orbit closure contains the orbit closure of the limiting linear series. In a suitable family, the limit of the orbit closure contains the orbit closure of a unique other linear series, which we now describe. Let $A$ be the multi-set of vanishing sequences of the points that collide and $B$ the multi-set of vanishing sequences of the remaining points. Let $c=(c_{0},\dots,c_{r})$ be the vanishing sequence of the limiting linear series at the point of collision. Set $c^{\prime}=(d-c_{r},\dots,d-c_{0})$. Denote by $\operatorname{WOrb}$ the weighted orbit closure, namely $\operatorname{WOrb}=|\Gamma|\cdot\operatorname{Orb}.$ Recall that we can treat the class $[\operatorname{WOrb}]$ as a function of the ramification profile. ###### Theorem 1.5 (Proposition 6.8). In the notation above, we have the equality of $\operatorname{GL}(r+1)\times\operatorname{GL}(2)$-equivariant classes $[\operatorname{WOrb}(A\cup B)]=[\operatorname{WOrb}(A\cup\\{c^{\prime}\\})]+[\operatorname{WOrb}(B\cup\\{c\\})].$ The statement above is a special case of a large family of linear relations among the classes $[\operatorname{WOrb}]$; see Section 6 for more. By repeated collisions of ramification points, we can express an arbitrary orbit class as a non-negative linear combination of orbit classes of linear series with 3 ramification points (the expression is not unique). For example, in the case of quartic pencils, Figure 1 shows a sequence of collisions and the resulting complementary orbits. As a result, we get the following equality $\begin{split}\operatorname{WOrb}((0,2)\times 6)&=\\\ \operatorname{WOrb}((0,3)&\times 2,(1,2))+2\operatorname{WOrb}((0,2)\times 2,(1,4))+\operatorname{WOrb}((0,2)\times 2,(2,3)).\end{split}$ In the non-equivariant Chow ring of the Grassmannian, the above evaluates to the identity $24\cdot\Omega_{3}+48\cdot\Omega_{2,1}=\left(8\Omega_{3}+20\cdot\Omega_{2,1}\right)+2\cdot\left(8\cdot\Omega_{3}+8\cdot\Omega_{2,1}\right)+12\cdot\Omega_{2,1}.$ $(0,2)$$(0,2)$$(0,2)$$(0,2)$$(0,2)$$(0,2)$$(0,3)$$+$$(0,3)$$+$$(1,2)$$+$$(0,2)$$(0,2)$$(1,4)$$(0,2)$$(0,2)$$(1,4)$$(0,2)$$(0,2)$$(2,3)$ Figure 1. An example of orbit specialisation for a generic quartic pencil. Collisions of ramification points create a new ramification profile and a unique complementary orbit shown underneath. ### 1.4. Ideas in the proof To find the class of the closure of the $G$-orbit of $m\in M$, we use a complete orbit parametrisation for $m$. A complete orbit parametrisation consists of a proper $G$-variety $X$ with a dense orbit and an equivariant map $\phi\colon X\to M$ whose image contains $m$. Then, up to a scalar, the class we seek is $\phi_{*}[X]$. We begin with ${\bf P}^{3}={\bf P}\operatorname{Hom}(\mathbf{k}^{2},\mathbf{k}^{2})$ with the natural $\operatorname{GL}(2)$ action on the target $\mathbf{k}^{2}$ and the equivariant rational map $\phi\colon{\bf P}^{3}\dashrightarrow{\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$ defined by $m\mapsto m(S)$. The map is only rational because when $m$ is degenerate, the image $m(S)\subset\operatorname{Sym}^{d}\mathbf{k}^{2}$ need not be $(r+1)$-dimensional. We prove (Theorem 4.1) that $\phi$ extends to a regular map on a simple blow up: (4) $\widetilde{{\bf P}}^{3}\to{\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2}).$ This is the same blow-up that Aluffi and Faber use in [1] for $r=0$. We compute the push-forward using equivariant localisation as in [13, Appendix B]. We upgrade from the orbit class in ${\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$ to the orbit class in $\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ using general principles, outlined in Section 2. ### 1.5. Organisation In Section 2, we discuss equivariant orbit classes in general and classes on Grassmannians in particular. We also discuss general results about one parameter specialisations of orbits. In Section 3, we recall vanishing sequences, the Wronskian, and the ramification divisor of a linear series. In Section 4, we begin the proof of the main theorem by constructing a complete parametrisation for a linear series. In Section 5, we finish the proof of the main theorem with a localisation computation. In Section 6, we discuss one- parameter specialisations of orbits of linear series, and the resulting additive decomposition of the orbit class. ### 1.6. Conventions Throughout, $\mathbf{k}$ denotes an algebraically closed field of characteristic 0 or of characteristic $>d$. All schemes are of finite type over $\mathbf{k}$. A _variety_ is a reduced scheme, separated over $\mathbf{k}$. Given a vector space $V$, the notation $\underline{V}$ denotes the trivial vector bundle with fiber $V$ on the scheme that is clear from the context. Given a vector bundle $V$, the _projectivisation_ ${\bf P}V$ is the space of one-dimensional sub-bundles. Unless mentioned otherwise, _point_ means a closed point. Given a variety $X$ with a $G$-action, $A^{i}_{G}(X)$ denotes the $G$-equivariant Chow group of $X$ of cycles of co-dimension $i$, as defined in [7]. ### 1.7. Acknowledgements We started working on the project when Anand P. visited Anand D. at the Institute for Computational and Experimental Research in Mathematics (ICERM) at Brown University during the semester program _Braids_ in 2022. We thank ICERM and the organisers of _Braids_. Anand D. thanks the Australian Research Council for the grant DE180101360 that supported a part of this project. ## 2\. Equivariant orbit classes Let $G$ be an algebraic group and $M$ a variety with a $G$ action. A _complete parametrisation_ of the $G$ orbit of $m\in M$ is a $G$-variety $X$ together with a proper $G$-equivariant map $\phi\colon X\to M$ such that 1. (1) there exists an $x\in X$ such that $\phi(x)=m$, 2. (2) the $G$-orbit of $x$ is dense in $X$, and 3. (3) the stabiliser of $x\in X$ is trivial. Let $\Gamma\subset G$ be the stabiliser of $m\in M$ and let $\phi\colon X\to M$ be a complete parametrisation of the $G$ orbit of $m$. Then we have (5) $\phi_{*}[X]=\begin{cases}|\Gamma|\operatorname{Orb}(m)&\text{ if $\Gamma$ is finite}\\\ 0&\text{otherwise}.\end{cases}$ We now specialise to two particular settings. In the first setting, we take $M=W$, where $W$ is a non-zero $G$-representation and the closely related setting of $M={\bf P}W$. Assume that the image of $G\to\operatorname{GL}(W)$ contains the scalar matrices. Suppose the map $G\to\operatorname{PGL}(W)$ factors through $G\to\overline{G}$. (In applications, $G$ will be a general linear group and $\overline{G}$ will be the corresponding projective general linear group.) Let $0\to O(-1)\to\underline{W}\to Q\to 0$ be the universal sequence on ${\bf P}W$. ###### Proposition 2.1. Let $w\in W$ be non-zero. Let $\Gamma\subset\overline{G}$ be the stabiliser of $[w]\in{\bf P}W$ and assume that it is finite. Let $\phi\colon X\to{\bf P}W$ be a complete parametrisation of the $\overline{G}$ orbit of $[w]\in{\bf P}W$. Then, in $A^{*}_{G}(W)=A^{*}_{G}(\bullet)$, we have $|\Gamma|[\operatorname{Orb}(w)]=\int_{X}c_{\rm top}(\phi^{*}Q).$ ###### Proof. Let $\pi\colon W^{*}\to{\bf P}W$ be the projection. Since the image of $G$ contains the scalars, we have $\operatorname{Orb}(w)=\pi^{*}\operatorname{Orb}([w]),$ and hence (6) $|\Gamma|\cdot\operatorname{Orb}(w)=|\Gamma|\cdot\pi^{*}\operatorname{Orb}([w]).$ Since $\phi$ is a complete parametrisation of the $\overline{G}$ orbit of $[w]$, we have $|\Gamma|\cdot\operatorname{Orb}([w])=\phi_{*}[X].$ We now compute the right hand side. On $X\times{\bf P}W$, let $\pi_{i}$ for $i=1,2$ be the two projections. Consider the composite $\pi_{2}^{*}O(-1)\to\underline{W}\to\pi_{1}^{*}\phi^{*}Q.$ Its vanishing locus is precisely the graph of $\phi\colon X\to{\bf P}W$. Therefore, we have (7) $\phi_{*}[X]={\pi_{2}}_{*}c_{\rm top}\operatorname{Hom}(\pi_{2}^{*}O(-1),\pi_{1}^{*}\phi^{*}Q).$ In (6), we substitute $\phi_{*}[X]$ from (7) and use the push-pull formula on the fiber square ${X\times W^{*}}$${W^{*}}$${X\times{\bf P}W}$${{\bf P}W.}$$\scriptstyle{\pi}$ Let $\widetilde{\pi}_{i}$ for $i=1,2$ be the two projections on $X\times W^{*}$. Noting that the pull-back of $O(-1)$ to $W^{*}$ is trivial, in $A^{*}_{G}(W^{*})$ we have (8) $\begin{split}|\Gamma|[\operatorname{Orb}(w)]&={\widetilde{\pi_{2}}}_{*}c_{\rm top}\operatorname{Hom}(O,\widetilde{\pi}_{1}^{*}\phi^{*}Q)\\\ &=\int_{X}c_{\rm top}(\phi^{*}Q).\end{split}$ Since $A^{*}_{G}(W^{*})=A^{*}_{G}(W)$ in co-dimension less than $\dim W$, equation (8) also holds in $A^{*}_{G}(W)$. ∎ In the next setting, we assume that $G=\operatorname{GL}(V)\times H$ and $M=\operatorname{Hom}(V,W)$, where $W$ is an $H$-representation. Let $\dim V=r+1$. A closely related setting is that of $M={\bf Gr}(r+1,W)$ with the induced action of $H$. Suppose the map $H\to\operatorname{PGL}(W)$ factors through $H\to\overline{H}$. (In applications, $H$ will be a general linear group and $\overline{H}$ will be the corresponding projective general linear group.) Let $0\to E\to\underline{W}\to Q\to 0$ be the universal sequence on the Grassmannian. ###### Proposition 2.2. Let $s\in\operatorname{Hom}(V,W)$ be an injective homomorphism with image $S$. Let $\Gamma\subset\overline{H}$ be the stabiliser of $[S]\in{\bf Gr}(r+1,W)$ and assume that it is finite. Let $\phi\colon X\to{\bf Gr}(r+1,W)$ be a complete parametrisation of the $\overline{H}$-orbit of $[S]$. Then, in $A^{*}_{G}(\operatorname{Hom}(V,W))=A^{*}_{G}(\bullet)$, we have $|\Gamma|[\operatorname{Orb}(s)]=\int_{X}c_{\rm top}\operatorname{Hom}(\underline{V},\phi^{*}Q).$ ###### Proof. Consider the $H$-equivariant bundle $\phi^{*}E$ on $X$. Set $F=\operatorname{Hom}(\underline{V},\phi^{*}E).$ Then $F$ admits an action of $G=\operatorname{GL}V\times H$. Set $Y={\bf P}F$; let $\pi\colon Y\to X$ be the projection; and let $\psi=\pi\circ\phi$. The composite (9) $O_{Y}(-1)\xrightarrow{i}\operatorname{Hom}(\underline{V},\psi^{*}E)\xrightarrow{j}\operatorname{Hom}(\underline{V},\underline{W})$ yields a $G$-equivariant map $Y\to{\bf P}\operatorname{Hom}(V,W).$ This map is a complete parametrisation of the $\operatorname{PGL}V\times\overline{H}$-orbit of $[s]$. It is easy to check that the stabiliser of $[s]$ in $\operatorname{PGL}V\times\overline{H}$ maps isomorphically to $\Gamma\subset\overline{H}$ under the second projection. By Proposition 2.1, in $A^{*}_{G}(\operatorname{Hom}(V,W))$ we have $|\Gamma|[\operatorname{Orb}(s)]=\int_{Y}c_{\rm top}(\operatorname{coker}(j\circ i)).$ From the sequence in (9), we have an exact sequence $0\to\operatorname{coker}i\to\operatorname{coker}(j\circ i)\to\operatorname{coker}j\to 0.$ and hence the equality of top Chern classes $c_{\rm top}(\operatorname{coker}j\circ i)=c_{\rm top}(\operatorname{coker}i)\cdot c_{\rm top}(\operatorname{coker}j).$ Observe that $\operatorname{coker}j=\pi^{*}\operatorname{Hom}(\underline{V},\phi^{*}Q).$ Since $Y\to X$ is a projective space bundle and $i$ is its universal sub- bundle, we have $\int_{Y/X}c_{\rm top}(\operatorname{coker}i)=[X].$ Now, the push-pull formula gives $\int_{Y/X}c_{\rm top}(\operatorname{coker}i)\cdot c_{\rm top}(\operatorname{coker}j)=c_{\rm top}(\operatorname{Hom}(\underline{V},\phi^{*}Q)).$ Integrating the two sides along $X$ gives the result. ∎ We now study specialisations of orbits in families. The results are essentially from [13, § 2.5], but we state and prove them more generally. Fix an action of an algebraic group $G$ on a variety $M$. For simplicity, assume that $G$ is irreducible. Given $m\in M$, define the $G$-equivariant cycle $\operatorname{WOrb}(m)=\begin{cases}|\operatorname{Stab}(m)|\cdot\operatorname{Orb}(m)&\text{ if $\operatorname{Stab}(m)$ is finite,}\\\ 0&\text{otherwise}.\end{cases}$ Let $\Delta$ be the spectrum of a DVR with general point $\eta$ and special point $0$. Fix a section $m_{\eta}{\colon}\eta\to M_{\eta}$. Let $\Sigma$ be the cycle on $M$ obtained as the limit of the cycle $\operatorname{WOrb}(m_{\eta})$ on $M_{\eta}$. Our goal is to understand $\Sigma$. We begin with an easy observation. ###### Proposition 2.3. Suppose $m_{\eta}$ has a limit $\overline{m}\in M$. Then the cycle $\Sigma-\operatorname{WOrb}(\overline{m})$ is effective. ###### Proof. Let $X\supset G$ be an equivariant compactification. Consider the map $\alpha\colon G_{\eta}\to M_{\eta}$ defined by $g\mapsto gm_{\eta}$, and let $Y\subset X\times M\times\Delta$ be the closure of the graph of $\alpha$. Then $Y$ is irreducible of dimension $\dim G+1$. Let $\pi\colon Y\to M$ be the projection. Observe that we have ${\pi}_{*}Y_{\eta}=\operatorname{WOrb}(m_{\eta})\text{ and }{\pi}_{*}Y_{0}=\Sigma.$ By construction, $Y$ is invariant under the $G$-action on $X\times M\times\Delta$. Since $y=(1,\overline{m},0)\in Y_{0}$, we have a copy of $G\subset Y_{0}$ given by $G\cdot y$. Let $Z\subset Y_{0}$ be its closure. Note that ${\pi}_{*}Z=\operatorname{WOrb}(\overline{m})$. Furthermore, since $\dim Z=\dim Y_{0}$, we see that $Z\subset Y_{0}$ is an irreducible component. Then $\Sigma-\operatorname{WOrb}(\overline{m})$ is the push-forward of the effective cycle $[Y_{0}]-[Z]$. ∎ We have the following strengthening of Proposition 2.3 (see [13, Proposition 2.13]). ###### Proposition 2.4. Consider sections $g_{1},\dots,g_{n}{\colon}\eta\to G_{\eta}$ such that for all $i\neq j$, the section $g_{i}g_{j}^{-1}$ does not extend to a section $\Delta\to G_{\Delta}$. Suppose $\overline{m}_{1},\dots,\overline{m}_{n}\in M$ are limits of $g_{1}m_{\eta},\dots,g_{n}m_{\eta}$. Then the cycle $\Sigma-\operatorname{WOrb}(\overline{m}_{1})-\dots-\operatorname{WOrb}(\overline{m}_{n})$ is effective. ###### Proof. Let $X\supset G$ be an equivariant compactification. Consider the map $\alpha\colon G_{\eta}\to G_{\eta}^{n}\times M_{\eta}$ given by $\alpha\colon g\mapsto(gg_{1}^{-1},\dots,gg_{n}^{-1},gm_{\eta})$ and let $Y\subset X^{n}\times M\times\Delta$ be the closure of its image. Denoting by $\pi$ the projection onto the $M$ factor, we have ${\pi}_{*}Y_{\eta}=\operatorname{WOrb}(m_{\eta})\text{ and }{\pi}_{*}Y_{0}=\Sigma.$ Let $\overline{g}_{i,j}\in X$ be the limit of $g_{i}g_{j}^{-1}$. Then $\overline{g}_{i,j}\not\in G$ for $i\neq j$. Then the limit of $\alpha(g_{i})$ is the point $y_{i}=(\overline{g}_{i,1},\dots,\overline{g}_{i,i-1},1,\overline{g}_{i,i+1},\overline{g}_{i,n},\overline{m}_{i})\in Y_{0}.$ Each $y_{i}$ gives a copy of $G\subset Y_{0}$, say $G_{i}$, defined by $G_{i}=G\cdot y_{i}$. Since $\overline{g}_{i,j}\not\in G$ for $i\neq j$, the $j$-th components of the points of $G_{i}$ do not lie in $G$ for $i\neq j$ and lie in $G$ for $i=j$. In particular, for $i\neq j$, the $G_{i}$ and $G_{j}$ are disjoint. Let $Z_{i}$ be the closure of $G_{i}$. Then the $Z_{i}$ are distinct irreducible components of $Y_{0}$. The cycle $\Sigma-\operatorname{WOrb}(\overline{m}_{1})-\dots-\operatorname{WOrb}(\overline{m}_{n})$ is the $\pi$-pushforward of the effective cycle $Y_{0}-Z_{1}-\dots-Z_{n}$. ∎ ## 3\. Vanishing sequences and the Wronskian Let $C$ be a smooth curve, $L$ a line bundle of degree $d$ on $C$, and $S\subset H^{0}(C,L)$ a subspace of dimension $r+1$. The pair $(S,L)$ is called a _linear series of rank $r$ and degree $d$_ or a $g^{r}_{d}$ on $C$. Given a point $p\in C$, there exists a unique sequence of integers $a_{0}(p)<\dots<a_{r}(p)$ in $\\{0,\dots,d\\}$ such that $S$ contains a section $s_{i}$ that vanishes to order exactly $a_{i}(p)$ at $p$. The sequence $a_{i}(p)$ is called the _vanishing sequence_ of $S$ at $p$. The sections $s_{0},\dots,s_{r}$ form a basis of $S$. Denote by $j_{r+1}L$ the bundle of jets of order $r+1$ of $L$; that is $j_{r+1}L={\pi_{2}}_{*}\left(\pi_{1}^{*}L\otimes O_{(r+1)\Delta}\right).$ Here $\Delta\to C\times C$ is the diagonal and the $\pi_{i}$ are the two projections. The linear series $S\subset H^{0}(C,L)$ induces an evaluation map $\underline{S}\to L,$ and also a map (10) $\underline{S}\to j_{r+1}L.$ The _Wronskian_ of $S$ is the determinant of (10). ###### Proposition 3.1. The Wronskian is non-zero and vanishes to order $\sum a_{i}(p)-i$ at $p$. ###### Proof. The proof is a part of [9, Proposition 1.1]. We give it for completeness. Let $s_{i}\in S$ vanish to order $a_{i}=a_{i}(p)$ at $p$. Denote by $\widehat{M}$ the completion of an $O_{C}$-module $M$ at $p$. Choose an isomorphism of $\mathbf{k}\llbracket t\rrbracket$-modules $\widehat{L}\cong\mathbf{k}\llbracket t\rrbracket$. Then we have an isomorphism $\widehat{j_{r+1}L}\cong\mathbf{k}\llbracket t,u\rrbracket/(u-t)^{r+1}.$ If the section $s\in\widehat{L}\cong\mathbf{k}\llbracket t\rrbracket$ is given by the power series $s(t)$, then the induced section $j_{r+1}(s)$ of $\widehat{j_{r+1}L}$ is given by $j_{r+1}(s)=s(t)+s^{\prime}(t)(u-t)+\dots+s^{(r)}(t)\frac{1}{r!}(u-t)^{r}.$ Let $s_{i}$ correspond to the power series $s_{i}(t)=t^{a_{i}}z_{i}\in\mathbf{k}\llbracket t\rrbracket$, where $z_{i}\in\mathbf{k}\llbracket t\rrbracket$ is a unit. With respect to the bases $s_{0},\dots,s_{r+1}$ of $S$ and $1,(u-t),\dots,(u-t)^{r}/r!$ of $\widehat{j_{r+1}L}$, the matrix of the map (10) is given by $\begin{pmatrix}t^{a_{0}}z_{0}&t^{a_{1}}z_{1}&\dots&t^{a_{r}}z_{r}\\\ (t^{a_{0}}z_{0})^{\prime}&(t^{a_{1}}z_{1})^{\prime}&\dots&(t^{a_{r}}z_{r})^{\prime}\\\ \vdots&\vdots&\ddots&\vdots\\\ (t^{a_{0}}z_{0})^{(r)}&(t^{a_{1}}z_{1})^{(r)}&\dots&(t^{a_{r}}z_{r})^{(r)}.\end{pmatrix}$ The lowest degree term in its determinant is $t^{\sum a_{i}-i}$ and its coefficient is the determinant of the $(r+1)\times(r+1)$ matrix $C$ with $C_{ij}=a_{i}(a_{i}-1)\cdots(a_{i}-j+1)$. It remains to show that the coefficient is non-zero. By a sequence of invertible row transformations, the matrix $C$ reduces to the Vandermonde matrix $V$ with $V_{ij}=a_{i}^{j}$. Since the $a_{i}$’s are distinct integers in $\\{0,\dots,d\\}$, their images are distinct in $\mathbf{k}$, which has characteristic 0 or characteristic $>d$. Hence the determinant of $V$, and hence of $C$, is non-zero. ∎ Let $\rho(S)\subset C$ be the vanishing locus of the Wronskian of $S$. We call $\rho(S)$ the _ramification divisor_ of $S$ and its points the _ramification points_. ###### Corollary 3.2. For all $p\in C-\rho(S)$, we have $a_{i}(p)=i$. In particular, for all but finitely many $p$, we have $a_{i}(p)=i$. ###### Proof. Follows immediately from Proposition 3.1. ∎ Note that $\rho(S)$ is the zero-locus of a section of $\det j_{r+1}L$. Recall that we have a sequence of surjective maps $j_{r+1}L\to j_{r}L\to\dots\to j_{2}L\to j_{1}L=L\to 0$ with $\ker(j_{i+1}L\to j_{i}L)=L\otimes\Omega_{C}^{i}.$ As a result, we get (11) $\det j_{r+1}L=L^{r+1}\otimes\Omega_{C}^{r(r+1)/2}.$ Consider the Grassmannian ${\bf Gr}={\bf Gr}(r+1,H^{0}(C,L))$ with the universal sub-bundle $S\subset\underline{H^{0}(C,L)}.$ On ${\bf Gr}\times C$, we have the universal evaluation map $\pi_{1}^{*}S\to\pi_{2}^{*}L,$ which induces a map $\pi_{1}^{*}S\to\pi_{2}^{*}j_{r+1}L.$ Taking the determinant of the last map and applying ${\pi_{1}}_{*}$ gives the following map on ${\bf Gr}$: $\det S\to\underline{H^{0}(C,\det j_{r+1}L)}.$ By Proposition 3.1, the map above is non-zero at every point of ${\bf Gr}$, and hence yields a morphism (12) $\rho\colon{\bf Gr}(r+1,H^{0}(C,L))\to{\bf P}H^{0}(C,\det j_{r+1}L).$ ###### Proposition 3.3. The map $\rho$ in (12) is finite. For $C={\bf P}^{1}$, it is also surjective. ###### Proof. Assume that $\dim H^{0}(C,L)>r+1$; otherwise, the statement is vacuous. Since the Grassmannian has Picard rank one, any morphism from it is either finite or constant. It is easy to see that $\rho$ is not constant; so it must be finite. For $C={\bf P}^{1}$ and $L=O(d)$, we have by (11) $\det j_{r+1}L\cong O((r+1)d)\otimes O(-r(r+1))=O((r+1)(d-r)).$ Hence, both the source and target of $\rho$ have the same dimension, namely $(r+1)(d-r)$. ∎ ## 4\. Complete parametrisation of the orbit of a linear series Fix a $2$-dimensional vector space $V$ and an $(r+1)$-dimensional subspace $S\subset\operatorname{Sym}^{d}V$. On ${\bf P}V\cong{\bf P}^{1}$, let $O(-1)\subset\underline{V}$ be the universal sub-bundle and $Q$ the quotient $\underline{V}/O(-1)$. Then we have a canonical identification $\operatorname{Sym}^{d}V=H^{0}({\bf P}V,Q^{d})$. We regard $S$ as a linear series associated to the line bundle $Q^{d}$. Let $W$ be another $2$-dimensional vector space. Consider the ${\bf P}^{3}={\bf P}\operatorname{Hom}(V,W).$ We have a rational map $\phi\colon\mathbf{P}^{3}\dashrightarrow{\bf Gr}(r+1,\operatorname{Sym}^{d}W)$ defined by (13) $m\mapsto[m(S)].$ Note that the map is equivariant with respect to the obvious $\operatorname{GL}W$ action. The goal of this section is to understand a $\operatorname{GL}W$-equivariant resolution of $\phi$. Let $B\subset\mathbf{P}V$ be the ramification divisor of $S$. For every $b\in B$, let $L_{b}\subset\mathbf{P}^{3}$ be the linear subspace (14) $L_{b}=\\{[m]\mid m(b)=0\\}.$ Then $L_{b}$ is a copy of ${\bf P}^{1}$ linearly embedded in ${\bf P}^{3}$, and for $b_{1}\neq b_{2}$, the lines $L_{b_{1}}$ and $L_{b_{2}}$ are disjoint. Let $\beta\colon\widetilde{{\bf P}}^{3}\to{\bf P}^{3}$ be the blow-up at $\bigsqcup_{b\in B}L_{b}$. Note that the lines $L_{b}$ are $\operatorname{GL}W$-invariant and hence the blow-up $\widetilde{{\bf P}}^{3}$ inherits an action by $\operatorname{GL}W$. See Figure 2 for a sketch of ${\bf P}^{3}$ and the lines $L_{b}$. Figure 2. The map $\phi\colon{\bf P}^{3}\dashrightarrow{\bf Gr}(r+1,\operatorname{Sym}^{d}W)$ becomes regular after blowing up the lines $L_{b}$ (red) on the discriminant quadric $\Delta$. Along the lines of the opposite ruling (blue, dashed), the map is constant. The key observation is the following consequence of a result of Aluffi and Faber [1]. ###### Theorem 4.1. The rational map $\phi\colon{\bf P}^{3}\dashrightarrow{\bf Gr}(r+1,\operatorname{Sym}^{d}W)$ extends to a regular map $\widetilde{\phi}\colon\widetilde{{\bf P}}^{3}\to{\bf Gr}(r+1,\operatorname{Sym}^{d}W).$ For the proof, we need a simple lemma. ###### Lemma 4.2. Let $\rho\colon X\to Y$ be a finite map between varieties. Let $P$ be an irreducible normal variety with a rational map $P\dashrightarrow X$. If the composite $P\dashrightarrow Y$ extends to a regular map $P\to Y$, then so does $P\dashrightarrow X$. ###### Proof. Let $\Gamma_{X}\subset P\times X$ and $\Gamma_{Y}\subset P\times Y$ be the closures of the graphs of $P\dashrightarrow X$ and $P\dashrightarrow Y$, respectively. Then the projections $\Gamma_{X}\to P$ and $\Gamma_{Y}\to P$ are birational. In fact, by the hypothesis on $P\dashrightarrow Y$, the projection $\Gamma_{Y}\to P$ is an isomorphism. The map $\Gamma_{X}\to\Gamma_{Y}$ induced by $\rho$ is finite and birational. Since the target $\Gamma_{Y}\cong P$ is normal, we conclude that $\Gamma_{X}\to\Gamma_{Y}$ is an isomorphism. Hence the projection $\Gamma_{X}\to P$ is also an isomorphism. We deduce that $P\dashrightarrow X$ extends to a regular map. ∎ ###### Proof of Theorem 4.1. Set ${\bf Gr}={\bf Gr}(r+1,\operatorname{Sym}^{d}W)$ and ${\bf P}={\bf P}H^{0}({\bf P}^{1},\det j_{r+1}O(d))={\bf P}\operatorname{Sym}^{(r+1)(d-r)}W.$ Recall that we have the finite map $\rho\colon{\bf Gr}\to{\bf P}$ defined by the Wronskian. By definition, $\rho$ is $\operatorname{GL}W$-equivariant. Let $R$ be the Wronksian of $S$. Consider the rational map ${\bf P}^{3}\dashrightarrow{\bf P}$ defined by $m\mapsto[m(R)].$ Then we have the commutative diagram ${\widetilde{\bf P}^{3}}$${{\bf Gr}}$${\widetilde{\bf P}^{3}}$${{\bf P}.}$$\scriptstyle{\rho}$ By [1, Proposition 1.2], the map $\widetilde{\bf P}^{3}\dashrightarrow{\bf P}$ extends to a regular map $\widetilde{\bf P}^{3}\to{\bf P}$. By Lemma 4.2, we conclude the same about $\widetilde{\bf P}^{3}\dashrightarrow{\bf Gr}.$ (In [1], the base field is assumed to have characteristic 0, but Proposition 1.2 and its proof hold in any characteristic.) ∎ Let $\Delta\subset{\bf P}^{3}={\bf P}\operatorname{Hom}(V,W)$ be the locus of degenerate maps. Then $\Delta$ is a quadric hypersurface, defined by the vanishing of the determinant. We have an isomorphism $\Delta\xrightarrow{\sim}{\bf P}V\times{\bf P}W$ given by (15) $m\mapsto(\ker m,\operatorname{im}m).$ Under this isomorphism, the line $L_{b}$ defined in (14) corresponds to the line $\\{b\\}\times{\bf P}W$. We now do a simple calculation that allows us to understand the map $\widetilde{\phi}\colon\widetilde{\bf P}\to{\bf Gr}(r+1,\operatorname{Sym}^{d}W).$ Choose bases $\langle v_{1},v_{2}\rangle$ for $V$ and $\langle w_{1},w_{2}\rangle$ for $W$. Let us write elements of $\operatorname{Hom}(V,W)$ as $2\times 2$-matrices with respect to these bases. Consider the one-parameter family of homomorphisms given by (16) $m_{t}=\begin{pmatrix}1&0\\\ 0&t\end{pmatrix},\quad t\in{\bf A}^{1}.$ For $t\neq 0$, let $S_{t}=m_{t}(S)\subset\operatorname{Sym}^{d}W,$ and $S_{0}=\lim_{t\to 0}S_{t}\subset\operatorname{Sym}^{d}W.$ The following proposition describes $S_{0}$. ###### Proposition 4.3. Let $S_{0}$ be defined as above. Let $b\in{\bf P}V$ be the point corresponding to $[v_{2}]\subset V$ and let $a_{0}<\dots<a_{r}$ be the vanishing sequence of $S$ at $b$. Then $S_{0}\subset\operatorname{Sym}^{d}W$ is given by $S_{0}=\langle w_{1}^{d-a_{0}}w_{2}^{a_{0}},\dots,w_{1}^{d-a_{r}}w_{2}^{a_{r}}\rangle.$ ###### Proof. Let $s_{i}\in S$ vanish to order $a_{i}$ at $b$. Then there exists a non-zero $c_{i}\in\mathbf{k}$ and a homogeneous polynomial $F_{i}(v_{1},v_{2})$ of degree $d-a_{i}-1$ such that $s_{i}=c_{i}v_{1}^{d-a_{i}}v_{2}^{a_{i}}+v_{2}^{a_{i}+1}F_{i}(v_{1},v_{2}).$ We see that $m_{t}(s_{i})=c_{i}t^{a_{i}}w_{1}^{d-a_{i}}w_{2}^{a_{i}}+t^{a_{i}+1}w_{2}^{a_{i}+1}F_{i}(w_{1},tw_{2}).$ Therefore, we get $\lim_{t\to 0}\langle m_{t}(s_{i})\rangle=\langle w_{1}^{d-a_{i}}w_{2}^{a_{i}}\rangle,$ and hence $\lim_{t\to 0}m_{t}(S)=\langle w_{1}^{d-a_{0}}w_{2}^{a_{0}},\dots,w_{1}^{d-a_{r}}w_{2}^{a_{r}}\rangle.$ ∎ ###### Corollary 4.4. If $b\in{\bf P}V$ is not a ramification point of $S$, then $S_{0}$ depends only on the subspace $\operatorname{im}m_{0}\subset W$. ###### Proof. In this case, the vanishing sequence at $b$ is $0<\dots<r$. From Proposition 4.3, we see that $S_{0}\subset\operatorname{Sym}^{d}W$ is the subspace of sections that vanish to order at least $d-r$ at $\operatorname{im}m_{0}=[w_{1}]\in{\bf P}W$. ∎ Corollary 4.4 implies that the map $\widetilde{\phi}$ is constant on the proper transforms of the lines ${\bf P}V\times\\{w\\}\subset\Delta$ (see Figure 2). ###### Theorem 4.5. Let $S^{\prime}\subset\operatorname{Sym}^{d}W$ be the image of $S\subset\operatorname{Sym}^{d}V$ under an isomorphism $V\to W$. The $\operatorname{GL}W$-orbit of $[S^{\prime}]\in{\bf Gr}(r+1,\operatorname{Sym}^{d}W)$ contains finitely many orbits in its closure. In addition to the orbit of $[S^{\prime}]$, these are the orbits of $\langle w_{1}^{d-a_{0}}w_{2}^{a_{0}},\dots,w_{1}^{d-a_{r}}w_{2}^{a_{r}}\rangle$ where $a_{0}<\dots<a_{r}$ is the vanishing sequence of $S$ at some $p\in{\bf P}V$. ###### Proof. The closure of the $\operatorname{GL}W$-orbit of $[S^{\prime}]$ is precisely the image of the morphism $\widetilde{\phi}\colon\widetilde{\bf P}\to{\bf Gr}(r+1,\operatorname{Sym}^{d}W).$ Let $\widetilde{\Delta}\subset\widetilde{\bf P}$ be the pre-image of $\Delta\subset{\bf P}^{3}$. Then $\widetilde{\bf P}-\widetilde{\Delta}$ is a copy of $\operatorname{PGL}_{3}$, and its image is the orbit of $[S^{\prime}]$. Consider $\widetilde{m}\in\widetilde{\Delta}$ and let $m\in\Delta$ be its image. Suppose $m$ does not lie on any line $L_{b}$ and let $\operatorname{im}m=[w_{1}]\in{\bf P}W$. Then by Corollary 4.4, the image $\widetilde{\phi}(\widetilde{m})\in{\bf Gr}(r+1,\operatorname{Sym}^{d}W)$ corresponds to (17) $\langle w_{1}^{d}w_{2}^{0},\dots,w_{1}^{d-r}w_{2}^{r}\rangle,$ where $w_{2}\in W$ is any vector linearly independent from $w_{1}$. Suppose $m$ lies on $L_{b}$. Let $w=\operatorname{im}m\in{\bf P}W$, so that $m\in\Delta$ corresponds to $(b,w)$ in the isomorphism (15). Suppose $v_{2}\in V$ is such that $b=[v_{2}]$ and $w_{1}\in W$ such that $w=[w_{1}]$. The point $\widetilde{m}$ corresponds to a direction in the normal bundle $N_{L_{b}/{\bf P}^{3}}$ at $p$. Consider the direction in $N_{L_{b}/{\bf P}^{3}}|_{p}$ corresponding to the line ${\bf P}V\times\\{w\\}\subset\Delta$ through $m$. By Corollary 4.4, the map $\widetilde{\phi}$ is constant on this line with value given by the subspace (17). If $\widetilde{m}$ corresponds to this direction, then $\widetilde{\phi}(\widetilde{m})$ corresponds to the same subspace. All the other directions in $N_{L_{b}/{\bf P}^{3}}|_{p}$ can be realised by tangent directions of families $m_{t}$ for suitable choices of $v_{1}\in V$ and $w_{2}\in W$. If $\widetilde{m}$ corresponds to the tangent direction given by $m_{t}$, then Proposition 4.3 implies that $\widetilde{\phi}(\widetilde{m})$ corresponds to the subspace (18) $\langle w_{1}^{d-a_{0}}w_{2}^{a_{0}},\dots,w_{1}^{d-a_{r}}w_{2}^{a_{r}}\rangle.$ In conclusion, we see that the image of every point of $\widetilde{\Delta}$ is of the form (17), which corresponds to the generic vanishing sequence, or (18), which corresponds to the vanishing sequence of a ramification point. ∎ ## 5\. Localisation The goal of this section is to compute the equivariant orbit class of a linear series using localisation on $\widetilde{{\bf P}}^{3}$. The calculation is similar to the one in [13, Appendix B]. Choose a basis $\langle w_{1},w_{2}\rangle$ of $W$ and let $T\subset\operatorname{GL}(W)$ be the diagonal torus with respect to this basis. Choose coordinates $t_{1},t_{2}$ on $T$ so that the action on $W$ is $(t_{1},t_{2})\colon(w_{1},w_{2})\mapsto(t_{1}w_{1},t_{2}w_{2}).$ Then the $T$-fixed locus of ${\bf P}^{3}={\bf P}\operatorname{Hom}(V,W)$ consists of two skew lines $\Lambda_{1}$ and $\Lambda_{2}$ given by $\Lambda_{i}={\bf P}\operatorname{Hom}(V,\langle w_{i}\rangle).$ Note that the lines $\Lambda_{i}\subset{\bf P}^{3}$ lie in the discriminant quadric $\Delta$, and are in the opposite ruling as the lines $L_{b}$ blown up to get $\widetilde{{\bf P}}^{3}$. The $T$-fixed locus of $\widetilde{{\bf P}}^{3}$ consists of the proper transforms $\widetilde{\Lambda}_{i}$ of $\Lambda_{i}$ and points $p_{i,b}$ for $i=1,2$ and $b\in B$. The point $p_{i,b}$ maps under the blow-down to the point $\Lambda_{i}\cap L_{b}$. We now describe it in local coordinates. Choose a basis $v_{1},v_{2}$ of $V$ such that $[v_{2}]=b\in{\bf P}V$. In the bases $v_{1},v_{2}$ of $V$ and $w_{1},w_{2}$ of $W$, we represent $m\in\operatorname{Hom}(V,W)$ by the matrix $m=\begin{pmatrix}X_{11}&X_{12}\\\ X_{21}&X_{22}\end{pmatrix}.$ The matrix entries $X_{ij}$ are functionals on $\operatorname{Hom}(V,W)$. The torus $T$-acts by $(t_{1},t_{2})\colon X_{ij}\mapsto t_{i}^{-1}X_{ij}.$ The line $\Lambda_{1}\subset{\bf P}^{3}$ is defined by $X_{21}=X_{22}=0$ and the line $L_{b}\subset{\bf P}^{3}$ by $X_{12}=X_{22}=0.$ Their intersection is the point $X_{12}=X_{21}=X_{22}=0$. Choose affine coordinates $x_{ij}=X_{ij}/X_{11}$ around this point. Then the blow-up of ${\bf P}^{3}$ in $L_{b}$ has the coordinate charts $(x_{21},x_{12},x_{22},u\mid x_{12}=ux_{22})\text{ and }(x_{21},x_{12},x_{22},v\mid x_{22}=vx_{12}).$ The point $p_{1,b}$ is contained in the second chart, and is given by (19) $p_{1,b}:x_{21}=x_{22}=v=0.$ Given $(m_{1},m_{2})\in\mathbf{Z}$, let $\chi(m_{1},m_{2})$ be the character of $T$ given by $\chi(m_{1},m_{2})\colon(t_{1},t_{2})\mapsto t_{1}^{m_{1}}\cdot t_{2}^{m_{2}}.$ The next two propositions give the normal bundles of the $T$-fixed loci. ###### Proposition 5.1. Let $N_{1,b}$ be the tangent space to $p_{1,b}\in\widetilde{{\bf P}}^{3}$. As a $T$-representation, we have $N_{1,b}\cong\chi(-1,1)\oplus\chi(-1,1)\oplus\chi(1,-1).$ ###### Proof. From (19), we see that the co-tangent space at $p_{1,b}$ is generated by $x_{21},x_{22},v$, on which $T$ acts by $(t_{1},t_{2})\colon x_{21}\mapsto t_{1}t_{2}^{-1}x_{21}\quad x_{22}\mapsto t_{1}t_{2}^{-1}x_{22}\quad v\mapsto t_{1}^{-1}t_{2}v.$ We now take the dual. ∎ ###### Proposition 5.2. Let $\widetilde{N}_{1}$ be the normal bundle of $\widetilde{\Lambda}_{1}\subset\widetilde{{\bf P}}^{3}$. In the $T$-equivariant Grothendieck group of $\widetilde{\Lambda}_{1}={\bf P}^{1}$ (with the trivial $T$-action), we have $[N_{1}]=\left([O]+[O\left(2-|B|\right)]\right)\otimes\chi(-1,1).$ ###### Proof. Let $N_{1}$ be the normal bundle of $\Lambda_{1}\subset{\bf P}^{3}$. The defining equations show that $\Lambda_{1}$ is the zero locus of sections of $O_{{\bf P}^{3}}(1)\otimes\chi(0,1)$ and the restriction of $O_{{\bf P}^{3}}(1)$ to $\Lambda_{1}$ is $O(1)\otimes\chi(-1,0)$. Therefore, we get $N_{1}\cong\left(O(1)\oplus O(1)\right)\otimes\chi(-1,1).$ Recall that $\beta\colon\widetilde{\bf P}^{3}\to{\bf P}^{3}$ is the blow-up along $\bigsqcup_{b}L_{b}$. We have the exact sequence $0\to\widetilde{N}_{1}\to\beta^{*}N_{1}\to Q\to 0,$ where $Q\cong O_{\operatorname{supp}B}\otimes\chi(-1,1)$. The statement follows. ∎ Let $\widetilde{S}$ be the pull-back to $\widetilde{\bf P}^{3}$ of the universal sub-bundle of ${\bf Gr}(r+1,\operatorname{Sym}^{d}W)$. The following two propositions describe $\widetilde{S}$ on the $T$-fixed locus. ###### Proposition 5.3. Let $p=p_{1,b}$ and let $a_{0}<\dots<a_{r}$ be the vanishing sequence of $S$ at $b$. As a $T$-representation, we have $\widetilde{S}|_{p}\cong\bigoplus_{i=0}^{r}\chi(d-a_{i},a_{i}).$ ###### Proof. Consider the family $m_{t}$ as in (16), where $[v_{2}]=b$. Then, in $\widetilde{\bf P}$ we have $\lim_{t\to 0}m_{t}=p.$ Therefore, we have $\widetilde{S}|_{p}=\lim_{t\to 0}m_{t}(S).$ Proposition 4.3 gives the limit and yields $\widetilde{S}|_{p}=\langle w_{1}^{d-a_{0}}w_{2}^{a_{0}},\dots,w_{1}^{d-a_{r}}w_{2}^{a_{r}}\rangle.$ The statement follows. ∎ ###### Proposition 5.4. As a $T$-equivariant sheaf on $\widetilde{\Lambda}_{1}$, we have $[\widetilde{S}]|_{\widetilde{\Lambda}_{1}}=\bigoplus_{i=0}^{r}\chi(d-i,i).$ ###### Proof. By Corollary 4.4, the map $\widetilde{\phi}$ is constant on $\widetilde{\Lambda}_{1}$, and its value corresponds to the subspace $\langle w_{1}^{d}w_{2}^{0},\dots,w_{1}^{d-r}w_{2}^{r}\rangle\subset\operatorname{Sym}^{d}W.$ The statement follows. ∎ We are now ready to prove the main theorem of the paper, which we restate. ###### Theorem 5.5. Let $s\in\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ be an injective map with image $S\subset\operatorname{Sym}^{d}\mathbf{k}^{2}$. Let $\\{(a_{i}(b))\mid b\in B\\}$ be the ramification profile of $S$. Let $\Gamma\subset\operatorname{PGL}(2)$ be the stabiliser of $S\subset\operatorname{Sym}^{d}\mathbf{k}^{2}$, and assume that it is finite. Let $\operatorname{Orb}(s)$ be the closure of the $\operatorname{GL}(r+1)\times\operatorname{GL}(2)$ orbit of $s$. Then the class $|\Gamma|\cdot[\operatorname{Orb}(s)]\in A^{(r+1)(d-r)-3}_{\operatorname{GL}(r+1)\times\operatorname{GL}(2)}(\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2}))$ is given by $|\Gamma|\cdot[\operatorname{Orb}(s)]=\frac{1}{(\omega_{1}-\omega_{2})^{3}}\sum_{b\in B}\psi_{a_{0}(b),\dots,a_{r}(b)}(\mu;\omega)+\frac{2-|B|}{(\omega_{1}-\omega_{2})^{3}}\cdot\psi_{0,\dots,r}(\mu;\omega).$ ###### Proof. Set $V=W=\mathbf{k}^{2}$ and consider the rational map ${\bf P}^{3}={\bf P}\operatorname{Hom}(V,W)\dashrightarrow{\bf Gr}(r+1,\operatorname{Sym}^{d}W)$ given by $m\mapsto m(S).$ Let $\phi\colon\widetilde{{\bf P}}^{3}\to{\bf Gr}(r+1,\operatorname{Sym}^{d}W)$ be its resolution provided by Theorem 4.1. Then $\phi$ is a complete parametrisation of the $\operatorname{PGL}W$-orbit of $[S]$. By Proposition 2.2, we have $|\Gamma|\cdot[\operatorname{Orb}(s)]=\int_{\widetilde{{\bf P}}^{3}}c_{(r+1)(d-r)}\operatorname{Hom}(\underline{\mathbf{k}}^{r+1},\phi^{*}Q),$ where $Q$ is the universal quotient bundle on the Grassmannian. In our case, we have $\phi^{*}Q=\operatorname{Sym}^{d}\underline{W}/\widetilde{S}.$ Let $T\subset\operatorname{GL}(2)$ be a maximal torus. By the equivariant localisation formula [8, Corollary 1], we have (20) $\int_{\widetilde{{\bf P}}^{3}}c_{(r+1)(d-r)}\operatorname{Hom}(\underline{\mathbf{k}}^{r+1},\phi^{*}Q)=\sum_{F}\int_{F}\frac{c_{(r+1)(d-r)}\operatorname{Hom}(\underline{\mathbf{k}}^{r+1},\phi^{*}Q)}{e(N_{F})},$ where the sum is over the $T$-fixed loci $F$ and $e(N_{F})$ is the top Chern class of the equivariant normal bundle of $F\subset\widetilde{{\bf P}}^{3}$. In our case, the fixed loci consist of the lines $\widetilde{\Lambda}_{i}$ for $i=1,2$ and the points $p_{i,b}$ for $i=1,2$ and $b\in B$. Let us compute the contribution from $F=\widetilde{\Lambda}_{1}$. Proposition 5.2 give us the numerator of the integrand: $\prod_{j=0}^{r}\prod_{i=r+1}^{d}((d-i)\omega_{1}+i\omega_{2}-\mu_{j})=\phi_{0,\dots,r}(\mu;\omega).$ Proposition 5.4 gives us the denominator of the integrand: $(\omega_{2}-\omega_{1})(\omega_{2}-\omega_{1}+(2-|B|)h),$ where $h=c_{1}(O(1))$. The integral is the coefficient of $h$ in the power series expansion of the quotient. We may compute it by formally differentiating with respect to $h$ and setting $h=0$. The result is (21) $\int_{\widetilde{\Lambda}_{1}}\frac{c_{(r+1)(d-r)}\operatorname{Hom}(\mathbf{k}^{r+1},\phi^{*}Q)}{e(N_{\widetilde{\Lambda}_{1}})}=\frac{2-|B|}{(\omega_{1}-\omega_{2})^{3}}\cdot\phi_{0,\dots,r}(\mu;\omega).$ The contribution from $F=\widetilde{\Lambda}_{2}$ is the same as above with $\omega_{1}$ and $\omega_{2}$ swapped. Let us compute the contribution from $F=p_{1,b}$. Proposition 5.3 gives us the numerator: $\prod_{j=0}^{r}\prod_{i}((d-i)\omega_{1}+i\omega_{2}-\mu_{j})=\phi_{a_{0}(b),\dots,a_{r}(b)}(\mu;\omega).$ On the left hand side above, the index $i$ ranges in the set $\\{0,\dots,d\\}-\\{a_{0}(b),\dots,a_{r}(b)\\}$. Proposition 5.1 gives us the denominator: $(\omega_{1}-\omega_{2})^{3}.$ Taking the integral is trivial in this case, so the result is (22) $\int_{p_{1,b}}\frac{c_{(r+1)(d-r)}\operatorname{Hom}(\mathbf{k}^{r+1},\phi^{*}Q)}{e(N_{p_{1,b}})}=\frac{1}{(\omega_{1}-\omega_{2})^{3}}\cdot\phi_{a_{0}(b),\dots,a_{r}(b)}(\mu;\omega).$ The contribution from $F=p_{2,b}$ is the same as above with $\omega_{1}$ and $\omega_{2}$ swapped. By combining (21), (22), and their analogues for $\widetilde{\Lambda}_{2}$ and $p_{2,b}$ we see that the integral in (20) is as claimed. ∎ ## 6\. Orbit specialisation The goal of this section is to understand limits of the orbit closure of a linear series under one-parameter specialisations. Recall that $\operatorname{Orb}$ denotes the orbit closure and $\operatorname{WOrb}$ the weighted orbit closure, namely the cycle obtained by multiplying the orbit closure by the order of the stabiliser group in $\operatorname{PGL}(2)$. Fix a positive integer $m$ and for each $j=1,\dots,m$, fix an increasing sequence of non-negative integers $a_{0}(j)<\dots<a_{r}(j)$. Assume that (23) $\sum_{j,i}(a_{i}(j)-i)=(r+1)(d-r).$ We consider one-parameter degenerations of linear series on ${\bf P}^{1}$ with ramification profile $\\{(a_{0}(j),\dots,a_{r}(j))\mid j=1,\dots,m\\}$. We use limit linear series, for which we refer the reader to [10]. Although [10] assumes characteristic 0, the theory holds without modification for characteristic $>d$; see [14]. Let $\Delta$ be the spectrum of a DVR with special point $0$ and generic point $\eta$. Let $p_{1,\eta},\dots,p_{m,\eta}$ be distinct sections $\eta\to{\bf P}^{1}_{\eta}$ and let $(S_{\eta},O(d))$ be a linear series that has vanishing sequence $a_{0}(j)<\dots<a_{r}(j)$ at $p_{j,\eta}$. Let $(\mathcal{X}\to\Delta,p_{1},\dots,p_{m})$ be a semi-stable extension of $({\bf P}^{1}_{\eta}\to\eta,p_{1,\eta},\dots,p_{m,\eta})$ such that $\mathcal{X}$ is non-singular. Then the central fiber $(X,\overline{p}_{1},\dots,\overline{p}_{m})$ is a semi-stable pointed rational curve. The linear series $S_{\eta}$ yields a refined limit linear series $\mathcal{S}$ on $\mathcal{X}$ (use the implication $3\implies 1$ of [10, Proposition 2.5]). Let $S$ be the reduction of $\mathcal{S}$ to the central fiber. We record the ramification profile of the limit linear series $S$ using a decorated graph. We begin with the dual graph $\Gamma$ of the pointed curve $(X,\overline{p}_{1},\dots,\overline{p}_{m})$. Recall that $\Gamma$ is obtained by drawing a vertex $\nu$ for every irreducible component $X_{\nu}$ of $X$, an edge $\mu$ to $\nu$ for every node of $X$ on $X_{\mu}$ and $X_{\nu}$, and a half-edge $i$ on $\nu$ if the marked point $\overline{p}_{i}$ lies on $X_{\nu}$. We think of the full-edge from $\mu$ to $\nu$ as a combination of two half-edges, one at $\mu$ and one at $\nu$. We use the term _dangling half-edge_ when we refer to a true half-edge, and not a part of a full-edge. On $\Gamma$ we record the ramification information of $S$ as follows. We label a half-edge $e$ incident to a vertex $\nu$ by the vanishing sequence of the $X_{\nu}$-aspect of $S$ at the point of $X_{\nu}$ represented by $e$. If $e$ is a full edge, joining $\mu$ and $\nu$ say, then it gets two labels, one at $\mu$, say $(a_{0},\dots,a_{r})$, and one at $\nu$, say $(b_{0},\dots,b_{r})$. Since $S$ is a refined series, the two labels are complementary; that is, they satisfy $a_{i}+b_{r-i}=d$. Figure 3 shows an example of a labelled dual graph $\Gamma$ arising from a limit linear series of rank $2$ and degree $4$. $(0,2)$$(0,2)$$(0,2)$$(0,2)$$(0,2)$$(0,2)$$(0,3)$$(1,4)$$(0,3)$$(1,4)$$(1,2)$$(2,3)$ Figure 3. We represent the ramification information of a refined limit $g^{r}_{d}$ by a labelled dual graph as above, where $r=1$ and $d=4$. The labels on the half-edges are vanishing sequences, and the two labels on a full-edge are complementary. For a vertex $\nu$ of $\Gamma$, let $S(\nu)$ be the $X_{\nu}$-aspect of the limit linear series $S$. ###### Theorem 6.1. In the notation above, let $\Sigma$ be the cycle on ${\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$ obtained as the limit of the cycle $\operatorname{WOrb}(S_{\eta})$. Then we have the equality of $\operatorname{GL}(2)$-invariant cycles $\Sigma=\sum_{\nu}\operatorname{WOrb}(S(\nu)).$ ###### Proof. We first show that the cycle $D=\Sigma-\sum_{\nu\in\Gamma}\operatorname{WOrb}(S(\nu))$ is effective using Proposition 2.4. For every vertex $\nu$ of $\Gamma$, let $\mathcal{X}_{\nu}$ be the surface obtained by blowing down all components of $X$ except $X_{\nu}$. Then we have an isomorphism $\mathcal{X}_{\nu}\cong{\bf P}^{1}\times\Delta$; fix such an isomorphism and identify the two. Choose an arbitrary vertex of $\Gamma$ and call it $0$. For every $\nu$, let $g_{\nu}\in\operatorname{PGL}_{2}(\eta)$ be the unique element such that $\mathcal{X}_{0,\eta}={\bf P}^{1}\times\eta\xrightarrow{g_{\nu}}{\bf P}^{1}\times\eta=\mathcal{X}_{\nu,\eta}$ represents the identity map $\mathcal{X}_{0,\eta}=\mathcal{X}_{\eta}\to\mathcal{X}_{\nu,\eta}=\mathcal{X}_{\eta}.$ For $\mu\neq\nu$, the identity map on the generic fiber does not extend to an isomorphism $\mathcal{X}_{\mu}\to\mathcal{X}_{\nu}$. Therefore, $g_{\mu}g_{\nu}^{-1}\in\operatorname{PGL}_{2}(\eta)$ does not extend to an element of $\operatorname{PGL}_{2}(\Delta)$. On $\mathcal{X}_{\nu}={\bf P}^{1}\times\Delta$, the limit of $[S_{\eta}]{\colon}\eta\to{\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$ is $[S(\nu)]\in{\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2}).$ Now, Proposition 2.4 applied to $M={\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$ and $G=\operatorname{PGL}_{2}$ implies that $D=\Sigma-\sum_{\nu}\operatorname{WOrb}(S(\nu))$ is effective. Let $N=(d-r)(r+1)$. We apply the Wronskian map $\rho{\colon}{\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})\to{\bf P}\operatorname{Sym}^{N}\mathbf{k}^{2}$ and see that the cycle $\rho_{*}D=\rho_{*}\Sigma-\sum_{\nu}\rho_{*}\operatorname{WOrb}(S(\nu))$ is effective. Let $R_{\eta}$ be the ramification divisor of $S_{\eta}$ and $R(\nu)$ the ramification divisor of $S(\nu)$. Then the cycle $\rho_{*}\Sigma$ on ${\bf P}\operatorname{Sym}^{N}\mathbf{k}^{2}$ is the limit of the cycle $\operatorname{WOrb}(R_{\eta})$ and $\rho_{*}\operatorname{WOrb}(S(\nu))$ is the cycle $\operatorname{WOrb}(R(\nu))$. Aluffi and Faber compute the degree of the weighted orbit closure for any divisor on ${\bf P}^{1}$ (see [1, Proposition 1.3] and note that in their paper this degree is called the “pre- degree”). Using their formula, it is straightforward to check that $\deg\rho_{*}D=0$ (see Remark 6.2). Since $\rho_{*}D$ is effective, we conclude that $\rho_{*}D=0$ and since $\rho$ is finite, that $D=0$. ∎ ###### Remark 6.2. We indicate how to check that the pre-degree of the orbit of $R_{\eta}$ is the sum of the pre-degrees of the orbits of $R(\nu)$ using [1, Proposition 1.3]. The pre-degree of the orbit closure of a divisor depends only on the multiplicities of the points in the divisor. More precisely, by [1, Proposition 1.3], the pre-degree for a divisor of degree $N$ with multiplicities $m_{1},\dots,m_{s}$ is given by (24) $\begin{split}p(m_{1},\dots,m_{s})&=N^{3}-3N\sum m_{i}^{2}+2\sum m_{i}^{3}\\\ &=\sum_{i\neq j\neq k}m_{i}m_{j}m_{k}.\end{split}$ To understand the multiplicities in the divisors $R(\nu)$, consider the labelled dual graph $\Gamma$ and replace the vanishing sequences by their weights, where the weight of the sequence $(a_{0},\dots,a_{r})$ is $\sum_{i}(a_{i}-i)$. Then the multiplicities in $R_{\eta}$ are the labels of the dangling half-edges and the multiplicities of $R(\nu)$ are the labels of the half-edges incident to $\nu$. The sum of the labels of all the half-edges incident to a vertex is $N=(d-r)(r+1)$, and the sum of the two labels of a full-edge is also $N$. Consider the operation of contracting an edge of $\Gamma$ as shown below $\leavevmode\hbox to248.02pt{\vbox to31.94pt{\pgfpicture\makeatletter\hbox{\hskip 212.2458pt\lower-22.05939pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }{}{{}}{}{{}}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{{}\pgfsys@moveto{4.71387pt}{0.0pt}\pgfsys@curveto{4.71387pt}{2.60342pt}{2.60342pt}{4.71387pt}{0.0pt}{4.71387pt}\pgfsys@curveto{-2.60342pt}{4.71387pt}{-4.71387pt}{2.60342pt}{-4.71387pt}{0.0pt}\pgfsys@curveto{-4.71387pt}{-2.60342pt}{-2.60342pt}{-4.71387pt}{0.0pt}{-4.71387pt}\pgfsys@curveto{2.60342pt}{-4.71387pt}{4.71387pt}{-2.60342pt}{4.71387pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ } }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{0.0pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {}{{}}{}{{}}{}{{}}{}{{}}{}{{}}{}{{}}{}\pgfsys@moveto{-28.45276pt}{-14.22638pt}\pgfsys@moveto{0.0pt}{-14.22638pt}\pgfsys@moveto{28.45276pt}{-14.22638pt}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-32.5058pt}{-15.10178pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$m_{1}$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-3.75pt}{-16.72638pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\ldots}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{24.46097pt}{-15.31139pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$m_{s}$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{-4.39507pt}{-2.19765pt}\pgfsys@lineto{-20.86671pt}{-10.43394pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{4.39507pt}{-2.19765pt}\pgfsys@lineto{20.92796pt}{-10.46457pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \par\pgfsys@beginscope\pgfsys@invoke{ }{{}} {}{{}}{}{{}}{}\pgfsys@moveto{-184.94292pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{{}\pgfsys@moveto{-180.22905pt}{0.0pt}\pgfsys@curveto{-180.22905pt}{2.60342pt}{-182.3395pt}{4.71387pt}{-184.94292pt}{4.71387pt}\pgfsys@curveto{-187.54634pt}{4.71387pt}{-189.65678pt}{2.60342pt}{-189.65678pt}{0.0pt}\pgfsys@curveto{-189.65678pt}{-2.60342pt}{-187.54634pt}{-4.71387pt}{-184.94292pt}{-4.71387pt}\pgfsys@curveto{-182.3395pt}{-4.71387pt}{-180.22905pt}{-2.60342pt}{-180.22905pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{-184.94292pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ } }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-184.94292pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {}{{}}{}{{}}{}{{}}{}{{}}{}{{}}{}{{}}{}\pgfsys@moveto{-204.85976pt}{-14.22638pt}\pgfsys@moveto{-184.94292pt}{-14.22638pt}\pgfsys@moveto{-165.02608pt}{-14.22638pt}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-208.9128pt}{-15.10178pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$m_{1}$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-188.69292pt}{-16.72638pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\ldots}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-168.91579pt}{-15.05276pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$m_{\ell}$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{-188.94135pt}{-2.85634pt}\pgfsys@lineto{-197.27371pt}{-8.80864pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{-180.94449pt}{-2.85634pt}\pgfsys@lineto{-172.44879pt}{-8.9253pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope \pgfsys@beginscope\pgfsys@invoke{ }{{}} {}{{}}{}{{}}{}\pgfsys@moveto{-113.81102pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{{}\pgfsys@moveto{-109.09715pt}{0.0pt}\pgfsys@curveto{-109.09715pt}{2.60342pt}{-111.2076pt}{4.71387pt}{-113.81102pt}{4.71387pt}\pgfsys@curveto{-116.41444pt}{4.71387pt}{-118.52489pt}{2.60342pt}{-118.52489pt}{0.0pt}\pgfsys@curveto{-118.52489pt}{-2.60342pt}{-116.41444pt}{-4.71387pt}{-113.81102pt}{-4.71387pt}\pgfsys@curveto{-111.2076pt}{-4.71387pt}{-109.09715pt}{-2.60342pt}{-109.09715pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{-113.81102pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ } }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-113.81102pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {}{{}}{}{{}}{}{{}}{}{{}}{}{{}}{}{{}}{}\pgfsys@moveto{-133.72786pt}{-14.22638pt}\pgfsys@moveto{-113.81102pt}{-14.22638pt}\pgfsys@moveto{-93.89418pt}{-14.22638pt}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-140.12201pt}{-14.88943pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$m_{\ell+1}$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-117.56102pt}{-16.72638pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\ldots}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-97.88597pt}{-15.31139pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$m_{s}$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{-117.80975pt}{-2.85597pt}\pgfsys@lineto{-125.49179pt}{-8.34254pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{-109.81259pt}{-2.85634pt}\pgfsys@lineto{-101.41898pt}{-8.85239pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{{}} {{{{{}}{}{}{}{}{{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}{}{{{}{}}}{}{}{{{}{}}}{}\pgfsys@moveto{-180.02905pt}{0.0pt}\pgfsys@lineto{-118.72488pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-169.59052pt}{3.533pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$v$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-133.58542pt}{3.533pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$w$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@beginscope\pgfsys@invoke{ }{{}} {{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-61.90552pt}{-1.8894pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{$\leadsto$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope }\pgfsys<EMAIL_ADDRESS> Since $N=v+w=m_{1}+\dots+m_{\ell}+v=m_{\ell+1}+\dots+m_{s}+w,$ we have $v=m_{\ell+1}+\dots+m_{s}$ and $w=m_{1}+\dots+m_{\ell}$. From (24) it follows that $p(m_{1},\dots,m_{\ell},v)+p(m_{\ell+1},\dots,m_{s},w)=p(m_{1},\dots,m_{s}).$ By contracting the edges one by one, we reduce the sum of the pre-degrees for $R(\nu)$ to a single sum, which is the pre-degree for $R_{\eta}$. ###### Corollary 6.3. In the setup of Theorem 6.1, in $A^{*}_{\operatorname{GL}(2)}({\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2}))$ we have $[\operatorname{WOrb}(S_{\eta})]=\sum_{\nu}[\operatorname{WOrb}(S(\nu))].$ ###### Proof. Since $[\operatorname{WOrb}(S_{\eta})]=[\Sigma]$, the statement follows from Theorem 6.1. ∎ Recall that the class $\operatorname{WOrb}(S)$ depends only on the ramification profile of $S$. The ramification profile of $S_{\eta}$ is recorded by the labels on the dangling half-edges of $\Gamma$. The ramification profile of $S(\nu)$ is recorded by the labels on the half-edges incident to $\nu$. Thus, the equality in Corollary 6.3 can be read off purely from the labelled graph $\Gamma$. A natural question is: which labelled graphs $\Gamma$ arise from degenerations of linear series? Here, the obvious necessary conditions are also sufficient. Let $\Gamma$ be a tree with $m$ dangling half-edges. Suppose all the half- edges of $\Gamma$, including the full-edges considered as two half-edges, are labelled with increasing $(r+1)$-tuples of integers in $\\{0,\dots,d\\}$. Assume that the two labels on a full-edge are complementary. ###### Proposition 6.4. Let $\Gamma$ be a labelled graph as above. Suppose for every vertex $\nu$ of $\Gamma$, there is a linear series $S(\nu)$ of rank $r$ and degree $d$ on ${\bf P}^{1}$ whose ramification profile agrees with the multi-set of labels of half-edges incident to $\nu$. Then $\Gamma$ arises from a degeneration. That is, there exists a $\Delta$, a semi-stable $m$-pointed curve $(\mathcal{X}\to\Delta,p_{1},\dots,p_{m})$ and a refined linear series $\mathcal{S}$ on $\mathcal{X}$ such that the labelled dual graph of its central fiber is $\Gamma$. ###### Proof. The proof is a simple application of the smoothing theorem for $g^{r}_{d}$’s [10, Theorem 3.4]. Let $(X,\overline{p}_{1},\dots,\overline{p}_{m})$ be a semi-stable curve with dual graph $\Gamma$. We assemble the linear series $S(\nu)$ together to get a refined limit linear series $S$ on $X$. Note that the expected dimension of the space $G^{r}_{d}$ on a rational curve with a prescribed ramification divisor is $0$. Since the Wronskian map is finite and the ramification divisors of all aspects of $S$ are prescribed, $S$ is an isolated point of $G^{r}_{d}(X,\overline{p}_{1},\dots,\overline{p}_{m})$. By [10, Theorem 3.4], there exists a $\Delta$ and a smoothing $(\mathcal{X}\to\Delta,p_{1},\dots,p_{m},\mathcal{S})$ of $(X,\overline{p}_{1},\dots,\overline{p}_{m},S)$. ∎ ###### Remark 6.5. The existence of a $g^{r}_{d}$ on ${\bf P}^{1}$ with prescribed ramification profile is a Schubert theoretic condition on ${\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$. Such a series exists if and only if the product of certain Schubert classes on the Grassmannian is non- zero. See the remark after [9, Theorem 2.3] for a precise statement. Theorem 6.1 asserts an equality of $\operatorname{GL}(2)$-invariant cycles on ${\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$. It can be easily upgraded to an equality of $\operatorname{GL}(r+1)\times\operatorname{GL}(2)$-invariant cycles on $\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$. ###### Proposition 6.6. In the notation of Theorem 6.1, let $s_{\eta}\colon\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})_{\eta}$ be a homomorphism with image $S_{\eta}$. Let $s_{\nu}\in\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ be homomorphisms with images $S(\nu)$. Let $\Sigma$ be the flat limit of $\operatorname{WOrb}(s_{\eta})$. Then, on $\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$, we have the equality of $\operatorname{GL}(r+1)\times\operatorname{GL}(2)$-invariant cycles $\Sigma=\sum_{\nu}\operatorname{WOrb}(s_{\nu}).$ ###### Proof. Let $U\subset\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ be the open subset of injective homomorphisms. The action of $\operatorname{GL}(r+1)$ on $U$ is free and the map $\operatorname{im}\colon U\to{\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$ taking a homomorphism to its image is the quotient by $\operatorname{GL}(r+1)$. On $U$, the assertions follow from Theorem 6.1. It suffices to prove that $\Sigma$ does contain any $(r+1)^{2}+3$-dimensional components in the complement of $U$. We claim that $\Sigma$ is the closure of $\Sigma\cap U$, from which the statement follows. To see this, take $x\in\Sigma$. Then, possibly after a base change, we can choose $x_{\eta}\in\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})_{\eta}$ in the $\operatorname{GL}(r+1)\times\operatorname{GL}(2)$-orbit of $s_{\eta}$ such that the flat limit of $x_{\eta}$ is $x$. Let $S\subset\operatorname{Sym}^{d}\mathbf{k}^{2}$ be the $(r+1)$-dimensional subspace corresponding to the flat limit of $\operatorname{im}(x_{\eta})$ in ${\bf Gr}(r+1,\operatorname{Sym}^{d}\mathbf{k}^{2})$. Then it is easy to see that $S$ contains the image of $x$. Let $\ell$ be the dimension of the image of $x$. By making a change of coordinates on $\mathbf{k}^{r+1}$ if necessary, assume that $x$ sends the last $r+1-\ell$ standard basis vectors to 0. Let $s^{\prime}\in\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ be a map that agrees with $x$ on the first $\ell$ basis vectors and whose image is $S$. Then $s^{\prime}\in\Sigma\cap U$. Consider the diagonal one parameter subgroup $\lambda_{t}\to\operatorname{GL}(r+1)$ with diagonal entries $(\underbrace{1,\dots,1}_{\ell},\underbrace{t,\dots,t}_{r+1-\ell}).$ Then, by construction, $\lim_{t\to 0}\lambda_{t}(s^{\prime})=x$. Since $\Sigma\cap U$ is $\operatorname{GL}(r+1)$-invariant, and $s^{\prime}\in\Sigma\cap U$, we get that $\lambda_{t}(s^{\prime})\in\Sigma\cap U$ for $t\neq 0$. We conclude that $x$ lies in the closure of $\Sigma\cap U$. ∎ ###### Corollary 6.7. In the setup of Proposition 6.6, in $A^{*}_{\operatorname{GL}(r+1)\times\operatorname{GL}(2)}(\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2}))$, we have $[\operatorname{WOrb}(s_{\eta})]=\sum_{\nu}[\operatorname{WOrb}(s_{\nu})].$ ###### Proof. Since $[\operatorname{WOrb}(s_{\eta})]=[\Sigma]$, the statement follows from Proposition 6.6. ∎ Finally, we discuss the specialisation mentioned in the introduction (Theorem 1.5). Let $p_{1},\dots,p_{m}$ be sections $\Delta\to{\bf P}^{1}\times\Delta$ that are distinct on the generic fiber but $p_{1}=\dots=p_{\ell}$ on the central fiber (with no other coincidences). Let $S_{\eta}$ be a linear series on the generic fiber whose ramification points are $p_{1},\dots,p_{m}$ and whose vanishing sequence at $p_{i}$ is $a(i)=(a_{0}(i),\dots,a_{r}(i))$. Let $S$ be the flat limit of $S_{\eta}$. It is easy to see that the vanishing sequence of $S$ at $p_{i}$ for $i>\ell$ is $a(i)$. Let $c=(c_{0},\dots,c_{r+1})$ be the vanishing sequence of $S$ at $p=p_{1}=\dots=p_{\ell}$. Let $c^{\prime}=(d-c_{r},\dots,d-c_{0})$ be the complementary sequence. ###### Proposition 6.8. In the setup above, there exists a linear series of rank $r$ and degree $d$ on ${\bf P}^{1}$ with ramification profile $\\{a(1),\dots,a(\ell),c^{\prime}\\}$. Furthermore, on $\operatorname{Hom}(\mathbf{k}^{r+1},\operatorname{Sym}^{d}\mathbf{k}^{2})$ we have the equality of $\operatorname{GL}(r+1)\times\operatorname{GL}(2)$-equivariant classes $[\operatorname{WOrb}(a(1),\dots,a(m))]=[\operatorname{WOrb}(a(1),\dots,a(\ell),c^{\prime})]+[\operatorname{WOrb}(a(\ell+1),\dots,a(m),c)].$ ###### Proof. The existence of the linear series can be deduced using Schubert calculus. We give an alternate geometric proof using the smoothing theorem. Consider the pointed stable limit of $({\bf P}^{1}_{\eta},p_{1},\dots,p_{m})$. The dual graph of limit, together with the labelling given by the limit linear series, has the form $\leavevmode\hbox to164.33pt{\vbox to42.54pt{\pgfpicture\makeatletter\hbox{\hskip 32.52437pt\lower-31.57028pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{{}\pgfsys@moveto{9.66846pt}{0.0pt}\pgfsys@curveto{9.66846pt}{5.3398pt}{5.3398pt}{9.66846pt}{0.0pt}{9.66846pt}\pgfsys@curveto{-5.3398pt}{9.66846pt}{-9.66846pt}{5.3398pt}{-9.66846pt}{0.0pt}\pgfsys@curveto{-9.66846pt}{-5.3398pt}{-5.3398pt}{-9.66846pt}{0.0pt}{-9.66846pt}\pgfsys@curveto{5.3398pt}{-9.66846pt}{9.66846pt}{-5.3398pt}{9.66846pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ } }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-3.61632pt}{-3.41666pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{$T$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{{}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{{}\pgfsys@moveto{104.85446pt}{0.0pt}\pgfsys@curveto{104.85446pt}{2.60342pt}{102.74402pt}{4.71387pt}{100.1406pt}{4.71387pt}\pgfsys@curveto{97.53717pt}{4.71387pt}{95.42673pt}{2.60342pt}{95.42673pt}{0.0pt}\pgfsys@curveto{95.42673pt}{-2.60342pt}{97.53717pt}{-4.71387pt}{100.1406pt}{-4.71387pt}\pgfsys@curveto{102.74402pt}{-4.71387pt}{104.85446pt}{-2.60342pt}{104.85446pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{100.1406pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ } }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{100.1406pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{ {}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{}}{} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-29.19136pt}{-21.7968pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$a(1)$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{{}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-2.625pt}{-26.83728pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize\ldots}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{ {}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ }}{ } {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{16.5468pt}{-21.7968pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$a(\ell)$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{-7.42091pt}{-6.50493pt}\pgfsys@lineto{-14.84631pt}{-13.0138pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{0.0pt}{-9.86845pt}\pgfsys@lineto{0.0pt}{-18.40428pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{7.3795pt}{-6.55222pt}\pgfsys@lineto{14.65704pt}{-13.0138pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{ {}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{}}{} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{61.64719pt}{-18.29337pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$a(\ell+1)$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{{}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{97.5156pt}{-21.88269pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize\ldots}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{ {}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ }}{ } {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{113.18396pt}{-18.29337pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$a(m)$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{96.0054pt}{-2.6545pt}\pgfsys@lineto{85.32564pt}{-9.51036pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{100.1406pt}{-4.91386pt}\pgfsys@lineto{100.1406pt}{-13.44969pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{103.9785pt}{-3.0686pt}\pgfsys@lineto{112.03514pt}{-9.51036pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{{}} {{{{{}}{}{}{}{}{{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}{}{{{}{}}}{}{}{{{}{}}}{}\pgfsys@moveto{9.86845pt}{0.0pt}\pgfsys@lineto{95.22673pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{24.8862pt}{3.533pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$c^{\prime}$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{76.6407pt}{3.533pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$c$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope }\pgfsys<EMAIL_ADDRESS> In this diagram, $T$ is a labelled tree with dangling half-edges labelled $a(1),\dots,a(\ell)$, and one of its vertices, say $t$, is connected to the remaining vertex, say $o$, by an edge that has the label $c^{\prime}$ near $t$ and $c$ near $o$. We apply Proposition 6.4 to $T$ together with the dangling edges labelled $a(1),\dots,a(\ell)$ and $c^{\prime}$. A geometric general fiber of the asserted degeneration gives a linear series of rank $r$ and degree $d$ on ${\bf P}^{1}$ with ramification profile $\\{a(1),\dots,a(\ell),c^{\prime}\\}$. By applying Proposition 6.4 to the two vertex graph $\leavevmode\hbox to155.87pt{\vbox to37.58pt{\pgfpicture\makeatletter\hbox{\hskip 29.02094pt\lower-26.61569pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{{}\pgfsys@moveto{4.71387pt}{0.0pt}\pgfsys@curveto{4.71387pt}{2.60342pt}{2.60342pt}{4.71387pt}{0.0pt}{4.71387pt}\pgfsys@curveto{-2.60342pt}{4.71387pt}{-4.71387pt}{2.60342pt}{-4.71387pt}{0.0pt}\pgfsys@curveto{-4.71387pt}{-2.60342pt}{-2.60342pt}{-4.71387pt}{0.0pt}{-4.71387pt}\pgfsys@curveto{2.60342pt}{-4.71387pt}{4.71387pt}{-2.60342pt}{4.71387pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ } }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{0.0pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{{}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{{}\pgfsys@moveto{99.89987pt}{0.0pt}\pgfsys@curveto{99.89987pt}{2.60342pt}{97.78943pt}{4.71387pt}{95.186pt}{4.71387pt}\pgfsys@curveto{92.58258pt}{4.71387pt}{90.47214pt}{2.60342pt}{90.47214pt}{0.0pt}\pgfsys@curveto{90.47214pt}{-2.60342pt}{92.58258pt}{-4.71387pt}{95.186pt}{-4.71387pt}\pgfsys@curveto{97.78943pt}{-4.71387pt}{99.89987pt}{-2.60342pt}{99.89987pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{95.186pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ } }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{95.186pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{ {}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{}}{} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-25.68793pt}{-18.29337pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$a(1)$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{{}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-2.625pt}{-21.88269pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize\ldots}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{ {}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ }}{ } {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{13.04337pt}{-18.29337pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$a(\ell)$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{-3.73637pt}{-3.1915pt}\pgfsys@lineto{-11.13383pt}{-9.51036pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{0.0pt}{-4.91386pt}\pgfsys@lineto{0.0pt}{-13.44969pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{3.71223pt}{-3.21947pt}\pgfsys@lineto{10.96613pt}{-9.51036pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{ {}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{}}{} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{56.6926pt}{-18.29337pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$a(\ell+1)$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{{}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{92.561pt}{-21.88269pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize\ldots}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{{{ {}}}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ }}{ } {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{108.22937pt}{-18.29337pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$a(m)$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{91.05081pt}{-2.6545pt}\pgfsys@lineto{80.37105pt}{-9.51036pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{95.186pt}{-4.91386pt}\pgfsys@lineto{95.186pt}{-13.44969pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}\pgfsys@moveto{99.02391pt}{-3.0686pt}\pgfsys@lineto{107.08055pt}{-9.51036pt}\pgfsys@stroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {{}}{}{}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}{}{{}} {{{{{}}{}{}{}{}{{}}}}}{}{{{{{}}{}{}{}{}{{}}}}}{{}}{}{}{}{}{}{{{}{}}}{}{}{{{}{}}}{}\pgfsys@moveto{4.91386pt}{0.0pt}\pgfsys@lineto{90.27214pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{ }\hbox{\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{19.93161pt}{3.533pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$c^{\prime}$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{71.68611pt}{3.533pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{\scriptsize$c$}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}},$ we see that the graph above arises from a degeneration. We now apply Corollary 6.7 to this degeneration. ∎ ## References * [1] P. Aluffi and C. Faber. Linear orbits of $d$-tuples of points in $\mathbb{P}^{1}$. J. Reine Angew. Math., 445:205–220, 1993. * [2] P. Aluffi and C. Faber. Linear orbits of smooth plane curves. J. Algebr. Geom., 2(1):155–184, 1993. * [3] P. Aluffi and C. Faber. Linear orbits of arbitrary plane curves. Michigan Math. J., 48:1–37, 2000. * [4] A. Berget and A. Fink. Equivariant Chow classes of matrix orbit closures. Transform. Groups, 22(3):631–643, 2017. * [5] A. S. Buch and R. Rimányi. A formula for non-equioriented quiver orbits of type $A$. J. Algebr. Geom., 16(3):531–546, 2007. * [6] A. Deopurkar, A. Patel, and D. Tseng. A universal formula for counting cubic surfaces, 2021. Pre-print, arXiv:2109.12672. * [7] D. Edidin and W. Graham. Equivariant intersection theory (With an appendix by Angelo Vistoli: The Chow ring of ${\mathcal{M}}_{2}$). Invent. Math., 131(3):595–644, 1998. * [8] D. Edidin and W. Graham. Localization in equivariant intersection theory and the Bott residue formula. Am. J. Math., 120(3):619–636, 1998. * [9] D. Eisenbud and J. Harris. Divisors on general curves and cuspidal rational curves. Invent. Math., 74:371–418, 1983. * [10] D. Eisenbud and J. Harris. Limit linear series: Basic theory. Invent. Math., 85:337–371, 1986. * [11] L. M. Fehér, R. Rimányi, and A. Weber. Characteristic classes of orbit stratifications, the axiomatic approach. In Schubert calculus and its applications in combinatorics and representation theory. Selected papers presented at the “International Festival in Schubert Calculus”, Guangzhou, China, November 6–10, 2017, pages 223–249. Singapore: Springer, 2020. * [12] M. Lee, A. Patel, H. Spink, and D. Tseng. Orbits in $(\mathbb{P}^{r})^{n}$ and equivariant quantum cohomology. Adv. Math., 362:79, 2020. * [13] M. Lee, A. Patel, and D. Tseng. Equivariant degenerations of plane curve orbits, 2019. Pre-print, arXiv:1903.10069. * [14] B. Osserman. A limit linear series moduli scheme. Ann. Inst. Fourier, 56(6):1165–1205, 2006.
# Top-Down Synthesis for Library Learning Matthew Bowers Massachusetts Institute of TechnologyCambridgeMAUSA <EMAIL_ADDRESS>, Theo X. Olausson Massachusetts Institute of TechnologyCambridgeMAUSA<EMAIL_ADDRESS>, Lionel Wong Massachusetts Institute of TechnologyCambridgeMAUSA<EMAIL_ADDRESS>, Gabriel Grand Massachusetts Institute of TechnologyCambridgeMAUSA<EMAIL_ADDRESS>, Joshua B. Tenenbaum Massachusetts Institute of TechnologyCambridgeMAUSA , Kevin Ellis Cornell UniversityIthacaNYUSA and Armando Solar-Lezama Massachusetts Institute of TechnologyCambridgeMAUSA (2023; 2022-07-07; 2022-11-07) ###### Abstract. This paper introduces corpus-guided top-down synthesis as a mechanism for synthesizing library functions that capture common functionality from a corpus of programs in a domain specific language (DSL). The algorithm builds abstractions directly from initial DSL primitives, using syntactic pattern matching of intermediate abstractions to intelligently prune the search space and guide the algorithm towards abstractions that maximally capture shared structures in the corpus. We present an implementation of the approach in a tool called Stitch and evaluate it against the state-of-the-art deductive library learning algorithm from DreamCoder. Our evaluation shows that Stitch is 3-4 orders of magnitude faster and uses 2 orders of magnitude less memory while maintaining comparable or better library quality (as measured by compressivity). We also demonstrate Stitch’s scalability on corpora containing hundreds of complex programs that are intractable with prior deductive approaches and show empirically that it is robust to terminating the search procedure early—further allowing it to scale to challenging datasets by means of early stopping. Program Synthesis, Library Learning, Abstraction Learning ††copyright: rightsretained††doi: 10.1145/3571234††journalyear: 2023††submissionid: popl23main-p278-p††journal: PACMPL††journalvolume: 7††journalnumber: POPL††article: 41††publicationmonth: 1††ccs: Software and its engineering Automatic programming ## 1\. Introduction Figure 1. (A) Given an initial DSL of lower-level primitives and a corpus of programs written in the initial DSL, (B) Stitch uses a fast and memory efficient corpus-guided top-down search algorithm to construct function abstractions which maximally capture shared structure across the corpus of programs. (C) Stitch automatically rewrites the initial corpus using the learned library functions. One way programmers manage complexity is by hiding functionality behind functional abstractions. For example, consider the graphics programs at the bottom of Fig. 1A. Each uses a generic set of drawing primitives and renders a technical schematic of a hardware component, shown to the left of each program. Faced with the task of writing more of these rendering programs, an experienced human programmer likely would not continue using these low-level primitives. Instead, they would introduce new functional abstractions, like the one at the bottom of Fig. 1B which renders a regular polygon given a size and number of sides. Useful abstractions like these allow more concise and legible programs to render the existing schematics. More importantly, well- written abstractions should generalize, making it easier to write new graphics programs for similar graphics tasks. Recently, the program synthesis community has introduced new approaches that can mimic this process, automatically building a library of functional abstractions in order to tackle more complex synthesis problems (Ellis et al., 2021, 2020; Shin et al., 2019; Lázaro-Gredilla et al., 2019; Dechter et al., 2013). One popular approach to library learning is to search for common tree fragments across a corpus of programs, which can be introduced as new abstractions (Shin et al., 2019; Lázaro-Gredilla et al., 2019; Dechter et al., 2013). Ellis et al. 2021, however, introduces an algorithm that reasons about variable bindings to abstract out well-formed functions instead of just tree fragments. While it produces impressive results, the system in Ellis et al. 2021 takes a deductive approach to library learning that is difficult to scale to larger datasets of longer and deeper input programs. This approach is deductive in that it uses semantics-preserving rewrite rules to attempt to refactor existing programs to expose shared structure. This requires representing and evaluating an exponentially large space of proposed refactorings to identify common functionality across the corpus. Prior work, such as Ellis et al. 2021, 2020, approaches this challenge by combining a dynamic bottom-up approach to refactoring with version spaces to more efficiently search over the refactored programs. However, these deductive approaches face daunting memory and search requirements as the corpus scales in size and complexity. This paper introduces an alternate approach to library learning, while preserving the focus on well-formed function abstractions from Ellis et al. 2021. Instead of taking a deductive approach based on refactoring the corpus with rewrite rules, we directly synthesize abstractions. We call this approach corpus-guided top-down synthesis, and it is based on the insight that when applied to the task of synthesizing abstractions, top-down search can be guided precisely towards discovering shared abstractions over a set of existing training programs. At every step of the search, syntactic comparisons between a partially constructed abstraction and the set of training programs can be used to strongly constrain the search space and direct the search towards abstractions that capture the greatest degree of shared syntactic structure. We implement this approach in Stitch, a corpus-guided top-down library learning tool written in Rust (Fig. 1). Stitch is open-source and available both as a Python package and a Rust crate; the code, installation instructions, and tutorials are available at the GitHub111https://github.com/mlb2251/stitch. We evaluate Stitch through a series of experiments (Section 6), and find that: Stitch learns libraries of comparable quality to those found by the algorithm of Ellis et al. 2021 on their iterative bootstrapped library learning task, while being 3-4 orders of magnitude faster and using 2 orders of magnitude less memory (Section 6.1); Stitch learns high quality libraries within seconds to single-digit minutes when run on corpora containing a few hundred programs with mean lengths between 76–189 symbols (sourced from Wong et al. 2022), while even the simplest of these corpora lies beyond the reach of the algorithm of Ellis et al. 2021 (Section 6.2); and that Stitch degrades gracefully when resources are constrained (Section 6.3). We also perform ablation studies to expose the relative impact of different optimizations in Stitch (Section 6.4). Finally, we show that Stitch is complementary to deductive rewrite approaches (Section 6.5). In summary, our paper makes the following contributions: 1. (1) Corpus-guided top-down synthesis (CTS): A novel, strongly-guided branch-and- bound algorithm for synthesizing function abstractions (Sec. 3). 2. (2) CTS for program compression: An instantiation of the CTS framework for utility functions favoring abstractions that compress a corpus of programs (Sec. 4). 3. (3) Stitch: A parallel Rust implementation of CTS for compression that achieves 3-4 orders of magnitude of speed and memory improvements over prior work, and the analysis of its performance, scaling, and optimizations through several experiments (Sec. 6). ## 2\. Overview Figure 2. (A) Schematic of corpus-guided top-down search (CTS). Partial abstractions can contain holes indicating unfinished subtrees, denoted ??, while complete abstractions do not contain holes. (B) Partial and complete abstractions match at locations in the corpus. The blue partial abstraction can be expanded into the green complete abstraction so the green abstraction matches in a subset of places that the blue abstraction does. For complete abstractions, matching indicates that the subtree can be rewritten to use the abstraction. In this section, we build intuition for the algorithmic insights that power Stitch. As a running example, we focus on learning a single abstraction from the following corpus of programs: (1) $\displaystyle\lambda x.\ \texttt{{\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}+ 3} ({\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}*} (+ 2 4) 2)}$ $\displaystyle\lambda xs.\ \texttt{map ($\lambda x.\ ${\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}+ 3} ({\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}*} 4 (+ 3 x))) xs}$ $\displaystyle\lambda x.\ \texttt{* 2 ({\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}+ 3} ({\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}*} x (+ 2 1)))}$ The notion of what a good abstraction is depends on the application, so our algorithm is generic over the utility function that we seek to maximize. Following prior work (Ellis et al., 2021; Shin et al., 2019; Dechter et al., 2013) we focus on compression as a utility function: a good abstraction is one which minimizes the size of the corpus when rewritten with the abstraction. The utility function used by Stitch is detailed in Section 4 and corresponds exactly to the compression objective, but at a high level the function seeks to maximize the product of the size of the abstraction and the number of locations where the abstraction can be used. This product balances two key features of a highly compressive abstraction: the abstraction should be general enough that it applies in many locations, but specific enough that it captures a lot of structure at each location. The optimal abstraction that maximizes our utility in this example is: (2) $\texttt{{\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}fn0}}=\lambda\alpha.\ \lambda\beta.\texttt{(+ 3 (* $\alpha$ $\beta$))}$ And the shared structure that this abstraction captures is highlighted in blue in Eq. 1. When rewritten to use this abstraction, the size of the resulting programs is minimized: (3) $\displaystyle\lambda x.\ \texttt{{\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}fn0} (+ 2 4) 2}$ $\displaystyle\lambda xs.\ \texttt{map ($\lambda x.\ ${\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}fn0} 4 (+ 3 x)) xs}$ $\displaystyle\lambda x.\ \texttt{* 2 ({\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}fn0} x (+ 2 1))}$ Stitch synthesizes the optimal abstraction directly through top-down search. Top-down search methods construct program syntax trees by iteratively refining partially-completed programs that have unfinished holes, until a complete program that meets the specification is produced (Feser et al., 2015; Polikarpova et al., 2016; Balog et al., 2016; Ellis et al., 2021; Nye et al., 2021; Shah et al., 2020). This kind of search is often made efficient by identifying branches of the search tree that can be pruned away because the algorithm can efficiently determine that none of the programs in that branch can be correct. We aim to apply a similar idea to the problem of synthesizing good abstractions; the idea is to explore the space of functions in the same top-down way, but in search of a function that maximizes the utility measure. The key observation is that this new objective affords even more aggressive pruning opportunities than the traditional correctness objective, allowing us to synthesize optimal abstractions very efficiently. We call this approach corpus-guided top-down search (CTS). ##### Corpus-Guided Top-Down Search. Like other top-down synthesis approaches, our algorithm explores the space of abstractions by repeatedly refining abstractions with holes, as illustrated in Fig. 2A. We call abstractions with holes partial abstractions in contrast to complete abstractions which have no holes. Our top-down algorithm searches over abstraction bodies, so to synthesize the optimal abstraction fn0 in our running example, we synthesize its body: (+ 3 (* $\alpha$ $\beta$)). We say that a partial abstraction matches at a subtree in the corpus if there’s a possible assignment to the holes and arguments that yields the subtree. For example, consider the subtree (+ 3 (* (+ 2 4) 2) from the first program in our running example, shown as a syntax tree at the bottom of figure 2B. The partial abstraction (+ 3 ??) shown in blue matches here with the hole $\texttt{??}=\texttt{(* (+ 2 4) 2)}$, and the complete abstraction (+ 3 (* $\alpha$ $\beta$)) matches here with $\alpha=\texttt{(+ 2 4)}$ and $\beta=\texttt{2}$. For complete abstractions, matching corresponds to being able to use the abstraction to rewrite this subtree, resulting in compression. We refer to the set of subtrees at which a complete or partial abstraction matches as the set of match locations. For example, the three match locations of (+ 3 (* $\alpha$ $\beta$)) are the subtree (+ 3 (* (+ 2 4) 2)) in the first program, (+ 3 (* 4 (+ 3 x))) in the second program, and (+ 3 (* x (+ 2 1))) in the third program. In traditional top-down synthesis, a branch of search can be safely pruned by proving that a program satisfying the specification cannot exist in that branch. In CTS, we can safely prune a branch of search if we can prove that it cannot contain the optimal abstraction. One way to prove this is by computing an upper bound on the utility of abstractions in the branch, and discarding the branch if we have previously found an abstraction with a utility that is higher than this bound. An efficient and conservative way to compute this upper bound is to overapproximate the set of match locations, then upper bound the compressive utility gain from each match location. Our key observation to compute this bound is that during search, expanding a hole in a partial abstraction yields an abstraction that is more precise and thus matches at a subset of the locations that the original matched at. This subset observation is depicted in Fig. 2 where the refinement of the partial abstraction (+ 3 ??) (blue) to the goal (+ 3 (* $\alpha$ $\beta$)) (green) results in a larger abstraction that matches at a subset of the locations (match locations shown at the top of Fig. 2B). An important consequence of this is that the match locations of a partial abstraction serves as an over- approximation of the match locations of any abstraction in this branch of top- down search. To upper bound the compressive utility gained by rewriting at a match location, notice that the most compressive abstraction that matches at a subtree is the constant abstraction corresponding to the subtree itself. For example, the largest possible abstraction at (+ 3 (* (+ 2 4) 2) is just (+ 3 (* (+ 2 4) 2). In other words, the size of the subtree is a strict upper bound on how much compression can be achieved at a match location. Thus we get the upper bound for some partial abstraction $A_{\texttt{??}}$: (4) $U_{\text{upperbound}}(A_{\texttt{??}})=\sum_{e\in\text{matches}(A_{\texttt{??}})}{\text{size}(e)}$ Equipped with this bound, our algorithm performs a branch-and-bound style top- down search, where at each step of search it discards all partial abstractions that have utility upper bounds that are less than the utility of the best complete abstraction found so far. Our full algorithm presented in Section 3 has some additional complexity. It handles rewriting in the presence of variables soundly, it uses an exact utility function that accounts for additional compression gained by using the same variable in multiple places, and it handles situations where match locations overlap and preclude one another. We also introduce two other important forms of pruning while maintaining the optimality of the algorithm, and we use the upper bound as a heuristic to prioritize more promising branches of search first. Figure 3. Visualization of Stitch library learning on the nuts-bolts subdomain from Wong et al. 2022. (A) From the base DSL primitives (top row), Stitch iteratively discovers a series of abstractions that compress programs in the domain. Arrows demonstrate how abstractions from selected iterations build on one another to achieve increasingly higher-level behaviors. (B) Rewriting a single item from the domain with the cumulative benefit of discovered abstractions yields increasingly compact expressions. Colors indicate correspondence between object parts and program fragments: orange = outer octagon, green = ring of six circles, purple = inner circle. ##### Building Up Abstraction Libraries. The top-down search algorithm described above yields a single abstraction. However, we can easily run this algorithm for multiple iterations on a corpus of programs to build up an entire library of abstractions. Fig. 3 illustrates the power of this kind of iterative library learning, which interleaves compression and rewriting. At each iteration, Stitch discovers a single abstraction that is used to rewrite the entire corpus of programs. Successive iterations therefore yield abstractions that build hierarchically on one another, achieving increasingly higher-level behaviors. As the library grows to contain richer and more complex abstractions, individual programs shrink into compact expressions. ##### Structure of the Paper. In the subsequent sections, we formalize our corpus-guided top-down search algorithm (Section 3), its application to the problem of compression (Section 4), and how it may be layered on top of data structures such as version spaces (Section 5). We then report experimental results (Section 6) showcasing both diverse library learning settings as well as ablations of Stitch’s search mechanisms. Finally, we conclude by situating Stitch within the landscape of related work (Section 7) and future work (Section 8) in the areas of library learning and program synthesis. ## 3\. Corpus-Guided Top-Down Search ### 3.1. Problem Setup In this section we provide the definitions necessary to understand our problem. ##### Grammar Our algorithm operates on lambda-calculus expressions with variables represented through de Bruijn indices (de Bruijn, 1972); expressions come from a context-free grammar of the form $e\Coloneqq\lambda.\ e^{\prime}\mid e^{\prime}\ e^{\prime\prime}\mid\$i\mid t$, where $t\in\mathcal{G}_{\texttt{sym}}$ refers to the set of built-in primitives in the domain-specific language. For example, in an arithmetic domain $\mathcal{G}_{\texttt{sym}}$ would include operators like + and constants like 3. We say that $e^{\prime}$ is a subexpression of $e$ if $e=C[e^{\prime}]$, where $C$ is a context as defined in the contextual semantics of Felleisen and Hieb 1992. An expression is closed if it has no free variables, in which case the expression is also a program. A set of programs $\mathcal{P}$ is a corpus. When representing variables through de Bruijn indices, $\$i$ refers to the variable bound by the $i$th closest lambda above it. Thus $\lambda x.\ \lambda y.\ (x\ y)$ is represented as $\lambda.\ \lambda.\ (\$1\ \$0)$. Beta reduction with de Bruijn indices requires upshifting: incrementing free variables in the argument when substitution recurses into a lambda, since the free variables must now point past one more lambda. Similarly, inverse beta reduction requires downshifting. ##### Abstraction Given a grammar $\mathcal{G}$, we define the abstraction grammar $\mathcal{G}_{A}$ as $\mathcal{G}$ extended to include abstraction variables, denoted by Greek letters $\alpha$, $\beta$, etc. Formally, $A\Coloneqq\lambda.\ A^{\prime}\mid A^{\prime}\ A^{\prime\prime}\mid\$i\mid t\mid\alpha$ where $t\in\mathcal{G}_{\texttt{sym}}$. A term $A$ from this grammar represents the body of an abstraction; e.g. the abstraction $\lambda\alpha.\ \lambda\beta.\ (\texttt{+}\ \alpha\ \beta)$ is simply represented by the term $(\texttt{+}\ \alpha\ \beta)$ from the language of $\mathcal{G}_{A}$. ##### Partial Abstraction A partial abstraction $A_{\texttt{??}}$ is an abstraction that can additionally include holes. A hole, denoted by ??, represents an unfinished subtree of the abstraction. Thus, $A_{\texttt{??}}\Coloneqq A\mid\texttt{??}\mid\lambda.\ A_{\texttt{??}}^{\prime}\mid A_{\texttt{??}}^{\prime}\ A_{\texttt{??}}^{\prime\prime}$. Any abstraction is thus also a partial abstraction. Given a grammar $\mathcal{G}$, the grammar of partial abstractions is denoted by $\mathcal{G}_{A_{\texttt{??}}}$. Each hole in a partial abstraction $A$ can be referred to by a unique index, which can be explicitly written as $\texttt{??}_{i}$. ##### Lambda-Aware Unification We introduce lambda-aware unification as a modification of traditional unification adapted to our algorithm. $\textsc{LambdaUnify}(A,e)$ returns a mapping (if one exists) from abstraction variables and holes to expressions $[\alpha_{i}\to e_{i}^{\prime},\ \ldots,\ \texttt{??}_{j}\to u_{j}^{\prime},\ \ldots]$ such that (5) $(\lambda\alpha_{i}.\ \ldots\ \lambda\texttt{??}_{j}.\ \ldots A)\ e_{i}^{\prime}\ \ldots\ u_{j}^{\prime}\ \ldots=e$ through beta reduction. A key difference from traditional unification is that the expression $e$ may be deep inside a program written using de Bruijn indices, so LambdaUnify must perform some index arithmetic in order to generate its output mapping; in particular, raising a subtree out of a lambda during this inverse beta reduction requires downshifting variables in it. The definition of LambdaUnify is presented in Fig. 4 (left). In U-App, $\textsc{merge}(l_{1},l_{2})$ merges two $[\alpha_{i}\to e_{i}]$ mappings to create a new mapping that includes all bindings from $l_{1}$ and $l_{2}$, and fails if the same abstraction variable $\alpha$ maps to different expressions in $l_{1}$ and $l_{2}$. The rule U-Same applies only when the abstraction argument to LambdaUnify is an expression (i.e. it contains no holes or abstraction variables) and is syntactically identical to the expression it is being unified with. In U-Lam, DownshiftAll returns a new mapping with all abstracted expressions $e$ and hole expressions $u$ downshifted using the $\downarrow$ operator presented in Fig. 4 (right). The $\downarrow$ operator has been modified from traditional downshifting to allow for partial abstractions that contain holes $\texttt{??}_{j}$ to break the rules of variable binding, because holes represent unfinished subtrees of the abstraction. In particular, a new syntactic form ”$\&i$” is used to represent a $\$i$ variable that has been downshifted further than traditional downshifting would permit. $\&i$ variables are created by $\downarrow$ when a traditional downshift would otherwise convert a free variable to a (different, incorrect) bound variable. These variables represent references to lambdas present within the body of the abstraction and are allowed in expressions $u_{j}^{\prime}$ bound to holes but not expressions $e_{i}^{\prime}$ bound to abstraction variables. To account for $\&i$ variables, the beta reduction used in Eq. 5, which is defined in Fig. 5, uses a modified upshift operator $\uparrow$ defined to be an inverse to the $\downarrow$ operator. With this modification of beta reduction, any $\&i$ variables with negative indices will ultimately be shifted back to positive indices during reduction. Furthermore, the beta reduction procedure is equivalent to traditional beta reduction when there are no holes and thus no $\&i$ variables. *[left=U-AbsVar] LambdaUnify(α, e) ↝[ α→e ] *[left=U-Hole] LambdaUnify(??_i, e) ↝[ ??_i →e ] *[left=U-App] LambdaUnify(A_1,e_1) ↝l_1 LambdaUnify(A_2,e_2) ↝l_2 l = merge(l_1,l_2) LambdaUnify((A_1 A_2),(e_1 e_2)) ↝l *[left=U-Lam] LambdaUnify(A, e) ↝l’ l = DownshiftAll(l’) LambdaUnify((λ. A),(λ. e)) ↝l *[left=U-Same] LambdaUnify(e,e) ↝[ ] $\displaystyle\textsc{DownshiftAll}([\alpha_{i}\to e_{i}^{\prime},\ \texttt{??}_{j}\to u_{j}^{\prime},\ ...])$ $\displaystyle\ \ \ \ =[\alpha_{i}\to\ \downarrow_{0}e_{i}^{\prime},\ \texttt{??}_{j}\to\ \downarrow_{0}u_{j}^{\prime},\ ...]$ $\displaystyle\downarrow_{d}\lambda.b=\lambda.\downarrow_{d+1}b$ $\displaystyle\downarrow_{d}(fx)=(\downarrow_{d}f\ \downarrow_{d}x)$ $\displaystyle\downarrow_{d}\$i=\begin{cases}\$i,&\text{if }i<d\\\ \$(i-1),&\text{if }i>d\\\ \&(i-1),&\text{if }i=d\\\ \end{cases}$ $\displaystyle\downarrow_{d}\&i=\&(i-1),$ $\displaystyle\downarrow_{d}t=t,\ \text{if $t$ is a primitive (i.e. $t\in\mathcal{G}_{\texttt{sym}}$) }$ Figure 4. (Left) Inference rules for lambda-aware unification. (Right) Definition of DownshiftAll. $\displaystyle(\lambda\alpha_{i}.\ \ldots\ \lambda\texttt{??}_{j}.\ \ldots A)\ e_{i}^{\prime}\ \ldots\ u_{j}^{\prime}\ \ldots$ $\displaystyle\ \ \ \ =[\alpha_{i}\to e_{i}^{\prime},\ \texttt{??}_{j}\to u_{j}^{\prime},\ ...]\circ A$ $\displaystyle l\circ\lambda.b=\lambda.\textsc{UpshiftAll}(l)\circ b$ $\displaystyle l\circ(fx)=(l\circ f)(l\circ x)$ $\displaystyle l\circ\$i=\$i$ $\displaystyle l\circ\alpha_{i}=l[\alpha_{i}]$ $\displaystyle l\circ\texttt{??}_{i}=l[\texttt{??}_{i}]$ $\displaystyle l\circ t=t,\ \text{for $t\in\mathcal{G}_{\texttt{sym}}$}$ $\displaystyle\textsc{UpshiftAll}([\alpha_{i}\to e_{i}^{\prime},\ \texttt{??}_{j}\to u_{j}^{\prime},\ ...])$ $\displaystyle\ \ \ \ =[\alpha_{i}\to\ \uparrow_{0}e_{i}^{\prime},\ \texttt{??}_{j}\to\ \uparrow_{0}u_{j}^{\prime},\ ...]$ $\displaystyle\uparrow_{d}\lambda.b=\lambda.\uparrow_{d+1}b$ $\displaystyle\uparrow_{d}(fx)=(\uparrow_{d}f\ \uparrow_{d}x)$ $\displaystyle\uparrow_{d}\$i=\begin{cases}\$i,&\text{if }i<d\\\ \$(i+1),&\text{if }i\geq d\\\ \end{cases}$ $\displaystyle\uparrow\&i=\begin{cases}\&(i+1),&\text{if }i+1\neq d\\\ \$(i+1),&\text{if }i+1=d\\\ \end{cases}$ $\displaystyle\uparrow_{d}t=t,\ \text{for $t\in\mathcal{G}_{\texttt{sym}}$}$ Figure 5. (Left) Definition of modified beta reduction and substitution ($\circ$). (Right) Definition of UpshiftAll. We provide a proof of the correctness of LambdaUnify with respect to Eq. 5 in Appendix B, as well as a Coq proof of the correctness in stitch.v in the supplemental material. However, to understand the key idea behind the proof we focus here on an example illustrating the different cases involving the $\downarrow$ operator and $\&i$ indices. Consider what happens when you run (6) $\textsc{LambdaUnify}(\lambda.\ f\ \texttt{??}_{0},\ \lambda.\ e)$ By Eq. 5, the goal is to produce a mapping $[\texttt{??}_{0}\rightarrow u^{\prime}_{0}]$ such than when replacing $\texttt{??}_{0}$ with $u^{\prime}_{0}$ it produces $(\lambda.\ e)$. In other words, $(\lambda.\ \lambda.\ f\ \$1)\ u^{\prime}_{0}=\lambda.\ e$. Now, suppose (7) $\textsc{LambdaUnify}(f~{}\texttt{??}_{0},\ e)\rightsquigarrow[\texttt{??}_{0}\rightarrow\lambda.\ \$i]$ This means that $(\lambda.\ f\ \$0)\ (\lambda.\ \$i)=e$ so $e=f\ (\lambda.\ \$i)$. There are three possibilities for $i$ that need to be considered, corresponding to the 3 cases in the definition of $\downarrow_{d}\$i$. In the first case, when $i=0$ and thus $e=f\ (\lambda.\ \$0)$, it means that $\$i$ in Eq. 7 is bound to the lambda in the return mapping. In this case, DownshiftAll in U-Lam should not do anything because if we can replace $\texttt{??}_{0}$ with $(\lambda.\ \$0)$ in $(f\ \texttt{??}_{0})$ to produce $e$, the same replacement in $(\lambda.\ f\ \texttt{??}_{0})$ will produce $(\lambda.\ e)$. The second case is when $i\geq 2$; this means that $e$ has some variables that were defined outside of it and they have been captured by $\texttt{??}_{0}$. For example, if $i=2$ and thus $e=f\ (\lambda.\ \$2)$, the solution to Eq. 6 is $[\texttt{??}_{0}\rightarrow\lambda.\ \$1]$, since $(\lambda.\ \lambda.\ f\ \$1)(\lambda.\ \$1)=\lambda.\ e$; in other words, when computing $\textsc{LambdaUnify}(\lambda.~{}f~{}\texttt{??}_{0},\ \lambda.\ e)$ from Eq. 7, the $\$i$ in the return mapping has to be downshifted to account for the fact that it will be substituted inside an additional lambda. The third case, when $i=1$, is more problematic. In this case, we have that $e=f\ (\lambda.\ \$1)$. This creates a problem, since there is no $i$ such that $(\lambda.\ \lambda.\ f\ \$1)(\lambda.\ \$i)=\lambda.\ e$; the $\$i$ needs to refer to the lambda surrounding $e$. The downshift operator addresses this through the special treatment of the $\&i$ index. LambdaUnify relates to prior work on unification modulo binders (Huet, 1975; Miller, 1991, 1992; Dowek et al., 1995, 1996) but focuses on a more efficiently solvable syntax-driven subset of the more general unification problems tackled in this prior work. We give a more detailed comparison to this prior work in Section 7. ##### Match Locations The set of match locations of a partial abstraction $A_{\texttt{??}}$ in a corpus $\mathcal{P}$, denoted $\textsc{Matches}(\mathcal{P},A_{\texttt{??}})$, is the maximal set of context-expression pairs $\\{(C_{1},e_{1}),(C_{2},e_{2}),\ \ldots,(C_{k},e_{k})\\}$ such that $\forall k.\ p=C_{k}[e_{k}]$ and $p\in\mathcal{P}$ and $\textsc{LambdaUnify}(e_{k},A_{\texttt{??}})$ succeeds. We discard locations where the mapping produced by LambdaUnify has &i indices in expressions bound by abstraction variables, while we allow them in expressions bound by holes as holes are allowed to violate variable binding rules. In Section 3.2 we explain how we maintain the set of matches incrementally. ##### Rewrite Strategies A corpus $\mathcal{P}$ can be rewritten to use a complete abstraction $A$ as follows. We introduce a new terminal symbol $t_{A}$ into the symbol grammar $\mathcal{G}_{\texttt{sym}}$ to represent the abstraction, and consider it semantically equivalent to $(\lambda\alpha_{0}.\ ...\ \lambda\alpha_{k}.\ A)$. We then re-express $\mathcal{P}$ in terms of $t_{A}$ as follows. We can replace the match location $(C,e)\in\textsc{Matches}(\mathcal{P},A)$ with $t_{A}$ applied with the argument assignments $[\alpha_{i}\to e_{i}]=\textsc{LambdaUnify}(A,e)$ to yield a semantically equivalent expression, $C[(t_{A}\ e_{0}\ ...\ e_{k})]$. Note that since this is a complete abstraction, it contains no holes. One complication is that rewriting at one match location may preclude rewriting at another match location if they overlap with each other. A rewrite strategy $\mathcal{R}$ is a procedure for selecting a subset of $\textsc{Matches}(\mathcal{P},A)$ to rewrite at, in which no match precludes another match. We refer to this set as $\textsc{RewriteLocations}_{\mathcal{R}}(\mathcal{P},A)$. We refer to rewriting a corpus $\mathcal{P}$ under an abstraction $A$ with rewrite strategy $\mathcal{R}$ by $\textsc{Rewrite}_{\mathcal{R}}(\mathcal{P},A)$. We refer to rewriting at only a single particular match location $m$ with abstraction $A$ by $\textsc{RewriteOne}(m,A)$. ##### Utility CTS optimizes a user-defined utility function $U_{\mathcal{P},\mathcal{R}}(A)$ which scores an abstraction $A$ given a corpus $\mathcal{P}$ and rewrite strategy $\mathcal{R}$. There are no strict constraints on the form or properties of the utility function. While in Section 4 we will focus on compressive utility functions, our framework does not generally require this. We say that a rewrite strategy is optimal with respect to a utility function if the rewrite strategy chooses locations to rewrite at such that the utility is maximized. A naïve optimal rewrite strategy exhaustively checks the utility of rewriting with each of the $2^{|\textsc{Matches}(\mathcal{P},A)|}$ possible subsets of $\textsc{Matches}(\mathcal{P},A)$, however depending on the specific utility function computing an optimal or approximately optimal strategy may be significantly more computationally tractable. In Section 4 we detail the rewrite strategy used by Stitch, which is an optimal, linear-time rewrite strategy for compressive utility functions based on dynamic programming. ### 3.2. Algorithm Given a corpus $\mathcal{P}$, rewrite strategy $\mathcal{R}$, and utility function $U_{\mathcal{P},\mathcal{R}}(A)$, the objective of CTS is to find the abstraction $A$ that maximizes the utility $U_{\mathcal{P},\mathcal{R}}(A)$. CTS takes a branch-and-bound approach (Land and Doig, 1960; Morrison et al., 2016) to this problem, as described in this section. ##### Expansion We construct the body of a partial abstraction $A_{\texttt{??}}$ through a series of top-down expansions starting from the trivial partial abstraction ??. Given a partial abstraction $A_{\texttt{??}}$ we can expand a hole in this partial abstraction using any production rule from the partial abstraction grammar $\mathcal{G}_{A_{\texttt{??}}}$, yielding a new abstraction $A_{\texttt{??}}^{\prime}$. We denote this expansion by $A_{\texttt{??}}\to A_{\texttt{??}}^{\prime}$. In Figure 2A our overall top-down search is depicted as a series of expansions. We say that a complete abstraction $A$ can be derived from a partial abstraction $A_{\texttt{??}}$, denoted $A_{\texttt{??}}\to^{*}A$, if $A$ is in the transitive reflexive closure of the expansion operation, i.e. there exists a sequence of expansions from $A_{\texttt{??}}$ to $A$. Any abstraction can be derived from ??. ##### A Naïve Approach The goal of our algorithm is to find the maximum-utility abstraction. One simple, inefficient approach to this is to simply enumerate the entire space of abstractions through top down synthesis and return the one with the highest utility. This naïve approach maintains a queue of partial abstractions initialized with just ??. At each step of the algorithm it pops an abstraction from the queue, chooses a hole in it and expands that hole using each possible production rule. Whenever an expansion produces a new partial abstraction it pushes it to the queue, and when it produces a complete abstraction it calculates its utility and updates the best abstraction found so far. Since an abstraction cannot match a program that is smaller than it, the algorithm can stop expanding when all expansions would lead to abstractions larger than the largest program in the corpus. This algorithm will enumerate all possible abstractions exactly once each. ##### Introducing Strict Dominance Pruning While the naïve approach will find the optimal abstraction, it is extremely inefficient. We can improve on this using an idea core to branch-and-bound algorithms: pruning. We allow the algorithm to choose to prune a partial abstraction instead of expanding it, in which case it simply discards the abstraction and chooses another from the queue. Of course, to maintain optimality we must be certain that we never prune the branch of search containing the optimal abstraction. To formalize safe pruning we will use the notion of strict dominance from branch-and-bound literature (Morrison et al., 2016; Ibaraki, 1977; Chu and Stuckey, 2015), which we explain below in terms of a notion of covering. A complete abstraction $A^{\prime\prime}$ covers another complete abstraction $A^{\prime}$, written $\text{covers}(A^{\prime\prime},A^{\prime})$, if the utility of $A^{\prime\prime}$ is greater than that of $A^{\prime}$. A partial abstraction $A_{\texttt{??}}^{\prime}$ is strictly dominated by another partial abstraction $A_{\texttt{??}}^{\prime\prime}$ if and only if every complete abstraction that is derivable from $A_{\texttt{??}}^{\prime}$ is covered by a complete abstraction that is derivable from $A_{\texttt{??}}^{\prime\prime}$. Formally, $A_{\texttt{??}}^{\prime\prime}$ strictly dominates $A_{\texttt{??}}^{\prime}$ iff $\forall A^{\prime}.\ A_{\texttt{??}}^{\prime}\to^{*}A^{\prime}\implies\exists A^{\prime\prime}.\ A_{\texttt{??}}^{\prime\prime}\to^{*}A^{\prime\prime}\land\text{covers}(A^{\prime\prime},A^{\prime})$. Equipped with this formalism, we claim that it is safe to prune an abstraction $A_{\texttt{??}}^{\prime}$ if we know that there exists some $A_{\texttt{??}}^{\prime\prime}$ which strictly dominates it. Note that $A_{\texttt{??}}^{\prime\prime}$ does not necessarily contain the optimum and may even be pruned in the search if it is strictly dominated by another abstraction. Knowing the existence of $A_{\texttt{??}}^{\prime\prime}$ is enough to enable pruning, regardless of whether it has been pruned or has not yet been enumerated during search. ###### Lemma 1. Naïve search augmented with strict dominance pruning finds the optimal abstraction. ###### Proof. We proceed with a proof by induction. We seek to prove the equivalent statement that the optimum is never pruned and thus it will be enumerated by the search. Our inductive hypothesis is that the optimum has not yet been pruned. In the base case this is trivially true, since no pruning has taken place yet. In the inductive step we must prove that a step of pruning maintains the validity of the inductive hypothesis. Recall that a step of pruning will discard some partial abstraction $A_{\texttt{??}}^{\prime}$ which is strictly dominated by another abstraction $A_{\texttt{??}}^{\prime\prime}$. Suppose then for the sake of contradiction that we did prune the optimum in this step; then the optimum must have been derivable from $A_{\texttt{??}}^{\prime}$. By strict dominance we know that all abstractions derivable from $A_{\texttt{??}}^{\prime}$ must be covered by some abstraction derivable from $A_{\texttt{??}}^{\prime\prime}$. Hence, there must exist some abstraction that covers the optimum. However, since the utility of the optimum is greater than or equal to that of all other abstractions, there is by definition no such abstraction; we have thus arrived at a contradiction. ∎ This lemma ensures the safety of composing together pruning strategies, as long as they all are forms of strict dominance pruning, since no instance of strict dominance pruning will remove the optimum. Determining which partial abstractions are strictly dominated by which others is specific to the utility function being used, and in Section 4 we identify two instances of strict dominance used when instantiating this framework for a compression-based utility. ##### Upper Bound Pruning We further improve this algorithm to employ upper bound based pruning, in which we bound the maximum utility that can be obtained in a branch of search, and discard the branch if we have previously found a complete abstraction with higher utility than this bound. This is the most common form of pruning used in branch-and-bound algorithms (see Morrison et al. 2016 Section 5.1 for a review). Our algorithm is generic over the upper bound function $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})$ which upper bounds the utility of any complete abstraction $A$ that can be derived from $A_{\texttt{??}}$. Formally, the bound must satisfy $\forall A.\ A_{\texttt{??}}\to^{*}A\implies U_{\mathcal{P},\mathcal{R}}(A)\leq\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})$. While this could trivially be satisfied with $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})=\infty$ for any choice of a utility function, a tighter bound will allow for more pruning. The soundness of upper bound pruning in maintaining the optimality of the solution trivially follows from the fact that the pruned branch only contains complete abstractions that are at most as high-utility as the best abstraction found so far. While we defer to Section 4 to construct the upper bound used by Stitch, it is worth mentioning a key insight that helps in constructing tight bounds for many utility functions. When bounding the utility of $A_{\texttt{??}}$, it is useful to bound the set of possible locations where rewrites can occur for any abstraction $A$ derived from $A_{\texttt{??}}$. The following lemma provides such a bound: ###### Lemma 2. $\textsc{Matches}(\mathcal{P},A_{\texttt{??}})$ is an upper bound on the set of locations where rewrites can occur in any $A$ derived from $A_{\texttt{??}}$. This follows from the fact that as partial abstractions are expanded they become more precise and thus match at a subset of the locations. Since rewriting only happens at a subset of match locations, $\textsc{Matches}(\mathcal{P},A_{\texttt{??}})$ serves as a bound on the set of locations where rewrites can occur. An important consequence of Lemma 2 is that if a partial abstraction matches at zero locations, then all abstractions derived from it will have zero rewriting locations, so no rewriting can occur. Such branches can therefore be safely pruned. ##### Improving the Search Order Finally, without sacrificing the optimality of this algorithm we can heuristically guide the order in which the space of abstractions is explored. The queue of partial abstractions used in top-down search can be replaced with a priority queue ordered by the upper bound. This way, the algorithm will first explore more promising, higher-bound branches of search, narrowing in on the optimal abstraction more quickly. In branch-and-bound literature this choice of a search order that uses a sound upper bound is sometimes referred to as best-bound search and makes the algorithm a form of A∗ search (Hart et al., 1968) since the upper bound is an admissible heuristic (Morrison et al., 2016). In preliminary experiments on a subset of benchmarks we found that search order had little effect on the overall runtime of the algorithm, but the best- bound ordering was moderately helpful in more quickly narrowing in on the optimal solution (i.e. useful when running the algorithm with a limited time budget), so this is used for all experiments. ##### Efficient Incremental Matching When expanding an abstraction $A_{\texttt{??}}$ to a new abstraction $A_{\texttt{??}}^{\prime}$, it is easy to compute $\textsc{Matches}(\mathcal{P},A_{\texttt{??}}^{\prime})$ since we know it is a subset of $\textsc{Matches}(\mathcal{P},A_{\texttt{??}})$. Thus, there is no need to perform matching from scratch against every subtree in the corpus to compute the match locations. Instead, we can simply inspect the relevant subtree at each match location of the original abstraction to see which match locations will be preserved by a given expansion. In fact, except when expanding into an abstraction variable $\alpha$, the sets of match locations obtained by different expansions of a hole will be disjoint. When expanding to an abstraction variable $\alpha$, if $\alpha$ is an existing abstraction variable then this is a situation where the same variable is being used in more than one place, as in the square abstraction $(\lambda\alpha.\ \texttt{*}\ \alpha\ \alpha)$. In this case we restrict the match locations to the subset of locations where within the location all instances of $\alpha$ are bound to syntactically identical subtrees. Additionally, if the user provides a maximum arity limit, then an expansion that causes the abstraction to exceed this limit can be discarded. ##### Algorithm Summary and Key Points The CTS algorithm combines all of the aforementioned optimizations in a top- down search algorithm with pruning. In summary, CTS takes as input a corpus $\mathcal{P}$, a rewrite strategy $\mathcal{R}$, a utility function $U_{\mathcal{P},\mathcal{R}}(A)$, and an upper bound function $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})$, and returns the optimal abstraction with respect to the utility function. CTS performs a top-down search over partial abstractions and prunes branches of search when partial abstractions match at zero locations or are eliminated by upper bound pruning or strict dominance pruning. CTS is made further efficient through the aforementioned search-order heuristic and efficient incremental matching. Importantly, none of these optimizations sacrifice the optimality of the abstraction found — all pruning is done soundly, as discussed in the prior sections on upper bound pruning and strict dominance pruning. We also note that CTS is amenable to parallelization without losing optimality. This can be implemented using a shared priority queue that is safely accessed by different worker threads through a locking primitive. Strict dominance pruning trivially remains sound as it doesn’t depend on any global information about the state of the search. Upper bound pruning also remains sound even if workers only occasionally synchronize their best- abstraction-so-far, as this will just mean that they occasionally have weaker upper bounds which are still sound. Since the algorithm maintains a best abstraction so-far, it can also be terminated early, making it an anytime algorithm. Finally, to learn a library of abstractions, CTS can be run repeatedly (much like DreamCoder), adding one abstraction at a time to the library and rewriting the corpus with each abstraction as it is learned before running CTS again. A listing of the full algorithm is provided in Appendix A. ## 4\. Applying Corpus-Guided Top-Down Search to Compression Having presented the general framework and algorithm of corpus-guided top-down search, we now instantiate this framework for optimizing a compression metric. ### 4.1. Utility In compression we seek to minimize some measure of the size, or more generally the cost, of a corpus of programs after rewriting them with a new abstraction. As in prior work (Ellis et al., 2021) we penalize large abstractions by including abstraction size in the utility. The compressive utility function for a corpus $\mathcal{P}$ and abstraction $A$ is given below. (8) $U_{\mathcal{P},\mathcal{R}}(A)\triangleq-\text{cost}(A)+\text{cost}(\mathcal{P})-\text{cost}(\textsc{Rewrite}_{\mathcal{R}}(\mathcal{P},A))$ Here, $\text{cost}(\cdot)$ is a cost function of the following form: (9) $\displaystyle\text{cost}(\lambda.\ e^{\prime})=\text{cost}_{\lambda}+\text{cost}(e^{\prime})$ $\displaystyle\text{cost}(e^{\prime}e^{\prime\prime})=\text{cost}_{\texttt{app}}+\text{cost}(e^{\prime})+\text{cost}(e^{\prime\prime})$ $\displaystyle\text{cost}(\$i)=\text{cost}_{\texttt{\$i}}$ $\displaystyle\text{cost}(t)=\text{cost}_{t}(t),\text{ for }t\in\mathcal{G}_{\texttt{sym}}$ $\displaystyle\text{cost}(\alpha)=\text{cost}_{\alpha}$ where $\text{cost}_{\lambda}$, $\text{cost}_{\texttt{app}}$, $\text{cost}_{\texttt{\$i}}$, and $\text{cost}_{\alpha}$ are non-negative constants, and $\text{cost}_{t}(t)$ is a mapping from grammar primitives to their (non-negative) costs. Finally, we introduce $\text{cost}_{\alpha=0}(\cdot)$ as a version of $\text{cost}(\cdot)$ where $\text{cost}_{\alpha}=0$. We can equivalently construct the utility in Eq. 8 by summing over the compression gained from performing each rewrite separately, as given below. (10) $U_{\mathcal{P},\mathcal{R}}(A)=-\text{cost}(A)+\sum_{e\in\textsc{RewriteLocations}_{\mathcal{R}}(\mathcal{P},A)}{\text{cost}(e)-\text{cost}(\textsc{RewriteOne}(e,A))}$ By reasoning about the way that rewriting transforms a program, this utility can be broken down even further: (11) $U_{\mathcal{P},\mathcal{R}}(A)=-\text{cost}(A)+\sum_{e\in\textsc{RewriteLocations}_{\mathcal{R}}(\mathcal{P},A)}{U_{local}(A,e)}$ (12) $U_{local}(A,e)=\\!\\!\underbrace{\text{cost}_{\alpha=0}(A)\vphantom{\sum_{\alpha\in\text{a}(A)}}}_{\text{abstraction size}}-\underbrace{(\text{cost}_{t}(t_{A})+\text{cost}_{\texttt{app}}\cdot\text{arity}(A))\vphantom{\sum_{\alpha\in\text{a}(A)}}}_{\text{application utility (negative)}}+\underbrace{\sum_{[\alpha\to e^{\prime}]\in\text{args}(A,e)}{\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!\\!(\text{usages}(\alpha)-1)\cdot\text{cost}(e^{\prime})}}_{\text{multiuse utility}}$ where $t_{A}$ is the new primitive corresponding to abstraction $A$. This form of the utility function can be efficiently computed without explicitly performing any rewrites; and it sheds light on the different sources of compression. The three main terms in this expression are (1) the abstraction size that comes from the shared structure that is removed, (2) the negative application utility that comes from introducing the new primitive and lambda calculus app nonterminals to apply it to each argument, and (3) the multiuse utility which comes from only needing to pass in a single copy of an argument that might be use in more than one place in the body. We emphasize that this form of the utility function is equivalent to the original definition based on rewriting given in Eq. 8. ### 4.2. Upper Bounding the Utility We seek an upper bound function $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})$ such that for any $A$ derived from $A_{\texttt{??}}$, $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})\geq U_{\mathcal{P},\mathcal{R}}(A)$. We begin from the decomposition of the utility function given in Eq. 10. Since costs are always non-negative, we can upper bound this by dropping the $-\text{cost}(A)$ term as well as the negative term within the sum: (13) $U_{\mathcal{P},\mathcal{R}}(A)\leq\sum_{e\in\textsc{RewriteLocations}_{\mathcal{R}}(\mathcal{P},A)}{\text{cost}(e)}$ Intuitively, dropping the negative term from the sum is equivalent to assuming we compressed the cost of this location all the way down to cost 0. We can also bound $\textsc{RewriteLocations}_{\mathcal{R}}(\mathcal{P},A)$ as $\textsc{Matches}(\mathcal{P},A_{\texttt{??}})$ using Lemma 2, yielding our final upper bound in terms of $A_{\texttt{??}}$: (14) $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})\triangleq\sum_{e\in\textsc{Matches}(\mathcal{P},A_{\texttt{??}})}{\text{cost}(e)}$ ### 4.3. Strict Dominance Pruning We identify two forms of strict dominance pruning compatible with a compressive utility. The first is redundant argument elimination: a partial abstraction can be dropped if it has two abstraction variables that always take the same argument as each other across all match locations, for example if it had variables $\alpha$ and $\beta$ and $\alpha=\beta=\texttt{(+ 3 5)}$ at one location, $\alpha=\beta=\texttt{2}$ at another location, and so on for all match locations. This abstraction would then be strictly dominated by the abstraction that doesn’t take $\beta$ as an argument and instead reuses $\alpha$ in place of $\beta$, and can hence be eliminated. Thus an abstraction like (+ (* $\alpha$ $\beta$) ??) is strictly dominated by an abstraction like (+ (* $\alpha$ $\alpha$) ??) if $\alpha=\beta$ across all match locations in the former abstraction. The second is argument capture: when a partial abstraction takes the same argument for at least one abstraction variable across all match locations, this abstraction can be discarded. This is because every abstraction derivable is covered by another abstraction which is identical except it has the argument inlined into the body. For example, if in the abstraction (+ (* $\alpha$ $\beta$) ??) the abstraction variable $\alpha$ takes the same argument $\alpha=$(+ 3 5) across every match location, there is a strictly dominating partial abstraction that simply has (+ 3 5) inlined in its body: (+ (* (+ 3 5) $\beta$) ??). Note that argument capture does not apply when the argument contains a free variable, as inlining would result in an invalid abstraction. A close reader might notice that if the number of times $\alpha$ appears in $A$ is greater than the number of rewrite locations, the abstraction with argument capture applied, $A^{\prime}$, will have slightly lower utility. Employing this pruning rule in our search means that we will find the optimal abstraction subject to the constraint that all possible argument captures have occurred. This utility difference from the optimal abstraction without argument capture is bounded by at most $\text{cost}(A^{\prime})$, since the difference comes from the contribution of the size of the abstraction body itself to the utility. 222 Specifically, for some $[\alpha\to e]$, the abstraction without argument capture is higher in utility by $-(\text{cost}(e)+\text{cost}_{\texttt{app}})*|\textsc{RewriteLocations}_{\mathcal{R}}(\mathcal{P},A)|+(\text{cost}(e)-\text{cost}_{\alpha})*\text{usages}(\alpha)$. For a cost function where $\text{cost}_{\texttt{app}}<\text{cost}_{\alpha}$, this is positive when the argument is used more times than the abstraction itself is used. In preliminary experiments across all of our experimental datasets, we never find this edge case to change which abstraction is chosen as optimal. ### 4.4. Rewrite Strategy Stitch employs a linear-time, optimal rewrite strategy for a compression objective. The goal of an optimal rewrite strategy is to efficiently select the optimal subset of $\textsc{Matches}(\mathcal{P},A)$ to perform rewrites at in order to maximize the utility. The main challenge is that when match locations overlap, the strategy must decide which of the two to accept. For example, consider the program (foo (foo (foo bar)) and the abstraction $t_{A}=\lambda\alpha.\ $(foo (foo $\alpha$)). This abstraction matches at the root of the program with $\alpha=$ (foo bar), resulting in the rewritten program ($t_{A}$ (foo bar)). However, though the abstraction also matched at the subtree (foo (foo bar) with $\alpha=$ bar in the original program, this match location is no longer present in the rewritten program, so only one of the two locations can be chosen by the rewrite strategy. Our approach is a bottom-up dynamic programming algorithm, which begins at the leaves of the program and proceeds upwards. At each node $e$, we compute the cumulative utility so far if we reject a rewrite here ($\text{util}_{R}$), if we accept a rewrite here ($\text{util}_{A}$), or if we choose the better of the two options ($\text{util}^{*}$): (15) $\displaystyle\text{util}_{R}[e]=\sum_{e^{\prime}\in children(e)}{\text{util}^{*}[e^{\prime}]}$ $\displaystyle\text{util}_{A}[e]=\left\\{\begin{array}[]{ll}0&\text{if }e\notin\textsc{Matches}(\mathcal{P},A)\\\ U_{local}(A,e)+\sum\limits_{e^{\prime}\in\text{args}(A,e)}{\text{util}^{*}[e^{\prime}]}&\text{otherwise}\\\ \end{array}\right.$ $\displaystyle\text{util}^{*}[e]=\text{max}(\text{util}_{R}[e],\text{util}_{A}[e])$ where $U_{local}(A,e)$ is the utility gained from a single rewrite as defined in Eq. 12. After calculating these quantities the rewrite strategy can start from the program root and recurse down the tree, rewriting at each node where rewriting is optimal (i.e. $\text{util}_{A}>\text{util}_{R}$) and then recursing into its arguments after rewriting. Since all arguments were originally subtrees in the program, these quantities will have been calculated for all of them, with the caveat that their de Bruijn indices may have been shifted. Shifting indices of a subtree does not change whether an abstraction can match there nor does it change the utility gained from using that abstraction, so this simply requires some extra bookkeeping to track. This algorithm is optimal by a simple inductive argument: at each step of the dynamic programming problem, we can assume that we know the cumulative utility of all children and potential arguments at this node, so we can use Eq. 15 to calculate the cumulative utility of either rejecting or accepting the rewrite at this node. ## 5\. Combining corpus-guided top-down search with deductive approaches In complex domains, assembling good libraries may involve more than just finding matching code-templates; sometimes, some initial refactoring is necessary to expose common structure. For example, consider learning the abstraction $\lambda\alpha.\ \texttt{(* 2 }\alpha\texttt{)}$ for doubling integers, given the expressions (* 2 8), (* 7 2), and (right-shift 3 1). This is only possible if the system can use the commutativity of multiplication to rewrite (* 7 2) into (* 2 7), and bitvector properties to rewrite (right-shift 3 1) into (* 2 3); such rewrites are not natively supported by CTS. In this section, we discuss how CTS can be combined with refactoring systems based on deductive rewrites to increase its expressivity further. The core idea is simple: run a rewrite system on the corpus to produce a set of refactorings of the corpus in a version space, then run CTS over the resulting data structure.333Note that while we have previously presented CTS as operating over syntax trees, the core notions of upper bounds and matching that CTS operates on are not restricted to program trees. Intuitively, this will lead to improved performance compared with a purely deductive approach since the cost of rewriting grows exponentially with the number iterations of rewrites that are applied. This is a problem for fully deductive approaches like Ellis et al. 2021 because extracting the abstractions often requires a long chain of rewrites, especially for higher-arity abstractions. However, a small number of rewrites is typically sufficient to expose the underlying commonalities, as it was in the example above; performing only a handful of rewrites and then using CTS to actually extract the library thus avoids the exponential blow-up of computing several rewrites in sequence, while still benefiting from the increase in expressivity afforded by the rewrites. ##### Version Spaces To illustrate this hybrid approach, we combine CTS with _version spaces_ (Lau et al., 2003; Polozov and Gulwani, 2015) representing sets of programs semantically equivalent under the $\beta$-inversion rewrite. Version spaces are represented as terms from a grammar obtained by extending the grammar of expressions with the union ($\uplus$) operator that represents a set of equivalent expressions: $v\Coloneqq\varnothing\mid v^{\prime}\ \uplus\ v^{\prime\prime}\mid\lambda.\ v^{\prime}\mid v^{\prime}\ v^{\prime\prime}\mid\$i\mid t$. We define a denotation operator, $\llbracket{v}\rrbracket$, mapping a version space $v$ to a set of terms: for the union operator, $\llbracket{v\uplus v^{\prime}}\rrbracket=\llbracket{v}\rrbracket\cup\llbracket{v^{\prime}}\rrbracket$; for applications, $\llbracket{v\ v^{\prime}}\rrbracket=\left\\{e\ e^{\prime}\,:\,\forall e,e^{\prime}\in\llbracket{v}\rrbracket\times\llbracket{v^{\prime}}\rrbracket\right\\}$; for lambda abstractions, $\llbracket{\lambda v}\rrbracket=\left\\{\lambda e\ \,:\,\forall e\in\llbracket{v}\rrbracket\right\\}$; for the empty set, $\llbracket{\varnothing}\rrbracket=\varnothing$; and for de Bruijn indices and terminals, $\llbracket{v}\rrbracket=v$. Ellis et al. 2021 gives a procedure called $I\beta(e)$, which take as input an expression $e$ and outputs a version space $v$ inverting one step of $\beta$-reduction: that is, $e^{\prime}\to^{\beta}e$ iff $e^{\prime}\in\llbracket{I\beta(e)}\rrbracket$. *[left=RU-AbsVar] RewriteUnify(α,v) ↝[ α→v ] *[left=RU-Hole] RewriteUnify(??_i,v) ↝[ ??_i →v ] *[left=RU-App] RewriteUnify(A_1,v_1) ↝l_1 RewriteUnify(A_2,v_2) ↝l_2 l = vmerge(l_1,l_2) RewriteUnify((A_1 A_2),(v_1 v_2)) ↝l *[left=RU-Lam] RewriteUnify(A,v) ↝l’ l = DownshiftAll(l’) RewriteUnify((λ. A),(λ. v)) ↝l *[left=RU-Same] RewriteUnify(v,v) ↝[ ] *[left=RU-Union] RewriteUnify(A,v_1) ↝l RewriteUnify(A,v_1⊎v_2) ↝l *[left=VM-Same] [α→v_1]∈l_1 [α→v_2]∈l_2 [α→(v_1∩v_2)]∈vmerge(l_1,l_2) *[left=VM-Diff] [α→v]∈l_1 α/∈args(l_2) [α→v]∈vmerge(l_1,l_2) Figure 6. Defining RewriteUnify, which takes as input an abstraction and a version space of possible refactorings, and yields multiple substitutions corresponding to all the ways that the abstraction’s holes and variables can match with programs encoded by the version space. To run CTS on top of this rewriting system requires a generalization of LambdaUnify: instead of unifying an abstraction with a term, yielding a single substitution, we unify against a _set of terms_ (a version space), yielding a set of candidate substitutions. The relation RewriteUnify (Fig. 6) accomplishes this, and equipped with this relation we can express the size of a subtree $e$ after expanding it into the version space $I\beta(e)$ and rewriting it with abstraction $A$ as: (16) $\text{cost}(\textsc{Rewrite}(A,e))=\min_{l:\;\textsc{RewriteUnify}(A,I\beta(e))\rightsquigarrow l}\text{cost}_{t}(t_{A})+\text{cost}_{\texttt{app}}\cdot\text{arity}(A)+\sum_{v\in\text{args}(l)}\min_{e^{\prime}\in\llbracket{v}\rrbracket}\text{cost}(e^{\prime})$ We then can approximate the utility of rewriting a corpus (based on Eq. 10) as follows (17) $U_{\mathcal{P},\mathcal{R}}(A)\approx-\text{cost}(A)+\sum_{p\in\mathcal{P}}{\;\max_{e\in\text{subtrees}(p)}{\text{cost}(e)-\text{cost}(\textsc{Rewrite}(A,e))}}$ This utility is exact when the optimal way of rewriting the corpus using $A$ has at most one rewrite location per program. This is because it considers the utility of the single best site at which to perform the rewrite instead of considering multiple simultaneous rewrites at different locations within a single program. We can bound this approximate utility for a partial abstraction by computing the approximate utility directly while treating any version space bound to a hole in $\text{args}(l)$ as having zero cost. ##### Example: Learning map from fold. Consider learning the higher-order function map $=\lambda\alpha.$(fold ($\lambda x.\lambda\ell.$ (cons ($\alpha$ $x$) $\ell$))) from two example programs: doubling a list of numbers, expressed as (fold ($\lambda x.\lambda\ell.$ (cons (+ $x$ $x$) $\ell$))), and decrementing a list of numbers, expressed as (fold ($\lambda x.\lambda\ell.$ (cons (- $x$ 1) $\ell$))). Matching these programs with the map function requires re- expressing them as (fold ($\lambda x.\lambda\ell.$ (cons (($\lambda z.$ (+ z z)) $x$) $\ell$))) and (fold ($\lambda x.\lambda\ell.$ (cons (($\lambda z.$ (- z 1)) $x$) $\ell$))), which matches the map function with the substitutions $\alpha=\texttt{($\lambda z.$ (+ z z))}$ and $\alpha=\texttt{($\lambda z.$ (- z 1))}$ respectively. These two re-expressions are performed by running a single round of $\beta$-inversion as a deductive rewriting step. While the majority of our experiments will focus on evaluating the approach laid out in Section 3 and Section 4, we implement and evaluate a prototype of integrating version spaces in this way in Section 6.5. ## 6\. Experiments In this section, we evaluate corpus-guided top-down search for library learning. Specifically, our evaluation focuses on five hypotheses about the performance of Stitch: 1. (1) Stitch learns libraries of comparable quality to those found by existing deductive library learning algorithms in prior work, while requiring significantly less resources. In Section 6.1 we run Stitch on the library learning tasks from (Ellis et al., 2021) and directly compare Stitch to DreamCoder, the deductive algorithm introduced in that work. We find that Stitch learns libraries which usually match or exceed the baseline in quality (measured via a compression metric), while improving the resource efficiency in terms of memory usage and runtime by 2 and 3-4 orders of magnitude compared to the baseline (respectively). 2. (2) Stitch scales to corpora of programs that contain more and longer programs than would be tractable with prior work. In Section 6.2, we evaluate Stitch’s ability to learn libraries within eight graphics domains from Wong et al. 2022, which are considerably larger and more complex than have been considered in previous work. We find that Stitch on average obtains a test set compression ratio of 2.55x-11.57x in 0.19s-60.16s, with a peak memory usage of 11.21MB-714.55MB in these domains. The problems are large enough to time out with the DreamCoder baseline. 3. (3) Stitch degrades gracefully when resource-constrained. In Section 6.3 we investigate Stitch’s performance when run as an anytime algorithm; i.e., one that can be terminated early for a best-effort result if a corpus is too large or there are limits on time or memory. We reuse the eight graphics domains from Wong et al. 2022 and find that with its heuristic guidance Stitch converges upon a set of high quality abstractions very early in search, doing so within 1% of the total search time in 3 out of 8 domains and within 10% in all except one. 4. (4) All the elements of Stitch matter. In Section 6.4, we carry out an ablation study on Stitch and find the argument capturing and upper bound pruning methods are essential to its performance, while redundant argument elimination also proves useful in certain domains. With all optimizations disabled, we find that Stitch cannot run in $\leq 90$ minutes and $\leq 50$GB of RAM on any of the domains from Wong et al. 2022. 5. (5) Stitch is complementary to deductive rewrite-based approaches to library learning. These prior experiments show the superior runtime performance of Stitch relative to deductive rewrite systems, but deductive systems have an important advantage over Stitch: the ability to incorporate arbitrary rewrite rules to expose more commonality among different programs and in that way discover better libraries. Such deductive approaches are especially more apt at learning higher-order abstractions. In Section 6.5, we give evidence that this expressivity gap can be reduced by running Stitch on top of a deductive rewrite system, allowing it to learn many new abstractions while still using $<$2% of the compute time. For all experiments, we parameterize Stitch’s $\text{cost}(e)$ function (as defined in Section 4.1) as follows: $\text{cost}_{\texttt{app}}=\text{cost}_{\lambda}=1$, $\text{cost}_{\texttt{\$i}}=\text{cost}_{\alpha}=\text{cost}_{t}(t)=100$. To avoid overfitting, DreamCoder prunes the abstractions that are only useful in programs from a single task. We add this to Stitch as well, treating each program as a separate task for datasets that don’t divide programs into tasks. We run all experiments on a machine with two AMD EPYC 7302 processors, 64 CPUs, and 256GB of RAM. We note however that Stitch itself runs exceptionally well on a more average machine. For example, on one author’s laptop (ThinkPad X1 Carbon Gen 8), the experiments from Section 6.2 can be replicated with nearly identical runtimes, and these are the most computationally intensive experiments outside of the ablation study. ### 6.1. Iterative Bootstrapped Library Learning #### Experimental setup Our first experiment is designed to replicate the experiments in DreamCoder, which is the state-of-the-art in deductive library learning. DreamCoder learns libraries iteratively: the system is initialized with a low-level DSL, and then alternates between synthesizing programs (via a neurally-guided enumerative search) that solve a training corpus of inductive tasks and updating the library of abstractions available to the synthesizer. Traces from the experiments carried out by Ellis et al. 2021 are publicly available444https://github.com/mlb2251/compression_benchmark and include all of the intermediate programs that were synthesized as well as the libraries learned from those programs. In this experiment, we take these traces and evaluate Stitch on each instance where library learning was performed, comparing the quality of the resulting library to the original one found by DreamCoder. We also re-run DreamCoder on these same benchmarks in order to evaluate its resource usage, capturing its runtime and memory usage in the same environment as Stitch. The library learning algorithm in Ellis et al. 2021 implements a stopping criterion to determine how many abstractions to retain on any given set of training programs. In our comparative experiments, we run the DreamCoder baseline first, and then match the number of abstractions learned by Stitch at each iteration to those learned by the baseline under its stopping criterion so that timing comparisons are fair. We record the total time spent both performing abstraction learning and rewriting for both Stitch and DreamCoder. We replicate experiments on five distinct domains from Ellis et al. 2021: * • Lists: A functional programming domain consisting of 108 total inductive tasks. * • Text: A string editing domain in the style of FlashFill (Gulwani et al., 2015) consisting of 128 total inductive tasks. * • LOGO: A graphics domain consisting of 80 total inductive tasks. * • Towers: A block-tower construction domain consisting of 56 total inductive tasks. * • Physics: A domain for learning equations corresponding to physical laws from observations of simulated data, consisting of 60 total inductive tasks. #### Assessing library quality with a compression metric The standard Stitch configuration optimizes a compression metric that minimizes the size of the programs after being rewritten to use the abstraction. This is a standard metric in program synthesis, since shorter programs are frequently easier to synthesize. Optimizing against this metric is equivalent to maximizing the likelihood of the rewritten programs under a uniform PCFG. The DreamCoder synthesizer is more sophisticated than simple enumeration; it takes as input a learned typed bigram PCFG and leverages it to synthesize programs more efficiently. When performing compression, it optimizes against this given PCFG in order to find abstractions that will be more profitable for its specific synthesizer. For the purpose of this evaluation, however, we restrict ourselves to the uniform PCFG because the one used by DreamCoder requires programs to be in a particular normal form. Another aspect of DreamCoder relevant to its compression metric is that DreamCoder synthesizes multiple programs that solve the same task and then selects the abstraction that works best on _some_ program for a given task. This is expressed formally in the equation below. (18) $\text{cost}(A)+\sum\limits_{\text{task}}\min\limits_{p\in\text{task}}\text{cost}\left(\textsc{Rewrite}(p,A)\right)$ It is trivial to implement this best-of-task metric in Stitch, so we use this for the comparisons with DreamCoder. Figure 7. Compression rates obtained when running Stitch on the five domains considered in 6.1 relative to those of DreamCoder. Higher is better for Stitch: a ratio above 1.0 indicates that Stitch achieves greater compression than DreamCoder. The specific compression metric used in the ratio is given in Eq. 18. Figure 8. Peak memory usage of Stitch and DreamCoder while running on the five domains considered in 6.1, averaged over all benchmarks. Lower is better; black lines indicate $\pm$ one standard deviation. Note the logarithmic y-axis. Figure 9. Wall-clock time required to find (and rewrite under) one abstraction in each of the five domains from 6.1, averaged over all benchmarks. Lower is better; black lines indicate $\pm$ one standard deviation. Note the logarithmic y-axis. #### Results We compare DreamCoder and Stitch for library quality and resource efficiency. Library quality. We first examine the quality of the libraries learned by both Stitch and DreamCoder using the compression metric in Eq. 18: Fig. 7 shows the ratio between them across all of the benchmarks for each domain. A ratio of 1.0 indicates that programs are the exact same length under the libraries learned by DreamCoder and Stitch, a ratio greater than 1.0 indicates that Stitch learns more compressive libraries, and a ratio less than 1.0 indicates that Stitch learns libraries which are less compressive. For example, a ratio of 1.1 indicates that DreamCoder rewrote to produce a corpus 10% larger than that of Stitch, so it achieved less compression. These results show that Stitch generally learns libraries of comparable and often greater quality than DreamCoder when matching the number of abstractions learned by the latter. In the logo, towers, text, and physics domains, Stitch always finds abstractions that are of equal—and often considerably greater—compressive quality than DreamCoder does; in the list domain, Stitch more often than not still obtains better compression, but sometimes loses out to DreamCoder. This is a result of the fact that Stitch cannot learn higher- order abstractions, which are useful in this domain; although we emphasize our focus is on scalability, we will later present an extension of Stitch capable of handling most higher-order functions in Section 6.5, based on the formalism developed in Section 5. Nonetheless, we conclude that Stitch learns libraries whose quality is comparable to and often better than those found by DreamCoder. Resource efficiency. In addition to the quality of the libraries found, we are interested in how the two methods compare in terms of time and space requirements. Since we a priori believe Stitch to be significantly faster, for this evaluation we allow DreamCoder to use 8 CPUs but limit Stitch to single threading555While this may seem unfair to Stitch, it is worth noting that it would be unlikely to benefit from multithreading when running on the order of milliseconds anyway; DreamCoder, on the other hand, would struggle greatly in these domains without the aid of parallelism.. The results are shown in Fig. 9 and Fig. 9; in summary, we find that Stitch takes tens of milliseconds to discover abstractions across all five domains—achieving a 3-4 order of magnitude speed-up over DreamCoder—while also requiring more than 2 orders of magnitude less memory. We thus conclude that Stitch is dramatically more efficient than the state-of-the-art deductive baseline when replicating the iterative library learning experiments of Ellis et al. 2021. ### 6.2. Large-Scale Corpus Library Learning #### Experimental setup. While the previous experiment allowed us to benchmark Stitch directly against a state-of-the-art deductive baseline, the iterated learning setting considered by Ellis et al. 2021 only evaluates library learning on relatively small corpora of short programs discovered by the synthesizer. Our second experiment is instead designed to evaluate Stitch in a more traditional learning setting, in which we aim to learn libraries of abstractions from a large corpus of existing programs all at once. We source our larger-scale program datasets from Wong et al. 2022, which present a series of datasets designed as a benchmark for comparing human-level abstraction learning and graphics program writing against automated synthesis and library learning models. These datasets are divided into two distinct high-level domains (technical drawings and block-tower planning tasks), each consisting of four distinct subdomains containing 250 programs: * • Technical drawing domains: nuts and bolts; vehicles; gadgets; furniture: CAD- like graphics programs that render technical drawings of common objects, written in an initial DSL consisting of looped transformations (scaling, translation, rotation) over simple geometric curves (lines and arcs). * • Tower construction domains: bridges; cities; houses; castles: Planning programs that construct complex architectures by placing blocks, written in an initial DSL that moves a virtual hand over a canvas and places horizontal and vertical bricks. We choose these datasets not only for their size and scale (full dataset statistics in Table 1), but also for the complexity of their potential abstractions: Wong et al. 2022 explicitly design their corpora to contain complex hierarchical structures throughout the programs, making them an interesting setting for library learning. Domain | #Programs | Average program length | Average program depth ---|---|---|--- nuts & bolts | 250 | $76.03\pm 24.22$ | $15.18\pm 2.13$ gadgets | 250 | $142.85\pm 87.32$ | $20.88\pm 2.37$ furniture | 250 | $171.74\pm 48.41$ | $31.83\pm 5.33$ vehicles | 250 | $141.70\pm 40.23$ | $21.22\pm 1.35$ bridges | 250 | $137.03\pm 59.71$ | $92.35\pm 39.80$ cities | 250 | $161.70\pm 55.56$ | $109.80\pm 37.66$ castles | 250 | $189.09\pm 60.18$ | $128.27\pm 40.77$ houses | 250 | $168.13\pm 55.75$ | $114.85\pm 37.60$ Table 1. Summary statistics about the domains from Wong et al. 2022. Program length is the number of terminal symbols in the program; program depth is the length of the longest path from root to leaf in the program tree. Both are reported as the mean over the entire dataset $\pm$ one standard deviation. When performing library learning in a synthesis setting by compressing a corpus of solutions, it’s desirable to find abstractions that would be useful for solving new tasks, as opposed to abstractions that overfit to the existing solutions. To evaluate how well the abstractions we learn apply to heldout programs in the domain, we split the corpora into train and test sets, running Stitch on the train set and evaluating its compression on the test set. Since Wong et al. 2022 do not present a train/test split of their datasets, we use stochastic cross validation to evaluate the generalization of the libraries found by Stitch. For each domain, we randomly sample 80% of the dataset to train on and reserve the last 20% as a held out test set; we repeat this procedure 50 times. We then ask Stitch to learn a library consisting of $\leq 10$ abstractions with a maximum arity of 3, and average the results across the different random seeds. To obtain a baseline to compare its performance against, we once again turned to DreamCoder (Ellis et al., 2021). However, we found that DreamCoder was unable to discover even a single abstraction when run directly on any of the datasets from Wong et al. 2022, despite being given hours of runtime and 256GB of RAM. We also experimented with heavily sub-sampling the training dataset before passing it to DreamCoder, but failed to find a configuration under which DreamCoder finds any interesting abstractions at all due to the fact that it immediately blows up on programs as long as these. As a result, we resort to presenting Stitch’s performance metrics without any baseline to compare against; we stress that this is a direct result of the fact that Stitch is the first library learning tool that scales to such a challenging setting. #### Results Domain | Compression Ratio | Runtime (s) | Peak mem. usage (MB) ---|---|---|--- Training set | Test set nuts & bolts | $12.00\pm 0.25$ | $11.57\pm 0.49$ | $0.24\pm 0.04$ | $11.10\pm 0.17$ gadgets | $4.03\pm 0.15$ | $3.91\pm 0.31$ | $1.83\pm 0.35$ | $30.11\pm 0.85$ furniture | $4.95\pm 0.07$ | $4.85\pm 0.26$ | $2.83\pm 0.49$ | $30.88\pm 0.62$ vehicles | $4.28\pm 0.14$ | $4.14\pm 0.25$ | $1.52\pm 0.29$ | $26.98\pm 0.88$ bridges | $4.36\pm 0.06$ | $3.78\pm 0.14$ | $17.58\pm 1.26$ | $189.08\pm 10.10$ cities | $3.15\pm 0.05$ | $3.06\pm 0.14$ | $50.74\pm 3.96$ | $413.63\pm 9.51$ castles | $2.57\pm 0.07$ | $2.55\pm 0.08$ | $77.26\pm 6.98$ | $683.89\pm 32.96$ houses | $8.92\pm 0.21$ | $8.85\pm 0.57$ | $15.54\pm 1.37$ | $241.17\pm 2.86$ Table 2. Results for the large-scale library learning experiment in Sec. 6.2. The compression ratio refers to how many times smaller the corpus is after rewriting under the learned library compared to the original corpus; higher is thus better. All results are given as the mean $\pm$ one standard deviation over 50 runs with different random seeds for the dataset splitting. The results are summarized in Table 2. We find that Stitch scales up to even the most complex sub-domains, running in 77 seconds with a peak memory usage below 1GB on castles. On four out of eight of the domains, Stitch finishes in single-digit seconds and consumes only tens of megabytes. This stands in stark contrast to DreamCoder, which we were unable to run on the very simple nuts & bolts domain even with 256GB RAM and several hours worth of compute budget. These results thus support our claim that Stitch scales to corpora of programs that would be intractable with prior library learning approaches. We hope that by providing our results on these datasets in full, future work in this field will benefit from having a directly comparable baseline. ### 6.3. Robustness to Early Search Termination #### Experimental setup This experiment is designed to evaluate how early into the search procedure Stitch finds what will eventually prove to be the optimal abstraction. This is highly relevant in settings where the set of training programs is too large to run the search to completion. The experiment showcases one of Stitch’s more subtle strengths: corpus-guided top-down abstraction search is an _anytime_ algorithm, and thus does not need to be run to completion to give useful results. We re-use the domains from Wong et al. 2022 and once again evaluate the quality of the library learned (measured in program compression), similarly to what was done in the previous experiment. However, since we are interested in how quickly Stitch finds _what it perceives to be_ the optimal abstraction, we measure compressivity of the training dataset itself (rather than a held-out test set) and capture the compression ratio obtained by each _candidate_ abstraction found during search (rather than just the compression ratio obtained when search has been run to completion). Thus, we are able to investigate how early on during the search procedure Stitch converges on a chosen library. We restrict Stitch to learning a single abstraction with a maximum arity of 3. #### Results Figure 10. Reduction in size obtained on the training set when rewritten under the best abstraction found thus far vs. the number of nodes expanded during search. The y-axis is normalized with respect to the optimal abstraction; the x-axis is normalized with respect to the total number of nodes explored by Stitch. Lines thus end earlier the quicker Stitch finds the optimal abstraction; however in all runs Stitch continues to search until reaching 100% on the x-axis, exploring the rest of the abstraction space without finding a new best abstraction. Note the logarithmic x-axis. The results are shown in Fig. 10. These results validate our hypothesis that Stitch is empirically robust to terminating the search procedure early: in every sub-domain except for _nuts & bolts_ Stitch converges to the optimal abstraction very early on, having only completed a tiny fraction of the total search.666Given that significant attention has already been given to wall- clock run-times of Stitch on similar workloads in 6.2, we here use the number of nodes explored instead of wall-clock time to ensure deterministic and easily reproducible results. We believe that this has great importance for Stitch’s applicability in data-rich settings since it suggests that a nearly- optimal abstraction can be found even if the search must be terminated early (e.g. after a fixed amount of time has passed), making early stopping an empirically useful way of speeding up the library learning process. ### 6.4. Ablation Study #### Experimental setup Stitch implements several different optimizations, which we have argued hasten the search for abstractions. To verify that this claim holds in practice, we now carry out a brief ablation study. Since the space of every possible combination of optimizations is too large to present succinctly, we focus our attention on four ablations: * • no-arg-capture (from Section 4.3), which disables the pruning of abstractions which are only ever used with the exact same set of arguments (and these arguments could therefore just be in-lined for greater compression). * • no-upper-bound, which disables the upper bound based pruning. * • no-redundant-args (from Section 4.3), which disables the pruning of multi- argument abstractions that have a redundant argument that could be removed because it is always the same as another argument. * • no-opts, which disables all of Stitch’s optimizations. To isolate the impact of disabling optimizations, we run a single iteration of abstraction learning on each of the 8 domains from 6.2 and collect the number of nodes explored during the search. The first iteration is generally the most challenging as the corpus is large and has not yet been compressed at all. Focusing on the number of nodes explored (rather than for example runtime) allows for deterministic results. We also fix the maximum arity of abstractions to 3 in all runs, aiming to strike a good balance between compute requirements and how much the optimization will be exposed. We limit each run to 50GB of virtual address space, as well as 90 minutes of compute. System Domain | bridges | castles | cities | gadgets | furniture | houses | nuts & bolts | vehicles ---|---|---|---|---|---|---|---|--- no-arg-capture | RAM | RAM | RAM | 16714.28 | RAM | RAM | 30.21 | 26.67 no-upper-bound | 12.15 | 23.71 | 27.90 | 207.76 | 119.00 | 36.18 | 179.12 | 179.99 no-redundant-args | 1.95 | 1.42 | 1.37 | 1.01 | 1.00 | 1.01 | 1.09 | 1.00 no-opt | RAM | RAM | RAM | | RAM | RAM | | RAM Table 3. Results from the ablation study. Each cell contains the ratio between the number of nodes explored during search by that particular system and the number of nodes explored by the baseline on the same domain. Lower is better; 1.00 means performance is identical to the baseline. Cells labeled RAM crashed due to reaching the 50GB virtual address space limit, while those labeled reached the 90 minute time restriction. #### Results The results are shown in Table 3. We first note as a sanity check that each ablation does indeed lead to reduced performance in general (i.e. explores a search space larger than the baseline does). Furthermore, the results suggest that upper bound based pruning is the most important in the nuts & bolts and vehicles domains, while pruning out argument capture abstractions is the most important in the other six domains; it is noteworthy that this latter ablation by itself causes Stitch to reach the memory limit on more than half of the domains. On the other hand, disabling the pruning of redundant arguments has a relatively modest impact on the size of the search space, but still leads to an almost 2x improvement in the bridge domain. Perhaps the most important takeaway from this ablation study is that when all optimizations are disabled, Stitch fails to find an abstraction within the resource budget on any of the domains. This verifies our hypothesis that corpus-guided pruning of the search space is the key factor involved in making top-down synthesis of abstractions tractable. ### 6.5. Learning Libraries of Higher-Order Functions #### Experimental setup Our experiments up till now have focused on performant and scalable library learning. This comes at the expense of some expressivity: deductive rewrite systems can, in principle, express broader spaces of refactorings. For example, a rewrite based on inverting $\beta$-reduction allows inventing auxiliary $\lambda$-abstractions, which helps with learning higher-order functions: in Ellis et al. 2021, DreamCoder is shown to recover higher-order functions such as map, fold, unfold, filter, and zip_with, starting from the Y-combinator. This works by constructing a version space which encodes every refactoring that is equivalent up to $\beta$-inversion rewrites. But DreamCoder’s coverage comes at a steep cost as inverting $\beta$-reduction is expensive. In this experiment, we seek to give evidence that it is possible to make Stitch recover all of these higher-order functions by layering it on top of the version space obtained after a single step of DreamCoder’s $\beta$-inversion, following the formalism outlined in Section 5. We then compare this modified version of Stitch with DreamCoder on its ability to learn these higher-order functions from programs generated by intermediate DreamCoder iterations, and measure the runtime of each approach. #### Probabilistic Re-ranking While the approach outlined in Section 5 should suffice to layer Stitch on top of deductive rewrite systems based on version spaces, some extra care needs to be taken to combine it with DreamCoder. This is because DreamCoder implements a probabilistic Bayesian objective for judging candidate abstractions (exploiting the connection between compression and probability (Shannon, 1948)), seeking the library $L$ which maximizes $P(L)\prod_{p\in\mathcal{P}}P(p|L)$ for a given or learned prior $P(L)$ and program likelihood $P(p|L)$. Stitch, on the other hand, effectively judges compression quality via a cost function capturing the (weighted) size of the programs as detailed in Section 4. To implement this probabilistic heuristic in Stitch, we simply run Stitch on the version spaces as-is but then re-score each complete abstraction popped off of the priority queue under the Bayesian objective, using DreamCoder’s models of the prior and likelihood. To make this integration easier, we re- implement Stitch in Python, giving a prototype version called pyStitch which is only used for this experiment. Our implementation employs strict dominance pruning, and prunes using the bound on the approximate utility given in Section 5. System | Fold | Unfold | Map | Filter | ZipWith | Time (s) ---|---|---|---|---|---|--- | Base | ✓ | $\times$ | ✓ | $\times$ | ✓ | 503 | +Bayes | ✓ | $\times$ | ✓ | $\times$ | ✓ | 817 | +VS | ✓ | ✓ | ✓ | $\times$ | ✓ | 3231 pyStitch | +Bayes+VS | ✓ | ✓ | ✓ | ✓ | ✓ | 3042 DreamCoder | Step 1 | $\times$ | $\times$ | ✓ | $\times$ | ✓ | 67 Step 2 | $\times$ | $\times$ | ✓ | ✓ | $\times$ | 116 Step 3 | $\times$ | $\times$ | ✓ | ✓ | ✓ | 2254 Step 4 | ✓ | ✓ | ✓ | ✓ | ✓ | 228048 Table 4. Comparing library learners on functional programming exercises. DreamCoder, $n$-steps: deductive baseline rewriting $n$ steps of $\beta$-reduction. pyStitch: Python reimplementation of Stitch, which enables better interoperability with DreamCoder’s version space algebra (+VS) and probabilistic models (+Bayes). pyStitch+VS+Bayes learns all of the same higher-order functions, using $<2\%$ of the compute. #### Results The results are shown in Table 4. We note first the large discrepancy in runtimes; running DreamCoder with 4 steps of rewriting takes roughly 2.5 days of compute on this domain, while even the slowest version of pyStitch still finishes in less than an hour. In terms of the number of higher-order abstractions found, DreamCoder only finds all five when run in its most expensive configuration; reducing the computation cost quickly decreases its expressivity. For pyStitch, the Bayesian re-ranking alone (pyStitch+Bayes) does not yield any improvements, while running it on the version spaces (pyStitch+VS) yields 4 out of 5 functions. However, it is only when these two adaptations are used in conjunction (pyStitch+Bayes+VS) that the Stitch-based method is able to find all of the higher-order functions. In summary, we find that running CTS after a single step of version space rewriting and then probabilistically re-ranking the results suffices to recover the core higher-order functions that DreamCoder learns, while using $<$2% of its compute. We thus conclude that running Stitch on top of a deductive rewrite system reduces the expressivity gap, while retaining superior performance. ## 7\. Related Work Stitch is related to two core ideas from prior work: deductive refactoring and library learning systems, which introduce the idea of learning abstractions that capture common structure across a set of programs, but have largely been driven by deductive algorithms; and guided top-down program synthesis systems, which use cost functions to guide top-down enumerative search over a space of programs, but have largely been used to synthesize whole programs for individual tasks in prior work. ### 7.1. Deductive Refactoring and Library Learning Recent work shares Stitch’s goal of learning libraries of program abstractions which capture reusable structure across a corpus of programs (Ellis et al., 2018, 2021; Dechter et al., 2013; Cropper, 2019; Shin et al., 2019; Allamanis and Sutton, 2014; Iyer et al., 2019; Wong et al., 2021; Jones et al., 2021). Several of these prior approaches also introduce a utility metric based on program compression in order to determine the most useful candidate abstractions to retain (Dechter et al., 2013; Ellis et al., 2021; Wong et al., 2021; Lázaro-Gredilla et al., 2019; Iyer et al., 2019). Much of this this prior work follows a bottom-up approach to abstraction learning, combining a bottom-up traversal across individual training programs with a second stage to extract shared abstractions from across the training corpus. This approach includes systems that work through direct memoization of subtrees across a corpus of programs (Dechter et al., 2013; Lin et al., 2014; Lázaro-Gredilla et al., 2019); antiunification (caching tree templates that can be unified with training program syntax trees) (Ellis et al., 2018; Henderson, 2013; Hwang et al., 2011; Iyer et al., 2019); or by more sophisticated refactoring using one or more rewrite rules to expose additional shared structure across training programs (Ellis et al., 2018, 2021; Chlipala et al., 2017; Liang et al., 2010). Many of the bottom up algorithms draw more generally on deductive synthesis approaches that apply local rewrite rules in a bottom-up fashion to program trees in order to refactor them — historically, to synthesize programs from a declarative specification of desired function (Burstall and Darlington, 1977; Manna and Waldinger, 1980). Deductive approaches to library learning, however, confront fundamental memory and search-time scaling challenges as the corpus size and depth of the training programs increases; prior deductive approaches such as (Ellis et al., 2018, 2021) use version spaces (Lau et al., 2003; Mitchell, 1977) to mitigate the memory usage during bottom-up abstraction proposal. Still, deductive approaches are challenging to bound and prune (unlike the top-down approach we take in Stitch), as they generally traverse individual program trees locally and must store possible abstraction candidates in memory before the extraction step. Some prior work (Shin et al., 2019; Allamanis and Sutton, 2014) takes an MCMC- based approach with better scaling behavior than deductive rewrites, but differs from the goals of Stitch as they don’t address binders and focus on common syntactic fragments instead of well-formed functions. ### 7.2. Guided Top-Down Program Synthesis Stitch uses a corpus-guided top-down approach to learning library abstractions that is closely related to recent guided enumerative synthesis techniques. This includes methods that leverage type-based constraints on holes (Feser et al., 2015; Polikarpova et al., 2016), over- and under-approximations of the behaviors of holes (Lee et al., 2016; Chen et al., 2020), and probabilistic techniques to heuristically guide the search (Balog et al., 2016; Ellis et al., 2020, 2021; Nye et al., 2021; Shah et al., 2020). These approaches have largely been applied in to synthesize entire programs based on input/output examples or another form of specification, in contrast to the abstraction- learning goal in our work (Allamanis et al., 2018; Balog et al., 2016; Chen et al., 2018; Ellis et al., 2018; Ganin et al., 2018; Koukoutos et al., 2017). Like Stitch however, these approaches sometimes use cost functions (such as the likelihood of a partially enumerated program under a hand-crafted or learned probabilistic generative model over programs) in order to direct search towards more desirable program trees. However, Stitch’s cost function leverages a more direct relationship between partially-enumerated candidate functions and the existing training corpus, unlike the cost functions typically applied in inductive synthesis, which must be estimated from input/output examples. ### 7.3. Lambda-Aware Unification The LambdaUnify procedure presented in Section 3 relates to prior work on unification modulo binders (Huet, 1975; Miller, 1991, 1992; Dowek et al., 1995, 1996). Notion of beta-equivalence. This prior work is concerned with more general notions of equivalence modulo beta-reduction, while LambdaUnify is based on a restricted but fast syntax-driven equivalence. For example, in Dowek et al. 1996 one might try to unify $(\alpha\ \texttt{foo})$ with $(\texttt{foo})$ and get the two solutions: $[\alpha\to(\lambda.\ \$0)]$ and $[\alpha\to(\lambda.\ \texttt{foo})]$. In contrast, LambdaUnify will never introduce a $\lambda-$abstraction to create a higher order argument and thus $(\alpha\ \texttt{foo})$ and $(\texttt{foo})$ would not unify at all, since one is an application while the other is a primitive. Instead, higher order abstraction arguments are handled by combining the algorithm with a deductive approach when desirable as in Section 5. We also note that in $\textsc{LambdaUnify}(A,e)$ it is assumed that $e$ is an expression and thus does not contain abstraction variables, which further simplifies the approach compared to this prior work. Handling of binders. Dowek et al. 1995 and Dowek et al. 1996 both employ de Bruijn indexed variables and therefore must similarly account for the shifting of variables in arguments when inverting beta reduction. While LambdaUnify handles this through DownshiftAll, Dowek et al. 1995 and Dowek et al. 1996 both converting their terms to the $\lambda\sigma-$calculus of explicit substitutions (Abadi et al., 1989) which allows them to insert upshifting operators at each abstraction variable location. This alternate handling of shifting is useful given the more general notion of equivalence they are considering, but is excessive for our simpler syntax-guided task. Handling of holes. In $\textsc{LambdaUnify}(A,e)$ we allow $A$ to contain holes ?? which are allowed to violate index shifting rules during unification as they are considered unfinished subtrees of the abstraction. This is handled through the use of $\&i$ indices. While the special-case handling of holes in expressions is not directly part of this prior work, there are similarities between it and the Skolemization (Skolem, 1920) done in Miller 1992. Skolemization allows for lifting an existential quantifier (i.e. an abstraction variable or hole) above a universal quantifier (i.e. a lambda) by turning the abstraction variable or hole into a function of its local context — in essence piping the local context into the hole. In the context of the lambda calculus this is essentially a form of lambda lifting (Johnsson, 1985), which is the process of lifting a local function that contains free variables by binding the free variables as additional arguments to the function and passing them in at each call site. In Miller 1992 this is used for abstraction variables (as there are no holes) and aids in the more general notion of beta- equivalence they are dealing with, while for our purposes the $\&i$ variables suffice and don’t require extra manipulations of lambdas to thread in the local context. ### 7.4. Upper Bounds in Network Motifs There is also an interesting connection between this work and prior work on finding network motifs (Milo et al., 2002), which are frequently-occurring subgraphs within a corpus of graphs. One approach to mining high-frequency motifs is based on growing a motif one edge at a time, much like we grow abstractions one node at a time during our search (Schreiber and Schwöbbermeyer, 2005; Kuramochi and Karypis, 2004, 2001). This approach makes use of the insight that the frequency of a motif will only decrease as it is grown, giving an upper bound on the frequency of any motif derived from it. While we employ a more complex compressive utility function than their frequency-based utility, our upper bound is based on a similar insight that larger abstractions will match at a subset of the locations. ### 7.5. Comparison to babble babble (Cao et al., 2023) is concurrent work in library learning that likewise adopts the compression objective from DreamCoder. While babble focuses primarily on expressive library learning through an algorithm that can reason over semantic equivalences represented through e-graphs, Stitch focuses primarily on efficient library learning through a parallel anytime branch-and- bound algorithm. Stitch synthesizes the maximally-compressive abstraction for a corpus of programs through a branch-and-bound top-down search, computing an upper bound on the compression of any partially-constructed abstraction to guide and prune the search. In contrast, babble operates on an e-graph representing many equivalent corpora and uses anti-unification to propose candidate abstractions and a custom e-graph beam-search extraction algorithm to select a maximally- compressive set of abstractions from the proposal. At a high level, Stitch aims to be efficient while babble aims to be expressive. The key differences between the algorithms are: * • Efficiency: On DreamCoder’s benchmarks, babble is 10-100x faster than DreamCoder while Stitch is 1,000-10,000x faster than DreamCoder. Unlike babble, Stitch is an anytime algorithm that can be stopped early for a best- so-far result. * • Expressivity: babble can learn libraries modulo equational theories, allowing it to find common semantic abstractions despite syntactic differences in the programs. Stitch only explores non-syntactic equivalences preliminarily in the PyStitch prototype with the DreamCoder beta-inversion rewrite rule (Section 5 and Section 6.5). * • Jointly learning a library: babble can learn multiple abstractions at once that jointly provide compression, while DreamCoder and Stitch repeatedly learn a single abstraction at a time. Stitch provides an optimality guarantee on each abstraction learned while babble approximates the joint objective through a beam search. These two approaches has advantages and disadvantages, and we believe that there is strong potential in combining the two. ## 8\. Conclusion and Future Work We have presented corpus-guided top-down synthesis (CTS)—an efficient new algorithm for synthesizing libraries of functional abstractions capturing common functionality within a corpus of programs. CTS directly synthesizes the abstractions, rather than exposing them through a series of rewrites as is done by deductive systems. Key to its performance is the usage of a guiding utility function, which allows CTS to effectively search (and prune out large portions of) the space of possible abstractions. We implement this algorithm in Stitch, an open-source library learning tool with an associated Python library and Rust crate.777https://github.com/mlb2251/stitch We evaluate Stitch across five experimental settings, demonstrating that it learns comparably compressive libraries with 2 orders of magnitude less memory and 3-4 orders of magnitude less time compared to the state-of-the-art deductive algorithm of Ellis et al. 2021. We also find that Stitch scales to learning libraries of abstractions from much larger datasets of deeper program trees than is possible with prior work, and that the anytime property of corpus-guided top-down search — abstractions discovered via top-down search are already compressive early in search and improve as it continues — offers the opportunity for high-quality library learning even in complex domains through early search termination. There remain many open problems and exciting directions in abstraction learning. One major direction is extending the CTS approach to handle reasoning over more general deductive rewrite systems, like those encoded in egg (Willsey et al., 2021) as babble does (Cao et al., 2023) (also see Section 7.5). The PyStitch prototype presented in Section 5 only works with a single rewrite rule (beta-inversion) and uses an approximate utility without an optimality guarantee. This is because it is challenging to adapt the upper bound used by Stitch to version spaces or e-graphs due to the trade-off between the compression gained from equational rewrites and compression gained from abstraction learning. The speedup afforded by Stitch may allow for new applications of abstraction learning as well, as it can now be used as an inexpensive subroutine that can be called much more frequently than prior algorithms. To further extend its range of applications, CTS could also be adapted to finding abstractions in data structures beyond program trees, such as dataflow DAGs and more general graph structures like molecules, if the bounds and matching can be adapted. Finally, while we work with fairly large lambda calculus programs in this work, there is a clear gap in scale between these programs and those in an actual codebase, so exploring applications of CTS to real-world code is an exciting direction. ## Data Availability Statement All source code can be found at https://github.com/mlb2251/stitch. An artifact for reproducing results in this work is available at https://github.com/mlb2251/stitch-artifact or alternatively a static version is available at https://zenodo.org/record/7151663 (Bowers et al., 2022). ###### Acknowledgements. We thank A. Lew and J. Andreas for helpful discussions, and J. Feser and I. Kuraj for feedback on the manuscript. M.B. and G.G are supported by the Sponsor National Science Foundation (NSF) Graduate Research Fellowship under Grant No. Grant #2141064. M.B. is also supported by the Sponsor Defense Advanced Research Projects Agency (DARPA) under the SDCPS Contract Grant #FA8750-20-C-0542. T.X.O. is supported by Herbert E. Grier (1933) and Dorothy J. Grier Fund Fellowship. L.W. and J.B.T. are supported by Sponsor AFOSR under grant number Grant #FA9550-19-1-0269, the MIT Quest for Intelligence, the MIT-IBM Watson AI Lab, ONR Science of AI, and DARPA Machine Common Sense. G.G. is also supported by the MIT Presidential Fellowship. A.S. is supported by the NSF under Grant No. Grant #1918839. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of sponsors. ## References * (1) * Abadi et al. (1989) Martin Abadi, Luca Cardelli, P-L Curien, and J-J Lévy. 1989. Explicit substitutions. In _Proceedings of the 17th ACM SIGPLAN-SIGACT symposium on Principles of programming languages_. 31–46. * Allamanis et al. (2018) Miltiadis Allamanis, Earl T Barr, Premkumar Devanbu, and Charles Sutton. 2018. A survey of machine learning for big code and naturalness. _ACM Computing Surveys (CSUR)_ 51, 4 (2018), 1–37. * Allamanis and Sutton (2014) Miltiadis Allamanis and Charles Sutton. 2014. Mining idioms from source code. In _Proceedings of the 22nd acm sigsoft international symposium on foundations of software engineering_. 472–483. * Balog et al. (2016) Matej Balog, Alexander L Gaunt, Marc Brockschmidt, Sebastian Nowozin, and Daniel Tarlow. 2016\. Deepcoder: Learning to write programs. _arXiv preprint arXiv:1611.01989_ (2016). * Bowers et al. (2022) Matthew Bowers, Olausson, Wong, Grand, Tenenbaum, Ellis, and Solar-Lezama. 2022\. _Artifact for ”Top-Down Synthesis For Library Learning”_. https://doi.org/10.5281/zenodo.7151663 * Burstall and Darlington (1977) Rod M Burstall and John Darlington. 1977. A transformation system for developing recursive programs. _Journal of the ACM (JACM)_ 24, 1 (1977), 44–67. * Cao et al. (2023) David Cao, Rose Kunkel, Chandrakana Nandi, Max Willsey, Zachary Tatlock, and Nadia Polikarpova. 2023. babble: Learning Better Abstractions with E-Graphs and Anti-Unification. _Proc. ACM Program. Lang._ POPL (2023). https://doi.org/10.1145/3571207 * Chen et al. (2020) Qiaochu Chen, Xinyu Wang, Xi Ye, Greg Durrett, and Isil Dillig. 2020. Multi-modal synthesis of regular expressions. In _Proceedings of the 41st ACM SIGPLAN International Conference on Programming Language Design and Implementation, PLDI 2020, London, UK, June 15-20, 2020_ , Alastair F. Donaldson and Emina Torlak (Eds.). ACM, 487–502. https://doi.org/10.1145/3385412.3385988 * Chen et al. (2018) Xinyun Chen, Chang Liu, and Dawn Song. 2018. Execution-guided neural program synthesis. In _International Conference on Learning Representations_. * Chlipala et al. (2017) Adam Chlipala, Benjamin Delaware, Samuel Duchovni, Jason Gross, Clément Pit-Claudel, Sorawit Suriyakarn, Peng Wang, and Katherine Ye. 2017. The end of history? Using a proof assistant to replace language design with library design. In _2nd Summit on Advances in Programming Languages (SNAPL 2017)_. Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik. * Chu and Stuckey (2015) Geoffrey Chu and Peter J. Stuckey. 2015. Dominance breaking constraints. _Constraints An Int. J._ 20, 2 (2015), 155–182. https://doi.org/10.1007/s10601-014-9173-7 * Cropper (2019) Andrew Cropper. 2019\. Playgol: Learning programs through play. _arXiv preprint arXiv:1904.08993_ (2019). * de Bruijn (1972) Nicolaas Govert de Bruijn. 1972\. Lambda calculus notation with nameless dummies, a tool for automatic formula manipulation, with application to the Church-Rosser theorem. In _Indagationes Mathematicae (Proceedings)_ , Vol. 75. Elsevier, 381–392. * Dechter et al. (2013) Eyal Dechter, Jonathan Malmaud, Ryan Prescott Adams, and Joshua B Tenenbaum. 2013. Bootstrap learning via modular concept discovery. In _Proceedings of the International Joint Conference on Artificial Intelligence_. AAAI Press/International Joint Conferences on Artificial Intelligence. * Dowek et al. (1995) Gilles Dowek, Thérèse Hardin, and Claude Kirchner. 1995. Higher-Order Unification via Explicit Substitutions. In _Proceedings of the Tenth Annual Symposium on Logic in Computer Science_ , D. Kozen (Ed.). IEEE Computer Society Press, San Diego, California, 366–374. * Dowek et al. (1996) Gilles Dowek, Thérèse Hardin, Claude Kirchner, and Frank Pfenning. 1996. Unification via Explicit Substitutions: The Case of Higher-Order Patterns. In _Proceedings of the Joint International Conference and Symposium on Logic Programming_ , M. Maher (Ed.). MIT Press, Bonn, Germany, 259–273. * Ellis et al. (2020) Kevin Ellis, Catherine Wong, Maxwell Nye, Mathias Sable-Meyer, Luc Cary, Lucas Morales, Luke Hewitt, Armando Solar-Lezama, and Joshua B Tenenbaum. 2020. Dreamcoder: Growing generalizable, interpretable knowledge with wake-sleep bayesian program learning. _arXiv preprint arXiv:2006.08381_ (2020). * Ellis et al. (2021) Kevin Ellis, Catherine Wong, Maxwell Nye, Mathias Sablé-Meyer, Lucas Morales, Luke Hewitt, Luc Cary, Armando Solar-Lezama, and Joshua B Tenenbaum. 2021. Dreamcoder: Bootstrapping inductive program synthesis with wake-sleep library learning. In _Proceedings of the 42nd ACM SIGPLAN International Conference on Programming Language Design and Implementation_. 835–850. * Ellis et al. (2018) Kevin M Ellis, Lucas E Morales, Mathias Sablé-Meyer, Armando Solar Lezama, and Joshua B Tenenbaum. 2018\. Library learning for neurally-guided bayesian program induction. (2018). * Felleisen and Hieb (1992) Matthias Felleisen and Robert Hieb. 1992. The Revised Report on the Syntactic Theories of Sequential Control and State. _Theor. Comput. Sci._ 103, 2 (sep 1992), 235–271. https://doi.org/10.1016/0304-3975(92)90014-7 * Feser et al. (2015) John K Feser, Swarat Chaudhuri, and Isil Dillig. 2015\. Synthesizing data structure transformations from input-output examples. _ACM SIGPLAN Notices_ 50, 6 (2015), 229–239. * Ganin et al. (2018) Yaroslav Ganin, Tejas Kulkarni, Igor Babuschkin, SM Ali Eslami, and Oriol Vinyals. 2018\. Synthesizing programs for images using reinforced adversarial learning. In _International Conference on Machine Learning_. PMLR, 1666–1675. * Gulwani et al. (2015) Sumit Gulwani, José Hernández-Orallo, Emanuel Kitzelmann, Stephen H Muggleton, Ute Schmid, and Benjamin Zorn. 2015. Inductive programming meets the real world. _Commun. ACM_ 58, 11 (2015), 90–99. * Hart et al. (1968) Peter E. Hart, Nils J. Nilsson, and Bertram Raphael. 1968\. A Formal Basis for the Heuristic Determination of Minimum Cost Paths. _IEEE Trans. Syst. Sci. Cybern._ 4, 2 (1968), 100–107. https://doi.org/10.1109/TSSC.1968.300136 * Henderson (2013) Robert John Henderson. 2013\. Cumulative learning in the lambda calculus. (2013). * Huet (1975) Gérard Huet. 1975\. A Unification Algorithm for Typed $\lambda$-Calculus. _Theoretical Computer Science_ 1 (1975), 27–57. * Hwang et al. (2011) Irvin Hwang, Andreas Stuhlmüller, and Noah D Goodman. 2011\. Inducing probabilistic programs by Bayesian program merging. _arXiv preprint arXiv:1110.5667_ (2011). * Ibaraki (1977) Toshihide Ibaraki. 1977\. The Power of Dominance Relations in Branch-and-Bound Algorithms. _J. ACM_ 24, 2 (1977), 264–279. https://doi.org/10.1145/322003.322010 * Iyer et al. (2019) Srinivasan Iyer, Alvin Cheung, and Luke Zettlemoyer. 2019\. Learning programmatic idioms for scalable semantic parsing. _arXiv preprint arXiv:1904.09086_ (2019). * Johnsson (1985) Thomas Johnsson. 1985\. Lambda Lifting: Treansforming Programs to Recursive Equations. In _Functional Programming Languages and Computer Architecture, FPCA 1985, Nancy, France, September 16-19, 1985, Proceedings_ _(Lecture Notes in Computer Science, Vol. 201)_ , Jean-Pierre Jouannaud (Ed.). Springer, 190–203. https://doi.org/10.1007/3-540-15975-4_37 * Jones et al. (2021) R Kenny Jones, David Charatan, Paul Guerrero, Niloy J Mitra, and Daniel Ritchie. 2021. ShapeMOD: macro operation discovery for 3D shape programs. _ACM Transactions on Graphics (TOG)_ 40, 4 (2021), 1–16. * Koukoutos et al. (2017) Manos Koukoutos, Mukund Raghothaman, Etienne Kneuss, and Viktor Kuncak. 2017. On repair with probabilistic attribute grammars. _arXiv preprint arXiv:1707.04148_ (2017). * Kuramochi and Karypis (2001) Michihiro Kuramochi and George Karypis. 2001. Frequent Subgraph Discovery. In _Proceedings of the 2001 IEEE International Conference on Data Mining, 29 November - 2 December 2001, San Jose, California, USA_ , Nick Cercone, Tsau Young Lin, and Xindong Wu (Eds.). IEEE Computer Society, 313–320. https://doi.org/10.1109/ICDM.2001.989534 * Kuramochi and Karypis (2004) Michihiro Kuramochi and George Karypis. 2004. Finding Frequent Patterns in a Large Sparse Graph. In _Proceedings of the Fourth SIAM International Conference on Data Mining, Lake Buena Vista, Florida, USA, April 22-24, 2004_ , Michael W. Berry, Umeshwar Dayal, Chandrika Kamath, and David B. Skillicorn (Eds.). SIAM, 345–356. https://doi.org/10.1137/1.9781611972740.32 * Land and Doig (1960) A. H. Land and A. G. Doig. 1960. An Automatic Method of Solving Discrete Programming Problems. _Econometrica_ 28, 3 (1960), 497–520. http://www.jstor.org/stable/1910129 * Lau et al. (2003) Tessa Lau, Steven A Wolfman, Pedro Domingos, and Daniel S Weld. 2003. Programming by demonstration using version space algebra. _Machine Learning_ 53, 1 (2003), 111–156. * Lázaro-Gredilla et al. (2019) Miguel Lázaro-Gredilla, Dianhuan Lin, J Swaroop Guntupalli, and Dileep George. 2019. Beyond imitation: Zero-shot task transfer on robots by learning concepts as cognitive programs. _Science Robotics_ 4, 26 (2019), eaav3150. * Lee et al. (2016) Mina Lee, Sunbeom So, and Hakjoo Oh. 2016. Synthesizing regular expressions from examples for introductory automata assignments. In _Proceedings of the 2016 ACM SIGPLAN International Conference on Generative Programming: Concepts and Experiences, GPCE 2016, Amsterdam, The Netherlands, October 31 - November 1, 2016_ , Bernd Fischer and Ina Schaefer (Eds.). ACM, 70–80. https://doi.org/10.1145/2993236.2993244 * Liang et al. (2010) Percy Liang, Michael I Jordan, and Dan Klein. 2010. Learning programs: A hierarchical Bayesian approach. In _Proceedings of the 27th International Conference on Machine Learning (ICML-10)_. 639–646. * Lin et al. (2014) Dianhuan Lin, Eyal Dechter, Kevin Ellis, Joshua B Tenenbaum, and Stephen H Muggleton. 2014\. Bias reformulation for one-shot function induction. (2014). * Manna and Waldinger (1980) Zohar Manna and Richard Waldinger. 1980. A deductive approach to program synthesis. _ACM Transactions on Programming Languages and Systems (TOPLAS)_ 2, 1 (1980), 90–121. * Miller (1991) Dale Miller. 1991\. A Logic Programming Language with Lambda-Abstraction, Function Variables, and Simple Unification. _Journal of Logic and Computation_ 1, 4 (1991), 497–536. * Miller (1992) Dale Miller. 1992\. Unification under a Mixed Prefix. _Journal of Symbolic Computation_ 14 (1992), 321–358. * Milo et al. (2002) R. Milo, S. Shen-Orr, S. Itzkovitz, N. Kashtan, D. Chklovskii, and U. Alon. 2002\. Network Motifs: Simple Building Blocks of Complex Networks. _Science_ 298, 5594 (2002), 824–827. https://doi.org/10.1126/science.298.5594.824 arXiv:https://www.science.org/doi/pdf/10.1126/science.298.5594.824 * Mitchell (1977) Tom M Mitchell. 1977\. Version spaces: A candidate elimination approach to rule learning. In _Proceedings of the 5th international joint conference on Artificial intelligence-Volume 1_. 305–310. * Morrison et al. (2016) David R. Morrison, Sheldon H. Jacobson, Jason J. Sauppe, and Edward C. Sewell. 2016. Branch-and-bound algorithms: A survey of recent advances in searching, branching, and pruning. _Discret. Optim._ 19 (2016), 79–102. https://doi.org/10.1016/j.disopt.2016.01.005 * Nye et al. (2021) Maxwell Nye, Yewen Pu, Matthew Bowers, Jacob Andreas, Joshua B Tenenbaum, and Armando Solar-Lezama. 2021. Representing Partial Programs with Blended Abstract Semantics. In _International Conference on Learning Representations_. * Polikarpova et al. (2016) Nadia Polikarpova, Ivan Kuraj, and Armando Solar-Lezama. 2016\. Program synthesis from polymorphic refinement types. _ACM SIGPLAN Notices_ 51, 6 (2016), 522–538. * Polozov and Gulwani (2015) Oleksandr Polozov and Sumit Gulwani. 2015. Flashmeta: A framework for inductive program synthesis. In _Proceedings of the 2015 ACM SIGPLAN International Conference on Object-Oriented Programming, Systems, Languages, and Applications_. 107–126. * Schreiber and Schwöbbermeyer (2005) Falk Schreiber and Henning Schwöbbermeyer. 2005\. Frequency Concepts and Pattern Detection for the Analysis of Motifs in Networks. _Trans. Comp. Sys. Biology_ 3 (2005), 89–104. https://doi.org/10.1007/11599128_7 * Shah et al. (2020) Ameesh Shah, Eric Zhan, Jennifer Sun, Abhinav Verma, Yisong Yue, and Swarat Chaudhuri. 2020\. Learning Differentiable Programs with Admissible Neural Heuristics. In _Advances in Neural Information Processing Systems_ , H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 4940–4952. https://proceedings.neurips.cc/paper/2020/file/342285bb2a8cadef22f667eeb6a63732-Paper.pdf * Shannon (1948) Claude Elwood Shannon. 1948\. A mathematical theory of communication. _The Bell system technical journal_ 27, 3 (1948), 379–423. * Shin et al. (2019) Eui Chul Shin, Miltiadis Allamanis, Marc Brockschmidt, and Alex Polozov. 2019. Program synthesis and semantic parsing with learned code idioms. _Advances in Neural Information Processing Systems_ 32 (2019). * Skolem (1920) Thoralf Skolem. 1920\. Logisch-kombinatorische Untersuchungen über die Erfüllbarkeit oder Bewiesbarkeit mathematischer Sätze nebst einem Theorem über dichte Mengen. (1920). * Willsey et al. (2021) Max Willsey, Chandrakana Nandi, Yisu Remy Wang, Oliver Flatt, Zachary Tatlock, and Pavel Panchekha. 2021\. Egg: Fast and Extensible Equality Saturation. _Proc. ACM Program. Lang._ 5, POPL, Article 23 (jan 2021), 29 pages. https://doi.org/10.1145/3434304 * Wong et al. (2021) Catherine Wong, Kevin M Ellis, Joshua Tenenbaum, and Jacob Andreas. 2021. Leveraging language to learn program abstractions and search heuristics. In _International Conference on Machine Learning_. PMLR, 11193–11204. * Wong et al. (2022) Catherine Wong, William McCarthy, Gabriel Grand, Jacob Andreas, Joshua B Tenenbaum, Robert Hawkins, and Judy Fan. 2022. Identifying concept libraries from language about object structure. In _CogSci_. To appear. ## Appendix A Corpus-guided top-down search: full algorithm In this appendix, we give an expansive description of the full corpus-guided top-down search (CTS) algorithm to ease re-implementation. The full algorithm is given in Algorithm 1. It takes as input a corpus of programs $\mathcal{P}$, a utility function $U_{\mathcal{P},\mathcal{R}}(A)$, and utility upper bound function $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})$. The algorithm searches for and returns the complete abstraction $A_{best}$ that maximizes the utility function. The algorithm maintains a priority queue $Q$ of partial abstractions ordered by their utility upper bound (or alternatively a custom priority function), a best complete abstraction so-far $A_{best}$, and a corresponding best utility so far $U_{best}$ for that abstraction. The priority queue is initialized to hold the single-hole partial abstraction ??, from which any abstraction can be derived. $A_{best}$ and $U_{best}$ are initialized as the best arity-zero abstraction and corresponding utility (line 4), since arity- zero abstractions can be quickly and completely enumerated (each unique, closed subexpression in the corpus is an arity-zero abstraction). We then proceed to the core loop of the algorithm from lines 6-31. At each step of this loop we pop the highest-priority partial abstraction $A_{\texttt{??}}$ off of the priority queue and process it. We discard $A_{\texttt{??}}$ if its utility upper bound doesn’t exceed our best utility found so far (lines 8-10). We then choose a hole $h$ in $A_{\texttt{??}}$ to expand with the procedure Choose-Hole. Choose-Hole can be a custom function; we find from preliminary experiments on a subset of the datasets that choosing the most recently introduced hole is effective in practice. The algorithm then uses the procedure Expansions to iterate over all possible single step expansions of the hole $h$ in $A_{\texttt{??}}$, such as replacing the hole with + or with (app ?? ??). For each expanded abstraction $A_{\texttt{??}}^{\prime}$, it is easy to compute its set of match locations $\textsc{Matches}(\mathcal{P},A_{\texttt{??}}^{\prime})$ since we know this will be a subset of the match locations of $A_{\texttt{??}}$ (and these will be disjoint subsets, except when expanding into an abstraction variable $\alpha$). We can easily inspect the relevant subtree at each match location of the original abstraction to see which expansions are valid and which match locations will be preserved by a given expansion. Note that an expansion to a free variable can be pruned immediately, as any resulting abstraction will not be well-formed. When expanding to an abstraction variable $\alpha$, if $\alpha$ is a new variable not already present in the partial abstraction, the set of match locations is unchanged. If $\alpha$ is an existing abstraction variable then this is a situation where the same variable is being used in more than one place, as in the square abstraction $(\lambda\alpha.\ \texttt{*}\ \alpha\ \alpha)$. In this case we restrict the match locations to the subset of locations where within a location all instances of $\alpha$ match against the same subtree. Additionally, if a maximum arity is provided then an expansion that causes the abstraction to exceed this limit is not considered. When considering each possible expanded abstraction $A_{\texttt{??}}^{\prime}$, there is room for strong corpus-guidance. First, we don’t need to consider any expansions that would result in zero match locations since all abstractions in this branch of search will have zero rewrite locations per Lemma 2 (lines 13-15). Furthermore, we use $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}}^{\prime})$ to upper bound the utility achievable in this branch of search and discard it if it is less than our best utility so far (lines 16-18). Since each $A_{\texttt{??}}^{\prime}$ covers a disjoint set of match locations (except in the case of expanding into an abstraction variable), the set of match locations often drops rapidly and can allow for the calculation of a tight upper bound depending on the utility function. As a final step of pruning, if we can identify that $A_{\texttt{??}}^{\prime}$ is strictly dominated by some other abstraction $A_{\texttt{??}}^{\prime\prime}$, we may discard $A_{\texttt{??}}^{\prime}$ (lines 19-21). Note additionally that arity-zero abstractions were precomputed, then abstractions that match at a single location are safe to prune as well (as long as they don’t have free variables) as arity-zero abstractions are always superior for single match locations. For any partial abstraction $A_{\texttt{??}}^{\prime}$ that has not been pruned, we then check whether it is a complete abstraction (there are no remaining holes) or whether it is still a partial abstraction. Partial abstractions get pushed to the priority queue ranked by their utility upper bound, and complete abstractions are used to update $A_{best}$ and $U_{best}$ if they have a higher utility than any prior complete abstractions. Once there are no more partial abstractions remaining in the priority queue, the algorithm terminates. Note that since the algorithm maintains a best abstraction so-far, it can also be terminated early, making it an any-time algorithm. We also note that this algorithm is amenable to parallelization as it can be easily parallelized over the while loop on lines 6-31, especially since the algorithm remains sound even when the upper bound $U_{best}$ used pruning is not always up to date. This algorithm can be iterated to build up a library of abstractions. Finally note that in our implementation, we employ structural hashing to simplify equality checks between subtrees and avoid re-doing work on multiple identical copies of a subtree. Our complete implementation of CTS for compression is provided at https://github.com/mlb2251/stitch. Algorithm 1 Corpus-guided top-down abstraction synthesis. Color-coded: Upper bound pruning, Zero-usage pruning, Strict dominance pruning 1:Corpus of input programs $\mathcal{P}$, utility function $U_{\mathcal{P},\mathcal{R}}(A)$, and utility upper bound function $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})$ 2:The maximally compressive abstraction $A_{best}$ 3:$Q\leftarrow$ Priority-Queue { ?? } $\triangleright$ New priority queue with the single partial abstraction ?? 4:$A_{best}\leftarrow\textsc{Best-Arity-Zero-Abstraction}(\mathcal{P})$ $\triangleright$ Initialize best abstraction so far 5:$U_{best}\leftarrow U_{\mathcal{P},\mathcal{R}}(A_{best})$ 6:while $\text{non-empty}(Q)$ do 7: $A_{\texttt{??}}\leftarrow\text{pop-max}(Q)$ $\triangleright$ Next partial abstraction to expand 8: if $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}})\leq U_{best}$ then 9: continue $\triangleright$ Upper bound pruning 10: end if 11: $h\leftarrow\textsc{Choose-Hole}(A_{\texttt{??}})$ $\triangleright$ Choose a hole to expand 12: for $(A_{\texttt{??}}^{\prime},M^{\prime})\in\textsc{Expansions}(A_{\texttt{??}},h,\mathcal{P})$ do $\triangleright$ abstraction $A_{\texttt{??}}^{\prime}$ and match locations $M^{\prime}$ 13: if $\text{length}(M^{\prime})==0$ then 14: continue $\triangleright$ No match locations in corpus 15: end if 16: if $\bar{U}_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}}^{\prime})\leq U_{best}$ then 17: continue $\triangleright$ Upper bound pruning 18: end if 19: if $\text{strictly-dominated}(A_{\texttt{??}}^{\prime},\mathcal{P})$ then 20: continue $\triangleright$ Strict dominance pruning 21: end if 22: if $\text{has-holes}(A_{e})$ then 23: $Q\leftarrow Q\cup A_{\texttt{??}}^{\prime}$ $\triangleright$ add partial abstraction to heap 24: else 25: if $U_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}}^{\prime})>U_{best}$ then 26: $U_{best}\leftarrow U_{\mathcal{P},\mathcal{R}}(A_{\texttt{??}}^{\prime})$ $\triangleright$ new best complete abstraction 27: $A_{best}\leftarrow A_{\texttt{??}}^{\prime}$ 28: end if 29: end if 30: end for 31:end while ## Appendix B proof of Correctness of LambdaUnify As introduced in Section 3, $\textsc{LambdaUnify}(A,e)$ returns a mapping from abstraction variables and holes to expressions $[\alpha_{i}\to e_{i}^{\prime},\ \ldots,\ \texttt{??}_{j}\to u_{j}^{\prime},\ \ldots]$ such that (19) $(\lambda\alpha_{i}.\ \ldots\ \lambda\texttt{??}_{j}.\ \ldots A)\ e_{i}^{\prime}\ \ldots\ u_{j}^{\prime}\ \ldots=e$ through beta reduction. The LambdaUnify procedure is given in Fig. 4 and the definition of beta reduction with a modified upshifting operator is given in Fig. 5. We can write Eq. 19 in terms of substitution as (20) $[\alpha_{i}\to e_{i}^{\prime},\ \ldots,\ \texttt{??}_{j}\to u_{j}^{\prime},\ \ldots]\ \circ A=e$ Which can in turn be written in terms of LambdaUnify as (21) $\textsc{LambdaUnify}(A,e)\circ A=e$ To prove the correctness of LambdaUnify, we must show that Eq. 21 holds for all $A$ and $e$. We do this in two ways: a written proof (below) of our correctness theorem (Theorem B.1), and a machine-verifiable proof in Coq. The Coq proof is given in stitch.v in the supplementary material and was proven with CoqIDE 8.15.2. The meanings of the definitions and lemmas in the Coq proof are commented. The Coq proof differs from the written proof in places because encoding the proof in Coq required coming up with some representations that worked well for Coq, though these differences are limited. For example, the Coq proof requires fully formalizing the behavior of Merge and proofs related to it which is more simply explained in the written proof. When lemmas in the written proof below correspond directly to lemmas in the Coq proof, the corresponding Coq lemma is indicated in the written proof below. ### B.1. Well-formedness We formalize a notion of well-formedness to capture the set of expressions that are accessible starting from an expression with no $\&i$ variables then applying any series of upshifts and downshifts. We define $\text{WF}_{d}(e)$, the well-formedness of an expression $e$ at a depth $d$, as follows: $\displaystyle\text{WF}_{d}\ \lambda.b=\text{WF}_{d+1}\ b$ $\displaystyle\text{WF}_{d}\ (fx)=\text{WF}_{d}\ f\land\text{WF}_{d}\ x$ $\displaystyle\text{WF}_{d}\ \$i=\top$ $\displaystyle\text{WF}_{d}\ \&i=\begin{cases}\top,&\text{if }i<d\\\ \bot,&\text{if }i\geq d\\\ \end{cases}$ $\displaystyle\text{WF}_{d}\ t=\top,\ \text{for $t\in\mathcal{G}_{\texttt{sym}}$}$ Note that trivially any expression $e$ without any $\&i$ variables is well formed, as $\&i$ variables are the only source of $\bot$ in this definition. If every expression in a mapping $l$ is well-formed, we say that the mapping itself is well-formed, written $(\text{WFMap}\ l)$. ### B.2. Relevant Lemmas The following two lemmas show that upshifting and downshifting preserve well- formedness. In Coq these lemmas are upshift_wf and downshift_wf. ###### Lemma 3. $\forall e.\ \forall d.\ \text{WF}_{d}\ e\implies\text{WF}_{d}\ \uparrow_{d}e$ ###### Proof. We proceed by induction on e. Case $e=\lambda.\ b$. Our goal is $\text{WF}_{d}\ \uparrow_{d}\lambda.b$ We can move $\uparrow_{d}$ and $\text{WF}_{d}$ into the lambda by unfolding their definitions to get $\text{WF}_{d+1}\ \uparrow_{d+1}b$ We also have the premise $\text{WF}_{d}\ \lambda.b$, where we can likewise move the $\text{WF}_{d}$ into the lambda to get $\text{WF}_{d+1}\ b$. The inductive hypothesis $\forall d.\ \text{WF}_{d}\ b\implies\text{WF}_{d}\ \uparrow_{d}b$ can be instantiated with $d$ as $d+1$ to get $\text{WF}_{d+1}\ b\implies\text{WF}_{d+1}\ \uparrow_{d+1}b$, the conclusion of which is precisely our goal and the premise of which is precisely our other premise. Case $e=f\ x$. Our goal is $\text{WF}_{d}\ \uparrow_{d}(f\ x)$ We can move $\uparrow_{d}$ and $\text{WF}_{d}$ into the application by unfolding their definitions to get $(\text{WF}_{d}\ \uparrow_{d}f)\ \land(\text{WF}_{d}\ \uparrow_{d}x)$ We can use our inductive hypotheses $\text{WF}_{d}\ f\implies\text{WF}_{d}\ \uparrow_{d}f$ to prove the left side of this conjunction if we can prove $\text{WF}_{d}\ f$. We have the premise that $\text{WF}_{d}\ (f\ x)$ from this overall induction case, which by definition of $\text{WF}_{d}$ can only be true if $\text{WF}_{d}\ f\land\text{WF}_{d}\ d$. Thus we can prove the left side of the conjunction, and the proof of the right side is identical with $f$ replaced by $x$ using the other inductive hypothesis. Case $e=\$i$. Our goal is $\text{WF}_{d}\ \uparrow_{d}\$i$ We will handle this in 2 cases: * • Case $i<d$. $\text{WF}_{d}\ \uparrow_{d}\$i$ $=\text{WF}_{d}\ \$i$ This is true because $i indices are always well-formed. * • Case $i\geq d$. $\text{WF}_{d}\ \uparrow_{d}\$i$ $=\text{WF}_{d}\ \$(i+1)$ This is true because $i indices are always well-formed. Case $e=\&i$. Our goal is $\text{WF}_{d}\ \uparrow_{d}\&i$ We will handle our goal in 2 cases: * • Case $i+1\neq d$. $\text{WF}_{d}\ \uparrow_{d}\$i$ $=\text{WF}_{d}\ \&(i+1)$ We know from our premise that $\text{WF}_{d}\ \&i$ which means $i<d$, so $i+1\leq d$. Combining this with our knowledge from the case branch that $i+1\neq d$, we know that $i+1<d$ which means that $\&(i+1)$ is well formed. * • Case $i+1=d$. $\text{WF}_{d}\ \uparrow_{d}\&i$ $=\text{WF}_{d}\ \$(i+1)$ This is true because $i indices are always well-formed. Case $e=t$. Our goal is $\text{WF}_{d}\ \uparrow_{d}t$ which simplifies by definition of $\uparrow_{d}$ to $\text{WF}_{d}\ t$ which is true by definition of $\text{WF}_{d}$. ∎ ###### Lemma 4. $\forall e.\ \forall d.\ \text{WF}_{d}\ e\implies\text{WF}_{d}\ \downarrow_{d}e$ ###### Proof. We proceed by induction on e. The cases $e=t$, $e=(f\ x)$ and $e=\lambda.b$ are identical to those in Lemma 3 but with $\uparrow$ replaced with $\downarrow$ so they will be omitted. Case $e=\$i$. Our goal is $\text{WF}_{d}\ \downarrow_{d}\$i$ We will handle this in 3 cases: * • Case $i<d$. $\text{WF}_{d}\ \downarrow_{d}\$i$ $=\text{WF}_{d}\ \$i$ This is true because $i indices are always well-formed. * • Case $i>d$. $\text{WF}_{d}\ \downarrow_{d}\$i$ $=\text{WF}_{d}\ \$(i-1)$ This is true because $i indices are always well-formed. * • Case $i=d$. $\text{WF}_{d}\ \downarrow_{d}\$i$ $=\text{WF}_{d}\ \&(i-1)$ Since $i=d$ we know that $i-1<d$ which means $\&(i-1)$ is well-formed. Case $e=\&i$. Our goal is $\text{WF}_{d}\ \downarrow_{d}\&i$ which simplifies to $\text{WF}_{d}\ \&(i-1)$ We know from our premise that $\text{WF}_{d}\ \&i$ so $i<d$ which means $i-1<d$ so $\&(i-1)$ is well-formed. ∎ The following lemma shows that upshifting is an inverse of downshifting. In Coq this lemma is upshift_downshift. ###### Lemma 5. $\forall e.\ \forall d.\ \text{WF}_{d}\ e\implies\uparrow_{d}\downarrow_{d}e=e$ ###### Proof. We proceed by induction on $e$. Case $e=\lambda.\ b$. Our goal is $\uparrow_{d}\downarrow_{d}\lambda.\ b=\lambda.\ b$ By the definitions of $\uparrow_{d}$ and $\downarrow_{d}$ we can propagate them into the body of a lambda while incrementing $d$: $\lambda.\ \uparrow_{d+1}\downarrow_{d+1}b=\lambda.\ b$ We also have the additional premise $\text{WF}_{d}\ \lambda.b$ which simplifies to $\text{WF}_{d+1}\ b$. Since our inductive hypothesis $\forall d.\ \text{WF}_{d}\ b\implies\uparrow_{d}\downarrow_{d}b=b$ is universally quantified over $d$ we can instantiate it with $d+1$ in place of $d$ and discharge its precondition with our other premise $\text{WF}_{d+1}\ b$ to get $\uparrow_{d+1}\downarrow_{d+1}b=b$ and apply this to our goal to get $\lambda.\ b=\lambda.\ b$ which is trivially true. Case $e=f\ x$. Our goal is $\uparrow_{d}\downarrow_{d}(f\ x)=f\ x$ By the definitions of $\uparrow_{d}$ and $\downarrow_{d}$ we can propagate them into both sides of the application: $(\uparrow_{d}\downarrow_{d}f)(\uparrow_{d}\downarrow_{d}x)=f\ x$ We also have the additional premise $\text{WF}_{d}\ (f\ x)$ can be broken into the two premises $\text{WF}_{d}\ f$ and $\text{WF}_{d}\ x$. We can then apply our inductive hypotheses $\forall d.\ \text{WF}_{d}\ f\implies\uparrow_{d}\downarrow_{d}f=f$ and $\forall d.\ \text{WF}_{d}\ x\implies\uparrow_{d}\downarrow_{d}x=x$ with these premises to rewrite our goal to $f\ x=f\ x$ which is trivially true. Case $e=\$i$. Our goal is $\uparrow_{d}\downarrow_{d}\$i=\$i$ We will handle this in 3 cases: * • Case $i<d$. $\uparrow_{d}\downarrow_{d}\$i=\$i$ Unfold definition of $\downarrow_{d}$ $\uparrow_{d}\$i=\$i$ Unfold definition of $\uparrow_{d}$ $\$i=\$i$ * • Case $i>d$. $\uparrow_{d}\downarrow_{d}\$i=\$i$ Unfold definition of $\downarrow_{d}$ $\uparrow_{d}\$(i-1)=\$i$ Unfold definition of $\uparrow_{d}$ noting that $(i-1)\geq d$ $\$((i-1)+1)=\$i$ Simplify $\$i=\$i$ * • Case $i=d$. $\uparrow_{d}\downarrow_{d}\$i=\$i$ Unfold definition of $\downarrow_{d}$ $\uparrow_{d}\&(i-1)=\$i$ Unfold definition of $\uparrow_{d}$ noting that $(i-1)+1=d$ $\$((i-1)+1)=\$i$ Simplify $\$i=\$i$ Case $e=\&i$. Our goal is $\uparrow_{d}\downarrow_{d}\&i=\&i$ Unfolding the definition of $\downarrow_{d}$ yields $\uparrow_{d}\&(i-1)=\&i$ Since we know that $\text{WF}_{d}\ \&i$ by our premise, we know that $i<d$ so $(i-1)+1<d$ so $(i-1)+1\neq d$ so $\uparrow_{d}$ unfolds to: $\&((i-1)+1)=\&i$ which simplifies to $\&i=\&i$ which is trivially true. Case $e=t$. Our goal is $\uparrow_{d}\downarrow_{d}t=t$ Neither $\uparrow_{d}$ nor $\downarrow_{d}$ have any effect on a grammar primitive $t$, so this is trivially true. ∎ The following lemma shows that any result of LambdaUnify is a well-formed mapping. In Coq this lemma is called lu_wf. ###### Lemma 6. $\forall A.\ \forall e.\text{WFMap}\ \textsc{LambdaUnify}(A,e)$ ###### Proof. This is a straightforward proof by induction. In the $\alpha_{i}$ and and $\texttt{??}_{i}$ cases LambdaUnify adds expressions from the original expression to the mapping, which are well-formed as they can’t contain $\&i$ variables. In the lambda case DownshiftAll shifts variables using $\downarrow$ which is guaranteed to preserve well-formedness by Lemma 4. In the application case merge simply combines mappings from the function and argument calls to LambdaUnify which are well-formed by induction, resulting in an overall well- formed mapping. ∎ The following lemma follows from Lemma 5 and shows that UpshiftAll is an inverse of DownshiftAll. In Coq this lemma is upshift_all_downshift_all. ###### Lemma 7. $\text{WFMap}\ l\implies\textsc{UpshiftAll}(\textsc{DownshiftAll}(l))=l$ ###### Proof. Writing $l$ as $[\alpha_{i}\to\ e_{i}^{\prime},\ \texttt{??}_{j}\to\ u_{j}^{\prime},\ ...]$ and unfolding the definitions of UpshiftAll DownshiftAll we get: $[\alpha_{i}\to\ \uparrow_{0}\downarrow_{0}e_{i}^{\prime},\ \texttt{??}_{j}\to\ \uparrow_{0}\downarrow_{0}u_{j}^{\prime},\ ...]=[\alpha_{i}\to\ e_{i}^{\prime},\ \texttt{??}_{j}\to\ u_{j}^{\prime},\ ...]$ This equality will hold if $\uparrow_{0}\downarrow_{0}u_{j}^{\prime}=u_{j}^{\prime}$ and $\uparrow_{0}\downarrow_{0}e_{i}^{\prime}=e_{i}^{\prime}$ for all $u_{j}^{\prime}$ and $e_{i}^{\prime}$. $\text{WFMap}\ l$ tells us that all expressions $u_{j}^{\prime}$ and $e_{i}^{\prime}$ are well-formed and thus satisfy the precondition of Lemma 5, which proves these equalities true. ∎ The following two lemmas allow us to ”forget” one side of a merge when it won’t have any effect on the substitution. In Coq the first lemma is merge_forget3 and the second is captured by the definition Compatible. ###### Lemma 8. If $l_{1}$ is a mapping that binds every abstraction variable and hole in $A$, then $\textsc{merge}(l_{1},l_{2})\circ A=l_{1}\circ A$. ###### Proof. $l_{1}$ already assigns to every abstraction variable and hole, and merging in $l_{2}$ can’t overwrite these assignments since merge fails when the same variable or hole is assigned to two different expressions. Thus merging in $l_{2}$ has no effect on the resulting expression from substitution. ∎ ###### Lemma 9. If $l_{2}$ is a mapping that binds every abstraction variable and hole in $A$, then $\textsc{merge}(l_{1},l_{2})\circ A=l_{2}\circ A$. ###### Proof. This proof is identical in structure to Lemma 8. $l_{2}$ already assigns to every abstraction variable and hole, and merging in $l_{1}$ can’t overwrite these assignments since merge fails when the same variable or hole is assigned to two different expressions. Thus merging in $l_{1}$ has no effect on the resulting expression from substitution. ∎ The following lemma shows that any result of LambdaUnify satisfies the precondition of Lemma 9 and Lemma 8. ###### Lemma 10. $\textsc{LambdaUnify}(A,e)$ returns a mapping that binds all abstraction variables $\alpha_{i}$ and all holes $\texttt{??}_{i}$ in $A$. ###### Proof. This is a straightforward proof by induction. In the $\alpha_{i}$ and and $\texttt{??}_{i}$ cases LambdaUnify binds the variable and hole respectively, in the lambda case DownshiftAll has no effect on which abstraction variables are bound (and it binds all the variables in the body by induction), and in the application case merge results in combining the function and argument mappings (which bind all variables by induction) to result in a mapping which binds all variables that appear in both the function and argument. ∎ ### B.3. Correctness Theorem In Coq this theorem is called correctness. ###### Theorem B.1. $\forall A.\ \forall e.\ \textsc{LambdaUnify}(A,e)\circ A=e$ ###### Proof. We proceed by induction over the proof tree of $\textsc{LambdaUnify}(A,e)$. Case U-AbsVar. Our goal is $\textsc{LambdaUnify}(A,\alpha)\circ\alpha=e$ which simplifies by definition of LambdaUnify to $[\alpha\to e]\circ\alpha=e$ which is true by the definition of substitution. Case U-Hole. Our goal is $\textsc{LambdaUnify}(A,\texttt{??})\circ\texttt{??}=e$ which simplifies by definition of LambdaUnify to $[\texttt{??}\to e]\circ\texttt{??}=e$ which is true by the definition of substitution. Case U-App. Our goal is $\textsc{LambdaUnify}(A_{1}\ A_{2},\ e_{1}\ e_{2})\circ(A_{1}\ A_{2})=e_{1}\ e_{2}$ This simplifies by definition of LambdaUnify to $\textsc{merge}(\textsc{LambdaUnify}(A_{1},e_{1}),\textsc{LambdaUnify}(A_{2},e_{2}))\circ(A_{1}\ A_{2})=e_{1}\ e_{2}$ We can distribute the substitution over the application (by definition of substitution in Fig. 5 (Left)) to get $\displaystyle($ $\displaystyle(\textsc{merge}(\textsc{LambdaUnify}(A_{1},e_{1}),\textsc{LambdaUnify}(A_{2},e_{2}))\circ A_{1})$ $\displaystyle(\textsc{merge}(\textsc{LambdaUnify}(A_{1},e_{1}),\textsc{LambdaUnify}(A_{2},e_{2}))\circ A_{2}))$ $\displaystyle=e_{1}\ e_{2}$ We can then forget the irrelevant part of each merge using Lemma 9 and Lemma 8 where the precondition (binding all abstraction variables and holes) is satisfied by Lemma 10: $\displaystyle($ $\displaystyle(\textsc{LambdaUnify}(A_{1},e_{1})\circ A_{1})$ $\displaystyle(\textsc{LambdaUnify}(A_{2},e_{2})\circ A_{2}))$ $\displaystyle=e_{1}\ e_{2}$ We can then finish the proof by directly applying our inductive hypotheses $\textsc{LambdaUnify}(A_{1},e_{1})\circ A_{1}=e_{1}$ and $\textsc{LambdaUnify}(A_{2},e_{2})\circ A_{2}=e_{2}$. Case U-Lam. Our goal is $\textsc{LambdaUnify}(\lambda.\ A,\lambda.\ e)\circ\lambda.\ A=\lambda.\ e$ This simplifies by definition of LambdaUnify to $\textsc{DownshiftAll}(\textsc{LambdaUnify}(A,e))\circ\lambda.\ A=\lambda.\ e$ We can move the substitution under the lambda while inserting UpshiftAll as per the definition of substitution $\lambda.\ \textsc{UpshiftAll}(\textsc{DownshiftAll}(\textsc{LambdaUnify}(A,e)))\circ A=\lambda.\ e$ Since we know that any result of LambdaUnify is well-formed by Lemma 6 we can apply Lemma 7 to get $\lambda.\ \textsc{LambdaUnify}(A,e)\circ A=\lambda.\ e$ We can then apply our inductive hypothesis $\textsc{LambdaUnify}(A,e)\circ A=e$ to get $\lambda.\ e=\lambda.\ e$ and we are done. Case U-Same. Our goal is $\textsc{LambdaUnify}(e,e)\circ e=e$ which simplifies by definition of LambdaUnify to $[]\circ e=e$ which is trivially true as the empty substitution has no effect. ∎
# Statistical Nonlocality of Dynamically Coherent Structures Andre N. Souza1<EMAIL_ADDRESS>Tyler Lutz2 Glenn R. Flierl1 1Massachusetts Institute of Technology, Cambridge, MA, United States 2 Yale University, New Haven, CT, United States ###### Abstract We introduce a class of stochastic advection problems amenable to analysis of turbulent transport. The statistics of the flow field are represented as a continuous time Markov process, a choice that captures the intuitive notion of turbulence as moving from one coherent structure to another. We obtain closed form expressions for the turbulent transport operator without invoking approximations. We recover the classical estimate of turbulent transport as a diffusivity tensor, the components of which are the integrated auto- correlation of the velocity field, in the limit that the operator becomes local in space and time. ## 1 Introduction The study of passive scalar transport is of fundamental importance in characterizing turbulence. Given that predicting a chaotic dynamical trajectory over long times is infeasible (Lorenz, 1963) one must switch to a statistical perspective to make headway on transport properties. From the analysis of anomalous dispersion by Taylor (1922), operator notions of mixing from Knobloch (1977), computations of “effective diffusivity” by Avellaneda & Majda (1991), simplified models of turbulence by Pope (2011), rigorous notions of mixing in terms of Sobolev norms by Thiffeault (2012), or upper bounds on transport as in Hassanzadeh et al. (2014), different approaches elucidate fundamental properties of turbulence. In addition to furthering our understanding of turbulence, there are practical applications for turbulence closures. In particular, Earth Systems Models require closure relations for the transport of unresolved motions Schneider et al. (2017); however, the closure relations are marred by structural and parametric uncertainty, requiring ad-hoc tuning to compensate for biases. There are structural biases associated with scaling laws and closure assumptions between turbulent fluxes and gradients. Modern studies are bridging the gap by incorporating more complex physics and novel scaling laws, Tan et al. (2018); Gallet & Ferrari (2020), but the chosen functional forms to represent fluxes remain unknown. The multi-scale nature of turbulent flows, the presence of coherent structures, as well as the interconnection of reacting chemical species and prognostic fields, instead suggests that fluxes are perhaps better modeled using nonlinear nonlocal (in space, time, and state) operators. Data-driven methods relying on flexible interpolants can significantly reduce structural bias, but often at the expense of interpretability, generalizability, or efficiency. Thus, understanding scaling laws and functional forms of turbulence closures is still necessary to physically constrain data-driven methods and decrease their computational expense. A promising avenue for significant progress, lying at the intersection of theory and practice, is the calculation of closure relations for passive scalars. The present work aims to characterize the transport of passive scalars by flow fields with known statistics. Our notion of transport is the ensemble mean flux, which we express as an operator that acts on the ensemble mean tracers in terms of the statistics of the flow field. We make arguments akin to those in Kraichnan (1968) to motivate the operator approach, but our method of calculation is fundamentally different. Given that the goal is to construct an operator, rather than estimating a diffusivity tensor acting on the ensemble mean gradients or deriving upper bounds, we take a field-theoretic perspective (Hopf, 1952). Doing so allows us to derive a coupled set of partial differential equations representing conditional mean tracers where the conditional averages are with respect to different flow states. The turbulent transport operator is then a Schur complement of the resulting linear system with respect to statistical “perturbation” variables. If the flow statistics are given by a continuous-time Markov process with a small finite state space, the Schur complement becomes tractable to compute analytically. Obtaining a closed form functional for turbulent fluxes allows for a direct statistical simulation approach similar to those of Venturi et al. (2013); Allawala & Marston (2016); Farrell & Ioannou (2019). The paper is organized as follows. In section 2 we formulate the closure problem and recast it as one of solving coupled partial differential equations. In section 3 we show how to explicitly solve the closure problem for a few flows with simple temporal structure but arbitrarily complex spatial structure. In section 4 we outline the general theory. Appendices supplement the body of the manuscript. Appendix A provides a direct field-theoretic derivation of arguments in section 2 and Appendix B provides a heuristic overview of obtaining continuous-time Markov processes and their statistics from deterministic or stochastic dynamical systems. ## 2 Problem Formulation We consider the advection and diffusion of an ensemble of passive scalars $\theta_{\omega}$ $\displaystyle\partial_{t}\theta_{\omega}+\nabla\cdot\left(\bm{u}_{\omega}\theta_{\omega}-\kappa\nabla\theta_{\omega}\right)$ $\displaystyle=s(\bm{x})$ (1) by a stochastic flow field $\bm{u}_{\omega}(\bm{x},t)$ where $\omega$ labels the ensemble member. Here $s$ is a deterministic mean zero source term and $\kappa$ is a diffusivity constant111For laboratory flows $\kappa$ would be the molecular diffusivity; for larger-scale problems, we rely on the fact that away from boundaries, the ensemble mean advective flux (but not necessarily other statistics) may still be much larger than the diffusive flux and thus the exact value of $\kappa$ will not matter.. Our target is to obtain a meaningful equation for the ensemble mean, $\displaystyle\partial_{t}\langle\theta\rangle+\nabla\cdot\left(\langle\bm{u}\theta\rangle-\kappa\nabla\langle\theta\rangle\right)$ $\displaystyle=s(\bm{x})$ (2) which requires a computationally amenable expression for the mean advective flux, $\langle\bm{u}\theta\rangle$, in terms of the statistics of flow field, $\bm{u}(\bm{x},t)$, and the ensemble average of the tracer, $\langle\theta\rangle$. Thus the closure problem is to find an operator $\mathcal{O}$ that relates the ensemble mean, $\langle\theta\rangle$, to the ensemble mean advective-flux, $\langle\bm{u}\theta\rangle$, i.e., $\displaystyle\mathcal{O}[\langle\theta\rangle]=\langle\bm{u}\theta\rangle.$ (3) We show how to define (and solve for) the operator $\mathcal{O}$. The operator will be linear with respect to its argument and depend on the statistics of the flow field. We assume all tracer ensemble members to have the same initial condition and thus the ensemble average here is with respect to different flow realizations. The only source of randomness comes from different flow realizations. Throughout the manuscript we assume homogeneous Neumann boundary conditions for the tracer and zero wall-normal flow for the velocity field when boundaries are present. These restrictions, combined with the assumption that the source term is mean zero, imply that the tracer average is conserved. For the statistics of the flow field, we consider a continuous time Markov process with $N$ states corresponding to steady flow fields $\bm{u}_{n}(\bm{x})$ where $n$ is the associated state index. We start with finitely many states for simplicity rather than necessity. Physically, we think of these states as representing coherent structures in a turbulent flow222This is viewed as a finite volume discretization in function space where the states are the “cell averages” of a control volume in function space.. A turbulent flow, by its very nature, chaotic and unpredictable over long time horizons, is modelled as being semi-unpredictable through our choice. Over short horizons, the probability of remaining in a given state is large. On the medium term, the flow is limited to moving to a subset of likely places in phase space. Over long time horizons, the most one can say about the flow is related to the likelihood of being found in the appropriate subset of phase space associated with the statistically steady state. Thus we proceed by characterizing the probability, $\mathbb{P}$, of transitioning from state $n$ to state $m$ by a transition matrix $\mathscr{P}(\tau)$, $\displaystyle\mathbb{P}\\{\bm{u}(\bm{x},t+\tau)=\bm{u}_{m}(\bm{x})|\bm{u}(\bm{x},t)=\bm{u}_{n}(\bm{x})\\}$ $\displaystyle=[\mathscr{P}(\tau)]_{mn}.$ (4) The transition probability is defined through its relation to the generator $\mathcal{Q}$, $\displaystyle\mathscr{P}(\tau)\equiv\exp(\mathcal{Q}\tau)$ (5) where $\exp(\mathcal{Q}\tau)$ is a matrix exponential. Each entry of $\mathscr{P}(\tau)$ must be positive. Furthermore, the column sum of $\mathscr{P}(\tau)$ for each $\tau$, sum to one since the total probability must sum to one. Similarly $\mathcal{Q}$’s off-diagonal terms must be positive333Indeed, to first order $\exp(\mathcal{Q}dt)=\mathbb{I}+\mathcal{Q}dt$. The positivity requirement of the transition probability $\mathscr{P}(dt)=\exp(\mathcal{Q}dt)$ necessitates the positivity of $\mathcal{Q}$’s off-diagonal terms as well as the negativity of the diagonal terms. and the column sum of $\mathcal{Q}$ must be zero. We denote the probability of being found at state $m$ at time $t$ by $\mathcal{P}_{m}(t)$, $\displaystyle\mathcal{P}_{m}(t)=\mathbb{P}\\{\bm{u}(\bm{x},t)=\bm{u}_{m}(\bm{x})\\}.$ (6) The evolution equation for $\mathcal{P}_{m}(t)$ is the master equation, $\displaystyle\frac{d}{dt}\mathcal{P}_{m}$ $\displaystyle=\sum_{n}\mathcal{Q}_{mn}\mathcal{P}_{n}.$ (7) We assume that Equation 7 has a unique steady state and denote the components of the steady state by $P_{m}$. We have used several ”P”s at this stage and their relation are: 1. 1. $\mathbb{P}$ denotes a probability. 2. 2. $\mathscr{P}(\tau)$ denotes the transition probability matrix for a time $\tau$ in the future. 3. 3. $\mathcal{P}_{m}(t)$ denotes the probability of being in state $m$ at time $t$. The algebraic relation $\sum_{m}P_{m}(t+\tau)\bm{\hat{e}}_{m}=\mathscr{P}(\tau)\sum_{n}P_{n}(t)\bm{\hat{e}}_{n}$ holds. 4. 4. $P_{m}$ is the statistically steady probability of being found in state $m$. In the limit $\lim_{t\rightarrow\infty}\mathcal{P}_{m}(t)=P_{m}.$ We exploit the given information about the flow field to infer the mean statistics of the passive tracer $\theta_{\omega}$. We do so by conditionally averaging the tracer field $\theta_{\omega}$ with respect to a given flow state $\bm{u}_{n}$. More precisely, given the stochastic partial differential equation, $\displaystyle\mathbb{P}\\{\bm{u}_{\omega}(\bm{x},t+\tau)=\bm{u}_{m}(\bm{x})|\bm{u}_{\omega}(\bm{x},t)=\bm{u}_{n}(\bm{x})\\}$ $\displaystyle=[\exp(\mathcal{Q}\tau)]_{mn},$ (8) $\displaystyle\partial_{t}\theta_{\omega}+\nabla\cdot\left(\bm{u}_{\omega}\theta_{\omega}-\kappa\nabla\theta_{\omega}\right)$ $\displaystyle=s(\bm{x}),$ (9) we shall obtain equations for probability weighted conditional means of $\theta_{\omega}$ defined by $\displaystyle\Theta_{m}(\bm{x},t)$ $\displaystyle\equiv\langle\theta_{\omega}\rangle_{\bm{u}(\bm{x},t)=\bm{u}_{m}(\bm{x})}\mathcal{P}_{m}(t).$ (10) We will show that the evolution equation for $\Theta_{m}$ is $\displaystyle\frac{d}{dt}\mathcal{P}_{m}$ $\displaystyle=\sum_{n}\mathcal{Q}_{mn}\mathcal{P}_{n},$ (11) $\displaystyle\partial_{t}\Theta_{m}+\nabla\cdot\left(\bm{u}_{m}\Theta_{m}-\kappa\nabla\Theta_{m}\right)$ $\displaystyle=s(\bm{x})\mathcal{P}_{m}+\sum_{n}\mathcal{Q}_{mn}\Theta_{n}.$ (12) The explicit dependence on the generator in Equation 12, as we shall see, yields considerable information. We recover the equation for the tracer ensemble mean, Equation 2, by summing Equation 12 over the index $m$, using $\langle\theta\rangle=\sum_{m}\Theta_{m}$, $\sum_{m}\mathcal{Q}_{mn}=\bm{0}$, and $\sum_{m}\mathcal{P}_{m}=1$, $\displaystyle\partial_{t}\sum_{m}\Theta_{m}+\nabla\cdot\left(\sum_{m}\bm{u}_{m}\Theta_{m}-\kappa\nabla\sum_{m}\Theta_{m}\right)$ $\displaystyle=s(\bm{x})$ (13) $\displaystyle\Leftrightarrow$ $\displaystyle\partial_{t}\langle\theta\rangle+\nabla\cdot\left(\langle\bm{u}\theta\rangle-\kappa\nabla\langle\theta\rangle\right)$ $\displaystyle=s(\bm{x}).$ (14) We comment that the presence of the generator when taking conditional averages is similar to the entrainment hypothesis in the atmospheric literature. See, for example, Tan et al. (2018) for its use in motivating a turbulence closure; however, here we derive the result from the direct statistical representation as opposed to hypothesize its presence from a dynamical argument. Most of the terms in Equation 12 are obtained by applying a conditional average to Equation 9, commuting with spatial derivatives when necessary, and then multiplying through by $\mathcal{P}_{m}$; however, the primary difficulty lies in proper treatment of the conditional average of the temporal derivative. We circumvent the problem in a roundabout manner: The strategy is to discretize the advection-diffusion equation, write down the resulting master equation, compute moments of the probability distribution, and then take limits to restore the continuum nature of the advection-diffusion equation. For an alternative derivation where we forego discretization see Appendix A and for a brief overview of the connection between the discrete, continuous, and mixed master equation see Appendix B or, in a simpler context, Hagan et al. (1989). A generic discretization (in any number of dimensions) of Equation 9 is of the form $\displaystyle\frac{d}{dt}\theta^{i}+\sum_{jkc}A_{ijk}^{c}u^{k,c}_{\omega}\theta^{j}-\sum_{j}D_{ij}\theta^{j}$ $\displaystyle=s^{i}$ (15) for some tensor $A_{ijk}^{c}$, representing advection, and matrix $D_{ij}$, representing diffusion. Here each $i,j$ and $k$ corresponds to a spatial location, and the index $c$ corresponds to a component of the velocity field $\bm{u}$. The variable $\theta^{i}$ is the value of the tracer at grid location $i$ and $u^{k,c}$ is the value of the $c$’th velocity component and grid location $k$. The master equation for the joint probability density for each component $\theta^{i}$ and Markov state $m$, $\rho_{m}(\bm{\theta})$, where $\bm{\theta}=(\theta^{1},\theta^{2},....)$ and the $m$-index denotes a particular Markov state, is a combination of the Liouville equation for 15 and the transition rate equation for 8, $\displaystyle\partial_{t}\rho_{m}$ $\displaystyle=\sum_{i}\frac{\partial}{\partial\theta^{i}}\left[\left(\sum_{jkc}A_{ijk}^{c}u^{k,c}_{m}\theta^{j}-\sum_{j}D_{ij}\theta^{j}-s^{i}\right)\rho_{m}\right]+\sum_{n}\mathcal{Q}_{mn}\rho_{n}.$ (16) Define the following moments, $\displaystyle\mathcal{P}_{m}$ $\displaystyle=\int d\bm{\theta}\rho_{m}\text{ and }\Theta_{m}^{j}=\int d\bm{\theta}\theta^{j}\rho_{m}.$ (17) We obtain an equation for $\mathcal{P}_{m}$ by integrating 130 by $d\bm{\theta}$ to yield $\displaystyle\frac{d}{dt}\mathcal{P}_{m}$ $\displaystyle=\sum_{n}\mathcal{Q}_{mn}\mathcal{P}_{n}$ (18) as expected from Equation 7. The equation for $\Theta_{m}^{\ell}$ is obtained by multiplying 130 by $\theta^{\ell}$ and then integrating with respect to $d\bm{\theta}$, $\displaystyle\frac{d}{dt}\Theta_{m}^{\ell}$ $\displaystyle=-\sum_{jkc}A_{\ell jk}^{c}u^{k,c}_{m}\Theta^{j}_{m}+\sum_{j}D_{\ell j}\Theta^{j}_{m}+s^{\ell}\mathcal{P}_{m}+\sum_{n}\mathcal{Q}_{mn}\Theta_{n}^{\ell},$ (19) where we integrated by parts on the $\int d\bm{\theta}\theta^{\ell}\partial_{\theta^{i}}\bullet$ term. Upon taking limits of 19 we have the following equations $\displaystyle\frac{d}{dt}\mathcal{P}_{m}$ $\displaystyle=\sum_{n}\mathcal{Q}_{mn}\mathcal{P}_{n}$ (20) $\displaystyle\partial_{t}\Theta_{m}+\nabla\cdot\left(\bm{u}_{m}\Theta_{m}-\kappa\nabla\Theta_{m}\right)$ $\displaystyle=s(\bm{x})\mathcal{P}_{m}+\sum_{n}\mathcal{Q}_{mn}\Theta_{n}.$ (21) We compare Equation 21 to the direct application of the conditional average to Equation 9 followed by multiplication with $\mathcal{P}_{m}$ to infer, $\displaystyle\langle\partial_{t}\theta_{\omega}\rangle_{\bm{u}(\bm{x},t)=\bm{u}_{m}(\bm{x})}\mathcal{P}_{m}$ $\displaystyle=\partial_{t}\Theta_{m}-\sum_{n}\mathcal{Q}_{mn}\Theta_{n}.$ (22) In summary, for an m-dimensional advection diffusion equation and $N$ Markov states, Equations 11-12 are a set of $N$-coupled m-dimensional advection diffusion equations with $N$ different steady velocities. When the statistics of the flow field are described by $c$ continuous variables, the resulting equation set becomes an $m+c$ dimensional system. Stated differently, if the statistics of $\bm{u}_{\omega}$ are characterized by a transitions between a continuum of states associated with a linear operator $\mathcal{F}_{\bm{\omega}}$ with variables $\bm{\omega}\in\mathbb{R}^{c}$, then $\displaystyle\partial_{t}\mathcal{P}$ $\displaystyle=\mathcal{F}_{\bm{\omega}}[\mathcal{P}]$ (23) $\displaystyle\partial_{t}\Theta+\nabla\cdot\left(\bm{u}\Theta-\kappa\nabla\Theta\right)$ $\displaystyle=s(\bm{x})\mathcal{P}+\mathcal{F}_{\bm{\omega}}[\Theta],$ (24) where $\mathcal{P}=\mathcal{P}(\bm{\omega},t)$, $\Theta=\Theta(\bm{x},\bm{\omega},t)$, and $\bm{u}=\bm{u}(\bm{x},\bm{\omega})$. Equations 20-21 are thought of as finite volume discretizations of flow statistics in Equations 23-24. Our primary concern in this work is to use Equations 11-12 to calculate meaningful expressions for $\langle\bm{u}\theta\rangle$; however, we shall first take a broader view to understand the general structure of the turbulent fluxes. The following argument is attributed to Weinstock (1969), but we use our own notation and make additional simplifications. Applying the Reynolds decomposition $\displaystyle\theta_{\omega}=\langle\theta\rangle+\theta^{\prime}_{\omega}\text{ and }\bm{u}_{\omega}=\langle\bm{u}\rangle+\bm{u}^{\prime}_{\omega}$ (25) yields $\displaystyle\partial_{t}\langle\theta\rangle+\nabla\cdot\left(\langle\bm{u}\rangle\langle\theta\rangle+\langle\bm{u}^{\prime}\theta^{\prime}\rangle-\kappa\nabla\langle\theta\rangle\right)$ $\displaystyle=s$ (26) $\displaystyle\partial_{t}\theta^{\prime}_{\omega}+\nabla\cdot\left(\langle\bm{u}\rangle\langle\theta\rangle-\langle\bm{u}^{\prime}\theta^{\prime}\rangle+\bm{u}_{\omega}\theta_{\omega}-\kappa\nabla\theta^{\prime}_{\omega}\right)$ $\displaystyle=0$ (27) The perturbation equation is rewritten as $\displaystyle\partial_{t}\theta^{\prime}_{\omega}+\nabla\cdot\left(\bm{u}_{\omega}^{\prime}\theta_{\omega}^{\prime}-\langle\bm{u}^{\prime}\theta^{\prime}\rangle+\langle\bm{u}\rangle\theta_{\omega}^{\prime}-\kappa\nabla\theta^{\prime}_{\omega}\right)$ $\displaystyle=-\nabla\cdot\left(\bm{u}_{\omega}^{\prime}\langle\theta\rangle\right)$ (28) This is an infinite system (or finite depending on the number of ensemble members) of coupled pde’s between the different ensemble members. The ensemble members are coupled due to the presence of the turbulent flux, $\langle\bm{u}^{\prime}\theta^{\prime}\rangle$. The key observation is to notice the terms on the left hand side involve the perturbation variables and not the ensemble mean of the gradients. Assuming it is possible to find the inverse, the Green’s function for the large linear system is used to yield $\displaystyle\theta^{\prime}_{\omega}(\bm{x},t)$ $\displaystyle=-\int d\bm{x}^{\prime}dt^{\prime}d\mu_{\alpha}\mathcal{G}_{\alpha\omega}(\bm{x},t|\bm{x}^{\prime},t^{\prime})\nabla\cdot\left(\bm{u}_{\alpha}^{\prime}\langle\theta\rangle\right)$ (29) where we also have to integrate with respect to the measure defining the different ensemble members through $d\mu_{\alpha}$. Notation wise this would means $\langle\theta\rangle=\int d\mu_{\omega}\theta_{\omega}$. We use this expression to rewrite the turbulent flux as $\displaystyle\langle\bm{u}^{\prime}\theta^{\prime}\rangle$ $\displaystyle=-\int d\bm{x}^{\prime}dt^{\prime}d\mu_{\omega}d\mu_{\alpha}\bm{u}^{\prime}_{\omega}(\bm{x},t)\mathcal{G}_{\alpha\omega}(\bm{x},t|\bm{x}^{\prime},t^{\prime})\left[\nabla\cdot\left(\bm{u}_{\alpha}^{\prime}(\bm{x}^{\prime},t^{\prime})\langle\theta\rangle(\bm{x}^{\prime},t^{\prime})\right)\right]$ (30) We make two simplifications for illustrative purposes. 1. 1. All ensemble averages are independent of time. 2. 2. The flow is incompressible, i.e., $\nabla\cdot\bm{u}=0$. Equation 30 becomes $\displaystyle\langle\bm{u}^{\prime}\theta^{\prime}\rangle$ $\displaystyle=-\int d\bm{x}^{\prime}dt^{\prime}d\mu_{\omega}d\mu_{\alpha}\left[\bm{u}^{\prime}_{\omega}(\bm{x},t)\mathcal{G}_{\alpha\omega}(\bm{x},t|\bm{x}^{\prime},t^{\prime})\bm{u}_{\alpha}^{\prime}(\bm{x}^{\prime},t^{\prime})\right]\cdot\nabla\langle\theta\rangle(\bm{x}^{\prime})$ (31) We perform the $t^{\prime},\alpha,\omega$ integrals first to define the turbulent-diffusivity tensor kernel as $\displaystyle\langle\bm{u}^{\prime}\theta^{\prime}\rangle$ $\displaystyle=-\int d\bm{x}^{\prime}\underbrace{\int dt^{\prime}d\mu_{\omega}d\mu_{\alpha}\left[\bm{u}^{\prime}_{\omega}(\bm{x},t)\otimes\bm{u}_{\alpha}^{\prime}(\bm{x}^{\prime},t^{\prime})\mathcal{G}_{\alpha\omega}(\bm{x},t|\bm{x}^{\prime},t^{\prime})\right]}_{\bm{\mathcal{K}}(\bm{x}|\bm{x}^{\prime})}\cdot\nabla\langle\theta\rangle(\bm{x}^{\prime})$ (32) $\displaystyle=-\int d\bm{x}^{\prime}\bm{\mathcal{K}}(\bm{x}|\bm{x}^{\prime})\cdot\nabla\langle\theta\rangle(\bm{x}^{\prime})$ (33) The independence of $\bm{\mathcal{K}}$ with respect to $t$ follows from the time-independence of $\langle\bm{u}^{\prime}\theta^{\prime}\rangle$ and $\langle\theta\rangle$. In total we see $\displaystyle\langle\bm{u}\theta\rangle=\langle\bm{u}\rangle\langle\theta\rangle-\int d\bm{x}^{\prime}\bm{\mathcal{K}}(\bm{x}|\bm{x}^{\prime})\cdot\nabla\langle\theta\rangle(\bm{x}^{\prime}).$ (34) An insight from Equation 34 is the dependence of turbulent fluxes $\langle\bm{u}^{\prime}\theta^{\prime}\rangle$ at location $\bm{x}$ as a weighted sum of gradients of the mean variable $\langle\theta\rangle$ at locations $\bm{x}^{\prime}$. The operator is linear and amenable to computation, even in turbulent flows, Bhamidipati et al. (2020). We consider the spectrum for the turbulent diffusivity operator $\int d\bm{x}^{\prime}\mathcal{K}(\bm{x}|\bm{x}^{\prime})\bullet$ as a characterization of turbulent-mixing by the flow field $\bm{u}(\bm{x},t)$. We comment that the operator $\int d\bm{x}^{\prime}\mathcal{K}(\bm{x}|\bm{x}^{\prime})\bullet$ is a mapping from vector fields to vector fields whereas the kernel $\mathcal{K}(\bm{x}|\bm{x}^{\prime})$ is a mapping from two positions to a tensor. For example, consider a one-dimensional problem in a periodic domain $x\in[0,2\pi)$. If $\mathcal{K}(x|x^{\prime})=\kappa_{e}\delta(x-x^{\prime})$ for some positive constant $\kappa_{e}$, the spectrum of the operator is flat and turbulent-mixing remains the same on every length scale. If $\mathcal{K}(x|x^{\prime})=-\kappa_{e}\partial_{xx}^{2}\delta(x-x^{\prime})$ then the rate of mixing increases with increasing wavenumber, one gets hyperdiffusion. And lastly, if $\int dx^{\prime}\mathcal{K}(x|x^{\prime})\bullet=(\kappa_{e}-\partial_{xx})^{-1}$, then the kernel is nonlocal and the rate of mixing decreases at smaller length scales. In the following section we calculate $\int d\bm{x}^{\prime}\mathcal{K}(\bm{x}|\bm{x}^{\prime})\bullet$ directly from the conditional equations and then discuss the general structure in Section 4. ## 3 Examples We now go through three examples to understand the implications of Equations 11-12. The three examples follow sequentially in increasing complexity. The first example considers transitions between two Markov states. There we introduce a generalizable approach to computing the turbulent diffusivity. We then apply the same approach to a slightly more complex problem, transitions between three Markov states. And finally, we conclude with a calculation involving transitions between four Markov states where we delve into details with cellular flow states. The generator for the two and three state systems are derived from a finite volume discretization of an Ornstein-Uhlenbeck process, see Appendix B.2. Although we present the general form of the kernels here, in Sections 3.1 and 3.2 one can consider the case where the advection-diffusion equation is one- dimensional, periodic, and with a flow field that is constant in space. In this case one can decompose the advection-diffusion equation into Fourier modes that are decoupled from one another and allows for alternative computations of the same result using more standard techniques. ### 3.1 Two State To start we consider the simplest time-dependent mean-zero incompressible stochastic flow field: the transition between two incompressible states $\bm{u}_{1}(\bm{x})=\bm{u}(\bm{x})$ and $\bm{u}_{2}(\bm{x})=-\bm{u}(\bm{x})$ where each state is equally likely in the statistically steady state. For this we use the generator $\displaystyle\mathcal{Q}$ $\displaystyle=\gamma\begin{bmatrix}-1&1\\\ 1&-1\end{bmatrix}$ (35) where $\gamma>0$. The eigenvector and eigenvalues of the generator are $\displaystyle\bm{v}^{1}=\begin{bmatrix}1/2\\\ 1/2\end{bmatrix}\text{ and }\bm{v}^{2}=\begin{bmatrix}1/2\\\ -1/2\end{bmatrix}$ (36) with respective eigenvalues $\lambda^{1}=0$ and $\lambda^{2}=-2\gamma$. The first eigenvector is the steady state probability, which we see has probability $1/2$ for each state, as expected. In what follows we denote the Laplacian by $\Delta$. Equations 11-12 for the two-state system are $\displaystyle\partial_{t}\mathcal{P}_{1}$ $\displaystyle=-\gamma\mathcal{P}_{1}+\gamma\mathcal{P}_{2}$ (37) $\displaystyle\partial_{t}\mathcal{P}_{2}$ $\displaystyle=-\gamma\mathcal{P}_{2}+\gamma\mathcal{P}_{1}$ (38) $\displaystyle\partial_{t}\Theta_{1}+\nabla\cdot\left(\bm{u}\Theta_{1}\right)$ $\displaystyle=\kappa\Delta\Theta_{1}+s(\bm{x})\mathcal{P}_{1}-\gamma\Theta_{1}+\gamma\Theta_{2}$ (39) $\displaystyle\partial_{t}\Theta_{2}-\nabla\cdot\left(\bm{u}\Theta_{2}\right)$ $\displaystyle=\kappa\Delta\Theta_{2}+s(\bm{x})\mathcal{P}_{2}-\gamma\Theta_{2}+\gamma\Theta_{1}.$ (40) We assume a statistically steady state so that the temporal derivatives vanish. Consequently, $\mathcal{P}_{1}=\mathcal{P}_{2}=1/2$ and the equations reduce to $\displaystyle\nabla\cdot\left(\bm{u}\Theta_{1}\right)$ $\displaystyle=\kappa\Delta\Theta_{1}+s(\bm{x})/2-\gamma\Theta_{1}+\gamma\Theta_{2}$ (41) $\displaystyle-\nabla\cdot\left(\bm{u}\Theta_{2}\right)$ $\displaystyle=\kappa\Delta\Theta_{2}+s(\bm{x})/2-\gamma\Theta_{2}+\gamma\Theta_{1}$ (42) We rewrite the above equation set into a mean and perturbation by changing basis to $\varphi_{i}$ variables according to $\displaystyle\begin{bmatrix}1/2&1/2\\\ 1/2&-1/2\end{bmatrix}\begin{bmatrix}\varphi_{1}\\\ \varphi_{2}\end{bmatrix}=\begin{bmatrix}\Theta_{1}\\\ \Theta_{2}\end{bmatrix}\Leftrightarrow\begin{bmatrix}\varphi_{1}\\\ \varphi_{2}\end{bmatrix}=\begin{bmatrix}1&1\\\ 1&-1\end{bmatrix}\begin{bmatrix}\Theta_{1}\\\ \Theta_{2}\end{bmatrix}$ (43) yielding, $\displaystyle\nabla\cdot\left(\bm{u}\varphi_{2}\right)$ $\displaystyle=\kappa\Delta\varphi_{1}+s(\bm{x})$ (44) $\displaystyle\bm{u}\cdot\nabla\varphi_{1}$ $\displaystyle=\kappa\Delta\varphi_{2}-2\gamma\varphi_{2}.$ (45) We used the incompressibility condition to yield the representation in Equation 45. Our choice of basis is no accident, we used the eigenvectors of the generator, $\mathcal{Q}$, to define the transformation in Equation 43. We recognize the variable $\varphi_{1}$ as the ensemble mean $\varphi_{1}=\langle\theta\rangle$ and $\varphi_{2}$ as a ”perturbation” variable. The turbulent flux term is $\langle\bm{u}^{\prime}\theta^{\prime}\rangle=\bm{u}\varphi_{2}$. We eliminate the dependence on the perturbation variable by inverting the Helmholtz operator in Equation 45. In total we have the following representation of the mean equation $\displaystyle\nabla\cdot\left(\bm{u}(\kappa\Delta-2\gamma)^{-1}[\bm{u}\cdot\nabla\langle\theta\rangle]\right)$ $\displaystyle=\kappa\Delta\langle\theta\rangle+s(\bm{x})$ (46) from whence we extract the turbulent diffusivity operator $\displaystyle\int d\bm{x}^{\prime}\mathcal{K}(\bm{x}|\bm{x}^{\prime})\bullet=\bm{u}(2\gamma-\kappa\Delta)^{-1}\bm{u}.$ (47) We point out a few salient features of Equation 47. The inverse Helmholtz operator, $\left(2\gamma-\kappa\Delta\right)^{-1}$, damps high spatial frequency components of ensemble mean gradients. Thus, the operator’s eigenvalues decrease as one examines increasingly fine-scale structure. Intuitively, as one examines a small-scale structure, the presence of diffusivity leads to lower turbulent fluxes, expressing the notion that it is difficult to transport something that immediately diffuses. The second observation pertains to the presence of the eigenvalue of the generator in the Helmholtz operator. If the flow field changes rapidly, transitioning between the disparate states, then $\gamma$ is large, and one can expect the turbulent-diffusivity to be local. In other words, the flow does not stay sufficiently long time near a coherent structure. In this example, the non-locality of the turbulent diffusivity is enabled by the presence of the regular diffusion operator. However, in the following example, we show that this need not be the case. ### 3.2 Three State For this example we consider a Markov process that transitions between three incompressible states $\bm{u}_{1}(\bm{x})=\bm{u}(\bm{x})$, $\bm{u}_{2}(\bm{x})=0$, and $\bm{u}_{3}(\bm{x})=-\bm{u}(\bm{x})$. For this let the generator be $\displaystyle\mathcal{Q}$ $\displaystyle=\gamma\begin{bmatrix}-1&1/2&0\\\ 1&-1&1\\\ 0&1/2&-1\end{bmatrix}$ (48) where $\gamma>0$. The eigenvectors of the generator are $\displaystyle\bm{v}^{1}=\begin{bmatrix}1/4\\\ 1/2\\\ 1/4\end{bmatrix}\text{ , }\bm{v}^{2}=\begin{bmatrix}1/2\\\ 0\\\ -1/2\end{bmatrix}\text{ and }\bm{v}^{3}=\begin{bmatrix}1/4\\\ -1/2\\\ 1/4\end{bmatrix}$ (49) with respective eigenvalues $\lambda^{1}=0$, $\lambda^{2}=-\gamma$, and $\lambda^{3}=-2\gamma$. The statistically steady three-state manifestation of Equations 11-12 are $\displaystyle\nabla\cdot\left(\bm{u}\Theta_{1}\right)$ $\displaystyle=\kappa\Delta\Theta_{1}+s(\bm{x})/4-\gamma\Theta_{1}+\gamma\Theta_{2}/2$ (50) $\displaystyle 0$ $\displaystyle=\kappa\Delta\Theta_{2}+s(\bm{x})/2-\gamma\Theta_{2}+\gamma\Theta_{1}+\gamma\Theta_{3}$ (51) $\displaystyle-\nabla\cdot\left(\bm{u}\Theta_{3}\right)$ $\displaystyle=\kappa\Delta\Theta_{3}+s(\bm{x})/4-\gamma\Theta_{3}+\gamma\Theta_{2}/2$ (52) Similar to before we define a transformation using the eigenvectors of the generator $\mathcal{Q}$, $\displaystyle\begin{bmatrix}1/4&1/2&1/4\\\ 1/2&0&-1/2\\\ 1/4&-1/2&1/4\end{bmatrix}\begin{bmatrix}\varphi_{1}\\\ \varphi_{2}\\\ \varphi_{3}\end{bmatrix}=\begin{bmatrix}\Theta_{1}\\\ \Theta_{2}\\\ \Theta_{3}\end{bmatrix}\Leftrightarrow\begin{bmatrix}\varphi_{1}\\\ \varphi_{2}\\\ \varphi_{3}\end{bmatrix}=\begin{bmatrix}1&1&1\\\ 1&0&-1\\\ 1&-1&1\end{bmatrix}\begin{bmatrix}\Theta_{1}\\\ \Theta_{2}\\\ \Theta_{3}\end{bmatrix}$ (53) The resulting equations are $\displaystyle\nabla\cdot\left(\bm{u}\varphi_{2}\right)$ $\displaystyle=\kappa\Delta\varphi_{1}+s(\bm{x})$ (54) $\displaystyle\frac{1}{2}\bm{u}\cdot\nabla\varphi_{1}+\frac{1}{2}\bm{u}\cdot\nabla\varphi_{3}$ $\displaystyle=\kappa\Delta\varphi_{2}-\gamma\varphi_{2}$ (55) $\displaystyle\bm{u}\cdot\nabla\varphi_{2}$ $\displaystyle=\kappa\Delta\varphi_{3}-2\gamma\varphi_{3}.$ (56) We again comment that $\varphi_{1}=\langle\theta\rangle$ and that $\varphi_{2}$ and $\varphi_{3}$ are thought of as perturbation variables. Furthermore the turbulent flux is $\langle\bm{u}^{\prime}\theta^{\prime}\rangle=\bm{u}\varphi_{2}$. We eliminate dependence on the perturbation variables $\varphi_{2}$ and $\varphi_{3}$ by first solving for $\varphi_{3}$ in terms of $\varphi_{2}$, $\displaystyle\varphi_{3}$ $\displaystyle=\left(\kappa\Delta-2\gamma\right)^{-1}\bm{u}\cdot\nabla\varphi_{2}$ (57) and then solving for $\varphi_{2}$ in terms of $\varphi_{1}$, $\displaystyle\varphi_{2}$ $\displaystyle=\left(\kappa\Delta-\gamma-\frac{1}{2}\bm{u}\cdot\nabla\left(\kappa\Delta-2\gamma\right)^{-1}\bm{u}\cdot\nabla\right)^{-1}\frac{1}{2}\bm{u}\cdot\nabla\varphi_{1}$ (58) And finally we write our equation for the ensemble mean as $\displaystyle\nabla\cdot\left(\bm{u}\left(\kappa\Delta-\gamma-\frac{1}{2}\bm{u}\cdot\nabla\left(\kappa\Delta-2\gamma\right)^{-1}\bm{u}\cdot\nabla\right)^{-1}\frac{1}{2}\bm{u}\cdot\nabla\langle\theta\rangle\right)$ $\displaystyle=\kappa\Delta\langle\theta\rangle+s(\bm{x})$ (59) from whence we extract the turbulent diffusivity operator $\displaystyle\int d\bm{x}^{\prime}\mathcal{K}(\bm{x}|\bm{x}^{\prime})\bullet$ $\displaystyle=\bm{u}\left(\gamma-\kappa\Delta+\frac{1}{2}\bm{u}\cdot\nabla\left(\kappa\Delta-2\gamma\right)^{-1}\bm{u}\cdot\nabla\right)^{-1}\frac{1}{2}\bm{u}.$ (60) We see that, unlike the two-state system, the $\kappa\rightarrow 0$ limit retains a nonlocal feature since $\displaystyle\int d\bm{x}^{\prime}\mathcal{K}(\bm{x}|\bm{x}^{\prime})\bullet\rightarrow\bm{u}\left(\gamma-\frac{1}{4\gamma}(\bm{u}\cdot\nabla)(\bm{u}\cdot\nabla)\right)^{-1}\frac{1}{2}\bm{u}$ (61) and the operator $(\bm{u}\cdot\nabla)(\bm{u}\cdot\nabla)$ can have a significant spatial structure. Similar to before, larger transition rates imply increasing local structures. In the $\kappa\rightarrow 0$ limit the non- dimensional parameter of interest is the characteristic timescale of the steady flow field, $LU^{-1}$, as it compares to the characteristic timescale for transitioning, $\gamma^{-1}$. For the last case, we emphasize the algebraic structure of Equations 11-12 by working in detail through a four-state system. ### 3.3 Four State Here we will consider a two-dimensional velocity field motivated by Flierl & McGillicuddy (2002). The flow is two-dimensional, periodic with $x\in[0,2\pi)$, and wall-bounded with $z\in[-1,1]$. Our Markov velocity states are defined through the stream-functions $\displaystyle\psi_{1}$ $\displaystyle=\sin(x)\cos\left(\frac{\pi}{2}z\right)\text{, }\psi_{2}=\cos(x)\cos\left(\frac{\pi}{2}z\right)\text{, }$ (62) $\displaystyle\psi_{3}$ $\displaystyle=-\sin(x)\cos\left(\frac{\pi}{2}z\right)\text{, and }\psi_{4}=-\cos(x)\cos\left(\frac{\pi}{2}z\right),$ (63) and the corresponding velocity states are $\bm{u}_{m}=(\partial_{z}\psi_{m},-\partial_{x}\psi_{m})$. The states are simply $\pi/2$ phase shifts of one another along the periodic $x$ direction and are thought of as describing a cellular flow that randomly propagates through a channel. The generator is given by $\displaystyle\mathcal{Q}$ $\displaystyle=\gamma\begin{bmatrix}-1&1/2&0&1/2\\\ 1/2&-1&1/2&0\\\ 0&1/2&-1&1/2\\\ 1/2&0&1/2&-1\end{bmatrix}+\omega\begin{bmatrix}0&1/2&0&-1/2\\\ -1/2&0&1/2&0\\\ 0&-1/2&0&1/2\\\ 1/2&0&-1/2&0\end{bmatrix}.$ (64) The parameter $\gamma$ is, as in the previous examples, related to the amount of time spent in a given state. When $\omega=0$ the cells move with equal likelihood to the left or the right, whereas $\omega\neq 0$ biases the cells to move in a particular direction. We require $|\omega|\leq\gamma$ for a probabilistic interpretation of results. We assume that the system is in a statistically steady state. Thus we start by observing that the normalized right eigenvectors of $\mathcal{Q}$ are $\displaystyle\bm{v}_{1}$ $\displaystyle=\begin{bmatrix}1/4\\\ 1/4\\\ 1/4\\\ 1/4\end{bmatrix}\text{ , }\bm{v}_{2}=\begin{bmatrix}1/2\\\ \iota/2\\\ -1/2\\\ -\iota/2\end{bmatrix}\text{ , }\bm{v}_{3}=\begin{bmatrix}1/2\\\ -\iota/2\\\ -1/2\\\ \iota/2\end{bmatrix}\text{ , and }\bm{v}_{4}=\begin{bmatrix}1/4\\\ -1/4\\\ 1/4\\\ -1/4\end{bmatrix}$ (65) with corresponding eigenvectors $\lambda_{1}=0,\lambda_{2}=-\gamma-\omega\iota,\lambda_{3}=-\gamma+\omega\iota,\lambda_{4}=-2\gamma$, respectively. Inspection of $\bm{v}^{1}$, the vector associated with the statistically steady state, reveals that each state is equally likely in the steady state. The steady-state equations are $\displaystyle\bm{u}_{1}\cdot\nabla\Theta_{1}$ $\displaystyle=\kappa\Delta\Theta_{1}+s/4-\gamma\Theta_{1}+\frac{\gamma+\omega}{2}\Theta_{2}+\frac{\gamma-\omega}{2}\Theta_{4}$ (66) $\displaystyle\bm{u}_{2}\cdot\nabla\Theta_{2}$ $\displaystyle=\kappa\Delta\Theta_{2}+s/4-\gamma\Theta_{2}+\frac{\gamma+\omega}{2}\Theta_{3}+\frac{\gamma-\omega}{2}\Theta_{1}$ (67) $\displaystyle\bm{u}_{3}\cdot\nabla\Theta_{3}$ $\displaystyle=\kappa\Delta\Theta_{3}+s/4-\gamma\Theta_{3}+\frac{\gamma+\omega}{2}\Theta_{4}+\frac{\gamma-\omega}{2}\Theta_{2}$ (68) $\displaystyle\bm{u}_{4}\cdot\nabla\Theta_{4}$ $\displaystyle=\kappa\Delta\Theta_{4}+s/4-\gamma\Theta_{4}+\frac{\gamma+\omega}{2}\Theta_{1}+\frac{\gamma-\omega}{2}\Theta_{3}.$ (69) As was done in prior sections we will change basis by utilizing the eigenvectors of $\mathcal{Q}$. We avoid the use of imaginary eigenvectors by instead using linear combinations $(v^{2}+v^{3})/2$ and $(v^{2}-v^{3})/(2\iota)$ at the cost of a non-diagonal generator in the resulting basis. Explicitly we use the similarity transformation defined by the matrices $\displaystyle S=\begin{bmatrix}1/4&1/2&0&1/4\\\ 1/4&0&1/2&-1/4\\\ 1/4&-1/2&0&1/4\\\ 1/4&0&-1/2&-1/4\end{bmatrix}\text{ and }S^{-1}=\begin{bmatrix}1.0&1.0&1.0&1.0\\\ 1.0&0.0&-1.0&0.0\\\ 0.0&1.0&0.0&-1.0\\\ 1.0&-1.0&1.0&-1.0\end{bmatrix}.$ (70) Thus we make the change of variables $\displaystyle\begin{bmatrix}1.0&1.0&1.0&1.0\\\ 1.0&0.0&-1.0&0.0\\\ 0.0&1.0&0.0&-1.0\\\ 1.0&-1.0&1.0&-1.0\end{bmatrix}\begin{bmatrix}\Theta_{1}\\\ \Theta_{2}\\\ \Theta_{3}\\\ \Theta_{4}\end{bmatrix}=\begin{bmatrix}\varphi_{1}\\\ \varphi_{2}\\\ \varphi_{3}\\\ \varphi_{4}\end{bmatrix}.$ (71) We observe $\varphi_{1}=\langle\theta\rangle$. With this change of variable and using $\bm{u}_{1}=-\bm{u}_{3}$ with $\bm{u}_{2}=-\bm{u}_{4}$, the resulting system of equations is written in block operator form as, $\displaystyle\begin{bmatrix}-\kappa\Delta&\nabla\cdot\left(\bm{u}_{1}\bullet\right)&\nabla\cdot\left(\bm{u}_{2}\bullet\right)&0\\\ 0&\gamma-\kappa\Delta&-\omega&\frac{1}{2}\bm{u}_{1}\cdot\nabla\\\ 0&\omega&\gamma-\kappa\Delta&-\frac{1}{2}\bm{u}_{2}\cdot\nabla\\\ 0&\bm{u}_{1}\cdot\nabla&-\bm{u}_{2}\cdot\nabla&2\gamma-\kappa\Delta\end{bmatrix}\begin{bmatrix}\varphi_{1}\\\ \varphi_{2}\\\ \varphi_{3}\\\ \varphi_{4}\end{bmatrix}$ $\displaystyle=\begin{bmatrix}s(x,z)\\\ -\frac{1}{2}\bm{u}_{1}\cdot\nabla\varphi_{1}\\\ -\frac{1}{2}\bm{u}_{2}\cdot\nabla\varphi_{1}\\\ 0\end{bmatrix}$ (72) where the dependence of the perturbation variables on $\varphi_{1}$ is included as a source rather than as a part of the block operator. The inverse of the $3\times 3$ lower right submatrix matrix of Equation 72, $\displaystyle\mathcal{G}=\begin{bmatrix}\gamma-\kappa\Delta&-\omega&\frac{1}{2}\bm{u}_{1}\cdot\nabla\\\ \omega&\gamma-\kappa\Delta&-\frac{1}{2}\bm{u}_{2}\cdot\nabla\\\ \bm{u}_{1}\cdot\nabla&-\bm{u}_{2}\cdot\nabla&2\gamma-\kappa\Delta\end{bmatrix}^{-1}$ (73) is the Green’s function associated with the perturbation variables $\varphi_{2}$, $\varphi_{3}$, $\varphi_{4}$. The perturbation Green’s function is used to represent the turbulent diffusivity operator as $\displaystyle\int d\bm{x}^{\prime}\mathcal{K}(\bm{x}|\bm{x}^{\prime})\bullet$ $\displaystyle=\begin{bmatrix}\bm{u}_{1}&\bm{u}_{2}&0\end{bmatrix}\begin{bmatrix}\gamma-\kappa\Delta&-\omega&\frac{1}{2}\bm{u}_{1}\cdot\nabla\\\ \omega&\gamma-\kappa\Delta&-\frac{1}{2}\bm{u}_{2}\cdot\nabla\\\ \bm{u}_{1}\cdot\nabla&-\bm{u}_{2}\cdot\nabla&2\gamma-\kappa\Delta\end{bmatrix}^{-1}\begin{bmatrix}\bm{u}_{1}/2\\\ \bm{u}_{2}/2\\\ 0\\\ \end{bmatrix}$ (74) Written this way, we emphasize the Schur-complement structure of the ensemble mean equations when eliminating dependence on the perturbation variables. To better understand the turbulent diffusivity operator for this example, we consider two local approximations. Assuming a scale separation we approximate $\displaystyle\int\mathcal{K}(\bm{x}|\bm{x}^{\prime})\cdot\nabla\langle\theta\rangle(\bm{x}^{\prime})d\bm{x}^{\prime}\approx\int\mathcal{K}(\bm{x}|\bm{x}^{\prime})d\bm{x}^{\prime}\cdot\nabla\langle\theta\rangle(\bm{x})=\bm{D}_{1}\cdot\nabla\langle\theta\rangle(\bm{x})$ (75) where we integrate the turbulent diffusivity kernel with respect to $\bm{x}^{\prime}$ $\displaystyle\bm{D}_{1}$ $\displaystyle\equiv\int\mathcal{K}(\bm{x}|\bm{x}^{\prime})d\bm{x}^{\prime}.$ (76) We comment that, upon discretization, the linear operator $\int d\bm{x}^{\prime}\mathcal{K}(\bm{x}|\bm{x}^{\prime})\bullet$ is represented as a matrix. Performing the integration with respect to $d\bm{x}^{\prime}$ amounts to performing a row sum on the matrix. A second local estimate of the turbulent diffusivity is obtained by neglecting the dissipation terms and perturbation gradients, i.e. $\displaystyle\begin{bmatrix}\gamma-\kappa\Delta&-\omega&\frac{1}{2}\bm{u}_{1}\cdot\nabla\\\ \omega&\gamma-\kappa\Delta&-\frac{1}{2}\bm{u}_{2}\cdot\nabla\\\ \bm{u}_{1}\cdot\nabla&-\bm{u}_{2}\cdot\nabla&2\gamma-\kappa\Delta\end{bmatrix}^{-1}\approx\begin{bmatrix}\gamma&-\omega&0\\\ \omega&\gamma&0\\\ 0&0&2\gamma\end{bmatrix}^{-1}.$ (77) Thus the local eddy diffusivity is $\displaystyle\bm{D}_{2}$ $\displaystyle=\begin{bmatrix}\bm{u}_{1}&\bm{u}_{2}&0\end{bmatrix}\begin{bmatrix}\gamma&-\omega&0\\\ \omega&\gamma&0\\\ 0&0&2\gamma\end{bmatrix}^{-1}\begin{bmatrix}\bm{u}_{1}/2\\\ \bm{u}_{2}/2\\\ 0\\\ \end{bmatrix}$ (78) $\displaystyle=\frac{1}{2(\gamma^{2}+\omega^{2})}\left[\gamma\left(\bm{u}_{1}\otimes\bm{u}_{1}+\bm{u}_{2}\otimes\bm{u}_{2}\right)+\omega\left(\bm{u}_{2}\otimes\bm{u}_{1}-\bm{u}_{1}\otimes\bm{u}_{2}\right)\right]$ (79) where we have interpreted products of vector fields as outer products. Using the velocity fields, $\displaystyle\bm{u}_{1}$ $\displaystyle=-\frac{\pi}{2}\sin(x)\sin\left(\frac{\pi}{2}z\right)\hat{x}-\cos(x)\cos\left(\frac{\pi}{2}z\right)\hat{z}$ (80) $\displaystyle\bm{u}_{2}$ $\displaystyle=-\frac{\pi}{2}\cos(x)\sin\left(\frac{\pi}{2}z\right)\hat{x}+\sin(x)\cos\left(\frac{\pi}{2}z\right)\hat{z},$ (81) we compute each outer product to obtain the components of the local turbulent diffusivity, $\displaystyle[\bm{D}_{2}]_{\hat{x}\otimes\hat{x}}$ $\displaystyle=\frac{\gamma\pi^{2}}{8(\gamma^{2}+\omega^{2})}\sin^{2}\left(\frac{\pi}{2}z\right)$ (82) $\displaystyle[\bm{D}_{2}]_{\hat{x}\otimes\hat{z}}$ $\displaystyle=-[\bm{D}_{2}]_{\hat{z}\otimes\hat{x}}=\frac{\omega\pi}{4(\gamma^{2}+\omega^{2})}\sin\left(\frac{\pi}{2}z\right)\cos\left(\frac{\pi}{2}z\right)$ (83) $\displaystyle[\bm{D}_{2}]_{\hat{z}\otimes\hat{z}}$ $\displaystyle=\frac{\gamma}{2(\gamma^{2}+\omega^{2})}\cos^{2}\left(\frac{\pi}{2}z\right)$ (84) In Section 4 we show, in general, the equivalence of neglecting dissipation terms and perturbation gradients, as was done in Equation 77, and estimating the diffusivity by computing the integrated auto-correlation of the statistically steady velocity field, $\displaystyle\bm{D}_{2}=\int_{0}^{\infty}\langle\bm{u}(\bm{x},t+\tau)\otimes\bm{u}(\bm{x},t)\rangle d\tau.$ (85) Figure 1: A comparison of two local diffusivity estimates. The analytic diffusivity estimate is in red and numerically computed diffusivity estimate uses blue dots. The z-axis is depth and the x-axis is the diffusivity amplitude. Here $\gamma=\omega=100$ and $\kappa=1/100$. Figure 2: A comparison of two local diffusivity estimates. The analytic diffusivity estimate is in red and numerically computed diffusivity estimate uses blue dots. The z-axis is depth and the x-axis is the diffusivity amplitude. Here $\gamma=\omega=1$ and $\kappa=1$. Figure 3: A representation of the turbulent diffusivity operator corresponding to the case in Figure 2. If the diffusivity was indeed local then each of the form matrices would only have a diagonal component. The x-z coordinate axis have been collapsed to a single index, hence the banded structure of the output. Thus the rows are the ”output” axis corresponding to the $(x,z)$ of $\int dx^{\prime}dz^{\prime}K(x,z|x^{\prime},z^{\prime})\bullet$ and the columns are the $(x^{\prime},z^{\prime})$ corresponding to input. The ordering is chosen such that each block-diagonal structure corresponds to a fixed $x_{i},x_{j}^{\prime}$ grid location. The similarity between different block rows follow from the periodicity of the $x$ coordinate. Here $\gamma=\omega=1$ and $\kappa=1$. We illustrate the local diffusivity calculation in two scenarios: 1. 1. $\gamma=\omega=100$ and $\kappa=1/100$ 2. 2. $\gamma=\omega=\kappa=1$ In the first case, we expect the local diffusivity estimate to work well, and the two diffusivity estimates to correspond to one another. In the latter case, there is no scale separation between transition rates, diffusive timescales, and advective timescales; we expect nonlocal effects to play a significant role. We discretize all operators using a collocation method as described by Trefethen (2000) to explore the nonlocality of the full turbulent flux operator, Equation 74. We use 65 Chebyshev modes in the wall-bounded direction and 8 Fourier modes in the periodic direction to approximate each operator, leading to a $1560\times 1560$ sized matrix representation of Equation 73. There are no significant changes upon halving or doubling the resolution in each direction. Furthermore, given the independence of the periodic direction in $\bm{D}_{2}$, we use the local estimate $\displaystyle\overline{\bm{D}}_{1}(z)$ $\displaystyle=\frac{1}{2\pi}\int\mathcal{K}(x,z|x^{\prime},z^{\prime})dx^{\prime}dz^{\prime}dx$ (86) for comparison between $\bm{D}_{1}$ and $\bm{D}_{2}$. For the first case, we compute the local diffusivities and display their result in Figure 1. We see that there is excellent agreement between the two approaches, except near the boundary in the $K^{xz}$ component of the diffusivity tensor. Here we note the influence of the homogenous Neumann boundary conditions in the diffusivity estimate. For the second case, we compute the local diffusivities and display their result in Figure 1. We see that there is poor agreement between the two approaches in each component of the diffusivity tensor. Given the lack of scale separation between the timescales of the problem, this comes as no surprise. To further explore the discrepancy, we show the full operator, $\int dx^{\prime}dz^{\prime}\mathcal{K}(x,z|x^{\prime},z^{\prime})\bullet$, in Figure 3. The current turbulent diffusivity operator is a four-dimensional object, which we represent with a two-dimensional heatmap. To do so, we flatten both the $x$ and $z$ dimensions into a single index and order them so that the $z$ values are sequential and $x$ values are separated upon the completion of a $z$-range. The structured pattern of the heatmap in Figure 3 is a consequence of the periodicity of the $x$ direction. There is a block structure in each of the components of the operator. These blocks correspond to the diffusivity operator at a fixed $x,x^{\prime}$ location and represent the variation in $z,z^{\prime}$. Thus the diagonal block component corresponds to the $x=x^{\prime}$ part of the operator. Each block row corresponds to a fixed $x$ value, and each block column corresponds to a fixed $x^{\prime}$ value. One can count 8 blocks appearing on a given row (and column), which corresponds to our choice of using 8 Fourier modes in the periodic direction. Each block row seems to be a periodic translation of one another; however, we emphasize that this is merely in appearance rather than actuality. The four states are not sufficient to guarantee translation invariance. The rich structure of the current operator stands in stark contrast to a local diffusivity operator. A local diffusivity operator is a diagonal matrix. We emphasize that the significant off-diagonal components imply that turbulent fluxes are not related solely to local gradients but must incorporate a weighted sum of gradients in a neighborhood of a given location. In the following section, we gather the approach used in the examples and generalize. ## 4 General Approach We have seen three examples that all follow a similar pattern: 1. 1. Compute the eigenvectors of the generator $\mathcal{Q}$. 2. 2. Transform the equations into a basis that diagonalizes $\mathcal{Q}$. 3. 3. Separate the mean equation from the perturbation equations. 4. 4. Solve for the perturbation variables in terms of the mean variable. Here we aim to gather the above procedure in the general case where we have access to the eigenvectors of $\mathcal{Q}$. Furthermore, in the last example, we claimed that the local turbulent diffusivity approximation as calculated by neglecting the effects of diffusion and perturbation gradients is equivalent to calculating the integrated auto-correlation of the Markov process. We justify that claim in Section 4.3. ### 4.1 Notation Let us establish a notation for the general procedure. We again let $\mathcal{Q}$ denote the generator with corresponding transition probability matrix $\mathscr{P}(\tau)$ given by the matrix exponential $\displaystyle\mathscr{P}(\tau)$ $\displaystyle=\exp(\tau\mathcal{Q}).$ (87) The entries of the matrix $[\mathscr{P}(\tau)]_{mn}$ denotes the transition probability of state $n$ to the state $m$. In each column of the transition matrix the sum of the entries is one. We assume a unique zero eigenvalue for $\mathcal{Q}$ with all other eigenvalues negative. We also assume that the eigenvalues can be ordered in such a way that they are decreasing, i.e. $\lambda_{1}=0$, $\lambda_{2}<0$ and $\lambda_{i}\leq\lambda_{j}$ for $i>j$ with $j\geq 2$. These choices result in a unique statistical steady state which we denote by the vector $\bm{v}_{1}$ with the property $\displaystyle\mathcal{Q}\bm{v}_{1}$ $\displaystyle=0\bm{v}_{1}\text{ and }\mathscr{P}(\tau)\bm{v}_{1}=\bm{v}_{1}\text{ for all }\tau.$ (88) and similarly for the left eigenvector, $\bm{w}_{1}$. We denote the entries of $\bm{v}_{1}$ and $\bm{w}_{1}$ by column vectors $\displaystyle\bm{v}_{1}$ $\displaystyle=\begin{bmatrix}P_{1}\\\ P_{2}\\\ \vdots\\\ P_{M}\end{bmatrix}\text{ and }\bm{w}_{1}=\begin{bmatrix}1\\\ 1\\\ \vdots\\\ 1\end{bmatrix}$ (89) where $M$ is the number of states. We assume that the eigenvector $\bm{v}_{1}$ is normalized such that $\sum_{m}P_{m}=1$. Consequently, $\bm{w}_{1}\cdot\bm{v}_{1}=\bm{w}_{1}^{T}\bm{v}_{1}=1$. We introduce unit vectors $\hat{\bm{e}}_{m}$ whose $m^{\prime}th$ entry is zero and all other entries are zero, e.g. $\displaystyle\hat{\bm{e}}_{1}$ $\displaystyle=\begin{bmatrix}1\\\ 0\\\ \vdots\\\ 0\end{bmatrix}\text{ , }\hat{\bm{e}}_{2}=\begin{bmatrix}0\\\ 1\\\ \vdots\\\ 0\end{bmatrix}\text{ and }\hat{\bm{e}}_{M}=\begin{bmatrix}0\\\ 0\\\ \vdots\\\ 1\end{bmatrix}.$ (90) Thus, $\bm{v}_{1}=\sum_{m}P_{m}\bm{\hat{e}}_{m}$, $\bm{w}_{1}=\bm{\hat{e}}_{m}$. Furthermore, $\bm{\hat{e}}_{m}\cdot\bm{v}_{1}=P_{m}$ and $\bm{\hat{e}}_{m}\cdot\bm{w}_{1}=1$ for each $m$. For the discussion that follows we will assume that the matrix $\mathcal{Q}$ has an eigenvalue decomposition. In general we denote the right eigenvectors of $\mathcal{Q}$ by $\bm{v}_{i}$ for $i=1,..,M$ and the left eigenvectors by $\bm{w}_{i}$ for $i=1,...,M$. These vectors are all associated with eigenvalues $\lambda_{i}$ for $i=1,...,M$ where $i=1$ denotes the unique eigenvalue $\lambda_{1}=0$. We recall that the left eigenvectors can be constructed from the right eigenvectors by stacking all the left eigenvectors in a matrix $V$, computing the inverse $V^{-1}$, and extracting the rows of the inverse. The aforementioned procedure guarantees the normalization $\bm{w}_{j}\cdot\bm{v}_{i}=\bm{w}_{i}^{T}\bm{v}_{j}=\delta_{ij}$. Thus we have the relations $\displaystyle\mathcal{Q}\bm{v}_{n}=\lambda_{n}\bm{v}_{n}\text{ and }\bm{w}^{T}_{n}\mathcal{Q}=\lambda_{n}\bm{w}^{T}_{n}.$ (91) With notation now in place, we observe that the operators $\mathcal{Q}$ and $\mathscr{P}(\tau)$ are characterized by their spectral decomposition $\displaystyle\mathcal{Q}$ $\displaystyle=\sum_{i}\lambda_{i}\bm{v}_{i}\bm{w}_{i}^{T}\text{ and }\mathscr{P}(\tau)=\sum_{i}e^{\tau\lambda_{i}}\bm{v}_{i}\bm{w}_{i}^{T}.$ (92) We remind the reader of the various use of ”P”s and their relation: 1. 1. $\mathbb{P}$ denotes a probability. 2. 2. $\mathscr{P}(\tau)$ denotes the transition probability matrix for a time $\tau$ in the future. 3. 3. $\mathcal{P}_{m}(t)$ denotes the probability of being in state $m$ at time $t$. The algebraic relation $\sum_{m}P_{m}(t+\tau)\bm{\hat{e}}_{m}=\mathscr{P}(\tau)\sum_{n}P_{n}(t)\bm{\hat{e}}_{n}$ holds. 4. 4. $P_{m}$ is the statistically steady probability of being found in state $m$. In the limit $\lim_{t\rightarrow\infty}\mathcal{P}_{m}(t)=P_{m}.$ We now introduce our Markov states as steady vector fields. The use of several vector spaces imposes a burden on notation: The vector spaces associated with Markov states, ensemble members, and the vector field $\bm{u}$. Instead of using overly decorated notation with an excessive number of indices, we introduce the convention that $\bm{u}$ will always belong to the vector space associated with the vector field, and all other vectors are associated with the vector space of Markov states. Effectively we let elements of our vector space associated with Markov states belong to a different algebra than real numbers. ### 4.2 The Spectral Representation With this notation now in place, the statistically steady equations $\displaystyle\nabla\cdot\left(\bm{u}_{m}\Theta_{m}\right)$ $\displaystyle=\kappa\Delta\Theta_{m}+P_{m}s+\sum_{n}\mathcal{Q}_{mn}\Theta_{n}$ (93) are represented as the matrix system $\displaystyle\sum_{m}\hat{\bm{e}}_{m}\nabla\cdot\left(\bm{u}_{m}\Theta_{m}\right)$ $\displaystyle=\sum_{m}\hat{\bm{e}}_{m}\kappa\Delta\Theta_{m}+s\bm{v}_{1}+\mathcal{Q}\left(\sum_{m}\hat{\bm{e}}_{m}\Theta_{m}\right)$ (94) where we made use of $\sum_{m}\bm{\hat{e}}_{m}P_{m}=\bm{v}_{1}$. We now re- express Equation 97 in terms of a basis that uses the eigenvectors of the transition matrix. Define components $\varphi_{n}$ by the change of basis formula $\displaystyle\sum_{m}\Theta_{m}\bm{\hat{e}}_{m}$ $\displaystyle=\sum_{n}\varphi_{n}\bm{v}_{n}\Leftrightarrow\sum_{n}\varphi_{n}\bm{\hat{e}}_{n}=\sum_{mn}(\bm{w}_{n}\cdot\bm{\hat{e}}_{m})\bm{\hat{e}}_{n}\Theta_{m}$ (95) We make the observation $\varphi_{1}=\langle\theta\rangle$. We have the following relations based on the general definitions of the left eigenvectors $\bm{w}_{n}$ and right eigenvectors $\bm{v}_{n}$, $\displaystyle\Theta_{n}$ $\displaystyle=\sum_{i}(\bm{\hat{e}}_{n}\cdot\bm{v}_{i})\varphi_{i}\text{ and }\varphi_{n}=\sum_{m}(\bm{w}_{n}\cdot\bm{\hat{e}}_{m})\Theta_{m}.$ (96) Multiplying Equation 97 by $\bm{w}^{T}_{j}$ and making use of Equation 96 we get $\displaystyle\sum_{n}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})\nabla\cdot\left(\bm{u}_{n}\Theta_{n}\right)$ $\displaystyle=\kappa\Delta\varphi_{j}+\delta_{1j}s+\lambda_{j}\varphi_{j}$ (97) $\displaystyle\Rightarrow$ $\displaystyle\sum_{n}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})\nabla\cdot\left(\bm{u}_{n}\left[\sum_{i}\bm{\hat{e}}_{n}\cdot\bm{v}_{i}\varphi_{i}\right]\right)$ $\displaystyle=\kappa\Delta\varphi_{j}+\delta_{1j}s+\lambda_{j}\varphi_{j}$ (98) $\displaystyle\Rightarrow$ $\displaystyle\sum_{in}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})(\bm{\hat{e}}_{n}\cdot\bm{v}_{i})\nabla\cdot\left(\bm{u}_{n}\varphi_{i}\right)$ $\displaystyle=\kappa\Delta\varphi_{j}+\delta_{1j}s+\lambda_{j}\varphi_{j}$ (99) We now wish to decompose Equation 99 into a mean equation, index $j=1$, and perturbation equations $j>1$. For the mean equation, we make use of the properties $\displaystyle\lambda_{1}=0\text{ , }\bm{w}_{1}\cdot\hat{\bm{e}}_{n}=1\text{ , and }\sum_{(i=1)n}(\bm{w}_{1}\cdot\hat{\bm{e}}_{n})(\bm{\hat{e}}_{n}\cdot\bm{v}_{i})\nabla\cdot\left(\bm{u}_{n}\varphi_{i}\right)=\nabla\cdot\langle\bm{u}\varphi_{1}\rangle$ (100) to arrive at (after changing summation index from $n$ to $m$), $\displaystyle\nabla\cdot\left(\langle\bm{u}\rangle\varphi_{1}\right)+\nabla\cdot\left[\sum_{(i\neq 1)m}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})\bm{u}_{m}\varphi_{i}\right]$ $\displaystyle=\kappa\Delta\varphi_{1}+s$ (101) We make the observation that the turbulent flux is $\displaystyle\langle\bm{u}^{\prime}\theta^{\prime}\rangle$ $\displaystyle=\sum_{(i\neq 1)m}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})\bm{u}_{m}\varphi_{i}.$ (102) The perturbation equations, indices $j>1$, are $\displaystyle\sum_{in}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})(\bm{\hat{e}}_{n}\cdot\bm{v}_{i})\nabla\cdot\left(\bm{u}_{n}\varphi_{i}\right)$ $\displaystyle=\kappa\Delta\varphi_{j}+\lambda_{j}\varphi_{j}\text{ for }j>1.$ (103) We isolate the dependence on the mean gradients by rearranging the above expression as follows for $j>1$ $\displaystyle\sum_{(i\neq 1)n}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})(\bm{\hat{e}}_{n}\cdot\bm{v}_{i})\nabla\cdot\left(\bm{u}_{n}\varphi_{i}\right)-\kappa\Delta\varphi_{j}-\lambda_{j}\varphi_{j}$ $\displaystyle=-\sum_{n}P_{n}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})\nabla\cdot\left(\bm{u}_{n}\varphi_{1}\right)$ (104) where we used $\bm{\hat{e}}_{n}\cdot\bm{v}_{i}=P_{n}$. Assuming that the operator on the left-hand side of Equation 104 is invertible, we introduce the Green’s function, $\mathcal{G}_{ij}$ to yield $\displaystyle\varphi_{i}=-\int d\bm{x}^{\prime}\sum_{(j\neq 1)n}\mathcal{G}_{ij}(\bm{x}|\bm{x}^{\prime})P_{n}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})\nabla_{\bm{x}^{\prime}}\cdot\left(\bm{u}_{n}(\bm{x}^{\prime})\varphi_{1}(\bm{x}^{\prime})\right)\text{ for }i\neq 1$ (105) Thus we represent our turbulent flux as $\displaystyle\langle\bm{u}^{\prime}\theta^{\prime}\rangle$ $\displaystyle=-\int d\bm{x}^{\prime}\sum_{(i\neq 1)(j\neq 1)mn}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})\bm{u}_{m}(\bm{x})\mathcal{G}_{ij}(\bm{x}|\bm{x}^{\prime})P_{n}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})\nabla_{\bm{x}^{\prime}}\cdot\left(\bm{u}_{n}(\bm{x}^{\prime})\varphi_{1}(\bm{x}^{\prime})\right).$ (106) For compressible flow, the eddy-flux depends on both the ensemble mean gradients and the ensemble mean value; otherwise, when each Markov state is incompressible, $\displaystyle\langle\bm{u}^{\prime}\theta^{\prime}\rangle$ $\displaystyle=-\int d\bm{x}^{\prime}\sum_{(i\neq 1)(j\neq 1)mn}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})\bm{u}_{m}(\bm{x})\mathcal{G}_{ij}(\bm{x}|\bm{x}^{\prime})P_{n}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})\bm{u}_{n}(\bm{x}^{\prime})\cdot\nabla_{\bm{x}^{\prime}}\varphi_{1}(\bm{x}^{\prime}),$ (107) in which case the turbulent diffusivity kernel is $\displaystyle\mathcal{K}(\bm{x}|\bm{x}^{\prime})$ $\displaystyle=\sum_{(i\neq 1)(j\neq 1)mn}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})\bm{u}_{m}(\bm{x})\mathcal{G}_{ij}(\bm{x}|\bm{x}^{\prime})P_{n}(\bm{w}_{j}\cdot\hat{\bm{e}}_{n})\bm{u}_{n}(\bm{x}^{\prime}).$ (108) The above expression completes the procedure that we enacted for the examples in Section 3. We now discuss local approximations to the turbulent diffusivity operator. ### 4.3 Local Approximation We start with the same local diffusivity approximation of Section 3.3 but using the spectral representation of Equations 11-12. In the perturbation equations, neglect the dissipation operator and perturbation gradients, e.g. only include index $i=1$, to yield the following reduction of Equation 103, $\displaystyle\sum_{n}(\bm{w}_{i}\cdot\hat{\bm{e}}_{n})P_{n}\left(\bm{u}_{n}\cdot\nabla\varphi_{1}\right)$ $\displaystyle=\lambda_{i}\varphi_{i}\text{ for }i>1$ (109) where we used $(\bm{\hat{e}}_{n}\cdot\bm{v}_{1})=P_{n}$ and have changed indices from $j$ to $i$. We solve for $\varphi_{i}$ for $i>1$ and focus on the perturbation flux term in Equation 101 $\displaystyle\langle\bm{u}^{\prime}\theta^{\prime}\rangle=\sum_{(i\neq 1)m}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})\bm{u}_{m}\varphi_{i}$ (110) to get the local turbulent-diffusivity estimate, $\displaystyle\sum_{(i\neq 1)m}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})\bm{u}_{m}\varphi_{i}$ $\displaystyle=\sum_{(i\neq 1)m}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})\bm{u}_{m}\left[\frac{1}{\lambda_{i}}\sum_{n}(\bm{w}_{i}\cdot\hat{\bm{e}}_{n})P_{n}\left(\bm{u}_{n}\cdot\nabla\varphi_{1}\right)\right]$ (111) $\displaystyle=\underbrace{\left[\sum_{(i\neq 1)mn}\frac{-1}{\lambda_{i}}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})(\bm{w}_{i}\cdot\bm{\hat{e}}_{n})\bm{u}_{m}\otimes P_{n}\bm{u}_{n}\right]}_{\bm{D}}\cdot(-\nabla\varphi_{1}).$ (112) We aim to show that the turbulent diffusivity from Equation 112 $\displaystyle\bm{D}$ $\displaystyle=\sum_{(i\neq 1)mn}\frac{-1}{\lambda_{i}}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})(\bm{w}_{i}\cdot\bm{\hat{e}}_{n})\bm{u}_{m}\otimes P_{n}\bm{u}_{n}$ (113) is equivalent to estimating the diffusivity by calculating the integral of the velocity perturbation autocorrelation in a statistically steady state, $\displaystyle\bm{D}=\int_{0}^{\infty}\langle\bm{u}^{\prime}(\bm{x},t+\tau)\otimes\bm{u}^{\prime}(\bm{x},t)\rangle d\tau$ (114) The above turbulent diffusivity is expected to work well in the limit that diffusive effects can be neglected and the velocity field transitions rapidly with respect to the advective timescale. Under such circumstances it is not unreasonable to think of velocity fluctuations as analogous to white noise with a given covariance structure. For example, letting $\bm{\xi}$ be a white noise process and $\bm{\sigma}$ be a variance vector, if $\displaystyle\bm{u}^{\prime}(\bm{x},t)\approx\bm{\sigma}(\bm{x})\xi\text{ where }\langle\xi(t+\tau)\xi(t)\rangle=\delta(\tau)$ (115) then a diffusivity is given by $\displaystyle\bm{D}(\bm{x})=\int_{0}^{\infty}\langle\bm{u}^{\prime}(\bm{x},t+\tau)\otimes\bm{u}^{\prime}(\bm{x},t)\rangle d\tau$ $\displaystyle=\bm{\sigma}(\bm{x})\otimes\bm{\sigma}(\bm{x}).$ (116) Indeed, we will show that the intuitive estimate, $\displaystyle\bm{D}(\bm{x})=\int_{0}^{\infty}\langle\bm{u}^{\prime}(\bm{x},t+\tau)\otimes\bm{u}^{\prime}(\bm{x},t)\rangle d\tau$ (117) does correspond to Equation 113. We begin with two observations. First, the statistically steady velocity field satisfies $\displaystyle\langle\bm{u}(\bm{x},t)\rangle=\sum_{m}P_{m}\bm{u}_{m}(\bm{x}),$ (118) where $\bm{u}_{m}(\bm{x})$ for each $m$ are the states of the Markov process. Second, recall that the vector $\mathscr{P}(\tau)\bm{\hat{e}}_{n}$ is a column vector of probabilities whose entries denote the probability of being found in state $m$ given that at time $\tau=0$ the probability of being found in state $n$ is one. Thus, the conditional expectation of $\bm{u}(\bm{x},t+\tau)$ given $\bm{u}(\bm{x},t)=\bm{u}_{n}(\bm{x})$ is $\displaystyle\langle\bm{u}(\bm{x},t+\tau)\rangle_{\bm{u}(\bm{x},t)=\bm{u}_{n}(\bm{x})}$ $\displaystyle=\left(\sum_{m}\bm{u}_{m}(\bm{x})\hat{\bm{e}}_{m}\right)^{T}\mathscr{P}(\tau)\bm{\hat{e}}_{n}$ (119) $\displaystyle=\sum_{im}e^{\tau\lambda_{i}}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})(\bm{w}_{i}\cdot\bm{\hat{e}}_{n})\bm{u}_{m}(\bm{x}).$ (120) Equation 119 expresses the conditional expectation as a weighted sum of Markov states $\bm{u}_{m}(\bm{x})$. We are now in a position to characterize the local turbulent-diffusivity estimate. The local turbulent-diffusivity is computed by taking the long time integral of a statistically steady flow field’s autocorrelation function, i.e. $\displaystyle\bm{D}(\bm{x})=\int_{0}^{\infty}\bm{R}(\bm{x},\tau)d\tau$ (121) where $\displaystyle\bm{R}(\bm{x},\tau)\equiv\langle\bm{u}(\bm{x},t+\tau)\otimes\bm{u}(\bm{x},t)\rangle-\langle\bm{u}(\bm{x},t+\tau)\rangle\otimes\langle\bm{u}(\bm{x},t)\rangle.$ (122) We calculate the second term under the statistically steady assumption of Equation 122, $\displaystyle\langle\bm{u}(\bm{x},t+\tau)\rangle\otimes\langle\bm{u}(\bm{x},t)\rangle=\left(\sum_{m}P_{m}\bm{u}_{m}(\bm{x})\right)\otimes\left(\sum_{n}P_{n}\bm{u}_{n}(\bm{x})\right).$ (123) For the first term of Equation 122 we decompose the expectation into conditional expectations, $\displaystyle\langle\bm{u}(\bm{x},t+\tau)\otimes\bm{u}(\bm{x},t)\rangle=\sum_{n}\langle\bm{u}(\bm{x},t+\tau)\otimes\bm{u}(\bm{x},t)\rangle_{\bm{u}(\bm{x},t)=\bm{u}_{n}(\bm{x})}P_{n}$ (124) Given that we are in a statistically steady state, we use Equation 119 to establish $\displaystyle\langle\bm{u}(\bm{x},t+\tau)\otimes\bm{u}(\bm{x},t)\rangle$ $\displaystyle=\sum_{n}\langle\bm{u}(\bm{x},t+\tau)\otimes\bm{u}(\bm{x},t)\rangle_{\bm{u}(\bm{x},t)=\bm{u}_{n}(\bm{x})}P_{n}$ (125) $\displaystyle=\sum_{imn}e^{\tau\lambda_{i}}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})(\bm{w}_{i}\cdot\bm{\hat{e}}_{n})\bm{u}_{m}\otimes P_{n}\bm{u}_{n}.$ (126) We isolate the $i=1$ index and use $\lambda_{1}=0$, $\bm{\hat{e}}_{m}\cdot\bm{v}_{1}=P_{m}$, and $\bm{w}_{1}\cdot\bm{\hat{e}}_{n}=1$ to arrive at $\displaystyle\sum_{mn}e^{\tau\lambda_{1}}(\bm{\hat{e}}_{m}\cdot\bm{v}_{1})(\bm{w}_{1}\cdot\bm{\hat{e}}_{n})\bm{u}_{m}\otimes P_{n}\bm{u}_{n}=\left(\sum_{m}P_{m}\bm{u}_{m}\right)\otimes\left(\sum_{n}P_{n}\bm{u}_{n}\right).$ (127) Equation 127 cancels with 123 so that in total we have the following characterization of Equation 122 $\displaystyle\bm{R}(\bm{x},\tau)$ $\displaystyle=\sum_{(i\neq 1)mn}e^{\tau\lambda_{i}}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})(\bm{w}_{i}\cdot\bm{\hat{e}}_{n})\bm{u}_{m}\otimes P_{n}\bm{u}_{n}.$ (128) Equation 128 is integrated to yield the local turbulent-diffusivity $\displaystyle\bm{D}(\bm{x})=\int_{0}^{\infty}\bm{R}(\bm{x},\tau)d\tau=\sum_{(i\neq 1)mn}\frac{-1}{\lambda_{i}}(\bm{\hat{e}}_{m}\cdot\bm{v}_{i})(\bm{w}_{i}\cdot\bm{\hat{e}}_{n})\bm{u}_{m}\otimes P_{n}\bm{u}_{n}$ (129) where we used $\lambda_{i}<0$ for $i>1$. A comparison of Equation 129 to Equation 113 reveals the correspondence. Thus we see that estimating the diffusivity through the velocity autocorrelation integral is equivalent to neglecting diffusive effects and perturbation gradients. ## 5 Conclusions We have introduced a class of stochastic partial differential equations amenable to analysis in this work. The class of problems falls under the umbrella of stochastic advection, where the flow state is modeled as a continuous time Markov process. We reformulated the problem of finding a turbulence closure for passive scalars advected by a stochastic flow field into solving a set of partial differential equations by conditionally averaging the passive scalar equation with respect to the flow state. The resulting dimensionality of the equations depended on the number of variables required to describe flow statistics and the dimensionality of the flow. A flow characterized by $m$ discrete variables leads to a set of $m$-coupled equations of the same dimensionality as the original. A flow characterized by a continuum of statistical variables can be discretized and reduced to the former. Eliminating the system’s dependence on all but the ensemble mean leads to an operator characterization of the turbulence closure, allowing for an exploration of closures that don’t invoke a scale separation hypothesis. We explored three examples of increasing complexity–Markov states characterized by two, three, and four states–and outlined a general approach to obtaining a closure based on the spectrum of the transition probability operator. In the examples, we examined the role of non-locality in determining a statistically steady turbulence closure. We calculated closures for all three systems and numerically evaluated a Green’s function for the four-state system. We also found the small velocity amplitude, weak scalar diffusivity, and fast transition rate limit reduce the closure to a spatially heterogeneous tensor acting on ensemble mean gradients. Furthermore, we related this tensor to the time-integrated auto-correlation of the stochastic flow field. We have not exhausted the number of examples offered by the formulation nor simplifications leading to analytically tractable results. Interesting future directions include using Markov states estimated directly from turbulence simulations, analyzing scale-separated flows, generalizing the advection- diffusion equation to reaction-advection-diffusion equations, and formulating optimal mixing problems. When the number of Markov states increases, the computational burden of estimating turbulent diffusivity operators becomes demanding; thus, there is a need to develop methods that exploit the structure of the problem as much as possible. Mathematically there are many challenges as well. All the arguments provided here are formal calculations, and the necessity of rigorous proofs remains. For example, a direct proof of the conditional averaging procedure is necessary. Ultimately, the goal is to reduce the stochastic-advection turbulence closure problem to one that can leverage theory from partial differential equations. [Supplementary data]Supplementary material and movies are available at ZENODO https://github.com/sandreza/StatisticalNonlocality [Acknowledgements]We would like to thank the 2018 Geophysical Fluid Dynamics summer school where much of this work was completed. We would also like to thank Tobias Bischoff, Simon Byrne, and Raffaele Ferrari for their encouragement and discussion with regards to the present manuscript. [Funding]Our work is supported by the generosity of Eric and Wendy Schmidt by recommendation of the Schmidt Futures program, and by the National Science Foundation under grant AGS-6939393. [Declaration of interests]The authors report no conflict of interest. ## Appendix A An Alternative Formal Derivation We wish to show that one can work directly with the continuous formulation of the advection-diffusion equations for the derivation of the conditional mean equations. Although we consider a finite (but arbitrarily large) number of Markov states here, considering a continuum follows mutatis mutandi. In Section 2 we wrote down the master equation for the discretized stochastic system as $\displaystyle\partial_{t}\rho_{m}$ $\displaystyle=\sum_{i}\frac{\partial}{\partial\theta^{i}}\left[\left(\sum_{jkc}A_{ijk}^{c}u^{k,c}_{m}\theta^{j}-\sum_{j}D_{ij}\theta^{j}-s^{i}\right)\rho_{m}\right]+\sum_{n}\mathcal{Q}_{mn}\rho_{n}.$ (130) We introduce the (spatial) volume element $\Delta\bm{x}_{i}$ to rewrite Equation 130 in the evocative manner, $\displaystyle\partial_{t}\rho_{m}$ $\displaystyle=\sum_{i}\Delta\bm{x}_{i}\frac{1}{\Delta\bm{x}_{i}}\frac{\partial}{\partial\theta^{i}}\left[\left(\sum_{jkc}A_{ijk}^{c}u^{k,c}_{m}\theta^{j}-\sum_{j}D_{ij}\theta^{j}-s^{i}\right)\rho_{m}\right]+\sum_{n}\mathcal{Q}_{mn}\rho_{n}.$ (131) We now take limits $\displaystyle\sum_{i}\Delta\bm{x}_{i}$ $\displaystyle\overset{``\text{lim}"}{=}\int d\bm{x},$ (132) $\displaystyle\frac{1}{\Delta\bm{x}_{i}}\frac{\partial}{\partial\theta^{i}}$ $\displaystyle\overset{``\text{lim}"}{=}\frac{\delta}{\delta\theta(\bm{x})},$ (133) $\displaystyle\sum_{jkc}A_{ijk}^{c}u^{k,c}_{m}\theta^{j}+\sum_{j}D_{ij}\theta^{j}-s^{i}$ $\displaystyle\overset{``\text{lim}"}{=}\bm{u}_{m}\cdot\nabla\theta-\kappa\Delta\theta-s,$ (134) to get the functional evolution equation for the probability density, $\displaystyle\partial_{t}\rho_{m}$ $\displaystyle=\int d\bm{x}\frac{\delta}{\delta\theta(\bm{x})}\left(\left[\bm{u}_{m}\cdot\nabla\theta-\kappa\Delta\theta-s\right]\rho_{m}\right)+\sum_{n}\mathcal{Q}_{mn}\rho_{n}$ (135) where $\bm{x}$ is a continuous index. As before we can derive the CM equations directly from the above equations. To do so we make the additional correspondence $\displaystyle d\bm{\theta}$ $\displaystyle\overset{``\text{lim}"}{=}\mathcal{D}[\theta].$ (136) We now define the same quantities as before, but using the field integral $\displaystyle\mathcal{P}_{m}$ $\displaystyle\equiv\int\mathcal{D}[\theta]\rho_{m}\text{ and }\Theta_{m}(\bm{y})\equiv\int\mathcal{D}[\theta]\theta(\bm{y})\rho_{m}.$ (137) We mention that the discrete indices $i,j,k$ from Equations 2 before get replaced by the continuous labels such as $\bm{x}$ and $\bm{y}$. We only make use of a few formal properties of the field integral, with direct correspondence the the $n$-dimensional integrals. We use linearity, i.e. for two mappings with compatible ranges $\mathcal{F}[\theta]$ and $\mathcal{H}[\theta]$, $\displaystyle\int\mathcal{D}[\theta]\left(\mathcal{F}[\theta]+\mathcal{H}[\theta]\right)$ $\displaystyle=\int\mathcal{D}[\theta]\mathcal{F}[\theta]+\int\mathcal{D}[\theta]\mathcal{H}[\theta]$ (138) We use the analogue to the divergence theorem, $\displaystyle\int\mathcal{D}[\theta]\int d\bm{x}\frac{\delta}{\delta\theta(\bm{x})}\left(\left[\bm{u}_{m}\cdot\nabla\theta-\kappa\Delta\theta-s\right]\rho_{m}\right)$ $\displaystyle=0$ (139) $\displaystyle\Leftrightarrow$ $\displaystyle\int d\bm{\theta}\nabla_{\bm{\theta}}\cdot(\bm{f}\rho)$ $\displaystyle=0$ (140) since the integral of a divergence should be zero if the probabilities vanish at infinity (i.e. that our tracer cannot have infinite values at a given point in space). We also make use of the integration by parts, i.e. for some functionals $\mathcal{F}$ and $\mathcal{H}$, $\displaystyle\int\mathcal{D}[\theta]\mathcal{H}\int d\bm{x}\frac{\delta}{\delta\theta(\bm{x})}\mathcal{F}$ $\displaystyle=-\int\mathcal{D}[\theta]\int d\bm{x}\frac{\delta\mathcal{H}}{\delta\theta(\bm{x})}\mathcal{F}$ (141) $\displaystyle\Leftrightarrow$ $\displaystyle\int d\bm{\theta}h\nabla_{\bm{\theta}}\cdot\bm{f}$ $\displaystyle=-\int d\bm{\theta}\left(\nabla_{\bm{\theta}}h\right)\cdot\bm{f}$ (142) And finally we also interchange sums and integrals, $\displaystyle\int\mathcal{D}[\theta](\Delta\theta)\rho_{m}$ $\displaystyle=\Delta\int\mathcal{D}[\theta]\theta\rho_{m}=\Delta\Theta_{m}$ (143) $\displaystyle\Leftrightarrow$ $\displaystyle\int d\bm{\theta}\left(\sum_{j}D_{\ell j}\theta^{j}\rho_{m}\right)$ $\displaystyle=\sum_{j}D_{\ell j}\int d\bm{\theta}\theta^{j}\rho_{m}=\sum_{j}D_{\ell j}\Theta^{j}_{m}.$ (144) We proceed similarly for the $\bm{u}_{m}\cdot\nabla$ term. We also use properties of the variational derivative such as, $\displaystyle\frac{\delta\theta(\bm{y})}{\delta\theta(\bm{x})}=\delta(\bm{x}-\bm{y})\Leftrightarrow\frac{\partial\theta^{\ell}}{\partial\theta^{i}}=\delta_{\ell i}.$ (145) Taken together one can directly obtain Equations 11 and 12 by first integrating Equation 135 with respect to $\mathcal{D}[\theta]$ to get $\displaystyle\partial_{t}\mathcal{P}_{m}$ $\displaystyle=\sum_{n}\mathcal{Q}_{mn}\mathcal{P}_{n}$ (146) and multiplying Equation 135 by $\theta(\bm{y})$ and then integrating with respect to $\mathcal{D}[\theta]$ to get $\displaystyle\partial_{t}\Theta_{m}(\bm{y},t)+\nabla_{\bm{y}}\cdot\left(\bm{u}_{m}(\bm{y})\Theta_{m}(\bm{y},t)-\kappa\nabla_{\bm{y}}\Theta_{m}(\bm{y},t)\right)$ $\displaystyle=s(\bm{y})\mathcal{P}_{m}+\sum_{n}\mathcal{Q}_{mn}\Theta_{n}(\bm{y},t).$ (147) In the above expression, removing explicit dependence of the position variable yields, $\displaystyle\partial_{t}\Theta_{m}+\nabla\cdot\left(\bm{u}_{m}\Theta_{m}-\kappa\nabla\Theta_{m}\right)$ $\displaystyle=sP_{m}+\sum_{n}\mathcal{Q}_{mn}\Theta_{n}.$ (148) Our reason for mentioning the above methodology is that it allows for expedited computations. There is no need to explicitly discretize, perform usual $n$-dimensional integral manipulations, and then take limits afterward. For example, computing the conditional two-moment equations defined by the variable $\displaystyle C_{m}(\bm{y},\bm{z},t)$ $\displaystyle\equiv\int\mathcal{D}[\theta]\theta(\bm{y})\theta(\bm{z})\rho_{m},$ (149) is obtained by multiplying Equation 135 by $\theta(\bm{y})$ and $\theta(\bm{z})$ and integrating with respect to $\mathcal{D}[\theta]$, $\displaystyle\partial_{t}C_{m}+\nabla_{\bm{y}}\cdot\left(\bm{u}_{m}(\bm{y})C_{m}-\kappa\nabla_{\bm{y}}C_{m}\right)+\nabla_{\bm{z}}\cdot\left(\bm{u}_{m}(\bm{z})C_{m}-\kappa\nabla_{\bm{z}}C_{m}\right)$ (150) $\displaystyle=s(\bm{z})\Theta_{m}(\bm{y})+s(\bm{y})\Theta_{m}(\bm{z})+\sum_{n}\mathcal{Q}_{mn}C_{n}.$ (151) In particular we note the source term on the right hand side and the appearance of the first conditional moment. In the derivation we used the product rule $\displaystyle\frac{\delta(\theta(\bm{y})\theta(\bm{z}))}{\delta\theta(\bm{x})}$ $\displaystyle=\delta(\bm{x}-\bm{y})\theta(\bm{z})+\delta(\bm{x}-\bm{z})\theta(\bm{y}).$ (152) If the advection-diffusion equation is $m$-dimensional and we have $N$ Markov states, the above equation is $2m+N$ dimensional. Indeed the equation for the $M^{\prime}th$ moment is a $M\times m+N$ dimensional partial differential equation. ## Appendix B A Heuristic Overview of the Master Equation and Discretizations In this section we provide an argument for the form of the master equation in the main text, Equation 130 in Section 2. Our starting point is Section B.1 where we use the Liouville equation for two continuous variables. We then apply the finite volume method to the Fokker-Planck equation of an Ornstein- Uhlenbeck process to derive the transition matrices used in the two-state and three-state systems in Section 3. We conclude with a formal argument for the use of discrete Markov states as an approximation to the compressible Euler equations in B.3. ### B.1 Two Variable System Suppose that we have an two variables $x,y\in\mathbb{R}$ governed by the equations, $\displaystyle\frac{dx}{dt}$ $\displaystyle=f(x)+\sqrt{2}\sigma\xi$ (153) $\displaystyle\frac{dy}{dt}$ $\displaystyle=g(x,y).$ (154) where $\xi$ is white noise. In this context we think of $x$ as being our flow field $\bm{u}$ and $y$ as the tracer $\theta$. The master equation implied by the dynamics is $\displaystyle\partial_{t}\rho$ $\displaystyle=-\partial_{x}\left(f(x)\rho-\sigma^{2}\partial_{x}\rho\right)-\partial_{y}\left(g(x,y)\rho\right).$ (155) We now discretize the equation with respect to the $x-$variable by partitioning space into non-overlapping cells, characterized by domains $\Omega_{m}$. First we start with the Fokker-Planck equation for $x$, which is independent of the $y-$variable, $\displaystyle\partial_{t}P$ $\displaystyle=-\partial_{x}\left(f(x)P-\sigma^{2}\partial_{x}P\right).$ (156) Observe the relation $\int\rho(x,y,t)dy=P(x,t)$. Define our coarse grained variable $\mathcal{P}_{m}$ as $\displaystyle\mathcal{P}_{m}$ $\displaystyle\equiv\int_{\Omega_{m}}P(x)dx$ (157) which is a probability. Thus the discretization of Equation 156 becomes $\displaystyle\partial_{t}\int_{\Omega_{m}}Pdx$ $\displaystyle=-\int_{\Omega_{m}}\partial_{x}\left(f(x)\rho-\sigma^{2}\partial_{x}\rho\right)dx$ (158) $\displaystyle\approx$ $\displaystyle\partial_{t}\mathcal{P}_{m}$ $\displaystyle=\sum_{n}\mathcal{Q}_{mn}\mathcal{P}_{n}$ (159) for a generator $\mathcal{Q}$ which we derive in B.2 with respect to a chosen numerical flux. Heuristically, going from Equation 158-159 is accomplished by observing that $\mathcal{P}_{m}$ is a probability and the operator $\mathcal{L}\equiv\partial_{x}\left(f(x)\bullet-\sigma^{2}\partial_{x}\bullet\right)$ is linear; thus, upon discretization, the operator is represented a matrix444It is, of course, possible to approximate using a nonlinear operator, but for simplicity we only consider the linear case. acting on the chosen coarse grained variables $\mathcal{P}_{n}$. The property $\sum_{m}\mathcal{Q}_{mn}=\bm{0}$ is the discrete conservation of probability. Going back to Equation 155, defining $\displaystyle\rho_{m}(y)$ $\displaystyle\equiv\int_{\Omega_{m}}\rho(x,y)dx,$ (160) introducing $x_{m}\in\Omega_{m}$, and performing the same discretization for the joint Markov system yields, $\displaystyle\partial_{t}\rho_{m}$ $\displaystyle=\sum_{n}\mathcal{Q}_{mn}\rho_{n}-\partial_{y}\left(g(x_{m},y)\rho_{m}\right),$ (161) where we used the approximation $\displaystyle\int_{\Omega_{m}}g(x,y)\rho(x,y)dx$ $\displaystyle\approx g(x_{m},y)\int_{\Omega_{m}}\rho(x,y)dx=g(x_{m},y)\rho_{m}(y).$ (162) The $x_{m}$ are the Markov states and the $\mathcal{Q}_{mn}$ serves as the specification for transitioning between different states. We also observe that one can simply start with the discrete states for $x$ and continuous variables for $y$ to directly obtain Equation 161 as was done in the main text. In what follows we give a concrete example of deriving a transition matrix $\mathcal{Q}$ from a finite-volume discretization of an Ornstein-Uhlenbeck (OU) process. We explicitly mention the kind of discretization that we use since retaining mimetic properties of the transition matrix $\mathcal{Q}$ is not guaranteed with other discretizations. Furthermore, using a finite volume discretization allows for the resulting discretization to be interpreted as a continuous time Markov process with a finite state space. ### B.2 Example Discretization Consider an Ornstein-Uhlenbeck process and the resulting Fokker-Planck equation, $\displaystyle\partial_{t}\rho$ $\displaystyle=-\partial_{x}\left(-x\rho-\partial_{x}\rho\right).$ (163) We discretize the above equation with $N+1$ cells, where $N=1$ and $N=2$ correspond to the two and three state systems, respectively. Using a finite volume discretization, we take our cells to be $\displaystyle\Omega_{m}$ $\displaystyle=[\Delta x\left(m-1/2-N/2\right),\Delta x\left(m+1/2-N/2\right)]$ (164) $\displaystyle\Delta x$ $\displaystyle=\frac{2}{\sqrt{N}}$ (165) for $m=0,1,...,N$. Our choice implies that cell centers (the discrete Markov states) are $\displaystyle x_{m}$ $\displaystyle=\Delta x(m-N/2)$ (166) for $m=0,...,N$ and cell faces are $\displaystyle x_{m}^{f}$ $\displaystyle=\Delta x(m-1/2-N/2)$ (167) for $m=0,...,N+1$. We define $\displaystyle\mathcal{P}_{m}$ $\displaystyle=\int_{\Omega_{m}}\rho dx\text{ and }\overline{\rho}_{m}\Delta x=\mathcal{P}_{m}.$ (168) Upon integrating with respect to the control volume we obtain, $\displaystyle\frac{d}{dt}\mathcal{P}_{m}$ $\displaystyle=-\left.\left(-x\rho-\partial_{x}\rho\right)\right|_{x=x_{m}^{f}}+\left.\left(-x\rho-\partial_{x}\rho\right)\right|_{x=x_{m+1}^{f}}$ (169) The numerical flux is chosen as follows, $\displaystyle\left.\left(-x\rho-\partial_{x}\rho\right)\right|_{x=x_{m}^{f}}$ $\displaystyle\approx-\frac{x_{m-1}\overline{\rho}_{m-1}+x_{m}\overline{\rho}_{m}}{2}-\frac{\overline{\rho}_{m}-\overline{\rho}_{m-1}}{\Delta x}$ (170) $\displaystyle=-\frac{x_{m-1}\mathcal{P}_{m-1}+x_{m}\mathcal{P}_{m}}{2\Delta x}-\frac{\mathcal{P}_{m}-\mathcal{P}_{m-1}}{(\Delta x)^{2}}$ (171) $\displaystyle=\frac{1}{2}\left((N-m+1)\mathcal{P}_{m-1}-m\mathcal{P}_{m}\right)$ (172) where we use the convention $\mathcal{P}_{-1}=\mathcal{P}_{N+1}=0$ so that boundaries, corresponding to indices $m=0$ and $m=N+1$, imply no flux conditions. Combining the flux estimates for both cell boundaries, the evolution equation for the probabilities $\mathcal{P}_{m}$ becomes $\displaystyle\partial_{t}\mathcal{P}_{m}$ $\displaystyle=\frac{1}{2}\left[(N-m+1)\mathcal{P}_{m-1}-N\mathcal{P}_{m}+(m+1)\mathcal{P}_{m+1}\right],$ (173) which implies the transition matrix $\displaystyle\mathcal{Q}_{mn}$ $\displaystyle=\frac{1}{2}\left(-N\delta_{mn}+n\delta_{(m+1)n}+(N-n)\delta_{(m-1)n}\right).$ (174) Equation 173 emphasize the row structure of the matrix whereas Equation 174 emphasizes the column structure. The steady state probability distribution is the binomial distribution555The continuous steady state distribution is a Normal distribution $\rho(x)=(2\pi)^{-1/2}\exp\left(-x^{2}/2\right)$. $\displaystyle P_{m}$ $\displaystyle=2^{-N}\binom{N}{m}.$ (175) Furthermore, the eigenvectors and eigenvalues of the matrix are in correspondence with the eigenfunctions and eigenvalues of the OU process as noted by Hagan et al. (1989). In particular, the cell centers, Equation 166, is a left eigenvector of $\mathcal{Q}_{mn}$ with eigenvalue $\lambda=-1$. This relation is useful for calculating the auto-correlation of the Markov process since Equation 119 only involves one eigenvalue. We used the transition matrix, Equation 174, in the construction of the two and three state systems. Similar to the construction in this section, four-state system transition matrix is obtained from discretizing a random walk in a periodic domain with a drift. The term proportional to $\gamma$ is attributed to diffusion and the term proportional to $\omega$ is attributed to drift. The resulting cell centers are then taken as the phase in the periodic direction of a fixed stream function. ### B.3 A Finite Volume Approximation in Function Space We start with the compressible Euler-Equations $\displaystyle\partial_{t}\rho+\nabla\cdot\left(\rho\bm{u}\right)$ $\displaystyle=0,$ (176) $\displaystyle\partial_{t}\rho\bm{u}+\nabla\cdot\left(\rho\bm{u}\otimes\bm{u}\right)+\nabla p$ $\displaystyle=0,$ (177) $\displaystyle\partial_{t}\rho e+\nabla\cdot\left(\bm{u}\left[\rho e+p\right]\right)$ $\displaystyle=0,$ (178) $\displaystyle p(\rho,\rho\bm{u},\rho e)$ $\displaystyle=p$ (179) where $\rho$ is density, $\rho\bm{u}$ is the momentum, $\bm{u}=\rho\bm{u}/\rho$ is the velocity, $\rho e$ is the total energy density, and $p$ is a thermodynamic pressure666For example, one could use the pressure for an ideal gas $p=(\gamma-1)(\rho e-\rho|\bm{u}|^{2}/2)$ with $\gamma=7/5$.. Here we introduce $Z$ as a probability density in function space for the state variables $S\equiv(\rho,\rho\bm{u},\rho e)$. In the notation of A, the evolution equation for the statistics $Z$ are $\displaystyle\partial_{t}Z$ $\displaystyle=\int d\bm{x}\left[\frac{\delta}{\delta\rho}\left(\nabla\cdot\left[\rho\bm{u}\right]Z\right)+\frac{\delta}{\delta\rho\bm{u}}\left(\nabla\cdot\left[\rho\bm{u}\otimes\bm{u}\right]Z+\nabla pZ\right)+\frac{\delta}{\delta\rho e}\left(\nabla\cdot\left(\bm{u}\left[\rho e+p\right]\right)Z\right)\right],$ (180) where we have suppressed the index $\bm{x}$ in the variational derivatives. Now consider a partition in function space into domains $\Omega_{m}$ and let $S_{m}$ denote a value of a state within the set $S_{m}$. In this case we define the probability as $\displaystyle\mathcal{P}_{m}$ $\displaystyle\equiv\int_{\Omega_{m}}\mathcal{D}[\rho]\mathcal{D}[\rho\bm{u}]\mathcal{D}[\rho e]Z.$ (181) In analogy with the calculations in Section B.2, integrating equation 180 with respect to a control volume $\Omega_{m}$ would result in an equation of the form $\displaystyle\partial_{t}\mathcal{P}_{m}=\sum_{n}\mathcal{Q}_{mn}\mathcal{P}_{n}$ (182) for some generator $\mathcal{Q}_{mn}$. The entries of the generator are functionals of the states $S_{m}\in\Omega_{m}$. Performing the necessary integrals and re-expressing it in this finite form is done indirectly through data-driven methods with time-series as in Klus et al. (2016), Fernex et al. (2021), or Maity et al. (2022). The difficulty of performing a discretization from first principles comes from choosing the subsets of function space to partition and carrying out the integrals in function space. Periodic orbits and fixed points of a flow serve as a natural skeleton for function space, but are typically burdensome to compute. We offer no solution, but hope that in the future such direct calculations are rendered tractable. In the meanwhile, indirect data-driven methods are the most promising avenue for the calculation of the generator $\mathcal{Q}$. ## References * Allawala & Marston (2016) Allawala, Altan & Marston, J. B. 2016 Statistics of the stochastically forced lorenz attractor by the fokker-planck equation and cumulant expansions. Phys. Rev. E 94, 052218. * Avellaneda & Majda (1991) Avellaneda, Marco & Majda, Andrew J. 1991 An integral representation and bounds on the effective diffusivity in passive advection by laminar and turbulent flows. Communications in Mathematical Physics 138 (2), 339–391. * Bhamidipati et al. (2020) Bhamidipati, Neeraja, Souza, Andre N. & Flierl, Glenn R. 2020 Turbulent mixing of a passive scalar in the ocean mixed layer. Ocean Modelling 149, 101615. * Farrell & Ioannou (2019) Farrell, B. F. & Ioannou, P. J. 2019 Statistical state dynamics: A new perspective on turbulence in shear flow. Zonal Jets Phenomenology, Genesis, and Physics. Ed. Boris Galpirin and Peter L. Read. Cambridge University Press 2019. pp. 380–400. * Fernex et al. (2021) Fernex, Daniel, Noack, Bernd R. & Semaan, Richard 2021 Cluster-based network modeling&#x2014;from snapshots to complex dynamical systems. Science Advances 7 (25), eabf5006, arXiv: https://www.science.org/doi/pdf/10.1126/sciadv.abf5006. * Flierl & McGillicuddy (2002) Flierl, G. R. & McGillicuddy, D. J. 2002 The Sea: Ideas and Observations on Progress in the Study of the Seas, , vol. 12, chap. Mesoscale and submesoscale physical-biological interactions, pp. 113–185. John Wiley and Sons. * Gallet & Ferrari (2020) Gallet, Basile & Ferrari, Raffaele 2020 The vortex gas scaling regime of baroclinic turbulence. Proceedings of the National Academy of Sciences 117 (9), 4491–4497, arXiv: https://www.pnas.org/doi/pdf/10.1073/pnas.1916272117. * Hagan et al. (1989) Hagan, Patrick S., Doering, Charles R. & Levermore, C. David 1989 Mean exit times for particles driven by weakly colored noise. SIAM Journal on Applied Mathematics 49 (5), 1480–1513, arXiv: https://doi.org/10.1137/0149090. * Hassanzadeh et al. (2014) Hassanzadeh, Pedram, Chini, Gregory P. & Doering, Charles R. 2014 Wall to wall optimal transport. Journal of Fluid Mechanics 751, 627–662. * Hopf (1952) Hopf, Eberhard 1952 Statistical hydromechanics and functional calculus. Indiana University Mathematics Journal 1, 87–123. * Klus et al. (2016) Klus, Stefan, Koltai, Péter & Schütte, Christof 2016 On the numerical approximation of the perron-frobenius and koopman operator. Journal of Computational Dynamics 3, 51 – 79. * Knobloch (1977) Knobloch, Edgar 1977 The diffusion of scalar and vector fields by homogeneous stationary turbulence. Journal of Fluid Mechanics 83 (1), 129–140. * Kraichnan (1968) Kraichnan, Robert H. 1968 Small‐scale structure of a scalar field convected by turbulence. The Physics of Fluids 11 (5), 945–953, arXiv: https://aip.scitation.org/doi/pdf/10.1063/1.1692063. * Lorenz (1963) Lorenz, E. N. 1963 Deterministic nonperiodic flow. Journal of the Atmospheric Sciences 20, 130–141. * Maity et al. (2022) Maity, Priyanka, Koltai, Péter & Schumacher, Jörg 2022 Large-scale flow in a cubic rayleigh&#x2013;b&#xe9;nard cell: long-term turbulence statistics and markovianity of macrostate transitions. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 380 (2225), 20210042, arXiv: https://royalsocietypublishing.org/doi/pdf/10.1098/rsta.2021.0042. * Pope (2011) Pope, Stephen B. 2011 Simple models of turbulent flows. Physics of Fluids 23 (1), 011301, arXiv: https://doi.org/10.1063/1.3531744. * Schneider et al. (2017) Schneider, Tapio, Lan, Shiwei, Stuart, Andrew & Teixeira, João 2017 Earth system modeling 2.0: A blueprint for models that learn from observations and targeted high-resolution simulations. Geophysical Research Letters 44 (24), 12,396–12,417, arXiv: https://agupubs.onlinelibrary.wiley.com/doi/pdf/10.1002/2017GL076101. * Tan et al. (2018) Tan, Zhihong, Kaul, Colleen M., Pressel, Kyle G., Cohen, Yair, Schneider, Tapio & Teixeira, João 2018 An extended eddy-diffusivity mass-flux scheme for unified representation of subgrid-scale turbulence and convection. Journal of Advances in Modeling Earth Systems 10 (3), 770–800, arXiv: https://agupubs.onlinelibrary.wiley.com/doi/pdf/10.1002/2017MS001162. * Taylor (1922) Taylor, G. I. 1922 Diffusion by continuous movements. Proceedings of the London Mathematical Society s2-20 (1), 196–212, arXiv: https://londmathsoc.onlinelibrary.wiley.com/doi/pdf/10.1112/plms/s2-20.1.196. * Thiffeault (2012) Thiffeault, Jean-Luc 2012 Using multiscale norms to quantify mixing and transport. Nonlinearity 25 (2), R1–R44. * Trefethen (2000) Trefethen, Lloyd N. 2000 Spectral Methods in MATLAB. Society for Industrial and Applied Mathematics, arXiv: https://epubs.siam.org/doi/pdf/10.1137/1.9780898719598. * Venturi et al. (2013) Venturi, D., Tartakovsky, D.M., Tartakovsky, A.M. & Karniadakis, G.E. 2013 Exact pdf equations and closure approximations for advective-reactive transport. Journal of Computational Physics 243, 323–343. * Weinstock (1969) Weinstock, Jerome 1969 Formulation of a statistical theory of strong plasma turbulence. The Physics of Fluids 12 (5), 1045–1058, arXiv: https://aip.scitation.org/doi/pdf/10.1063/1.2163666.
∎ 11institutetext: 1 Theoretical and Applied Mechanics, Northwestern University, 2145 Sheridan Rd, Evanston, 60208, Illinois, USA. 2 Department of Mechanical Engineering, Northwestern University, 2145 Sheridan Rd, Evanston, 60208, Illinois, USA. 3 Department of Mathematics, Princeton University, Washington Road, Princeton, 08540, New Jersey, USA. †Corresponding Author: Wing Kam Liu, 11email<EMAIL_ADDRESS> # Deep Learning Discrete Calculus (DLDC): A Family of Discrete Numerical Methods by Universal Approximation for STEM Education to Frontier Research Sourav Saha1 Chanwook Park2 Stefan Knapik2 Jiachen Guo1 Owen Huang3 Wing Kam Liu1,2 ###### Abstract The article proposes formulating and codifying a set of applied numerical methods, coined as Deep Learning Discrete Calculus (DLDC), that uses the knowledge from discrete numerical methods to interpret the deep learning algorithms through the lens of applied mathematics. The DLDC methods aim to leverage the flexibility and ever increasing resources of deep learning and rich literature of numerical analysis to formulate a general class of numerical method that can directly use data with uncertainty to predict the behavior of an unknown system as well as elevate the speed and accuracy of numerical solution of the governing equations for known systems. The article is structured in two major sections. In the first section, the building blocks of the DLDC methods are presented and deep learning structures analogous to traditional numerical methods such as finite difference and finite element methods are constructed with a view to incorporate these techniques in Science, Technology, Engineering, Mathematics (STEM) syllabus for K-12 students. The second section builds upon the building blocks of the previous discussion, and proposes new solution schemes for differential and integral equations pertinent to multiscale mechanics. Each section is accompanied with mathematical formulation of the numerical methods, analogous DLDC formulation, and suitable examples. ###### Keywords: Numerical Methods Discrete Calculus Kernel Learning Partial Differential Equation Convolution ## 1 Introduction The problems of engineering and physical science can be categorized into three types saha2021hierarchical , type 1 problems are the problems with limited physical understanding and a lot of experimental data, type 2 problems are those with incomplete physical understanding and some experimental data, and type 3 problems are those for which there is sufficient knowledge about the system but solving for the system response is computationally challenging. An example of type 1 problem is relating the spatio-temporal variation of temperature to the resulting mechanical properties in metal additive manufacturing xie2021mechanistic , a type 2 problem would be calibrating the heat source and other relevant models for computational fluid dynamics simulation for metal additive manufacturing process gan2021benchmark , and a type 3 problem would be relating the microstructure and fatigue life of a 3D printed metallic part xie2021mechanistic ; kafka2021image . Although the classification is not a strict one, it points to two facts: a) even with numerical modeling, aid of data is needed at some level, and b) by combining data science and numerical method one can cover the spectrum of solving for completely unknown phenomenon to challenging problems. Hence, the question naturally arises, is there a way we can come up with a new method that gives us the best of both worlds? Establishing this connection between numerical methods and deep learning is the most sought after prize in the literature nowadays. The quest started with solving the dynamic system directly using data science and deep learning methods sirignano2018dgm ; raissi2018deep ; han2017deep . The deep learning methods primarily include deep neural network yu2018deep , recurrent neural network xiao2018nonlinear , convolutional neural network zhu2018convolutional , and residual neural network kani2017dr . Other techniques from data science such as unsupervised learning are also heavily used to analyze the data and extract meaningful features that may or may not have explicit physical meaning (sometimes called the latent variables) karl2016deep ; chakraverty2017artificial . However, it was apparent that using directly deep learning methods results in lack of generalization magill2018neural . The neural networks are essentially highly non-linear interpolation functions with parameters being trained on observations liu2021mechanistic . Hence, to predict the behaviour of a system outside the training range becomes challenging. Moreover, the neural networks in such methods may become very complex and it is often impossible to interpret the inner working of the neural network in use. This results in a ”black box” method which we can just use without proper know-how. Finally, these methods are data-hungry i.e., it requires a lot of data to implement these. As a solution to the lack of generalization, combining the knowledge of mechanics with data science is proposed as Mechanistic Data Science (MDS) liu2021mechanistic . The knowledge of mechanics can be incorporated in selection of features and curing of data to be used as features, or, governing principles such as conservation of energy can be directly used as an optimization constraint. The later method is also termed as the Physics Informed Neural Network (PINN) raissi2019physics ; karniadakis2021physics ; pang2019fpinns . These methods have become extremely popular as these serve perfectly to solve type 3 problems krishnapriyan2021characterizing ; li2021physics ; leung2021nh . Moreover, these methods only partially solve the issue of the requirement of large datasets. However, neither of these theories solves the problem of lack of interpretation as the neural network is still there. Finally, for type 1 problems (where we only have data) the idea of adding governing principle does not work. Data science techniques to discover the underlying governing equation from data have been proposed brunton2016discovering ; kaiser2018sparse ; kaheman2020sindy . Another example is Dimension-Net which discovers the underlying non-dimensional numbers directly from data gan2021universal ; xie2021data . While these techniques are exciting, one needs to have some idea about the system behavior to use these techniques as these are primarily regression techniques working on a library of candidate functions. Therefore, there is still a need for a data science framework that can directly use the experimental data to solve for the dynamics of the system and at the same time interpret the technique. Efforts towards merging numerical methods with deep learning have gained increased attention in the last few years. The initial efforts were to establish the parallels between a subset of numerical methods with a subset of deep learning techniques chen2018neural ; zhang2021hierarchical . For example, how a recurrent neural network behaves like an ordinary differential equation niu2019recurrent such as wave equation hughes2019wave , or, how some finite difference techniques or multi-grid methods have parallels with convolutional neural network he2019mgnet . While these works are important as an interpretation of the deep learning techniques, they do not suggest how to improve the numerical solvers using data science or vice-versa. One of many early efforts that tries to leverage deep learning to improve existing numerical methods is called Hierarchical Deep Learning Finite Element Method (HiDeNN-FEM) zhang2021hierarchical where using custom neural networks different linear and non-linear interpolation shape functions are obtained. A recent breakthrough in interpretable general deep learning methods for solving partial differential equation came in the form of operator learning kovachki2021neural . Two major methods were proposed concurrently for operator learning called Fourier Neural Operator (FNO) li2020fourier , and DeepONet lu2019deeponet . There have been studies showing comparisons between these two techniques lu2022comprehensive . Keeping the similarities and differences aside, the philosophy of both of these methods are same: mapping functions from input space to output space instead of mapping data. Based on the same philosophy, several other structures are being proposed including graph kernel network and non-local kernel network anandkumar2020neural ; you2022nonlocal . Figure 1: The key features of the Deep Learning Discrete Calculus (DLDC) method. Aside from solving challenging problems in scientific and engineering research, contemporary deep learning methods have a huge pedagogical potential as well. In order to develop a competent workforce, big tech companies and entrepreneurs are now focusing on integrating artificial intelligence (AI) in high school curriculum tucker2020exploring . There is increasing need for skilled workforce over outstanding academic achievers in the industries. Naturally, the question arises how to train the young students in AI and at the same time, make the curriculum interesting for them. If the AI is introduced in a traditional way through teaching computer science, the diversity in the student body will become very difficult to attain. Due to the socioeconomic background of the students, direct training in machine learning algorithms may become very hard to attain. Thus untested introduction to AI for middle and high school students will create another bias. On the other end of the spectrum, the way mathematics, especially, calculus is taught in the high schools may become too abstract for some students to grasp. It is natural for the young students to be interested to learn a concept with a clear idea about how to implement it. Usually, in the school curriculum the calculus is taught through some textbook examples which often make the students disinterested in the topic wang2022perspective . This lack of interest results in reduced number of students choosing Science, Technology, Engineering, and Mathematics (STEM) majors. There have been a lot research touretzky2022artificial ; touretzky2022machine ; yin2022ai4all on how to incorporate data science and AI into high school education. However, most of these studies are at strategic level and do not provide a clear pathway to merge two broad fields of calculus and AI. It is the authors’ opinion that introducing AI through deep learning and interpretation of deep learning via elementary calculus is the most optimized way to teach calculus at the high school level. This has two-way benefit: one, implementing calculus through advanced computational tools with real-life data will be possible for young students. It will increase their interest and understanding on the basic calculus. Two, interpretability of deep learning algorithms can be improved if the algorithms are thought from the first principle. With the broad scope in mind, in this article, the authors propose an applied numerical method that aims to combine the mechanistic knowledge, classical numerical methods, and deep learning algorithms called Deep Learning Discrete Calculus (DLDC) (see, Figure 1). A formal definition of Deep Learning Discrete Calculus can be given as: The Deep Learning Discrete Calculus is a discipline to integrate fundamental definitions of calculus with numerical methods using deep learning neural network. Deep learning is an extremely powerful tool for feature extraction and prediction, but it requires massive and cleansed data sets to perform well. For many engineering applications, the collection of so much data is prohibitively expensive or even impossible, making traditional deep learning approaches nonviable. Incorporating discrete calculus and numerical method tools into neural network architectures can empower deep learning to address engineering problems where datasets may be small and noisy. Additionally, numerical method-informed deep learning frameworks sometimes have the potential to extrapolate beyond available observations, which is something traditional deep learning implementations are infamously incapable of. The synthesis of numerical analysis ideas with machine learning can reduce the expenses of collecting data, pre-processing data, and training models, while improving the interpretability, accuracy, and generalizability of deep neural networks. Figure 1 shows the key features of the DLDC method. The DLDC aims to become a general method for solving the three different types of problems discussed before. The article has two major parts. The first part aims to introduce the concept of DLDC through building blocks of calculus, such as, differential and integral calculus for STEM education. In later sections, the article discusses how to apply the DLDC methods to solve differential and integral equations, and how it has advantages over the traditional numerical methods. Each demonstration is accompanied by an example to further reinforce the ideas presented in the article. ## 2 Formulating DLDC for STEM Education Figure 2: A neural network model with input, hidden, and output layers. In this section, the basic mathematical foundations of the DLDC methods are presented. This section will be particularly useful for implementing in STEM education. Before going into too much detail, a brief discussion on neural network is required. For extensive discussion, the readers can refer to liu2021mechanistic . A neural network has three types of layers, input, hidden, and output. The input variables of interest go into input layer, and the output variables are obtained through the output layer. The hidden layer(s) takes input variable and pass those through a non-linear function, called activation function, with optimization parameters. Mathematically, if $x$ is the input variable, $\mathcal{A}$ is the activation function, $y$ is the output variable, the equation for a neural network can be written as, $\centering y=\mathcal{A}(Wx+b)\@add@centering$ (1) where $W$ is the weight, and $b$ is the bias. By varying the weight and bias to minimize the error between predicted output and training data, one can achieve an approximation of any linear and non-linear functional relationship. A schematic relationship showing the details of a neural network with one input, one hidden, and one output layer is shown in Figure 2. The convention of super and sub-scripts for weights and biases are shown in Figure 2. Following this convention, the output can be written as a function of the trainable weights, biases, and given input as $\centering y=\mathcal{A}\left[W_{1,1}^{2,3}(\mathcal{A}(W_{1,1}^{1,2}x+b_{1}^{2}))+W_{2,1}^{2,3}\mathcal{A}(W_{1,2}^{1,2}x+b_{2}^{2})\right]\@add@centering$ (2) The trainable $\bm{W}$ and $\bm{b}$ parameters are varied to minimize the following loss function: $Lossfunction=\frac{1}{P}\sum_{n=1}^{P}(y-y^{*})^{2}$ (3) ### 2.1 Differential Calculus Based on this basic definition of neural network, we proceed to show how to build a neural network from scratch to mimic the finite difference methods. Let us consider a function $f(x)$ as shown in Figure 2(a). If we want to have an approximation of first derivative of $f(x)$ at A, the forward difference method will give us chapra2011numerical , $\centering\frac{dy}{dx}=\frac{f(x_{j+1})-f(x_{j})}{x_{j+1}-x_{j}}\@add@centering$ (4) Here, the index $j$ indicates point A, and $j+1$ indicates the next point on the function. The central difference method gives, $\centering\frac{dy}{dx}=\frac{f(x_{j+1})-f(x_{j-1})}{x_{j+1}-x_{j-1}}\@add@centering$ (5) A higher order approximation of the first derivative involving three points looks like, $\centering\frac{dy}{dx}=\frac{-3f(x_{j})+4f(x_{j+1})-f(x_{j+2})}{x_{j+1}-x_{j-1}}\@add@centering$ (6) Therefore, from observation, we can simply write the general form of these equations as, $\centering\frac{dy}{dx}=\sum_{j=1}^{n}w_{j}f(x_{j})\@add@centering$ (7) Figure 3: (a) A schematic diagram showing a continuous function $f(x)$ and discrete sampling data, (b) The DLDC structure for the first derivative (forward difference method). This form for numerical differentiation is famously known as differential quadrature method in numerical analysis bellman1971differential ; bert1996differential ; shu2012differential . If we compare the form of Eqn. 7 with Eqn. 1, a clear parallel can be observed. The finite difference formulas can be obtained using linear activation functions with zero bias. Inspired from this observation, a first-order derivative equivalent neural network is proposed in Figure 2(b). In the figure, the weights and biases are expressed in index notation for convenience. In $W_{i,j}^{l-1,l}$, the subscripts $i$ and $j$ denotes the incoming neuron and outgoing neuron, respectively. The superscript $l$ identifies the layers. The index $l-1$ indicates the incoming layer and $l$ indicates the outgoing layer. The network has one hidden layer, rectified linear unit activation function, and zero bias. The weights between input and hidden layer are constrained to be 1, and the weights between the hidden and output layer (red marked) are optimized. All the activation functions are linear and biases are set to zero. The cost function can be written as, $\centering\frac{1}{N}\sum_{i=1}^{N}\left(\frac{dy}{dx}^{true}-\frac{dy}{dx}^{prediction}\right)^{2}_{i}\@add@centering$ (8) Here, $N$ is the sample size. It is interesting to note that after optimization, the trainable weights becomes $W_{1,1}^{2,3}=\frac{1}{x_{j+1}-x{j}}$ and $W_{2,3}^{2,1}=\frac{-1}{x_{j+1}-x{j}}$. This value resembles the forward difference method. If a neural network like this is trained once against data generated by a known function and it’s derivative, the trained network can be used to predict the first derivative of any function provided the sampling remains the same. This concept has been applied to formulate the first order derivative of different types of known function. Some of the results are presented in Figure 4. In case of Figure 4(a), the neural network is trained for a simple sinusoidal function in a range between $0-360$ degrees. With the trained neural network, we tried to predict the first-order derivative of cosine function which gives very good prediction. In the next example, see, Figure 4(b), a simple polynomial was used to generate training data for the neural network. Once the network is trained, the same network is applied to predict the derivative of a higher-order polynomial (Figure 4(c)) and a relatively complicated function (Figure 4(d)). Figure 4: The DLDC results for first-order derivative. (a) Performance for predicting the first derivative of cosine function. The training was done with sine function. (b) Training performance for simple polynomial, and (c), (d) performance of the network to predict derivatives of more complicated functions. ### 2.2 Integral Calculus The next building block is to solve an integration problem. Numerically, an integral of a function with respect to a variable is an estimation of area under the curve representing that function when plotted against the said variable. Based on this simple concept, numerical methods such as trapezoidal or Simpson’s method are in use. However, engineers are mostly interested to solve time dependent ordinary or partial differential equation. The first method that comes to mind for such computation is the Euler’s method. Let us suppose we want to solve an ordinary differential equation. $\centering\frac{dy}{dt}=f(y,t)\@add@centering$ (9) where $y$ is the dependent variable and $t$ is time. A DLDC equivalent structure for explicit Euler method and associated prediction is shown in Figure 5. The generalized alpha method for Euler’s method is: $\centering y^{n+1}=y^{n}+\Delta t\left([1-\alpha].f^{n}+[\alpha].f^{n+1}\right),\@add@centering$ (10) where, $n$ is the time step, $\Delta t$ is step size, and $\alpha$ is a factor that indicates an explicit method if the value is 0 and implicit method if the value is 1. The structure of the DLDC for Euler integration is exactly the same as forward difference method, except the inputs are the values of the integrand and dependent variable at the previous step and the output is the value of the dependent variable in the next time step. In the neural network shown in Fig. 5 (a), the biases are set to zero and activation functions are linear. The black weights are fixed to a value of 1 and the red weights are only allowed to train. The values of the red marked weights in the figure are the final values of these weights after training. It turns out that after the training, the weights finally get to a value that makes the output mimic the equation Eq. 10. Figure 5: (a) The DLDC structure for the explicit Euler formulation, (b) A comparison between the analytical and the DLDC prediction for the time integration. The numerical prediction in the example was made by solving $\frac{dy}{dt}=15t^{2}+8t$ with $y=0$ initial condition. The predicted and analytical solutions diverge slightly near 1 second. But this error is coming from the Euler method itself as the weights of the trained DLDC is exactly same as the explicit Euler method. Similar network can be constructed for implicit Euler method as well. Figure 6: The DLDC structure for 2-point Gauss quadrature method. The red marked terms are to be optimized during training. The next attempt was made to construct a DLDC method for Gauss quadrature integration method chapra2011numerical . In brief, the Gauss quadrature method can perform integration by converting integral to fixed limit between -1 to 1. The equations look like, $\centering\int_{a}^{b}f(x)dx=\int_{-1}^{1}f\left(\frac{b-a}{2}\xi+\frac{a+b}{2}\right)\frac{dx}{d\xi}d\xi,\@add@centering$ (11) $\centering\int_{-1}^{1}g(x)dx=\sum_{i=1}^{n}w_{i}g(x_{i}).\@add@centering$ (12) Here, $n$ is the number of quadrature points and $w_{i}$ are weights assigned to each point. $g(x_{i})$ is the value of the converted function at each quadrature points. The DLDC structure for the 2-point Gauss quadrature is shown in Figure 6. In order to train the network, a third order polynomial $g(x,\textbf{a})=a_{0}+a_{1}x+a_{2}x^{2}+a_{3}x^{3}$ is assumed and the corresponding Gauss integral $\int_{-1}^{1}g(x,\textbf{a})dx=2a_{0}+\frac{2}{3}a_{2}$. The training data is generated for different values of a and the integral. In the DLDC neural network, the activation function is rectified linear unit between input and first hidden layer and at the last layer, and linear for the second hidden layer. If we train the neural network with respect to $x_{1},x_{2},w_{11}^{34},w_{21}^{34}$ with following cost function, $\centering\frac{1}{N}\sum_{n=1}^{N}\left(\sigma_{1}^{4}(x_{1},x_{2},w_{11}^{34},w_{21}^{34};\textbf{a}^{n})-\int_{-1}^{1}g(x,\textbf{a})dx\right)^{2},\@add@centering$ (13) where $\centering\sigma_{1}^{4}(x_{1},x_{2},w_{11}^{34},w_{21}^{34};\textbf{a}^{n})=\sum_{j=1}^{N_{i-1}}w_{jk}^{i-1,i}\sigma_{j}^{i-1}+b_{k}^{i}=c_{1}g(x_{1};\textbf{a}^{n})+c_{2}g(x_{2},\textbf{a}^{n}),\@add@centering$ (14) after the training the quadrature coordinates $x_{1}$, $x_{2}$, and the weights $c_{1}$ and $c_{2}$ have exact same value as theoretical Gauss quadrature method. The structure was extended to 3 and 4-point rules and for all the cases the trained values for the co-ordinates and weights match well with the analytical value (see Figure 7). Just like the Euler method and finite difference approximation, the Gauss quadrature can be trained with a known integration can be used for any other integration without retraining. Figure 7: A table showing the comparison between different quadrature methods and optimized values for the DLDC. ## 3 DLDC Formulation for Frontier Research This section builds upon the ideas of the DLDC formulation for STEM education and proposes discrete calculus-inspired deep learning methods to solve partial differential equations and integral equations which are at the core of applied mechanics. ### 3.1 Solving Partial Differential Equations #### 3.1.1 Convolution Hierarchical Deep-learning Neural Network (C-HiDeNN) In this section, we use DLDC to formulate a new interpolation theory based on convolution, called Convolution Hierarchical Deep-learning Neural Network (C-HiDeNN). C-HiDeNN interpolants are smoother and more accurate than those of FEM. They are designed to have arbitrary reproducing polynomial orders, resulting in superior convergence behavior without using higher order elements. To understand C-HiDeNN, we should review HiDeNN-FEM first saha2021hierarchical ; zhang2021hierarchical . HiDeNN-FEM is a neural network representation of FEM. HiDeNN-FEM shape functions are small neural networks that are designed to be mathematically equivalent to FEM shape functions. Therefore, the weight and bias of HiDeNN-FEM shape functions are represented as a function of nodal coordinates and nodal field variables. The element-wise shape functions are then gathered to form a global neural network and finally, a loss function based on the principle of minimum potential energy is minimized. In HiDeNN-, the solution is obtained by optimizing nodal coordinates and nodal field variables with respect to the potential energy loss function. There are two solution schemes for the HiDeNN-FEM. First, the nodal coordinates are kept fixed and only the nodal variables are updated. This makes the HiDeNN-FEM mathematically equivalent to FEM, thus the solution accuracy and computation time of HiDeNN-FEM is on the same order as FEM. Second, both the nodal coordinates and nodal variables are updated, thus making it equivalent to r-adaptive FEM. For detailed discussions, readers may refer to zhang2021hierarchical . Figure 8: C-HiDeNN formulation for 1D Poisson problem. Light blue and orange terms are weights and biases of the neural network, respectively. If there is no weight or bias assigned for a neuron, it will have fixed negligible weight=1 and bias=0. Functions inside neurons with blue edges represent activation functions while those with black edges represent inputs (green color) and outputs (white color) of the neuron. (a) shows nodal coordinates in both physical and parametric space, focusing on the element of interest $e_{I}$. Nodal patch domains and other terminologies are defined below. (b) represents the C-HiDeNN shape function of element $e_{I}$, which constitutes the hierarchical DeNN layer of the global neural network (d). (c) is the convolution patch function that can be found at the greed dotted box in (b). This figure is taken from park2023convolution . Since HiDeNN-FEM is still based on finite element interpolations, the improvement of global solution accuracy is insignificant even with the automatic r-adaptivity. This prompted the idea of convolution to answer the question: ”Can we incorporate convolution filters to HiDeNN-FEM shape functions to achieve highly smooth and accurate interpolants?”. This idea is written as (in 1D): $u^{h,e}(\xi)=\sum_{i\in A^{e}}N_{i}(\xi)\sum_{j\in A^{i}_{s}}W^{x_{i}}_{a,p,j}\left(x^{h,e}(\xi)\right)u_{j}=\sum_{k\in A^{e}_{s}}\tilde{N}_{k}(\xi)u_{k},$ (15) with elementwise mapping between natural coordianates and physical coordinates: $x^{h,e}(\xi)=\sum_{i\in A^{e}}N_{i}(\xi)x_{i}.$ (16) In Eq. 15, two interpolants appear: 1) general polynomial interpolants, $N_{i}(\xi)$, defined over element $A^{e}$; 2) convolution patch functions, $W^{x_{i}}_{a,p,j}\left(x^{h,e}(\xi)\right)$, defined over nodal patch domain $A^{i}_{s}$. The elemental patch domain is visualized in Fig.8(a) where the integer-valued patch size $s$ refers to the number of element layers surrounding node $i$. That is, $A^{i=x_{I}}_{s=2}$ contains node $x_{I-2}$ through node $x_{I+2}$ while $A^{i=x_{I+1}}_{s=2}$ contains node $x_{I-1}$ through node $x_{I+3}$. The general polynomial interpolants $N_{i}(\xi)$ can be thought as any finite element shape functions that are compactly supported and satisfy Kronecker delta and partition of unity. The convolution patch functions $W^{x_{i}}_{a,p,j}\left(x^{h,e}(\xi)\right)$ are also compactly supported supported interpolation functions that satisfy Kronecker delta and reproducing conditions. Here, the dilation parameter $a$ determines the size of support domain and $p$ is the reproducing polynomial order. For the convolution patch functions, we borrow from well-developed meshfree techniques. In this study, the radial point interpolation method is adopted because this method returns stable interpolants that satisfy Kronecker delta property and reproducing conditions. For details, readers may refer to liu2001local ; liu2005introduction . Finally, the double summation in Eq. 15 is combined into a single summation over the elemental patch domain $A^{e}_{s}=\cup_{i\in A^{e}}A^{i}_{s}$, and the resulting convolution interpolants are $\tilde{N}_{k}(\xi)$ where $k\in A^{e}_{s}$. The convolution interpolants therefore satisfy compact supportness, partition of unity, Kronecker delta, and reproducing conditions, making it easy to apply boundary conditions. In terms of DLDC, the C-HiDeNN shape function can be written as a neural network illustrated in Fig.8(b). The first two hidden layers refer to the finite element shape functions, and the third hidden layer is for the convolution patch functions. Note that the number of neurons is determined by the number of nodes in nodal patch domains. That is, larger patch size $s$ leads to larger number of neurons or higher connectivity. This is analogous to the convolutional neural network (CNN) kernels. In C-HiDeNN, higher reproducing polynomial order $p$ is achieved by setting large patch size $s$. Based on our preliminary study, $s\geq p$ to get stable convolution patch functions. In other words, C-HiDeNN can achieve higher order $p$ by increasing the connectivity of nodes while using the same linear elements. This is completely different in FEM where higher order elements must be used to build higher order shape functions. Therefore, the global degrees of freedom (DOFs) of C-HiDeNN are the same as linear FEM, but the bandwidth of the global stiffness matrix becomes larger due to increased nodal connectivity. The weights of the third hidden layer in Fig.8(b) are built form another sub- neural network shown in Fig.8(c). This follows the radial point interpolation theory and details are discussed in park2023convolution . Finally, the elementwise shape functions are assembled to form a global neural network and potential energy loss function shown in Fig.8. To demonstrate superior accuracy of C-HiDeNN compared to FEM, a 2-D Poisson problem has been solved: $\begin{split}\nabla\cdot(\nabla u(\textbf{x}))+b_{f}(\textbf{x})&=0\quad\textrm{in}\ \Omega\\\ u&=0\quad\textrm{on}\ \Gamma\end{split}$ (17) where $b_{f}(\textbf{x})$ is the body force and a square domain $\Omega$ whose lower left corner is at $(0,0)$ and upper right cornder is at (10,10) is used. $\Gamma$ is the boundary of the domain $\Omega$. We set the analytical field variable as: $u(\textbf{x})=\frac{1}{625}(x^{2}-10x)(y^{2}-10y)\left(2e^{-2((x-3)^{2}+(y-3)^{2})}+e^{-2((x-7)^{2}+(y-7)^{2})}\right).$ (18) L2 and H1 error estimators are used: $\begin{split}\lVert e\rVert_{L_{2}}&=\lVert u-u^{h}\rVert_{L_{2}}=\frac{\left(\int_{\Omega}(u-u^{h})^{2}dx\right)^{1/2}}{\left(\int_{\Omega}u^{2}dx\right)^{1/2}}\\\ \lVert e\rVert_{H_{1}}&=\lVert u-u^{h}\rVert_{H_{1}}=\frac{\left(\int_{\Omega}(u-u^{h})^{2}dx+\int_{\Omega}\lVert\nabla u-\nabla u^{h}\rVert^{2}_{2}dx\right)^{1/2}}{\left(\int_{\Omega}u^{2}dx+\int_{\Omega}\lVert\nabla u\rVert^{2}_{2}dx\right)^{1/2}}\end{split}$ (19) Figure 9: Convergence plot. (a, b) are for $L_{2}$ norm error and (c, d) are for $H_{1}$ norm error estimation. (a, c) are the error vs. mesh size plots and (b, d) are the error vs. degrees of freedom plots. Numbers on the graph are the convergence rates (italic font for C-HiDeNN). For C-HiDeNN, patch size $s=3$ and dilation parameter $a=30$. The graphs are colored by the reproducing polynomial order $p$. FEM uses dashed lines and C-HiDeNN uses solid lines. Fig.9 shows C-HiDeNN-FEM can have arbitrary convergence rates depending on the reproducing polynomial order $p$. That is, for the same $p$, the convergence rates (slope of Fig.9(a,c)) of FEM and C-HiDeNN are asymptotically the same. However, their y-intercepts are different. When $p=1$, C-HiDeNN is around two orders of magnitude more accurate than FEM for a given mesh size. The gap decreases as $p$ goes up and the FEM turns around when $p=3$ (i.e., blue dotted curves are lower than blue solid curves in Fig.9(a,c)). However, this is not a fair comparison as higher order FEM uses higher order elements that has more DOFs while C-HiDeNN still uses the linear elements. The errors are plotted with respect to DOFs in Fig.9(b,d). These two plots show that C-HiDeNN curves are always lower than those of FEM for the same $p$. That is, for the same DOFs, C-HiDeNN is always more accurate than FEM. In other words, to achieve the same level of error, FEM needs more DOFs than C-HiDeNN for the same $p$. When $p=3$, for example, FEM requires 5 to 10 times more DOFs than C-HiDeNN-FEM for a given accuracy (see Fig.9(b,d)). It is important to note that the convolution operations over the local patch domains are the same as those performed in convolutional neural network (CNN). The larger the patch size $s$, the more neurons connectivity. The dilation parameter $a$ acts as a convolution operator, dictating the feature extraction of the data. Thus, many of the well-developed high-performance computing algorithms utilizing parallel architectures (such as GPU and TPU) in CNN can be integrated with C-HiDeNN. park2023convolution discusses in detail on how the parallel programming can accelerate C-HiDeNN computation and demonstrates that it can be as fast as commercial FEM software running on CPU. #### 3.1.2 Space Time Finite Element Method with DLDC This section proposes a DLDC strategy for solving partial differential equations based on a space-time finite element formulation hughes1988space ; hulbert1990space ; wang2017time . Space-time finite element methods are numerical discretization schemes to predict the spatiotemporal responses of dynamic systems. When used to approximate solutions to partial differential equations, they require assumed values of the underlying governing parameters. For structural problems, these parameters may correspond to material properties such as density, stiffness, and damping capacity. For many applications these assumptions are reasonable, but for many others the distribution of properties throughout a material domain is unknown. For the latter, experimental observation is required for the accurate calibration of model parameters. Consider an elastic one-dimensional bar governed by the following equation. $\centering EA\frac{\partial^{2}u}{\partial x^{2}}-\rho A\frac{\partial^{2}u}{\partial t^{2}}+f=0.\@add@centering$ (20) Here, $t$ is time, $x$ is location, $u$ is displacement, $E$ is elastic stiffness, $A$ is cross-sectional area, $\rho$ is density of the material, and $f$ is external excitation. The discretization of the bar must be done through the space and time dimensions simultaneously, and here it is done with linear quadrilateral elements. Details of the Galerkin formulation are included in the Appendix. The nodes are numbered counting upwards through space in the same order for each time instant, beginning with the first time instant. This results in a highly structured space-time stiffness matrix that will be exploited in the developed DLDC method. For this problem with one spatial dimension, $N$ nodes through space and $T$ nodes through time will result in a total of $N\times T$ degrees of freedom. Besides the enforcement of initial and boundary conditions, the stiffness matrix retains a diagonal structure such that the $N$ equations corresponding to a given time instant only have $3N$ non-zero coefficients. This corresponds to the displacement field at time $t$ being only directly related to the displacement fields at times $t-1$ and $t+1$. For a system discretized with linear quadrilateral elements, a piece of the matrix equation’s larger diagonal structure is depicted below with subscripts denoting sub-matrix size and superscripts denoting time step. $\centering\begin{bmatrix}\mathbf{A}_{N\times N}&\mathbf{B}_{N\times N}&\mathbf{C}_{N\times N}&0&0\\\ 0&\mathbf{A}_{N\times N}&\mathbf{B}_{N\times N}&\mathbf{C}_{N\times N}&0\\\ 0&0&\mathbf{A}_{N\times N}&\mathbf{B}_{N\times N}&\mathbf{C}_{N\times N}\end{bmatrix}\begin{bmatrix}\mathbf{u}_{N\times 1}^{t-2}\\\ \mathbf{u}_{N\times 1}^{t-1}\\\ \mathbf{u}_{N\times 1}^{t}\\\ \mathbf{u}_{N\times 1}^{t+1}\\\ \mathbf{u}_{N\times 1}^{t+2}\\\ \end{bmatrix}=\begin{bmatrix}\mathbf{f}_{N\times 1}^{t-1}\\\ \mathbf{f}_{N\times 1}^{t}\\\ \mathbf{f}_{N\times 1}^{t+1}\\\ \end{bmatrix}\@add@centering$ (21) Here, $\bf{A}$, $\bf{B}$, and $\bf{C}$ represent repeated sub-matrices within the global stiffness matrix. For the elastic bar governed by an equation with no damping term, it is noted that $\bf{A}$ and $\bf{C}$ are identical. The solution of this matrix equation can be done sequentially, solving only $N$ simultaneous equations at a time before progressing to the next time instant. These computations can be formulated as an autoregressive neural network. Autoregressive neural networks are sequential feed-forward models used to predict sequence values based on prior observations of the sequence taskaya2005comparative ; triebe2019ar-net . Generally, this can be represented by the following equation with $ARNN$ denoting the autoregressive neural network as a function of prior observation data, where $\mathbf{X}^{t}$ and $\bm{\varepsilon}^{t}$ represent the observation and prediction error respectively at step $t$ in a sequence of data. $\centering\mathbf{X}^{t}=ARNN(\mathbf{X}^{t-1},\mathbf{X}^{t-2},\mathbf{X}^{t-3},...,\mathbf{X}^{t-p})+\bm{\varepsilon}^{t}\@add@centering$ (22) Here, the autoregressive model would be considered of order $p$, since the previous $p$ observations are used to calculate a prediction. Predictions spanning multiple time steps may be achieved by using predicted sequence values as inputs for the prediction of further steps in the sequence. By inspection of the space-time finite element matrix equation 21, the following equivalent second order autoregressive neural network can be constructed. $\centering\mathbf{u}^{t+1}=ARNN(\mathbf{f}^{t},\mathbf{u}^{t},\mathbf{u}^{t-1})=\mathbf{C}^{-1}\left[\mathbf{f}^{t}-\begin{bmatrix}\mathbf{A}&\mathbf{B}\end{bmatrix}\begin{bmatrix}\mathbf{u}^{t-1}\\\ \mathbf{u}^{t}\end{bmatrix}\right]\@add@centering$ (23) At first glance, this re-framing of the space-time finite element formulation as a neural network provides no apparent advantage. The computations required to solve the partial differential equation remain unchanged. However, recalling that the space-time stiffness matrix elements depend on the coefficients of the partial differential equation, it follows that the neural network provides a functional mapping of these coefficients to predicted sequence values. Sub-matrices, $\bf{A}$, $\bf{B}$, and $\bf{C}$, of the space- time stiffness matrix are functions of the differential equation coefficients according to the finite element assembly. In the context of the elastic bar example, given force and prior displacement fields, the constructed neural network maps the elastic stiffness, cross-sectional area, and density of the bar to the displacement field at the next time step. $\centering\mathbf{u}^{t+1}=ARNN(E,A,\rho:\mathbf{f}^{t},\mathbf{u}^{t},\mathbf{u}^{t-1})+\bm{\varepsilon}^{t+1}=\mathbf{C}^{-1}(E,A,\rho)\left[\mathbf{f}^{t}-\begin{bmatrix}\mathbf{A}(E,A,\rho)&\mathbf{B}(E,A,\rho)\end{bmatrix}\begin{bmatrix}\mathbf{u}^{t-1}\\\ \mathbf{u}^{t}\end{bmatrix}\right]+\bm{\varepsilon}^{t+1}\@add@centering$ (24) This empowers the automatic learning of the differential equation coefficients from observed spatiotemporal data. If $E$, $A$, and $\rho$ are taken as the trainable parameters of the autoregressive neural network, they can be optimized such that the error term, $\bm{\varepsilon}$, is minimized over all time steps. By embedding the space-time finite element method in the architecture of the neural network, variable time step and spatial mesh sizes can be readily accommodated unlike with black-box neural network implementation. Furthermore, while traditional dense or deep neural network models are prone to over-fitting, much less data is needed to accurately predict system responses with this FEM-informed auto-regressive neural network since the number of trainable parameters is limited to the number of equation coefficients. Additionally, unlike traditional deep learning implementations, extrapolation beyond the training data set to new initial and boundary conditions can be expected since the computations performed during a forward pass retain the form of the underlying numerical method. To numerically demonstrate the proposed finite element-informed auto- regressive neural network, we consider a spring-mass-damper system governed by the following differential equation. $m\frac{\partial^{2}u}{{\partial t}^{2}}+c\frac{\partial u}{\partial t}+ku=f_{0}sin(\omega t)$ (25) (a) Exact Training Data and Predicted Solution (b) Convergence of the Training Procedure Figure 10: Training the STFEM Auto-Regressive Neural Network with Exact Data. Here, $t$ is time, $u$ is displacement, $m$ is mass, $c$ is damping coefficient, $k$ is spring stiffness, and $f_{0}$ and $\omega$ are the maximum magnitude and frequency of an external sinusoidal force. Training data is generated by solving the equation using 150 finite elements spanning 3 time units for the following parameter values and initial conditions: $m=1,c=10,k=100,f_{0}=10,\omega=2\pi,u(t=0)=1$, and $\frac{\partial u}{\partial t}(t=0)=1$. A STFEM autoregressive neural network is implemented with $m$, $c$, and $k$ as the trainable parameters. The sum of the absolute differences, also known as the L1 loss function, between the true displacements, $u^{t}$, and their predicted values are taken as the objective function to be minimized. The Adam optimization algorithm is used to iteratively update the learnable parameters. A convergence plot of the training procedure is depicted in Fig. 10(b), which shows that the learnable parameters in the neural network approach the true values of the equation coefficients as the sum of the absolute errors, $|{\varepsilon}^{t}|$, in the prediction of ${u}^{t}$ is minimized, resulting in accurate prediction of the equation solution visualized in Fig. 10(a). The learned values of these coefficients within the neural network can then be used to make predictions for problems with new initial conditions and boundary conditions, generalizing and extrapolating beyond training data like a traditional numerical method while remaining a data-driven machine learning framework. As shown in Fig. 11, the model is able to accurately predict the behavior of the spring-mass-damper system for a variety of problems with different boundary and initial conditions even though it was only trained on a single case. Figure 11: Prediction for New Initial Conditions and Boundary Conditions from Exact Training Data. Gaussian noise of mean 0 and variance 0.001 is added to a new set of training data for the same problem, which is now solved with a coarser 50 element mesh spanning 3 time units. The resulting time series is shown in Fig. 12, and Fig. 13 illustrates that even when provided with a very small and noisy training data set the STFEM structured auto-regressive neural network can accurately capture the system behavior across a variety of new conditions. The ability of the STFEM neural network to produce accurate predictions for cases beyond the training data set is not typical of traditional neural networks, which typically only perform well when applied to cases similar to those seen during training. Figure 12: True Solution and Generated Noisy Training Data. Figure 13: Prediction for New Initial Conditions and Boundary Conditions from Noisy Training Data. Dense and deep neural networks offer great function representation capacity, which make them powerful tools for predicting complex phenomena if they are provided with sufficiently large and diverse datasets to learn from. With limited data, neural networks exhibit overfitting due to their large number of trainable parameters leading to poor interpolation capability. Additionally, the extrapolation capability of neural networks to make good predictions beyond the range of their training data is generally very poor with “black box” implementation. These qualities preclude neural networks from solving engineering problems where the collection of data is expensive or cannot span the space of desired prediction capability. By designing an autoregressive neural network architecture consistent with the space-time finite element method, the generalizability and interpretability of the numerical method can be preserved while harnessing the advantage of data-driven machine learning to reduce the need to make potentially incorrect parameter assumptions. The STFEM neural network may be trained on video data for only a single loading case to produce reasonably accurate parameter values. This intelligently designed neural network, which captures mesh connectivity and spatiotemporal discretization, both reduces the cost of model training and increases robustness to data noise. The simple neural network architecture, with limited learnable parameters, eliminates the need for collecting a massive quantity of data. In fact, with exact data, only a few time observations are required to eventually converge to the true system parameters, which could then be used to make predictions across a wide variety of boundary conditions in the same manner as finite element analysis. Since the functional structure of the neural network is specified as the finite element method, the collection of additional data only serves to reduce inaccuracies arising due to noise in the observations. This contrasts traditional deep-learning models which seek to learn a functional structure from scratch, hence their massive data requirement. ### 3.2 Solving Integral Equations The section will discuss how the DLDC methods can be used to solve the integral equations. For brevity and the underlying context of the article, this section will focus on how to solve the Lippmann-Schwinger equation lippmann1950variational efficiently with the understanding of the numerical methods through DLDC. This form of integral equation has myriads of applications including scattering gopal2022accelerated and multiscale mechanics of materials zecevic2022new ; moulinec1998numerical . A general integral equation appears as, $f(x)=\int_{a}^{b}\mathcal{K}(x,y)u(y)dy,$ (26) where $u$ is an unknown function, $f$ is a known function, and $\mathcal{K}$ is the kernel function. Solution of the integral equation depends on determining the kernel function. The importance of this kernel function will be clear if discussed in the context of solving a differential equation through Green’s function. For example, consider the following equation, $\mathcal{L}u(x)=f(x),\forall x\in\mathcal{\mathbf{R}}$ (27) where, $\mathcal{L}$ is a linear differential operator, $u(x)$ is the unknown function, and $f(x)$ is the forcing function. The solution of this differential equation is given by, $u(x)=\int\mathcal{G}(x-y)f(y)dy,$ (28) Here, $\mathcal{G}(\cdot)$ is the Green’s function, and the following relationship holds true, $\mathcal{L}\mathcal{G}(x,y)=\delta(x-y),$ (29) where $\delta(\cdot)$ is the Delta function. If one compares Eq. 26 and Eq. 28, it will be apparent that for linear differential operator $\mathcal{L}(\cdot)$, the Green’s function is the kernel function. If one solves for this Green’s function, the differential equation will be solved automatically. It is very hard to analytically solve for this Green’s function. Usually, this Green’s function is determined numerically. In recent operator learning paradigm kovachki2021neural ; li2020fourier , this kernel function is replaced by a convolutional neural network. This neural network or kernel is trained by adding subsequent convolutional layers. In order to achieve resolution independence, in Fourier Neural Network (FNO) the input space of the function variables is taken to Fourier domain and trained li2020fourier . Another approach is using graph kernels li2020neural where a subset of the domain points from input and the output functional space are considered to construct a graph and convolutional kernel is employed for training. While these networks are shown to solve complex problems including Navier-Stokes equation li2020fourier , when the domain to solve becomes very large for example large-scale simulation of 3D microstructure as shown in saha2021macroscale ; saha2021microscale , the number of training parameters may explode and a lot of offline training efforts are required. In this section, using the concepts of mechanics and DLDC, the article gives formulations for two methods that can reduce the computational burden when solving large-scale engineering problems through convolutional kernels. Figure 14: The idea of solving a problem with SCA is explained at the top row of the figure. The 2D representative volume element with $100\times 100$ mesh is discretized into 8 clusters. In DLDC method, each cluster is assumed to be a node of the graph and mechanistic functional spaces are taken as nodal attributes. The interaction matrix is similar to nodal interaction among the clusters. As a demonstration, the article chooses to solve resort to the self-consistent clustering analysis liu2016self . The SCA algorithm has demonstrated considerable success in modeling damage, failure, and mechanical response of multiscale materials system yu2019self ; kafka2021image . The idea of SCA is to group similar material points together in a structure based on their elastic response, and solve the Lippmann-Schwinger equation only for the group or cluster of material points instead of the entire domain. More details can be found here liu2016self . The following section will touch the basics of SCA for convenience of discussion. The homogenization problem for a representative volume element (RVE) can be modeled as the following integral equation. $\bm{\epsilon(\mathbf{x})}=\bm{\epsilon^{0}(x)}-\int_{\Omega}\mathcal{G}(x,y):\left[\bm{\sigma(y)}-\bm{C^{0}(y):\epsilon(y)}\right]dy,$ (30) where $\bm{\epsilon(x)}$ is local strain at material point $x$, $\Omega$ is the physical domain (RVE), $\bm{\epsilon^{0}(x)}$ is background strain, $\bm{\sigma(y)}$ is the local stress, $C^{0}$ is the reference material stiffness. With SCA, the Eq. 30 is not solved for all the material points rather on clustered domain. The clustering is done by K-means clustering hartigan1979algorithm . At first, a small, elastic load is applied to the RVE and the resulting strain concentration matrix is stored. Based on this strain concentration matrix, the RVE is clustered (see Figure 14). On this clustered domain the Eq. 30 becomes $\Delta\bm{\epsilon^{I}}=\Delta\bm{\epsilon^{0}}-\sum_{j=1}^{k}\left[\frac{1}{c^{I}|\Omega|}\int_{\Omega}\int_{\Omega}\chi^{I}(x)\chi^{J}(y)\mathcal{G}(x,y)dxdy\right]:\left[\Delta\sigma^{I}(y)-C^{0}(y):\Delta\epsilon^{J}(y)\right].$ (31) In the equation, $c$ is cluster identifier, $k$ is the number of clusters, $I,J$ are the cluster indices, and $\chi^{I}(x)$ is the cluster variables which assumes value 1 when $x$ is in cluster $I$ and zero otherwise. If one compares Eqn. 30 and 31, it will be apparent that the Green’s function is replaced by the convolution operation also known as the interaction tensor. $\bm{D}^{IJ}=\frac{1}{c^{I}|\Omega|}\int_{\Omega}\int_{\Omega}\chi^{I}(x)\chi^{J}(y)\mathcal{G}(x,y)dxdy.$ (32) In conventional SCA method, this interaction tensor is pre-computed as so- called ”offline” database. Based on this pre-computed kernel/interaction tensor, the strain on each cluster is computed from Eq. 32. With the DLDC method, the article will show how to determine this kernel on the cluster centroids so that the number of parameters to be trained can be reduced. Using graph kernel networks, the Eq. 31 can be solved on reduced functional space. The justification comes from the mathematical form of the graph kernel neural networks. The graph kernel network has the following form for variable update: $v_{t+1}(x)=\sigma\left(Wv_{t}(x)+\int_{\Omega}\mathcal{K}(x,y,a(x),a(y))v_{t}(y)dy\right).$ (33) Here, $v_{t+1}$ is the output variable in a transformed functional space at $(t+1)$-th time step, $\sigma$ is the activation function, $W$ is weight, $(x,y,a(x),a(y)$ is the edge index in the input function space where $x$ and $y$ are coordinates, and $a(x)$ and $a(y)$ are values of input function. More details can be found here []. The discretized form of the Eq. 33 is: $v_{t+1}(x)=\sigma\left(Wv_{t}(x)+\frac{1}{|N(x)|}\sum_{y\in N}\mathcal{K}(x,y,a(x),a(y))v_{t}(y)dy\right).$ (34) Figure 15: Problem statement for the one-dimensional self-consistent clustering analysis. a) The one-dimensional problem domain for training the network, b) The cluster centroids after k-means clustering, c) Comparison between analytical and SCA solution for 33 clusters, d) Evolution of error during training. In this equation, $N(x)$ defines a neighborhood inside radius $r$. Drawing similarities with Eq. 31, it is apparent that the interaction tensor and the summation term in Eq. 34 have similarities. The only difference is that, in graph kernel network the neighborhood nodes can be selected by changing the radius of whereas the SCA is inherently global when computing the interaction tensor. As an example, let us consider a one-dimensional domain of length 10 as shown in Figure 15 (a) discretized into 1000 points in equal intervals. The material is a linear elastic one with stiffness varying as $C(x)=\frac{1}{1+x^{2}}$. On this domain, the overall applied strain $\epsilon$ is varied from 0.05 to 0.5 and the domain is clustered starting from 2 to 128 clusters. For each such discretization, the cluster centroids are taken as the input samples. Essentially, 1000 material points are reduced 2-128 material points with the assumption that the local strain has a constant value inside each of these clusters. A distribution of possible cluster centroids with 33 clusters is shown in Figure 15(b). To generate the data, the solution with SCA is validated against the analytical solution (Figure 15(c)). The training is performed with a modified version of the graph kernel network with 6 iterative layers and a radius of 2 (i.e., the interaction is established between only 2 neighbor clusters). The training performance of the network is shown in Figure 15. The figure suggests that the training performance is quite good as the normalized mean squared error is going down with number of epochs. A more thorough analysis on the efficacy of such construction is shown in a companion paper owensahagao2022 where it is shown how varying the hyperparameters of the neural network such as the radius of influence $r$ or number of layers for training. Apart from the reduction of training parameters and domain points, the method can also extrapolate to some extent. It is often identified as the ”resolution independence”. An example of such extrapolation is shown in Figure 16.The figure shows the prediction of the kernel learning algorithm and SCA for 300 clusters and 0.2 applied strain. This particular case is outside the domain of training and testing datasets. In spite of that, we can see that the kernel learning algorithm is as good as SCA. However, further testing is required before claiming that kernel learning method has resolution independence. Nevertheless, the results are promising and shows a way how we can overcome the limitations of the current graph kernel networks while solving three-dimensional engineering problems. Figure 16: Comparison between kernel learning prediction and SCA for 300 clusters. ## 4 Future Developments While the structure of the neural networks proposed in DLDC shows the potential of a general framework of solving fundamental calculus and applied science problems, there are still some limitations that need to be overcome. For example, the forward difference or central difference neural networks presented in the differential calculus section works well only same discretization or sampling intervals are used. This is apparent from the construction of neural networks as well. However, in real-life applications, unstructured data or uneven sampling is quite common. The next step is to extend the DLDC structures similar to differential quadrature method to obtain the derivative of unstructured data. The weights can be learned via either convolutional kernel (similar to C-HiDeNN) or fully-connected neural network or its variant (such as autoencoder). While the DLDC structure gives the exact reproduction of the Gauss quadrature method, it is to be explored if the flexibility of deep neural network can be used to see if it can solve integration that hard to solve with Gauss methods, such as integration around singularities. The authors are currently working to extend C-HiDeNN into a more general form and to solve multiresolution problems such as mcveigh2006multiresolution ; mcveigh2009multiresolution . While the proposed integral equation solver can significantly reduce the number of training parameters for graph kernels, we still need to train the method with some data. The authors are currently working to extend C-HiDeNN method to solve the integral Lippmann-Schwinger equation so that the convolutional patch function can capture interaction among the cluster centroids. ## 5 Conclusions The proposed DLDC method aims to combine the rich knowledge of numerical methods with modern data science techniques and create a new perspective on solving challenging problems in engineering and physical sciences. The building blocks of the DLDC are still in the making. The purpose of building the tools is to make a general and easy to follow process of making a deep neural network to solve any governing differential equation. The vision is to combine all the benefits from different numerical methods and propose a unified approach to solve engineering problems. At the same time, the DLDC methods can be used as a teaching tool for calculus in K-12 education system. These tools will simultaneously introduce the students with deep learning and applied calculus. This is expected to increase participation in STEM programs from students. Moreover, the DLDC methods try to avoid multiple training and can be more flexible using the deep neural network. The DLDC version of finite element method i.e., C-HiDeNN shows higher accuracy compared to FEM with same degrees of freedom. The proposed integral equation solver potentially avoids large number of training parameters that was the bottleneck for using graph kernel network. However, further research is required to establish the DLDC as viable method to solve for general problems in science and engineering. ###### Acknowledgements. The authors would like to acknowledge the support of National Science Foundation (NSF, USA) grants CMMI-1762035 and CMMI-1934367 and AFOSR, USA grant FA9550-18-1-0381. C. Park and S. Saha would like to thank the Division of Orthopaedic Surgery and Sports Medicine at Ann and Robert H. Lurie Children’s Hospital for their philanthropic grant. The authors would like to acknowledge the contribution of Alberto Ciampaglia, Department of Mechanical and Aerospace Engineering, Politecnico di Torino, Italy and Visiting Researcher, Department of Mechanical Engineering, Northwestern University to this article. ## 6 Appendix 1: Space-Time Finite Element Formulation for a 1D Elastic Bar The governing equation is: $EA\frac{\partial^{2}u}{\partial x^{2}}-\rho A\frac{\partial^{2}u}{\partial t^{2}}=0.$ (35) For a bar with left end at $x=L$ and right end at $x=R$, vibrating from time $t=0$ to time $t=T$, the above equation can be converted to a weak form by applying the principle of virtual work. $\int\limits^{T}_{0}\int\limits^{R}_{L}\tilde{v}\left(EA\frac{\partial^{2}u}{\partial x^{2}}-\rho A\frac{\partial^{2}u}{\partial t^{2}}\right)\partial x\partial t=0$ (36) $\int\limits^{T}_{0}\int\limits^{R}_{L}\left(EA\frac{\partial\tilde{v}}{\partial x}\frac{\partial^{2}u}{\partial x^{2}}-EA\frac{\partial\tilde{v}}{\partial x}\frac{\partial u}{\partial x}-\rho A\frac{\partial\tilde{v}}{\partial t}\frac{\partial^{2}u}{\partial t^{2}}+\rho A\frac{\partial\tilde{v}}{\partial t}\frac{\partial u}{\partial t}\right)\partial x\partial t=0$ (37) $\left[\int_{0}^{T}EA\tilde{v}\frac{\partial u}{\partial x}\,\partial t\right]_{x=L}^{x=R}-\int\limits^{T}_{0}\int\limits^{R}_{L}EA\frac{\partial\tilde{v}}{\partial x}\frac{\partial u}{\partial x}\partial x\partial t-\left[\int_{L}^{R}\rho A\tilde{v}\frac{\partial u}{\partial t}\,\partial x\right]_{t=0}^{t=T}+\int\limits^{T}_{0}\int\limits^{R}_{L}\rho A\frac{\partial\tilde{v}}{\partial t}\frac{\partial u}{\partial t}\partial x\partial t=0$ (38) Finite element shape function matrices can be used to discretize the simplified weak form. For an element spanning $x=x_{0}^{e}$ to $x=x_{1}^{e}$ through space and $t=t_{0}^{e}$ to $t=t_{1}^{e}$ through time, the element matrices correspond to the following expressions. $\text{Element Stiffness Matrix }[K]^{e}=\int\limits^{t_{1}^{e}}_{t_{0}^{e}}\int\limits^{x_{1}^{e}}_{x_{0}^{e}}EA\frac{\partial\tilde{v}}{\partial x}\frac{\partial u}{\partial x}\partial x\partial t$ (39) $\text{Element Mass Matrix }[M]^{e}=\int\limits^{t_{1}^{e}}_{t_{0}^{e}}\int\limits^{x_{1}^{e}}_{x_{0}^{e}}\rho A\frac{\partial\tilde{v}}{\partial t}\frac{\partial u}{\partial t}\partial x\partial t$ (40) The global boundary conditions are prescribed via the other terms in Eq. 38. $\text{Space Boundary Conditions }=\left[\int_{0}^{T}EA\tilde{v}\frac{\partial u}{\partial x}\,\partial t\right]_{x=L},\left[\int_{0}^{T}EA\tilde{v}\frac{\partial u}{\partial x}\,\partial t\right]_{x=R}$ (41) $\text{Time Boundary Conditions }=\left[\int_{L}^{R}\rho A\tilde{v}\frac{\partial u}{\partial t}\,\partial x\right]_{t=0},\left[\int_{L}^{R}\rho A\tilde{v}\frac{\partial u}{\partial t}\,\partial x\right]_{t=T}$ (42) The time boundary condition at $t=0$ corresponds to the enforcement of initial velocity. The time boundary condition at the final time, $t=T$, can be eliminated from the global matrix equation in favor of equations that enforce initial displacement at $t=0$. ## 7 Appendix 2: Time Finite Element Formulation for a Spring Mass Damper System The governing equation is: $m\frac{d^{2}u}{dt^{2}}+c\frac{du}{dt}+ku=f_{0}sin(\omega t).$ (43) For a mass oscillating from time $t=0$ to time $t=T$, the above equation can be converted to a weak form by applying the principle of virtual work. $\int_{0}^{T}\tilde{v}\left(m\frac{d^{2}u}{dt^{2}}+c\frac{du}{dt}+ku- f_{0}sin(\omega t)\right)\,dt=0$ (44) $\left[m\tilde{v}\frac{du}{dt}\right]_{t=0}^{t=T}-\int_{0}^{T}m\frac{d\tilde{v}}{dt}\frac{du}{dt}\,dt+\int_{0}^{T}c\tilde{v}\frac{du}{dt}\,dt+\int_{0}^{T}k\tilde{v}u\,dt-\int_{0}^{T}\tilde{v}f_{0}sin(\omega t)\,dt=0$ (45) Finite element shape function matrices can be used to discretize the simplified weak form. For an element spanning $t=t_{0}^{e}$ to $t=t_{1}^{e}$ through time, the element matrices correspond to the following expressions. $\text{Element Mass Matrix }[M]^{e}=\int_{t_{0}^{e}}^{t_{1}^{e}}m\frac{d\tilde{v}}{dt}\frac{du}{dt}\,dt$ (46) $\text{Element Damping Matrix }[C]^{e}=\int_{t_{0}^{e}}^{t_{1}^{e}}c\tilde{v}\frac{du}{dt}\,dt$ (47) $\text{Element Stiffness Matrix }[K]^{e}=\int_{t_{0}^{e}}^{t_{1}^{e}}k\tilde{v}u\,dt$ (48) $\text{Element Force Matrix }[F]^{e}=\int_{t_{0}^{e}}^{t_{1}^{e}}\tilde{v}f_{0}sin(\omega t)\,dt$ (49) The element force matrices account for the force boundary condition due to the external sinusoidal excitation of the sprung mass. The temporal boundary conditions are prescribed via the other term in Eq. 45. $\text{Time Boundary Conditions }=\left[m\tilde{v}\frac{du}{dt}\right]_{0},\left[m\tilde{v}\frac{du}{dt}\right]_{T}$ (50) The time boundary condition at $t=0$ corresponds to the enforcement of initial velocity. The time boundary condition at the final time, $t=T$, can be eliminated from the global matrix equation in favor of equations that enforce initial displacement at $t=0$. ## References * (1) S. Saha, Z. Gan, L. Cheng, J. Gao, O.L. Kafka, X. Xie, H. Li, M. Tajdari, H.A. Kim, W.K. Liu, Computer Methods in Applied Mechanics and Engineering 373, 113452 (2021) * (2) X. Xie, J. Bennett, S. Saha, Y. Lu, J. Cao, W.K. Liu, Z. Gan, npj Computational Materials 7(1), 1 (2021) * (3) Z. Gan, K.K. Jones, Y. Lu, W.K. Liu, Integrating Materials and Manufacturing Innovation 10(2), 177 (2021) * (4) O.L. Kafka, K.K. Jones, C. Yu, P. Cheng, W.K. Liu, Journal of the Mechanics and Physics of Solids 150, 104350 (2021) * (5) J. Sirignano, K. Spiliopoulos, Journal of computational physics 375, 1339 (2018) * (6) M. Raissi, The Journal of Machine Learning Research 19(1), 932 (2018) * (7) J. Han, A. Jentzen, et al., Communications in mathematics and statistics 5(4), 349 (2017) * (8) B. Yu, et al., Communications in Mathematics and Statistics 6(1), 1 (2018) * (9) L. Xiao, B. Liao, S. Li, K. Chen, Neural Networks 98, 102 (2018) * (10) M. Zhu, B. Chang, C. Fu, arXiv preprint arXiv:1802.08831 (2018) * (11) J.N. Kani, A.H. Elsheikh, arXiv preprint arXiv:1709.00939 (2017) * (12) M. Karl, M. Soelch, J. Bayer, P. Van der Smagt, arXiv preprint arXiv:1605.06432 (2016) * (13) S. Chakraverty, S. Mall, _Artificial neural networks for engineers and scientists: solving ordinary differential equations_ (CRC Press, 2017) * (14) M. Magill, F. Qureshi, H. de Haan, Advances in Neural Information Processing Systems 31 (2018) * (15) W.K. Liu, Z. Gan, M. Fleming, Springer. Lubliner, J., Oliver, J., Oller, S. and Oñate, E.(1989). A plastic-damage model for concrete. International Journal of solids and structures 25(3), 299 (2021) * (16) M. Raissi, P. Perdikaris, G.E. Karniadakis, Journal of Computational physics 378, 686 (2019) * (17) G.E. Karniadakis, I.G. Kevrekidis, L. Lu, P. Perdikaris, S. Wang, L. Yang, Nature Reviews Physics 3(6), 422 (2021) * (18) G. Pang, L. Lu, G.E. Karniadakis, SIAM Journal on Scientific Computing 41(4), A2603 (2019) * (19) A. Krishnapriyan, A. Gholami, S. Zhe, R. Kirby, M.W. Mahoney, Advances in Neural Information Processing Systems 34, 26548 (2021) * (20) R. Li, E. Lee, T. Luo, Materials Today Physics 19, 100429 (2021) * (21) W.T. Leung, G. Lin, Z. Zhang, arXiv preprint arXiv:2108.12942 (2021) * (22) S.L. Brunton, J.L. Proctor, J.N. Kutz, Proceedings of the national academy of sciences 113(15), 3932 (2016) * (23) E. Kaiser, J.N. Kutz, S.L. Brunton, Proceedings of the Royal Society A 474(2219), 20180335 (2018) * (24) K. Kaheman, J.N. Kutz, S.L. Brunton, Proceedings of the Royal Society A 476(2242), 20200279 (2020) * (25) Z. Gan, O.L. Kafka, N. Parab, C. Zhao, L. Fang, O. Heinonen, T. Sun, W.K. Liu, Nature communications 12(1), 1 (2021) * (26) X. Xie, W.K. Liu, Z. Gan, arXiv preprint arXiv:2111.03583 (2021) * (27) R.T. Chen, Y. Rubanova, J. Bettencourt, D.K. Duvenaud, Advances in neural information processing systems 31 (2018) * (28) L. Zhang, L. Cheng, H. Li, J. Gao, C. Yu, R. Domel, Y. Yang, S. Tang, W.K. Liu, Computational Mechanics 67(1), 207 (2021) * (29) M.Y. Niu, L. Horesh, I. Chuang, arXiv preprint arXiv:1904.12933 (2019) * (30) T.W. Hughes, I.A. Williamson, M. Minkov, S. Fan, Science advances 5(12), eaay6946 (2019) * (31) J. He, J. Xu, Science china mathematics 62(7), 1331 (2019) * (32) N. Kovachki, Z. Li, B. Liu, K. Azizzadenesheli, K. Bhattacharya, A. Stuart, A. Anandkumar, arXiv preprint arXiv:2108.08481 (2021) * (33) Z. Li, N. Kovachki, K. Azizzadenesheli, B. Liu, K. Bhattacharya, A. Stuart, A. Anandkumar, arXiv preprint arXiv:2010.08895 (2020) * (34) L. Lu, P. Jin, G.E. Karniadakis, arXiv preprint arXiv:1910.03193 (2019) * (35) L. Lu, X. Meng, S. Cai, Z. Mao, S. Goswami, Z. Zhang, G.E. Karniadakis, Computer Methods in Applied Mechanics and Engineering 393, 114778 (2022) * (36) A. Anandkumar, K. Azizzadenesheli, K. Bhattacharya, N. Kovachki, Z. Li, B. Liu, A. Stuart, in _ICLR 2020 Workshop on Integration of Deep Neural Models and Differential Equations_ (2020) * (37) H. You, Y. Yu, M. D’Elia, T. Gao, S. Silling, arXiv preprint arXiv:2201.02217 (2022) * (38) C. Tucker, K. Jackson, J.J. Park, in _American Society of Engineering Education_ (2020) * (39) N. Wang, P. Tonko, N. Ragav, M. Chungyoun, J. Plucker, Technology & Innovation (2022) * (40) D.S. Touretzky, C. Gardner-McCune, Computational Thinking Education in K-12: Artificial Intelligence Literacy and Physical Computing pp. 153–180 (2022) * (41) D. Touretzky, C. Gardner-McCune, D. Seehorn, International Journal of Artificial Intelligence in Education pp. 1–34 (2022) * (42) Y. Yin, Ai4all: Ai education for k-12. Tech. rep., EasyChair (2022) * (43) S.C. Chapra, R.P. Canale, et al., _Numerical methods for engineers_ , vol. 1221 (Mcgraw-hill New York, 2011) * (44) R. Bellman, J. Casti, Journal of Mathematical Analysis and Applications 34(2), 235 (1971) * (45) C.W. Bert, M. Malik, Applied Mechanics Review (1996) * (46) C. Shu, _Differential quadrature and its application in engineering_ (Springer Science & Business Media, 2012) * (47) C. Park, Y. Lu, S. Saha, T. Xue, J. Guo, S. Mojumder, G. Wagner, W. Liu, Computaitonal Mechanics (2023) * (48) G. Liu, Y. Gu, Journal of Sound and vibration 246(1), 29 (2001) * (49) G.R. Liu, Y.T. Gu, _An introduction to meshfree methods and their programming_ (Springer Science & Business Media, 2005) * (50) T.J. Hughes, G.M. Hulbert, Computer methods in applied mechanics and engineering 66(3), 339 (1988) * (51) G.M. Hulbert, T.J. Hughes, Computer methods in applied mechanics and engineering 84(3), 327 (1990) * (52) L. Wang, H. Zhong, Applied Mathematical Modelling 41, 445 (2017) * (53) T. Taskaya-Temizel, M.C. Casey, Neural Networks 18(5-6), 781 (2005) * (54) O. Triebe, N. Laptev, R. Rajagopal, arXiv preprint arXiv:1911.12436 (2018) * (55) B.A. Lippmann, J. Schwinger, Physical Review 79(3), 469 (1950) * (56) A. Gopal, P.G. Martinsson, Advances in Computational Mathematics 48(4), 1 (2022) * (57) M. Zecevic, R.A. Lebensohn, L. Capolungo, Mechanics of Materials 166, 104208 (2022) * (58) H. Moulinec, P. Suquet, Computer methods in applied mechanics and engineering 157(1-2), 69 (1998) * (59) Z. Li, N. Kovachki, K. Azizzadenesheli, B. Liu, K. Bhattacharya, A. Stuart, A. Anandkumar, arXiv preprint arXiv:2003.03485 (2020) * (60) S. Saha, O.L. Kafka, Y. Lu, C. Yu, W.K. Liu, Integrating Materials and Manufacturing Innovation 10(3), 360 (2021) * (61) S. Saha, O.L. Kafka, Y. Lu, C. Yu, W.K. Liu, Integrating Materials and Manufacturing Innovation 10(2), 142 (2021) * (62) Z. Liu, M.A. Bessa, W.K. Liu, Computer Methods in Applied Mechanics and Engineering 306, 319 (2016) * (63) C. Yu, O.L. Kafka, W.K. Liu, Computer Methods in Applied Mechanics and Engineering 349, 339 (2019) * (64) J.A. Hartigan, M.A. Wong, Journal of the royal statistical society. series c (applied statistics) 28(1), 100 (1979) * (65) O. Huang, S. Saha, J. Guo, W.K. Liu, Computational Mechanics (submitted) (2023) * (66) C. McVeigh, F. Vernerey, W.K. Liu, L.C. Brinson, Computer Methods in Applied Mechanics and Engineering 195(37-40), 5053 (2006) * (67) C. McVeigh, W.K. Liu, Journal of the Mechanics and Physics of Solids 57(2), 244 (2009)
# Holonomic Control of Arbitrary Configurations of Docked Modboats Zhijie Qiao, Gedaliah Knizhnik, and Mark Yim The authors are with the GRASP Laboratory, University of Pensylvannia, Philadelphia, PA 19104. <EMAIL_ADDRESS> ###### Abstract The Modboat is a low-cost, underactuated, modular robot capable of surface swimming, docking to other modules, and undocking from them using only a single motor and two passive flippers. Undocking is achieved by causing intentional self-collision between the tails of neighboring modules in certain configurations; this becomes a challenge, however, when collective swimming as one connected component is desirable. Prior work has developed controllers that turn arbitrary configurations of docked Modboats into steerable vehicles, but they cannot counteract lateral forces and disturbances. In this work we present a centralized control strategy to create holonomic vehicles out of arbitrary configurations of docked Modboats using an iterative potential-field based search. We experimentally demonstrate that our controller performs well and can control surge and sway velocities and yaw angle simultaneously. ## I Introduction Aquatic modular self-reconfigurable robotic systems (MSRRs) are of great interest to researchers and industry; they can be used for monitoring ocean environments, performing exploration tasks, and collecting flow information[1], while adapting to changing conditions and scales of interest. Conventional wisdom has been that such MSRRs must be built from modules capable of holonomic motion [2, 3, 4, 5], which has limited development due to increased complexity. Recent work by the authors, however, has shown that effective aquatic MSRRs can be built from underactuated surface-swimming modules [6, 7, 8, 9, 10]. The underactuated modules used in this prior work — the Modboats — use passive flippers and an inertial rotor powered by a single motor to generate thrust and steering [6]. They are capable of docking and reconfiguration through permanent-magnet based docks and can undock from one another using mechanical self-collision of protruding tails (see Fig. 1), all while using only one motor [7]. While this passive docking setup and mechanical undocking method greatly reduces actuation complexity for the system, it also introduces a major constraint when attempting to swim collectively: any collective behavior must constantly avoid self-collisions between neighboring tails, which would cause the docked configuration to disintegrate. Prior work addressed this concern by introducing restrictions on the phase and thrust direction allowed for the Modboat modules [8, 9], which are technically capable of thrusting in any direction [10]. By limiting the thrust direction to the surge axis of the configuration and requiring in-phase swimming, these approaches were able to guarantee no unintentional self-collisions and maintain steerability for any arbitrary structure [8, 9]. But the resulting controllers could not produce force along the configurations’ sway axes, which made them highly susceptible to noise and external disturbances. It is also reasonable to suspect that a structure of $N\geq 3$ modules should be holonomic in the plane, given a reasonable control law, but thus far such a controller has not been developed due to the complexity of the collision constraint [9]. In this work, we propose a holonomic collective control approach for an arbitrary configuration of docked Modboats, by using an iterative potential field search to find collision-free desirable movements. While iterative path planning has been widely adopted in autonomous robot navigation [11, 12, 13, 14], little work has been done to explore its use in the avoidance of internal collision constraints. Our controller uses such an approach to relax the assumption in [8, 9] of thrust only along the surge axis, while using a collision checker to avoid any unintended undocking. The rest of this work is organized as follows: in Section II we formulate the problem and present an approach for finding collision-free motions that generate desired forces. Section III presents the control approach for determining desired forces for given motions, and Section IV presents the experimental evaluation of our approach. The results are discussed in Section V. Figure 1: Example structure of three docked Modoats in a parallel configuration, with the structure frame given by $S$ and the surge ($y$) and sway ($x$) axes marked. For each boat, the top body is shown in black, and the tail is shown in color, each pointing at $-\pi/2~{}$\mathrm{r}\mathrm{a}\mathrm{d}$$. The passive flippers are not shown but connect to the tails and do not protrude from the top body footprint. More details of the Modboat structure can be found in [6]. ## II Methodology Consider a configuration of docked Modboat modules labeled $i\in[1,N]$. To move, the modules collectively perform a series of swim cycles, in which each module executes the waveform given in (1), where $\phi_{i}(t)$ is the motor angle of module $i$ and $\phi_{0,i}$, $A_{i}$, and $T$ are the centerline, amplitude, and period of the oscillation, respectively. The period is constant for all modules for concurrency111In this work we use $T=1.5~{}$\mathrm{s}$$, which has been empirically determined to be an effective period for the system [15, 9], although the methodology is applicable to any period., while the centerline and amplitude vary for each module for control. Parameters are determined at the beginning of each cycle and executed for a single period. Since the final pose of one cycle and the initial pose of the next cycle may not be coincident, an additional transition cycle is allowed between swim cycles to prepare, after which point the process repeats. $\phi_{i}(t)=\phi_{0,i}+A_{i}\cos{\left(2\pi t/T\right)},\quad t\in[0,T)$ (1) $F(A)=0.022\mathinner{\\!\left\lvert A\right\rvert}-0.019,\quad\mathinner{\\!\left\lvert A\right\rvert}\in[0.9,2.6]$ (2) In prior work we have shown that the use of waveform (1) results in a linear relationship between force and amplitude, given in (2) for a period of $1.5~{}$\mathrm{s}$$[8, 9]. Unlike in prior work [8, 9], however, in this work we allow $\phi_{0}$ to take on any value. We also allow amplitude to take on both positive and negative values for numerical flexibility; negative amplitudes result in a phase shift of $\pi~{}$\mathrm{r}\mathrm{a}\mathrm{d}$$, but do not affect the generated force, as per (2). Figure 2: Numerically simulated phase-space with collision region, with $\phi_{1}$ being the left boat and $\phi_{2}$ being the right boat. Collided configurations are shown in red, and free space in green, and top-body neighbors display an offset version of this relationship. The blue line represents a sample motion, with boat $1$ having $(\phi_{0},A)=(-1.5,0.5)$, and boat $2$ having $(\phi_{0},A)=(2.0,1.0)$ ###### Problem 1 (Valid Swim Cycles). Given a configuration of docked Modboats $i\in[1,N]$ and a set of desired forces $\vec{F}_{des}=\begin{bmatrix}F_{x,des}&F_{y,des}&\tau_{des}\end{bmatrix}^{T}$ find a set $\Phi$ of pairs $(\phi_{0},A)_{i}$ for $i\in[1,N]$ such that the generated forces $\vec{F}=\begin{bmatrix}F_{x}&&F_{y}&&\tau\end{bmatrix}^{T}$ equal (as closely as possible) the desired forces, i.e. $\vec{F}=\vec{F}_{des}$, while avoiding tail collisions between neighboring modules. ###### Problem 2 (Valid Transition Cycles). Given multiple sets of motion pairs $\Phi_{k}$, $k\in\mathbb{Z}^{+}$, where $k$ represents the swim cycle for which the solution is used, find a set of motions $\Psi$ to transition from the final pose in $\Phi_{k}$ to the initial pose in $\Phi_{k+1}$ while avoid tail collisions between neighboring modules. The goal of this work, then is to find solutions to 1 and 2 under these conditions, i.e. to find a valid set of swim and transition cycles that generate a desired set of forces without causing internal collisions. This will allow the configuration of docked Modboats to function as a holonomic vehicle. ### II-A Collision Region The first step in solving 1 is to identify what set of poses causes a collision between neighbors. We can construct a graph of this collision region by numerically solving for the intersection of the tails for any pair of boats, as described in [9]. The resulting graph is shown in Fig. 2 for a single pair of boats; in this phase space any motion of the two neighboring boats is a straight line, as proven in [9]. Figure 3: Distance to Collision (DoC) analysis for four different intersecting cases, using the collision region shown in Fig. 2. $AB$ is the phase-space trajectory and $CDEF$ is the collision region. The DoC is shown in red in each case. The critical question is how to define the distance to collision (DoC) given the phase-space representation in Fig. 2 and linear trajectories. For any pair of boats, we can define the DoC as $DoC=D(\phi_{i},A_{i},\phi_{j},A_{j})$. Alternatively, given a line $AB$ representing a trajectory in phase-space and a polygon with boundary $CDEF$ representing the collision region, we propose to define DoC as follows and as illustrated in Fig. 3: 1. 1. If $AB$ is completely outside $CDEF$, the DoC is the minimum distance from $AB$ to $CDEF$. 2. 2. If $AB$ intersects $CDEF$ on point $G$ with $B$ interior to $CDEF$, the DoC is $-|BG|$. 3. 3. If $AB$ is inside $CDEF$ with extensions intersecting $CDEF$ on $G$ and $H$, the DoC is $-\left(|AB|+\min[|AG|,|BH|]\right)$. 4. 4. If $AB$ intersects $CDEF$ on $G$ and $H$, the DoC is $-\left(|GH|+\min[|AG|,|BH|]\right)$. In summary, the DoC captures the distance along the phase-space trajectory to move into or out of the collision region. While more optimal strategies (i.e. moving sideways) exist, this strategy is physically meaningful in the context of the Modboats. Computing the DoC is also expensive, since the calculation needs to be performed for each pair of boats at every step in the solution process, so we precompute a DoC table with a discretization of $0.1~{}$\mathrm{r}\mathrm{a}\mathrm{d}$$ for the centerline and amplitude. Note that in practice Fig. 2 must be extended to $[-2\pi,2\pi]$ to account for angle wrapping, which creates several identical but shifted collision regions. In this case the DoC must be calculated for every collision region and the most conservative value is used. ### II-B Attractive Field Given the collision-space representation developed in Section II-A, we begin to solve 1 by creating an attractive potential to drive the values for generated structural forces $\vec{F}$ to their desired values $\vec{F}_{des}$, which can be computed based on the gradient of an error term. Let our error vector be given by $\vec{e}$ in (3), and our weight vector be given by $\vec{w}$ in (4), which accounts for the relative unit scale of force and torque (we have heuristically determined that $[\begin{matrix}1&1&10\end{matrix}]^{T}$ is effective). Then the error in generated forces is given by $E=\vec{e}\cdot\vec{w}$. $\vec{e}=\vec{F}_{des}-\vec{F}$ (3) $\vec{w}=\begin{bmatrix}w_{x}&w_{y}&w_{\tau}\end{bmatrix}^{T}$ (4) Eq. (1) is parameterized by $\phi_{0,i}$ and $A_{i}$ for each boat, or $\vec{\phi}_{0}$ and $\vec{A}$ for all boats. Then (5) and (6) give the gradient in terms of those quantities. $\displaystyle\nabla_{\vec{\phi}_{0}}E$ $\displaystyle=-\left(\nabla_{\vec{\phi}_{0}}\vec{F}\right)\vec{w}$ (5) $\displaystyle\nabla_{\vec{A}\hphantom{{}_{|}}}E$ $\displaystyle=-\left(\nabla_{\vec{A}\hphantom{{}_{|}}}\vec{F}\right)\vec{w}$ (6) From (2) and Fig. 1, we can define the forces produced by each boat as in (7), which combine to form the configuration’s generated forces as $\vec{F}=\sum_{i}\vec{F}_{i}$. Note that the generated force for each boat points opposite the centerline direction of the tail tip, and $F_{i}=F(A_{i})$ from (2). $\vec{F}_{i}=\begin{bmatrix}F_{x}\\\ F_{y}\\\ \tau\end{bmatrix}_{i}=\begin{bmatrix}-F_{i}\cos(\phi_{0,i})\\\ -F_{i}\sin(\phi_{0,i})\\\ -F_{i}\sin(\phi_{0,i})x_{i}+F_{i}\cos(\phi_{0,i})y_{i}\end{bmatrix}$ (7) Figure 4: Repulsive field strength segmented function vs.distance to collision. This stepped function determines the strength of the repulsive field and prioritizes the worst collisions. The gradient of the force produced by each boat can be taken as in (5) with respect to its centerline, and (6) with respect to its amplitude. These gradients then form the rows of the gradients in (5) and (6) with respect to $\vec{\phi}_{0}$ and $\vec{A}$, respectively. $\displaystyle\nabla_{\phi_{0,i}}\vec{F}_{i}$ $\displaystyle=\begin{bmatrix}\hphantom{-}F_{i}\sin(\phi_{0,i})\\\ -F_{i}\cos(\phi_{0,i})\\\ -F_{i}\cos(\phi_{0,i})x_{i}-F_{i}\sin(\phi_{0,i})y_{i}\end{bmatrix}^{T}$ (8) $\displaystyle\nabla_{A_{i}}\vec{F}_{i}$ $\displaystyle=\begin{bmatrix}-0.022\cos(\phi_{0,i})\\\ -0.022\sin(\phi_{0,i})\\\ -0.022\left[\sin(\phi_{0,i})x_{i}-\cos(\phi_{0,i})y_{i}\right]\end{bmatrix}^{T}$ (9) Eqs. (5) and (6) can then be used — with values obtained from (8) and (9) — to obtain the desired step direction from the attractive field as in (10). However, since the size of this step may be arbitrarily small, we take a fixed size step in the direction given by the gradient, where we use the $\operatorname{sgn}$ function rather than taking the unit vector to account for the discretized phase space. $\displaystyle\begin{split}\vec{d}_{\vec{\phi}_{0}}&=\Delta d\hphantom{|}\operatorname{sgn}{(\nabla_{\vec{\phi}_{0}}E)}\\\ \vec{d}_{\vec{A}\hphantom{{}_{|}}}&=\Delta d\hphantom{|}\operatorname{sgn}{(\nabla_{\vec{A}\hphantom{{}_{|}}}E)}\end{split}$ (10) ### II-C Repulsive Field To solve 1 we also need a repulsive field to drive each Modboat’s tail away from collisions with any of its neighbors, which occur at motor angles given by Fig. 2. Because of the non-standard nature of distance to collision in this scenario, we take the following approach: for each boat, let $D_{0}$ represent the DoC at the current state, while $D_{\phi}$ and $D_{A}$ represent the DoC after updating $\phi_{0,i}$ and $A_{i}$ while neighboring parameters (subscript $j$) remain constant, as in (11). $\begin{bmatrix}D_{0}\\\ D_{\phi}\\\ D_{A}\end{bmatrix}_{i,j}=\begin{bmatrix}D(\phi_{0,i}\hphantom{xxd_{\phi_{0,i}}},A_{i}\hphantom{xxd_{A,i}},\phi_{0,j},A_{j})\\\ D(\phi_{0,i}+d_{\phi_{0,i}},A_{i}\hphantom{xxd_{A,i}},\phi_{0,j},A_{j})\\\ D(\phi_{0,i}\hphantom{xxd_{\phi_{0,i}}},A_{i}+d_{A,i},\phi_{0,j},A_{j})\end{bmatrix}$ (11) The repulsive field strength $U_{i}$ is constructed in (12) for each boat using a segmented expression based on the DoC, as shown in Fig. 4, which has been shown to work well in our experiments, and its sign is adjusted based on whether the attractive field step results in “more” or “less” collision. $\begin{bmatrix}U_{\phi}\\\ U_{A}\end{bmatrix}_{i}=\sum_{j}\left(\begin{bmatrix}\operatorname{sgn}{(D_{\phi}-D_{0})}\\\ \operatorname{sgn}{(D_{A}-D_{0})}\end{bmatrix}S(D_{0})\right)_{i,j}$ (12) The repulsive field outputs in (12) are calculated by summing over all occupied neighbor sites $j$ (maximum 4). If $U_{i}$ is positive, the update in (10) is performed for boat $i$; otherwise, no update is performed. This allows each boat to prioritize the worst collision case and reach a balance among all its neighbors. ### II-D Overall Procedure for _$n_{epoch}\in[1,N]$_ do for _$i\in[1,N_{1}]$_ do Apply attractive field; for _$i\in[N_{1},N_{2}]$_ do Apply attractive field if repulsive field $>$ 0; for _$i\in[N_{2},N_{3}]$_ do Apply -attractive field if repulsive field $<$ 0; Algorithm 1 Potential field algorithm for generating collision free solutions to 1. The attractive field in Section II-B and repulsive field in Section II-C together should be sufficient to solve 1. The repulsive field formulation in Section II-C has two issues, however: (a) it does not guarantee the final solution will be collision free, and (b) the Modboats may be unable to traverse through collided phase-space to reach a desired solution. To address these issues, we apply a three-stage iterative approach to solving 1 as presented in Algorithm 1: 1. 1. Applying the attractive field alone prioritizes finding a solution to the desired forces. 2. 2. Applying the attractive field and repulsive field together searches for a valid solution. 3. 3. Applying only the repulsive field alone prioritizes ensuring the solution is valid. This approach is quite effective, as shown in Section IV, although it still does not guarantee the final solution is fully collision free or globally optimal. Since collision avoidance is critical, the final solution used is the most optimal collision-free solution found along the way. When applied to all modules $i\in[1,N]$ concurrently, Algorithm 1 produces a collision-free set $\Phi$ of $(\phi_{0},A)_{i}$ $\forall i\in[1,N]$, solving 1. Each Modboat then executes (1) with those parameters for a single swim cycle (i.e. a single period), before repeating the process for the next cycle and set of desired forces. ### II-E Transition Solver Figure 5: To make finding valid transition sets more tractable, we decompose the structure into horizontal and vertical sub-problems. For each sub-problem, only interactions in the given direction are considered, and solution sets are found via AC [16] for each boat. The full collision-free solution is the intersection of the solution sets for each boat in the sub-problems. Although the procedure in Section II-D solves 1 and produces collision free movements, when used to generate a sequence of movements $\Phi_{k}$ for $k\in\mathbb{Z}^{+}$ it does not guarantee that the transition from the last position of one movement to the start of the next is itself collision free. An additional transition solver is needed to solve 2 and find a collision free set of transition paths $\Psi$. We assume, for ease of computation, that transitions take the entirety of $t_{trans}$ for all boats, and occur at a constant speed. Any Modboat can take either a clockwise or counterclockwise path from the last position of its previous cycle to the first position of its next. For any pair of modules, then, there are four possible transitions, each of which is a line in the phase-space in Fig. 2 and can be simply checked for collision. Of these, at least one is likely to be collision-free, but to increase the number of available options we also consider negating the amplitude of all boats in the next cycle, which shifts the start locations of the next cycle but does not otherwise affect the solution. This set of possible transitions is then run through the arc consistency (AC) algorithm [16] to find a valid collision- free transition for each boat pair. Because Modboat configurations are two-dimensional there many interactions between neighbors, since diagonal neighbors — while not directly neighbors — influence each other through shared connections. To make the search space more tractable, we split the spatial constraint by considering horizontal interactions and vertical interactions as two separate problems, as in Fig. 5. When AC is run on each row (column) of the horizontal (vertical) sub-problem, it generates a set of valid transitions for each boat. The full solution is then generated by intersecting the solution sets for each boat from the horizontal and vertical sub-solutions, which recovers the full spatial constraint. In our testing with simulated random inputs and prior locations, a valid transition set exists for all boats in over $99\%$ of cases for square configurations of Modboats with 2–5 boats to a side. However, since the number of possible sets to evaluate scales exponentially with the number of boats, this approach becomes intractable for larger structures; a more efficient strategy will need to be developed for larger configurations, but is left to future work. A critical thing to note is that the transition requires a finite $t_{trans}$ between cycles of (1), which means that decisions about generated forces are made at intervals of $T+t_{trans}$. Using a small $t_{trans}$ maintains the responsiveness of the overall controller, but the fast transitions introduce unwanted dynamic disturbances. Using a large $t_{trans}$ minimizes the dynamic disturbances, but slows the response of the overall controller. We minimize this impact by selecting the solution set that results in the minimum overall distance travelled, but we also expect most transitions to be small during normal operation, since the control generated in Section III should be relatively continuous unless sharp maneuvers are needed. ## III Control The methodology of Section II provides the parameters for (1) — namely $\phi_{0}$ and $A$ — for each boat $i$ in a structure given its shape and a set of desired values $\vec{F}_{des}$. To determine these desired values and implement holonomic control for the configuration as a whole, PID control is applied to the equations of motion as derived in [8]. Control for a desired yaw angle $\Theta$ is given in Eqs. 13, 14 and 15, where $\Omega$ is the observed angular velocity of the structure, $T$ is the period of (1), $I$ is the Modboat configuration’s moment of inertia, and $C_{R}$ is a drag coefficient [9]. Note that (13) includes a prediction of the yaw at the end of the current cycle, which accounts for the delay introduced by discrete control during sharp yaw maneuvers. $e_{\Theta}=\Theta_{des}-\left(\Theta_{obs}+\Omega T\right)$ (13) $\alpha=K_{p,\Theta}e_{\Theta}+K_{d\Theta}\frac{de_{\Theta}}{dt}$ (14) $\tau_{des}=I\alpha+C_{R}|\Omega|\Omega$ (15) Velocity control in [8, 9] — which presented steerable vehicles — considered the desired velocity as a surge (i.e. body-fixed frame) velocity value. Since the method presented in this paper results in a holonomic vehicle, we consider the desired values as $\vec{v}_{des}=[\begin{matrix}v_{x,des}&v_{y,des}\end{matrix}]^{T}$ expressed in the world frame. Just as in [8] an artificial linear acceleration is computed based on the observed error and then integrated into the commanded velocity $\vec{v}_{c}$, as in Eqs. 16, 17 and 18. $\vec{e}_{v}=\vec{v}_{des}-\vec{v}_{obs}$ (16) $\vec{a}_{y}=K_{p,v}\vec{e}_{v}+K_{d,v}\frac{d\vec{e}_{v}}{dt}$ (17) $\vec{v}_{c}=\vec{v}_{des}+(\gamma^{n-1}\sum_{i=0}^{n-1}\vec{a}_{y}+\vec{a}_{y})T$ (18) A diminishing coefficient $\gamma<1$ is added to (17) to prevent control lag due to excessive error accumulation. The commanded velocity is then converted to desired force values using the quadratic drag relationship [8] and converted to the boat frame using (19). $\vec{F}_{des}$ can then be constructed from (19) and (15) and decomposed into $\vec{\phi}_{0}$ and $\vec{A}$ using the methodology in Section II. $\begin{bmatrix}F_{x,des}\\\ F_{y,des}\end{bmatrix}=C_{L}\begin{bmatrix}\hphantom{-}\sin(\Theta_{obs})&\cos(\Theta_{obs})\\\ -\cos(\Theta_{obs})&\sin(\Theta_{obs})\end{bmatrix}\left(\vec{v}_{c}\odot\vec{v}_{c}\right)$ (19) ## IV Experiments Experiments were conducted in a $4.5~{}$\mathrm{m}$\times 3.0~{}$\mathrm{m}$\times 1.2~{}$\mathrm{m}$$ tank of still water equipped with an OptiTrack motion capture system that provides the real-time position, velocity, and orientation data for each boat at $120~{}$\mathrm{H}\mathrm{z}$$. The control methodology in Sections II and III was computed in Python on an offboard PC and, the resulting parameters for (1) and transitions were sent to each Modboat via WiFi at the beginning of each cycle. Experimental evaluation of the methodology presented in this work is ongoing, but preliminary results show great potential. The following four evaluations have so far been conducted using three boats in a parallel configuration, and the results shown in Fig. 6: 1. 1. $[\begin{matrix}v_{x}&v_{y}&\Theta\end{matrix}]_{des}=[\begin{matrix}0.04~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&\pi/2~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$ for a duration of $90~{}$\mathrm{s}$$. 2. 2. $[\begin{matrix}v_{x}&v_{y}&\Theta\end{matrix}]_{des}=[\begin{matrix}0.04~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$ for a duration of $90~{}$\mathrm{s}$$. 3. 3. $[\begin{matrix}v_{x}&v_{y}&\Theta\end{matrix}]_{des}=[\begin{matrix}0.03~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0.01~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&\pi/2~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$ for a duration of $90~{}$\mathrm{s}$$. 4. 4. $[\begin{matrix}v_{x}&v_{y}&\Theta\end{matrix}]_{des}=[\begin{matrix}0.04~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&\pi/2~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$ for a duration of $60~{}$\mathrm{s}$$, then $[\begin{matrix}v_{x}&v_{y}&\Theta\end{matrix}]_{des}=[\begin{matrix}0~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0.04~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&\pi/2~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$ for another $60~{}$\mathrm{s}$$. (a) (b) (c) (d) Figure 6: Experimental trajectories for the four preliminary tests. $[\begin{matrix}v_{x}&v_{y}&\Theta\end{matrix}]_{des}=$ (a) $[\begin{matrix}0.04~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&\pi/2~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$ (b) $[\begin{matrix}0.04~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$ (c) $[\begin{matrix}0.03~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0.01~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&\pi/2~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$. (d) $[\begin{matrix}0.04~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&\pi/2~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$, then $[\begin{matrix}0~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&0.04~{}$\mathrm{m}\mathrm{/}\mathrm{s}$&\pi/2~{}$\mathrm{r}\mathrm{a}\mathrm{d}$\end{matrix}]$. The COM pose is shown in orange. ## V Discussion The results in Fig. 6 show excellent preliminary performance. The Modboat configuration is able to travel in a desired direction in either a head-on (Fig. 6a) or sideways (Fig. 6b) orientation, and even mix velocities in the world frame (Fig. 6c). In a significant stress test, the controller is also able to provide excellent 90 degree turning performance, as evidenced by the trajectory in Fig. 6d, which shows little overshoot and good direction tracking during both legs of the trajectory. Notably, our controller has a difficult time maintaining a steady orientation (all tests in Fig. 6), but nevertheless maintains reasonable directional swimming. ## VI Conclusion In this work we have presented a potential-field based control approach to allow a group of three or more docked Modboats to function as a holonomic vehicle, improving on prior work [8, 9] that could only create a steerable vehicle. This method works for arbitrary structures of docked modules and has the potential to be scaled to large structures containing many modules. Preliminary experimental results have shown that this approach is effective in controlling the velocity and somewhat effective at controlling orientation in a few different testing scenarios. The major limitation of this strategy is the need for a fixed transition time outside the main swim cycle, which is necessary to avoid collisions but creates undesirable dynamics and slows down the overall controller response time. Future work will consider a transition strategy that minimizes this impact, as well as solution strategies for the swim cycle that minimize the need for transitions. Future work will also consider testing with larger numbers of modules and expanding our evaluation to non-controlled environments, like lakes or rivers. This will stress the ability of our controller to reject temporary disturbances. Developing this approach to work with larger configurations will also necessitate developing a more efficient transition solver that displays better scaling. ## Acknowledgment We thank Dr. M. Ani Hsieh for the use of her instrumented water basin in obtaining all of the testing data. ## References * [1] M. R. Palmer, Y. W. Shagude, M. J. Roberts, E. Popova, J. U. Wihsgott, S. Aswani, J. Coupland, J. A. Howe, B. J. Bett, K. E. Osuka, C. Abernethy, S. Alexiou, S. C. Painter, J. N. Kamau, N. Nyandwi, and B. Sekadende, “Marine robots for coastal ocean research in the western indian ocean,” _Ocean & Coastal Management_, vol. 212, p. 105805, 2021. * [2] J. Paulos, N. Eckenstein, T. Tosun, J. Seo, J. Davey, J. Greco, V. Kumar, and M. Yim, “Automated Self-Assembly of Large Maritime Structures by a Team of Robotic Boats,” _IEEE Transactions on Automation Science and Engineering_ , vol. 12, no. 3, pp. 958–968, 2015. * [3] I. O’Hara, J. Paulos, J. Davey, N. Eckenstein, N. Doshi, T. Tosun, J. Greco, J. Seo, M. Turpin, V. Kumar, and M. Yim, “Self-assembly of a swarm of autonomous boats into floating structures,” in _2014 IEEE International Conference on Robotics and Automation (ICRA)_ , Hong Kong, 2014, pp. 1234–1240. * [4] W. Wang, L. A. Mateos, S. Park, P. Leoni, B. Gheneti, F. Duarte, C. Ratti, and D. Rus, “Design, Modeling, and Nonlinear Model Predictive Tracking Control of a Novel Autonomous Surface Vehicle,” in _2018 IEEE International Conference on Robotics and Automation (ICRA)_ , Brisbane, Australia, 2018, pp. 6189–6196. * [5] W. Wang, T. Shan, P. Leoni, D. Fernández-Gutiérrez, D. Meyers, C. Ratti, and D. Rus, “Roboat II: A Novel Autonomous Surface Vessel for Urban Environments,” in _2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_ , Las Vegas, NV (Virtual), 2020, pp. 1740–1747. * [6] G. Knizhnik, “Modboat: A single-motor modular self-reconfigurable robot,” Feb 2022, https://www.modlabupenn.org/modboats/. * [7] G. Knizhnik and M. Yim, “Docking and Undocking a Modular Underactuated Oscillating Swimming Robot,” in _2021 IEEE International Conference on Robotics and Automation (ICRA)_ , Xi’an, China, 5 2021, pp. 6754–6760. * [8] G. Knizhnik and M. Yim, “Amplitude Control for Parallel Lattices of Docked Modboats,” in _2022 International Conference on Robotics and Automation (ICRA)_ , Philadelphia, PA, 5 2022, pp. 3027–3033. * [9] G. Knizhnik and M. Yim, “Collective Control for Arbitrary Configurations of Docked Modboats,” _arXiv:2209.04000 [cs.RO]_ , 9 2022. * [10] G. Knizhnik and M. Yim, “Thrust Direction Control of an Underactuated Oscillating Swimming Robot,” in _2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_ , Prague, Czech Republic (Virtual), 2021, pp. 8665–8670. * [11] K. Okumura, Y. Tamura, and X. Défago, “Iterative refinement for real-time multi-robot path planning,” in _2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_ , 2021, pp. 9690–9697. * [12] W. Yao, N. Qi, N. Wan, and Y. Liu, “An iterative strategy for task assignment and path planning of distributed multiple unmanned aerial vehicles,” _Aerospace Science and Technology_ , vol. 86, pp. 455–464, 2019. * [13] A. Richardson and E. Olson, “Iterative path optimization for practical robot planning,” in _2011 IEEE/RSJ International Conference on Intelligent Robots and Systems_ , 2011, pp. 3881–3886. * [14] S. Kopriva, D. Sislak, D. Pavlicek, and M. Pechoucek, “Iterative accelerated a* path planning,” in _49th IEEE Conference on Decision and Control (CDC)_ , 2010, pp. 1201–1206. * [15] G. Knizhnik, P. de Zonia, and M. Yim, “Pauses Provide Effective Control for an Underactuated Oscillating Swimming Robot,” _IEEE Robotics and Automation Letters_ , vol. 5, no. 4, pp. 5075–5080, 10 2020. * [16] A. K. Mackworth, “Consistency in networks of relations,” _Artificial Intelligence_ , vol. 8, no. 1, pp. 99–118, 1977.
# } [ ###### keywords: Finite Elements, MFEM library, Lagrange, Raviart-Thomas, Taylor-Hood, Laplace Equation, Navier-Stokes Equations. ###### keywords: Elementos Finitos, Librería MFEM, Lagrange, Raviart-Thomas, Taylor-Hood, Ecuación de Laplace, Ecuaciones de Navier-Stokes. [firstpage = 1, volume = 0, number = 0, month = 00, year = 1900, day = 00, monthreceived = 0, yearreceived = 1900, monthaccepted = 0, yearaccepted = 1900] authors[] authors department = Departamento de Matemáticas, institution = Universidad Nacional de Colombia, city = Bogotá D.C., country = Colombia ] We revise the finite element formulation for Lagrange, Raviart-Thomas, and Taylor-Hood finite element spaces. We solve Laplace equation in first and second order formulation, and compare the solutions obtained with Lagrange and Raviart-Thomas finite element spaces by changing the order of the shape functions and the refinement level of the mesh. Finally, we solve Navier- Stokes equations in a two dimensional domain, where the solution is a steady state, and in a three dimensional domain, where the system presents a turbulent behaviour. All numerical experiments are computed using MFEM library, which is also studied. Revisamos la formulación de elementos finitos para los espacios de elementos finitos de Lagrange, Raviart-Thomas y Taylor-Hood. Solucionamos la ecuación de Laplace en su formulación de primer y segundo orden, y comparamos las soluciones obtenidas con los espacios de elementos finitos de Lagrange y Raviart-Thomas al cambiar el orden de las funciones base y el nivel de refinamiento de la malla. Finalmente, resolvemos las ecuaciones de Navier- Stokes en un dominio bidimensional, donde la solución es un estado estable, y en un dominio tridimensional, donde el sistema presenta un comportamiento turbulento. Todos los experimentos numéricos se realizan utilizando la librería MFEM, la cual es también estudiada. ###### Contents 1. 1 Preliminaries 1. 1.1 Partial Differential Equations 1. 1.1.1 Laplace equation 2. 1.1.2 Navier-Stokes equations 2. 1.2 The Finite Element Method 1. 1.2.1 FEM for elliptic problems 2. 1.2.2 Lagrange finite elements 3. 1.2.3 Lagrange spaces of higher order 4. 1.2.4 Raviart-Thomas finite elements 5. 1.2.5 Taylor-Hood finite elements 3. 1.3 MFEM Library 1. 1.3.1 Information about the library 2. 1.3.2 Overview 3. 1.3.3 Code structure 2. 2 Lagrange vs. Raviart-Thomas finite elements 1. 2.1 Problem 2. 2.2 Code 3. 2.3 Tests 4. 2.4 Results 5. 2.5 Analysis 6. 2.6 Some other examples 3. 3 Numerical Experiments with NS 1. 3.1 2D Experiment: Steady State 2. 3.2 3D Experiment: Turbulence 4. 4 Conclusion and Perspectives 5. 5 Appendices 1. 5.1 Appendix A : Code for comparison 2. 5.2 Appendix B : Numerical values of the comparison 3. 5.3 Appendix C : MiniApp Code for Navier-Stokes ## 1 Preliminaries In this section we are going to recall the theoretical background needed in the rest of the paper. First, we are going to review the finite element methods used for the Laplace equation in second and first order form. We write the strong and weak form of the problem and introduce the Lagrange and mixed finite element spaces. We also introduce the MFEM library by giving an overview of its main characteristics and the general structure of a finite element code in MFEM. ### 1.1 Partial Differential Equations For the scope of this work, partial differential equations are of the form $F\left(u,p,\frac{\partial p}{\partial x_{i}},\frac{\partial u}{\partial x_{i}},\frac{\partial^{2}p}{\partial x_{i}^{2}},\frac{\partial^{2}u}{\partial x_{i}^{2}},f\right)=0;\ \ i=1,2,3,4$ where $u:\Omega\subseteq\mathbb{R}^{3}\rightarrow U\subseteq\mathbb{R}^{3}$, $p:\Omega\subseteq\mathbb{R}^{3}\rightarrow P\subseteq\mathbb{R}$, $f$ is a restriction parameter and $F$ is any mathematical expression within its variables. It is well known that PDEs have multiple solutions, in fact, there is a vectorial space consisting of all the solutions for a given equation. For this reason, the equation is usually presented with a boundary condition that forces a unique solution for the equation [2]. If some problem is being solved in a given domain $\Omega\subseteq\mathbb{R}^{3}$, whose boundary is $\Gamma$, then the problem has the form $\left\\{\begin{split}&F\left(u,p,\frac{\partial p}{\partial x_{i}},\frac{\partial u}{\partial x_{i}},\frac{\partial^{2}p}{\partial x_{i}^{2}},\frac{\partial^{2}u}{\partial x_{i}^{2}},f\right)=0\text{ in }\Omega,\\\ &F_{0}\left(u,p,\frac{\partial p}{\partial x_{i}},\frac{\partial u}{\partial x_{i}},\frac{\partial^{2}p}{\partial x_{i}^{2}},\frac{\partial^{2}u}{\partial x_{i}^{2}},g\right)=0\text{ in }\Gamma.\end{split}\right.$ The two main equations that we treat are the Laplace equation and the Navier- Stokes equation, whose represent a fluids phenomenon in real life. Note that $x_{1}=x,x_{2}=y,x_{3}=z,x_{4}=t$ (in time-space interpretation) and that, $p$ is the fluid’s pressure and $u$ is the fluid’s velocity. #### 1.1.1 Laplace equation Laplace equation consists on finding $p:\Omega\subseteq\mathbb{R}^{3}\rightarrow\mathbb{R}$ such that $-\Delta p=f,\text{ in }\Omega$ (1) where $f:\Omega\rightarrow\mathbb{R}$ is a given function and $\Delta p=\nabla\cdot\nabla p=\frac{\partial^{2}p}{\partial x^{2}}+\frac{\partial^{2}p}{\partial y^{2}}+\frac{\partial^{2}p}{\partial z^{2}}$ [4]. It is clear that the equation is a partial differential equation of the form presented before in Section 1.1 because it only depends on the second order derivatives of $p$. Now, we present two common ways of imposing a boundary condition to the problem: Dirichlet and Neumann. Let $\Gamma$ be the boundary of $\Omega$. Dirichlet boundary condition The Dirichlet boundary condition is $p=g\text{ in }\Gamma$ (2) where $g:\Gamma\rightarrow\mathbb{R}$ is a given function [4]. If $g=0$, the condition is called homogeneous Dirichlet boundary condition. Neumann boundary conditions The Neumann boundary condition is $-\nabla p\cdot\eta=h\text{ in }\Gamma$ (3) where $h:\Gamma\rightarrow\mathbb{R}$ is a given function and $\eta$ is the boundary’s normal vector [4]. As seen on [4], problem (1) can be stated with the two types of boundary condition (2) and (3), by imposing each condition on different parts of $\Gamma$. However, the version of the problem presented later on Section 1.2.2 has the homogeneous Dirichlet boundary condition. Finally, the Laplace equation (1) can be formulated on two different ways: first order formulation and second order formulation. The second order formulation is the one presented already on (1), which involves second order derivatives. Now, if we set $u=\nabla p$, the problem can be stated [4] as $\left\\{\begin{split}\mbox{div}(u)=-f\text{ in }\Omega,\\\ u=\nabla p\text{ in }\Omega,\end{split}\right.$ (4) which is the first order formulation for the problem and, notice that it only involves first order derivatives of $u$ and $p$. On Section 2, we compare both formulations of Laplace equation (1) and (4). We use Lagrange finite elements to solve (1) and Raviart-Thomas finite elements to solve (4). #### 1.1.2 Navier-Stokes equations In this section we revise incompressible Stokes and Navier-Stokes equations with Dirichlet boundary condition. First, incompressible Navier-Stokes equations consist on finding the velocity $u:\Omega\rightarrow\mathbb{R}^{3}$ and the pressure $p\rightarrow\mathbb{R}$ that solve the system of equations (5) [8]. $\begin{split}\frac{\partial u}{\partial t}+(u\cdot\nabla)u-\nu\Delta u+\nabla p=f,\text{ in }\Omega,\\\ \nabla\cdot u=0,\text{ in }\Omega,\\\ u=g,\text{ in }\Gamma,\end{split}$ (5) where $\Omega\subseteq\mathbb{R}^{3}$ is the spatial domain, $\Gamma=\partial\Omega$ the boundary of the domain, $\nu$ is called the kinematic viscosity (more information below), $f$ is the forcing term (given function), and $g$ is the specified Dirichlet boundary condition. On the other hand, when removing the non-linear term $(u\cdot\nabla)u$ from the first equation, we get Stokes linear equations which is system of equations (6) [8]. $\begin{split}\frac{\partial u}{\partial t}-\nu\Delta u+\nabla p=f,\text{ in }\Omega,\\\ \nabla\cdot u=0,\text{ in }\Omega,\\\ u=g,\text{ in }\Gamma.\end{split}$ (6) Take into account that $u(x,y,z,t)=(u_{x},u_{y},u_{z})$ is a vector for each point in time-space and that $p(x,y,z,t)=p$ is a scalar for each point in time-space. For real life fluid problems modeled by these equations, $(x,y,z)$ is the position of a fluid’s particle in space and $t$ is the time. Also, the equation $\nabla\cdot u=0$ is the one that establishes the incompressible condition for the fluid. Kinematic Viscosity On the modeling of fluid flow, as presented on [7], a very important parameter appears and its called Reynolds number, $Re$. On [7], it is defined as $Re=\frac{\rho LU}{\mu},$ where $\rho$ is the density of the fluid, $L$ is the characteristic linear dimension of the domain of the flow, $U$ is some characteristic velocity, and $\mu$ is the fluid’s viscosity. For example, as seen on [7], $L$ is defined supposing that $\Omega=[0,L]^{n}$, i.e., $L$ is the length of the domain on each direction, and $U$ can be taken as the square root of the average initial kinetic energy in $\Omega$. However, for the purpose of this work we neglect the parameters $L$ and $U$ in such way that $Re=\frac{\rho}{\mu}$. Therefore, the kinematic viscosity of a fluid is, according to [7], $\nu:=\frac{\mu}{\rho}=\frac{1}{Re}.$ (7) Notice that if the viscosity of the fluid, $\mu$, is higher and the density of the fluid, $\rho$, is lower, then, the kinematic viscosity of the fluid, $\nu$, is higher. This parameter quantifies the resistance that a fluid imposes to movement due to an external force, like gravity [7]. On Section 3, we do some numerical experiments in 2D and 3D using the Navier Miniapp of MFEM library, which solves (5), by using the formulation presented on Section 1.2.5. ### 1.2 The Finite Element Method First, on Section 1.2.1 we show how the solution for a differential equation is also a solution for a minimization problem and a variational problem. This serves as a basic case for showing that partial differential equations are, in fact, solved via minimization or variational problems; which are the ones solved with finite element methods. Then, in Sections 1.2.2 and 1.2.4 we study two finite element methods. On both of them, the following procedure was applied: 1. 1. Consider the problem of solving Laplace equation (23) with homogeneous Dirichlet boundary condition. 2. 2. Multiply by some function (test function) and integrate by parts. Apply boundary conditions. 3. 3. Discretize the domain and select finite-dimensional function spaces for the solution and the test functions. 4. 4. Produce a matrix system to solve for solution weights in the linear combination representation of the approximated solution. The basis functions that form part of the finite-dimensional spaces are called shape functions. In Lagrange formulation, those are the functions in $V_{h}$, and in mixed formulation, those are the functions in $H_{h}^{k}$ and $L_{h}^{k}$, where the parameter $h$ denotes the size of the elements in the triangulation of the domain. Moreover, in Lagrange formulation the boundary condition is essential, and in mixed formulation, it is natural. Furthermore, on Section 1.2.3 we show high order shape functions used in Lagrange finite element method and, on Section 1.2.5 we study the most common finite elements used to approximate the solution for Navier-Stokes equations. #### 1.2.1 FEM for elliptic problems Let $\mathbb{D}$ be the two-point boundary value problem (8) taken from [2]. $\begin{split}&-u^{\prime\prime}(x)=f(x),\ x\in(0,1),\\\ &u(0)=u(1)=0,\end{split}$ (8) where $f$ is a given continuous function and $u^{\prime\prime}(x)=\frac{d^{2}u(x)}{dx^{2}}$. Notice that this problem is a partial differential equation of the form presented on section 1.1 but with all the functions being of the form $\mathbb{R}\rightarrow\mathbb{R}$ and $\Omega=(0,1)$ being a 1D domain, with homogeneous boundary condition ($0$ on $\Gamma=\\{0,1\\}$). By integrating $-u^{\prime\prime}(x)=f(x)$ twice, it is clear that the problem (8) has a unique solution $u$ [2]. For example, if $f(x)=e^{x}$ $\begin{split}&-u^{\prime\prime}(x)=e^{x}\\\ \implies&-u^{\prime}(x)=e^{x}+c_{1}\\\ \implies&-u(x)=e^{x}+c_{1}x+c_{2}\end{split}$ and by applying boundary conditions, $\left\\{\begin{split}&u(0)=0\implies c_{2}=-1,\\\ &u(1)=0\implies c_{1}=1-e.\end{split}\right.$ It follows that $u(x)=-e^{x}+(e-1)x+1$ is the unique solution. Recall that the boundary conditions force the problem to have a unique solution, as mentioned previously on the work. Now, following [2], define the linear space $V$ of all continuous functions on $[0,1]$ that vanish at $\\{0,1\\}$, whose derivative is piecewise continuous and bounded on $[0,1]$. Also, define the linear functional $F:V\rightarrow\mathbb{R}$, by $F(v)=\frac{1}{2}\int_{0}^{1}\left[v^{\prime}(x)\right]^{2}dx-\int_{0}^{1}f(x)v(x)dx.$ With this settled, let $\mathbb{M}$ be the optimization problem of finding $u\in V$ such that $F(u)\leq F(v)$ (9) for all $v\in V$, and let $\mathbb{V}$ be the variational problem of finding $u\in V$ such that $\int_{0}^{1}u^{\prime}(x)v^{\prime}(x)dx=\int_{0}^{1}f(x)v(x)dx$ (10) for all $v\in V$. Remark that if $u,v\in V$ then $w=u+\alpha v\in V$ for any $\alpha\in\mathbb{R}$, because $w$ is continuous in $[0,1]$; $w(1)=w(0)=u(0)+\alpha v(0)=0$, i.e. vanishes on $\\{0,1\\}$; and $w^{\prime}=u^{\prime}+\alpha v^{\prime}$ is piecewise continuous and bounded on $[0,1]$. The rest of this section is focused on showing that $\mathbb{D}$, $\mathbb{M}$ and $\mathbb{V}$ are equivalent problems [2]. Equivalence $\mathbb{D}\iff\mathbb{V}$ Let $u_{\mathbb{D}}$ be the solution for $\mathbb{D}$. Then, multiply $-u_{\mathbb{D}}^{\prime\prime}(x)=f(x)$ on both sides by some $v\in V$ (this $v$ is called a test function) and integrate to obtain $-\int_{0}^{1}u_{\mathbb{D}}^{\prime\prime}(x)v(x)dx=\int_{0}^{1}f(x)v(x)dx.$ (11) Using the formula for integration by parts, $\int_{0}^{1}a(x)b^{\prime}(x)dx=a(1)b(1)-a(0)b(0)-\int_{0}^{1}a^{\prime}(x)b(x)dx,$ (12) with $a(x)=v(x)$ and $b(x)=u_{\mathbb{D}}^{\prime}(x)$, and applying the fact that $v(0)=v(1)=0$ we get $\int_{0}^{1}u_{\mathbb{D}}^{\prime\prime}(x)v(x)dx=-\int_{0}^{1}v^{\prime}(x)u_{\mathbb{D}}^{\prime}(x)dx.$ (13) Then, replacing (13) on equation (11) we get that $\int_{0}^{1}v^{\prime}(x)u_{\mathbb{D}}^{\prime}(x)dx=\int_{0}^{1}f(x)v(x)dx.$ (14) Notice that (14) is the equation associated to the variational problem $\mathbb{V}$ (see (10)), and as $v\in V$ was arbitrary, we have that $u_{\mathbb{D}}$ satisfies (14) (and so, (10)) for all $v\in V$. Therefore, $u_{\mathbb{D}}$ is also a solution for $\mathbb{V}$.✓ On the other hand, let $u_{\mathbb{V}}\in V$ be the solution for $\mathbb{V}$ (on [2], it is shown that the solution for $\mathbb{V}$ is unique). Then, we have by (10) that, for all $v\in V$, $\int_{0}^{1}u_{\mathbb{V}}^{\prime}(x)v^{\prime}(x)dx-\int_{0}^{1}f(x)v(x)dx=0.$ (15) Now, applying integration by parts (12) in the same way as before, we obtain (13). Replacing (13) on (15) and unifying the integral, we get $-\int_{0}^{1}[u_{\mathbb{V}}^{\prime\prime}(x)+f(x)]v(x)dx=0$ (16) for all $v\in V$. Then, by (18) we have that $u_{\mathbb{V}}^{\prime\prime}(x)+f(x)=0$ for $x\in(0,1)$. In other words, $-u_{\mathbb{V}}^{\prime\prime}(x)=f(x),\ x\in(0,1).$ (17) Notice that (17), along with the fact that $u_{\mathbb{V}}(0)=u_{\mathbb{V}}(1)=0$, is the equation associated to the differential problem $\mathbb{D}$ (see (8)). Therefore, $u_{\mathbb{V}}$ is also a solution for $\mathbb{V}$ as long as $u^{\prime\prime}(x)$ exists and is continuous (regularity assumption). But, this last condition for $u_{\mathbb{V}}$ holds, as seen on [2], so, the result holds.✓ To complete the proof, we have to prove (18), which is exercise 1.1 from [2]: $\begin{split}\text{If $w$ is continuous on $[0,1]$ and}\\\ \int_{0}^{1}w(x)v(x)dx=0\text{ for all }v\in V,\\\ \text{then }w(x)=0\text{ for all }x\in(0,1).\end{split}$ (18) Proof: By contradiction, suppose that $w(x)\not=0$ for some $x\in(0,1)$. Then, $w(x_{0})=c\in\mathbb{R}$ for $x_{0}\in(0,1)$. Without loss of generality, suppose that $c>0$ (for $c<0$ the argument is analogous). As $w$ is continuous, there is an interval centered at $x_{0}$, $I=(x_{0}-\delta,x_{0}+\delta)$, such that $f(x)>0$ for $x\in I$. Now, define $v(x)=\left\\{\begin{split}0,\ if\ &x\not\in I,\\\ \frac{c}{\delta}[x+(\delta- x_{0})],\ if\ &x\in(x_{0}-\delta,x_{0})\\\ -\frac{c}{\delta}[x-(\delta+x_{0})],\ if\ &x\in(x_{0},x_{0}+\delta),\end{split}\right.$ which is a function sketched on Figure 1. Figure 1: Sketch of the proof for (18) involving an arbitrary function $w(x)$ (solid black line) and the built function $v\in V$ (solid red line). Notice that $v(0)=v(1)=0$ because there is always a $\delta>0$ such that $0<x_{0}-\delta$ and $x_{0}+\delta<1$, and $v(x)=0$ for $x\not\in I$. Also, $v(x)$ is continuous on $[0,1]$ by construction (it was built with 4 lines that have connections in $(x_{0}-\delta,0),(x_{0},c)$ and $(x_{0}+\delta,0)$). Moreover, $v^{\prime}(x)=\left\\{\begin{split}0,\ if\ &x\not\in I,\\\ \frac{c}{\delta},\ if\ &x\in(x_{0}-\delta,x_{0}),\\\ -\frac{c}{\delta},\ if\ &x\in(x_{0},x_{0}+\delta),\end{split}\right.$ is clearly piecewise continuous and bounded on $[0,1]$. Therefore, $v\in V$. However, as $w(x)v(x)>0$ for $x\in I$ and $w(x)v(x)=0$ for $x\not\in I$, then $\int_{0}^{1}w(x)v(x)dx=\int_{x_{0}-\delta}^{x_{0}+\delta}w(x)v(x)dx>0.$ In other words, if $w(x)\not=0$ for some $x\in(0,1)$, then we found $v\in V$ such that $\int_{0}^{1}w(x)v(x)dx\not=0$, which contradicts the hypothesis that $\int_{0}^{1}w(x)v(x)dx=0$ for all $v\in V$. Therefore, $w(x)=0$ for all $x\in(0,1)$.∎ Equivalence $\mathbb{V}\iff\mathbb{M}$ Let $u_{\mathbb{V}}\in V$ be the solution for $\mathbb{V}$. Take some $v\in V$ and set $w=v-u_{\mathbb{V}}\in V$. Then $\begin{split}&F(v)=F(u+w)\\\ =&\frac{1}{2}\int_{0}^{1}\left[u_{\mathbb{V}}^{\prime}(x)+w^{\prime}(x)\right]^{2}dx-\int_{0}^{1}f(x)[u_{\mathbb{V}}(x)+w(x)]dx\\\ =&\frac{1}{2}\left(\int_{0}^{1}[u_{\mathbb{V}}^{\prime}(x)]^{2}dx+\int_{0}^{1}2u_{\mathbb{V}}^{\prime}(x)w^{\prime}(x)dx+\int_{0}^{1}[w^{\prime}(x)]^{2}dx\right)\\\ &-\int_{0}^{1}f(x)u_{\mathbb{V}}(x)dx-\int_{0}^{1}f(x)w(x)dx\\\ =&\frac{1}{2}\int_{0}^{1}[u_{\mathbb{V}}^{\prime}(x)]^{2}dx-\int_{0}^{1}f(x)u_{\mathbb{V}}(x)dx+\int_{0}^{1}u_{\mathbb{V}}^{\prime}(x)w^{\prime}(x)dx\\\ &-\int_{0}^{1}f(x)w(x)dx+\frac{1}{2}\int_{0}^{1}[w^{\prime}(x)]^{2}dx\\\ \stackrel{{\scriptstyle\eqref{equiv1:eq4}}}{{=}}&\frac{1}{2}\int_{0}^{1}[u_{\mathbb{V}}^{\prime}(x)]^{2}dx-\int_{0}^{1}f(x)u_{\mathbb{V}}(x)dx+\frac{1}{2}\int_{0}^{1}[w^{\prime}(x)]^{2}dx\\\ =&F(u_{\mathbb{V}})+\frac{1}{2}\int_{0}^{1}[w^{\prime}(x)]^{2}dx\geq F(u_{\mathbb{V}}).\end{split}$ In other words, we have that $F(v)\geq F(u_{\mathbb{V}}).$ (19) Notice that (19) is the equation associated to the optimization problem $\mathbb{M}$ (see (9)), and as $v\in V$ was arbitrary, we have that $u_{\mathbb{V}}$ satisfies (19) (and so, (9)) for all $v\in V$. Therefore, $u_{\mathbb{V}}$ is also a solution for $\mathbb{M}$.✓ On the other hand, let $u_{\mathbb{M}}\in V$ be a solution for $\mathbb{M}$. Then, for any $v\in V$ and any $\alpha\in\mathbb{R}$ we have that $u_{\mathbb{M}}+\alpha v\in V$, and so $F(u_{\mathbb{M}})\leq F(u_{\mathbb{M}}+\alpha v)$ by (9). Therefore, the minimum for $F(u_{\mathbb{M}}+\alpha v)$ is achieved at $\alpha=0$. Now, define $g(\alpha)=F(u_{\mathbb{M}}+\alpha v)$, which is a differentiable function [2]. We have that $\begin{split}&g(\alpha)=F(u_{\mathbb{M}}+\alpha v)\\\ =&\frac{1}{2}\int_{0}^{1}\left[u_{\mathbb{M}}^{\prime}(x)+\alpha v^{\prime}(x)\right]^{2}dx-\int_{0}^{1}f(x)[u_{\mathbb{M}}(x)+\alpha v(x)]dx\\\ =&\frac{1}{2}\int_{0}^{1}[u_{\mathbb{M}}^{\prime}(x)]^{2}dx+\alpha\int_{0}^{1}u_{\mathbb{M}}^{\prime}(x)v^{\prime}(x)dx+\frac{\alpha^{2}}{2}\int_{0}^{1}[v^{\prime}(x)]^{2}dx\\\ &-\int_{0}^{1}f(x)u_{\mathbb{M}}(x)dx-\alpha\int_{0}^{1}f(x)v(x)dx.\end{split}$ (20) Since $g(\alpha)=F(u_{\mathbb{M}}+\alpha v)$ has a minimum at $\alpha=0$, then $g^{\prime}(0)=0$. Therefore, taking the derivative on (20) and replacing $\alpha=0$, we get $\begin{split}&g^{\prime}(\alpha)=\int_{0}^{1}u_{\mathbb{M}}^{\prime}(x)v^{\prime}(x)dx+\alpha\int_{0}^{1}[v^{\prime}(x)]^{2}dx-\int_{0}^{1}f(x)v(x)dx\\\ \implies&g^{\prime}(0)=\int_{0}^{1}u_{\mathbb{M}}^{\prime}(x)v^{\prime}(x)dx-\int_{0}^{1}f(x)v(x)dx.\end{split}$ (21) Finally, as $g^{\prime}(0)=0$, we get from (21), that $\int_{0}^{1}u_{\mathbb{M}}^{\prime}(x)v^{\prime}(x)dx=\int_{0}^{1}f(x)v(x)dx.$ (22) Notice that (22) is the equation associated to the variational problem $\mathbb{V}$ (see (10)), and as $v\in V$ was arbitrary, we have that $u_{\mathbb{M}}$ satisfies (22) (and so, (10)) for all $v\in V$. Therefore, $u_{\mathbb{M}}$ is also a solution for $\mathbb{V}$.✓ As mentioned before, on [2], it is shown that the solution for $\mathbb{V}$ is unique. Therefore, as $\mathbb{V}$ and $\mathbb{M}$ are equivalent, then the solution for $\mathbb{M}$ is also unique. #### 1.2.2 Lagrange finite elements In this section we consider the Laplace equation [2], $\begin{split}-\Delta&p=f\text{ in }\Omega,\\\ &p=0\text{ in }\Gamma,\end{split}$ (23) where $\Omega\subseteq\mathbb{R}^{2}$ is an open-bounded domain with boundary $\Gamma$, $f$ is a given function and $\Delta p=\frac{\partial^{2}p}{\partial x^{2}}+\frac{\partial^{2}p}{\partial y^{2}}$. Following [2], consider the space $V=\\{v:v\text{ continuous on }\Omega,\frac{\partial v}{\partial x},\frac{\partial v}{\partial y}\text{ piecewise continuous on }\Omega\text{ and }v=0\text{ on }\Gamma\\}.$ Alternatively we can work in the Sobolev space (see [2, 4, 3]) $H^{1}(\Omega)=\\{v\in L^{2}(\Omega)\ \Big{|}\ \frac{\partial v}{\partial x},\frac{\partial v}{\partial y}\in L^{2}(\Omega)\\}.$ Here $L^{2}(\Omega)=\\{v:\Omega\rightarrow\mathbb{R}\ \Big{|}\int_{\Omega}v^{2}<\infty\\}$. We multiply the first equation of (23) by some $v\in V$ (referred to as test function) and integrate over $\Omega$ to obtain $-\int_{\Omega}\Delta p\ v=\int_{\Omega}f\ v.$ (24) Applying divergence theorem we obtain the Green’s formula ([2]), $-\int_{\Omega}\Delta p\ v=\int_{\Omega}\nabla v\cdot\nabla p-\int_{\Gamma}v\ \nabla p\cdot\eta,$ (25) where $\eta$ is the outward unit normal to $\Gamma$. Since $v=0$ on $\Gamma$, the third integral equals $0$. Note that the boundary integral does not depend on $p$’s value on $\Gamma$ but rather on the normal derivative of $p$ in $\Gamma$. Due to this fact the boundary condition $p=0$ on $\Gamma$ is know as an essential boundary condition. Then, replacing (25) on (24), we get, $\int_{\Omega}\nabla v\cdot\nabla p=\int_{\Omega}f\ v.$ (26) This holds for all $v\in V$. This is called weak formulation of the Laplace equation (23). We remark that, according to [2], if $p\in V$ satisfies (26) for all $v\in V$ and is sufficiently regular, then $p$ also satisfies (23), i.e., it’s a (classical) solution for our problem. For more details see [2] and references therein. In order to set the problem for a computer to solve it, we are going to discretize it and encode it into a linear system. First, consider a triangulation $T_{h}$ of the domain $\Omega$. This is, $T_{h}=\\{K_{1},\dots,K_{m}\\}$ a set of non-overlapping triangles such that $\Omega=K_{1}\cup\dots\cup K_{m}$ and no vertex ($N_{i}$) of one triangle lies on the edge of another triangle, as seen on Figure 2. Figure 2: A triangulation for a given domain $\Omega$ showing a node $N_{i}$, and formed by some triangles $K_{j}$. Note: Triangles have been separated in the edges to take a better look, but the triangulation has no empty spaces. Visualization: [6]. The $h$ in the notation $T_{h}$ is a measure of the size of mesh, it usually refers to a typical element diameter or perhaps to the largest element diameter in the triangulation. In this manuscript $h$ is defined by $h=\max\\{\mbox{diam}(K):K\in T_{h}\\}$ where $\mbox{diam}(K)=\text{longest side of }K$. Now, let $V_{h}=\\{v:v\text{ continuous on }\Omega,v|_{K}\text{ linear for }K\in T_{h},\ v=0\text{ on }\Gamma\\}$. We consider the nodes ($N_{1},\dots,N_{M}$) of the triangulation that are not on the boundary, because $p=0$ there, and we define some functions $\varphi_{j}\in V_{h}$ in such way that $\varphi_{j}(N_{i})=\left\\{\begin{array}[]{lcc}1&,\ i=j\\\ \\\ 0&,\ i\not=j\\\ \end{array}\right.$ for $i,j=1,\dots,M$. See Figure 3 for an illustration of $\varphi_{j}\in V_{h}$. Figure 3: Illustration of the function $\varphi_{j}$ produced with MFEM library. On the left picture we plot the function $\varphi_{j}$. On the right picture we show the same plot depicting the elements of the underlying triangulation. Visualization: [6]. With this, $V_{h}=\mbox{span}\\{\varphi_{i}:i=1,\dots,M\\}$ and for any given $v\in V_{h}$ we have $v(x)=\sum_{j=1}^{M}\xi_{j}\varphi_{j}(x),$ with $\xi_{j}=v(N_{j})$ and $x\in\Omega\cup\Gamma$. So, $V_{h}$ is a finite- dimensional subspace of $V$. See [2] for details. Then, if $p_{h}\in V_{h}$ satisfies (26) for all $v\in V_{h}$, in particular, $\int_{\Omega}\nabla p_{h}\cdot\nabla\varphi_{j}=\int_{\Omega}f\ \varphi_{j},\ \ j=1,\dots,M.$ (27) Since $\nabla p_{h}=\sum_{i=1}^{M}\xi_{i}\nabla\varphi_{i}$ with $\xi_{i}=p_{h}(N_{i})$, replacing on (27) we get, $\sum_{i=1}^{M}\xi_{i}\int_{\Omega}\nabla\varphi_{i}\cdot\nabla\varphi_{j}=\int_{\Omega}f\ \varphi_{j},\ \ j=1,\dots,M.$ (28) Finally, (28) is a linear system of $M$ equations and $M$ unknowns ($\xi_{1},\dots,\xi_{M}$), which can be written as, $A\xi=b,$ (29) where $A[i,j]=\int_{\Omega}\nabla\varphi_{i}\cdot\nabla\varphi_{j}$, $\xi[i]=p_{h}(N_{i})$ and $b[i]=\int_{\Omega}f\ \varphi_{i}$. We can solve (29) with MFEM library as done on Section 2 (Example#1). Before continuing with the next section, let us show some theorems regarding the error between the solution $p$ for problem (23) and its approximation $p_{h}$. The theorems are presented on the general form but, after the proof, we show how is it used on our particular problem. For the following theorem, $A$ is a bilinear form on $V\times V$ and $L$ is a linear form on $V$ such that 1. 1. $A$ is continuous ($\mathcal{C}$) There is a constant $\gamma>0$ such that $|A(v,w)|\leq\gamma||v||_{V}||w||_{V},\ \forall v,w\in V.$ 2. 2. $A$ is $V$-elliptic ($V_{\epsilon}$) There is a constant $\alpha>0$ such that $\alpha||v||_{V}^{2}\leq A(v,v).$ ###### Theorem 1.1. [2] Céa Lemma If $p\in V$ is the solution for $A(p,v)=L(v),\ \forall v\in V$ and $p_{h}\in V_{h}\subset V$ is the solution for $A(p_{h},v_{h})=L(v_{h}),\ \forall v_{h}\in V_{h}$ then, $||p-p_{h}||_{V}\leq\frac{\gamma}{\alpha}||p-v_{h}||_{V},\ \forall v_{h}\in V_{h}.$ ###### Proof 1.2. Using the hypothesis that $A(p,v)=L(v)$ for all $v\in V$, along with the fact that $V_{h}\subset V$, we have that $A(p,v_{h})=L(v_{h})$ for all $v_{h}\in V_{h}$. Now, we subtract the last equation with the one given as hypothesis, $A(p_{h},v_{h})=L(v_{h})$, to get, $A(p-p_{h},v_{h})=L(v_{h})-L(v_{h})=0$ for all $v_{h}\in V_{h}$. For an arbitrary $w\in V_{h}$, let $v_{h}=p_{h}-w\in V_{h}$. Then, $\begin{split}&\alpha||p-p_{h}||^{2}_{V}\\\ \leq&A(p-p_{h},p-p_{h})+0\quad(\text{See }V_{\epsilon})\\\ =&A(p-p_{h},p-p_{h})+A(p-p_{h},w)\\\ =&A(p-p_{h},p-p_{h}+w)\\\ =&A(p-p_{h},p-p_{h}+p_{h}-v_{h})\\\ =&A(p-p_{h},p-v_{h})\\\ \leq&\gamma||p-p_{h}||_{V}||p-v_{h}||_{V}\quad(\text{See }\mathcal{C}).\end{split}$ In other words, we have that $\alpha||p-p_{h}||^{2}_{V}\leq\gamma||p-p_{h}||_{V}||p-v_{h}||_{V}$ for all $v_{h}\in V_{h}$. Dividing by $\alpha||p-p_{h}||_{V}$ on both sides, we get, $||p-p_{h}||_{V}\leq\frac{\gamma}{\alpha}||p-v_{h}||_{V}$ for all $v_{h}\in V_{h}.$ Note that $||p-p_{h}||_{V}\not=0$ because $p_{h}$ is supposed to be an approximation for $p$, and not the exact solution. On the particular case of (26), we have that $A(p,v)=\int_{\Omega}\nabla v\cdot\nabla p$ and $L(v)=\int_{\Omega}f\ v.$ Using Cauchy-Schwarz Inequality for Integrals we have that $A(p,v)=\int_{\Omega}\nabla v\cdot\nabla p\leq\sqrt{\int_{\Omega}|\nabla p|^{2}}\cdot\sqrt{\int_{\Omega}|\nabla v|^{2}}=1\cdot||p||_{V}||v||_{V}.$ In other words, the parameter for the continuity of the bilinear form of our particular case is $\gamma=1$. Also, notice that $1\cdot||v||_{V}^{2}=\left(\sqrt{\int_{\Omega}|\nabla v|^{2}}\right)^{2}=\int_{\Omega}|\nabla v|^{2}=\int_{\Omega}\nabla v\nabla v=A(v,v).$ That is, the parameter for the $V$-ellipticity of the bilinear form for our particular case is $\alpha=1$. Therefore, Céa Lemma ensures that $\sqrt{\int_{\Omega}|p-p_{h}|^{2}}\leq\sqrt{\int_{\Omega}|p-v_{h}|^{2}}$ for all $v_{h}\in V_{h}$. Now, before presenting the second theorem, define the operator $\mathcal{I}^{h}:\mathcal{C}(\Omega)\rightarrow V_{h}$ that associates every continuous function whose domain is $\Omega$, $f\in\mathcal{C}(\Omega)$, with a function $\mathcal{I}^{h}f\in V_{h}$ [4]. This operator is an interpolation operator defined by the nodes of the triangulation of $\Omega$: if $T_{h}=\\{K_{1},\dots,K_{m}\\}$ is a triangulation of $\Omega$, then $\mathcal{I}^{h}f(K_{i})=f(K_{i}),\ i=1,\dots,m$. ###### Theorem 1.3. [4] Let $\Omega\subseteq\mathbb{R}^{2}$ be a polygonal domain. Let $\\{T_{h_{i}}\\}$ be a family of triangulations of $\Omega$, with $T_{h_{i}}$ being a quasi-uniform triangulation. Then, $||\mathcal{I}^{h}p-p||_{1}\leq c\cdot h\cdot||p||_{2},$ where $||f||_{2}=\left(\int_{\Omega}f(x)^{2}+|\nabla f(x)|^{2}+\sum_{ij}(\partial_{ij}f)^{2}dx\right)^{1/2}.$ The proof of this theorem is out of the scope for this work. However, it can be checked on [4]. In summary, for our case, Céa Lemma states that Laplace problem can be approximated by the space $V_{h}$, and Theorem 1.3 states that the approximation is a good one. #### 1.2.3 Lagrange spaces of higher order This short section has the purpose of explaining Lagrange finite element spaces of higher order. Previously, on Section 1.2.2, when introducing Lagrangian elements, the shape function’s degree was set to one. Better approximations can be obtained by using polynomials of higher order. One can define, for a fixed order $k$, $\begin{split}V^{k}_{h}=\\{v:&v\text{ continuous on }\Omega,\\\ &v|_{K}\text{ polynomial of order at most }k,K\in T_{h},\ v=0\text{ on }\Gamma\\}.\end{split}$ For example, as seen in [4], the space of Bell triangular finite elements for a given triangulation $T_{h}$ is the space of functions that are polynomials of order 5 when restricted to every triangle $K\in T_{h}$. That is, if $v$ is in this space, then, $v|_{K}(x,y)=a_{1}x^{5}+a_{2}y^{5}+a_{3}x^{4}y+a_{4}xy^{4}+\dots+a_{16}x+a_{17}y+a_{18}$ for all $K\in T_{h}$. Here, the constants $a_{i},\ i=1,\dots,18$ correspond to $v$’s DOF (degrees of freedom). On Figures 4 and 5, we present a visualization of some shape functions of different orders. We encourage the reader to compare them with Figure 3 and notice the degree of the polynomial in the nonzero part of the shape functions. Figure 4: Illustration of finite element basis (shape) functions of order 2. On the left picture we show one continuous basis function. On the right we also show the underlying triangulation. Visualization: [6]. Figure 5: Illustration of finite elements basis function of orders 5 (left) and 10 (right). Visualization: [6]. #### 1.2.4 Raviart-Thomas finite elements First, let’s define some important spaces, where $\Omega$ is a bounded domain in $\mathbb{R}^{2}$ and $\Gamma$ its boundary. See [2, 3, 4] and references therein for details. The space of all square integrable functions, $L^{2}(\Omega)=\\{v:\Omega\rightarrow\mathbb{R}\ \Big{|}\int_{\Omega}v^{2}<\infty\\}.$ We also use the first order Sobolev space, $H^{1}(\Omega)=\\{v\in L^{2}(\Omega)\ \Big{|}\ \frac{\partial v}{\partial x},\frac{\partial v}{\partial y}\in L^{2}(\Omega)\\}$ and the subspace of $H^{1}(\Omega)$ of functions with vanishing value on the boundary, $H_{0}^{1}(\Omega)=\\{v\in H^{1}(\Omega)\ |\ v=0\ on\ \Gamma\\}.$ We also introduce the space of square integrable vector functions with square integrable divergence, $H(\mbox{div};\Omega)=\\{\mathbf{v}\in L^{2}(\Omega)\times L^{2}(\Omega)\ |\ \mbox{div}(\mathbf{v})\in L^{2}(\Omega)\\}.$ As above, let $\Omega\in\mathbb{R}^{2}$ be a bounded domain with boundary $\Gamma$ and consider problem (23). This time we require explicitly that $f\in L^{2}(\Omega)$. Recall (from Section 1.2.2) that this problem can be reduced to $\int_{\Omega}\nabla v\cdot\nabla p=\int_{\Omega}f\ v,\text{ for all $v\in V$},$ where Dirichlet boundary condition ($p=0\ in\ \Gamma$) is essential. Recall that we can take $V=H_{0}^{1}(\Omega)$ as seen in [2, 3]. Let $u=\nabla p$ in $\Omega$. Then, problem (23) can be written as the following system of fist order partial differential equations, $\begin{split}u&=\nabla p\text{ in }\Omega\\\ \mbox{div}(u)&=-f\text{ in }\Omega\\\ p&=0\text{ in }\Gamma,\end{split}$ (30) because $\Delta p=\mbox{div}(\nabla p)$. Now, following a similar procedure as in Section 1.2.2, multiply the first equation of (30) by some $\mathbf{v}\in H(\mbox{div};\Omega)$ and integrate both sides to obtain, $\int_{\Omega}u\ \mathbf{v}=\int_{\Omega}\nabla p\cdot\mathbf{v}.$ (31) Consider Green’s identity [3], $\int_{\Omega}\mathbf{v}\cdot\nabla p+\int_{\Omega}p\ \mbox{div}(\mathbf{v})=\int_{\Gamma}(\mathbf{v}\cdot\eta)p,$ (32) where $\eta$ is the normal vector exterior to $\Gamma$. Replacing (32) in (31), and considering the third equation of (30), we get, $\int_{\Omega}u\ \mathbf{v}+\int_{\Omega}p\ \mbox{div}(\mathbf{v})=\int_{\Gamma}(\mathbf{v}\cdot\eta)p,$ (33) On the other hand, we can multiply the second equation of problem (30) by some $w\in L^{2}(\Omega)$, integrate and obtain, $\int_{\Omega}w\ \mbox{div}(u)=-\int_{\Omega}f\ w.$ (34) Note that the boundary integral depends directly on the value of $p$ in $\Gamma$. And, this is referred to as the case of a natural boundary condition. Observe that the same boundary condition appeared as an essential boundary condition in the second order formulation considered before (Section 1.2.2). In this first order formulation it showed up as a natural boundary condition. Finally, applying boundary condition $p=0\ \text{in }\Gamma$ into (33), and joining (33) and (34). We get the following problem deduced from (30), $\begin{split}&\int_{\Omega}u\ \mathbf{v}+\int_{\Omega}p\ \mbox{div}(\mathbf{v})=0\\\ &\int_{\Omega}w\ \mbox{div}(u)=-\int_{\Omega}f\ w.\end{split}$ (35) For this problem, which is a variational formulation of (30), the objective is to find $(u,p)\in H(\mbox{div};\Omega)\times L^{2}(\Omega)$ such that it is satisfied for all $\mathbf{v}\in H(\mbox{div};\Omega)$ and all $w\in L^{2}(\Omega)$. For the discretized problem related to (35), in [3] the following spaces are defined for a triangulation $T_{h}$ of the domain $\Omega$ and a fixed integer $k\geq 0$, $\begin{split}&H_{h}^{k}:=\\{\mathbf{v_{h}}\in H(\mbox{div};\Omega)\ |\ \mathbf{v_{h}}|_{K}\in RT_{k}(K)\text{ for all }K\in T_{h}\\},\text{ and}\\\ &L_{h}^{k}:=\\{w_{h}\in L^{2}(\Omega)\ |\ w_{h}|_{K}\in\mathbb{P}_{k}(K)\text{ for all }K\in T_{h}\\},\end{split}$ where $\begin{split}&\mathbb{P}_{k}(K)=\\{p:K\rightarrow\mathbb{R}\ |\ p\text{ is a polynomial of degree }\leq k\\},\text{ and}\\\ &RT_{k}(K)=[\mathbb{P}_{k}(K)\times\mathbb{P}_{k}(K)]+\mathbb{P}_{k}(K)x.\end{split}$ Note that $\mathbf{p}\in RT_{k}(K)$ if and only if there are some $p_{0},p_{1},p_{2}\in\mathbb{P}_{k}(K)$ such that $\mathbf{p}(x)=\begin{pmatrix}p_{1}(x)\\\ p_{2}(x)\end{pmatrix}+p_{0}(x)\begin{pmatrix}x\\\ y\end{pmatrix}\text{ for all }\begin{pmatrix}x\\\ y\end{pmatrix}\in K,$ and, also note that $\mathbf{p}$ has a degree of $k+1$. Then, (35) gives the following discrete problem: find $(u_{h},p_{h})\in H_{h}^{k}\times L_{h}^{k}$ such that $\begin{split}&\int_{\Omega}u_{h}\ \mathbf{v}_{h}+\int_{\Omega}p_{h}\ \mbox{div}(\mathbf{v}_{h})=0\\\ &\int_{\Omega}w_{h}\ \mbox{div}(u_{h})=-\int_{\Omega}f\ w_{h},\end{split}$ (36) for all $\mathbf{v}_{h}\in H_{h}^{k}$ and all $w_{h}\in L_{h}^{k}$. As spaces $H_{h}^{k}$ and $L_{h}^{k}$ are finite dimensional, they have a finite basis. That is, $H_{h}^{k}=\mbox{span}\\{\varphi_{i}:i=1,\dots,M\\}$ and $L_{h}^{k}=\mbox{span}\\{\psi_{j}:j=1,\dots,N\\}$. Then, $u_{h}=\sum_{i=1}^{M}u_{i}\varphi_{i}$ and $p_{h}=\sum_{j=1}^{N}p_{j}\psi_{j}$, where $u_{i}$ and $p_{j}$ are scalars. In particular, as $\varphi_{k}\in H_{h}^{k}$ and $\psi_{l}\in L_{h}^{k}$, we have that problem (36) can be written as, $\begin{split}&\int_{\Omega}\left(\sum_{i=1}^{M}u_{i}\varphi_{i}\right)\varphi_{k}+\int_{\Omega}\left(\sum_{j=1}^{N}p_{j}\psi_{j}\right)\mbox{div}(\varphi_{k})=0\\\ &\int_{\Omega}\psi_{l}\mbox{div}\left(\sum_{i=1}^{M}u_{i}\varphi_{i}\right)=\int_{\Omega}f\psi_{l},\end{split}$ (37) for $k=1,\dots,M$ and $l=1,\dots,N$. Which is equivalent to the following, by rearranging scalars, $\begin{split}&\sum_{i=1}^{M}u_{i}\int_{\Omega}\varphi_{i}\cdot\varphi_{k}+\sum_{j=1}^{N}p_{j}\int_{\Omega}\psi_{j}\mbox{div}(\varphi_{k})=0\\\ &\sum_{i=1}^{M}u_{i}\int_{\Omega}\psi_{l}\mbox{div}(\varphi_{i})=\int_{\Omega}f\psi_{l},\end{split}$ (38) for $k=1,\dots,M$ and $l=1,\dots,N$. The problem (38) can be formulated into the following matrix system $\begin{pmatrix}A&B\\\ B^{t}&0\end{pmatrix}\begin{pmatrix}U\\\ P\end{pmatrix}=\begin{pmatrix}0\\\ F\end{pmatrix},$ (39) where $A$ is a $N\times N$ matrix, $B$ is a $M\times N$ matrix with $B^{t}$ it’s transpose, $U$ is a $M$-dimensional column vector and $P,F$ are $N$-dimensional column vectors. The entries of these arrays are $A[i,j]=\int_{\Omega}\varphi_{i}\cdot\varphi_{j}$, $B[i,j]=\int_{\Omega}\psi_{j}\mbox{div}(\varphi_{i})$, $U[i]=u_{i}$, $P[i]=p_{i}$ and $F[i]=\int_{\Omega}f\psi_{i}$. The linear system (39) can be solved for $(U,P)$ with a computer using MFEM library. Note that with the entries of $U$ and $P$, the solution $(u_{h},p_{h})$ of (36) can be computed by their basis representation. The spaces defined to discretize the problem are called Raviart-Thomas finite element spaces. The fixed integer $k$ is also called the order of the shape functions or the order of the finite element space. The parameter $h$ is the same as in Section 1.2.2, which is a measure of size for $T_{h}$. See [3] for details. #### 1.2.5 Taylor-Hood finite elements In this section, we show the spatial discretization done in [8], for Stokes equations (40), $\begin{split}\frac{\partial u}{\partial t}-\nu\Delta u+\nabla p=f,\text{ in }\Omega,\\\ \nabla\cdot u=0,\text{ in }\Omega,\\\ u=g,\text{ in }\Gamma,\end{split}$ (40) and then mention the corresponding spatial discretization for Navier-Stokes equations (41), as an extension of the previous one, $\begin{split}\frac{\partial u}{\partial t}+(u\cdot\nabla)u-\nu\Delta u+\nabla p=f,\text{ in }\Omega,\\\ \nabla\cdot u=0,\text{ in }\Omega,\\\ u=g,\text{ in }\Gamma.\end{split}$ (41) When applying a numerical method to solve both systems of equations, (40) and (41), time has to be discretized too. However, time discretization is out of the scope of this paper (it can be found on section 4 of [8]). First of all, let $T_{h}=\\{K_{1},\dots,K_{m}\\}$ be a discretization of the domain $\Omega$. That is, $\Omega=K_{1}\cup\dots\cup K_{m}$, and $K_{1},\dots,K_{m}$ don’t overlap between them, and no vertex of one of them lies on the edge of another (check triangulation on Section 1.2.2). If $\Omega$ is in 2D, $K_{i}$ is a quadrilateral, and if $\Omega$ is in 3D, $K_{i}$ is a a hexahedron. Then, define the following finite element function spaces on $T_{h}$, where $d\in\\{2,3\\}$ is the dimension of $\Omega$ [8]. $\begin{split}&U_{h}^{k}=\\{v\in(H^{1}(\Omega))^{d}\ :\ v(K)\in(\mathcal{Q}_{k}(K))^{d}\text{ for all }K\in T_{h}\\},\\\ &P_{h}^{k}=\\{s\in H^{1}(\Omega)\ :\ s(K)\in\mathcal{Q}_{k}(K)\text{ for all }K\in T_{h}\\},\end{split}$ where $\mathcal{Q}_{k}(K)$ is the set of all polynomials whose degree on each of their variables is less or equal than $k$, with domain $K$. For example, if $K$ is two dimensional, $\mathcal{Q}_{2}(K)=\\{a_{0}+a_{1}x+a_{2}y+a_{3}x^{2}+a_{4}y^{2}+a_{5}xy+a_{6}x^{2}y+a_{7}xy^{2}+a_{8}x^{2}y^{2}:a_{i}\in\mathbb{R},\ i=1,\dots,8\\}$. Notice that the total degree of $a_{8}x^{2}y^{2}$ is $4$, but the degree on each variable ($x$ or $y$) is just $2$, as desired. As done on previous sections, multiply the equations of (40) by some test function $v\in U_{h}^{k}$ and $s\in P_{h}^{k}$, respectively, to obtain $\begin{split}&\int_{\Omega}\frac{\partial u}{\partial t}v-\nu\int_{\Omega}\Delta u\ v+\int_{\Omega}\nabla p\cdot v=\int_{\Omega}fv,\\\ &\int_{\Omega}\left(\nabla\cdot u\right)s=0.\end{split}$ (42) Now, applying Green’s formula (25) with homogeneous boundary condition ($u=g=0$ in $\Gamma$), as done on Section 1.2.2, the term $-\nu\int_{\Omega}\Delta u\ v$ becomes $\nu\int_{\Omega}\nabla u\cdot\nabla v$. Also, it is usual to multiply the second equation by $-1$, therefore, (42) becomes the finite element formulation (43) found on [8]. The idea is to find $(u,p)\in(U_{h}^{k},P_{h}^{k})$, for all $(v,s)\in(U_{h}^{k},P_{h}^{k})$, such that $\begin{split}&\int_{\Omega}\frac{\partial u}{\partial t}v-\nu\int_{\Omega}\nabla u\cdot\nabla v+\int_{\Omega}\nabla p\cdot v=\int_{\Omega}fv,\\\ &-\int_{\Omega}\left(\nabla\cdot u\right)s=0.\end{split}$ (43) Let $\\{\phi_{i}:i=1,\dots,n\\}$ be a basis for $U_{h}^{k}$ and $\\{\psi_{j}:j=1,\dots,m\\}$ be a basis for $P_{h}^{k}$. Therefore, $u(x,t)=\sum_{i=1}^{n}u_{i}(t)\phi_{i}(x)$ and $p(x,t)=\sum_{j=1}^{m}p_{j}(t)\psi_{j}(x)$ where $u_{i}$ and $p_{j}$ are functions depending only on $t$. Replacing these representations on (43) and noticing that $(\phi_{i},\psi_{j})\in(U_{h}^{k},P_{h}^{k})$, we get the system of equations (44), with $I=1,\dots,n$ and $J=1,\dots,m$. $\begin{split}&\int_{\Omega}\frac{\partial\left(\sum_{i=1}^{n}u_{i}\phi_{i}\right)}{\partial t}\phi_{I}-\nu\int_{\Omega}\nabla\left(\sum_{i=1}^{n}u_{i}\phi_{i}\right)\cdot\nabla\phi_{I}\\\ +&\int_{\Omega}\nabla\left(\sum_{j=1}^{m}p_{j}\psi_{j}\right)\cdot\phi_{I}=\int_{\Omega}f\phi_{I},\\\ &-\int_{\Omega}\left(\nabla\cdot\left(\sum_{i=1}^{n}u_{i}\phi_{i}\right)\right)\psi_{J}=0.\end{split}$ (44) After rearranging scalars, using properties of the dot product and the operator $\nabla$, and noting that $\phi_{i}$ does not depend on $t$, (44) can be formulated as (45). $\begin{split}&\int_{\Omega}\left(\sum_{i=1}^{n}\phi_{i}\frac{\partial u_{i}}{\partial t}\right)\phi_{I}-\nu\int_{\Omega}\left(\sum_{i=1}^{n}u_{i}\nabla\phi_{i}\right)\cdot\nabla\phi_{I}\\\ +&\int_{\Omega}\left(\sum_{j=1}^{m}p_{j}\nabla\psi_{j}\right)\cdot\phi_{I}=\int_{\Omega}f\phi_{I},\ \ I=1,\dots,n\\\ &-\int_{\Omega}\left(\sum_{i=1}^{n}u_{i}\nabla\cdot\phi_{i}\right)\psi_{J}=0,\ J=1,\dots,m.\end{split}$ (45) Finally, (45) can be formulated as (46) after swapping integrals with summations and rearranging integration scalars. $\begin{split}&\sum_{i=1}^{n}\frac{\partial u_{i}}{\partial t}\int_{\Omega}\phi_{i}\phi_{I}-\sum_{i=1}^{n}u_{i}\int_{\Omega}\nu\nabla\phi_{i}\cdot\nabla\phi_{I}\\\ +&\sum_{j=1}^{m}p_{j}\int_{\Omega}\phi_{I}\cdot\nabla\psi_{j}=\int_{\Omega}f\phi_{I},\ \ I=1,\dots,n\\\ &-\sum_{i=1}^{n}u_{i}\int_{\Omega}\psi_{J}\nabla\cdot\phi_{i}=0,\ J=1,\dots,m.\end{split}$ (46) As before, the problem can be reduced to the matrix system (47), which is the semi-discrete Stokes problem [8]. $\begin{split}M\dot{u}+Lu+Gp=f,\\\ -Du=0,\end{split}$ (47) where $M$ and $L$ are $n\times n$ matrices, $G$ is a $n\times m$ matrix, $D$ is a $m\times n$ matrix, $u$ and $f$ are $n$-dimensional vectors, $p$ is a $m$-dimensional vector and $\dot{u}$ is the notation used for the partial derivate of $u$ with respect to time $t$. The entries of these arrays are $M[i,j]=\int_{\Omega}\phi_{i}\phi_{j}$, $L[i,j]=\int_{\Omega}\nu\nabla\phi_{i}\cdot\nabla\phi_{j}$, $G[i,j]=\int_{\Omega}\phi_{i}\cdot\nabla\psi_{j}$, $D[i,j]=\int_{\Omega}\psi_{i}\nabla\cdot\phi_{j}$, $f[i]=\int_{\Omega}f\phi_{i}$, $p[i]=p_{i}$, $u[i]=u_{i}$ and $\dot{u}[i]=\frac{\partial u_{i}}{\partial t}$. As mentioned on [8], for the steady Stokes problem, $\dot{u}=0$ is taken. In such case, (47) becomes the linear matrix system (48). $\begin{pmatrix}L&G\\\ -D&0\end{pmatrix}\begin{pmatrix}u\\\ p\end{pmatrix}=\begin{pmatrix}f\\\ 0\end{pmatrix}$ (48) Furthermore, for the Navier-Stokes equations (41), the semi-discrete formulation is [8]: $\begin{split}M\dot{u}+Lu+\mathcal{N}(u)+Gp=f,\\\ -Du=0,\end{split}$ (49) where $\mathcal{N}(u)[i]=\bigintss_{\Omega}\begin{pmatrix}u_{1}&\dots&u_{n}\end{pmatrix}\begin{pmatrix}(\phi_{1}\cdot\nabla)\phi_{1}&\dots&(\phi_{1}\cdot\nabla)\phi_{n}\\\ \vdots&\ddots&\vdots\\\ (\phi_{n}\cdot\nabla)\phi_{1}&\dots&(\phi_{n}\cdot\nabla)\phi_{n}\\\ \end{pmatrix}\begin{pmatrix}u_{1}\\\ \vdots\\\ u_{n}\end{pmatrix}\phi_{i}$ is the discretized nonlinear vector-convection term, of size $n\times n$. Finally, recall that the spaces $U_{h}^{k}$ and $P_{h}^{k}$ have a given order $k$. As mentioned on [8], Taylor-Hood finite element space is the tuple $(U_{h}^{k},P_{k-1})$, which is used to solve steady and unsteady Stokes problem (convergence is optimal and stable for $k\geq 2$). And, for Navier- Stokes equations, the finite element space used is $(U_{h}^{k},P_{h}^{k})$, called $P_{N}P_{N}$ space. According to MFEM documentation [5], the implementation for the solution of Navier-Stokes equations is done following [8], which is the theory presented on this section. ### 1.3 MFEM Library In this manuscript, we worked with MFEM’s Example#1 and Example#5 which can be found on [5]. Example#1 uses standard Lagrange finite elements and Example#5 uses Raviart-Thomas mixed finite elements. Further, in Section 2.1, we find the parameters so that both problems are equivalent and then (Section 2.4), we compare the solutions. We finally mention that for a fair comparison between Lagrange and mixed finite element’s approximation, Lagrange shape functions of order $k-1$ will be compared to the corresponding (mixed) approximation obtained by using $RT_{k}(K)$. Afterwards, we worked with MFEM’s miniapp for solving Navier-Stokes equations, which corresponds to the experiments done on Section 3. On Section 3.1 we worked in a 2-dimensional domain, and on Section 3.2 we worked in a 3-dimensional domain. #### 1.3.1 Information about the library According to it’s official site [5], MFEM is a free, lightweight, scalable C++ library for finite element methods that can work with arbitrary high-order finite element meshes and spaces. MFEM has a serial and a parallel version. The serial version is the one recommended for beginners, and is used in Section 2. On the other hand, the parallel version provides more computational power and enables the use of some MFEM mini-apps, like the Navier-Stokes mini app, which is used in Section 3. Moreover, the Modular Finite Element Method (MFEM) library is developed by the MFEM Team at the Center for Applied Scientific Computing (CASC), located in the Lawrence Livermore National Laboratory (LLNL), under the BSD licence. However, as it is open source, the public repository can be found at github.com/mfem in order for anyone to contribute. Also, since 2018, a wrapper for Python (PyMFEM) is being developed in order to use MFEM library among with Python code, which demonstrates the wide applicability that the library can achieve. And, in 2021, the first community workshop was hosted by the MFEM Team, which encourages the use of the library and enlarges the community of MFEM users. Finally, take into account that the use of the library requires a good manage of C++ code, which is a programming language that’s harder to use compared to other languages, such as Python. This understanding of C++ code is important because some parts of the library are not well documented yet, and, by checking the source code, the user may find a way of implementing what is required. #### 1.3.2 Overview The main classes (with a brief and superficial explanation of them) that we are going to use in the code are: * • Mesh: domain with the partition. * • FiniteElementSpace: space of functions defined on the finite element mesh. * • GridFunction: mesh with values (solutions). * • $\\_$Coefficient: values of GridFunctions or constants. * • LinearForm: maps an input function to a vector for the rhs. * • BilinearForm: used to create a global sparse finite element matrix for the lhs. * • $\\_$Vector: vector. * • $\\_$Solver: algorithm for solution calculation. * • $\\_$Integrator: evaluates the bilinear form on element’s level. * • NavierSolver: class associated to the navier mini-app, which is used to solve Navier-Stokes equations. The ones that have $\\_$ are various classes whose name ends up the same and work similarly. Note: lhs: left hand side of the linear system. rhs: right hand side of the linear system. #### 1.3.3 Code structure A MFEM general code has the following steps (directly related classes with the step are written): 1. 1. Receive an input file (.msh) with the mesh and establish the order for the finite element spaces. 2. 2. Create a mesh object, get the dimension, and refine the mesh (refinement is optional). Mesh 3. 3. Define the finite element spaces required. FiniteElementSpace 4. 4. Define the coefficients, functions, and boundary conditions of the problem. XCoefficient 5. 5. Define the LinearForm for the rhs and assemble it. LinearForm, XIntegrator 6. 6. Define the BilinearForm for the lhs and assemble it. BilinearForm, XIntegrator 7. 7. Solve the linear system. XSolver, XVector 8. 8. Recover solution. GridFunction 9. 9. Show solution with a finite element visualization tool like GLVis [6] (optional). And, for the general code structure of the navier mini-app, we have the following steps: 1. 1. Receive an input file (.msh) with the mesh, create a parallel mesh object and refine the mesh (refinement is optional). ParMesh 2. 2. Create the flow solver by stating the order of the finite element spaces and the parameter for kinematic viscosity $\nu$. NavierSolver 3. 3. Establish the initial condition, the boundary conditions and the time step $dt$. 4. 4. Iterate through steps in time with the NavierSolver object and save the solution for each iteration in a parallel GridFunction. ParGridFunction 5. 5. Show the solution with a finite element visualization tool like ParaView [10] (optional). Notice that the Mesh and GridFunction classes used in the mini-app are for the parallel version of MFEM. The reason for this, is that the navier mini-app is available for the parallel version of MFEM only. Also, the mini-app is coded in such way that the code is simple (see Appendix 5.3). ## 2 Lagrange vs. Raviart-Thomas finite elements In this section, we take examples 1 and 5 from [5], define their problem parameters in such way that they’re equivalent, create a code that implements both of them at the same time and compares both solutions ($L_{2}$ norm), run the code with different orders, and analyse the results. Some considerations to have into account for a fair comparison are that, the order for the Mixed method should be 1 less than the order for Lagrange method, because, with this, both shape functions would have the same degree. Also, we will compare pressures and velocities with respect to the order of the shape functions and the size of the mesh ($h$ parameter). Furthermore, for the problem, the exact solution is known, so, we will use it for comparison. And, the maximum order and refinement level to be tested is determined by our computational capacity (as long as solvers converge fast). ### 2.1 Problem As mentioned before, we have to find the parameters for example 1 and 5 from [5], in such way that both problems are equivalent. This step is important because example # 1 is solved using Lagrange finite elements, while example # 5 is solved using mixed finite elements. Therefore, in order to make the comparison, the problem must be the same for both methods. Example#1 [5]: Compute $p$ such that $\begin{split}-\Delta&p=1\text{ in }\Omega\\\ &p=0\text{ in }\Gamma.\end{split}$ (50) Example#5 [5]: Compute $p$ and $\mathbf{u}$ such that $\begin{split}&k\mathbf{u}+\nabla p=f\text{ in }\Omega\\\ &-\mbox{div}(\mathbf{u})=g\text{ in }\Omega\\\ &-p=p_{0}\text{ in }\Gamma.\end{split}$ (51) From the first equation of (51), $\mathbf{u}=\frac{f-\nabla p}{k}.$ (52) Then, replacing (52) on the second equation of (51), $-\mbox{div}\left(\frac{f-\nabla p}{k}\right)=g.$ (53) If we set $k=1;\ f=0\ and\ g=-1$ in (53), we get $-\Delta p=1,$ (54) which is the first equation of (50). So, setting ($*$) $p_{0}=0,\ k=1;\ f=0\ and\ g=-1$ in (51), we get, $\begin{split}&\mathbf{u}+\nabla p=0\text{ in }\Omega\\\ &-\mbox{div}(\mathbf{u})=-1\text{ in }\Omega\\\ &-p=0\text{ in }\Gamma.\end{split}$ (55) Notice that from the first equation we get that $\mathbf{u}=-\nabla p$. This is important because in problem (50) we don’t get the solution for $\mathbf{u}$ from the method, so, we will have to find it from $p$’s derivatives. In the code, we will set the value of the parameters in the way shown here, so that both problems are the same. As seen in (52)-(54), problem (51) is equivalent to problem (50) with the values assigned for coefficients and functions at ($*$). ### 2.2 Code The first part of the code follows the structure mentioned in Section 1.3.3, but implemented for two methods at the same time (and with some extra lines for comparison purposes). Also, when defining boundary conditions, the essential one is established different from the natural one. And, after getting all the solutions, there’s a second part of the code where solutions are compared between them and with the exact one. Note: The complete code with explanations can be found on the Appendix A. However, before taking a look into it, the reader may have the following into account. The following table shows the convention used for important variable names along the code: Variable Name | Object ---|--- X_space | Finite element space X X_mixed | Variable assigned to a mixed method related object u | Velocity solution p | Pressure solution X_ex | Variable assigned to an exact solution object ### 2.3 Tests The tests of the two methods presented previously, were run on MFEM library on the domain shown on Figure 6. Figure 6: Illustration of the star domain used for the numerical tests. Visualization: [6]. Each run test is determined by the order of the Lagrange shape functions and the h parameter of the mesh. Remember that mixed shape functions have order equal to $\textit{order}-1$. The parameter order is changed directly from the command line, while the parameter h is changed via the number of times that the mesh is refined ($h=h(\\#refinements)$). As we refine the mesh more times, finite elements of the partition decrease their size, and so, the parameter $h$ decreases. Tests were run with: $order=1,\dots,N$ and $refinements=0,\dots,M$, where $N,M$ depend on the computation capacity. The star domain was partitioned using quads (instead of triangles), and such partition is shown on Figure 7. Figure 7: Initial mesh used for numerical tests (no refinements). Visualization: [6]. Results on Section 2.4 are presented in graphs. However, all the exact values that were computed can be found in the Appendix B. ### 2.4 Results In Figure 8 we show the computed solution when running the code with $\textit{order}=2$ and $\\#Refinements=3$. We use the visualization tool [6]. We mention that, at the scale of the plot, Lagrange and Mixed solutions look the same. Figure 8: Illustration of computed pressure and velocities. GLVis ([6]) is used for this visualization. Pressure (left) and $L^{2}$ norm of the vector velocity (right). Visualization: [6]. In the following results, if $u=(u_{x},u_{y})$ is the solution obtained by the mixed or Lagrange finite element method and $u_{ex}=(u_{x_{ex}},u_{y_{ex}})$ is the exact solution for the problem, then, $U_{error}=\frac{\sqrt{\left(||u_{x}-u_{x_{ex}}||_{L^{2}}\right)^{2}+\left(||u_{y}-u_{y_{ex}}||_{L^{2}}\right)^{2}}}{||u_{ex}||_{L^{2}}}.$ Figure 9: Variation of error with respect to the refinement level for the approximation of the solution of problem (50) with Lagrangian finite elements of order 1 and problem (51) with mixed Raviart-Thomas finite elements of order 0. Figure 10: Variation of error with respect to the refinement level for the approximation of the solution of problem (50) with Lagrangian finite elements of order 2 and problem (51) with mixed Raviart-Thomas finite elements of order 1. Figure 11: Variation of error with respect to the refinement level for the approximation of the solution of problem (50) with Lagrangian finite elements of order 3 and problem (51) with mixed Raviart-Thomas finite elements of order 2. Figure 12: Variation of error with respect to the refinement level for the approximation of the solution of problem (50) with Lagrangian finite elements of order 4 and problem (51) with mixed Raviart-Thomas finite elements of order 3. ### 2.5 Analysis In this section we comment and analyze the results in tables presented on the Appendix B. To understand the information presented, take into account that the exact solution would have value $0$ in X err. Also, if the two solutions obtained (Lagrange and Mixed) are exactly the same, the value in P comp and U comp would be $0$. And, lower values of $h$ mean more mesh refinements, ie, smaller partition elements. As expected, computational time increases as order and refinements increase. Here are the most relevant observations that can be obtained after analysing the data corresponding to absolute errors. For fixed order, absolute errors have little variation when reducing $h$ (max variation is $4.722$e$-03$ in $Uerr$ order 1); $Perr$ increases as $h$ decreases, while $Pmx\ err$ decreases as $h$ decreases; and, $Uerr$ increases as $h$ decreases, while $Umx\ err$ decreases as $h$ decreases. Absolute errors variation (respect to refinement) is lower when order is higher. For example; in order 2, $Perr$ is the same for each $h$ (up to three decimal places); while in order 6, $Perr$ is the same for each $h$ (up to five decimal places). For fixed $h$, absolute errors remain almost constant between orders. Moreover, $Perr$ (absolute error obtained for pressure with Lagrange) is always lower than $Pmx\ err$ (absolute error obtained for pressure with mixed) and $Uerr$ (absolute error obtained for velocity with Lagrange) is always lower than $Umx\ err$ (absolute error obtained for velocity with mixed). As order increases, pressure and velocity absolute errors tend to be the same. In order 10, the difference between $Perr$ and $Pmx\ err$ is $0.000001$ and the difference between $Uerr$ and $Umx\ err$ is $<0.0000009$. However, notice that in all the cases, the absolute error was higher than $1$. In $L_{2}$ norm, this value is pretty little, and shows that we are only getting approximations of the exact solution. Now, the most relevant observations that can be obtained after analysing the data corresponding to comparison errors. First of all, comparison error tends to $0$; and comparison errors $Ucomp$ and $Pcomp$ decrease as $h$ decreases. For a fixed order, comparison error can be similar to a higher order comparison error, as long as enough refinements are made. Moreover, when order increases, comparison errors are lower for fixed $h$. Pressure comparison error lowers faster than velocity comparison error. Maximum comparison errors were found at order 1 with no refinements, where $Pcomp\approx 7.5$e$-02$ and $Ucomp\approx 3.7$e$-02$, and minimum comparison errors were found at order 10 with 1 refinement (higher refinement level computed for order 10), where $Pcomp\approx 5.1$e$-06$ and $Ucomp\approx 9.8$e$-04$. It can be seen that $Pcomp$ improved in almost four decimal places while $Ucomp$ improved in just 2. ### 2.6 Some other examples In this section we show three of the examples that MFEM library provides [5]. We only show the problem, a brief verification of the exact solution and the solution obtained using MFEM, without going into details of any type. The purpose is to show the wide variety of applications that MFEM library can have and let the reader familiarize with the visualization of some finite element method solutions. Example 1 This example is Example # 3 of [5] and consists on solving the second order definite Maxwell equation $\nabla\times\nabla\times E+E=f$ with Dirichlet boundary condition. In the example, the value for $f$ is given by $f\begin{pmatrix}x\\\ y\\\ z\end{pmatrix}=\begin{pmatrix}\left(1+\pi^{2}\right)\sin(\pi y)\\\ \left(1+\pi^{2}\right)\sin(\pi z)\\\ \left(1+\pi^{2}\right)\sin(\pi x)\end{pmatrix}.$ The exact solution for $E$ is $E\begin{pmatrix}x\\\ y\\\ z\end{pmatrix}=\begin{pmatrix}\sin(\pi y)\\\ \sin(\pi z)\\\ \sin(\pi x)\end{pmatrix},$ and can be verified by computing $\begin{split}&\nabla\times\nabla\times E+E\\\ =&\begin{pmatrix}\frac{\partial}{\partial x}\\\ \frac{\partial}{\partial y}\\\ \frac{\partial}{\partial z}\end{pmatrix}\times\begin{pmatrix}\frac{\partial}{\partial x}\\\ \frac{\partial}{\partial y}\\\ \frac{\partial}{\partial z}\end{pmatrix}\times\begin{pmatrix}\sin(\pi x)\\\ \sin(\pi z)\\\ \sin(\pi y)\end{pmatrix}+\begin{pmatrix}\sin(\pi y)\\\ \sin(\pi z)\\\ \sin(\pi x)\end{pmatrix}\\\ =&\begin{pmatrix}\frac{\partial}{\partial x}\\\ \frac{\partial}{\partial y}\\\ \frac{\partial}{\partial z}\end{pmatrix}\times\begin{pmatrix}-\pi\cos(\pi z)\\\ -\pi\cos(\pi x)\\\ -\pi\cos(\pi y)\end{pmatrix}+\begin{pmatrix}\sin(\pi y)\\\ \sin(\pi z)\\\ \sin(\pi x)\end{pmatrix}\\\ =&\begin{pmatrix}\pi^{2}\sin(\pi y)\\\ \pi^{2}\sin(\pi z)\\\ \pi^{2}\sin(\pi x)\end{pmatrix}+\begin{pmatrix}\sin(\pi y)\\\ \sin(\pi z)\\\ \sin(\pi x)\end{pmatrix}=f.\end{split}$ The solution for $E$, $E_{h}$, computed with MFEM library is presented on Figure 13. The error for the approximation is $\left|\left|E_{h}-E\right|\right|_{L^{2}}=0.39154.$ Figure 13: Visualization of the norm of the solution for the electromagnetic diffusion problem corresponding to the second order definite Maxwell equation. Solution for $E$, $E_{h}$, obtained using MFEM library. Visualization: [6]. Example 2 This example is Example # 4 of [5] and consists on solving the diffusion problem corresponding to the second order definite equation $-\nabla(\alpha\mbox{div}(F))+\beta F=f$ with Dirichlet boundary condition, and, with parameters $\alpha=1$ and $\beta=3$. In the example, the value for $f$ is given by $f\begin{pmatrix}x\\\ y\end{pmatrix}=\begin{pmatrix}(3+2\pi^{2})\cos(\pi x)\sin(\pi y)\\\ (3+2\pi^{2})\cos(\pi y)\sin(\pi x)\end{pmatrix}.$ The exact solution for $F$ is $F\begin{pmatrix}x\\\ y\end{pmatrix}=\begin{pmatrix}\cos(\pi x)\sin(\pi y)\\\ \cos(\pi y)\sin(\pi x)\end{pmatrix},$ and can be verified by computing $\begin{split}&-\nabla(\alpha\mbox{div}(F))+\beta F\\\ =&-\nabla\left(\mbox{div}\begin{pmatrix}\cos(\pi x)\sin(\pi y)\\\ \cos(\pi y)\sin(\pi x)\end{pmatrix}\right)+3\begin{pmatrix}\cos(\pi x)\sin(\pi y)\\\ \cos(\pi y)\sin(\pi x)\end{pmatrix}\\\ =&-\nabla\left(-2\pi\sin(\pi x)\sin(\pi y)\right)+3\begin{pmatrix}\cos(\pi x)\sin(\pi y)\\\ \cos(\pi y)\sin(\pi x)\end{pmatrix}\\\ =&-\begin{pmatrix}-2\pi^{2}\cos(\pi x)\sin(\pi y)\\\ -2\pi^{2}\sin(\pi x)\cos(\pi y)\end{pmatrix}+3\begin{pmatrix}\cos(\pi x)\sin(\pi y)\\\ \cos(\pi y)\sin(\pi x)\end{pmatrix}\\\ =&\begin{pmatrix}2\pi^{2}\cos(\pi x)\sin(\pi y)\\\ 2\pi^{2}\cos(\pi y)\sin(\pi x)\end{pmatrix}+\begin{pmatrix}3\cos(\pi x)\sin(\pi y)\\\ 3\cos(\pi y)\sin(\pi x)\end{pmatrix}=f.\end{split}$ The solution for $F$, $F_{h}$, computed with MFEM library is presented on Figure 14. The domain is a square with a circular hole in the middle. On the visualization, the triangular elements of the mesh are shown. The error for the approximation is $\left|\left|F_{h}-F\right|\right|_{L^{2}}=5.55372\times 10^{-6}.$ Figure 14: Visualization of the norm of the solution for the diffusion problem showing the mesh elements. Solution for $F$, $F_{h}$, obtained using MFEM library. Visualization: [6]. Example 3 This example is Example # 7 of [5] and consists on solving the Laplace problem with mass term corresponding to the equation $-\Delta u+u=f.$ In the example, the value for $f$ is given by $f\begin{pmatrix}x\\\ y\\\ z\end{pmatrix}=\frac{7xy}{x^{2}+y^{2}+z^{2}}.$ The exact solution for $u$ is $u\begin{pmatrix}x\\\ y\\\ z\end{pmatrix}=\frac{xy}{x^{2}+y^{2}+z^{2}},$ and can be verified by computing $\begin{split}&-\Delta u+u\\\ =&-\mbox{div}\left(\nabla\left(\frac{xy}{x^{2}+y^{2}+z^{2}}\right)\right)+\frac{xy}{x^{2}+y^{2}+z^{2}}\\\ =&-\mbox{div}\left(\frac{1}{(x^{2}+y^{2}+z^{2})^{2}}\begin{pmatrix}y(-x^{2}+y^{2}+z^{2})\\\ x(x^{2}-y^{2}+z^{2})\\\ -2xyz\end{pmatrix}\right)+\frac{xy}{x^{2}+y^{2}+z^{2}}\\\ =&-\left(-\frac{6xy}{x^{2}+y^{2}+z^{2}}\right)+\frac{xy}{x^{2}+y^{2}+z^{2}}\\\ =&\frac{7xy}{x^{2}+y^{2}+z^{2}}=f.\end{split}$ The solution for $u$, $u_{h}$, computed with MFEM library is presented on Figure 15. The error for the approximation is $\left|\left|u_{h}-u\right|\right|_{L^{2}}=0.00236119.$ Figure 15: Visualization of the solution for the Laplace problem with mass term. Solution for $u$, $u_{h}$, obtained using MFEM library. Visualization: [6]. These examples show that MFEM can solve several types of equations that can include divergence, curl, gradient and Laplacian operators. Also, the given parameter $f$ was a function of the form $\mathbb{R}^{3}\rightarrow\mathbb{R}^{3}$, $\mathbb{R}^{2}\rightarrow\mathbb{R}^{2}$ and $\mathbb{R}^{3}\rightarrow\mathbb{R}$, which shows that MFEM can work with scalar and vectorial functions in a 2D or 3D domain. Although the shown visualizations were norms of the solution, using another visualization tool, such as [10], the vectors of the solution can be seen. ## 3 Numerical Experiments with NS In this section we run some computational experiments solving Navier-Stokes equations using MFEM’s Navier Stokes mini app. One of the experiments is in a 2D domain (Section 3.1) and the other one is in a 3D domain (Section 3.2). The two dimensional experiment converges to a steady state when the initial condition is the steady state with a small perturbation. In such experiment, we compare graphically the pressure solution obtained when changing the order and the refinement level. On the other hand, in the three dimensional experiment, we revise a graphical solution for a dynamical system, where turbulence is present. ### 3.1 2D Experiment: Steady State The two dimensional domain is a $2\times 4$ rectangle whose vertex coordinates are $(-0.5,1.5)$, $(-0.5,-0.5)$, $(1,1.5)$ and $(1,-0.5)$, as shown on Figure 16. Figure 16: Two dimensional domain used in the 2D Navier-Stokes experiment. It is a $2\times 4$ rectangle. Visualization: [10]. Also, the default mesh (with no refinements) is presented on figure 17. Figure 17: Two dimensional domain mesh used in the 2D Navier-Stokes experiment. Visualization: [10]. For this experiment, $Re=40$ and the velocity boundary condition was settled to be $u_{0}(x,y)=\begin{pmatrix}1-e^{2\left(10-\sqrt{100+\pi^{2}}\right)x}\cdot\cos\left(2\pi y\right)\\\ \frac{2\left(10-\sqrt{100+\pi^{2}}\right)}{2\pi}\cdot e^{2\left(10-\sqrt{100+\pi^{2}}\right)x}\cdot\sin\left(2\pi y\right)\end{pmatrix}.$ (56) If the initial condition is picked to be $u_{0}$, then the system is already on a steady state. As we wanted the experiment to reach the steady state, then initial condition was ensured to be $u_{i}(x,y)=u_{0}(x,y)+\delta\begin{pmatrix}(x+0.5)(x-1)(y+0.5)(y-1.5)\\\ (x+0.5)(x-1)(y+0.5)(y-1.5)\end{pmatrix},$ (57) where the $\delta$ parameter in the initial condition gives the magnitude of the perturbation from the steady state. The parameter was picked to be $\delta=0.001$. Notice that the term after $\delta$ vanishes in the boundary of the domain. The expected result in pressure (computed with order 6 and 5 refinements) is presented on figure 18, which corresponds to the steady state reached by the system. Figure 18: Expected solution for the pressure in the 2D experiment, computed with order 6 and 5 refinements. It is the steady state of the system. Visualization: [10]. For illustration of how the system changed, the initial condition for the pressure is shown in figure 19. Figure 19: Initial condition for pressure in the 2D experiment, computed with order 6 and 5 refinements. Visualization: [10]. Notice that the initial condition has a higher pressure on the upper-right part of the domain and it’s not uniform along any of the axes. However, after the system reaches the steady state pressure is constant for a fixed $x$ value. Now, the experiment consists on iterating through different orders and changing the refinement level for each of the orders, in order to check differences in the graphical solution. The experiment was done using a time step of $dt=0.001$ with a total time of $T=0.05$. It was computed with the parallel version of MFEM library (with 4 cores), using the Navier Miniapp [5]. The system was solved 36 times, corresponding to $order=1,2,3,4,5,6$, and for each of them, $\\#refinements=0,1,2,3,4,5$. After checking all of the results in ParaView [10], we summarized the general behaviour of the steady state solutions, as shown on figure 20. As notation, each of the results has a corresponding $(k,r)$ value, where $k$ denotes the order and $r$ the amount of refinements made. Figure 20: Summary of the steady state solutions for the 2D experiment as order and refinement level changes. The orders and refinement levels that don’t appear are almost identical to the expected steady state. Visualization: [10]. From the summary it is clear that with low order and low refinement levels, the approximation for the solution is not a good one. It’s interesting to notice that increasing $1$ in the order can be similar to increasing $1$ in the number of refinements, as seen on the steady state of $(1,1)\sim(2,0)$, $(1,2)\sim(2,1)$ and $(3,1)\sim(4,0)$. Furthermore, with a high order, the solution converges as expected even if the mesh is not refined. Finally, note that for the lowest order or for the lowest refinement level, an increase of $5$ in the counterpart achieves a good approximation (the steady states that are not shown are the ones that already achieve a good approximation). ### 3.2 3D Experiment: Turbulence The three dimensional domain is a $0.4\times 0.4\times 2.5$ parallelepiped with a vertex on the origin $(0,0,0)$ and the other ones on the positive part of the coordinate system. It has a cylindrical hole, parallel to the $yx$-plane, whose cross section center is located at $(0.5,0.2)$ and has a radius of $0.05$. The domain is shown on Figure 21. Figure 21: Three dimensional domain used in the 3D Navier-Stokes experiments. It is a $0.4\times 0.4\times 2.5$ parallelepiped with a cylindrical hole. Visualization: [10]. For the experiment, both the initial condition and the boundary condition for velocity were settled to be the same: $u_{0}\begin{pmatrix}x\\\ y\\\ z\\\ t\end{pmatrix}=u_{i}\begin{pmatrix}x\\\ y\\\ z\\\ t\end{pmatrix}=\begin{pmatrix}u_{x}(x,y,z,t)\\\ 0\\\ 0,\end{pmatrix}$ (58) where $u_{x}(x,y,z,t)=\left\\{\begin{split}\frac{36yz}{0.41^{4}}\sin\left(\frac{\pi t}{8}\right)(0.41-y)(0.41-z),\text{ if }x\leq 10^{-8},\\\ 0,\text{ if }x>10^{-8}.\end{split}\right.$ (59) Notice that the velocity condition simulates a system where the fluid is entering through the squared face of the domain near the hole. Also, the boundary condition was forced only on the inlet and on the walls of the domain. Furthermore, the experiment was computed with $4$-th order elements, using a time step of $dt=0.001$, a total time of $T=8$ and the parameter of kinematic viscosity for the fluid being $\nu=0.001$. For the visualization of the solutions, we used ParaView’s [10] stream tracer functionality, which shows the stream lines of the system at a given moment. Recall that the stream lines are tangent to the vector field of the velocity, and they show the trajectory of particles through the field (at a given instant of time). Also, there are 5 visualizations, corresponding to times $t=0,2,4,6,8$, and have the value of pressure codified in a color scale. At $t=0$, the stream lines show a laminar flow (all lines are almost parallel and follow the same direction) that avoids the obstacle, and the pressure is high on the left because the fluid is entering though that part of the domain, as shown on figure 22. Figure 22: Stream lines for the velocity of the 3D experiment at $t=0$. Visualization: [10]. Then, at $t=2$, more fluid is coming in, therefore, the pressure on the left increases. However, the value of pressure after the obstacle starts to have some variations, which will cause the turbulence later. Figure 23: Stream lines for the velocity of the 3D experiment at $t=2$. Visualization: [10]. At $t=4$, the turbulence starts to show up. As seen on figure 24, after the obstacle, some of the stream lines have spiral forms. Figure 24: Stream lines for the velocity of the 3D experiment at $t=4$. Visualization: [10]. Then, at $t=6$, the pressure on the left finally starts to lower (with a lot of variation) and the pressure on the right starts to increase, because the fluid is already passing through the obstacle and no more fluid is coming in.. More spiral-shaped stream lines show up after the obstacle. Figure 25: Stream lines for the velocity of the 3D experiment at $t=6$. Visualization: [10]. Finally, at $t=T=8$, the pressure on the left is low and on the right is high. However, the high variation of pressure in previous time steps generated a lot of turbulence near the obstacle. The fluid presents two states, a laminar one, and a turbulent one. The turbulent state, characterized by spiral movement and swirls, is presented near the obstacle; while the laminar state, characterized by straight lines, is presented away from the obstacle. As seen on figure 26. Figure 26: Stream lines for the velocity of the 3D experiment at $t=T=8$. Visualization: [10]. ## 4 Conclusion and Perspectives The MFEM library allows us to approximate the solution of partial differential equations in a versatile way. Moreover, the library has a lot of potential because it can compute with high order elements without requiring a very powerful computer, for example, we could run experiments with elements of order 10 (which is a relatively high order). Also, the navier mini app of the library provides a simple, precise and efficient way for simulating dynamical systems that involve incompressible fluids. Furthermore, it is important to use a good visualization tool, preferably one that allows the visualization of vector fields and stream lines when working with fluid equations. Finally, recall that finite element methods enable the study of systems that depend on difficult partial differential equations. Following this work, some study can be made on some of the following topics: * • Fluid modeling via partial differential equations. * • Error theorems for finite element methods. * • Effects of the mesh in the solution. * • Solutions for the equations when the parameters are not constants, but functions. * • Discretization of time. * • Picking of the time step $dt$, in order to achieve an appropriate solution. * • Turbulence and laminar flow. ## References * [1] Felipe Cruz. Comparing Lagrange and Mixed finite element methods using MFEM library. Beyond Research work at National University of Colombia. Arxiv: https://arxiv.org/submit/3724279/view * [2] Claes Johnson. Numerical Solution of Partial Differential Equations by the Finite Element Method. ISBN10 048646900X. Dover Publications Inc. 2009. * [3] Gabriel N. Gatica. A Simple Introduction to the Mixed Finite Element Method. Theory and Applications. ISBN 978-3-319-03694-6. Springer. 2014. * [4] Juan Galvis & Henrique Versieux. Introdução à Aproximação Numérica de Equações Diferenciais Parciais Via o Método de Elementos Finitos. ISBN: 978-85-244-325-5. 28 Colóquio Brasileiro de Matemática. 2011. * [5] R. Anderson and J. Andrej and A. Barker and J. Bramwell and J.-S. Camier and J. Cerveny V. Dobrev and Y. Dudouit and A. Fisher and Tz. Kolev and W. Pazner and M. Stowell and V. Tomov and I. Akkerman and J. Dahm and D. Medina and S. Zampini, MFEM: A modular finite element methods library, Computers & Mathematics with Applications 81 (2021), 42-74. Main Page: mfem.org. DOI 10.11578/dc.20171025.1248 * [6] GLVis - OpenGL Finite Element Visualization Tool. Main Page: glvis.org. DOI 10.11578/dc.20171025.1249 * [7] Kalita, P. & Łukaszewicz, G. (2016). Navier-Stokes Equations. An Introduction with Applications. Springer. DOI 10.1007/978-3-319-27760-8. * [8] Michael Franco, Jean-Sylvain Camier, Julian Andrej, Will Pazner (2020) High-order matrix-free incompressible flow solvers with GPU acceleration and low-order refined preconditioners (https://arxiv.org/abs/1910.03032) * [9] Schneider, T., et al. (2018). Decoupling Simulation Accuracy from Mesh Quality. New York University, USA. 2018 Association for Computing Machinery. https://doi.org/10.1145/3272127.3275067. * [10] Ayachit, Utkarsh, The ParaView Guide: A Parallel Visualization Application, Kitware, 2015, ISBN 978-1930934306. ## 5 Appendices ### 5.1 Appendix A : Code for comparison Here, the code used for Section 2 (written in C++) is shown, with a brief explanations of it’s functionality. $\triangleright$Include the required libraries (including MFEM) and begin main function. ⬇ #include "mfem.hpp" #include <fstream> #include <iostream> using namespace std; using namespace mfem; int main(int argc, char *argv[]){ $\triangleright$Parse command-line options (in this project we only change "order" option) and print them. ⬇ const char *mesh_file = "../data/star.mesh"; int order = 1; bool visualization = true; OptionsParser args(argc, argv); args.AddOption(&mesh_file, "-m", "–mesh", "Mesh␣file␣to␣use."); args.AddOption(&order, "-o", "–order", "Finite␣element␣order␣(polynomial␣degree)."); args.AddOption(&visualization, "-vis", "–visualization", "-no-vis", "–no- visualization", "Enable␣or␣disable␣GLVis␣visualization."); args.Parse(); if (!args.Good()){ args.PrintUsage(cout); return 1; } args.PrintOptions(cout); $\triangleright$Create mesh object from the star.mesh file and get it’s dimension. ⬇ Mesh *mesh = new Mesh(mesh_file,1,1); int dim = mesh->Dimension(); $\triangleright$Refine the mesh a given number of times (UniformRefinement). ⬇ int ref_levels; cout << "Refinements:␣"; cin >> ref_levels; for (int l = 0; l < ref_levels; l++){ mesh->UniformRefinement(); } $\triangleright$Get size indicator for mesh size (h_max) and print it. ⬇ double mesh_size, h = 0; for (int i=0;i<mesh->GetNE();i++){ mesh_size = mesh->GetElementSize(i,2); if(mesh_size>h){ h = mesh_size; } } cout << "h:␣" << h << endl; $\triangleright$Define finite element spaces. For mixed finite element method, the order will be one less than for Lagrange finite element method. The last one is a vector $L^{2}$ space that we will use later to get mixed velocity components. ⬇ FiniteElementCollection *H1 = new H1_FECollection(order, dim); FiniteElementSpace *H1_space = new FiniteElementSpace(mesh, H1); FiniteElementCollection *hdiv_coll(new RT_FECollection(order-1, dim)); FiniteElementCollection *l2_coll(new L2_FECollection(order-1, dim)); FiniteElementSpace *R_space = new FiniteElementSpace(mesh, hdiv_coll); FiniteElementSpace *W_space = new FiniteElementSpace(mesh, l2_coll); FiniteElementSpace *V_space = new FiniteElementSpace(mesh, l2_coll, 2); $\triangleright$Define the parameters of the mixed problem. C functions are defined at the end. Boundary condition is natural. ⬇ ConstantCoefficient k(1.0); void fFun(const Vector & x, Vector & f); VectorFunctionCoefficient fcoeff(dim, fFun); double gFun(const Vector & x); FunctionCoefficient gcoeff(gFun); double f_bound(const Vector & x); FunctionCoefficient fbndcoeff(f_bound); $\triangleright$Define the parameters of the Lagrange problem. Boundary condition is essential. ⬇ ConstantCoefficient one(1.0); Array<int> ess_tdof_list; if (mesh->bdr_attributes.Size()){ Array<int> ess_bdr(mesh->bdr_attributes.Max()); ess_bdr = 1; H1_space->GetEssentialTrueDofs(ess_bdr, ess_tdof_list); } $\triangleright$Define the exact solution. C functions are defined at the end. ⬇ void u_ex(const Vector & x, Vector & u); double p_ex(const Vector & x); double u_ex_x(const Vector & x); double u_ex_y(const Vector & x); $\triangleright$Get space dimensions and create vectors for the right hand side. ⬇ Array<int> block_offsets(3); block_offsets[0] = 0; block_offsets[1] = R_space->GetVSize(); block_offsets[2] = W_space->GetVSize(); block_offsets.PartialSum(); BlockVector rhs_mixed(block_offsets); Vector rhs(H1_space->GetVSize()); $\triangleright$Define the right hand side. These are LinearForm objects associated to some finite element space and rhs vector. "f" and "g" are for the mixed method and "b" is for the Lagrange method. Also, "rhs" vectors are the variables that store the information of the right hand side. ⬇ LinearForm *fform(new LinearForm); fform->Update(R_space, rhs_mixed.GetBlock(0), 0); fform->AddDomainIntegrator(new VectorFEDomainLFIntegrator(fcoeff)); fform->AddBoundaryIntegrator(new VectorFEBoundaryFluxLFIntegrator(fbndcoeff)); fform->Assemble(); LinearForm *gform(new LinearForm); gform->Update(W_space, rhs_mixed.GetBlock(1), 0); gform->AddDomainIntegrator(new DomainLFIntegrator(gcoeff)); gform->Assemble(); LinearForm *b(new LinearForm); b->Update(H1_space, rhs, 0); b->AddDomainIntegrator(new DomainLFIntegrator(one)); b->Assemble(); $\triangleright$Create variables to store the solution. "x" is the Vector used as input in the iterative method. ⬇ BlockVector x_mixed(block_offsets); GridFunction u_mixed(R_space), p_mixed(W_space), ux_mixed(W_space), uy_mixed(W_space), ue(V_space); Vector x(H1_space->GetVSize()); GridFunction ux(W_space),uy(W_space),p(H1_space); $\triangleright$Define the left hand side for mixed method. This is the BilinearForm representing the Darcy matrix (39). VectorFEMMassIntegrator is asociated to $k*u-\nabla p$ and VectorFEDDivergenceIntegrator is asociated to $\mbox{div}(u)$. ⬇ BilinearForm *mVarf(new BilinearForm(R_space)); MixedBilinearForm *bVarf(new MixedBilinearForm(R_space, W_space)); mVarf->AddDomainIntegrator(new VectorFEMassIntegrator(k)); mVarf->Assemble(); mVarf->Finalize(); SparseMatrix &M(mVarf->SpMat()); bVarf->AddDomainIntegrator(new VectorFEDivergenceIntegrator); bVarf->Assemble(); bVarf->Finalize(); SparseMatrix & B(bVarf->SpMat()); B *= -1.; SparseMatrix *BT = Transpose(B); BlockMatrix D(block_offsets); D.SetBlock(0,0, &M); D.SetBlock(0,1, BT); D.SetBlock(1,0, &B); $\triangleright$Define the left hand side for Lagrange method. This is the BilinearForm asociated to the laplacian operator. DiffusionIntegrator is asociated to $\Delta u$. The method FormLinearSystem is only used to establish the essential boundary condition. ⬇ OperatorPtr A; Vector XX,BB; BilinearForm *a(new BilinearForm(H1_space)); a->AddDomainIntegrator(new DiffusionIntegrator(one)); a->Assemble(); a->FormLinearSystem(ess_tdof_list, p, *b, A, XX, BB); $\triangleright$Solve linear systems with MINRES (for mixed) and CG (for Lagrange). SetOperator method establishes the lhs. Mult method executes the iterative algorithm and receives as input: the rhs and the vector to store the solution. ⬇ int maxIter(10000); double rtol(1.e-6); double atol(1.e-10); MINRESSolver Msolver; Msolver.SetAbsTol(atol); Msolver.SetRelTol(rtol); Msolver.SetMaxIter(maxIter); Msolver.SetPrintLevel(0); Msolver.SetOperator(D); x_mixed = 0.0; Msolver.Mult(rhs_mixed, x_mixed); if (Msolver.GetConverged()) std::cout << "MINRES␣converged␣in␣" << Msolver.GetNumIterations() << "␣iterations␣with␣a␣residual␣norm␣of␣" << Msolver.GetFinalNorm() << ".\n"; else std::cout << "MINRES␣did␣not␣converge␣in␣" << Msolver.GetNumIterations() << "␣iterations.␣Residual␣norm␣is␣" << Msolver.GetFinalNorm() << ".\n"; CGSolver Lsolver; Lsolver.SetAbsTol(atol); Lsolver.SetRelTol(rtol); Lsolver.SetMaxIter(maxIter); Lsolver.SetPrintLevel(0); Lsolver.SetOperator(*A); x = 0.0; Lsolver.Mult(rhs,x); if (Lsolver.GetConverged()) std::cout << "CG␣converged␣in␣" << Lsolver.GetNumIterations() << "␣iterations␣with␣a␣residual␣norm␣of␣" << Lsolver.GetFinalNorm() << ".\n"; else std::cout << "CG␣did␣not␣converge␣in␣" << Lsolver.GetNumIterations() << "␣iterations.␣Residual␣norm␣is␣" << Lsolver.GetFinalNorm() << ".\n"; $\triangleright$Save the solution into GridFunctions, which are used for error computation and visualization. ⬇ u_mixed.MakeRef(R_space, x_mixed.GetBlock(0), 0); p_mixed.MakeRef(W_space, x_mixed.GetBlock(1), 0); p.MakeRef(H1_space,x,0); $\triangleright$Get missing velocities from the solutions obtained. Remember that $u=-\nabla p$. Mixed components are extracted using the auxiliary variable "ue" defined before. ⬇ p.GetDerivative(1,0,ux); p.GetDerivative(1,1,uy); ux *= -1; uy *= -1; VectorGridFunctionCoefficient uc(&u_mixed); ue.ProjectCoefficient(uc); GridFunctionCoefficient ux_mixed_coeff(&ue,1); GridFunctionCoefficient uy_mixed_coeff(&ue,2); ux_mixed.ProjectCoefficient(ux_mixed_coeff); uy_mixed.ProjectCoefficient(uy_mixed_coeff); $\triangleright$Create the asociated Coefficient objects for error computation. ⬇ GridFunction* pp = &p; GridFunctionCoefficient p_coeff(pp); GridFunction* uxp = &ux; GridFunction* uyp = &uy; GridFunctionCoefficient ux_coeff(uxp); GridFunctionCoefficient uy_coeff(uyp); FunctionCoefficient pex_coeff(p_ex); VectorFunctionCoefficient uex_coeff(dim,u_ex); FunctionCoefficient uex_x_coeff(u_ex_x); FunctionCoefficient uex_y_coeff(u_ex_y); $\triangleright$Define integration rule. ⬇ int order_quad = max(2, 2*order+1); const IntegrationRule *irs[Geometry::NumGeom]; for (int i=0; i < Geometry::NumGeom; ++i){ irs[i] = &(IntRules.Get(i, order_quad)); } $\triangleright$Compute exact solution norms. Here, the parameter $2$ in ComputeLpNorm makes reference to the $L^{2}$ norm. ⬇ double norm_p = ComputeLpNorm(2., pex_coeff, *mesh, irs); double norm_u = ComputeLpNorm(2., uex_coeff, *mesh, irs); double norm_ux = ComputeLpNorm(2., uex_x_coeff, *mesh, irs); double norm_uy = ComputeLpNorm(2., uex_y_coeff, *mesh, irs); $\triangleright$Compute and print absolute errors. ⬇ double abs_err_u_mixed = u_mixed.ComputeL2Error(uex_coeff,irs); printf("Velocity␣Mixed␣Absolute␣Error:␣%e\n", abs_err_u_mixed / norm_u); double abs_err_p_mixed = p_mixed.ComputeL2Error(pex_coeff,irs); printf("Pressure␣Mixed␣Absolute␣Error:␣%e\n", abs_err_p_mixed / norm_p); double abs_err_p = p.ComputeL2Error(pex_coeff,irs); printf("Pressure␣Absolute␣Error:␣%e\n", abs_err_p / norm_p); double abs_err_ux = ux.ComputeL2Error(uex_x_coeff,irs); double abs_err_uy = uy.ComputeL2Error(uex_y_coeff,irs); double abs_err_u = pow(pow(abs_err_ux,2)+pow(abs_err_uy,2),0.5); printf("Velocity␣Absolute␣Error:␣%e\n", abs_err_u / norm_u); $\triangleright$Compute and print comparison errors. ⬇ double err_ux = ux_mixed.ComputeL2Error(ux_coeff,irs); double err_uy = uy_mixed.ComputeL2Error(uy_coeff,irs); double err_u = pow(pow(err_ux,2)+pow(err_uy,2),0.5); printf("Velocity␣Comparison␣Error:␣%e\n", err_u / norm_u); double err_p = p_mixed.ComputeL2Error(p_coeff, irs); printf("Pressure␣Comparison␣Error:␣%e\n", err_p / norm_p); $\triangleright$Visualize the solutions and the domain. GLVis visualization tool uses port $19916$ to receive data. ⬇ char vishost[] = "localhost"; int visport = 19916; if(visualization){ Vector x_domain(H1_space->GetVSize()); GridFunction domain(H1_space); x_domain=0.0; domain.MakeRef(H1_space,x_domain,0); socketstream dom_sock(vishost, visport); dom_sock.precision(8); dom_sock << "solution\n" << *mesh << domain << "window_title␣’Domain’" << endl; socketstream um_sock(vishost, visport); um_sock.precision(8); um_sock << "solution\n" << *mesh << u_mixed << "window_title␣’Velocity␣Mixed’" << endl; socketstream pm_sock(vishost, visport); pm_sock.precision(8); pm_sock << "solution\n" << *mesh << p_mixed << "window_title␣’Pressure␣Mixed’" << endl; socketstream uxm_sock(vishost, visport); uxm_sock.precision(8); uxm_sock << "solution\n" << *mesh << ux_mixed << "window_title␣’X␣Velocity␣Mixed’" << endl; socketstream uym_sock(vishost, visport); uym_sock.precision(8); uym_sock << "solution\n" << *mesh << uy_mixed << "window_title␣’Y␣Velocity␣Mixed’" << endl; socketstream p_sock(vishost, visport); p_sock.precision(8); p_sock << "solution\n" << *mesh << p << "window_title␣’Pressure’" << endl; socketstream ux_sock(vishost, visport); ux_sock.precision(8); ux_sock << "solution\n" << *mesh << ux << "window_title␣’X␣Velocity’" << endl; socketstream uy_sock(vishost, visport); uy_sock.precision(8); uy_sock << "solution\n" << *mesh << uy << "window_title␣’Y␣Velocity’" << endl; }} $\triangleright$Define C functions. ⬇ void fFun(const Vector & x, Vector & f){ f = 0.0; } double gFun(const Vector & x){ return -1.0; } double f_bound(const Vector & x){ return 0.0; } void u_ex(const Vector & x, Vector & u){ double xi(x(0)); double yi(x(1)); double zi(0.0); u(0) = - exp(xi)*sin(yi)*cos(zi); u(1) = - exp(xi)*cos(yi)*cos(zi); } double u_ex_x(const Vector & x){ double xi(x(0)); double yi(x(1)); double zi(0.0); return -exp(xi)*sin(yi)*cos(zi); } double u_ex_y(const Vector & x){ double xi(x(0)); double yi(x(1)); double zi(0.0); return -exp(xi)*cos(yi)*cos(zi); } double p_ex(const Vector & x){ double xi(x(0)); double yi(x(1)); double zi(0.0); return exp(xi)*sin(yi)*cos(zi); } ### 5.2 Appendix B : Numerical values of the comparison The order parameter will be fixed for each table and $h$ parameter is shown in the first column. To interpret the results take into account that P refers to pressure, U refers to velocity, mx refers to mixed (from mixed finite element method), err refers to absolute error (compared to the exact solution), and comp refers to comparison (the error between the two solutions obtained by the two different methods). Order = 1 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 7.549479e-02 | 1.021287e+00 | 1.025477e+00 | 3.680827e-02 | 1.029378e+00 | 1.037635e+00 0.286032 | 3.627089e-02 | 1.022781e+00 | 1.023990e+00 | 1.727281e-02 | 1.032760e+00 | 1.035055e+00 0.143016 | 1.791509e-02 | 1.023236e+00 | 1.023596e+00 | 9.222996e-03 | 1.033725e+00 | 1.034369e+00 0.0715079 | 8.922939e-03 | 1.023372e+00 | 1.023480e+00 | 5.111295e-03 | 1.033999e+00 | 1.034182e+00 0.035754 | 4.455715e-03 | 1.023412e+00 | 1.023445e+00 | 2.859769e-03 | 1.034077e+00 | 1.034130e+00 0.017877 | 2.226845e-03 | 1.023424e+00 | 1.023435e+00 | 1.603788e-03 | 1.034100e+00 | 1.034115e+00 Order = 2 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 8.069013e-03 | 1.023329e+00 | 1.023554e+00 | 1.399079e-02 | 1.033924e+00 | 1.034255e+00 0.286032 | 2.138257e-03 | 1.023391e+00 | 1.023470e+00 | 7.845012e-03 | 1.034056e+00 | 1.034146e+00 0.143016 | 5.704347e-04 | 1.023417e+00 | 1.023442e+00 | 4.400448e-03 | 1.034093e+00 | 1.034120e+00 0.0715079 | 1.537926e-04 | 1.023426e+00 | 1.023434e+00 | 2.469526e-03 | 1.034104e+00 | 1.034112e+00 0.035754 | 4.194302e-05 | 1.023428e+00 | 1.023431e+00 | 1.385966e-03 | 1.034107e+00 | 1.034110e+00 Order = 3 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 8.691241e-04 | 1.023389e+00 | 1.023471e+00 | 8.745151e-03 | 1.034060e+00 | 1.034143e+00 0.286032 | 2.477673e-04 | 1.023417e+00 | 1.023443e+00 | 4.911967e-03 | 1.034094e+00 | 1.034120e+00 0.143016 | 7.316263e-05 | 1.023426e+00 | 1.023434e+00 | 2.756849e-03 | 1.034104e+00 | 1.034112e+00 0.0715079 | 2.178864e-05 | 1.023428e+00 | 1.023431e+00 | 1.547232e-03 | 1.034108e+00 | 1.034110e+00 Order = 4 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 3.199774e-04 | 1.023412e+00 | 1.023448e+00 | 6.119857e-03 | 1.034088e+00 | 1.034124e+00 0.286032 | 9.547574e-05 | 1.023424e+00 | 1.023435e+00 | 3.434952e-03 | 1.034103e+00 | 1.034114e+00 0.143016 | 2.862666e-05 | 1.023428e+00 | 1.023431e+00 | 1.927814e-03 | 1.034107e+00 | 1.034111e+00 Order = 5 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 1.552006e-04 | 1.023420e+00 | 1.023439e+00 | 4.578518e-03 | 1.034099e+00 | 1.034117e+00 0.286032 | 4.658038e-05 | 1.023427e+00 | 1.023433e+00 | 2.569749e-03 | 1.034106e+00 | 1.034112e+00 0.143016 | 1.406993e-05 | 1.023429e+00 | 1.023431e+00 | 1.442205e-03 | 1.034108e+00 | 1.034110e+00 Order = 6 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 8.612580e-05 | 1.023424e+00 | 1.023435e+00 | 3.584133e-03 | 1.034103e+00 | 1.034114e+00 0.286032 | 2.600417e-05 | 1.023428e+00 | 1.023431e+00 | 2.011608e-03 | 1.034107e+00 | 1.034111e+00 0.143016 | 7.897631e-06 | 1.023429e+00 | 1.023430e+00 | 1.128989e-03 | 1.034109e+00 | 1.034110e+00 Order = 7 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 5.243187e-05 | 1.023426e+00 | 1.023433e+00 | 2.899307e-03 | 1.034105e+00 | 1.034112e+00 0.286032 | 1.589631e-05 | 1.023429e+00 | 1.023431e+00 | 1.627221e-03 | 1.034108e+00 | 1.034110e+00 Order = 8 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 3.409225e-05 | 1.023427e+00 | 1.023432e+00 | 2.404311e-03 | 1.034107e+00 | 1.034111e+00 0.286032 | 1.037969e-05 | 1.023429e+00 | 1.023430e+00 | 1.349427e-03 | 1.034108e+00 | 1.034110e+00 Order = 9 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 2.328387e-05 | 1.023428e+00 | 1.023431e+00 | 2.033288e-03 | 1.034107e+00 | 1.034110e+00 0.286032 | 7.124397e-06 | 1.023429e+00 | 1.023430e+00 | 1.141177e-03 | 1.034109e+00 | 1.034110e+00 Order = 10 h | P comp | P err | Pmx err | U comp | U err | U mx err ---|---|---|---|---|---|--- 0.572063 | 1.664200e-05 | 1.023429e+00 | 1.023431e+00 | 1.746755e-03 | 1.034108e+00 | 1.034110e+00 0.286032 | 5.085321e-06 | 1.023429e+00 | 1.023430e+00 | 9.803705e-04 | 1.034109e+00 | 1.034109e+00 ### 5.3 Appendix C : MiniApp Code for Navier-Stokes Here, the code used for Section 3 (written in C++) is shown, with a brief explanations of it’s functionality. $\triangleright$Include the required libraries (including navier miniapp). ⬇ #include "navier_solver.hpp" #include <fstream> using namespace mfem; using namespace navier; using namespace std; $\triangleright$Define the context for the problem. In this case, we define parameters for the 2D and 3D experiment. In general, the parameters to define should be $\nu$, $dt$, $T$, $k$ and $\\#refinements$. ⬇ struct NavierContext{ //Parameters for 2D experiment int max_order_steady = 6; int max_refinements_steady = 5; double kinvis_steady = 1.0 / 40.0; double t_final_steady = 50 * 0.001; double dt_steady = 0.001; double delta = 0.001; //Parameters for 3D experiment double kinvis_3d = 0.001; double t_final_3d = 8.0; double dt_3d = 1e-3; } ctx; $\triangleright$Define a velocity as a C function. This function represents the initial and boundary conditions for the velocity in the 3D experiment. ⬇ void vel_3d(const Vector &x, double t, Vector &u){ double xi = x(0); double yi = x(1); double zi = x(2); double U = 2.25; if(xi <= 1e-8){ u(0) = 16.0 * U * yi * zi * sin(M_PI * t / 8.0) * (0.41 - yi) * (0.41 - zi) / pow(0.41, 4.0); }else{ u(0) = 0.0; } u(1) = 0.0; u(2) = 0.0;} $\triangleright$Define another velocity as a C function. This function represents the boundary condition, $u_{0}$, for the 2D experiment. ⬇ void vel_steady(const Vector &x, double t, Vector &u){ double reynolds = 1.0 / ctx.kinvis_steady; double lam = 0.5 * reynolds - sqrt(0.25 * reynolds * reynolds + 4.0 * M_PI * M_PI); double xi = x(0); double yi = x(1); u(0) = 1.0 - exp(lam * xi) * cos(2.0 * M_PI * yi); u(1) = lam / (2.0 * M_PI) * exp(lam * xi) * sin(2.0 * M_PI * yi);} $\triangleright$Define a third velocity as a C function. This function represents the initial condition, $u_{i}$, for the 2D experiment. ⬇ void vel(const Vector &x, double t, Vector &u){ double reynolds = 1.0 / ctx.kinvis_steady; double lam = 0.5 * reynolds - sqrt(0.25 * reynolds * reynolds + 4.0 * M_PI * M_PI); double xi = x(0); double yi = x(1); double delta = ctx.delta; u(0) = 1.0 - exp(lam * xi) * cos(2.0 * M_PI * yi); u(1) = lam / (2.0 * M_PI) * exp(lam * xi) * sin(2.0 * M_PI * yi); u(0) = u(0)+delta*(xi+0.5)*(xi-1)*(yi+0.5)*(yi-1.5); u(1) = u(1)+delta*(xi+0.5)*(xi-1)*(yi+0.5)*(yi-1.5); } $\triangleright$Begin a function that receives an order and the amount of refinements. This function will compute the 2D experiment. First, it defines the mesh shown on figure 17. ⬇ void NS_steady(int order, int refinement){ Mesh *mesh = new Mesh(2, 4, Element::QUADRILATERAL, false, 1.5, 2.0); mesh->EnsureNodes(); GridFunction *nodes = mesh->GetNodes(); *nodes -= 0.5; $\triangleright$Refine the mesh and create the parallel version of the mesh. ⬇ for (int i = 0; i < refinement; ++i){ mesh->UniformRefinement();} auto *pmesh = new ParMesh(MPI_COMM_WORLD, *mesh); delete mesh; $\triangleright$Create the NavierSolver object (receives the mesh, $k$, and $\nu$ as parameters). Set the initial condition, the function vel. And set the boundary condition, the function vel_steady. ⬇ //Create the flow solver NavierSolver flowsolver(pmesh, order, ctx.kinvis_steady); flowsolver.EnablePA(true); //Set the initial condition ParGridFunction *u_ic = flowsolver.GetCurrentVelocity(); VectorFunctionCoefficient u_excoeff(pmesh-> Dimension(), vel); u_ic->ProjectCoefficient(u_excoeff); //Add Dirichlet boundary conditions to velocity space Array<int> attr(pmesh->bdr_attributes.Max()); attr = 1; flowsolver.AddVelDirichletBC(vel_steady, attr); $\triangleright$Set up the problem ($dt$) and define the ParGridFunctions to store the solution. ⬇ double t = 0.0; bool last_step = false; flowsolver.Setup(ctx.dt_steady); ParGridFunction *u_gf = flowsolver.GetCurrentVelocity(); ParGridFunction *p_gf = flowsolver.GetCurrentPressure(); $\triangleright$Create the ParaView file and associate the variables to save the solution. ⬇ ParaViewDataCollection pvdc("STEADY"+to_string(order)+to_string(refinement),pmesh); pvdc.SetDataFormat(VTKFormat::BINARY32); pvdc.SetHighOrderOutput(true); pvdc.SetLevelsOfDetail(order); pvdc.SetCycle(0); pvdc.SetTime(t); pvdc.RegisterField("velocity",u_gf); pvdc.RegisterField("pressure",p_gf); pvdc.Save(); $\triangleright$Iterate from $t=0$ to $t=T$. For each iteration, the step is taken and saved in the ParaView file. ⬇ for(int step = 0; !last_step; ++step){ //Check for final step if (t + ctx.dt_steady >= ctx.t_final_steady - ctx.dt_steady / 2){last_step = true;} //Do the step flowsolver.Step(t, ctx.dt_steady, step); //Save paraview information pvdc.SetCycle(step); pvdc.SetTime(t); pvdc.Save();} delete pmesh;} $\triangleright$Begin a function that receives an order and the amount of refinements. This function will compute the 3D experiment. First, it defines the mesh associated to the domain shown on figure 21. This part of the code was planned to be used with different orders and refinement levels, however, it was only run with one case. ⬇ void NS_3D(int order, int refinement){ Mesh *mesh = new Mesh("box-cylinder.mesh"); $\triangleright$Refine the mesh and create the parallel version of the mesh. ⬇ for (int i = 0; i < refinement; ++i){ mesh->UniformRefinement();} auto *pmesh = new ParMesh(MPI_COMM_WORLD, *mesh); delete mesh; $\triangleright$Create the NavierSolver object. Set the initial condition, the function vel_3d. And set the boundary condition, the function vel_3d too. Notice that the boundary condition is only applied to some parts of the mesh. ⬇ //Create the flow solver NavierSolver flowsolver(pmesh, order, ctx.kinvis_3d); flowsolver.EnablePA(true); //Set the initial condition ParGridFunction *u_ic = flowsolver.GetCurrentVelocity(); VectorFunctionCoefficient u_excoeff(pmesh-> Dimension(), vel_3d); u_ic->ProjectCoefficient(u_excoeff); //Add Dirichlet boundary conditions to velocity space restricted to selected attributes on the mesh Array<int> attr(pmesh->bdr_attributes.Max()); attr[0] = 1; //Inlet attr[2] = 1; //Walls flowsolver.AddVelDirichletBC(vel_3d, attr); $\triangleright$Set up the problem ($dt$) and define the ParGridFunctions to store the solution. ⬇ double t = 0.0; bool last_step = false; flowsolver.Setup(ctx.dt_3d); ParGridFunction *u_gf = flowsolver.GetCurrentVelocity(); ParGridFunction *p_gf = flowsolver.GetCurrentPressure(); $\triangleright$Create the ParaView file and associate the variables to save the solution. ⬇ ParaViewDataCollection pvdc("3D"+to_string(order)+to_string(refinement), pmesh); pvdc.SetDataFormat(VTKFormat::BINARY32); pvdc.SetHighOrderOutput(true); pvdc.SetLevelsOfDetail(order); pvdc.SetCycle(0); pvdc.SetTime(t); pvdc.RegisterField("velocity", u_gf); pvdc.RegisterField("pressure", p_gf); pvdc.Save(); $\triangleright$Iterate from $t=0$ to $t=T$. For each iteration, the step is taken and saved in the ParaView file. ⬇ for(int step = 0; !last_step; ++step){ //Check for final step if(t + ctx.dt_3d >= ctx.t_final_3d - ctx.dt_3d / 2){last_step = true;} //Do the step flowsolver.Step(t, ctx.dt_3d, step); //Save paraview information every 10 steps if (step % 10 == 0){ pvdc.SetCycle(step); pvdc.SetTime(t); pvdc.Save();}} delete pmesh;} $\triangleright$Finally, define the main function with a MPI session (for parallel computation). The 2D experiment corresponds to the use of the function NS_steady while iterating through orders and refinement levels. And, the 3D experiment is running the function NS_3D with order $4$. No refinements were done because the original mesh (done by MFEM [5]) gives a good solution already. ⬇ int main(int argc, char *argv[]){ //Define mpi sessions (parallel programming) MPI_Session mpi(argc,argv); //Steady experiment for(int order=1;order<=ctx.max_order_steady;order++){ for(int refinements=0;refinements<=ctx.max_refinements_steady;refinements++){ NS_steady(order,refinements); } } //3D experiment NS_3D(4,0); } Note: This code was run using the command mpirun -n 4, which uses 4 cores of the computer.
# UV/vis-to-IR Photonic Down Conversion Mediated by Excited State Vibrational Polaritons Connor K. Terry<EMAIL_ADDRESS>Department of Chemistry, Northwestern University, Evanston, Illinois 60208-3113, USA. Justin<EMAIL_ADDRESS>Department of Chemistry, Northwestern University, Evanston, Illinois 60208-3113, USA. Emily A. <EMAIL_ADDRESS>Department of Chemistry, Northwestern University, Evanston, Illinois 60208-3113, USA. Roel <EMAIL_ADDRESS>Department of Chemistry, Northwestern University, Evanston, Illinois 60208-3113, USA. ###### Abstract This work proposes a new photophysical phenomenon whereby UV/vis excitation of a molecule involving a Franck-Condon (FC) active vibration yields infrared (IR) emission by strong coupling to an optical cavity. The resulting UV/vis- to-IR photonic down conversion process is mediated by vibrational polaritons in the electronic excited state potential. It is shown that the formation of such excited state vibrational polaritons (ESVP) via UV/vis excitation only occurs with molecules having vibrational modes with both a non-zero FC activity and IR activity in the excited state. Density functional theory calculations are shown to effectively identify a candidate molecule, 1-Pyreneacetic acid (PAA), with this property and the dynamics of ESVP are modeled using the truncated Wigner approximation. Overall, this work presents a new avenue of polariton chemistry where excited state dynamics, driven by photoexcitation, are influenced by the formation of vibrational polaritons. Along with this, the photonic down conversion is potentially useful in both the sensing of excited state vibrations and in quantum transduction schemes. ## 1 Introduction Most light-matter interactions involved in spectroscopy are in the weak coupling regime such that the intramolecular Coulombic fields are significantly stronger than the interacting radiation fields.[1] As such, perturbative theories accurately describe many spectroscopic phenomena.[2] These theories break down in the “strong coupling” regime where light and matter states hybridize, resulting in energy splittings that can be spectroscopically observed.[3] The quanta of such hybrid light-matter states are called “polaritons”, the formation of which is commonly realized by the use of high finesse optical cavities with quantized radiation modes that can be coupled to dipole carrying transitions within states of matter. Polaritons formed between infrared (IR)-active molecular or condensed phase vibrations and an optical cavity are, naturally, called “vibrational polaritons”. In recent experiments, it has been demonstrated that product selectivity and rates of ground state reactions may be modulated by the formation of vibrational polaritons through coupling an optical cavity to the vibrational coordinate on which a reaction proceeds.[4, 5, 6, 7, 8, 9, 10] Although there has been significant experimental and theoretical work towards understanding such ground state vibrational polaritons and their effect on reactions along with other physiochemical phenomena,[11, 3, 12, 13, 14, 15, 16, 17, 18] analogies on the electronic excited state potential have remained underexplored.[19, 20] Here, we propose a novel photonic down conversion process that results from the formation of excited state vibrational polaritons (ESVP) via UV/vis electronic excitation of a molecule containing a vibrational mode that is coupled to an IR cavity. Under vibronic coupling, the UV/vis excitation of the molecule induces a wavepacket consisting of ESVP, while Rabi oscillations occur between occupied vibrational and photonic states as time evolves. These photons in turn may leak from the cavity, leading to IR light emission.[21] The UV/vis-to-IR photonic down conversion process proposed in this work has potential applications in quantum information science, where the vibrational wave-packet, formed by a UV/vis excitation of the cavity-embedded molecule, can be directly transduced into a photonic wavepacket through coupling to the optical cavity. Hence, molecules with different vibronic couplings and excited state potential energy surfaces produce different coherences within the photonic wavepacket emitted via the ESVP photonic down conversion process. This could allow for the manipulation of coherent light by molecular design. Furthermore, the UV/vis-to-IR photonic down conversion process implies that an optical cavity can be used to enhance the emission of excited state vibrations, analogous to the Purcell effect in the weak light-matter coupling regime.[22, 23] The effective detection of excited state vibrations through this process would be a powerful measurement tool for identifying vibrational modes induced by electronic transitions—information that has been difficult to obtain experimentally,[24] and that often requires a determination through electronic structure calculations and molecular dynamics simulations.[25] Detection of photonic emission from vibrations is typically impossible due to their non-radiative relaxation being on the order of picoseconds,[26, 27] which out-competes any non-enhanced emission that occurs on a $>$ 1 µs time scale.[28, 29] It was experimentally indicated by Raschke and coworkers that ground state vibrational emission could be enhanced by coupling molecules to an optical resonator, yielding a 50% decrease in the vibrational dephasing lifetime for poly(methylmethacrylate),[29] although photonic emission from vibrations was not actually observed. Another potential application of ESVP is in the modulation of photophysical processes that occur via transitions between excited electronic states, and which can be mediated by vibrational coherences. For example, vibrational coherences have been shown to mediate singlet fission in certain molecular systems,[30, 31, 32, 33, 34, 35] promote charge and energy transfer in biological light harvesting systems,[36, 37, 38, 39] as well as drive the photochemical reaction of vision.[40, 41, 42] The hybridization of cavity photons with relevant vibrational modes could modulate such coherences as well as change excited state potential energy surfaces (PES), altering reaction pathways and rates similarly to what has been shown for ground state reactions. In this work, we utilize the truncated Wigner approximation (TWA)[43] to theoretically predict the time evolution of quantum mechanical observables in a combined cavity-molecule system, either without dissipation or when coupled to dissipative harmonic baths. The TWA is exact for systems governed by linear or quadratic potentials, i.e., systems consisting of harmonic oscillators. Furthermore, the TWA benefits from linear scaling of complexity with increasing number of degrees of freedom (DOF), as opposed to the exponential increase in complexity of the Hilbert space of the system. Therefore, the TWA enables exact modeling of the evolution of polariton wavepackets coupled to baths of harmonic oscillators without requiring a truncated Hilbert space and without tracing out the bath DOF, as is done in standard quantum master equation techniques.[44, 45] We are therefore able to concurrently and explicitly monitor the population transfer into both a radiative and non- radiative bath, giving insights into how different interaction parameters affect photonic emission from ESVP in the presence of non-radiative vibrational relaxation. ## 2 Results ### 2.1 Photonic Down Conversion Mechanism Figure 1: Schematic of the UV/vis-to-IR photonic down conversion process mediated by ESVP. An impulsive excitation by UV/vis light (blue arrow) creates ESVP due to vibronic interactions of the cavity coupled vibrational mode. The ESVP then leak IR radiation from the cavity (red arrows). Figure 1 shows a schematic of the ESVP mediated photonic down conversion process. A molecule is positioned inside an optical cavity, initially in its ground electronic state. A mode of the optical cavity is assumed to be near resonant with a spectrally isolated vibrational mode of the molecule. We treat the cavity as transparent to high optical frequencies allowing a UV/vis pulse to promote the molecule into an electronically excited state. We also assume the FC approximation where vibronic coupling is characterized by a linear shift in the excited state nuclear PES, as shown in Figure 1, with $Q_{0}^{(g)}$ and $Q_{0}^{(e)}$ being the ground and excited state equilibrium nuclear configurations of the vibrational mode of interest, respectively. Assuming a non-zero Huang-Rhys (HR) factor (vibronic coupling constant), a vibrational wavepacket will be created upon electronic excitation of the molecule. In the absence of the cavity, this wavepacket will have an expected occupation number, $\langle\hat{N}_{\textrm{vib}}\rangle$, equal to the HR factor ($S$),[46, 47] and its energy expectation value equals $S\hbar\omega_{v}$, which is called the reorganization energy, where $\omega_{v}$ is the vibrational frequency. When the cavity is present and the excited state vibrational transition dipole moment ($\bm{\mu}_{\textrm{IR}}^{(e)}$) is non-zero for the cavity resonant vibrational mode, the vibrational wavepacket will comprise a superposition of polariton eigenstates and Rabi oscillations will occur between vibrations and cavity photons as time evolves. Furthermore, given a finite quality factor ($\mathcal{Q}$) of the cavity, which is a measure of its radiative energy loss,[48] the cavity photons involved in such ESVP will leak out of the cavity resulting in IR emission that can be detected and/or harnessed.[21] Altogether, these steps constitute a photonic down conversion process where a single incident UV/vis photon is converted into one or more IR photons. In this work, we only consider dissipation of cavity photons through the cavity mirrors so that $\mathcal{Q}$ is dependent only on the amount of detectable radiative emission. In general, other dissipation pathways should also be accounted for such as absorption by the cavity mirrors and bound mode photoluminescence,[21] which would also decrease $\mathcal{Q}$. Therefore, the results presented herein give an idealized estimate of the amount of photonic down conversion that can occur within a lossy cavity. ### 2.2 Theoretical Model We use the Pauli-Fierz Hamiltonian as a basis for constructing a model of the ESVP mediated photonic down conversion process.[49, 50, 51] Accordingly, the cavity-molecule system is described as $\hat{H}_{\textrm{system}}=\hat{H}_{\textrm{mol}}+\frac{\hat{p}^{2}}{2}+\frac{\omega_{c}^{2}}{2}(\hat{q}+\bm{A}\cdot\bm{\hat{\mu}})^{2},$ (1) where $\hat{p}$ and $\hat{q}$ are the cavity mode momentum and position operators, respectively, and $\omega_{c}$ is the frequency of the cavity mode. Moreover, $\bm{A}=\sqrt{2/\hbar\omega_{c}}A_{0}\bm{\epsilon}_{c}$, where $A_{0}=\sqrt{\hbar/2\omega_{c}\varepsilon V}$ is the amplitude of the cavity mode vector potential, with $\varepsilon$ being the effective permittivity of the cavity and $V$ being the cavity mode volume, and $\bm{\epsilon}_{c}$ is the cavity mode polarization unit vector. Lastly, $\bm{\hat{\mu}}$ is the molecular dipole operator and $\hat{H}_{\textrm{mol}}$ describes the molecule. We express $\hat{H}_{\textrm{system}}$ in the molecular diabatic basis at the ground state equilibrium, that is, the basis of electronic eigenstates, $|\alpha\rangle$, that satisfy $\hat{H}_{\textrm{mol}}^{\textrm{el}}(Q_{0}^{(g)})|\alpha\rangle=E_{\alpha}(Q_{0}^{(g)})|\alpha\rangle.$ (2) Here, $\hat{H}_{\textrm{mol}}^{\textrm{el}}(Q_{0}^{(g)})$ is the component of $\hat{H}_{\textrm{mol}}$ containing a dependence on the electronic DOF, parameterized by $Q$, which is taken to be fixed at $Q=Q_{0}^{(g)}$. We truncate this basis to only include the ground and a single excited diabatic state denoted as $|g\rangle$ and $|e\rangle$, respectively. In this basis, $\hat{H}_{\textrm{mol}}$, including a single harmonic vibrational mode, can be written as $\hat{H}_{\textrm{mol}}=\frac{\hat{P}^{2}}{2}+\frac{1}{2}\omega_{v}^{2}\hat{Q}^{2}+\left(E_{\textrm{vert}}-\omega_{v}^{2}Q_{0}^{(e)}\hat{Q}\right)|e\rangle\langle e|,$ (3) where $\hat{P}$ and $\hat{Q}$ are the mass-weighted momentum and position operators of the vibrational mode, respectively, $E_{\textrm{vert}}=\hbar\Omega_{e,g}+S\hbar\omega_{v}$ is the energy of the vertical transition from the ground to excited electronic PES at the point of the ground state nuclear equilibrium configuration (the FC point) with $\Omega_{e,g}$ being the transition frequency between the ground and excited electronic states at their respective nuclear equilibrium position. Furthermore, in Eq. (3) we have taken the ground state nuclear equilibrium configuration to be at the origin so that $Q^{(g)}_{0}=0$, and it is assumed that there is no coupling between diabatic states. We have also assumed that the vibrational frequency is independent of electronic state. Eq. (1) and Eq. (3) effectively describe a system of coupled harmonic oscillators: the vibrational mode and the cavity, which are coupled through an interaction of the cavity field with the molecular dipole. Expanding the quadratic term in Eq. (1) gives $\hat{H}_{\textrm{system}}=\hat{H}_{\textrm{mol}}+\hat{H}_{\textrm{cav}}+\hat{H}_{\textrm{int}},$ (4) where $\hat{H}_{\textrm{cav}}=\frac{\hat{p}^{2}}{2}+\frac{\omega_{c}^{2}}{2}\hat{q}^{2}$ (5) and $\hat{H}_{\text{int}}=\omega_{c}^{2}\bm{A}\cdot\bm{\hat{\mu}}\,\hat{q}+\frac{\omega_{c}^{2}}{2}(\bm{A}\cdot\bm{\hat{\mu}})^{2}.$ (6) As is commonly done with Jaynes-Cummings like models,[21, 52, 13] we ignore the 2nd term of $\hat{H}_{\text{int}}$, which is the dipole self energy, reserving a survey of its effects to a follow up study (as it may become significant in ultra strong coupling[51]). Once $\hat{H}_{\textrm{system}}$ is expressed in the diabatic basis, each element of the dipole operator in this basis is expanded to the first order with respect to $\hat{Q}$ about the ground state nuclear equilibrium position ($Q_{0}^{(g)}=0$) in order to account for linear coupling between the cavity and the vibrational mode. This gives rise to two cavity-vibration interaction terms, $\hat{H}_{\textrm{int}}=\hat{H}_{\text{PFS}}+\hat{H}_{\text{cav-vib}},$ (7) where $\hat{H}_{\text{PFS}}=\omega_{c}^{2}\bm{A}\cdot\bm{\mu}^{0}_{e}\,\hat{q},$ (8) and where $\hat{H}_{\text{cav- vib}}=\omega_{c}^{2}\bm{A}\cdot\bm{\mu}_{e}^{\prime}\hat{Q}\,\hat{q}.$ (9) Here, $\bm{\mu}_{e}^{0}$ is the permanent dipole moment of the excited state at $Q_{0}^{(g)}$ and $\bm{\mu}_{e}^{\prime}$ is the derivative of the nuclear dipole with respect to $Q$ (see “Methods”). The term $\hat{H}_{\text{PFS}}$ polarizes the bare cavity Fock states (hence, denoted PFS for polarizes Fock states),[53] and its effect on the ESVP mediated down conversion process is discussed in section 3 of the SI. However, the term $\hat{H}_{\text{cav-vib}}$ is the dominant interaction term and mixes cavity and vibrational DOF. Therefore, we only consider $\hat{H}_{\textrm{system}}=\hat{H}_{\textrm{mol}}+\hat{H}_{\textrm{cav}}+\hat{H}_{\textrm{cav- vib}}$. For the application of the TWA we keep $\hat{H}_{\textrm{system}}$ in terms of phase space operators; nonetheless, we believe it valuable to connect our model with the coupling constant ($g$) used in Jaynes-Cummings like models.[21, 52, 13] This is done by substituting $\hat{Q}=\sqrt{\frac{\hbar}{2\omega_{v}}}(\hat{b}+\hat{b}^{\dagger})$ and $\hat{q}=\sqrt{\frac{\hbar}{2\omega_{c}}}(\hat{a}+\hat{a}^{\dagger})$ into Eq. (9), as well as taking the rotating wave approximation (RWA), giving rise to the expression $g=\omega_{c}A_{0}\bm{\epsilon}_{c}\cdot\,\bm{\mu_{\textrm{IR}}^{(e)}},$ (10) where $\bm{\mu}_{\textrm{IR}}^{(e)}=\bm{\mu}_{e}^{\prime}\sqrt{\frac{\hbar}{2\omega_{v}}}$ (11) is the vibrational transition dipole moment, which can be found through electronic structure calculations or from experimental IR spectra.[54, 55]. ### 2.3 Relevant Molecular Parameters Figure 2: Pictorial representation of the normal mode decomposition for an optical cavity coupled to a molecular vibration. The cavity and vibration harmonic oscillators are depicted as springs. The cavity, with position coordinate $q$, and vibration, with position coordinate $Q$, are coupled by $G$. The cavity and vibrational coordinates are then rotated into a coordinate system of two uncoupled oscillators called the polariton basis, with coordinates $R_{-}$ and $R_{+}$. Using the model Hamiltonian described in section 2.2, we determine the relevant molecular properties that are required for the photonic down conversion to occur through ESVP. This is achieved by monitoring the cavity photon occupancy ($\hat{N}_{\text{cav}}$), which is representative of the amount of photons that can be harnessed as IR emission if the cavity is leaky (finite $\mathcal{Q}$). Rather than representing the Hilbert space by a truncated basis of cavity photon and vibrational modes, we apply the TWA to arrive at an exact analytical expression for $\langle\hat{N}_{\text{cav}}\rangle$ (see SI section 2 for more details). Note that the RWA is not invoked here. To apply the TWA, we rotate $\hat{H}_{\textrm{system}}$ into a basis of two uncoupled oscillators called the polariton normal mode basis (see Figure 2 and “Methods”). Due to the displacement of the excited state PES, upon a vertical excitation of the molecule both the upper and lower polariton harmonic oscillators are displaced along their respective coordinates, as depicted in Figure 2. This displacement of excited state polariton oscillators creates a non-equilibrium initial condition that drives the dynamics. At finite temperature the resulting photon occupancy, following the UV/vis excitation, is given by $\begin{split}\langle\hat{N}_{\text{cav}}(t)\rangle=\frac{S\omega_{v}^{3}}{2(C^{2}+1)}&\bigg{[}\frac{1}{2\omega_{c}}\left(\frac{\sin(\Omega_{-}t)}{\Omega_{-}}-\frac{\sin(\Omega_{+}t)}{\Omega_{+}}\right)^{2}\\\ &+2\omega_{c}\left(\frac{\sin^{2}(\Omega_{-}t/2)}{\Omega_{-}^{2}}-\frac{\sin^{2}(\Omega_{+}t/2)}{\Omega_{+}^{2}}\right)^{2}\bigg{]}+N_{\beta},\end{split}$ (12) where $C=\frac{\omega_{v}^{2}-\omega_{c}^{2}}{4\frac{g}{\hbar}\sqrt{\omega_{v}\omega_{c}}}$, and where the polariton frequencies are given by $\Omega_{\pm}=\sqrt{A\pm B}$ with $A=\frac{1}{2}(\omega_{v}^{2}+\omega_{c}^{2})$ and $B=\frac{1}{2}\sqrt{(\omega_{v}^{2}-\omega_{c}^{2})^{2}+16\frac{g^{2}}{\hbar^{2}}\omega_{v}\omega_{c}}$. The effect of finite temperature is captured by $N_{\beta}=\frac{1}{4}\sum_{\gamma}\left(c_{c}^{(\gamma)2}\left(\frac{\Omega_{\gamma}}{\omega_{c}}+\frac{\omega_{c}}{\Omega_{\gamma}}\right)\coth(\beta\hbar\Omega_{\gamma}/2)\right)-\frac{1}{2}$, which describes the thermal distribution of vibrational polaritons on the electronic ground state, before electronic excitation of the molecule, where $\gamma$ denotes the upper ($\gamma=+$) or lower ($\gamma=-$) polariton mode, $\beta$ is the inverse temperature, and $c_{c}^{(\gamma)}$ is the projection of the cavity mode onto the $\gamma$ polariton mode. In the following analysis, we take the zero temperature limit ($\beta\rightarrow\infty$) and assume the cavity to be in perfect resonance with the vibration so that $\omega_{c}=\omega_{v}$, yielding $c_{c}^{(\gamma)}=1/\sqrt{2}$ for $\gamma=+$ and $\gamma=-$. This results in $N_{\beta}=\frac{1}{8}\sum_{\gamma}\left(\frac{\Omega_{\gamma}}{\omega_{c}}+\frac{\omega_{c}}{\Omega_{\gamma}}\right)-\frac{1}{2}$. For weak coupling, where $\Omega_{\gamma}\approx\omega_{c}=\omega_{v}$, it is apparent that $N_{\beta}\approx 0$. But, larger coupling strengths result in $N_{\beta}>0$, which is concurrent with the breakdown of the RWA. In that case, the polaritonic ground state within the RWA, which has no photons or vibrational quanta, mixes with the state containing a single cavity photon and single vibration, and therefore, even at zero temperature there will be a non- zero number of cavity photons if the coupling is large. A key result from Eq. (12) is that $\langle\hat{N}_{\text{cav}}(t)\rangle$ is proportional to $S$. This can be understood by considering that $S$ is proportional to the squared displacement of the excited state PES. If the displacement is zero, all FC factors will be zero indicating that no vibrational transitions will be induced upon the UV/vis excitation. Increasing $S$ increases the displacement, and consequentially, the number of vibrational quanta created upon electronic excitation. Given a non-zero $g$ value, these vibrational quanta will be converted into cavity photons. Figure 3: (a) Time evolution of the cavity photon occupancy, $\langle\hat{N}_{\text{cav}}\rangle$, for a single intramolecular vibration coupled to a cavity following an electronic excitation of the molecule, for different values of $g$ and for a fixed HR factor ($S=0.75$). (b) The maximum value of the $\langle\hat{N}_{\text{cav}}\rangle$ time trace (Max$\\{\langle\hat{N}_{\text{cav}}\rangle\\}$) as a function of both $g$ and $S$. The white dashed line represents $S=0.75$, with each point on the line corresponding to the $S$ and $g$ values used for the traces in (a). The cavity frequency was set in resonance with the vibration: $\omega_{c}=\omega_{v}=1600\textrm{ cm}^{-1}$. Figure 3a shows the time evolution of $\langle\hat{N}_{\text{cav}}\rangle$ given by Eq. (12) for a cavity mode in perfect resonance with the vibrational mode, i.e., $\omega_{c}=\omega_{v}=7.3$ mHa or 1600 $\textrm{cm}^{-1}$. Results are shown for five evenly spaced $g$ values between 0 and 0.9 mHa. At time zero, $\langle\hat{N}_{\text{cav}}\rangle=0$ for all values of $g$ since the UV/vis excitation acts exclusively on the electronic DOF, inducing vibrations in the excited state. This vibrational wavepacket has zero projection onto the cavity mode at time zero and is not an eigenstate of $\hat{H}_{\textrm{system}}$. As time evolves, Rabi oscillations occur as the wavepacket gains a non-zero projection onto the cavity mode giving rise to the $\langle\hat{N}_{\text{cav}}\rangle$ traces observed. The multi-oscillatory behavior observed in the $\langle\hat{N}_{\text{cav}}\rangle$ traces is due to interference between the contributing polariton frequencies. It can be seen that as $g$ increases so does the maximum value $\langle\hat{N}_{\text{cav}}\rangle$ can have (Max$\\{\langle\hat{N}_{\text{cav}}\rangle\\}$). For small $g$ values, Max$\\{\langle\hat{N}_{\text{cav}}\rangle\\}$ is approximately equal to the expected number of vibrational quanta produced upon electronic excitation, $S$. However, the energy stored in the cavity-vibration coupling will also contribute to $\langle\hat{N}_{\text{cav}}\rangle$, and as $g$ becomes less negligible in comparison to the cavity and vibration energies, $\langle\hat{N}_{\text{cav}}\rangle$ will become noticeably greater than $S$, as observed in Figure 3a. Figure 3b shows a heat map of Max$\\{\langle\hat{N}_{\text{cav}}\rangle\\}$ as a function of $g$ and $S$. From this, it is clear that maximizing both $g$ and $S$ will lead to the largest Max$\\{\langle\hat{N}_{\text{cav}}\rangle\\}$ value. Accordingly, a molecule that has a vibrational mode with a significant $\bm{\mu}_{\textrm{IR}}^{(e)}$ value, and hence a large $g$ value (see Eq. (10)), along with having significant vibronic coupling, i.e., large $S$, will make a good candidate for use in ESVP mediated photonic down conversion. ### 2.4 Radiative and Non-Radiative Dissipation of ESVP So far our analysis has assumed the molecule and cavity to be fully isolated, apart from their mutual coupling, without including emission from the cavity, which is an integral step of the ESVP mediated photonic down conversion process. Moreover, practical implementations of this process will suffer from non-radiative relaxation of the intramolecular vibration. It is therefore necessary to consider photonic emission and non-radiative dissipation from the cavity-vibration system. To this end, we are interested in how the photonic emission can be maximized throughout the parameter space of the system. It is evident that a large $S$ value will increase the probability of photonic down conversion as this will produce more excited state vibrational quanta. Other parameters, however, such as the non-radiative relaxation rate, the cavity- vibration coupling, and the cavity emission rate (related to $\mathcal{Q}$) have a non-trivial relationship to the percentage of incoming UV/vis photons converted to emitted IR photons. To study the influence of these parameters on the photonic down conversion process, we add two harmonic baths to our model: 1) a bath representing an external electromagnetic field that accounts for cavity emission, and 2) a bath of (solvent) phonon modes that account for non-radiative vibrational relaxation. Accordingly, the total system-bath Hamiltonian is given by $H_{\textrm{total}}=H_{\textrm{system}}+\sum_{i=1}^{E+B}\left(\frac{\hat{p}_{i}^{2}}{2}+\omega_{i}^{2}\frac{\hat{r}_{i}^{2}}{2}\right)+\sum_{i=1}^{E}\kappa_{i}\,\hat{r}_{i}\hat{q}+\sum_{i=E+1}^{E+B}\chi_{i}\,\hat{r}_{i}\hat{Q},$ (13) where $\hat{r}_{i}$ is the position operator for the $1\leq i\leq E$ external field modes and $E<i\leq E+B$ non-radiative bath modes, $\kappa_{i}$ is the coupling constant between the $i$th external field mode and the cavity, and $\chi_{i}$ is the coupling constant between the $i$th non-radiative bath mode and the vibration. It should be noted that, from Eq. (26) and Eq. (31) in “Methods”, $\kappa_{i}=\sqrt{\frac{\omega_{i}}{2\pi\mathcal{Q}D_{E}}},$ (14) where $D_{E}$ is the density of states of the external electromagnetic field, chosen to be constant as a function of $\omega_{i}$.[56] Hence, the cavity- external field coupling, $\kappa_{i}$, and $\mathcal{Q}$ are inversely related. This is an intuitive result because as $\mathcal{Q}$ decreases, cavity photons are expected to dissipate more efficiently due to a stronger coupling of the cavity to the external field. Figure 4: (a) Schematic showing the total Hamiltonian, $\hat{H}_{\textrm{total}}$, including the cavity-molecule system, the non- radiative bath modes, and the external-field modes that constitute emission. (b-d) Dynamics of different occupation numbers for different quality factor values ($\mathcal{Q}$) at $g=0.04$ mHa: Black = cavity photons; blue = intramolecular vibrations; orange = external field photons; green = non- radiative bath. (e) A heat map of the % emission of IR photons as a function of $g$ and $\mathcal{Q}^{1/2}$, where % emission is evaluated at 18 ps. (f) % emission as a function of $\mathcal{Q}^{1/2}$ evaluated at different $g$ values. These traces correspond to % emission along the dashed lines in (e). The labeled points along the $g=0.04$ mHa trace correspond to the $\mathcal{Q}$ and $g$ values of the dynamics shown in (b-d). The cavity frequency was set in resonance with the vibration: $\omega_{c}=\omega_{v}=1600\textrm{ cm}^{-1}$. The schematic in Figure 4a represents the total Hamiltonian given in Eq. (13). All components of the ensemble are comprised of harmonic oscillators and the red arrows show the coupling between components. Following a similar procedure as was done for the calculation of $\langle\hat{N}_{\text{cav}}\rangle$ in the isolated cavity-vibration case, we used the TWA to explicitly describe the time evolution of the system, non-radiative bath, and external field. This time, normal modes of the $\hat{H}_{\textrm{total}}$ Hessian matrix were determined. The time evolution was monitored by means of the occupation numbers of the different components. Figures 4b-d show the resulting occupation numbers following a UV/vis excitation of the molecule for different $\mathcal{Q}$ values and for a fixed $g$ value of 0.04 mHa. These data show that by changing $\mathcal{Q}$, regimes with quantitatively different behaviors are obtained. For the relatively small value $\mathcal{Q}$ $=9$ (Figure 4b), the cavity occupancy is significantly dampened due the strong coupling to the external field. Therefore, the cavity dissipation lifetime is shorter than the Rabi oscillation period. It can also be seen that the non-radiative relaxation (green) out-competes photonic emission (orange) for this case. For the value $\mathcal{Q}$ $=100$ (Figure 4c), short lived Rabi-oscillations are observed. The presence of Rabi- oscillations are indicative of strong coupling between the cavity and the molecular vibrational mode. In this case, the photonic emission outcompetes the non-radiative relaxation. For the value $\mathcal{Q}$ $=1600$ (Figure 4d), longer-lived Rabi-oscillations are observed between the cavity and the molecular vibration mode. In this case, similar to the $\mathcal{Q}$ $=9$ case, the non-radiative relaxation outcompetes the photonic emission. The change in dynamic behavior with $\mathcal{Q}$ observed in Figures 4b-d are the result of $\mathcal{Q}$ affecting the cavity decay lifetime (contingent on both the radiative and non-radiative dissipation rates) relative to $g$. The strong coupling regime, indicated by the presence of Rabi-oscilations between the cavity and vibration as well as peak splitting in the emission spectrum, occurs when the cavity decay lifetime is long compared to the cavity-vibration Rabi oscillation period.[57] Figure 4e shows a heat map of the percentage of vibrations induced by the UV/vis excitation converted to photonic emission (% emission) as a function of $g$ and $\mathcal{Q}$. The % emission is calculated as the ratio of external field photons at equilibrium normalized to the initial number of vibrations created by the UV/vis excitation, given by $S$, that is, $\textrm{\% emission}=\frac{1}{S}\sum_{i}^{E}\langle\hat{N}_{\textrm{ext}}^{\,i}(t_{\textrm{eq}})\rangle.$ (15) Here, $\hat{N}_{\textrm{ext}}^{\,i}$ is the $i$th external field occupation number operator and $t_{\textrm{eq}}$ is the time at which equilibrium is reached, here taken to be $t_{\textrm{eq}}=18$ ps, which suffices for all values of $g$ and $\mathcal{Q}$ used. By using this definition the % emission is independent of $S$. For example, the results in Figure 4e were computed with $S=0.5$ but would be identical for any $S$ value used. Determining the net amount of photons produced in the photonic down conversion process simply requires multiplying the % emission by $S$. Figure 4f shows slices of the % emission heat map along $\mathcal{Q}$ for different $g$ values given by the dashed lines in Figure 4e. The labeled points in 4f indicate the related dynamics in Figure 4b-d that give rise to those % emission values. It can be seen from Figure 4f that the % emission increases with $g$ for all values of $\mathcal{Q}$, which indicates that increasing $g$ will always increase the photonic down conversion yield. Furthermore, all traces in Figure 4 have a maximum along $\mathcal{Q}$, which we denote as $\mathcal{Q}^{\textrm{max}}$. This is due to the competition between radiative and non-radiative decay. For $\mathcal{Q}$ values larger than $\mathcal{Q}^{\textrm{max}}$, the probability of cavity photons leaking from the cavity into the external field becomes small enough that non- radiative dissipation outcompetes emission. On the other hand, for $\mathcal{Q}$ values smaller than $\mathcal{Q}^{\textrm{max}}$, the cavity mode is coupled strong enough to the external field modes that it begins to decouple to the vibration, and again, non-radiative dissipation outcompetes emission. In fact, in the limit where $\mathcal{Q}$ becomes very small, the cavity mode mixes into the external field to the extent that the system effectively becomes a vibration interacting with a free-space electromagnetic field, and radiative emission from the vibration is not expected due to the fast non-radiative relaxation. $\mathcal{Q}^{\textrm{max}}$ should be applicable for any polariton system where the matter component has a non-radiative relaxation pathway, implying that a finite cavity $\mathcal{Q}$ value is needed to maximize the photonic emission from such systems. This motivates cavity design principles for measuring and utilizing ESVP mediated photonic down conversion, as well as for any system that uses an optical resonator to enhance emission. ### 2.5 Calculation of Relevant Parameters for Pyrene and PAA The identification of molecules that allow for the formation of ESVP through a UV/vis excitation is critical for all potential applications of the ESVP mediated photonic down conversion process. Following the results of Figures 3 and 4, molecules must contain vibrational modes that are both FC active (i.e., having non-zero HR factor) and IR active within the excited state for the ESVP photonic down conversion to occur. According to the rule of mutual exclusion, such modes exist only in molecules lacking inversion symmetry (vide infra). Furthermore, the extent to which the photonic down conversion process can occur in non-centrosymmetric molecules is dependent on the strength of their vibronic couplings and excited state IR activity. This section focuses on extracting these molecular properties in specific molecules while making a connection to our theoretical model. Figure 5: DFT calculations of the vibronic couplings and excited state IR dipoles for pyrene and 1-pyreneacetic acid (PAA). Calculated absorption spectra for (a) pyrene and (b) PAA were performed at different temperatures (red = 300 K; blue = 0.1 K) and with different standard deviations ($\sigma$) for the Gaussian distribution of electronic energies taken to represent inhomogeneous broadening (red $=200\textrm{ cm}^{-1}$; blue $=10\textrm{ cm}^{-1}$). Experimental spectra (black) were taken in cyclohexane for pyrene and toluene for PAA at room temperature ($\sim$ 293 K). Calculated HR factors ($S$) and excited state IR dipole moments ($\mu^{(e)}_{\textrm{IR}}$) for vibrational modes of (c) pyrene and (d) PAA are plotted as a function of the wavenumber of each mode. $\mu^{(e)}_{\textrm{IR}}$ is plotted “on top of” the $S$ values, and are only shown for modes with $S>0$, while its value is indicated above each bar. We have performed DFT and time-dependent-DFT (TD-DFT) calculations for two molecules: one with and one without inversion symmetry. Figure 5 shows results for both pyrene and 1-pyreneacetic acid (PAA), which are rigid chromophores with clear vibronic progressions due to vibronic coupling. Pyrene is centrosymmetric and should not have vibrational modes that are both IR and Raman active in accordance with the rule of mutual exclusion.[58] This is relevant as Raman activity implies FC activity. In centrosymmetric molecules, all electronic states contain an inversion symmetry such that electronically exciting them will only induce symmetric vibrations, meaning the excited state PES is displaced only along these symmetric modes. This displacement allows ground state vibrational overtones to be reached in the second order Raman process making them Raman active.[59] Importantly, since only symmetric modes are induced upon electric excitation, which are IR inactive, no ESVP should form in centrosymetric molecules through UV/vis excitation. PAA, on the other hand, has a carboxylic acid functional group that breaks the inversion symmetry of the pyrene moiety and should contain IR and FC active vibrational modes. By comparing these two molecules we can test whether simple violations of the rule of mutual exclusion can serve to design molecules for ESVP mediated photonic down conversion. In order to determine the accuracy of the calculations, absorption spectra were calculated and compared to experimental spectra (Figure 5a and 5b). These spectra were calculated using the TWA as detailed in other work,[60] and are exact within the FC and harmonic approximation. Results were calculated at finite temperature by taking a Boltzmann population on the ground state vibrations and also include inhomogeneous broadening due to a Gaussian distribution ($\sigma$ denoting the standard deviation) of the electronic transition energy. At room temperature, calculated spectra (red) with $\sigma=200\textrm{ cm}^{-1}$ show excellent qualitative agreement with the experimental spectra (black) indicating that the calculations capture the correct vibronic couplings and normal modes within the molecules. The calculated spectra at 0.1 K and $\sigma=10\textrm{ cm}^{-1}$ remove the majority of the inhomogenous broadening as well as broadening due to transitions from hot ground state vibrations that occur at room temperature. This gives “the skeleton” of the vibronic progression with peaks corresponding only to transitions from ground vibrational states on the electronic ground state. A renormalization of the DFT-calculated electronic transition energy was applied for our calculations to match the experimental data. Figure 5c and 5d show $S$ (blue) and $\bm{\mu}_{\textrm{IR}}^{(e)}$ (green) for the different vibrational modes as a function of each mode’s wavenumber for both pyrene and PAA. Although both molecules contain many modes with non- zero $\bm{\mu}_{\textrm{IR}}^{(e)}$ values, we only show $\bm{\mu}_{\textrm{IR}}^{(e)}$ for modes with non-zero $S$ values as these are the only vibrational modes that are excited by electronic transitions. $S$ values are calculated for ground state vibrations and $\bm{\mu}_{\textrm{IR}}^{(e)}$ values are calculated for excited state normal modes. Because Duschinsky rotations can occur,[61] the overlap squared of each ground state mode with each excited state mode was taken in order to match them. In both molecules, there was generally minimal mixing of ground state normal modes in the excited state, validating our application of the FC approximation. Figure 5c shows that each FC active vibrational mode has a zero $\bm{\mu}_{\textrm{IR}}^{(e)}$ value for pyrene, in accordance with the rule of mutual exclusion. On the other hand, Figure 5d shows that each FC active vibrational mode in PAA has a non-zero $\bm{\mu}_{\textrm{IR}}^{(e)}$ value due to the broken inversion symmetry of the pyrene moiety. Therefore, an optical cavity may be tuned to one of the FC active vibrational modes of PAA shown in Figure 5d—a good choice being the highest frequency mode at 1680 $\textrm{cm}^{-1}$—and upon UV/vis excitation of the PAA molecule, ESVP would form. The results from the DFT calculations presented here can be connected to the theoretical model presented in sections 2.3 and 2.4 through the calculated $S$ and $g$ values. However, for a single molecule coupled to a cavity mode, $g$ is typically very small when compared to values required for the UV/vis-to-IR photonic down conversion process. For example, for the mode at 1680 $\textrm{cm}^{-1}$, $\bm{\mu}_{\textrm{IR}}^{(e)}=0.57$ Debye, which corresponds to a $g$ value of $1.8\times 10^{-6}$ mHa. This is assuming a cavity volume of $V=1.06\times 10^{-16}\textrm{ m}^{3}$ using $V=\lambda^{3}/2$, where $\lambda$ is the wavelength of the cavity mode. Comparing this $g$ value with Figure 4e, there would be negligible photonic down conversion for all $\mathcal{Q}$ values. The small $g$ value and photonic down conversion yield are due to having only a single molecule coupled to the cavity. Experimental vibrational-polariton measurements rely on an ensemble of molecules to be coupled to a cavity mode, and the Rabi splittings observed correspond to larger coupling strengths than would be expected from each individual molecule.[62] This is refereed to as collective strong coupling for which it is known that the Rabi-splitting increases with the square root of the number of cavity coupled molecules ($N$).[13, 63] As a rough estimate, an effective $g$ value that scales with $\sqrt{N}$ can, therefore, be considered. Using the concentration of molecules present in the optical cavity and knowing the molecule’s absorption cross section, one could estimate the number of excited molecules within the cavity and calculate an effective coupling strength from the DFT results. This would allow the % emission from photonic down conversion of the molecule to be determined as a function of the $\mathcal{Q}$. It should be noted, however, that collective strong coupling remains underexplored, especially in the presence of multiple quanta such as involved in the ESVP mediated photonic down conversion process, the topic of which we reserve for future work. ## 3 Discussion In this work, a novel UV/vis-to-IR photonic down conversion process is shown to be driven by ESVP for a molecule in an optical cavity. Accordingly, a UV/vis excitation of the molecule induces vibrational polaritons on the molecule’s electronic excited state potential. Such ESVP may then leak IR light, resulting in a photonic down conversion that can potentially be exploited for applications in sensing, quantum information, and in the manipulation of photo-induced reactions. The mechanism and relevant molecular parameters that allow for the down conversion process to occur are the HR factor and the excited state IR dipole, which both need to be non-zero. According to the rule of mutual exclusion, this implies that the photonic down conversion process can only occur within molecules lacking inversion symmetry. By studying the parameter space of the ESVP mediated photonic down conversion while including both radiative and non-radiative dissipation, we determined the optimal parameters to maximize the down conversion yield, facilitating experimental and practical implementations. Due to the inclusion of non- radiative dissipation, the photonic down conversion is maximized at a specific cavity quality factor. With excess quality factors, the probability of cavity photon leakage becomes small and non-radiative processes will out-compete emission. However, with lower than optimal quality factors, the cavity mode becomes highly delocalized through mixing with free-space electromagnetic field modes outside of the cavity inhibiting strong coupling with the molecular vibration. These two limiting behaviors give rise to an intermediate quality factor, $\mathcal{Q}^{\textrm{max}}$, that maximizes the down conversion process. In the previous subsection, we demonstrated how to use DFT calculations to predict if ESVP will form on a molecule when it is excited with UV/vis light within a cavity. We believe this to be useful for designing molecules that maximize the ESVP mediated photonic down conversion, allowing for applications in quantum information and sensing to be realized. Such DFT calculations can also determine if an optical cavity could be used to modulate a molecule’s excited state dynamics following a UV/vis excitation with ramifications for photo-induced reactions. The importance of the rule of mutual exclusion is emphasized by our evaluation of both pyrene, containing and inversion symmetry, and 1-pyreneacetic acid, lacking an inversion symmetry. Our results show that only 1-pyreneacetic acid has both FC active (non-zero HR factor) and excited state IR active vibrational modes, which permit the formation of ESVP. We note that a potential exception to the requirement of molecular inversion symmetry for the formation of ESVP induced by a UV/vis excitation would be the occurrence of excited state symmetry breaking due to solvent interactions.[64] This results in ground state vibrational modes of different symmetries mixing in the excited state and could allow excited state IR active vibrations to be produced on ground state centrosymmetric molecules following a UV/vis excitation of the molecule. These vibrations could, therefore, couple to a cavity, opening further ways to mediate the photonic down conversion. ## 4 Methods ### 4.1 Diabatic Expansion of the Dipole Operator The dipole operator can be separated into components that depend on electronic and nuclear DOF as $\bm{\hat{\mu}}=\bm{\hat{\mu}_{\textrm{el}}}+\bm{\hat{\mu}_{\textrm{nu}}}.$ (16) When expanding $\hat{H}_{\textrm{system}}$ in the diabatic basis, each element of the dipole operator in this basis becomes $\bm{\hat{\mu}_{\alpha\beta}}(\hat{Q})=\langle\alpha|\bm{\hat{\mu}_{\textrm{el}}}|\beta\rangle+\bm{\hat{\mu}_{\textrm{nu}}}(\hat{Q})\delta_{\alpha\beta},$ (17) which act on the nuclear Hilbert space of $\hat{Q}$ and not on the electronic Hilbert space. Note that $\bm{\hat{\mu}_{\textrm{nu}}}$ does not act on the diabatic basis states as they do not depend on $Q$ (see Eq. (2)). The linear expansion of each element of the dipole operator is given by $\bm{\hat{\mu}}_{\alpha\beta}(\hat{Q})=\bm{\mu_{\alpha\beta}}(Q_{0}^{(g)})+\frac{\partial\bm{\mu_{\alpha\beta}}(Q)}{\partial Q}\bigg{|}_{Q_{0}^{(g)}}(\hat{Q}-Q_{0}^{(g)}),$ (18) with the first coefficient being $\bm{\mu_{\alpha\beta}}(Q_{0}^{(g)})=\langle\alpha|\bm{\hat{\mu}_{\textrm{el}}}|\beta\rangle+\bm{\mu_{\textrm{nu}}}(Q_{0}^{(g)})\delta_{\alpha\beta},$ (19) where $\langle\alpha|\bm{\hat{\mu}_{\textrm{el}}}|\beta\rangle$ is the permanent electic dipole of state $\alpha$ when $\alpha=\beta$ and the transition dipole between the states $\alpha$ and $\beta$ when $\alpha\neq\beta$. Further, $\bm{\mu_{\textrm{nu}}}(Q_{0}^{(g)})$ is the dipole moment of the nuclei at configuration $Q_{0}^{(g)}$. It follows that the second coefficient in Eq. (18) is given by the derivative of the nuclear dipole moment with respect to $Q$ at configuration $Q_{0}^{(g)}$, $\frac{\partial\bm{\mu_{\alpha\beta}}(Q)}{{\partial Q}}\bigg{|}_{Q_{0}^{(g)}}=\frac{\partial\bm{\mu_{\textrm{nu}}}(Q)}{{\partial Q}}\bigg{|}_{Q_{0}^{(g)}}\delta_{\alpha\beta}.$ (20) We ignored coupling between the cavity and the molecule’s electronic DOF as the cavity mode is assumed to be near resonant with the vibrational mode of interest making it significantly off-resonant from the electronic transition. Accordingly, we set elements of the dipole operator where $\alpha\neq\beta$ to zero. Because of this, the electronic ground and excited state contributions to $\hat{H}_{\textrm{system}}$ are uncoupled and can be treated separately. For the purpose of this study, we focused on the excited state contribution where $\alpha=\beta=e$. To simplify notation we defined $\bm{\mu}_{e}^{\prime}\equiv\frac{\partial\bm{\mu_{ee}}(Q)}{\partial Q}\bigg{|}_{Q_{0}^{(g)}},$ (21) and $\bm{\mu}_{e}^{0}\equiv\bm{\mu_{ee}}(Q_{0}^{(g)})-\bm{\mu}_{e}^{\prime}Q_{0}^{(g)}.$ (22) Note that $\bm{\mu}_{e}^{0}$ is the permanent dipole moment of the excited state at the ground state equilibrium configuration, which was taken to be $Q_{0}^{(g)}=0$. ### 4.2 Polariton Normal Mode Basis $\hat{H}_{\textrm{system}}$ can be rotated into a basis of two uncoupled oscillators called the polariton normal mode basis (see Figure 2). This basis is found by determining the eigenvectors of the cavity-vibration Hessian matrix, $\bm{H}=\begin{pmatrix}\omega_{c}^{2}&G\\\ G^{*}&\omega_{v}^{2}\end{pmatrix}.$ (23) The polariton harmonic oscillators have upper frequency ($\Omega_{+}$) and lower frequency ($\Omega_{-}$) given by the square root of the eigenvalues of $\bm{H}$. The off diagonal element of the Hessian matrix, $G$, is the cavity- vibration bilinear coupling constant given by $G=\frac{2}{\hbar}\sqrt{\omega_{c}\omega_{v}}g.$ (24) The analytical expression for $\langle\hat{N}_{\text{cav}}(t)\rangle$ was determined by rotating the cavity photon occupancy operator, $\hat{N}_{\text{cav}}=\frac{1}{2\hbar\omega_{c}}(\omega_{c}^{2}\hat{q}^{2}+\hat{p}^{2})-1/2,$ (25) into the polariton basis by expanding $\hat{q}$ and $\hat{p}$ in terms of polariton phase space operators, $\hat{R}_{-}$ and $\hat{R}_{+}$. The Weyl symbol of $\hat{N}_{\text{cav}}$, in the polariton basis, was determined by replacing the polariton phase space operators with phase space variables that can be evolved according to the classical equations of motion for two uncoupled harmonic oscillators. From this, we calculate the expectation value of the time evolved Weyl symbol of $\hat{N}_{\text{cav}}$ acting on the time independent Wigner distribution of the initial position and momentum of the system.[65, 43] ### 4.3 System-Bath Couplings An ohmic spectral density was taken for both the non-radiative bath and the external field, $J_{E}(\omega)=\eta_{E}\,\omega$ and $J_{B}(\omega)=\eta_{B}\,\omega$, respectively, where $\eta_{E}$ ($\eta_{B}$) is the cavity (vibration) damping constant due to coupling to the external field (non-radiative bath). By taking a large but finite number of non- radiative bath and external field modes (we used 500 modes for each bath), $\hat{H}_{\textrm{total}}$ approximates the thermodynamic number of bath modes under experimental conditions.[66] In our modeling, we used a constant density of states ($D_{B}$ for the non-radiative bath; $D_{E}$ for the external field) and frequency dependent coupling constants, $\kappa_{i}=\sqrt{\frac{\eta_{E}\,\omega_{i}}{D_{E}}},$ (26) and $\chi_{i}=\sqrt{\frac{\eta_{B}\,\omega_{i}}{D_{B}}}.$ (27) In all of our calculations, the vibrational damping constant, $\eta_{B}$, was fixed such that an excited vibration would have a lifetime of 2 ps for $g=0$, which is typical of condensed-phase molecular vibrations (we study different non-radiative lifetimes in section 5 of the SI).[26, 27] Furthermore, $\eta_{E}$ is related to $\mathcal{Q}$, which is defined as the ratio of the cavity center frequency to its emission linewidth, $\Delta\omega_{c}$,[67] $\mathcal{Q}=\frac{\omega_{c}}{\Delta\omega_{c}}.$ (28) An expression for the emission profile linewidth of an optical cavity coupled to a bath of external field modes can be extracted from the Lindblad master equation as[45] $\Delta\omega_{c}=2\pi D_{E}|\kappa_{\omega_{c}}|^{2},$ (29) where $\kappa_{\omega_{c}}$ is the cavity-external field coupling at the cavity frequency. Substituting equation 29 into equation 28 gives $\mathcal{Q}=\frac{\omega_{c}}{2\pi D_{E}|\kappa_{\omega_{c}}|^{2}}.$ (30) Setting $\kappa_{i}=\kappa_{\omega_{c}}$ and $\omega_{i}=\omega_{c}$ in equation 26 and then comparing the result to equation 30, it follows that $\eta_{E}=\frac{1}{2\pi\mathcal{Q}}.$ (31) It should be noted that the Lindblad equation invokes the RWA and therefore will give inaccurate $\mathcal{Q}$ values in the ultrastrong cavity-external field coupling regime. ### 4.4 DFT Calculations of Huang-Rhys Factors and Excited State IR Dipoles DFT calculations were performed using the Qchem software (version 5.3). All DFT and TD-DFT calculations utilized the B3LYP functional and $\textrm{6-31+G}^{*}$ basis set. In order to determine $S$ and $\bm{\mu}_{\textrm{IR}}^{(e)}$ for pyrene and PAA, first the ground state equilibrium geometry was calculated for each molecule in its singlet ground state, followed by a normal mode analysis about this ground state minimum to obtain the intramolecular vibrations, both done via DFT calculations. TD-DFT calculations were then used to determine the forces felt by the nuclei due to the molecule being excited into the first, bright excited singlet state. This force is due to the electron density of the excited state acting on the ground state nuclear configuration, shifting the equilibrium nuclear configuration. Excited state force vectors from the TD-DFT calculations were projected onto the ground state vibrational modes. These forces determine the shift in the excited state PES and were used to calculate HR factors for each vibrational mode. Finally, we performed TD-DFT excited state normal mode analyses at the FC point to calculate the excited state intramolecular vibrations and their corresponding $\bm{\mu}_{\textrm{IR}}^{(e)}$ values. No description of solvent interactions was included in the computed spectra. ## 5 Acknowledgments The authors thank Jonathan D. Schultz for the helpful discussions regarding this work. This work was supported as part of the Center for Molecular Quantum Transduction (CMQT), an Energy Frontier Research Center funded by the U.S. Department of Energy, Office of Science, Basic Energy Sciences under Award No. DE-SC0021314. This material is based upon work supported by the National Science Foundation Graduate Research Fellowship under Grant No. DGE-2234667. ## Declarations The authors declare no competing interests. ## References * [1] Andrews, D. L., Craig, D. P. & Thirunamachandran, T. Molecular quantum electrodynamics in chemical physics. _International Reviews in Physical Chemistry_ 8 (4), 339–383 (1989). * [2] Mukamel, S. _Principles of Nonlinear Optical Spectroscopy_ (Oxford University Press, 1995). * [3] Ebbesen, T. W. Hybrid Light-Matter States in a Molecular and Material Science Perspective. _Accounts of Chemical Research_ 49 (11), 2403–2412 (2016). * [4] Thomas, A. _et al._ Ground-State Chemical Reactivity under Vibrational Coupling to the Vacuum Electromagnetic Field. _Angewandte Chemie - International Edition_ 55 (38), 11462–11466 (2016). * [5] Vergauwe, R. M. _et al._ Modification of Enzyme Activity by Vibrational Strong Coupling of Water. _Angewandte Chemie - International Edition_ 58 (43), 15324–15328 (2019). * [6] Thomas, A. _et al._ Tilting a ground-state reactivity landscape by vibrational strong coupling. _Science_ 363 (6427), 615–619 (2019). * [7] Pang, Y. _et al._ On the Role of Symmetry in Vibrational Strong Coupling: The Case of Charge‐Transfer Complexation. _Angewandte Chemie_ 132 (26), 10522–10526 (2020). * [8] Hirai, K., Takeda, R., Hutchison, J. A. & Uji-i, H. Modulation of Prins Cyclization by Vibrational Strong Coupling. _Angewandte Chemie - International Edition_ 59 (13), 5332–5335 (2020). * [9] Lather, J. & George, J. Improving Enzyme Catalytic Efficiency by Co-operative Vibrational Strong Coupling of Water. _Journal of Physical Chemistry Letters_ 12 (1), 379–384 (2021). * [10] Sau, A. _et al._ Modifying Woodward–Hoffmann Stereoselectivity Under Vibrational Strong Coupling. _Angewandte Chemie_ 133 (11), 5776–5781 (2021). * [11] Dunkelberger, A. D., Spann, B. T., Fears, K. P., Simpkins, B. S. & Owrutsky, J. C. Modified relaxation dynamics and coherent energy exchange in coupled vibration-cavity polaritons. _Nature Communications_ 7, 1–10 (2016). * [12] Saurabh, P. & Mukamel, S. Two-dimensional infrared spectroscopy of vibrational polaritons of molecules in an optical cavity. _Journal of Chemical Physics_ 144 (12) (2016). * [13] Ribeiro, R. F., Martínez-Martínez, L. A., Du, M., Campos-Gonzalez-Angulo, J. & Yuen-Zhou, J. Polariton chemistry: controlling molecular dynamics with optical cavities. _Chemical Science_ 9 (30), 6325–6339 (2018). * [14] Simpkins, B. S., Dunkelberger, A. D. & Owrutsky, J. C. Mode-Specific Chemistry through Vibrational Strong Coupling (orA Wish Come True). _Journal of Physical Chemistry C_ 125 (35), 19081–19087 (2021). * [15] Li, X., Mandal, A. & Huo, P. Cavity frequency-dependent theory for vibrational polariton chemistry. _Nature Communications_ 12 (1), 1–9 (2021). * [16] Dunkelberger, A. D., Simpkins, B. S., Vurgaftman, I. & Owrutsky, J. C. Vibration-Cavity Polariton Chemistry and Dynamics. _Annual Review of Physical Chemistry_ 73, 429–451 (2022). * [17] Li, T. E., Nitzan, A. & Subotnik, J. E. Energy-efficient pathway for selectively exciting solute molecules to high vibrational states via solvent vibration-polariton pumping. _Nature Communications_ 13 (1), 4203 (2022). * [18] Lindoy, L. P., Mandal, A. & Reichman, D. R. Quantum Dynamics of Vibrational Polariton Chemistry (2022). * [19] Shalabney, A. _et al._ Coherent coupling of molecular resonators with a microcavity mode. _Nature Communications_ 6 (Umr 7006), 1–6 (2015). * [20] Strashko, A. & Keeling, J. Raman scattering with strongly coupled vibron-polaritons. _Physical Review A_ 94 (2), 1–10 (2016). * [21] Herrera, F. & Spano, F. C. Absorption and photoluminescence in organic cavity QED. _Physical Review A_ 95 (5), 1–24 (2017). * [22] Purcell, E. M. in Spontaneous Emission Probabilities at Radio Frequencies 839–839 (1995). * [23] Lalanne, P., Yan, W., Vynck, K., Sauvan, C. & Hugonin, J. P. Light Interaction with Photonic and Plasmonic Resonances. _Laser and Photonics Reviews_ 12 (5), 1–38 (2018). * [24] Jonas, D. M. Vibrational and Nonadiabatic Coherence in 2D Electronic Spectroscopy, the Jahn-Teller Effect, and Energy Transfer. _Annual Review of Physical Chemistry_ 69, 327–352 (2018). * [25] Nelson, T. R. _et al._ Non-adiabatic Excited-State Molecular Dynamics: Theory and Applications for Modeling Photophysics in Extended Molecular Materials. _Chemical Reviews_ 120 (4), 2215–2287 (2020). * [26] Laubereau, A., Von Der Linde, D. & Kaiser, W. Direct measurement of the vibrational lifetimes of molecules in liquids. _Physical Review Letters_ 28 (18), 1162–1165 (1972). * [27] Oxtoby, D. W. in Vibrational Population Relaxation in Liquids , Vol. 47 487–519 (2007). * [28] Mark, F. _Quantum Optics: An Introduction_ (Oxford University Press, 2006). * [29] Metzger, B. _et al._ Purcell-Enhanced Spontaneous Emission of Molecular Vibrations. _Physical Review Letters_ 123 (15), 153001 (2019). * [30] Bakulin, A. A. _et al._ Real-time observation of multiexcitonic states in ultrafast singlet fission using coherent 2D electronic spectroscopy. _Nature Chemistry_ 8 (1), 16–23 (2016). * [31] Monahan, N. R. _et al._ Dynamics of the triplet-pair state reveals the likely coexistence of coherent and incoherent singlet fission in crystalline hexacene. _Nature Chemistry_ 9 (4), 341–346 (2017). * [32] Fujihashi, Y., Chen, L., Ishizaki, A., Wang, J. & Zhao, Y. Effect of high-frequency modes on singlet fission dynamics. _Journal of Chemical Physics_ 146 (4) (2017). * [33] Tempelaar, R. & Reichman, D. R. Vibronic exciton theory of singlet fission . III . How vibronic coupling and thermodynamics promote rapid triplet generation in pentacene crystals 244701, 1–9 (2018). * [34] Duan, H. G. _et al._ Intermolecular vibrations mediate ultrafast singlet fission. _Science Advances_ 6 (38) (2020). * [35] Schultz, J. D. _et al._ Influence of Vibronic Coupling on Ultrafast Singlet Fission in a Linear Terrylenediimide Dimer. _Journal of the American Chemical Society_ (2021). * [36] Womick, J. M. & Moran, A. M. Vibronic enhancement of exciton sizes and energy transport in photosynthetic complexes. _Journal of Physical Chemistry B_ 115 (6), 1347–1356 (2011). * [37] Tiwari, V., Peters, W. K. & Jonas, D. M. Electronic resonance with anticorrelated pigment vibrations drives photosynthetic energy transfer outside the adiabatic framework. _Proceedings of the National Academy of Sciences of the United States of America_ 110 (4), 1203–1208 (2013). * [38] Fuller, F. D. _et al._ Vibronic coherence in oxygenic photosynthesis. _Nature Chemistry_ 6 (8), 706–711 (2014). * [39] Cao, J. _et al._ Quantum biology revisited (April), 1–12 (2020) . * [40] Wang, Q. _et al._ Vibrationally Coherent Photochemistry in the Femtosecond Primary Event of Vision Published by : American Association for the Advancement of Science Stable URL : http://www.jstor.org/stable/2885325 Your use of the JSTOR archive indicates your acceptance of. _Science_ 266 (5184), 422–424 (1994) . * [41] Johnson, P. J. _et al._ Local vibrational coherences drive the primary photochemistry of vision. _Nature Chemistry_ 7 (12), 980–986 (2015). * [42] Johnson, P. J. _et al._ The Primary Photochemistry of Vision Occurs at the Molecular Speed Limit. _Journal of Physical Chemistry B_ 121 (16), 4040–4047 (2017). * [43] Polkovnikov, A. Phase space representation of quantum dynamics. _Annals of Physics_ 325 (8), 1790–1852 (2010). * [44] Breuer, H.-P. & Petruccione, F. _The Theory of Open Quantum Systems_ Vol. 15 (Oxford University Press, 2007). * [45] Carmichael, H. J. _Statistical Methods in Quantum Optics 1_ (Springer Berlin Heidelberg, Berlin, Heidelberg, 1999). * [46] Huang, K. & Rhys, A. Theory of light absorption and non-radiative transitions in F-centres. _Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences_ 204 (1078), 406–423 (1950). * [47] De Jong, M., Seijo, L., Meijerink, A. & Rabouw, F. T. Resolving the ambiguity in the relation between Stokes shift and Huang-Rhys parameter. _Physical Chemistry Chemical Physics_ 17 (26), 16959–16969 (2015). * [48] Lalanne, P., Sauvan, C. & Hugonin, J. P. Photon confinement in photonic crystal nanocavities. _Laser and Photonics Reviews_ 2 (6), 514–526 (2008). * [49] Flick, J., Ruggenthaler, M., Appel, H. & Rubio, A. Atoms and molecules in cavities, from weak to strong coupling in quantum-electrodynamics (QED) chemistry. _Proceedings of the National Academy of Sciences of the United States of America_ 114 (12), 3026–3034 (2017). * [50] Schäfer, C., Ruggenthaler, M. & Rubio, A. Ab initio nonrelativistic quantum electrodynamics: Bridging quantum chemistry and quantum optics from weak to strong coupling. _Physical Review A_ 98 (4) (2018). * [51] Rokaj, V., Welakuh, D. M., Ruggenthaler, M. & Rubio, A. Light-matter interaction in the long-wavelength limit: No ground-state without dipole self-energy. _Journal of Physics B: Atomic, Molecular and Optical Physics_ 51 (3) (2018). * [52] Herrera, F. & Spano, F. C. Theory of Nanoscale Organic Cavities: The Essential Role of Vibration-Photon Dressed States. _ACS Photonics_ 5 (1), 65–79 (2018). * [53] Mandal, A., Montillo Vega, S. & Huo, P. Polarized Fock States and the Dynamical Casimir Effect in Molecular Cavity Quantum Electrodynamics. _Journal of Physical Chemistry Letters_ 11 (21), 9215–9223 (2020). * [54] Thomas, M., Brehm, M., Fligg, R., Vöhringer, P. & Kirchner, B. Computing vibrational spectra from ab initio molecular dynamics. _Physical Chemistry Chemical Physics_ 15 (18), 6608–6622 (2013). * [55] Neugebauer, J., Reiher, M., Kind, C. & Hess, B. A. Quantum chemical calculation of vibrational spectra of large molecules - Raman and IR spectra for Buckminsterfullerene. _Journal of Computational Chemistry_ 23 (9), 895–910 (2002). * [56] Nemati, S., Henkel, C. & Anders, J. Coupling function from bath density of states. _Epl_ 139 (3) (2022). * [57] Carlson, C., Salzwedel, R., Selig, M., Knorr, A. & Hughes, S. Strong coupling regime and hybrid quasinormal modes from a single plasmonic resonator coupled to a transition metal dichalcogenide monolayer. _Physical Review B_ 104 (12), 1–13 (2021). * [58] Wilson, E. B., Decius, J. C., Cross, P. C. & Sundheim, B. R. Molecular Vibrations: The Theory of Infrared and Raman Vibrational Spectra. _Journal of The Electrochemical Society_ 102 (9), 235C (1955). * [59] Williams, S. O. & Imre, D. G. Raman Spectroscopy : Time-Dependent Pictures (6), 3363–3374 (1988) . * [60] Provazza, J., Tempelaar, R. & Coker, D. F. Analytic and numerical vibronic spectra from quasi-classical trajectory ensembles. _Journal of Chemical Physics_ 155 (1) (2021). * [61] Peng, Q., Yi, Y., Shuai, Z. & Shao, J. Excited state radiationless decay process with Duschinsky rotation effect: Formalism and implementation. _Journal of Chemical Physics_ 126 (11) (2007). * [62] George, J., Shalabney, A., Hutchison, J. A., Genet, C. & Ebbesen, T. W. Liquid-phase vibrational strong coupling. _Journal of Physical Chemistry Letters_ 6 (6), 1027–1031 (2015). * [63] Nagarajan, K., Thomas, A. & Ebbesen, T. W. Chemistry under Vibrational Strong Coupling. _Journal of the American Chemical Society_ 143 (41), 16877–16889 (2021). * [64] Szakács, Z. & Vauthey, E. Excited-State Symmetry Breaking and the Laporte Rule. _Journal of Physical Chemistry Letters_ 12 (16), 4067–4071 (2021). * [65] Case, W. B. Wigner functions and Weyl transforms for pedestrians. _American Journal of Physics_ 76 (10), 937–946 (2008). * [66] May, V. & Kühn, O. _Charge and Energy Transfer Dynamics in Molecular Systems_ (Wiley, 2011). * [67] Auffèves, A. _et al._ Controlling the dynamics of a coupled atom-cavity system by pure dephasing. _Physical Review B - Condensed Matter and Materials Physics_ 81 (24), 1–10 (2010).
# Analysis of UAV Radar and Communication Network Coexistence with Different Multiple Access Protocols Sung Joon Maeng, Jaehyun Park, , and İsmail Güvenç, This work has been supported in part by the NSF award CNS-1910153. This work has been also supported in part by National Research Foundation of Korea under the framework of international cooperation program (2022K2A9A2A06035926).S. J. Maeng, İ. Güvenç are with the Department of Electrical and Computer Engineering, North Carolina State University, Raleigh, NC 27606 USA (e-mail: <EMAIL_ADDRESS>iguvenc@ncsu.edu).J. Park with the Department of Electronic Engineering, Pukyong National University, Busan 608-737, South Korea (e-mail: jaehyun@pknu.ac.kr). ###### Abstract Unmanned aerial vehicles (UAVs) are expected to be used extensively in the future for various applications, either as user equipment (UEs) connected to a cellular wireless network, or as an infrastructure extension of an existing wireless network to serve other UEs. Next generation wireless networks will consider the use of UAVs for joint communication and radar and/or as dedicated radars for various sensing applications. Increasing number of UAVs will naturally result in larger number of communication and/or radar links that may cause interference to nearby networks, exacerbated further by the higher likelihood of line-of-sight signal propagation from UAVs even to distant receivers. With all these, it is critical to study network coexistence of UAV- mounted base stations (BSs) and radar transceivers. In this paper, using stochastic geometry, we derive closed-form expressions to characterize the performance of coexisting UAV radar and communication networks for spectrum overlay multiple access (SOMA) and time-division multiple access (TDMA). We evaluate successful ranging probability (SRP) and the transmission capacity (TC) and compare the performance of TDMA and SOMA. Our results show that SOMA can outperform TDMA on both SRP and TC when the node density of active UAV- radars is larger than the node density of UAV-comms. ###### Index Terms: Coexistence, guard zone, HPPP, multiple access, sensing and communication, stochastic geometry, UAV communication, UAV radar detection. ## I Introduction Recently, various different applications of cellular-connected unmanned aerial vehicles (UAVs) have been getting significant attention due to their cost- efficient deployment and controllable mobility. UAVs are utilized in many fields such as environmental monitoring and surveillance [1], public safety [2], video broadcasting [3], and delivery [4]. Moreover, UAV-mounted base stations (UAV-BS) and user equipment (UAV-UE), as well as UAV-mounted radars (UAV-radar) are commonly considered in sensing and communication applications, since the UAVs are available to quickly change the position to serve users at the outage area, and/or surveil/track the location of detected moving targets. Joint design of the radar and communication systems is considered as one of the key research areas for wireless networks beyond 5G systems which can benefit significantly from the use of autonomous UAVs. In the meanwhile, as the demand for using wider bandwidths has been increasing to support higher throughput and massive connectivity, the band of operation for broadband wireless networks has been moving to higher frequencies such as millimeter-wave (mmWave) and sub/THz bands that are also commonly used by radar systems. Some traditional radar bands, including certain bands below 6 GHz, are also being opened for shared use with communication networks due to the increasing congestion in the dedicated spectrum for cellular networks. All these developments call for rigorously studying the coexistence scenarios for radar and communication networks and coming up with strategies for effective spectrum sharing [5]. Stochastic geometry-based techniques are commonly used in the literature for obtaining closed-form expressions on the performance of wireless networks where transmit sources are randomly deployed in the spatial domain [6]. For example, the analysis of accumulated interference from multiple nodes following a homogeneous Poisson point process (HPPP) is useful to evaluate the capacity of wireless networks [6]. In this paper, we specifically investigate UAV radar sensing and communication network coexistence scenarios. In particular, we consider scenarios where radar transmission and data transmission are coordinated by two different multiple-access protocols: spectrum overlay multiple access (SOMA) and time-division multiple access (TDMA). In SOMA, radar sensing and data communication share the same spectrum so that the spectrum is overlapped. On the other hand, in TDMA, radar detection and communication are separated by time. We utilize stochastic geometry-based analysis where UAVs are randomly located in 3D space following a two-dimensional homogeneous Poisson point process (HPPP). We individually analyze the radar detection performance and the data communication performance using the successful ranging probability (SRP) and the transmission capacity (TC), respectively. Table I: Literature review for stochastic geometry based wireless network performance analysis. Ref. | Analysis objective | Application | Radar | Networks ---|---|---|---|--- [7] | TC with different spatial diversity techniques | Terrestrial | ✗ | Ad hoc [8] | Information and energy outage probability and area harvested energy | Terrestrial | ✗ | SWIPT in ad hoc networks [9] | Channel outage and packet loss probability | Terrestrial | ✗ | DL URLLC communications [10, 11] | SRP | Terrestrial | ✓ | Radar with road scenario [12] | SRP | Terrestrial | ✓ | Radar networks [13, 14] | Coverage probability with different antenna patterns and directivity | Terrestrial | ✗ | DL cellular [15] | Connection and secrecy probability | Terrestrial | ✗ | DL secure communication cellular [16] | Coverage probability with underlay and overlay protocols | UAV | ✗ | UAV-to-UAV and UL terrestrial [17] | Coverage probability and spectral efficiency | UAV | ✗ | Two-tiers cellular [18, 19] | Coverage probability and spectral efficiency | UAV | ✗ | DL communication [20] | Successful transmission probability, energy and SINR coverage | UAV | ✗ | DL SWIPT and UL communication [21] | Coverage probability | UAV | ✗ | UAV-aided DL and UL communication [22] | LoS probability | UAV | ✗ | BS-to-UAV link [23] | Coverage probability, motion energy, and flight time | UAV | ✗ | UAV path planning [24] | Connection, secrecy, and energy-information outage probability | UAV | ✗ | Secure communication in SWIPT This work | SRP and TC | UAV | ✓ | Radar and communication coexistence Contributions of this paper can be summarized as follows: * – We derive closed-form expressions for SRP and TC on the UAV radar and communication coexistence scenario where UAVs are placed following HPPP with a guard zone. * – We analyze the performance of SRP and TC in SOMA and TDMA, respectively. We also investigate behaviors of SRP and TC depending on the node density, radius of the guard zone, power splitting factor in SOMA, and time division factor in TDMA. * – We analytically compare TDMA and SOMA on SRP and TC and show that TDMA outperforms SOMA on SRP while SOMA is better than TDMA on TC in the general condition. Furthermore, we analyze the condition that SOMA can be superior to TDMA on both SRP and TC metrics. The rest of this paper is organized as follows. Section II presents the literature review. In Section III, we describe the UAV radar and communication network coexistence design. In Section IV, we provide the signal propagation model when UAVs are distributed by HPPP. In Section V, we derive the closed- form expressions of the SRP and the TC. In Section VI, we analyze SRP depending on the system parameters and the multiple access protocols. In Section VII, we analyze the TC depending on the system parameters and the multiple access protocols. In Section VIII, we compare the SRP and the TC performance of SOMA and TDMA. In Section IX, we show the simulation results to verify the analysis in previous sections, and Section X provides concluding remarks. ## II Literature Review The operation of UAVs on BSs and radar detectors have been investigated in the literature. A flying UAV-BS can maximize the capacity or minimize the outage of networks by optimizing UAV trajectory [25, 26]. In [27, 28], the trajectory and precoder of UAV-BS are optimized to maximize physical layer secrecy. In [29], a UAV-radar is used in measuring the depth of the snow on the sea. Human detection and classification by a UAV-radar have been studied in [30]. Target detection using radar imaging from UAV-radar has been investigated in [31]. In [32], the feasibility of a surveillance system using a UAV-radar has been explored. The study of coexistence networks has been explored in the literature. In [33], a beamforming approach has been studied to facilitate the coexistence between downlink (DL) multi-user-multiple-input-multiple-output (MU-MIMO) communication and MIMO radar system. In [34], the joint design of the radar and communication system for the coexistence of MIMO radar and MIMO communication has been studied. Moreover, UAV communication and radar sensing network coexistence that utilizes the spectrum for both purposes has been investigated for an efficient and flexible system design [35]. In [36], joint UAV communication and cooperative sensing network based on beam sharing scheme has been explored. Stochastic geometry based network analysis has been thoroughly investigated in the literature. TC is analyzed in ad hoc networks with different spatial diversity techniques where transmitting nodes are distributed by an HPPP [7], and this work is extended to the wireless information and power transfer (SWIPT)-based ad hoc networks in [8]. In [9], packet loss probability depending on the packet size, packet duration, and SINR are derived in downlink ultra-reliable and low-latency communications (URLLC) scenarios where distributed antenna ports are randomly placed following an HPPP. In [10, 11, 12], the effect of radar interference on the radar detection performance is analyzed and SRP is evaluated using stochastic geometry. More specifically, the geometric layout of vehicles on a road where the locations of vehicles on a certain lane are decided by unidimensional HPPP model is investigated in [10, 11]. In [11], radar cross-section (RCS) characteristics are modeled and analyzed using HPPPs for automotive radar network scenarios. In [13, 14], the effects of different directional antenna patterns, node densities, and antenna array sizes on coverage probability are studied for mmWave networks. The locations of BSs and the eavesdroppers are randomly distributed by independent HPPPs in [15], and closed-form expression of secrecy probability for secure communications is explored. Closed-form analysis of network performance using stochastic geometry techniques have also been studied in UAV networks in [16, 17, 18, 19, 20, 21, 22, 23, 24]. In [16], coexisting UAV-to-UAV links and uplink (UL) ground-BS to ground-user links are considered. Then, coverage of two different scenarios are studied, where the spectrum for each link is either reused, or it is allocated in a dedicated manner. The literature review with representative works related to stochastic geometry-based wireless network performance analysis is summarized in Table I. To the best of our knowledge, the study of radar networks based on stochastic geometry has been limited, and UAV communication and radar network coexistence scenario has not been investigated yet. Table II: Key symbols and notations used in this paper. Symbol | Definition ---|--- ${\lambda_{\rm r}}^{\prime}$ | Node density of UAV-radars ${\lambda_{\rm d}}^{\prime}$ | Node density of UAV-comms ${\bar{\lambda}}_{\rm r}$ | Active node density of UAV-radars ${\lambda_{\rm r}}$ | Effective node density of active UAV-radars in HPPP ${\lambda_{\rm d}}$ | Effective node density of UAV-comms in HPPP $\mathsf{h}_{\rm UAV}$ | UAVs height $r_{0}$ | Radius of guard zone $\phi$ | Power splitting factor $\tau$ | Time division factor $\delta$ | Duty circle $\mathsf{P}_{\rm Tx}$ | Transmit power $\mathsf{G}_{\rm t}$ | Tx antenna gain $\alpha$ | Path-loss exponent $\alpha_{\rm I}$ | Path-loss exponent from the interference $\mathsf{G}_{\rm r}$ | Receiver antenna gain $\mathsf{G}_{\rm rI}$ | Receiver antenna gain from the interference $\bar{\sigma}$ | Average RCS $\sigma$ | RCS $S_{\rm e}$ | Effective aperture of radar receiver $\mathsf{G}_{\rm p}$ | Processing gain of radar receiver $f_{\rm c}$ | Carrier frequency $R_{0}$ | Target distance $r_{i}$ | Distance from the interferer $c$ | Speed of light $h_{0}$ | Small-scale fading $h_{i}$ | Small-scale fading from the interference $\beta_{\rm th}$ | Target SINR threshold for outage probability $\gamma{\rm th}$ | Target SINR threshold for successful range probability $\beta_{0}$ | SINR of the received data signal $\gamma{0}$ | SINR of the received radar signal $N_{0}$ | Noise power ## III System Model We consider UAV networks where radar nodes and communication nodes coexist. Radar-mounted UAVs (UAV-radars) detect and track a target on the ground by transmitting radar signals and receiving the reflected signals from the target. On the other hand, UAVs that are equipped with a BS (UAV-comms) communicate with a ground user. We assume that UAV-radars and UAV-comms follow a two-dimensional HPPP independently where the node densities are ${\lambda_{\rm r}}^{\prime}$ and ${\lambda_{\rm d}}^{\prime}$ respectively. All UAVs fly at a fixed identical height $\mathsf{h}_{\rm UAV}$. Fig. 1 describes two different network representations of the radar detection scenario and the communication scenario in the HPPP model. The guard zone with radius $r_{0}$ is considered between two UAVs, or between a UAV and a user to protect them from potential strong interference. The distance between a UAV- radar and a target in the radar detection scenario and the distance between a UAV-comm and a served user in the communication scenario is $R_{0}$. UAV- radars are assigned to active UAV-radar by the random spectrum access with the duty cycle $\delta$, and the rest of UAV-radars remain inactive UAV-radars in the networks. (a) (b) Figure 1: UAV radar and communication network coexistence. A blue UAV is either: (a) a typical UAV-radar that detects a target using radar transmission; or (b) a serving UAV that communicates with a typical user. The green UAV-comms or the red (active) UAV-radars can interfere with radar detection or communication signals. The black (inactive) UAV-radars do not transmit any interference signals. To avoid strong interference, a guard zone is with a radius $r_{0}$ established between UAVs, and between a UAV and a user. Figure 2: Two different multiple access schemes: SOMA (left) and TDMA (right). In the radar and communication coexistence, UAV-radars and UAV-comms need to coordinate the time and spectrum resources for radar and data transmissions. We consider two different multiple access schemes: SOMA where radar signal and data signal share the same spectrum during transmission time, and TDMA where the time for radar and data transmission is scheduled at separate time slots. Fig. 2 illustrates different radar and data allocations depending on multiple access schemes. The power allocated to radar and communication is determined by power splitting factor $\phi$ for SOMA, and the time duration assigned to radar and communication is decided by time division factor $\tau$ for TDMA. Note that the interference behavior in this network coexistence is dependent on the multiple access. In this paper, we focus on the analysis and comparison of SOMA and TDMA on data transmission and radar detection. The key parameters are summarized in Table II. ## IV Signal Propagation Models in HPPP In this section, we describe signal and interference models in HPPP when SOMA and TDMA are adopted respectively. Throughout this paper, we denote SOMA and TDMA as $\rm s.o.$ and $\rm t.d.$ at the superscript. ### IV-A Radar and Data Signal Models #### IV-A1 SOMA We place a typical UAV-radar with the origin $(0,\leavevmode\nobreak\ 0,\leavevmode\nobreak\ \mathsf{h}_{\rm UAV})$ and distance from the target at $(x_{\rm t},y_{\rm t},0)$ is $R_{0}=\sqrt{x_{\rm t}^{2}+y_{\rm t}^{2}+(\mathsf{h}_{\rm UAV})^{2}}$ as in Fig. 1(a). We assume that the height of a UAV-radar $\mathsf{h}_{\rm UAV}$ is sufficiently high so that line-of- sight is secured to detect the target and the free-space path loss model can be considered [37]. The power of the received signal that is reflected back from the target can be expressed as [10], $\displaystyle\mathsf{P}_{\rm r}^{\rm s.o.}=\left(\frac{(1-\phi)\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}}{4\pi R_{0}^{\alpha}}\right)\left(\frac{\sigma S_{\rm e}}{4\pi R_{0}^{\alpha}}\right)\mathsf{G}_{\rm p},$ (1) where $\mathsf{P}_{\rm Tx}$, $\mathsf{G}_{\rm t}$, $\alpha$ indicate transmit power, Tx antenna gain, and path loss exponent, and $\sigma$, $S_{\rm e}$, $\mathsf{G}_{\rm p}$ denote radar cross-section (RCS) of the target, the effective aperture of radar receiver, and the processing gain. Swerling I model is considered for the RCS and the RCS of the target follows the exponential distribution, $\sigma\sim\frac{1}{\bar{\sigma}}e^{-\frac{\sigma}{\bar{\sigma}}}$ [11]. The effective area is given by $\displaystyle S_{\rm e}=\frac{\mathsf{G}_{\rm r}c^{2}}{4\pi f_{\rm c}^{2}},$ (2) where $c$, $f_{\rm c}$ denote the speed of light and the carrier frequency. Then, the power of the reflected back radar signal in (1) can be rewritten as $\displaystyle\mathsf{P}_{\rm r}^{\rm s.o.}=\frac{(1-\phi)\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}c^{2}\sigma}{(4\pi)^{3}f_{\rm c}^{2}R_{0}^{2\alpha}}.$ (3) In the communication scenario as in Fig. 1(b), a typcial user is located at the origin and it receives the signal from the serving UAV-comm at $(x_{\rm t},y_{\rm t},\mathsf{h}_{\rm UAV})$, which is at a distance of $R_{0}=\sqrt{x_{\rm t}^{2}+y_{\rm t}^{2}+(\mathsf{h}_{\rm UAV})^{2}}$ from the user. The power of the received signal of the user is given by $\displaystyle\mathsf{P}_{\rm d}^{\rm s.o.}=\frac{\phi\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}c^{2}}{(4\pi)^{2}f_{\rm c}^{2}R_{0}^{\alpha}}h_{0},$ (4) where $h_{0}\sim\exp(1)$ represents Rayleigh small-scale fading. Note that the allocated power of the radar and the data signals are split by the power splitting factor $\phi$ in SOMA as illustrated in Fig 2. #### IV-A2 TDMA The received signal power of the radar signal and the data signal in TDMA are expressed as $\displaystyle\mathsf{P}_{\rm r}^{\rm t.d.}=\frac{\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}c^{2}\sigma}{(4\pi)^{3}f_{\rm c}^{2}R_{0}^{2\alpha}},$ (5) $\displaystyle\mathsf{P}_{\rm d}^{\rm t.d.}=\frac{\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}c^{2}}{(4\pi)^{2}f_{\rm c}^{2}R_{0}^{\alpha}}h_{0}.$ (6) Since the radar detection and communication are separately conducted in different time slots in TDMA, transmit power is not adjusted as in SOMA. ### IV-B Effective Radar and Communication Node Densities The PPP in this network introduces the guard zone and it can be modeled by the Matérn hard-core point processes (MHCPP) type-II, which can be further approximated by the HPPP model. #### IV-B1 SOMA The approximated effective node density of the radar and the communication from MHCPP type-II can be written as [38] $\displaystyle\lambda_{\rm r}^{\rm s.o.}=\frac{1-e^{-\bar{\lambda}_{\rm r}^{\rm s.o.}\pi r_{0}^{2}}}{\pi r_{0}^{2}},\leavevmode\nobreak\ \lambda_{\rm d}^{\rm s.o.}=\frac{1-e^{-\lambda^{\prime}_{\rm d}\pi r_{0}^{2}}}{\pi r_{0}^{2}},$ (7) where ${\bar{\lambda}}_{\rm r}^{\rm s.o.}=\delta\lambda^{\prime}_{\rm r}$ is the active UAV-radar node density, and $\delta$ is the duty cycle. #### IV-B2 TDMA The effective node density of the radar and the communication from MHCPP type- II can be written as $\displaystyle\lambda_{\rm r}^{\rm t.d.}=\frac{1-e^{-\bar{\lambda}^{\rm t.d.}_{\rm r}\pi r_{0}^{2}}}{\pi r_{0}^{2}},\leavevmode\nobreak\ \lambda_{\rm d}^{\rm t.d.}=\frac{1-e^{-\lambda^{\prime}_{\rm d}\pi r_{0}^{2}}}{\pi r_{0}^{2}},$ (8) where ${\bar{\lambda}}_{\rm r}^{\rm t.d.}=\frac{\delta}{1-\tau}\lambda^{\prime}_{\rm r}$ is the active UAV-radar node density. Note that the duty cycle $\delta$ in SOMA increases to $\frac{\delta}{1-\tau}$ in TDMA as much as the reduced radar transmission time by $\tau$, since it is assumed that the total number of active UAV-radar nodes during the time period is the same for both SOMA and TDMA. Since the effective node density of UAV-comms for TDMA and SOMA is equal, we merge the notation of the node density as $\lambda_{\rm d}=\lambda_{\rm d}^{\rm s.o.}=\lambda_{\rm d}^{\rm t.d.}$. ### IV-C Interference Models In this subsection, we obtain the power of interference coming from nearby active UAV-radars and UAV-comms in the HPPP model. #### IV-C1 SOMA Since the radar detection and communication occupy the same spectrum band at the same time, the aggregated interference power from UAV-radars and UAV-comms can be expressed as $\displaystyle\mathsf{I}^{\rm s.o.}$ $\displaystyle=\underbrace{\sum_{r_{i}\in\Phi(\lambda_{\rm d})\backslash r_{0}}\frac{\phi\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm rI}c^{2}}{(4\pi)^{2}f_{\rm c}^{2}r_{i}^{\alpha_{\rm I}}}h_{i}}_{\text{interfence from nearby UAV-comms }(\mathsf{I}_{1})}$ $\displaystyle+\underbrace{\sum_{r_{j}\in\Phi(\lambda_{\rm r}^{\rm s.o.})\backslash r_{0}}\frac{(1-\phi)\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm rI}c^{2}}{(4\pi)^{2}f_{\rm c}^{2}r_{j}^{\alpha_{\rm I}}}h_{j}}_{\text{interfence from nearby UAV-radars }(\mathsf{I}_{2})},$ (9) where $r\in\Phi(\lambda)\backslash r_{0}$ means a two-dimensional HPPP with a density $\lambda$ and $r>r_{0}$, $\mathsf{G}_{\rm rI}$ and $\alpha_{\rm I}$ are Rx antenna gain and path-loss exponent from the interfering signals, respectively, $h_{i}$, $h_{j}$ represent small-scale fading from the interference, and $r_{i}$, $r_{j}$ are distance between a typical UAV-radar or user and an interferer. Since the radar detection and communication are carried out together in SOMA, the interference power at both the typical UAV- radar (Fig. 1(a)) and the typical information receiver (Fig. 1(b)) is the same: $\mathsf{I}^{\rm s.o.}=\mathsf{I}^{\rm s.o.}_{\rm d}=\mathsf{I}^{\rm s.o.}_{\rm r}$. #### IV-C2 TDMA The interfence comes only from the UAV-radars in the radar detection scenario. Likewise, the interference comes only from UAV-comms in communication scenario. Then the aggregated interference power for each case can be expressed as $\displaystyle\mathsf{I}^{\rm t.d.}_{\rm r}$ $\displaystyle=\sum_{r_{i}\in\Phi(\lambda_{\rm r}^{\rm t.d.})\backslash r_{0}}\frac{\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm rI}c^{2}}{(4\pi)^{2}f_{\rm c}^{2}r_{i}^{\alpha_{\rm I}}}h_{i},$ (10) $\displaystyle\mathsf{I}^{\rm t.d.}_{\rm d}$ $\displaystyle=\sum_{r_{i}\in\Phi(\lambda_{\rm d})\backslash r_{0}}\frac{\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm rI}c^{2}}{(4\pi)^{2}f_{\rm c}^{2}r_{i}^{\alpha_{\rm I}}}h_{i}.$ (11) ## V Performance Analysis of Radar Detection and Data Communication In this section, we discuss performance evaluation metrics for two different scenarios. Specifically, we derive the SRP in the radar detection scenario and the TC in the communication scenario. ### V-A Successful Ranging Probability SRP is the probability that a UAV-radar succeeds in detecting the target, which is decided by the signal-to-interference-plus-noise ratio (SINR). SRP is defined by $\displaystyle\mathsf{Pr_{s.r.}}(\gamma_{\rm th})=\mathsf{Pr}(\gamma_{0}>\gamma_{\rm th}),$ (12) where $\gamma_{0}=\frac{\mathsf{P}_{\rm r}}{\mathsf{I}_{\rm r}+\mathsf{N}_{0}}$ and $\mathsf{N}_{0}$ are SINR of the received radar signal and the noise power, respectively. In addition, $\gamma_{\rm th}$ denotes SINR threshold where the target is successfully detected. In what follows, we derive the closed-form expression of SRP in both SOMA and TDMA. #### V-A1 SOMA SRP in (12) can be derived from (3), (IV-C1) as follows: $\displaystyle\mathsf{Pr_{s.r.}^{\rm s.o.}}(\gamma_{\rm th})=\mathsf{Pr}\left(\frac{\mathsf{P}^{\rm s.o.}_{\rm r}}{\mathsf{I}^{\rm s.o.}+\mathsf{N}_{0}}>\gamma_{\rm th}\right)$ $\displaystyle=\mathsf{Pr}\left(\frac{(1-\phi)\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}c^{2}\sigma}{\mathsf{I}^{\rm s.o.}+\mathsf{N}_{0}}>(4\pi)^{3}f_{\rm c}^{2}R_{0}^{2\alpha}\gamma_{\rm th}\right)$ $\displaystyle\approx\mathsf{Pr}\left(\sigma>\frac{(4\pi)^{3}f_{\rm c}^{2}R_{0}^{2\alpha}\gamma_{\rm th}\mathsf{I}^{\rm s.o.}}{(1-\phi)\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}c^{2}}\right)$ $\displaystyle=\int_{0}^{\infty}\left\\{1-F_{\sigma}\left(\frac{(4\pi)^{3}f_{\rm c}^{2}R_{0}^{2\alpha}\gamma_{\rm th}y}{(1-\phi)\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}c^{2}}\right)\right\\}f_{\mathsf{I}^{\rm s.o.}}(y){\rm d}y$ $\displaystyle=\int_{0}^{\infty}e^{-\frac{(4\pi)^{3}f_{\rm c}^{2}R_{0}^{2\alpha}\gamma_{\rm th}y}{(1-\phi)\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}c^{2}\bar{\sigma}}}f_{\mathsf{I}^{\rm s.o.}}(y){\rm d}y$ $\displaystyle=\mathcal{L}_{\mathsf{I}^{\rm s.o.}}\left(\frac{(4\pi)^{3}f_{\rm c}^{2}R_{0}^{2\alpha}\gamma_{\rm th}}{(1-\phi)\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}c^{2}\bar{\sigma}}\right)$ (13) where the approximation comes from the interference limit regime assumption, $F_{\sigma}(X)=1-e^{-\frac{X}{\bar{\sigma}}}$ is the cumulative distribution function (CDF) of $\sigma$, and $f_{\mathsf{I}^{\rm s.o.}}(\rm x)$ is the probability density function (PDF) of $\mathsf{I}^{\rm s.o.}$, $\mathcal{L}_{\mathsf{I}^{\rm s.o.}}(z)$ indicates Laplace transform of the PDF of $\mathsf{I}^{\rm s.o.}$. $\mathcal{L}_{\mathsf{I}^{\rm s.o.}}(z)$ can be derived as follows. The interference term can be rewritten as $\mathsf{I}^{\rm s.o.}=\mathsf{I}_{1}+\mathsf{I}_{2}$ where $\mathsf{I}_{1}$, $\mathsf{I}_{2}$ are the first and the second terms in (IV-C1) respectively. Then, we can obtain [9] $\displaystyle\mathcal{L}_{\mathsf{I}_{1}}(z)$ $\displaystyle=\exp{\left\\{-2\pi\lambda_{\rm d}{\rm A}_{1}(z)\frac{(z\phi{\rm K}_{1})^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}},$ $\displaystyle\mathcal{L}_{\mathsf{I}_{2}}(z)$ $\displaystyle=\exp{\left\\{-2\pi\lambda_{\rm r}^{\rm s.o.}{\rm A}_{2}(z)\frac{(z(1-\phi){\rm K}_{1})^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}},$ (14) where ${\rm A}_{1}(z)={\rm B}\left(\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)-{\rm B}\left(\frac{1}{1+z\phi{\rm K}_{1}r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)$, ${\rm A}_{2}(z)={\rm B}\left(\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)-{\rm B}\left(\frac{1}{1+z(1-\phi){\rm K}_{1}r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)$, ${\rm K_{1}}=\frac{\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm rI}c^{2}}{(4\pi)^{2}f_{\rm c}^{2}}$, ${\rm B}(a,b)$ is the beta function, and ${\rm B}(x;a,b)=\int_{0}^{x}u^{a-1}(1-u)^{b-1}{\rm d}u$ is the incomplete beta fucntion. Then, we can derive $\displaystyle\mathcal{L}_{\mathsf{I}^{\rm s.o.}}(z)=\mathcal{L}_{\mathsf{I}_{1}}(z)\mathcal{L}_{\mathsf{I}_{2}}(z)=$ $\displaystyle\exp{\left\\{-2\pi\left(\phi^{\frac{2}{\alpha_{\rm I}}}\lambda_{\rm d}{\rm A}_{1}(z)+\left(1-\phi\right)^{\frac{2}{\alpha_{\rm I}}}\lambda_{\rm r}^{\rm s.o.}{\rm A}_{2}(z)\right)\frac{(z{\rm K}_{1})^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}}.$ (15) Then, the closed-form expression of SRP in SOMA can be obtained as (V-A1) at the top of the next page. $\displaystyle\mathsf{Pr_{s.r.}^{\rm s.o.}}(\gamma_{\rm th})=\exp\left\\{-2\pi\left(\left(\frac{\phi}{1-\phi}\right)^{\frac{2}{\alpha_{\rm I}}}\lambda_{\rm d}\underbrace{\left\\{{\rm B}\left(\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)-{\rm B}\left(\frac{1}{1+\left(\frac{\phi 4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{(1-\phi)\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)\right\\}}_{C_{1}}+\right.\right.$ $\displaystyle\left.\left.\lambda_{\rm r}^{\rm s.o.}\underbrace{\left\\{{\rm B}\left(\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)-{\rm B}\left(\frac{1}{1+\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)\right\\}}_{C_{2}}\right)\frac{\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}$ (16) $\displaystyle\mathsf{Pr_{s.r.}^{\rm t.d.}}(\gamma_{\rm th})=\exp\left\\{-2\pi\lambda_{\rm r}^{\rm t.d.}\underbrace{\left\\{{\rm B}\left(\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)-{\rm B}\left(\frac{1}{1+\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)\right\\}}_{C_{2}}\frac{\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}$ (17) $\displaystyle C^{\rm s.o.}={\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left\\{-2\pi\left(\lambda_{\rm d}\underbrace{\left\\{{\rm B}\left(\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)-{\rm B}\left(\frac{1}{1+\left(\frac{{\rm G}_{\rm rI}R_{0}^{\alpha}\beta_{\rm th}}{\mathsf{G}_{\rm r}}\right)r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)\right\\}}_{C_{3}}+\right.\right.$ $\displaystyle\left.\left.\left(\frac{1-\phi}{\phi}\right)^{\frac{2}{\alpha_{\rm I}}}\lambda_{\rm r}^{\rm s.o.}\underbrace{\left\\{{\rm B}\left(\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)-{\rm B}\left(\frac{1}{1+\left(\frac{(1-\phi){\rm G}_{\rm rI}R_{0}^{\alpha}\beta_{\rm th}}{\phi\mathsf{G}_{\rm r}}\right)r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)\right\\}}_{C_{4}}\right)\frac{\left(\frac{{\rm G}_{\rm rI}R_{0}^{\alpha}\beta_{\rm th}}{\mathsf{G}_{\rm r}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}$ (18) $\displaystyle C^{\rm t.d.}=\tau{\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left\\{-2\pi\lambda_{\rm d}\underbrace{\left\\{{\rm B}\left(\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)-{\rm B}\left(\frac{1}{1+\left(\frac{{\rm G}_{\rm rI}R_{0}^{\alpha}\beta_{\rm th}}{\mathsf{G}_{\rm r}}\right)r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)\right\\}}_{C_{3}}\frac{\left(\frac{{\rm G}_{\rm rI}R_{0}^{\alpha}\beta_{\rm th}}{\mathsf{G}_{\rm r}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}$ (19) #### V-A2 TDMA SRP can be derived from (5), (10) as follow: $\displaystyle\mathsf{Pr_{s.r.}^{\rm t.d.}}(\gamma_{\rm th})$ $\displaystyle=\mathsf{Pr}\left(\frac{\mathsf{P}^{\rm t.d.}_{\rm r}}{\mathsf{I}^{\rm t.d.}_{\rm r}+\mathsf{N}_{0}}>\gamma_{\rm th}\right),$ $\displaystyle=\mathcal{L}_{\mathsf{I}^{\rm t.d.}_{\rm r}}\left(\frac{(4\pi)^{3}f_{\rm c}^{2}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}c^{2}\bar{\sigma}}\right),$ (20) $\displaystyle\mathcal{L}_{\mathsf{I}^{\rm t.d.}_{\rm r}}(z)$ $\displaystyle=\exp{\left\\{-2\pi\lambda_{\rm r}^{\rm t.d.}{\rm A}_{3}(z)\frac{(z{\rm K}_{1})^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}},$ (21) where ${\rm A}_{3}(z)={\rm B}(\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}})-{\rm B}(\frac{1}{1+z{\rm K}_{1}r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}})$. Note that detailed mathematical steps are skipped, since many steps are similar to (V-A1), (V-A1). Then, the closed-form expression of SRP in TDMA can be expressed as (V-A2) at the top of the next page. ### V-B Transmission Capacity TC is defined by the achievable data rate given an outage constraint multiplied by the spatial density and the data transmission time duration [7, 8]. At first, the outage probability can be expressed as $\displaystyle\mathsf{Pr_{out}}(\beta_{\rm th})=\mathsf{Pr}(\beta_{0}<\beta_{\rm th}),$ (22) where $\beta_{0}=\frac{\mathsf{P}_{\rm d}}{\mathsf{I}_{\rm d}+\mathsf{N}_{0}}$ is SINR of the received data signal, and $\beta_{\rm th}$ is a target SINR. Then, TC is given as $\displaystyle C^{\rm s.o.}$ $\displaystyle={\lambda}_{\rm d}(1-\mathsf{Pr_{out}}(\beta_{\rm th}))\log(1+\beta_{\rm th}),$ (23) $\displaystyle C^{\rm t.d.}$ $\displaystyle=\tau{\lambda}_{\rm d}(1-\mathsf{Pr_{out}}(\beta_{\rm th}))\log(1+\beta_{\rm th}),$ (24) where $C^{\rm s.o.},\leavevmode\nobreak\ C^{\rm t.d.}$ denote transmission capacity of SOMA and TDMA, respectively. Next, we derive the closed-form expression of TC in both SOMA and TDMA. #### V-B1 SOMA Outage probability in (22) can be derived as $\displaystyle\mathsf{Pr_{out}^{s.o.}}(\beta_{\rm th})=1-\mathsf{Pr}\left(\frac{\mathsf{P}^{\rm s.o.}_{\rm d}}{\mathsf{I}^{\rm s.o.}+\mathsf{N}_{0}}>\beta_{\rm th}\right)$ $\displaystyle=1-\int_{0}^{\infty}\left\\{1-F_{h_{0}}\left(\frac{(4\pi)^{2}f^{2}R_{0}^{\alpha}\beta_{\rm th}y}{\phi\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}c^{2}}\right)\right\\}f_{\mathsf{I}^{\rm s.o.}}(y){\rm d}y$ $\displaystyle=1-\mathcal{L}_{\mathsf{I}^{\rm s.o.}}\left(\frac{(4\pi)^{2}f^{2}R_{0}^{\alpha}\beta_{\rm th}}{\phi\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}c^{2}}\right),$ (25) where $F_{h_{0}}(X)=1-e^{-X}$. From (V-A1), (23), and (V-B1), the closed-from expression of the TC in SOMA is given as (V-A1) at the top of the next page. #### V-B2 TDMA Outage probability in TDMA can be derived as $\displaystyle\mathsf{Pr_{out}^{t.d.}}(\beta_{\rm th})$ $\displaystyle=1-\mathsf{Pr}\left(\frac{\mathsf{P}^{\rm t.d.}_{\rm d}}{\mathsf{I}^{\rm t.d.}_{\rm d}+\mathsf{N}_{0}}>\beta_{\rm th}\right)$ $\displaystyle=1-\mathcal{L}_{\mathsf{I}^{\rm t.d.}_{\rm d}}\left(\frac{(4\pi)^{2}f^{2}R_{0}^{\alpha}\beta_{\rm th}}{\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm r}c^{2}}\right).$ (26) The Laplace transform of $\mathsf{I}^{\rm t.d.}_{\rm d}$ can be derived as $\displaystyle\mathcal{L}_{\mathsf{I}^{\rm t.d.}_{\rm d}}(z)$ $\displaystyle=\exp{\left\\{-2\pi\lambda_{\rm d}\int_{r_{0}}^{\infty}\mathbb{E}_{h}\left[1-e^{-z{\rm K}_{1}hr^{-\alpha_{\rm I}}}\right]r{\rm d}r\right\\}}$ $\displaystyle=\exp{\left\\{-2\pi\lambda_{\rm d}{\rm A}_{3}(z)\frac{(z{\rm K}_{1})^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}}.$ (27) From (23), (V-B2), (V-B2), the closed-form expression of the TC in TDMA can be derived as (19) at the top of the page. ## VI Network Design Strategy for Successful Ranging Probability In this section, we discuss how network parameters such as node densities, radius of guard zone, power splitting factor, and time division factor are determined from the analysis in Section V for a given SRP constraint. ### VI-A Node Densities The density of the node in the networks affects the power of the interference signal. Specifically, as the UAV-radar node density $\lambda_{\rm r}^{\prime}$ increases, the interference power at the typical UAV-radar increases and therefore, the SINR of the received radar signal decreases, which results in the lower SRP. In SOMA, SINR is also affected by the UAV-comm node density $\lambda_{\rm d}^{\prime}$ due to the simultaneous transmission of data and radar signals. Therefore, one can be interested in finding the maximum node density given a target SRP ($\mathsf{\bar{Pr}_{s.r.}}$) and SINR threshold $\gamma_{\rm th}$ of the SRP. #### VI-A1 SOMA When SINR threshold $\gamma_{\rm th}$ and the target SRP are given, we can rearrange (V-A1) such that only the terms that are related to the node densities $\lambda_{\rm d}$ and $\lambda_{\rm r}^{\rm s.o.}$ are placed to the left side of the equation. Then, we obtain inequality as follows: $\displaystyle\left(\frac{\phi}{1-\phi}\right)^{\frac{2}{\alpha_{\rm I}}}C_{1}\lambda_{\rm d}+C_{2}\lambda_{\rm r}^{\rm s.o.}\leq\frac{-\log\mathsf{\bar{Pr}_{s.r.}^{\rm s.o.}}\alpha_{\rm I}}{2\pi\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)^{\frac{2}{\alpha_{\rm I}}}},$ (28) where $C_{1}$ and $C_{2}$ are indicated in (V-A1). If we assume a condition $\lambda_{\rm d}=\lambda_{\rm r}^{\rm s.o.}$, the above inequality can be rewritten as $\displaystyle\lambda_{\rm d}=\lambda_{\rm r}^{\rm s.o.}\leq\frac{-\log\mathsf{\bar{Pr}_{s.r.}^{\rm s.o.}}\alpha_{\rm I}}{2\pi\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)^{\frac{2}{\alpha_{\rm I}}}\left(\left(\frac{\phi}{1-\phi}\right)^{\frac{2}{\alpha_{\rm I}}}C_{1}+C_{2}\right)}.$ (29) The maximum node densities $\lambda_{\rm d}^{\star}$ and $\lambda_{\rm r}^{\rm s.o.\star}$ can be obtained when (29) goes to equality. Note that this analysis can be easily extended to the condition that $\lambda_{\rm d}$ and $\lambda_{\rm r}^{\rm s.o.}$ are given by the different ratio ($\lambda_{\rm d}\propto\lambda_{\rm r}^{\rm s.o.}$) to find the maximum node densities. #### VI-A2 TDMA In the same manner of obtaining (29) for SOMA, the maximum node densities of the UAV-radar $\lambda_{\rm r}^{\rm t.d.\star}$ with the given target SRP and SINR threshold can be expressed from (V-A2) as $\displaystyle\lambda_{\rm r}^{\rm t.d.\star}=\frac{-\log\mathsf{\bar{Pr}_{s.r.}^{\rm s.o.}}\alpha_{\rm I}}{2\pi\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)^{\frac{2}{\alpha_{\rm I}}}C_{2}}.$ (30) ### VI-B Radius of Guard Zone Guard zone constrains the minimum distance between nodes to avoid strong interference. As the minimum distance increases, the power of the interference decreases. This implies that SRP is reduced as radius of guard zone $r_{0}$ increases. When we design networks with target SRP ($\mathsf{\bar{Pr}_{s.r.}}$) and the SINR threshold $\gamma_{\rm th}$, the minimum radius of the guard zone $r_{0}$ that satisfies the target performance can be obtain by solving (V-A1) in SOMA and (V-A2) in TDMA for $r_{0}$. ### VI-C Power Splitting Factor $\phi$ in SOMA Power splitting factor $\phi$ determines the transmit power ratio between UAV- comms and UAV-radars in SOMA where the radar signal power proportionally decreases as $\phi$ increases. From the closed-form expression of the SRP in (V-A1), the terms that are affected by $\phi$ are $\left(\frac{\phi}{1-\phi}\right)^{\frac{2}{\alpha_{\rm I}}}$ and ${\rm B}\left(\frac{1}{1+\left(\frac{\phi 4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{(1-\phi)\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)$. Since $\frac{\phi}{1-\phi}$ and incomplete beta function are monotonic increasing functions, it is easily proved that SRP is decreasing function with respect to the power splitting factor $\phi$. This can be intuitively interpreted as higher transmit power of UAV-comms increasing the power of the interference signal. ###### Proposition 1 When $0\leq\phi<0.5$, the impact of the UAV-comm node density $\lambda_{\rm d}$ on SRP is less than the UAV-radar node density $\lambda_{\rm r}^{\rm s.o.}$. When $\phi=0.5$, the impact of the communication node density $\lambda_{\rm d}$ on SRP is equal to the radar node density $\lambda_{\rm r}^{\rm s.o.}$, while when $0.5<\phi\leq 1$, the impact of the communication node density $\lambda_{\rm d}$ on SRP is greater than the radar node density $\lambda_{\rm r}^{\rm s.o.}$. ###### Proof: From (V-A1), we can observe that varying $\phi$ only affects the node density of UAV-comm $\lambda_{\rm d}$ term, not the node density of UAV-radar $\lambda_{\rm r}^{\rm s.o.}$ term. Then, when $\phi=0.5$, $\frac{\phi}{1-\phi}$ becomes 1, which leads to the result that the impact of $\lambda_{\rm d}$ becomes the same as the impact of $\lambda_{\rm r}^{\rm s.o.}$. On the other hand, when $\phi$ is greater than 0.5, $\frac{\phi}{1-\phi}$ becomes greater than 1 as well, which makes the multiplying term by $\lambda_{\rm d}$ becomes greater than the multiplying term by $\lambda_{\rm r}^{\rm s.o.}$. In the same way, when $\phi$ is less than 0.5, the multiplying term by $\lambda_{\rm d}$ becomes less than the multiplying term by $\lambda_{\rm r}^{\rm s.o.}$. ∎ Propostion 1 implies that SRP is affected by the ratio between the node density of UAV-comm ($\lambda_{\rm d}$) and the UAV-radar ($\lambda_{\rm r}^{\rm s.o.}$) and when $\phi$ is given, a different ratio of UAV-comm and UAV-radar node density can improve SRP, which is observed in Fig. 6 of Section IX. ### VI-D Time Division Factor $\tau$ in TDMA As we mention in Section IV-B, the increase in $\tau$ reduces radar transmission time and increase the duty cycle, which results in higher node density of the active UAV-radar $\bar{\lambda}^{\rm t.d.}_{\rm r}$. The effective UAV-radar node density $\lambda_{\rm r}^{\rm t.d.}$ in the HPPP approximation is proportionally increased by $\bar{\lambda}^{\rm t.d.}_{\rm r}$ in (8). ## VII Network Design Strategy for Transmission Capacity In this section, we analyze TC depending on network design parameters. We find the node densities that maximize the TC and we investigate the impact of the radius of guard zone. We also investigate the effect of the power splitting factor and the time division factor on the TC. ### VII-A Node Densities As the node density of the UAV-comm $\lambda_{\rm d}$ increases, SINR is decreased by the larger number of interferers but the higher node density can increase the capacity of the unit area. Because of this trade-off, we can find the maximum node density $\lambda_{\rm d}$ that maximizes TC. #### VII-A1 SOMA When target SINR $\beta_{\rm th}$, the UAV-radar node density $\lambda_{\rm r}^{\rm s.o.}$, radius of guard zone $r_{0}$, and the power splitting factor $\phi$ are given, we can find the $\lambda_{\rm d}$ that maximizes the TC from (V-A1). The term in (V-A1) that is affected by $\lambda_{\rm d}$ are written as $\displaystyle D_{1}={\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left(-2\pi{\lambda}_{\rm d}C_{3}\frac{\left(\frac{{\rm G}_{\rm rI}R_{0}^{\alpha}\beta_{\rm th}}{\mathsf{G}_{\rm r}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right),$ (31) where $C_{3}$ is indicated in (V-A1). From (31), the first and the second derivative of transmission capacity with respective to ${\lambda}_{\rm d}$ can be expressed as $\displaystyle(C^{\rm s.o.})^{{}^{\prime}}$ $\displaystyle=\log(1+\beta_{\rm th})\exp\left(-2\pi{\lambda}_{\rm d}C_{3}^{\prime}\right)\left(1-2\pi{\lambda}_{\rm d}C_{3}^{\prime}\right),$ (32) $\displaystyle(C^{\rm s.o.})^{{}^{\prime\prime}}$ $\displaystyle=4\pi C_{3}^{\prime}\log(1+\beta_{\rm th})\exp\left(-2\pi{\lambda}_{\rm d}C_{3}^{\prime}\right)\left(\pi{\lambda}_{\rm d}C_{3}^{\prime}-1\right),$ (33) where $C_{3}^{\prime}=C_{3}\frac{\left(\frac{{\rm G}_{\rm rI}R_{0}^{\alpha}\beta_{\rm th}}{\mathsf{G}_{\rm r}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}$. Then, transmission capacity is maximized at $\displaystyle{\lambda}_{\rm d}^{\star}=\frac{1}{2\pi C_{3}^{\prime}}\quad\quad\text{(SOMA)}.$ (34) In SOMA, the node density of UAV-radar $\lambda_{\rm r}^{\rm s.o.}$ also increases the power of interference, which reduces the TC. In $\eqref{eq:tc_so_ana}$, the terms that include the radar node density $\lambda_{\rm r}^{\rm s.o.}$ are given as $\displaystyle D_{2}={\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left(-2\pi\left(\frac{1-\phi}{\phi}\right)^{\frac{2}{\alpha_{\rm I}}}{\lambda}_{\rm r}^{\rm s.o.}C_{4}^{\prime}\right),$ (35) where $C_{4}^{\prime}=C_{4}\frac{\left(\frac{{\rm G}_{\rm rI}R_{0}^{\alpha}\beta_{\rm th}}{\mathsf{G}_{\rm r}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}$. From the above equation, it can be found that TC is a decreasing function of the ${\lambda}_{\rm r}^{\rm s.o.}$. #### VII-A2 TDMA Similarly to SOMA, we can optimize the UAV-comm node density $\lambda_{\rm d}$ in TDMA. From $\eqref{eq:tc_td_ana}$, the TC can be rewritten as $\displaystyle C^{\rm t.d.}=\tau{\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left(-2\pi{\lambda}_{\rm d}C_{3}^{\prime}\right).$ (36) Then, the optimal UAV-comm node density that maximizes TC can be derived as $\displaystyle{\lambda}_{\rm d}^{\star}=\frac{1}{2\pi C_{3}^{\prime}}\quad\quad\text{(TDMA)}.$ (37) In addition, the TC in TDMA is not affected by $\lambda_{\rm r}^{\rm t.d.}$. ###### Remark 1 From the above analysis, TC is maximized at ${\lambda}_{\rm d}^{\star}=\frac{1}{2\pi C_{3}^{\prime}}$ for both SOMA and TDMA. On the other hand, the TC in SOMA decreases as $\lambda_{\rm r}^{\rm s.o.}$ increases, while TC in TDMA is independent of $\lambda_{\rm r}^{\rm t.d.}$. ### VII-B Radius of Guard Zone Guard zone improves SINR and it reduces the effective node density $\lambda_{\rm d}$ from (7). Therefore, as radius of guard zone, $r_{0}$, increases TC would be either improved by higher SINR or degraded by the lower node density. Since it is mathematically intractable to obtain the first and the second derivatives of TC with respect to $r_{0}$ in (V-A1) and (19), we observe the effect of $r_{0}$ by simulations. From simulation results in Fig. 3(b), it is observed that the maximum TC decreases as the $r_{0}$ increases from 5 m to 25 m, which implies that the TC is a decreasing function of $r_{0}$ in a typical parameter setup. ### VII-C Power Splitting Factor $\phi$ and Time Division Factor $\tau$ In SOMA, TC is improved as $\phi$ increases since the transmit power of UAV- comm becomes higher, which improves SINR. The terms in (V-A1) that are affected by $\phi$ are $\left(\frac{1-\phi}{\phi}\right)^{\frac{2}{\alpha_{\rm I}}}$ and ${\rm B}\left(\frac{1}{1+\left(\frac{(1-\phi){\rm G}_{\rm rI}R_{0}^{\alpha}\beta_{\rm th}}{\phi\mathsf{G}_{\rm r}}\right)r_{0}^{-\alpha_{\rm I}}};\frac{2}{\alpha_{\rm I}},1-\frac{2}{\alpha_{\rm I}}\right)$. Since $\frac{1-\phi}{\phi}$ is a decreasing function and an incomplete beta function is a monotonic increasing function, it is easily proved that TC is an increasing function in terms of $\phi$. In TDMA, $\tau$ decides the time duration of the data transmission, and larger $\tau$ increases TC. From (19), it is observed that TC is linearly increasing with respect to $\tau$. ## VIII Performance Comparison of SOMA and TDMA We compare the performance of SRP and transmission capacity between two different multiple access strategies to give intuition in the design of UAV radar sensing and communication network coexistence. We consider two different scenarios when $\phi=\tau=0.5$: case 1 and case 2. In case 1, we analyze the condition where the node density of UAV-comms and active UAV-radars is equal, and we compare SOMA with TDMA by SRP and TC. In case 2, we analyze the condition that the node density of UAV-radars is greater than that of UAV- comms, and we find the condition that both SRP and TC of SOMA are higher than those of TDMA. ### VIII-A Case 1: $\lambda_{\rm d}^{\prime}=\bar{\lambda}_{\rm r}\neq 0$ We first analyze a special case that $\lambda_{\rm d}^{\prime}=\bar{\lambda}_{\rm r}\neq 0$ and $\phi=\tau=0.5$ where the active UAV-radar and the UAV-comm node density are equal and the resources allocation of the data transmission and the radar detection are the same. In this condition, SRP of SOMA and TDMA can be rewritten from (V-A1) and (V-A2) as $\displaystyle\mathsf{Pr_{s.r.}^{\rm s.o.}}(\gamma_{\rm th})$ $\displaystyle=\exp\left\\{-4\pi\lambda_{\rm r}^{\rm s.o.}C_{2}\frac{\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\},$ $\displaystyle\mathsf{Pr_{s.r.}^{\rm t.d.}}(\gamma_{\rm th})$ $\displaystyle=\exp\left\\{-2\pi\lambda_{\rm r}^{\rm t.d.}C_{2}\frac{\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}.$ (38) Then, we can derive the following proposition. ###### Proposition 2 In case 1, SRP of TDMA is always greater than SOMA: $\mathsf{Pr_{s.r.}^{\rm s.o.}}(\gamma_{\rm th})<\mathsf{Pr_{s.r.}^{\rm t.d.}}(\gamma_{\rm th})$. ###### Proof: From (VIII-A), proposition 2 is proved if $2\lambda_{\rm r}^{\rm s.o.}>\lambda_{\rm r}^{\rm t.d.}$ holds. From (7), (8), the statement can be derived as $\displaystyle 2-2e^{-\bar{\lambda}_{\rm r}^{\rm s.o.}\pi r_{0}^{2}}>1-e^{-\bar{\lambda}_{\rm r}^{\rm t.d.}\pi r_{0}^{2}}$ $\displaystyle\to$ $\displaystyle 2-2e^{-\lambda_{\rm r}^{\prime}\delta\pi r_{0}^{2}}>1-e^{-2\lambda_{\rm r}^{\prime}\delta\pi r_{0}^{2}}$ $\displaystyle\to$ $\displaystyle\left(e^{-\lambda_{\rm r}^{\prime}\delta\pi r_{0}^{2}}\right)^{2}-2e^{-\lambda_{\rm r}^{\prime}\delta\pi r_{0}^{2}}+1>0$ $\displaystyle\to$ $\displaystyle\left(e^{-\lambda_{\rm r}^{\prime}\delta\pi r_{0}^{2}}-1\right)^{2}>0.$ (39) ∎ Next, in this special case, TC can be rewritten from (V-A1), (19) as $\displaystyle C^{\rm s.o.}$ $\displaystyle={\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left\\{-4\pi\lambda_{\rm d}C_{3}^{\prime}\right\\},$ $\displaystyle C^{\rm t.d.}$ $\displaystyle=\frac{1}{2}{\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left\\{-2\pi\lambda_{\rm d}C_{3}^{\prime}\right\\}.$ (40) Then, we can derive the following proposition. ###### Proposition 3 In case 1, transmission capacity of SOMA is greater than TDMA, when outage probability $\mathsf{Pr_{out}^{t.d.}}(\beta_{\rm th})<\frac{1}{2}$, $\mathsf{Pr_{out}^{s.o.}}(\beta_{\rm th})<\frac{3}{4}$. ###### Proof: From (VIII-A), $C^{\rm s.o.}>C^{\rm t.d.}$, if the following inequality holds: $\displaystyle{\lambda}_{\rm d}\log(1+\beta_{\rm th})\left(\exp\left\\{-2\pi\lambda_{\rm d}C_{3}^{\prime}\right\\}\right)^{2}$ $\displaystyle>\frac{1}{2}{\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left\\{-2\pi\lambda_{\rm d}C_{3}^{\prime}\right\\}$ $\displaystyle\to$ $\displaystyle\quad\quad\exp\left\\{-2\pi\lambda_{\rm d}C_{3}^{\prime}\right\\}>\frac{1}{2},$ $\displaystyle\to$ $\displaystyle\quad\quad 1-\mathsf{Pr_{out}^{t.d.}}(\beta_{\rm th})>\frac{1}{2},$ $\displaystyle\to$ $\displaystyle\quad\quad\mathsf{Pr_{out}^{t.d.}}(\beta_{\rm th})<\frac{1}{2},$ $\displaystyle\to$ $\displaystyle\quad\quad\mathsf{Pr_{out}^{s.o.}}(\beta_{\rm th})=1-\left(\exp\left\\{-2\pi\lambda_{\rm d}C_{3}^{\prime}\right\\}\right)^{2}<\frac{3}{4}.$ (41) ∎ Note that the condition that outage probability is greater than $\frac{3}{4}$ is a generally desirable condition. Therefore, in case 1 ($\lambda_{\rm d}^{\prime}=\bar{\lambda}_{\rm r}\neq 0$ and $\phi=\tau=0.5$), TDMA outperforms SOMA for SRP, but SOMA is better than TDMA for TC. ### VIII-B Case 2: $\lambda_{\rm d}^{\prime}<\bar{\lambda}_{\rm r}$ We can also analyze another special case where $\lambda_{\rm d}^{\prime}<\bar{\lambda}_{\rm r}$, $\phi=\tau=0.5$, and an additional condition that the UAV-radar node density $\lambda_{\rm r}^{\prime}$ is sufficiently small. Then, the effective UAV-radar node densities $\lambda_{\rm r}^{\rm s.o.}$, $\lambda_{\rm r}^{\rm t.d.}$ in (7), (8) can be approximated by the first order Taylor expansion at $\lambda_{\rm r}^{\prime}=0$ as $\lambda_{\rm r}^{\rm s.o.}\approx\delta\lambda_{\rm r}^{\prime}$ and $\lambda_{\rm r}^{\rm t.d.}\approx 2\delta\lambda_{\rm r}^{\prime}$. In this condition, we can have the following proposition. ###### Proposition 4 In case 2 where $\lambda_{\rm d}^{\prime}<\bar{\lambda}_{\rm r}$, the radar node density $\lambda_{\rm r}^{\prime}$ is sufficient small, and $\phi=\tau=0.5$, SRP of SOMA is greater than TDMA. ###### Proof: From (V-A1) and (V-A2), we can obtain SRP in case 2 as follows: $\displaystyle\mathsf{Pr_{s.r.}^{\rm s.o.}}(\gamma_{\rm th})$ $\displaystyle=\exp\left\\{-2\pi(\lambda_{\rm d}+\delta\lambda_{\rm r}^{\prime})C_{2}\frac{\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\},$ $\displaystyle\mathsf{Pr_{s.r.}^{\rm t.d.}}(\gamma_{\rm th})$ $\displaystyle=\exp\left\\{-2\pi(2\delta\lambda_{\rm r}^{\prime})C_{2}\frac{\left(\frac{4\pi{\rm G}_{\rm rI}R_{0}^{2\alpha}\gamma_{\rm th}}{\mathsf{G}_{\rm r}\mathsf{G}_{\rm p}\bar{\sigma}}\right)^{\frac{2}{\alpha_{\rm I}}}}{\alpha_{\rm I}}\right\\}.$ (42) Then, we can easily prove that $\mathsf{Pr_{s.r.}^{\rm s.o.}}>\mathsf{Pr_{s.r.}^{\rm t.d.}}(\gamma_{\rm th})$ if $\lambda_{\rm d}^{\prime}<\delta\lambda_{\rm r}^{\prime}=\bar{\lambda}_{\rm r}$. ∎ Next, we can also obtain the following proposition regarding TC. ###### Proposition 5 In the case that $\phi=\tau=0.5$, TC of SOMA is greater than TDMA, when $\mathsf{Pr}(\beta_{1}<\beta_{\rm th})<\frac{1}{2}$, where $\beta_{1}=\frac{\mathsf{P}^{\rm s.o.}_{\rm r}}{\mathsf{I}_{1}}$ denotes SIR in SOMA considering the interference only comes from the active UAV-radar nodes ( $\mathsf{I}_{1}$ in (IV-C1)). ###### Proof: In the case that $\phi=\tau=0.5$, transmission capacity can be rewritten from (V-A1), (19) as $\displaystyle C^{\rm s.o.}$ $\displaystyle={\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left\\{-2\pi\lambda_{\rm d}C_{3}^{\prime}\right\\}\exp\left\\{-2\pi\lambda_{\rm r}^{\rm s.o.}C_{3}^{\prime}\right\\},$ $\displaystyle C^{\rm t.d.}$ $\displaystyle=\frac{1}{2}{\lambda}_{\rm d}\log(1+\beta_{\rm th})\exp\left\\{-2\pi\lambda_{\rm d}C_{3}^{\prime}\right\\}.$ (43) Then, $C^{\rm s.o.}>C^{\rm t.d.}$ holds, when $\displaystyle 1-\exp\left\\{-2\pi\lambda_{\rm r}^{\rm s.o.}C_{3}^{\prime}\right\\}<\frac{1}{2}$ $\displaystyle\to\leavevmode\nobreak\ $ $\displaystyle\mathsf{Pr}(\beta_{1}<\beta_{\rm th})<\frac{1}{2},$ (44) where $\mathsf{I}_{1}=\sum_{r_{j}\in\Phi(\lambda_{\rm r}^{\rm s.o.})\backslash r_{0}}\frac{\frac{1}{2}\mathsf{P}_{\rm Tx}\mathsf{G}_{\rm t}\mathsf{G}_{\rm rI}c^{2}}{(4\pi)^{2}f_{\rm c}^{2}r_{j}^{\alpha_{\rm I}}}h_{j}$. ∎ ###### Remark 2 From Proposition 4 and Proposition 5, SOMA can outperform TDMA in both SRP and TC, if the conditions in Proposition 4 and Proposition 5 are satisfied. This implies that the active UAV-radar node density is greater than the UAV-comm node density ($\lambda_{\rm d}^{\prime}<\bar{\lambda}_{\rm r}$) while the outage probability considering the interference only from the active UAV- radars is less than 0.5. Moreover, when the first condition holds, the second condition is generally desirable since the target outage probability is mostly less than 0.5 and the interference coming from the UAV-comms is smaller than the active UAV-radars. Table III: Parameter settings for UAV radar and communication network coexistence analysis. Parameter | Value ---|--- Transmit power ($\mathsf{P}_{\mathsf{Tx}}$) | $20$ dBm Transmitter antenna gain ($\mathsf{G}_{\rm t}$) | 10 dBi Receiver antenna gain ($\mathsf{G}_{\rm r}$) | 10 dBi Receiver antenna gain from the interference ($\mathsf{G}_{\rm rI}$) | -10 dBi Target distance ($R_{0}$) | 50 m Average RCS ($\bar{\sigma}$) | 30 dBsm Path-loss exponent ($\alpha$) | 2.0 Path-loss exponent from the interference ($\alpha_{\rm I}$) | 2.5 Processing gain ($\mathsf{G}_{\rm p}$) | 10 dBi Duty cycle ($\delta$) | 0.1 Carrier frequency ($f_{\rm c}$) | $35$ GHz (a) (b) Figure 3: SRP and TC depending on SINR threshold and radius of guard zone ($r_{0}$) where $\phi=0.5$, $\tau=0.5$, $\lambda^{\prime}_{\rm d}=0.01$, $\lambda^{\prime}_{\rm r}=0.1$. In case 1, TDMA outperforms SOMA on SRP while SOMA is superior to TDMA on TC. ## IX Simulation Results In this section, we evaluate the performance of UAV radar and communication network coexistence based on simulation and analysis. SRP and TC with SOMA and TDMA are presented with the change of the different parameters. We consider 35 GHz carrier frequency for mmWave communication and Ka-band radar. The key parameters are listed in Table III. ### IX-A SRP and TC Dependence SINR Threshold and Radius of Guard Zone (a) (b) Figure 4: The SRP and the TC depending on SINR threshold where $\lambda^{\prime}_{\rm d}=0.00025$, $\lambda^{\prime}_{\rm r}=0.005$, $r_{0}=5$ m, $\phi=0.5$, $\tau=0.5$. In case 2, SOMA outperforms TDMA on both the SRP and the TC. In this subsection, we compare SOMA and TDMA by SRP and TC depending on radius of guard zone and SINR threshold. In Fig. 3, we show SRP and TC of both SOMA and TDMA with $\lambda^{\prime}_{\rm d}$, $\lambda^{\prime}_{\rm r}$ and $\phi$, $\tau$ by case 1 in Section VIII-A. As we discuss in Proposition 2 and Proposition 3, TDMA outperforms SOMA in SRP while SOMA is superior to TDMA in the TC. We also observe that as radius of guard zone $r_{0}$ increases, SRP improves but TC degrades, which is matched to the analysis in Section VI-B and Section VII-B. Fig. 4 shows SRP and TC with a system configuration in case 2 in Section VIII-B. It is observed that both SRP and TC could be better in SOMA if we consider case 2, which is mentioned in Remark 2. Note that in a general system parameter setting, we obtain SRP and the TC performance of case 1. ### IX-B SRP and TC Dependence Power Splitting Factor and Time Division Factor (a) (b) Figure 5: Change of SRP and TC as $\phi$ in SOMA and $\tau$ in TDMA increases where $\lambda^{\prime}_{\rm d}=0.01$, $\lambda^{\prime}_{\rm r}=0.1$, $\beta_{\rm th}=-5$ dB, $\gamma_{\rm th}=-10$ dB. Figure 6: Change of SRP as $\phi$ increases with different ratio of the node density of the UAV-radar and the UAV-comm where $\gamma_{\rm th}=-10$ dB, which is analyzed in Propostion 1. In this subsection, we evaluate SRP and TC depending on $\phi$ in SOMA and $\tau$ in TDMA. Fig. 5(a) shows that as $\phi$ increases TC improves but SRP decreases as we discuss in Section VI-C and in Section VII-C, which represents the impact of the different power ratio between the radar signal and the data signal on SRP and TC. In, Fig. 5(b), it is observed that TC increases as $\tau$ becomes large. On the other hand, the SRP slowly decreases as $\tau$ increases when we compare it with $\phi$ in SOMA in Fig. 5(a). Fig. 6 show the effect of the different radio of the active UAV-radar node density $\bar{\lambda}_{\rm r}^{\rm s.o.}$ and the UAV-comm node density $\lambda_{\rm d}^{\prime}$ on the SRP in SOMA. It is observed that when $0.1<\phi<0.5$, higher UAV-comm node density achieves higher SRP while when $0.5<\phi<1$, higher active UAV-radar node density achieves higher SRP, which can be interpreted by Proposition 1. ### IX-C SRP and TC Dependence Node Density of UAV-radar and UAV-comm (a) (b) Figure 7: Change of SRP and TC depending on the UAV-radar node density where $\lambda^{\prime}_{\rm d}=0.01$, $r_{0}=5$ m, $\beta_{\rm th}=0$ dB, $\gamma_{\rm th}=-10$ dB, $\phi=0.5$, $\tau=0.5$. (a) (b) Figure 8: Change of SRP and TC depending on the UAV-comm node density where $\lambda^{\prime}_{\rm r}=0.01$, $r_{0}=5$ m, $\beta_{\rm th}=0$ dB, $\gamma_{\rm th}=-10$ dB, $\phi=0.5$, $\tau=0.5$. In this subsection, we simulate the dependence of SRP and TC on the node density of the UAV-radar ($\lambda_{\rm r}^{\prime}$) and the UAV-comm ($\lambda_{\rm d}^{\prime}$). As we discuss in Section VI-A and Section VII-A, Fig. 7(a) shows that SRP is a decreasing function of $\lambda_{\rm r}^{\prime}$ for both SOMA and TDMA. In addition, in Fig. 7(b), it is observed that TC decreases as $\lambda_{\rm r}^{\prime}$ increases in SOMA, however, the TC is not affected by $\lambda_{\rm r}^{\prime}$ in TDMA. In Fig. 8(a), we observe that SRP is a decreasing function of $\lambda_{\rm d}^{\prime}$ in SOMA, while SRP is not affected by $\lambda_{\rm r}^{\prime}$ in TDMA. Fig. 8 shows that TC is maximized at ${\lambda}_{\rm d}^{\prime\star}=0.0115$ for both SOMA and TDMA, which can be derived from Remark 1 and (7). ## X Conclusion In this paper, we investigate the coexistence of UAV radar and communication network. We deploy UAV-radars and UAV-comms by using HPPP where UAV-radars detect and track targets and UAV-comms communicate with their serving users in the same frequency band. We take into account two different multiple-access protocols, SOMA, and TDMA, to operate both radar signals and data signals simultanously. We analyze the performance of SRP in the radar detection scenario and TC in the data communication scenario. We show that in general, TDMA outperforms SOMA on SRP, while SOMA outperforms TDMA on TC. However, SOMA can achieve higher SRP as well as higher TC when the node density of UAV- radars is higher than that of UAV-comms (i.e., $\lambda_{\rm d}^{\prime}<\bar{\lambda}_{\rm r}$), for $\phi=\tau=0.5$, and $\mathsf{Pr}(\beta_{1}<\beta_{\rm th})<\frac{1}{2}$. We also find the UAV-comm node density that maximizes TC by derving the first and the second derivative of its analytic form and analyze the behavior of SRP and TC depending on the node density, the radius of the guard zone, the power splitting factor, and time division factor. ## References * [1] S. Hayat, E. Yanmaz, and R. Muzaffar, “Survey on Unmanned Aerial Vehicle Networks for Civil Applications: A Communications Viewpoint,” _IEEE Commun. Surveys Tuts._ , vol. 18, no. 4, pp. 2624–2661, Apr. 2016. * [2] A. Merwaday and I. Guvenc, “UAV Assisted Heterogeneous Networks for Public Safety Communications,” in _Proc. IEEE Wireless Commun. Netw. Conf. (WCNC)_ , New Orleans, LA, USA, Mar. 2015, pp. 329–334. * [3] P. Chandhar, D. Danev, and E. G. Larsson, “Massive MIMO for Communications With Drone Swarms,” _IEEE Trans. Wireless Commun._ , vol. 17, no. 3, pp. 1604–1629, Mar. 2018. * [4] S. Ortiz, C. T. Calafate, J.-C. Cano, P. Manzoni, and C. K. Toh, “A UAV-Based Content Delivery Architecture for Rural Areas and Future Smart Cities,” _IEEE Internet Computing_ , vol. 23, no. 1, pp. 29–36, Jan. 2019. * [5] L. Zheng, M. Lops, Y. C. Eldar, and X. Wang, “Radar and Communication Coexistence: An Overview: A Review of Recent Methods,” _IEEE Sig. Proc. Mag._ , vol. 36, no. 5, pp. 85–99, Sep. 2019. * [6] H. ElSawy, A. Sultan-Salem, M.-S. Alouini, and M. Z. Win, “Modeling and Analysis of Cellular Networks Using Stochastic Geometry: A Tutorial,” _IEEE Commun. Surveys Tuts._ , vol. 19, no. 1, pp. 167–203, 2017. * [7] A. M. Hunter, J. G. Andrews, and S. Weber, “Transmission Capacity of Ad Hoc Networks with Spatial Diversity,” _IEEE Trans. Wireless Commun._ , vol. 7, no. 12, pp. 5058–5071, Dec. 2008. * [8] J. Park, B. Clerckx, C. Song, and Y. Wu, “An Analysis of the Optimum Node Density for Simultaneous Wireless Information and Power Transfer in Ad Hoc Networks,” _IEEE Trans. Veh. Technol._ , vol. 67, no. 3, pp. 2713–2726, Mar. 2018. * [9] J. Park, J.-P. Hong, W. Shin, and S. Kim, “Performance Analysis of Distributed Antenna System for Downlink Ultrareliable Low-Latency Communications,” _IEEE Syst. J._ , vol. 15, no. 1, pp. 518–525, Mar. 2021\. * [10] A. Al-Hourani, R. J. Evans, S. Kandeepan, B. Moran, and H. Eltom, “Stochastic Geometry Methods for Modeling Automotive Radar Interference,” _IEEE Trans. Intell. Transp. Syst._ , vol. 19, no. 2, pp. 333–344, Feb. 2018. * [11] Z. Fang, Z. Wei, X. Chen, H. Wu, and Z. Feng, “Stochastic Geometry for Automotive Radar Interference With RCS Characteristics,” _IEEE Wireless Commun. Lett._ , vol. 9, no. 11, pp. 1817–1820, Nov. 2020. * [12] A. Munari, L. Simić, and M. Petrova, “Stochastic Geometry Interference Analysis of Radar Network Performance,” _IEEE Commun. Lett._ , vol. 22, no. 11, pp. 2362–2365, Nov. 2018. * [13] M. Rebato, J. Park, P. Popovski, E. De Carvalho, and M. Zorzi, “Stochastic Geometric Coverage Analysis in mmWave Cellular Networks With Realistic Channel and Antenna Radiation Models,” _IEEE Trans. Commun._ , vol. 67, no. 5, pp. 3736–3752, May 2019. * [14] X. Yu, J. Zhang, M. Haenggi, and K. B. Letaief, “Coverage Analysis for Millimeter Wave Networks: The Impact of Directional Antenna Arrays,” _IEEE J. Sel. Areas Commun._ , vol. 35, no. 7, pp. 1498–1512, July 2017\. * [15] C. Wang and H.-M. Wang, “Physical Layer Security in Millimeter Wave Cellular Networks,” _IEEE Trans. Wireless Commun._ , vol. 15, no. 8, pp. 5569–5585, Aug. 2016. * [16] M. M. Azari, G. Geraci, A. Garcia-Rodriguez, and S. Pollin, “UAV-to-UAV Communications in Cellular Networks,” _IEEE Trans. Wireless Commun._ , vol. 19, no. 9, pp. 6130–6144, Sep. 2020. * [17] C.-H. Liu, K.-H. Ho, and J.-Y. Wu, “MmWave UAV Networks With Multi-Cell Association: Performance Limit and Optimization,” _IEEE J. Sel. Areas Commun._ , vol. 37, no. 12, pp. 2814–2831, Dec. 2019. * [18] V. V. Chetlur and H. S. Dhillon, “Downlink Coverage Analysis for a Finite 3-D Wireless Network of Unmanned Aerial Vehicles,” _IEEE Trans. Commun._ , vol. 65, no. 10, pp. 4543–4558, Oct. 2017. * [19] W. Yi, Y. Liu, Y. Deng, and A. Nallanathan, “Clustered UAV Networks With Millimeter Wave Communications: A Stochastic Geometry View,” _IEEE Trans. Commun._ , vol. 68, no. 7, pp. 4342–4357, July 2020. * [20] X. Wang and M. C. Gursoy, “Coverage Analysis for Energy-Harvesting UAV-Assisted mmWave Cellular Networks,” _IEEE J. Sel. Areas Commun._ , vol. 37, no. 12, pp. 2832–2850, Dec. 2019. * [21] W. Yi, Y. Liu, E. Bodanese, A. Nallanathan, and G. K. Karagiannidis, “A Unified Spatial Framework for UAV-Aided MmWave Networks,” _IEEE Trans. Commun._ , vol. 67, no. 12, pp. 8801–8817, Dec. 2019. * [22] M. Gapeyenko, D. Moltchanov, S. Andreev, and R. W. Heath, “Line-of-Sight Probability for mmWave-Based UAV Communications in 3D Urban Grid Deployments,” _IEEE Trans. Wireless Commun._ , vol. 20, no. 10, pp. 6566–6579, Oct. 2021. * [23] Z. Huang, C. Chen, and M. Pan, “Multiobjective UAV Path Planning for Emergency Information Collection and Transmission,” _IEEE Internet Things J._ , vol. 7, no. 8, pp. 6993–7009, Aug. 2020. * [24] X. Sun, W. Yang, and Y. Cai, “Secure Communication in NOMA-Assisted Millimeter-Wave SWIPT UAV Networks,” _IEEE Internet Things J._ , vol. 7, no. 3, pp. 1884–1897, Mar. 2020. * [25] M. M. U. Chowdhury, S. J. Maeng, E. Bulut, and I. Guvenc, “3-D Trajectory Optimization in UAV-Assisted Cellular Networks Considering Antenna Radiation Pattern and Backhaul Constraint,” _IEEE Trans. Aerosp. Electron. Syst._ , vol. 56, no. 5, pp. 3735–3750, Oct. 2020. * [26] Y. Pan, K. Wang, C. Pan, H. Zhu, and J. Wang, “UAV-Assisted and Intelligent Reflecting Surfaces-Supported Terahertz Communications,” _IEEE Wireless Commun. Lett._ , vol. 10, no. 6, pp. 1256–1260, June 2021. * [27] M. Cui, G. Zhang, Q. Wu, and D. W. K. Ng, “Robust Trajectory and Transmit Power Design for Secure UAV Communications,” _IEEE Trans. Veh. Technol._ , vol. 67, no. 9, pp. 9042–9046, Sep. 2018. * [28] S. J. Maeng, Y. Yapici, I. Guvenc, H. Dai, and A. Bhuyan, “Precoder Design for mmWave UAV Communications with Physical Layer Security,” in _Proc. IEEE Int. Workshop Signal Process. Adv. Wireless Commun. (SPAWC)_ , Atlanta, GA, USA, May 2020, pp. 1–5. * [29] R. O. R. Jenssen, M. Eckerstorfer, and S. Jacobsen, “Drone-Mounted Ultrawideband Radar for Retrieval of Snowpack Properties,” _IEEE Trans. Instrum. Meas._ , vol. 69, no. 1, pp. 221–230, Jan. 2020. * [30] S. Z. Gurbuz, U. Kaynak, B. Ozkan, O. C. Kocaman, F. Kiyici, and B. Tekeli, “Design Study of a Short-range Airborne UAV Radar for Human Monitoring,” in _Proc. IEEE Asilomar Conf. on Signals, Syst., and Comput._ , Pacific Grove, CA, USA, Nov. 2014, pp. 568–572. * [31] G. Fasano, A. Renga, A. R. Vetrella, G. Ludeno, I. Catapano, and F. Soldovieri, “Proof of Concept of Micro-UAV-based Radar Imaging,” in _Proc. International Conference on Unmanned Aircraft Systems (ICUAS)_ , Miami, FL, USA, June 2017, pp. 1316–1323. * [32] E. Vinogradov, D. A. Kovalev, and S. Pollin, “Simulation and Detection Performance Evaluation of a UAV-mounted Passive Radar,” in _Proc. IEEE Int. Symp. Personal, Indoor, Mobile Radio Commun. (PIMRC)_ , Bologna, Italy, Sep. 2018, pp. 1185–1191. * [33] F. Liu, C. Masouros, A. Li, and T. Ratnarajah, “Robust MIMO Beamforming for Cellular and Radar Coexistence,” _IEEE Wireless Commun. Lett._ , vol. 6, no. 3, pp. 374–377, June 2017. * [34] J. Qian, M. Lops, L. Zheng, and X. Wang, “Joint Design for Co-existence of MIMO Radar and MIMO Communication System,” in _Proc. IEEE Asilomar Conf. on Signals, Syst., and Comput._ , Pacific Grove, CA, USA, Oct. 2017, pp. 568–572. * [35] Q. Wu, J. Xu, Y. Zeng, D. W. K. Ng, N. Al-Dhahir, R. Schober, and A. L. Swindlehurst, “A Comprehensive Overview on 5G-and-Beyond Networks With UAVs: From Communications to Sensing and Intelligence,” _IEEE J. Sel. Areas Commun._ , vol. 39, no. 10, pp. 2912–2945, Oct. 2021. * [36] X. Chen, Z. Feng, Z. Wei, F. Gao, and X. Yuan, “Performance of Joint Sensing-Communication Cooperative Sensing UAV Network,” _IEEE Trans. Veh. Technol._ , vol. 69, no. 12, pp. 15 545–15 556, Dec. 2020. * [37] A. Al-Hourani, S. Kandeepan, and S. Lardner, “Optimal LAP Altitude for Maximum Coverage,” _IEEE Wireless Commun. Lett._ , vol. 3, no. 6, pp. 569–572, Dec. 2014. * [38] M. Haenggi, “Mean Interference in Hard-Core Wireless Networks,” _IEEE Commun. Lett._ , vol. 15, no. 8, pp. 792–794, Aug. 2011.
# Strong two-meson decays of light and charmed vector mesons Roberto Correa da Silveira LFTC, Universidade Cidade de São Paulo, Rua Galvão Bueno 868, São Paulo, SP 01506-000, Brazil Fernando E. Serna LFTC, Universidade Cidade de São Paulo, Rua Galvão Bueno 868, São Paulo, SP 01506-000, Brazil Departamento de Física, Universidad de Sucre, Carrera 28 No. 5-267, Barrio Puerta Roja, Sincelejo 700001, Colombia Bruno El-Bennich LFTC, Universidade Cidade de São Paulo, Rua Galvão Bueno 868, São Paulo, SP 01506-000, Brazil ###### Abstract We calculate the strong decay couplings for $\rho\to\pi\pi$, $\phi\to KK$, $K^{*}\to K\pi$ and $D^{*}\to D\pi$ in a unified and consistent approach based on the impulse approximation, nonperturbative solutions of the quark-gap equation and the Poincaré invariant Bethe-Salpeter amplitudes of vector and pseudoscalar mesons. In particular, we obtain the coupling $g_{D^{*}\\!D\pi}=17.24^{+3.06}_{-2.30}$ in very good agreement with the experimental value by CLEO, which corresponds to a strong effective coupling between heavy vector and pseudoscalar mesons to the pion of $\hat{g}=0.58^{+0.10}_{-0.08}$. ## I Introduction One of the challenges in hadron physics is to understand the spectrum, constituent composition and momentum distribution of quarks and gluons within the hadrons. To obtain deeper insight into the hadron’s structure, their excitations have been intensively investigated in the past decades. This includes radial excitations, higher angular momentum states and exotic states containing constituent gluons that contribute to the total angular momentum of the hadron. The vector mesons, being the lowest spin excitations of the pseudoscalars, offer a first glimpse into an electromagnetic excitation of a $\bar{q}q$ pair. This is because neutral vector mesons can directly couple to the photon via an electromagnetic current since their quantum numbers, $J^{PC}=1^{--}$, are those of the photon. Naturally, they have been much studied and from the viewpoint of functional approaches to Quantum Chromodynamics (QCD) they were helpful to establish the ladder truncation of the Bethe-Salpeter equation (BSE) Maris:1999nt , at least for the ground states of lighter vector mesons. Of course, beyond the masses of the pseudoscalar and vector mesons, their electromagnetic and electroweak properties are of fundamental interest and there is no lack of studies dedicated to weak decay constants, elastic and transition form factors Maris:2000sk ; deMelo:1997hh ; Bakker:2002mt ; deMelo:2012hj ; DeMelo:2018bim ; deMelo:2014gea ; Chang:2013nia ; Raya:2015gva ; Ding:2018xwy ; Xu:2019ilh ; daSilva:2012gf ; El-Bennich:2012mkr ; El- Bennich:2008dhc ; Ivanov:2007cw ; El-Bennich:2009gbu . Beyond these observables, the strong decays of vector mesons into two light(er) mesons provide another source of information on the nonperturbative dynamics complementary to electromagnetic interactions and weak decays. They are the simplest possible decays that proceed via strong interactions and since the vector meson decays via a $P$-wave interaction, the Bethe-Salpeter amplitude (BSA) is probed differently than in the electroweak sector. In here, our main object of interest is the reaction $D^{*}\to D\pi$ which we studied in Refs. El-Bennich:2010uqs ; El-Bennich:2011tme ; El-Bennich:2012hom ; El- Bennich:2016bno within the limitations of not having the BSA of charmed mesons at hand but motivated by the first measurement of the $D^{*}$ width, $\Gamma(D^{*+})=96\pm 4\pm 22$ keV CLEO:2001sxb . This result is of great interest, as it is one of the few quantities in flavor physics that does not probe electroweak properties of heavy mesons and which opens a window on nonperturbative QCD in mesons with two very distinct mass scales. Moreover, the strong coupling $g_{D^{*}\\!D\pi}$ one can extract from the decay width is related to a putative universal coupling $\hat{g}$ between heavy-light mesons and a low-momentum pion in the heavy-meson chiral Lagrangian Casalbuoni:1996pg ; El-Bennich:2021ldv ; Braghin:2021qmu . At leading order in the $1/m_{D}$ expansion this relation is $g_{D^{*}\\!D\pi}=2\sqrt{m_{D}^{*}m_{D}}\,\hat{g}/f_{\pi}$. These calculations were based on one-covariant models of the $D$ and $D^{*}$ wave functions and were therefore not Poincaré invariant, so that the momentum partition parameters had to be chosen according to some semi-classical criterium El-Bennich:2010uqs . This, of course, was not satisfactory and the motivation remained to compute the decay amplitude guided by Ref. Jarecke:2002xd which dealt with the decays $\rho\to\pi\pi$, $\phi\to\bar{K}K$ and $K^{*}\to K\pi$. In this work we close this gap and compute the vector- meson decay with the complete Poincaré covariant structure of the BSA for the $D$, $D^{*}$ and pion. Along the way, we also obtain the strong couplings considered in Ref. Jarecke:2002xd which we update. The remainder of this paper is composed of five sections: in Section II we explain the framework in which the strong couplings are calculated and define the decay kinematics; in Section III we describe the functional approach to QCD we use to calculate the quark propagators and BSA of the mesons within a given truncation scheme in Euclidean space; in Section IV the numerical method to solve the BSE is summarized and the meson’s masses and weak decay constants are calculated. Finally, in Section V we present our results for the strong couplings and wrap up with final remarks in Section VI. ## II Strong Decay Amplitude In what follows, we limit ourselves to the _impulse approximation_ of the strong decays depicted in Fig. 1. As argued in Ref. Jarecke:2002xd , since the $\rho$ and $\phi$ appear as resonance poles in the timelike electromagnetic form factors of the pion and kaon, the pole residues are proportional to the respective coupling constants $g_{\rho\pi\pi}$ and $g_{\phi KK}$. Hence, if these form factors are obtained in impulse approximation, so will be the couplings. Now, the impulse approximation for the electromagnetic coupling to mesons conserves the current as long as the meson’s BSA and quark-photon vertex are calculated in the ladder and the quark propagators in rainbow truncation, respectively, and the resulting electromagnetic form factors are in excellent agreement with experiment Maris:2000sk ; Chang:2013nia ; Raya:2015gva ; Ding:2018xwy . In the time-like region, on the other hand, the ladder truncation of the BSE fails to produce the $\rho$ pole in $e^{+}e^{-}\to\gamma^{*}\to\pi^{+}\pi^{-}$ and was amended to include effective pion degrees of freedom in the BSE scattering kernel Williams:2018adr ; Miramontes:2019mco ; Miramontes:2021xgn . Therefore, we expect that the impulse approximation for the strong decays of lighter mesons misses some of the relevant physics, in particular in the case of the $\rho$ meson whose decay width is almost 20% of its mass. Going beyond this approximation, not merely in the BSE kernel but also in the decay amplitude, is a technically and numerically challenging task. As the main aim here is to improve on earlier calculations of $D^{*}\to D\pi$, we deliberately ignore these corrections. The strong decay coupling for a process $V\to PP$ is defined as, $\langle P(p_{2})P(q)|V(p_{1},\lambda)\rangle:=\,g_{V\\!PP}\,\epsilon^{\lambda}\\!\cdot q\ ,$ (1) where the initial state is a vector meson with transverse polarization $\epsilon^{\lambda}_{\mu}$ and momentum $p_{1}^{2}=-m_{V}^{2}$, while the light(er) mesons have on-shell momenta $p_{1}^{2}=-m_{P}^{2}$, $q^{2}=-m_{P}^{2}$, with $q=p_{1}-p_{2}$, and can have different flavor content. The decay amplitude in impulse approximation can be expressed by the loop integral: $\displaystyle g_{V\\!PP}\;\epsilon^{\lambda}\\!\cdot q\,$ $\displaystyle=\int^{\Lambda}\\!\frac{d^{4}k}{(2\pi)^{4}}\,\mathrm{Tr_{CD}}\,\big{[}\epsilon^{\lambda}\\!\cdot\Gamma_{V}(k_{V},p_{1})$ $\displaystyle S_{f}(k_{1})$ $\displaystyle\bar{\Gamma}_{P}(k_{P},-p_{2})S_{f}(k_{2})\bar{\Gamma}_{P}(k_{P}^{\prime},-q)S_{f}(k_{3})\big{]}\ .$ (2) In here, $\Gamma(k,P)$ are the BSAs of the mesons, $S(k_{i})$ are the quarks propagators and the trace is over color and Dirac indices. Following the momentum flow in Fig. 1, the quark momenta are defined as, $\displaystyle k_{1}$ $\displaystyle=k+w_{1}p_{1}\,,$ (3) $\displaystyle k_{2}$ $\displaystyle=k+w_{1}p_{1}-p_{2}\,,$ (4) $\displaystyle k_{3}$ $\displaystyle=k-w_{2}p_{1}\,.$ (5) with the constraint $w_{1}+w_{2}=1$ on the partition parameters due to momentum conservation. The relative BSA momenta are given by, $\displaystyle k_{V}$ $\displaystyle=k+\tfrac{1}{2}(w_{1}-w_{2})p_{1}\,,$ (6) $\displaystyle k_{P}$ $\displaystyle=k+w_{1}p_{1}-\tfrac{1}{2}\,p_{2}\,,$ (7) $\displaystyle k_{P}^{\prime}$ $\displaystyle=k+\tfrac{1}{2}(w_{1}-w_{2})p_{1}-\tfrac{1}{2}\,p_{2}\,.$ (8) Note that the relative momentum of the vector meson is only real if $w_{1}=w_{2}$, as in the meson’s rest frame $p_{1}=(\mathbf{0},im_{V})$ in the Euclidean-space formulation we use. This will be discussed in more detail in Section V. Figure 1: Decay diagram depicting a generic strong decay $V\to PP$ in the impulse approximation of Eq. (2). The shaded ovals represent vector ($V$) and pseudoscalar ($P$) meson BSAs (15), while the dark-shaded circles symbolize dressed quark propagators (34) and the double-lined arrows describe the incoming vector-meson momentum and outgoing pseudoscalar-meson momenta. We conclude this section by mentioning some definitions with respect to the charge when one of the final mesons is an isovector state. We follow Ref. Bracco:2011pg and define the generic $D^{*}D\pi$ coupling as the one containing the neutral meson: $\displaystyle g_{D^{*}\\!D\pi}$ $\displaystyle=\,g_{D^{*\pm}D\mp\pi^{0}}=g_{D^{*0}D^{0}\pi^{0}}$ $\displaystyle=\,\tfrac{1}{\sqrt{2}}\,g_{D^{*-}D^{0}\pi^{+}}=\tfrac{1}{\sqrt{2}}\,g_{D^{*+}D^{0}\pi^{-}}\ .$ (9) Likewise, considering SU(3) flavor algebra one has, $g_{K^{*}\\!K\pi}=\sqrt{3}\,g_{K^{*+}K^{+}\pi^{0}}=\sqrt{\tfrac{3}{2}}\,g_{K^{*+}K^{0}\pi^{+}}\ ,$ (10) and moreover $g_{\rho\pi\pi}=g_{\rho^{0}\pi^{+}\pi^{-}}$, $g_{\phi KK}=g_{\phi K^{+}K^{-}}$. In Section III we describe how the ingredients of the strong decay amplitude (2), namely the quark propagators and BSAs, are obtained from solving the quark-gap equation and the BSE. ## III Pseudoscalar and Vector Meson Bound States ### III.1 Bethe-Salpeter Equation The relativistic initial and final bound states in the decay amplitude (2) are described by Poincaré covariant BSAs, $\Gamma^{fg}_{P}(k,P)$ and $\Gamma^{fg}_{V\nu}(k,P)$, which are the solutions of the homogeneous BSE in the $J^{PC}=0^{-+}$ and $J^{PC}=1^{--}$ channels, respectively Bashir:2012fs : $\displaystyle\Gamma^{fg}_{P}(k,P)=\int^{\Lambda}\\!\frac{d^{4}q}{(2\pi)^{4}}\,K^{fg}(k,q,P)\,\chi^{fg}_{P}(q,P)\ ,$ (11) $\displaystyle\Gamma^{fg}_{V\nu}(k,P)=\int^{\Lambda}\\!\frac{d^{4}q}{(2\pi)^{4}}\,K^{fg}(k,q,P)\,\chi^{fg}_{V\nu}(q,P)\ .$ (12) In these BSEs, $k$ is the relative quark-antiquark momentum, $P$ is the meson momentum and $K^{fg}(k,q,P)$ is the fully amputated scattering kernel which sums up all possible quark-antiquark interactions. The Bethe-Salpeter wave functions, $\chi^{fg}_{P}(k,p)$ and $\chi^{fg}_{V\nu}(k,P)$, are obtained by attaching the quark propagators to the BSA, $\displaystyle\chi^{fg}_{P}(k,P)$ $\displaystyle=\ S_{f}(k_{\eta})\,\Gamma^{fg}_{P}(k,P)S_{g}(k_{\bar{\eta}})\ ,$ (13) $\displaystyle\chi^{fg}_{V\nu}(k,P)$ $\displaystyle=\ S_{f}(k_{\eta})\,\Gamma^{fg}_{V\nu}(k,P)S_{g}(k_{\bar{\eta}})\ ,$ (14) with the shorthands, $k_{\eta}=k+\eta P$ and $k_{\bar{\eta}}=k-\bar{\eta}P$, that define momentum-partition parameters: $\eta+\bar{\eta}=1$. The BSA has the most general Poincaré covariant form that can be composed of the Dirac matrices and the relative and total momenta consistent with the quantum numbers $P$ and $C$ of a given meson, $\Gamma_{M}^{fg}(k,P)=\sum_{i=1}^{N}\,T^{i}(k,P)\,\mathcal{F}_{i}^{fg}\big{(}k,P,z_{k}\big{)}\ ,$ (15) where $T^{i}(k,P)$ are Dirac covariants, $\mathcal{F}_{i}^{fg}$ are scalar Lorentz-invariant amplitudes and $z_{k}=k\cdot P/|k\|P|$ is an angle between $k$ and $P$. In case of pseudoscalar mesons, we choose the usual $N=4$ covariants, $\displaystyle T^{1}(k,P)$ $\displaystyle=i\gamma_{5}\,$ (16) $\displaystyle T^{2}(k,P)$ $\displaystyle=\gamma_{5}\,\gamma\cdot P\ ,$ (17) $\displaystyle T^{3}(k,P)$ $\displaystyle=\gamma_{5}\,\gamma\cdot k\,k\cdot P\ ,$ (18) $\displaystyle T^{4}(k,P)$ $\displaystyle=\gamma_{5}\,\sigma_{\mu\nu}\,k_{\mu}P_{\nu}\ ,$ (19) and for a vector meson $N=8$ covariant vector components are required: $\displaystyle T^{1}_{\nu}(k,P)$ $\displaystyle=\ i\gamma^{T}_{\nu}\ ,$ (20) $\displaystyle T^{2}_{\nu}(k,P)$ $\displaystyle=\ i\left[3k^{T}_{\nu}\gamma\cdot k^{T}-\gamma^{T}_{\nu}\big{(}k^{T}\big{)}^{\\!2}\right]\ ,$ (21) $\displaystyle T^{3}_{\nu}(k,P)$ $\displaystyle=\ ik\cdot P\,\gamma\cdot P\,k^{T}_{\nu}\ ,$ (22) $\displaystyle T^{4}_{\nu}(k,P)$ $\displaystyle=\ i\left[\gamma^{T}_{\nu}\gamma\cdot P\,\gamma\cdot k^{T}+k^{T}_{\nu}\,\gamma\cdot P\right]\ ,$ (23) $\displaystyle T^{5}_{\nu}(k,P)$ $\displaystyle=\ k^{T}_{\nu}\ ,$ (24) $\displaystyle T^{6}_{\nu}(k,P)$ $\displaystyle=\ k\cdot P\left[\gamma^{T}_{\nu}\gamma\cdot k^{T}-\gamma\cdot k^{T}\,\gamma^{T}_{\nu}\right]\ ,$ (25) $\displaystyle T^{7}_{\nu}(k,P)$ $\displaystyle=\ \gamma^{T}_{\nu}\gamma\cdot P-\gamma\cdot P\,\gamma^{T}_{\nu}-2\,T_{\nu}^{8}(k,P)\ ,$ (26) $\displaystyle T_{\nu}^{8}(k,P)$ $\displaystyle=\ \hat{k}^{T}_{\nu}\,\gamma\cdot\hat{k}^{T}\,\gamma\cdot P\ .$ (27) The transverse projections are $V^{T}_{\nu}=V_{\nu}-P_{\nu}(V\cdotp P)/P^{2}$ with $P\cdot V^{T}=0$ for any four-vector $V_{\nu}$ and $\hat{k}^{T}\cdot\hat{k}^{T}=1$. Note that the $T^{i}_{\mu}(k,P)$ in Eqs. (20) to (27) form an orthogonal basis Maris:1999nt ; Gao:2014bca with respect to the Dirac trace. In order to calculate the meson’s weak decay constant, one has to normalize the meson’s BSA. We do so with the derivative of the eigenvalue trajectory, $\lambda(P^{2})$, of the BSE Nakanishi:1965zz ; Nakanishi:1965zza : $\displaystyle\left(\frac{\partial\ln\lambda}{\partial P^{2}}\right)^{\\!-1}\\!\\!$ $\displaystyle=\ \operatorname{tr}_{\mathrm{CD}}\int\frac{d^{4}k}{(2\pi)^{4}}\,\bar{\Gamma}^{fg}_{M}(k;-P)$ $\displaystyle\times\ S_{f}(k_{\eta})\Gamma^{fg}_{M}(k;P)S_{g}(k_{\bar{\eta}})\,.$ (28) With this we calculate the weak decay constant of the pseudoscalar meson defined by $f_{P}P_{\mu}=\langle 0\,|\bar{q}_{g}\gamma_{5}\gamma_{\mu}q_{f}|P(k,P)\rangle\ ,$ (29) which can be expressed by the integral: $f_{P}P_{\mu}=\frac{\mathcal{Z}_{2}N_{c}}{\sqrt{2}}\int^{\Lambda}\\!\frac{d^{4}k}{(2\pi)^{4}}\,\operatorname{Tr}_{\mathrm{D}}\left[\gamma_{5}\gamma_{\mu}\,\chi_{P}^{fg}(k,P)\right]\,.$ (30) Likewise, the weak decay constant of a vector meson is defined by the amplitude, $f_{V}m_{V}\,\epsilon^{\lambda}_{\mu}=\langle 0\,|\bar{q}_{g}\gamma_{\mu}q_{f}|V(k,P,\lambda)\rangle\ ,$ (31) where $m_{V}$ is the vector-meson mass and $\epsilon^{\lambda}_{\mu}(P)$ is the polarization vector of the transverse vector meson of helicity $\lambda$ which satisfies $\epsilon^{\lambda}\cdot P=0$ and is normalized as ${\epsilon^{\lambda}}^{*}\\!\cdot\epsilon^{\lambda}=3$. This can again be expressed by a loop integral: $f_{V}m_{V}=\frac{\mathcal{Z}_{2}N_{c}}{3\sqrt{2}}\int^{\Lambda}\\!\\!\frac{d^{4}k}{(2\pi)^{4}}\,\operatorname{Tr_{D}}\left[\gamma_{\mu}\,\chi^{fg}_{V\mu}(k,P)\right]\ .$ (32) In both expressions for the decay constants we define $\mathcal{Z}_{2}(\mu,\Lambda)=\surd Z_{2}^{f}\surd Z_{2}^{g}$, as in Section III.3, and $N_{c}=3$. ### III.2 Quark Gap Equation Amongst the Green functions that enter the BSE, whether in Eq. (11) or Eq. (12), are the flavor-dependent dressed quark propagators described by Schwinger functions we obtain as solutions of the Dyson-Schwinger equation (DSE), $\displaystyle S^{-1}_{f}(p)=$ $\displaystyle\ Z_{2}^{f}\\!\left(i\,\gamma\cdot p+m^{\mathrm{bm}}_{f}\right)$ $\displaystyle+\ Z_{1}^{f}$ $\displaystyle g^{2}\\!\\!\int^{\Lambda}\\!\frac{d^{4}k}{(2\pi)^{4}}\,D^{ab}_{\mu\nu}(q)\frac{\lambda^{a}}{2}\gamma_{\mu}S_{f}(k)\,\Gamma^{b}_{\nu,f}(k,p)\,,$ (33) where $m^{\textrm{bm}}_{f}$ is the bare current-quark mass, $Z_{1}^{f}(\mu,\Lambda)$ and $Z_{2}^{f}(\mu,\Lambda)$ are the vertex and wave- function renormalization constants at the renormalization point $\mu$, respectively. The integral in Eq. (33) represents the self-energy of the quark and involves the dressed-quark propagator $S_{f}(k)$, the dressed-gluon propagator $D_{\mu\nu}(q)$ with momentum $q=k-p$ and the quark-gluon vertex, $\Gamma^{a}_{\mu}(k,p)=\frac{1}{2}\,\lambda^{a}\Gamma_{\mu}(k,p)$ Albino:2018ncl ; Albino:2021rvj ; El-Bennich:2022obe , where the SU(3) color matrices $\lambda^{a}$ are in the fundamental representation. The Poincaré- invariant regularization scale is $\Lambda\gg\mu$ and can be taken to infinity. The solution of the DSE can be cast in the most general covariant form as, $\displaystyle S_{f}(p)$ $\displaystyle=\,-i\gamma\cdot p\,\sigma_{\rm v}^{f}(p^{2})+\sigma_{\rm s}^{f}(p^{2})$ $\displaystyle=\,Z_{f}(p^{2})/\left[i\gamma\cdot p+M_{f}(p^{2})\right]\ .$ (34) In this DSE, $Z_{f}(p^{2})$ defines the wave function and $M_{f}(p^{2})$ is the running mass of the quark. The scalar functions $\sigma_{\rm s}^{f}(p^{2})$ and $\sigma_{\rm v}^{f}(p^{2})$ thus depend on $Z_{f}(p^{2})$ and $M_{f}(p^{2})$. In a subtractive renormalization scheme the two renormalization conditions, $\displaystyle Z_{f}(\mu^{2})$ $\displaystyle=\,1\ ,$ (35) $\displaystyle S^{-1}_{f}(\mu^{2})$ $\displaystyle=\,i\gamma\cdot p\ +m_{f}(\mu)\ ,$ (36) are imposed, where $m_{f}(\mu)$ is the renormalized current-quark mass related to the bare mass by, $Z_{4}^{f}(\mu,\Lambda)\,m_{f}(\mu)=Z_{2}^{f}(\mu,\Lambda)\,m_{f}^{\rm bm}(\Lambda)\ ,$ (37) and $Z_{4}^{f}(\mu,\Lambda)$ is the renormalization constant that pertains to the mass term in the QCD Lagrangian. ### III.3 Truncation Scheme The rainbow-ladder (RL) truncation of the integral equation (33) and of the BSE kernel has proven to be a robust and successful symmetry-preserving approximation and allows for the description of light ground-state mesons in the isospin-nonzero pseudoscalar and vector channels. The RL truncation is realized by restricting the fully dressed quark gluon vertex to the perturbative vertex: $\Gamma_{\nu,f}\to Z_{2}^{f}\gamma_{\nu}$. The DSE kernel then reduces to Serna:2018dwk , $Z_{1}^{f}g^{2}D_{\mu\nu}(q)\Gamma_{\nu,f}(k,p)=\big{(}Z^{f}_{2}\big{)}^{\\!2}\mathcal{G}_{f}(q^{2})D_{\mu\nu}^{\mathrm{free}}(q)\frac{\lambda^{a}}{2}\gamma_{\nu}\,,$ (38) in which an Abelianized Ward identity is enforced that leads to $Z_{1}^{f}=Z_{2}^{f}$ Bashir:2012fs and implies the omission of the three- gluon interaction in $\Gamma_{\mu}(k,p)$. An additional factor $Z_{2}^{f}$ in Eq. (38) ensures multiplicative renormalizability of the DSE and therefore the mass function $M_{f}(p^{2})$ is a renormalization-point invariant quantity Bloch:2002eq . We work in Landau gauge in which the free gluon propagator is transverse, $D_{\mu\nu}^{\mathrm{free}}(q):=\delta^{ab}\left(\delta_{\mu\nu}-\frac{q_{\mu}q_{\nu}}{q^{2}}\right)\\!\frac{1}{q^{2}}\ ,$ (39) and introduce the flavor-dependent interaction, $\frac{\mathcal{G}_{f}(q^{2})}{q^{2}}=\,\mathcal{G}_{f}^{\mathrm{IR}}(q^{2})+4\pi\tilde{\alpha}_{\mathrm{PT}}(q^{2})\ ,$ (40) where we deliberately absorb a factor $1/q^{2}$ from the gluon propagator (39). The dressing function $\mathcal{G}_{f}(q^{2})$ consists of a term that dominates in the infrared domain, $|k|<\Lambda_{\mathrm{QCD}}$, and is suppressed at large momenta, and a second term that implements the regular continuation of the perturbative QCD coupling and dominates large momenta. We use the model of Ref. Qin:2011dd given by, $\displaystyle\mathcal{G}_{f}^{\mathrm{IR}}(q^{2})$ $\displaystyle=\frac{8\pi^{2}}{\omega^{4}_{f}}D_{f}\,e^{-q^{2}/\omega^{2}_{f}}$ (41) $\displaystyle 4\pi\tilde{\alpha}_{\mathrm{PT}}(q^{2})$ $\displaystyle=\frac{8\pi^{2}\gamma_{m}\,\mathcal{E}(q^{2})}{\ln\left[\tau+\left(1+q^{2}/\Lambda^{2}_{\textrm{\tiny QCD}}\right)^{\\!2}\right]}\ ,$ (42) in which $\gamma_{m}=12/(33-2N_{f})$ is the anomalous mass dimension and $N_{f}$ is the active flavor number, $\Lambda_{\textrm{\tiny QCD}}=0.234$ GeV, $\tau=e^{2}-1$, $\mathcal{E}(q^{2})=[1-\exp(-q^{2}/4m^{2}_{t})]/q^{2}$ and $m_{t}=0.5$ GeV. The flavor dependence of the interaction $\mathcal{G}_{f}^{\mathrm{IR}}(q^{2})$ was introduced in Refs. Serna:2017nlr ; Serna:2020txe ; Chen:2019otg ; Serna:2022yfp to accommodate the strong flavor-symmetry breaking effects that led to complications in the calculation $D$\- and $B$-meson properties Rojas:2014aka ; Mojica:2017tvh ; see Refs. Serna:2020txe ; Serna:2022yfp ; Qin:2019oar for the details of the implementation of the BSE kernels (11) and (12) consistent with Eq. (38). Suffice to say that herein we employ $\mathcal{G}_{u}(q^{2})=\mathcal{G}_{d}(q^{2})=\mathcal{G}_{s}(q^{2})\neq\mathcal{G}_{c}(q^{2})\neq\mathcal{G}_{b}(q^{2})$. Flavor | $m_{f}(19\,\mathrm{GeV})$ | $m_{f}(2\,\mathrm{GeV})$ | $\omega_{f}$ | $\kappa$ | $M^{E}_{f}$ ---|---|---|---|---|--- $u,d$ | 0.0034 | 0.018 | 0.500 | 0.80 | 0.408 $s$ | 0.082 | 0.166 | 0.500 | 0.80 | 0.562 $c$ | 0.903 | 1.272 | 0.698 | 0.60 | 1.342 $b$ | 3.741 | 4.370 | 0.640 | 0.56 | 4.259 Table 1: Parameters of the interaction model in Eqs. (41) and (45): $m_{f}(\mu)$, $\omega_{f}$ and $\kappa=(\omega D_{f})^{1/3}$ (in GeV). $M^{E}_{f}$ is the Euclidean constituent quark mass: $M^{E}_{f}=\\{p^{2}|p^{2}=M^{2}(p^{2})\\}$. The form of the quark-antiquark ladder kernel we therefore employ differs somewhat from the usual one in case of heavy-flavored mesons, $K_{fg}(k,q,P)=-\mathcal{Z}_{2}^{2}\,\frac{\mathcal{G}_{fg}(l^{2})}{l^{2}}\,D_{\mu\nu}^{\mathrm{free}}(l)\,\frac{\lambda^{a}}{2}\gamma_{\mu}\frac{\lambda^{a}}{2}\gamma_{\nu}\ ,$ (43) in which the relative momentum is $l=k-q$. In other words, in Eq. (43) we combine the wave-function renormalization constants of both quarks, $\mathcal{Z}_{2}(\mu,\Lambda)=\surd{Z_{2}^{f}}\surd{Z_{2}^{g}}$, and use the averaged interaction, $\frac{\mathcal{G}_{fg}(l^{2})}{l^{2}}=\mathcal{G}_{fg}^{\mathrm{IR}}(l^{2})+4\pi\tilde{\alpha}_{\mathrm{PT}}(l^{2})\,,$ (44) which leads to a different treatment of the light and heavy quarks. The interaction in the low-momentum domain is given by the Gaussian form, $\mathcal{G}_{fg}^{\mathrm{IR}}(l^{2})=\frac{8\pi^{2}}{(\omega_{f}\omega_{g})^{2}}\sqrt{D_{f}\,D_{g}}\,e^{-l^{2}/(\omega_{f}\omega_{g})}\,,$ (45) while $4\pi\tilde{\alpha}_{\mathrm{PT}}(q^{2})$ is as in Eq. (42). The parameters of this interaction model are listed in Table 1 Serna:2020txe . | $m_{M}$ | $m^{\mathrm{exp}}_{M}$ | $\epsilon_{m_{M}}$ [%] | $f_{M}$ | $f^{\mathrm{exp/LQCD}}_{M}$ | $\epsilon_{f_{M}}$ [%] ---|---|---|---|---|---|--- $\pi(u\bar{d})$ | 0.140 | 0.138 | 1.45 | $0.094$ | 0.092(1) | 2.17 $K(u\bar{s})$ | 0.494 | 0.494 | 0.0 | $0.110$ | 0.110(2) | 0.0 $D(c\bar{d})$ | $1.867$ | 1.864 | 0.11 | $0.144$ | 0.150 (0.5) | 4.00 $\rho(u\bar{u})$ | 0.730 | 0.775 | 5.81 | 0.145 | 0.153(1) | 5.23 $\phi(s\bar{s})$ | 1.070 | 1.019 | 5.20 | 0.187 | 0.168(1) | 11.31 $K^{*}(u\bar{s})$ | 0.883 | 0.896 | 1.45 | 0.163 | 0.159(1) | 2.55 $D^{*}(c\bar{u})$ | 2.021 | 2.009 | 0.60 | 0.165 | 0.158(6) | 4.43 Table 2: Masses and weak decay constants [in GeV] of ground-state pseudoscalar and vector mesons, $M=P,V$. The experimental mass values are taken from the Particle Data Group [PDG] ParticleDataGroup:2022pth and the leptonic decay constants for the $\rho$, $K^{*}$ and $\phi$ mesons are derived from their experimental decay width via $f_{V}^{2}=\frac{3m_{V}}{4\pi\alpha^{2}Q^{2}}\,\Gamma_{V\rightarrow e^{+}e^{-}}$. The decay constant of the $D^{*}$ meson is a lattice-QCD prediction by the ETM collaboration Lubicz:2017asp . The relative deviations from experimental values are given by $\epsilon_{v}=|v^{\textrm{exp.}}-v^{\textrm{th.}}|/v^{\textrm{exp.}}$. The only missing ingredient now is the quark propagator for complex momenta, $S_{f}(q_{\eta})=-i\gamma\cdot q_{\eta}\,\sigma_{\rm v}^{f}(q_{\eta}^{2})+\sigma_{\rm s}^{f}(q_{\eta}^{2})\,,$ (46) and likewise for $S_{f}(q_{\bar{\eta}})$, as in Euclidean space the arguments $q_{\eta}^{2}$ and $q_{\bar{\eta}}^{2}$ define parabolas on the complex plane, $\displaystyle q^{2}_{\eta}$ $\displaystyle=$ $\displaystyle q^{2}-\eta^{2}m^{2}_{M}+2i\eta\,m_{M}|q|z_{q}\ ,$ $\displaystyle q^{2}_{\bar{\eta}}$ $\displaystyle=$ $\displaystyle q^{2}-{\bar{\eta}}^{2}m^{2}_{M}-2i\bar{\eta}\,m_{M}|q|z_{q}\,,$ (47) where $z_{q}=q\cdot P/|q||P|$. We apply Cauchy’s integral theorem as described, e.g., in Ref. Krassnigg:2008bob and obtain the solutions of the DSE on the complex plane with the contour parametrization of the parabola defined in Ref. Rojas:2014aka ; see Refs. El-Bennich:2016qmb ; Serna:2020txe for graphic visualizations of $\sigma_{\rm s}^{u}(q_{\eta}^{2})$. ## IV Pseudoscalar and Vector Meson Properties Using the quark propagators on the complex momentum plane (47) and the BSE kernel (43) in Eqs. (11) and (12), we treat the BSE as an eigenvalue problem El-Bennich:2015kja ; El-Bennich:2017brb . For instance, in case of the vector mesons, the covariant decomposition in Eq. (15) along with the orthogonality of the basis in Eqs. (20) to (27) allow to recast the homogeneous BSE (12) with the kernel (43) in a set of eight coupled-integral equations, $\displaystyle\mathcal{F}_{i}^{fg}$ $\displaystyle\big{(}k,P,z_{k}\big{)}=-\tfrac{4}{3}\,\mathcal{Z}_{2}^{2}\int^{\Lambda}_{q}\\!\mathcal{G}_{fg}\big{(}l^{2}\big{)}D^{\mathrm{free}}_{\mu\nu}(l)\mathcal{F}_{j}^{fg}(q,P,z_{q})$ $\displaystyle\times\,\mathrm{Tr}_{\mathrm{D}}\left[T^{i}_{\rho}(k,P)\gamma_{\mu}S_{f}(q_{\eta})T^{j}_{\rho}(q,P)S_{g}(q_{\bar{\eta}})\gamma_{\nu}\right],$ (48) where the mnemonic shortcut for the integral represents the same integral with Poincaré-invariant cut-off as before. In solving this equation system numerically, we expand the scalar amplitudes in terms of Chebyshev moments, $\mathcal{F}_{im}^{fg}(k,P)$, $\mathcal{F}_{i}^{fg}(k,P,z_{k})=\sum_{m=0}^{\infty}\mathcal{F}_{im}^{fg}(k,P)\,U_{m}(z_{k})\ ,$ (49) which allows for a faster convergence. We consider $m=3$ Chebyshev polynomials, $U_{m}(z_{p})$, of second kind. The eigenvalue problem for the vector $\bm{\mathcal{F}}:=\\{\mathcal{F}_{i},\ldots,\mathcal{F}_{8}\\}$ is then solved by means of Arnoldi factorization in the ARPACK library Lehoucq1998 . Details of the practical implementation of ARPACK in a numerical treatment of the BSE are reviewed, for instance, in Refs. Blank:2010bp ; Rojas:2014aka . Figure 2: Real and imaginary parts of the scalar amplitudes $\mathcal{F}_{1}(k,P)$ and $\mathcal{F}_{2}(k,P)$ of the $D^{*}$ meson. Figure 3: Real and imaginary parts of the leading BSA of the $D$ meson analytically continued on the complex momentum plane with the representation in Eq. (52). Note that the amplitude is plotted on the parabola spanned by $k_{P}^{2}$ (7) whose vertex lies in the time-like region. On the real axis the BSA normalization is $E_{D}(k_{p}=0,p_{2})=1$. The masses, $m_{P}$ and $m_{V}$, of the ground state pseudoscalar and vector mesons are the solutions of the eigenvalue trajectory $\lambda(P^{2}=-m^{2}_{P,V})=1$. They are listed along with the leptonic decay constants in Table 2 and are in very good agreement with experimental data, when available, or lattice-QCD results otherwise. As a byproduct we obtain the BSA of the mesons, which we illustrate with the real and imaginary parts of the dominant scalar functions of the $D^{*}$ meson in Fig. 2. ## V Results As we work in Euclidean space, the relative momenta, $k_{P}$ and $k_{P}^{\prime}$ in the decay, Eqs. (7) and (8), of the final pseudoscalar mesons and of the $D^{*}$ (6) are complex. This is because in the center of mass of the initial vector meson its four-momentum is $p_{1}=(\mathbf{0},im_{V})$ and thus $k_{V}=k+(w_{1}-w_{2})p_{1}/2$ is only real if $w_{1}=w_{2}$. In case of the final-state mesons, the relative momentum is inevitably complex. In principle, due to the Poincaré invariance of the BSAs, our calculations are independent of the choice for the partition parameters. Practically, though, we are limited by numerical constraints as choosing $w_{2}=0.5$ in the case of $m_{V}=m_{D^{*}}$ implies probing the light-quark propagator at large time-like momenta, much larger than the light quark’s mass. In this region, the solutions of the quark propagator on the complex momentum plane are characterized by branch cuts and/or complex- conjugate poles El-Bennich:2016qmb ; Eichmann:2021vnj and a contour deformation is not trivial. Having in hand the numerical BSA for real momenta, we therefore parametrize it with a Nakanishi type of representation which allows for an analytical continuation of the $\mathcal{F}_{i}^{fg}(k,P)$ (49) in the complex plane. In order to do so, we split the BSA in even and odd components, $\mathcal{F}_{i}(k,P)=\mathcal{F}^{0}_{i}(k,P)+k\cdot P\,\mathcal{F}^{1}_{i}(k,P)\,,$ (50) in which $\mathcal{F}^{0,1}_{i}(k,P)$ are even under $k\cdot P\rightarrow-k\cdot P$ and where we henceforth suppress the flavor indices $fg$. As discussed, for instance, in Ref. Serna:2020txe , $\mathcal{F}^{1}_{i}(k,P)\equiv 0$ for flavorless pseudoscalar mesons, such as the neutral pion, as they are eigenstates of the charge-conjugation operator defined as, $\Gamma_{M}(k,P)\,\stackrel{{\scriptstyle C}}{{\longrightarrow}}\ \bar{\Gamma}_{M}(k,P):=C\,\Gamma^{T}_{M}(-k,P)C^{T}\ .$ (51) The constraint that the covariant basis (15) satisfies $\bar{\Gamma}_{M}(k,P)=\lambda_{c}\Gamma_{M}(k,P)$ with $\lambda_{c}=+1$ for pseudoscalar mesons and $\lambda_{c}=-1$ for vector mesons, respectively, therefore imposes a definite parity of the scalar amplitudes $\mathcal{F}_{i}(k,P)$ under $k\cdot P\rightarrow-k\cdot P$. For the neutral vector mesons, $\rho$ and $\phi$, this implies that the $\mathcal{F}_{i}(k,P)$ are necessarily even and $\mathcal{F}^{1}_{i}(k,P)\equiv 0$ again. In case of the $K$, $K^{*}$, $D$ and $D^{*}$ mesons, which are not eigenstates of $C$, both amplitudes in Eq. (50) must be considered. $g_{V\\!PP}$ | $(V,P)=(8,4)$ | $(V,P)=(5,4)$ | $(V,P)=(1,1)$ | Reference | $\epsilon_{g_{V\\!P\\!P}}\ [\%]$ ---|---|---|---|---|--- $g_{\rho\pi\pi}$ | $5.13^{+0.24}_{-0.25}$ | 5.14 | 7.99 | $5.94\pm 0.44$ | 13.6 $g_{\phi KK}$ | $5.02^{+0.12}_{-0.22}$ | 5.03 | 10.12 | $5.53\pm 0.31$ | 9.2 $g_{K^{*}\\!K\pi}$ | $5.10^{+0.22}_{-0.26}$ | 5.25 | 9.08 | $5.47\pm 0.99$ | 6.8 $g_{D^{*}\\!D\pi}$ | $17.24^{+3.06}_{-2.30}$ | 16.41 | 37.22 | $17.9\pm 0.3\pm 1.90$ | 3.7 Table 3: Strong couplings $g_{V\\!PP}$ for the decay channels $\rho\to\pi\pi$, $\phi\to\bar{K}K$, $K^{*}\to K\pi$ and $D^{*}\to D\pi$. The pair $(V,P)$ denotes the number of scalar amplitudes employed in the BSA of the vector and pseudoscalar mesons, respectively. The theoretical errors stem from the fit to Nakanishi representations of the BSA. The reference couplings are derived from the experimental decay widths ParticleDataGroup:2022pth via $\Gamma=g_{V\\!P_{1}\\!P_{2}}^{2}k^{3}/6\pi m_{V}^{2}$ with $k^{2}=[m_{V}^{2}-(m_{P_{1}}+m_{P_{2}})^{2}][m_{V}^{2}-(m_{P_{1}}-m_{P_{2}})^{2}]/4m_{V}^{2}$. The experimental $D^{*}\\!D\pi$ coupling was obtained by the CLEO collaboration CLEO:2001sxb . The relative deviations from experimental values are defined as $\epsilon_{g_{V\\!P\\!P}}=|g_{V\\!PP}^{\textrm{exp.}}-g_{V\\!PP}^{\textrm{th.}}|/g_{V\\!PP}^{\textrm{exp.}}$. When the relative momentum of the meson is complex, we choose the following analytic representation of the scalar amplitudes for $l=0,1$, $\mathcal{F}_{i}^{l}(k,P)=\sum_{j=1}^{N}\int_{-1}^{1}\\!d\alpha\,\rho_{j}(\alpha)\,\frac{U_{j}\Lambda^{2n_{j}}}{\Delta^{n_{j}}(k,\alpha,\Lambda)}\,,$ (52) where $\Delta=k^{2}+\alpha k\cdot P+\Lambda^{2}$ and the spectral density $\rho_{j}(\alpha)$ is given by, $\rho_{j}(\alpha)=\tfrac{1}{2}\\!\left(C^{\nicefrac{{1}}{{2}}}_{0}(\alpha)+\sigma_{j}C^{\nicefrac{{1}}{{2}}}_{2}(\alpha)\right)\,.$ (53) $C^{\nicefrac{{1}}{{2}}}_{0}(\alpha)$ and $C^{\nicefrac{{1}}{{2}}}_{2}(\alpha)$ are Gegenbauer polynomials of order $1/2$. The parameters $U_{j}$, $\Lambda$, $n_{j}$ and $\sigma_{j}$ are listed in Tables 4, 5 and 7 of Appendix A for the pion, kaon, $D$ and $D^{*}$ mesons. We use _real_ numerical BSA solutions for the remaining mesons considered in this work, namely the $\rho$, $\phi$ and $K^{*}$. This approach, frequently employed in calculations of distribution amplitudes Shi:2014uwa ; Serna:2020txe , differs from the method in Ref. Jarecke:2002xd based on a 2nd order Taylor expansion of $\mathcal{F}_{i}(k,P)$ about the closest point on the positive real axis to the complex-valued relative momenta $k$. We checked that the BSA parameterization of Eq. (52) produces the correct weak decay constants of the pseudoscalar and vector mesons. An illustration of the analytic continuation of the $D$-meson’s dominante amplitude with the parametrization of Eq. (52) is given in Fig. 3. With these technical considerations taken into account, we can take the trace and calculate the loop integral in Eq. (2) for different initial and final states in the strong two-body decay of a vector meson. The couplings, $g_{V\\!PP}$, we obtain are listed in Table 3, where the errors are due to the fit uncertainties of the Nakanishi representations. We remind that this error would increase if we included a systematic error of the ladder truncation, as explored in Ref. Serna:2020txe for example. More precisely, it is known that some typical observables, such as the pion and kaon masses and weak decay constants, are insensitive to a range $\omega_{i}\pm\Delta\omega_{i}$, $i=u,d,s$, of the interaction parameter in Eq. (41). Varying $\omega_{i}$ alters the BSA of the mesons and this can add to the uncertainty in the strong $V\to PP$ decay amplitude. However, modifying $\omega_{u,d}$ in the $D$ and $D^{*}$ mesons is a numerically delicate matter, as no solution of the BSA is found for the uncertainties $\pm\Delta\omega_{u,d}$. We therefore abstain from including this source of error. Figure 4: Comparison of theoretical values for $g_{D^{*}\\!D\pi}$ with the experimental coupling extracted from the $D^{*}$ decay width by the CLEO collaboration CLEO:2001sxb (shaded band). The couplings are taken from Ref. El-Bennich:2010uqs (DSE-BSE), Refs. Becirevic:2009xp and Abada:2002xe (LQCD), Ref. Bracco:2011pg (QCDSR) and Ref. Melikhov:2001zv (DQM). We consider the cases of limiting the BSA to the dominant covariant in Eq. (15), namely $\gamma_{5}$ and $\gamma_{\mu}$ for the pseudoscalar and vector mesons, respectively. Using merely the dominant amplitude, $g_{\rho\pi\pi}$ is 56% larger than the value obtained with the complete BSA. We find a difference of 102% for the $g_{\phi KK}$ coupling which increases to 116% for $g_{D^{*}\\!D\pi}$. Clearly, this leading approximation is not adequate even for couplings that involve only light quarks. On the other hand, including the next four leading covariants of the vector meson’s BSA, the dominant physics in the impulse approximation is captured and the couplings are within 1–5% of the values obtained with $(V,P)=(8,4)$. Our values for the strong couplings mostly agree with reference values, except for the $\rho\to\pi\pi$ coupling which is found to be the 13.6% smaller than the experimental coupling. As mentioned in Section II, this is expected given our limitation to the impulse and ladder approximation which omits intermediate $\pi\pi$, $\bar{K}K$ and $K\pi$ channels. Including explicit two- pion exchange in the BSE kernel, the decay width of the $\rho$-meson can be determined from the imaginary part of the resonance pole from which one deduces a coupling constant $g_{\rho\pi\pi}=5.7$ Williams:2018adr . The widths of the $\phi$, $K^{*}$ and $D^{*}$, on the other hand, are much smaller than their masses and our approximation is more accurate, though we still notice a deviation of 9.2% for $g_{\phi KK}$. As our focus is on the $D^{*}\to D\pi$ decay, we also compare our result with couplings obtained with lattice QCD (LQCD), QCD sum rules (QCDSR) and a dispersion relation quark model (DQM) in Fig. 4. Our calculation is a significant improvement on earlier work El-Bennich:2010uqs ; El- Bennich:2011tme ; El-Bennich:2012hom ; El-Bennich:2016bno which also considered the impulse approximation but employed model wave functions for the mesons, based on the dominant covariant term of the BSA, and a simplified, constant-mass propagator for the charm quark. This, as we noted in Table 3, has detrimental effects on translational invariance and the couplings depend on a _suitable_ choice of the partition parameters $w_{1}$ and $w_{2}$, see the discussion in Ref. El-Bennich:2010uqs . Since our calculation is fully Poincaré covariant, our decay amplitudes are independent of the momentum distribution, as we verified with variations of $w_{1}$ and $w_{2}$ up to a critical limit where we encounter singularities in the quark propagators on the complex plane. ## VI Conclusion We revisited the strong decays of vector mesons into two pseudoscalar mesons within the framework of the DSE and BSE, having in mind the particular decay $D^{*}\to D\pi$. As we argued, these decays are the simplest hadronic observables beyond the meson’s masses, weak decay constants and electromagnetic form factors, and thus provide additional information about the dynamics of QCD in the nonperturbative regime. In particular, the vector- meson decay to a pair of pseudoscalars proceeds via a $P$-wave interaction and therefore involves the BSA differently than the meson’s weak decay constants. The strong $D^{*}$ decay is then even more interesting, as it probes nonperturbative QCD simultaneously at two distinct scales, namely the light- and charm-quark masses. We limited ourselves to the impulse approximation for the aforementioned reason: our BSE kernel in ladder truncation is too simple to include $\pi\pi$, $\bar{K}K$ and $K\pi$ channels in these decays and this is most likely the largest source of error in our calculation of the $\rho\to\pi\pi$ coupling. Nonetheless, this calculation represents an important theoretical and numerical improvement over the simpler approaches in Ref El-Bennich:2010uqs ; El-Bennich:2011tme ; El-Bennich:2012hom ; El-Bennich:2016bno , as the full Poncaré invariant BSA structure of all mesons is included and the quark propagators are calculated on the complex momentum plane for all flavors. This present calculation can also serve as a guidance to reevaluate off-shell space-like couplings between the $\rho$ and $D$\- and $D^{*}$ mesons in Ref. El-Bennich:2016bno without resorting to model wave functions. Our final value for $g_{D^{*}\\!D\pi}$ is 3.7% lower than that extracted from the experimental decay width and well within the experimental errors. It corresponds to a universal coupling in a chiral heavy meson Lagrangian which at leading order in the heavy-mass expansion is given by: $\hat{g}=\frac{g_{D^{*}\\!D\pi}}{2\sqrt{m_{D}^{*}m_{D}}}\,f_{\pi}=0.58^{+0.10}_{-0.08}\ .$ (54) A consensus seems to be growing that the most recent theoretical couplings are in good agreement with the CLEO value CLEO:2001sxb extracted from the $D^{*}$ decay width. Future improvements ought to consider strong $\pi\pi$ interactions, likely along the lines presented in Refs. Williams:2018adr ; Miramontes:2019mco ; Miramontes:2021xgn , in the BSE kernels and to go beyond the impulse approximation. ###### Acknowledgements. B.E., F.E.S. and R.C.S. participate in the Brazilian network project _INCT- Física Nuclear e Aplicações_ , no. 464898/2014-5. This work was supported by the São Paulo Research Foundation (FAPESP), grant no. 2018/20218-4, and by the National Council for Scientific and Technological Development (CNPq), grant no. 428003/2018-4. F.E.S. is a CAPES-PNPD postdoctoral fellow financed by grant no. 88882.314890/2013-01. We appreciated helpful communication with Peter Tandy. ## Appendix A Parameters of Bethe-Salpeter Amplitude Representation $\mathcal{F}_{i}^{l}$ | $\Lambda$ | $U_{1}$ | $U_{2}$ | $U_{3}$ | $\sigma_{1}$ | $\sigma_{2}$ | $\sigma_{3}$ | $n_{1}$ | $n_{2}$ | $n_{3}$ ---|---|---|---|---|---|---|---|---|---|--- $E_{\pi}^{0}$ | 1.280 | 2.558 | $-1.559$ | 0.0 | 1.810 | 1.548 | 0.0 | 4 | 5 | 0 $F_{\pi}^{0}$ | 1.150 | 1.838 | $-0.948$ | $-0.381$ | $-2.679$ | $-2.547$ | $-5.107$ | 4 | 5 | 3 $G_{\pi}^{0}$ | 1.106 | 2.402 | $-1.950$ | 0.0 | $-0.4590$ | $-0.474$ | 0.0 | 6 | 7 | 0 $H_{\pi}^{0}$ | 1.056 | 1.253 | $-0.857$ | $-0.140$ | $-0.696$ | $-0.634$ | $-2.663$ | 5 | 6 | 3 Table 4: Parameters of the BSA representation in Eq. (52) for the pion. In the isospin limit, $m_{u}=m_{d}$, only $\mathcal{F}_{i}^{0}$ contributes which we fit to the sum of the 0th and 2nd Chebyshev moments. $\mathcal{F}_{i}^{l}$ | $\Lambda$ | $U_{1}$ | $U_{2}$ | $U_{3}$ | $\sigma_{1}$ | $\sigma_{2}$ | $\sigma_{3}$ | $n_{1}$ | $n_{2}$ | $n_{3}$ ---|---|---|---|---|---|---|---|---|---|--- $E^{0}$ | 1.557 | 2.590 | $-1.590$ | 0.0 | 1.342 | 0.891 | 0.0 | 5 | 6 | 0 $E^{1}$ | 1.495 | $-0.810$ | 3.769 | $-2.251$ | $-1.039$ | $-0.680$ | $-0.680$ | 5 | 6 | 7 $F^{0}$ | 1.514 | 2.756 | $-3.558$ | 1.220 | $-0.527$ | $-0.173$ | 0.323 | 7 | 8 | 9 $F^{1}$ | 1.604 | 3.150 | $-5.480$ | 2.537 | $-1.074$ | $-0.881$ | $-0.718$ | 10 | 11 | 12 $G^{0}$ | 1.631 | $-0.613$ | 1.139 | $-0.522$ | 0.595 | 2.134 | 2.651 | 8 | 10 | 12 $G^{1}$ | 1.229 | 2.949 | $-4.542$ | 1.880 | $-0.558$ | $-0.747$ | $-0.920$ | 8 | 10 | 12 $H^{0}$ | 1.727 | $-0.276$ | 1.031 | $-0.594$ | 1.722 | 0.661 | 0.303 | 8 | 10 | 12 $H^{1}$ | 1.443 | $-1.564$ | 4.434 | $-2.663$ | $-0.8104$ | $-0.905$ | $-0.938$ | 8 | 9 | 10 Table 5: Parameters of the BSA representation in Eq. (52) for the kaon. Both, the $\mathcal{F}_{i}^{0}$ and $\mathcal{F}_{i}^{1}$ amplitudes contribute, as the kaon is not an eigenstate of charge conjugation. In the fit we include the sum of the 1st and 3rd Chebyshev moments in the odd component of the BSA. $\mathcal{F}_{i}^{l}$ | $\Lambda$ | $U_{1}$ | $U_{2}$ | $U_{3}$ | $\sigma_{1}$ | $\sigma_{2}$ | $\sigma_{3}$ | $n_{1}$ | $n_{2}$ | $n_{3}$ ---|---|---|---|---|---|---|---|---|---|--- $E^{0}$ | 1.750 | 2.078 | $-1.077$ | 0.0 | $-1.274$ | $-1.126$ | 0.0 | 5 | 6 | 0 $E^{1}$ | 2.146 | $-0.207$ | 0.209 | 0.0 | $-1.115$ | $-1.115$ | 0.0 | 6 | 9 | 0 $F^{0}$ | 2.222 | 0.060 | 0.155 | 0.0 | $-0.934$ | $-2.211$ | 0.0 | 6 | 9 | 0 $F^{1}$ | 2.583 | 0.003 | $-0.043$ | 0.0 | $-1.372$ | $-1.757$ | 0.0 | 6 | 9 | 0 $G^{0}$ | 1.543 | $-1.596$ | 2.008 | $-0.427$ | $-1.439$ | $-1.596$ | $-1.795$ | 6 | 7 | 10 $G^{1}$ | 1.423 | 0.197 | 0.201 | $-0.233$ | $-1.289$ | $-2.657$ | $-2.097$ | 5 | 6 | 9 $H^{0}$ | 1.711 | $-0.249$ | 0.868 | $-0.530$ | $-1.535$ | $-1.618$ | $-1.621$ | 8 | 9 | 10 $H^{1}$ | 1.155 | $-0.481$ | 0.937 | $-0.470$ | $-1.917$ | $-2.011$ | $-2.086$ | 6 | 7 | 8 Table 6: Parameters of the BSA representation in Eq. (52) for the $D$ meson. Even and odd components of the BSA are in terms of Chebyshev moments as described in Tables 4 and 5. $\mathcal{F}_{i}^{l}$ | $\Lambda$ | $U_{1}$ | $U_{2}$ | $U_{3}$ | $\sigma_{1}$ | $\sigma_{2}$ | $\sigma_{3}$ | $n_{1}$ | $n_{2}$ | $n_{3}$ ---|---|---|---|---|---|---|---|---|---|--- $\mathcal{F}_{1}^{0}$ | 1.942 | 3.510 | $-2.509$ | 0.0 | $-1.666$ | $-1.851$ | 0.0 | 7 | 8 | 0 $\mathcal{F}_{1}^{1}$ | 0.998 | 2.441 | $-1.533$ | $-0.680$ | $-1.856$ | $-2.027$ | $-1.167$ | 4 | 5 | 3 $\mathcal{F}_{2}^{0}$ | 1.107 | 0.440 | -0.216 | 0.0 | $-1.656$ | $-1.312$ | 0.0 | 4 | 6 | 0 $\mathcal{F}_{2}^{1}$ | 0.902 | 2.493 | $-3.754$ | 1.278 | $-2.500$ | $-2.522$ | $-2.550$ | 7 | 8 | 10 $\mathcal{F}_{3}^{0}$ | 1.445 | $-1.536$ | 1.372 | 0.0 | $-1.838$ | $-1.831$ | 0.0 | 7 | 8 | 0 $\mathcal{F}_{3}^{1}$ | 1.005 | 0.099 | $-0.032$ | 0.0 | $-1.030$ | $-4.365$ | 0.0 | 3 | 9 | 0 $\mathcal{F}_{4}^{0}$ | 1.411 | $-1.235$ | 0.837 | 0.0 | $-1.844$ | $-1.833$ | 0.0 | 5 | 6 | 0 $\mathcal{F}_{4}^{1}$ | 1.554 | 1.192 | $-2.367$ | 1.181 | $-1.949$ | $-2.040$ | $-2.10$ | 7 | 8 | 9 $\mathcal{F}_{5}^{0}$ | 1.895 | $-4.279$ | 2.958 | 0.0 | $-1.775$ | $-1.787$ | 0.0 | 7 | 8 | 0 $\mathcal{F}_{5}^{1}$ | 1.730 | 2.657 | $-4.797$ | 2.212 | $-1.831$ | $-1.932$ | $-2.00$ | 7 | 8 | 9 $\mathcal{F}_{6}^{0}$ | 1.384 | 0.815 | $-0.639$ | 0.0 | $-1.916$ | $-1.942$ | 0.0 | 6 | 8 | 0 $\mathcal{F}_{6}^{1}$ | 1.250 | $-3.728$ | 2.097 | 0.0 | $-2.229$ | $-3.453$ | 0.0 | 7 | 12 | 0 $\mathcal{F}_{7}^{0}$ | 1.183 | $-0.557$ | 0.971 | $-0.412$ | $-0.711$ | $-1.517$ | $-1.927$ | 4 | 5 | 6 $\mathcal{F}_{7}^{1}$ | 1.091 | $-0.190$ | 0.076 | 0.080 | $-1.453$ | $-2.005$ | $-0.520$ | 4 | 8 | 3 $\mathcal{F}_{8}^{0}$ | 1.316 | 0.601 | $-0.746$ | 0.147 | $-1.917$ | $-2.011$ | $-2.086$ | 4 | 5 | 8 $\mathcal{F}_{8}^{1}$ | 0.909 | 0.073 | $-0.647$ | 0.501 | $-2.127$ | $-2.312$ | $-2.331$ | 4 | 5 | 6 Table 7: Parameters of the BSA representation in Eq. (52) for the $D^{*}$ meson. Even and odd components of the BSA are in terms of Chebyshev moments as described in Tables 4 and 5. ## References * (1) P. Maris and P. C. Tandy, Phys. Rev. C 60 (1999), 055214 doi:10.1103/PhysRevC.60.055214 * (2) P. Maris and P. C. Tandy, Phys. Rev. C 62 (2000), 055204 doi:10.1103/PhysRevC.62.055204 * (3) J. P. B. C. de Melo and T. Frederico, Phys. Rev. C 55 (1997), 2043 doi:10.1103/PhysRevC.55.2043 * (4) B. L. G. Bakker, H. M. Choi and C. R. Ji, Phys. Rev. D 65 (2002), 116001 doi:10.1103/PhysRevD.65.116001 * (5) J. P. B. C. de Melo and T. Frederico, Phys. Lett. B 708 (2012), 87-92 doi:10.1016/j.physletb.2012.01.021 * (6) J. P. B. C. De Melo, Phys. Lett. B 788 (2019), 152-160 doi:10.1016/j.physletb.2018.11.003 * (7) J. P. B. C. de Melo, K. Tsushima, B. El-Bennich, E. Rojas and T. Frederico, Phys. Rev. C 90 (2014) no.3, 035201 doi:10.1103/PhysRevC.90.035201 * (8) L. Chang, I. C. Cloët, C. D. Roberts, S. M. Schmidt and P. C. Tandy, Phys. Rev. Lett. 111 (2013) no.14, 141802 doi:10.1103/PhysRevLett.111.141802 * (9) K. Raya, L. Chang, A. Bashir, J. J. Cobos-Martínez, L. X. Gutiérrez-Guerrero, C. D. Roberts and P. C. Tandy, Phys. Rev. D 93 (2016) no.7, 074017 doi:10.1103/PhysRevD.93.074017 * (10) M. Ding, K. Raya, A. Bashir, D. Binosi, L. Chang, M. Chen and C. D. Roberts, Phys. Rev. D 99 (2019) no.1, 014014 doi:10.1103/PhysRevD.99.014014 * (11) Y. Z. Xu, D. Binosi, Z. F. Cui, B. L. Li, C. D. Roberts, S. S. Xu and H. S. Zong, Phys. Rev. D 100 (2019) no.11, 114038 doi:10.1103/PhysRevD.100.114038 * (12) E. O. da Silva, J. P. B. C. de Melo, B. El-Bennich and V. S. Filho, Phys. Rev. C 86 (2012), 038202 doi:10.1103/PhysRevC.86.038202 * (13) B. El-Bennich, J. P. B. C. de Melo and T. Frederico, Few Body Syst. 54 (2013), 1851-1863 doi:10.1007/s00601-013-0682-5 * (14) B. El-Bennich, J. P. B. C. de Melo, B. Loiseau, J. P. Dedonder and T. Frederico, Braz. J. Phys. 38 (2008), 465-471 doi:10.1590/S0103-97332008000400016 * (15) M. A. Ivanov, J. G. Körner, S. G. Kovalenko and C. D. Roberts, Phys. Rev. D 76 (2007), 034018 doi:10.1103/PhysRevD.76.034018 * (16) B. El-Bennich, M. A. Ivanov and C. D. Roberts, Nucl. Phys. B Proc. Suppl. 199 (2010), 184-190 doi:10.1016/j.nuclphysbps.2010.02.026 * (17) B. El-Bennich, M. A. Ivanov and C. D. Roberts, Phys. Rev. C 83 (2011), 025205 doi:10.1103/PhysRevC.83.025205 * (18) B. El-Bennich, G. Krein, L. Chang, C. D. Roberts and D. J. Wilson, Phys. Rev. D 85 (2012), 031502 doi:10.1103/PhysRevD.85.031502 * (19) B. El-Bennich, C. D. Roberts and M. A. Ivanov, PoS QCD-TNT-II (2012), 018 doi:10.22323/1.136.0018 * (20) B. El-Bennich, M. A. Paracha, C. D. Roberts and E. Rojas, Phys. Rev. D 95 (2017) no.3, 034037 doi:10.1103/PhysRevD.95.034037 * (21) A. Anastassov et al. [CLEO], Phys. Rev. D 65 (2002), 032003 doi:10.1103/PhysRevD.65.032003 * (22) R. Casalbuoni, A. Deandrea, N. Di Bartolomeo, R. Gatto, F. Feruglio and G. Nardulli, Phys. Rept. 281 (1997), 145-238 doi:10.1016/S0370-1573(96)00027-0 * (23) B. El-Bennich and F. E. Serna, PoS CHARM2020 (2021), 025 doi:10.22323/1.385.0025 * (24) F. L. Braghin, Phys. Rev. D 105 (2022) no.5, 054009 doi:10.1103/PhysRevD.105.054009 * (25) D. Jarecke, P. Maris and P. C. Tandy, Phys. Rev. C 67 (2003), 035202 doi:10.1103/PhysRevC.67.035202 * (26) R. Williams, Phys. Lett. B 798 (2019), 134943 doi:10.1016/j.physletb.2019.134943 * (27) Á. S. Miramontes and H. Sanchis-Alepuz, Eur. Phys. J. A 55 (2019) no.10, 170 doi:10.1140/epja/i2019-12847-6 * (28) Á. S. Miramontes, H. Sanchis-Alepuz and R. Alkofer, Phys. Rev. D 103 (2021) no.11, 116006 doi:10.1103/PhysRevD.103.116006 * (29) M. E. Bracco, M. Chiapparini, F. S. Navarra and M. Nielsen, Prog. Part. Nucl. Phys. 67 (2012), 1019-1052 doi:10.1016/j.ppnp.2012.03.002 * (30) A. Bashir, L. Chang, I. C. Cloët, B. El-Bennich, Y. X. Liu, C. D. Roberts and P. C. Tandy, Commun. Theor. Phys. 58 (2012), 79-134 doi:10.1088/0253-6102/58/1/16 * (31) F. Gao, L. Chang, Y. X. Liu, C. D. Roberts and S. M. Schmidt, Phys. Rev. D 90 (2014) no.1, 014011 doi:10.1103/PhysRevD.90.014011 * (32) N. Nakanishi, Phys. Rev. 138 (1965), B1182-B1192 doi:10.1103/PhysRev.138.B1182 * (33) N. Nakanishi, Phys. Rev. 139 (1965), B1401-B1406 doi:10.1103/PhysRev.139.B1401 * (34) L. Albino, A. Bashir, L. X. Gutiérrez-Guerrero, B. EL Bennich and E. Rojas, Phys. Rev. D 100 (2019) no.5, 054028 doi:10.1103/PhysRevD.100.054028 * (35) L. Albino, A. Bashir, B. El-Bennich, E. Rojas, F. E. Serna and R. C. da Silveira, JHEP 11 (2021), 196 doi:10.1007/JHEP11(2021)196 * (36) B. El-Bennich, F. E. Serna, R. C. da Silveira, L. A. F. Rangel, A. Bashir and E. Rojas, Rev. Mex. Fis. Suppl. 3 (2022) no.3, 0308092 doi:10.31349/SuplRevMexFis.3.0308092 * (37) F. E. Serna, C. Chen and B. El-Bennich, Phys. Rev. D 99 (2019) no.9, 094027 doi:10.1103/PhysRevD.99.094027 * (38) J. C. R. Bloch, Phys. Rev. D 66 (2002), 034032 doi:10.1103/PhysRevD.66.034032 * (39) S. x. Qin, L. Chang, Y. x. Liu, C. D. Roberts and D. J. Wilson, Phys. Rev. C 84 (2011), 042202 doi:10.1103/PhysRevC.84.042202 * (40) F. E. Serna, B. El-Bennich and G. Krein, Phys. Rev. D 96 (2017) no.1, 014013 doi:10.1103/PhysRevD.96.014013 * (41) M. Chen and L. Chang, Chin. Phys. C 43 (2019) no.11, 114103 doi:10.1088/1674-1137/43/11/114103 * (42) F. E. Serna, R. C. da Silveira, J. J. Cobos-Martínez, B. El-Bennich and E. Rojas, Eur. Phys. J. C 80 (2020) no.10, 955 doi:10.1140/epjc/s10052-020-08517-3 * (43) F. E. Serna, R. C. da Silveira and B. El-Bennich, Phys. Rev. D 106 (2022), L091504 doi.org/10.1103/PhysRevD.106.L091504 * (44) E. Rojas, B. El-Bennich and J. P. B. C. de Melo, Phys. Rev. D 90 (2014), 074025 doi:10.1103/PhysRevD.90.074025 * (45) F. F. Mojica, C. E. Vera, E. Rojas and B. El-Bennich, Phys. Rev. D 96 (2017) no.1, 014012 doi:10.1103/PhysRevD.96.014012 * (46) P. Qin, S. x. Qin and Y. x. Liu, Phys. Rev. D 101 (2020) no.11, 114014 doi:10.1103/PhysRevD.101.114014 * (47) A. Krassnigg, PoS CONFINEMENT8 (2008), 075 doi:10.22323/1.077.0075 * (48) B. El-Bennich, G. Krein, E. Rojas and F. E. Serna, Few Body Syst. 57 (2016) no.10, 955-963 doi:10.1007/s00601-016-1133-x * (49) B. El-Bennich and E. Rojas, EPJ Web Conf. 113 (2016), 05003 doi:10.1051/epjconf/201611305003 * (50) B. El-Bennich, EPJ Web Conf. 172 (2018), 02005 doi:10.1051/epjconf/201817202005 * (51) R. B. Lehoucq, D. C. Sorensen, C. Yang, _ARPACK Users’ Guide: Solution of Large-Scale Eigenvalue Problems with Implicitly Restarted Arnoldi Methods_ (Society for Industrial & Applied Mathematics, 1998). * (52) M. Blank and A. Krassnigg, Comput. Phys. Commun. 182 (2011), 1391-1401 doi:10.1016/j.cpc.2011.03.003 * (53) R. L. Workman et al. [Particle Data Group], PTEP 2022 (2022), 083C01 doi:10.1093/ptep/ptac097 * (54) V. Lubicz et al. [ETM], Phys. Rev. D 96 (2017) no.3, 034524 doi:10.1103/PhysRevD.96.034524 * (55) G. Eichmann, E. Ferreira and A. Stadler, Phys. Rev. D 105 (2022) no.3, 034009 doi:10.1103/PhysRevD.105.034009 * (56) C. Shi, L. Chang, C. D. Roberts, S. M. Schmidt, P. C. Tandy and H. S. Zong, Phys. Lett. B 738 (2014), 512-518 doi:10.1016/j.physletb.2014.07.057 * (57) D. Bećirević and B. Haas, Eur. Phys. J. C 71 (2011), 1734 doi:10.1140/epjc/s10052-011-1734-y * (58) A. Abada, D. Bećirević, P. Boucaud, G. Herdoiza, J. P. Leroy, A. Le Yaouanc, O. Pène and J. Rodríguez-Quintero, Phys. Rev. D 66 (2002), 074504 doi:10.1103/PhysRevD.66.074504 * (59) D. Melikhov, Eur. Phys. J. direct 4 (2002) no.1, 2 doi:10.1007/s1010502c0002
# Machine learning potentials from transfer learning of periodic correlated electronic structure methods: Application to liquid water with AFQMC, CCSD, and CCSD(T) Michael S. Chen Department of Chemistry, Stanford University, Stanford, California, 94305, USA Joonho Lee Department of Chemistry, Columbia University, New York, New York 10027, USA Hong-Zhou Ye Department of Chemistry, Columbia University, New York, New York 10027, USA Timothy C. Berkelbach<EMAIL_ADDRESS>Department of Chemistry, Columbia University, New York, New York 10027, USA Center for Computational Quantum Physics, Flatiron Institute, New York, New York 10010, USA David R. Reichman <EMAIL_ADDRESS>Department of Chemistry, Columbia University, New York, New York 10027, USA Thomas E. Markland<EMAIL_ADDRESS>Department of Chemistry, Stanford University, Stanford, California, 94305, USA ###### Abstract Obtaining the atomistic structure and dynamics of disordered condensed phase systems from first principles remains one of the forefront challenges of chemical theory. Here we exploit recent advances in periodic electronic structure to show that, by leveraging transfer learning starting from lower tier electronic structure methods, one can obtain machine learned potential energy surfaces for liquid water from the higher tier AFQMC, CCSD, and CCSD(T) approaches using $\leq$200 energies. By performing both classical and path integral molecular dynamics simulations on these machine learned potential energy surfaces we uncover the interplay of dynamical electron correlation and nuclear quantum effects across the entire liquid range of water while providing a general strategy for efficiently utilizing periodic correlated electronic structure methods to explore disordered condensed phase systems. ## I Introduction Ab initio molecular dynamics (AIMD) simulations, where the forces and energies are generated at each time-step by performing an electronic structure calculation, provide an appealing route to simulate reactive chemical dynamics. However, for disordered condensed phase systems an accurate description typically requires many 100’s of atoms to obtain a chemically reasonable description of bulk systems (e.g., water) and this grows into the 1000’s for more heterogeneous systems (e.g., those with interfaces). Since AIMD simulations require an electronic structure calculation to be performed at each time-step, to statistically converge even simple thermodynamic properties necessitates many tens of thousands of ab initio calculations (10’s of picosecond timescale at a $\sim$1 fs time step) and for slower converging properties many millions are needed (nanosecond timescale). The computational expense of these simulations is further compounded if one wants to incorporate nuclear quantum effects (NQE) via ab initio path integral molecular dynamics simulations (PIMD). Due to its reasonable compromise between accuracy and efficiency, density functional theory (DFT) is currently the most frequently employed electronic structure method in condensed phase AIMD studies. However, the results depend—sometimes sensitively—on the choice of the exchange- correlation functional and the inclusion of dispersion corrections[1, 2]. This issue motivates the use of beyond-DFT electronic structure theories, such as those based on the many-electron wavefunction. For example, work performed almost a decade ago demonstrated the AIMD simulations of liquid water using second-order Möller-Plesset perturbation theory (MP2) [3]. However, the high cost of more accurate methods precludes their direct use in AIMD. Machine learned potentials (MLPs) have emerged as an extremely promising approach to accurately model ab initio potential energy surfaces of condensed phase systems while being orders of magnitude more computationally efficient to evaluate. For liquid water, MLPs have been successfully developed at various levels of electronic structure ranging from different levels of DFT[4, 5, 6, 7, 8, 9, 10] to more recently using the random phase approximation (RPA)[11] and MP2[12, 13]. The modeling of liquid water and other molecular systems with more accurate electronic structure methods, such as coupled- cluster theory or quantum Monte Carlo, has been so far limited to training on finite clusters of molecules[14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25]. Recent advances have opened the door to efficiently calculating the properties of periodic condensed phase systems using higher-level methods like coupled cluster singles and doubles without and with perturbative triples (CCSD and CCSD(T))[26, 27] and phaseless auxiliary-field quantum Monte Carlo (AFQMC)[28]. However, while these advances allow the energies of many 100’s of condensed phase configurations to be evaluated, this is still considerably less than what would be typically required to train an accurate MLP. Here, we demonstrate that by using an approach based on transfer learning starting from a variety of lower tier electronic structure methods one can generate a highly data-efficient approach to training MLPs using high level electronic structure methods. With this approach, we show that only 200 high- quality energies obtained from small periodic boxes containing 16 water molecules provide sufficient data for training our MLPs. Specifically, we train MLPs to periodic electronic structure calculations performed with AFQMC, CCSD, and CCSD(T). The MLPs are then used to perform AIMD and PIMD simulations of larger water boxes for the long times necessary to statistically converge static and dynamic properties using both classical and quantum mechanical treatment of nuclei. This allows us to achieve a careful comparison of the quality of the underlying electronic structure theories in describing water across its entire liquid temperature range and uncover the changes in the properties of liquid water as the dynamical electron correlation captured is increased. In addition, we provide a set of MLPs and a curated training set that can form the basis of future studies of water and aqueous systems. Our data-efficient approach provides a route to accurately obtain the properties of other disordered condensed-phase systems by combining high-level electronic structure theory and machine learning. ## II Methods ### II.1 Machine learning To develop a highly data efficient strategy that only requires a minimal number of energies from periodic high level electronic structure methods, such as CCSD, CCSD(T), and AFQMC, and produce accurate MLPs, we exploited a combination of approaches including an active learning procedure for curating a small but comprehensive training set, the training of MLPs to energies for small periodic boxes (16 waters) and showing that they reproduce properties when used for larger simulations (64 waters) and a transfer learning approach that leverages the transferability of physics from lower tier electronic structure methods. #### II.1.1 Developing an efficient training set by leveraging a committee of machine learning potentials We first applied an active learning procedure to curate a data efficient set of configurations to train our MLPs. We invoked the commonly employed Query- by-Committee (QbC)[29, 30, 8] approach to iteratively add configurations to the training set. At every iteration, the current dataset of configuration energies was used to train a committee of 8 MLPs, each a Behler-Parrinello neural network potential[31, 4] with a different random initialization of weights and random train-validation (90-10) splits of the dataset. To mitigate overfitting, we applied early stopping to each of these MLPs in the committee, monitoring the energy prediction error over a validation set. The mean potential energy surface obtained for this committee MLP was used to run a short MD simulation (SI Sec. LABEL:sec:si-md-details) that was terminated either when the system becomes unphysical (i.e., reaching a temperature greater than 400K) or when 50 ps of simulation trajectory was generated. We then selected the 10 configurations where the committee MLP had the largest standard deviation in its potential energy prediction to recalculate at the target level of electronic structure theory and add to the training set for the next iteration of this procedure. To prevent selected configurations from being overly correlated with one another, no two selected configurations were closer than 100 fs apart. The initial dataset of 50 configurations used to initialize this procedure was also selected via an iterative QbC procedure, as applied to a single 100 ps SCC-DFTB[32] (SI Sec. LABEL:sec:si-md-details) trajectory, that started with 10 randomly selected configurations and added 10 additional configurations for each of 4 subsequent iterations. We generated 5 different 200 configuration datasets by running 5 instances of our QbC active learning scheme where the target level of electronic structure theory was DFT using the revPBE0-D3[33, 34, 35] functional (SI Sec. LABEL:sec:si-dft-hf-dftb) which was chosen due to being computationally efficient compared to the high level electronic structure methods and since it has previously been shown to produce the properties of liquid water accurately when combined with path integral simulations[36]. To select the generated dataset on which the higher tier electronic structure methods would be used we then used each of the 5 datasets to train a committee MLP using revPBE0-D3. We then selected the one that gave the lowest force prediction error (RMSE) when evaluated on a test set of 1000 water configurations (64 water molecules) drawn from previously published AIMD simulations[36]. To check the transferability of the selected 200 configuration dataset, we recalculated the energies and forces for the generated datasets and the test set using the BLYP[37, 38] functional (SI Sec. LABEL:sec:si-dft-hf-dftb). After training and evaluating a new set of BLYP trained MLPs, the same dataset that resulted in the lowest error revPBE0-D3 trained MLP also gave the lowest error BLYP trained MLP (SI Fig. LABEL:fig:trainingset_force_lc). Given the transferability of the relative utility of this dataset, we used this same 200 configuration dataset to train our CCSD, CCSD(T), and AFQMC MLPs. #### II.1.2 Training machine learned potentials on configurations of small systems Figure 1: Oxygen-oxygen RDFs for liquid water when running classical and PIMD simulations using NNPs fitted to either CCSD, CCSD(T), and AFQMC energies at both 300K and 370K. The PIMD oxygen-oxygen RDFs when using the CCSD(T) and AFQMC correspond closely with the experimental results shaded in grey at both 295K and 366K. Given the significant scaling of CCSD, CCSD(T), and AFQMC’s computational cost with system size, we sought to reduce the system size of the configurations used in our training set. We investigated the feasibility of training our committee MLP to the energies of a set configurations for small periodic boxes to accurately predict for properties associated with a larger simulation box. For liquid water running periodic molecular dynamics simulations of small water boxes (32 water molecules or fewer) leads to significant artifacts even in simple properties such as radial distribution functions (RDFs) when compared to larger system sizes[39]. However, as demonstrated in SI Sec. LABEL:sec:si-revpbe0-refit-tests for the revPBE0-D3 functional, if one trains a MLP on energies of small (16 molecule) periodic water configurations and then uses the resulting model to perform dynamics of a larger system (64 molecule), then the results obtained are in excellent agreement with those obtained from performing a AIMD simulation at the larger system size. Based on this demonstration we therefore trained the MLPs for the higher level methods on periodic configurations of 16 water molecules using the transfer learning approach described in the following section and then report the properties obtained in Sec. III by performing MD and PIMD simulations using 64 water molecules. #### II.1.3 Transfer learning approach to train the MLPs With only hundreds of energies from the higher tier periodic electronic structure methods available, to make efficient use of this data we employed a transfer learning approach[40]. To achieve this we first trained a committee MLP at a lower level of electronic structure theory, DFT or Hartree Fock (HF) on 531 configurations using both energies and gradients to improve the fitting. The parameters obtained for these MLPs were then used as the starting point for the fits to the higher level methods. This strategy exploits the idea that while the lower level methods may not reach the levels of chemical accuracy required for some applications they do contain fundamental physical information (e.g., about the fact that O and H when in close proximity form a high frequency chemical bond) that can be used to structure the neural network that underlies the MLP. Hence while the high level data is used to tune the accuracy of the MLP, it is leveraging the physics learnt by the initial training to the lower level method. While using transfer learning to make efficient use of very small amounts of high level electronic structure data has distinct advantages over starting from a randomly initialized MLP, one must be mindful of the risk of hysteresis in the final MLP, i.e., that by biasing the weights by taking them from a model trained on a low level electronic structure method, that the final MLP will incorrectly contain remnants of the failures of the low level method. Hence to assess our transfer learning approach’s ability to produce a final model that accurately reproduces the properties of the target high level electronic structure method, we initialized the procedure starting from a range of different low level methods that each give very different properties of liquid water. By comparing the structural and dynamical properties of liquid water obtained by performing MD and PIMD simulations on the final models obtained from these different starting points, one can thus evaluate which properties are obtained universally across the models initialized from different low level methods and thus accurately reflect the high level electronic structure approach. In practice, following the convergence of the lower level fits to both energy and gradient data (see SI Sec. LABEL:sec:si-ml-details), we retrained the committee MLP (8 separate MLPs) to the energies of the higher level method employing the extended Kalman filter optimizer as implemented in the n2p2[41] package and using a 90-10 train-validation split in order to monitor the prediction error over the validation set for early-stopping each individual fit to prevent overfitting. Before applying this transfer learning approach to the high level CCSD, CCSD(T) and AFQMC methods for which AIMD is not possible for the timescales required, in SI Sec. LABEL:sec:si-revpbe0-refit-tests we tested it by training an MLP with revPBE0-D3 as the higher level method and BLYP and HF as the lower level methods from which the transfer learning was performed. We chose DFT with the revPBE0-D3 exchange-correlation functional as the higher level in this benchmark due to its accurate description of the properties of liquid water and since we can compare the results of the transfer learned MLPs to AIMD and AI-PIMD trajectories that we have previously obtained [36]. We chose BLYP and HF as the lower level methods since the former gives an incredibly overstructured and dynamically sluggish description of water while the latter gives the opposite. As shown in SI Sec. LABEL:sec:si-revpbe0-refit-tests, MD simulations using the final transfer learned MLP models of revPBE0-D3 using only 200 energies starting from either BLYP or HF correctly capture the target RDFs and VDOS for liquid water at 300K obtained from AIMD simulations, which are both markedly different from those given by low level methods themselves (BLYP water has a more structured oxygen-oxygen RDF and higher wavenumber hydrogen VDOS bend and stretch peak positions than revPBE0-D3 water, and vice-versa for HF water). The agreement with the reference revPBE0-D3 results for both models is particularly strong for the RDFs. For the VDOS, the BLYP-initialized model outperforms the HF-initialized model in capturing the high frequency O-H stretch peak. Both models also accurately reproduce the RDFs obtained from AI- PIMD simulations of revPBE0-D3 for liquid water[36] at 300K (SI Fig. LABEL:fig:revpbe0-D3_refits_rdfs_pimd_300K) with the only discrepancy again being in the VDOS (SI Fig. LABEL:fig:revpbe0-D3_refits_vdos_pimd_300K_370K) where the BLYP-initialized MLP captures the full spectrum whereas HF shows an overstructured and blue shifted OH stretch region. Additionally, we compared our transfer learning approach to two common alternative machine learning approaches: directly training a committee MLP on the high level energies starting from randomly initialized weights and training a committee delta-learning model that corrects from the lower to higher level method. To allow for a fair comparison, all three types of models were trained to the same 200 configuration training set of energies (revPBE0-D3) for liquid water using the same MLP architecture and optimization settings. The randomly initialized model resulted in a potential that was unstable for the purposes of running MD for a periodic simulation box of 64 water molecules, with the instantaneous temperature drifting severely after the first simulation step. For the delta-learning model, we trained a committee MLP to capture the energy difference between revPBE0-D3 and another committee MLP trained to BLYP. The delta-learning model was similarly as unstable as the randomly initialized model. Hence for liquid water with 200 energies at the higher level we found that the transfer learning approach we have detailed above provides the most accurate results. Given the demonstrated efficacy of our transfer learning procedure, in Sec. III we applied this approach to train committee MLPs to CCSD, CCSD(T), and AFQMC energies for liquid water using three different sets of MLP initializations: HF, BLYP, and revPBE0-D3. ### II.2 Correlated electronic structure methods We perform correlated electronic structure calculations of liquid water with periodic boundary conditions at the $\Gamma$-point using AFQMC, CCSD, and CCSD(T). Periodic CCSD and CCSD(T) calculations were performed using PySCF [42, 43] where electron-repulsion integrals are handled using the range- separated density fitting method [44, 45], and AFQMC calculations were performed using QMCPACK[46, 47] and ipie[48]. These calculations all began with a periodic spin-restricted HF calculation also perfomed using PySCF. We provide brief details here, and further information can be found in SI Secs. LABEL:sec:si-ccsd-ccsdt and LABEL:sec:si-afqmc. AFQMC is a projector Monte Carlo method where the ground state of a given Hamiltonian is obtained via imaginary time evolution. Without any approximations, it scales exponentially as the system size grows due to the fermionic sign problem. We use the phaseless approximation [28] to obtain an algorithm that scales with the system size $N$ as $O(N^{3})-O(N^{4})$ for each sample at the expense of introducing errors in the final ground state energy estimate. The phaseless approximation sets a boundary condition in imaginary time evolution using a priori chosen wavefunction called the trial wavefunction. The bias from this approximation becomes smaller as one improves the quality of trial wavefunctions. In this work we employ the simplest trial wavefunction, the spin-restricted Hartree-Fock determinant. A recent benchmark study examined the accuracy of AFQMC with Hartree-Fock trial wavefunctions over 1004 data points [49], and based on these results at this level of approximation we expect the accuracy of AFQMC to be between CCSD and CCSD(T) for the problems considered here. Coupled cluster (CC) parameterizes the electronic wavefunction using an exponential ansatz, $\ket{\Psi_{\textrm{CC}}}=\mathrm{e}^{\hat{T}}\ket{\Phi_{\textrm{HF}}}$, where $\hat{T}$ creates all possible particle-hole excitations from the HF reference and is determined by iteratively solving a set of coupled non-linear equations[50]. CCSD approximates the full CC ansatz by truncating the $T$-operator at single and double-excitation levels, $\hat{T}=\hat{T}_{1}+\hat{T}_{2}$[26]. CCSD(T) improves upon CCSD by further including the contribution from triple excitations in a perturbative (non- iterative) manner[27] and is often referred to as the “gold standard” in the quantum chemistry of simple molecules. The computational cost of CCSD and CCSD(T) scales more steeply than that of AFQMC, as $O(N^{6})$ and $O(N^{7})$, respectively. In this work, we avoid the high computational cost of full CCSD and full CCSD(T) by systematically compressing the virtual space using the frozen natural orbital (FNO) approximation [51, 52], which we confirmed to introduce a negligible error (SI Sec. LABEL:sec:si-ccsd-ccsdt). ## III Results and Discussion Figure 2: Tetrahedral order parameter $q$ distributions as sampled via classical and PIMD simulations via NNPs fitted to either CCSD, CCSD(T), and AFQMC energies at both 300K and 370K. Having established the applicability of our data efficient approach to training MLPs that accurately reproduce the target potential energy surfaces on DFT, we now apply it to obtain the static and dynamic properties of liquid water using three correlated methods: CCSD, CCSD(T), and AFQMC. We trained a committee MLP, which as described in Sec. II.1.1 is formed of the mean of 8 independently trained MLPs, for each correlated electronic structure method. By performing MD and PIMD simulations in the NVT ensemble at 300 K and 370 K using the committee MLP for each electronic structure method, here we evaluate how different treatments of dynamical electron correlation affect the selected properties when the nuclei are treated classically or quantum mechanically. To provide an assessment of the accuracy of our transfer learned committee MLPs on different structural and dynamical properties as discussed in Sec. II.1.3 we can compare the consistency of the results obtained from models that have been initially trained to different lower level electronic structure methods. Hence, for each of the correlated methods we trained three transfer learned committee MLPs starting from HF, BLYP, and revPBE0-D3 as our lower level methods which are known to understructure, overstructure, and accurately reproduce the properties of water respectively. Performing classical MD at 300K from MLPs starting from these three different methods SI Fig. LABEL:fig:afqmc-rdfs-vdos-classical shows that regardless of the lower level that the transfer learning was initialized with, the properties obtained at the high level, AFQMC in this case, coincide closely although the O-H stretch for the HF-initialized model is blue-shifted slightly. This demonstrates that our training set of 200 energies of periodic boxes consisting of 16 water molecules is large and diverse enough to train an accurate model of liquid water under these conditions using our transfer learning protocol. When nuclear quantum effects are included by peforming PIMD simulations using the committee MLPs, the oxygen-oxygen RDF and VDOS for the revPBE0-D3 and BLYP initialized transfer learning models remain consistent with one another (SI Fig. LABEL:fig:afqmc-rdfs-vdos-rpmd). In this section we show the results from the committee MLP generated using transfer learning from revPBE0-D3 for each of the correlated methods since, as shown in SI Fig. LABEL:tab:si_electronic_structure_correlations, revPBE0-D3 shows the strongest correlation with the training set energies of CCSD, CCSD(T), and AFQMC out of the three initialization methods we used. For our AFQMC results, it is important to note that the AFQMC energies contain stochastic error, with each of our N=200 training set AFQMC energies having a corresponding estimate for the standard error that ranges from 1-2 mH depending on the specific training set configuration. To evaluate how this level of error might affect our reported properties, we employed a test where we sample new training sets where the same N=200 configuration are used but a random value is added to each AFQMC energy to reflect the uncertainty of our AFQMC energies (see SI Sec. LABEL:sec:si-energy-noise-tests for details). In total, 12 training sets were sampled and a transfer learned committee MLP was fit to each. SI Figs. LABEL:fig:afqmc_noise_rdf and LABEL:fig:afqmc_noise_vdos serve to quantify the variations in the oxygen-oxygen RDF and hydrogen VDOS obtained from the 12 separate training sets due to the stochastic error in the AFQMC energies, with the grey shading representing the standard deviations. From this test we found that the stochastic error in our AFQMC calculations does introduce noticeable variations in the OH stretch peak of the hydrogen VDOS, particularly around the top of the peak, but the RDFs and VDOS are otherwise consistent for the different training sets. ### III.1 Static properties of water from correlated electronic structure methods We first compare the static equilibrium properties for liquid water at 300 K and 370 K obtained via classical MD and PIMD simulations using the committee MLP for each correlated electronic structure method. For these properties PIMD exactly includes the NQEs for distinguishable particles, which is a highly reliable assumption for nuclei at this temperature. Figure 1 shows the oxygen- oxygen RDFs for each of CCSD, CCSD(T), and AFQMC as compared to the experimental results at 295 K[53] and 366 K[54]. At 300 K, classical MD CCSD(T) and AFQMC both give a first peak that is slightly higher than observed experimentally suggesting the liquid is overstructured. However, once NQEs are accounted for in the PIMD simulations both RDFs become slightly less structured and show better agreement with experiment, with that of CCSD(T) coinciding quantitatively. CCSD gives good agreement with experiment when used in classical MD simulations but is understructured when NQEs are included which is consistent with this electronic structure approach starting from a HF reference which gives a severely understructured liquid with the additional dynamical correlation added through the tiers of CC theory progressively structuring the liquid. At 370 K when used in PIMD simulations all methods give good agreement with the experiment with AFQMC again exhibiting a first peak that is higher than CCSD, CCSD(T) and experiment. SI Fig. LABEL:fig:mbpol_comp_rdfs_classical_300K and Fig. LABEL:fig:mbpol_comp_rdfs_pimd_300K show the hydrogen-hydrogen and oxygen- hydrogen RDFs at 300 K sampled via classical MD and PIMD, respectively, where all three electronic structure methods give similar results but AFQMC again exhibits a more structured hydrogen bond network with the first intermolecular OH peak at $\sim$1.85 $\AA$, which corresponds to hydrogen bonds, being slightly higher than the other methods. The tetrahedral order parameter provides a measure of higher order structural correlations within water’s hydrogen bond network beyond the purely radial information encoded in the RDFs. The tetrahedral order parameter $q$ is defined for a given water molecule as[55], $q=1-\frac{3}{8}\sum_{j=1}^{3}\sum_{k=j+1}^{4}\left(\cos{\theta_{jk}}+\frac{1}{3}\right),$ (1) where $\theta_{jk}$ is the angle that a given water molecule’s oxygen atom makes with two neighboring oxygen atoms $j$ and $k$. The tetrahedral order parameter thus ranges from 0 to 1 with higher values indicating that the hydrogen bond network possesses angles closer to that of a perfect tetrahedral arrangement of the four nearest neighbour oxygen atoms around a central water molecule. As shown in Fig. 2 at 300 K for both classical MD and PIMD this property further highlights the understructured hydrogen bond network of CCSD compared to the more accurate correlated methods, CCSD(T) and AFQMC, that are in close agreement. At 370 K the distribution of the tetrahedral order parameter for all three methods shifts to lower values. The comparison of these static equilibrium properties suggests that the inclusion of higher order electron correlation contributions in methods like CCSD(T) and AFQMC, as compared to CCSD or HF, results in a greater degree of structuring in liquid water at 300 K and 370 K. Given the directional nature of this additional structuring, as seen from the greater probability density at higher $q$ in Figure 2, this arises from slightly stronger hydrogen bonds being formed when using the two higher level methods. Our comparisons of the oxygen-oxygen RDFs obtained from classical and PIMD simulations at 300 K also show that for these correlated methods the inclusion of NQEs works to slightly destructure liquid water. The relatively subtle overall effect of NQEs on liquid water around 300 K is known to arise from the close balance of competing quantum effects in this system[56]. ### III.2 Dynamical properties of water from correlated electronic structure methods | AFQMC | CCSD(T) | CCSD ---|---|---|--- Classical T=300K ($10^{-9}$ m2/s) | 2.09 (0.05) | 2.21 (0.06) | 2.60 (0.08) TRPMD T=300K ($10^{-9}$ m2/s) | 2.16 (0.09) | 2.30 (0.08) | 2.80 (0.11) TRPMD T=370K ($10^{-9}$ m2/s) | 7.09 (0.11) | 8.16 (0.14) | 8.29 (0.14) Table 1: Diffusion coefficients for liquid water when running classical and TRPMD simulations using NNPs fitted to either CCSD, CCSD(T), and AFQMC energies at both 300K and 370K. Mean diffusion coefficients and standard errors of the mean are reported and are calculated using 20 ps length trajectories taken from 1 ns classical MD or 500 ps TRPMD trajectories. The experimental diffusion coefficient for water at 300 K and 370 K are 2.41$\pm$0.05[57] and 8.26$\pm$0.02 ($10^{-9}$ m2/s)[58], respectively. We now turn our attention to how the different correlated electronic structure methods behave when used to compute dynamical properties of liquid water, namely the self diffusion coefficient and VDOS. For these properties we compare the results obtained from classical MD and TRPMD. For these properties, since real time quantum dynamics is intractable for such a large atomistic condensed phase system for the timescales required to compute these properties, we use TRPMD to approximate the role of NQEs. TRPMD has previously been shown to be an accurate way for treating NQEs in the dynamics of condensed phase systems, however it is known to spuriously broaden high- frequency vibrational modes[59, 36]. The diffusion coefficients obtained for the three correlated methods are shown in Table 1 at 300 K and 370 K. Unlike the other properties we report, diffusion coefficients exhibit a notable scaling with system size that must be corrected for to make comparisons with experiment. Hence, as described in SI Sec. LABEL:sec:si-diffusion-coefficients, the diffusion coefficients were computed using simulations of 64 water molecules and then extrapolated to the infinite system size limit using the previously derived system size scaling relation[60] and the experimental viscosity of water[61]. At 300 K, where the experimentally measured diffusion coefficient is 2.41$\pm$0.05$\times 10^{-9}$ m2/s [57], when classical MD is used AFQMC and CCSD(T) yield smaller diffusion coefficients than observed experimentally, 2.09$\pm$0.05 and 2.21$\pm$0.06 $\times 10^{-9}$ m2/s respectively, while CCSD gives a larger value 2.60$\pm$0.08 $\times 10^{-9}$ m2/s. This behavior is consistent with the trends observed for the electronic structure approaches in the structural properties, where CCSD formed a slightly understructured hydrogen bond network compared to the more accurate correlated methods, which here leads to faster dynamics. Upon including NQEs using TRPMD, the diffusion coefficients for all three electronic methods increase, consistent with the disruption of the hydrogen bond network upon including zero-point energy, which brings the CCSD(T) result (2.30$\pm$0.08 $\times 10^{-9}$ m2/s) to within the statistical error bar of the experimentally observed value. Even with NQEs included the AFQMC diffusion coefficient is lower (2.16$\pm$0.09 $\times 10^{-9}$ m2/s) than experiment, consistent with it forming a more structured liquid. However, it should be noted that the discrepancy in the diffusion coefficient is very small; to change water’s diffusion coefficient from that observed at 300 K via classical MD using our CCSD(T) model to the value obtained by performing TRPMD dynamics would require less than a 2 K change in the temperature of the liquid[58]. In addition, for dynamical properties TRPMD only approximately includes NQEs and hence the better agreement of CCSD(T) with the experimental value could be changed if a different approach was used to treat the quantum dynamics of the nuclei. At 370 K, when TRPMD is used CCSD and CCSD(T) give similar results (8.29$\pm$0.14 and 8.16$\pm$0.14 $\times 10^{-9}$ m2/s, respectively), both of which are close to the experimental value of 8.26$\pm$0.02 $\times 10^{-9}$ m2/s[58]. The diffusion of AFQMC water is again considerably slower, which is consistent with the relatively greater degree of structure we saw in the latter’s respective oxygen-oxygen RDF and $q$ distribution. Figure 3: The hydrogen VDOS for liquid water when running classical and TRPMD simulations using NNPs fitted to either CCSD, CCSD(T), and AFQMC energies at both 300K and 370K. The VDOS in Figure 3 provides more information on the frequency dependence of the dynamics of water for the three electronic structure methods, since the diffusion coefficient is simply proportional to its zero frequency limit. All three methods give qualitatively similar VDOS around the lower frequency librational band ($\sim$500 cm-1) and peak associated with the H-O-H bending mode ($\sim$1600 cm-1). At 300 K using classical MD, the main qualitative difference lies in the OH stretch peak (3000-4000 cm-1), with CCSD(T) giving a peak that is redshifted by $\sim$80 cm-1 with respect to the CCSD peak, while the AFQMC peak is slightly broader than the others and is centered closer to the CCSD result. At the low-frequency end of the O-H stretch peak, the VDOS for CCSD(T) and AFQMC coincide with one another. Low frequency O-H stretches are typically associated with stronger hydrogen bonds and the frequency of the O-H stretch peak has previously been demonstrated to be inversely correlated with the tetrahedrality parameter[5]. Hence, the consistency between CCSD(T) and AFQMC at the low frequency part of their respective O-H stretch peaks is consistent with the structural evidence in Sec. III.1 showing that these two methods similarly structure water via slightly stronger hydrogen bonds as compared to CCSD. The TRPMD simulations at 300 K, which include NQEs via TRPMD simulations, lead to a $\sim$120 cm-1 red-shift and broadening of the stretch peak and a $\sim$100 cm-1 shift in the bend for all three methods, which is consistent with observations from previous studies[59, 36]. Although some of the broadening likely arises from physical effects, TRPMD is known to introduce spurious broadening of high-frequency vibrational modes[59, 36]. At 370 K the most significant difference is the shift in the zero frequency intensity, which we expect since this is directly related to the self- diffusion coefficient and otherwise the VDOS is qualitatively similar with respect to the 300 K results. This is expected since for the high frequency modes the zero point energy greatly exceeds the thermal energy in the the mode, i.e. $k_{B}T<<\hbar\omega/2$ where $k_{B}$ is the Boltzmann constant, $T$ is the temperature and $\omega$ is the frequency of the mode. Overall, the trends we observe in the dynamical properties largely mirror the evidence we presented for the structural properties showing that CCSD results in an understructured description of liquid water at 300 K, as compared to CCSD(T) and AFQMC, while at 300 K and 370 K AFQMC overstructures water. The differences in the diffusion coefficients in Table 1 reflect this trend, with CCSD overestimating the experimental diffusion coefficient at 300 K with its understructured description of water and AFQMC underpredicting the diffusion coefficient at 370 K. With the VDOS the main differences between the three correlated methods manifest in the O-H stretch peak positions and breadth, with peaks given by MD at 300 K for both CCSD(T) and AFQMC skewed more to lower frequencies that are associated with stronger hydrogen bonds. ## IV Conclusion In summary, we leveraged developments in high-level periodic electronic structure theory and exploited methods to improve the data efficiency of fitting MLPs to investigate the static and dynamical properties of liquid water at the level of CCSD, CCSD(T), and AFQMC. We devised a data efficient protocol for training MLPs that uses small periodic boxes of water (16 molecules) sampled judiciously via an iterative QbC active learning procedure. To make the most out of the few configuration energies we can afford to compute, we also employed a transfer learning approach that leverages the transferability of physics between lower level electronic structure methods (e.g. DFT with the BLYP functional, revPBE0-D3 functional, or HF) and our target higher-level methods, using MLPs fit to the former to initialize a fine-tuning transfer learning fit to the latter. Using this approach we showed that we can train stable MLPs with as few as 50 configuration energies, capture the RDFs with 100 (SI Fig. LABEL:fig:revpbe0-D3_refits_rdfs_classical_300K), and with 200 configuration energies obtain both accurate static and dynamical properties such as the diffusion constant and VDOS (SI Fig. LABEL:fig:revpbe0-D3_refits_vdos_classical_300K). In contrast, with these same 200 energies we were unable to train stable models using delta learning or using random initialization of the model. We used our MLPs trained to CCSD, CCSD(T), and AFQMC to examine how different static and dynamical properties of liquid water are affected by the level of dynamic electron correlation accounted for and the inclusion of NQEs. Our results indicate that CCSD tends to understructure liquid water and overpredict the diffusion coefficient, as compared to experiment. On the other hand, both CCSD(T) and AFQMC give oxygen-oxygen RDFs and diffusion coefficients that are more consistent with experimental values at 300 K, suggesting that the more accurate treatment of dynamical correlation present in these methods is sufficient for describing liquid water. The inclusion of NQEs for our 300 K simulations bring the CCSD(T) and AFQMC results in even closer agreement with experiment and seems to generally manifest as a slight destructuring of liquid water for all three electronic structure descriptions. This small destructuring upon including NQEs for these correlated methods is in contrast to some DFT exchange correlation functionals where due to the overprediction of the anharmonicity of the O-H coordinate the inclusion of NQEs works to structure the liquid phase[36]. Ultimately, we envision that the configurations and energies that form the training dataset, the resulting MLPs, and the protocols we employed here will be useful in their own separate respects for future work in modeling potential energy surfaces for condensed phase systems. ## Acknowledgments T.E.M and M.S.C were supported by the National Science Foundation under Grant No. CHE-2154291. This research also used resources of the National Energy Research Scientific Computing Center (NERSC), a U.S. Department of Energy Office of Science User Facility operated under Contract No. DE-AC02-05CH11231. This work was also supported by the US Air Force Office of Scientific Research under Grant No. FA9550-21-1-0400 (H.-Z.Y. and T.C.B.). We acknowledge the computing resources provided by Columbia University’s Shared Research Computing Facility project, which is supported by NIH Research Facility Improvement Grant No. 1G20RR030893-01, and associated funds from the New York State Empire State Development, Division of Science Technology and Innovation (NYSTAR) Contract No. C090171, both awarded April 15, 2010. The Flatiron Institute is a division of the Simons Foundation. ## References * DiStasio _et al._ [2014] R. A. DiStasio, B. Santra, Z. Li, X. Wu, and R. Car, “The individual and collective effects of exact exchange and dispersion interactions on the ab initio structure of liquid water,” The Journal of Chemical Physics 141, 084502 (2014), https://doi.org/10.1063/1.4893377 . * Gillan, Alfè, and Michaelides [2016] M. J. Gillan, D. Alfè, and A. Michaelides, “Perspective: How good is DFT for water?” Journal of Chemical Physics 144 (2016), 10.1063/1.4944633. * Del Ben _et al._ [2013] M. Del Ben, M. Schönherr, J. Hutter, and J. Vandevondele, “Bulk liquid water at ambient temperature and pressure from mp2 theory,” Journal of Physical Chemistry Letters 4, 3753–3759 (2013). * Morawietz _et al._ [2016] T. Morawietz, A. Singraber, C. Dellago, and J. Behler, “How van der Waals interactions determine the unique properties of water,” Proceedings of the National Academy of Sciences 113, 8368–8373 (2016), arXiv:1606.07775 . * Morawietz _et al._ [2018] T. Morawietz, O. Marsalek, S. R. Pattenaude, L. M. Streacker, D. Ben-Amotz, and T. E. Markland, “The Interplay of Structure and Dynamics in the Raman Spectrum of Liquid Water over the Full Frequency and Temperature Range,” Journal of Physical Chemistry Letters 9, 851–857 (2018). * Zhang _et al._ [2018] L. Zhang, J. Han, H. Wang, R. Car, and W. E, “Deep potential molecular dynamics: A scalable model with the accuracy of quantum mechanics,” Phys. Rev. Lett. 120, 143001 (2018). * Cheng _et al._ [2019] B. Cheng, E. A. Engel, J. Behler, C. Dellago, and M. Ceriotti, “Ab initio thermodynamics of liquid and solid water,” Proceedings of the National Academy of Sciences 116, 1110–1115 (2019), https://www.pnas.org/doi/pdf/10.1073/pnas.1815117116 . * Schran, Brezina, and Marsalek [2020] C. Schran, K. Brezina, and O. Marsalek, “Committee neural network potentials control generalization errors and enable active learning,” The Journal of Chemical Physics 153, 104105 (2020). * Yao and Kanai [2020] Y. Yao and Y. Kanai, “Temperature dependence of nuclear quantum effects on liquid water via artificial neural network model based on scan meta-gga functional,” The Journal of Chemical Physics 153, 044114 (2020), https://doi.org/10.1063/5.0012815 . * Zhang _et al._ [2021] C. Zhang, F. Tang, M. Chen, J. Xu, L. Zhang, D. Y. Qiu, J. P. Perdew, M. L. Klein, and X. Wu, “Modeling liquid water by climbing up jacob’s ladder in density functional theory facilitated by using deep neural network potentials,” The Journal of Physical Chemistry B 125, 11444–11456 (2021), pMID: 34533960, https://doi.org/10.1021/acs.jpcb.1c03884 . * Yao and Kanai [2021] Y. Yao and Y. Kanai, “Nuclear Quantum Effect and Its Temperature Dependence in Liquid Water from Random Phase Approximation via Artificial Neural Network,” Journal of Physical Chemistry Letters 12, 6354–6362 (2021). * Lan _et al._ [2021] J. Lan, D. M. Wilkins, V. Rybkin, M. Iannuzzi, and J. Hutter, “Quantum Dynamics of Water from Møller-Plesset Perturbation Theory via a Neural Network Potential,” chemRxiv (2021), 10.26434/chemrxiv-2021-n32q8-v2. * Liu, Lan, and He [2022] J. Liu, J. Lan, and X. He, “Toward high-level machine learning potential for water based on quantum fragmentation and neural networks,” The Journal of Physical Chemistry A 126, 3926–3936 (2022), pMID: 35679610, https://doi.org/10.1021/acs.jpca.2c00601 . * Bukowski _et al._ [2007] R. Bukowski, K. Szalewicz, G. C. Groenenboom, and A. van der Avoird, “Predictions of the properties of water from first principles,” Science 315, 1249–1252 (2007), https://www.science.org/doi/pdf/10.1126/science.1136371 . * Babin, Leforestier, and Paesani [2013] V. Babin, C. Leforestier, and F. Paesani, “Development of a “first principles” water potential with flexible monomers: Dimer potential energy surface, vrt spectrum, and second virial coefficient,” Journal of Chemical Theory and Computation 9, 5395–5403 (2013), pMID: 26592277, https://doi.org/10.1021/ct400863t . * Babin, Medders, and Paesani [2014] V. Babin, G. R. Medders, and F. Paesani, “Development of a “first principles” water potential with flexible monomers. ii: Trimer potential energy surface, third virial coefficient, and small clusters,” Journal of Chemical Theory and Computation 10, 1599–1607 (2014), pMID: 26580372, https://doi.org/10.1021/ct500079y . * Medders, Babin, and Paesani [2014] G. R. Medders, V. Babin, and F. Paesani, “Development of a “first-principles” water potential with flexible monomers. iii. liquid phase properties,” Journal of Chemical Theory and Computation 10, 2906–2910 (2014), pMID: 26588266, https://doi.org/10.1021/ct5004115 . * Reddy _et al._ [2016] S. K. Reddy, S. C. Straight, P. Bajaj, C. Huy Pham, M. Riera, D. R. Moberg, M. A. Morales, C. Knight, A. W. Götz, and F. Paesani, “On the accuracy of the mb-pol many-body potential for water: Interaction energies, vibrational frequencies, and classical thermodynamic and dynamical properties from clusters to liquid water and ice,” The Journal of Chemical Physics 145, 194504 (2016), https://doi.org/10.1063/1.4967719 . * Schran, Behler, and Marx [2020] C. Schran, J. Behler, and D. Marx, “Automated Fitting of Neural Network Potentials at Coupled Cluster Accuracy: Protonated Water Clusters as Testing Ground,” Journal of Chemical Theory and Computation 16, 88–99 (2020), arXiv:1908.08734 . * Nandi _et al._ [2021] A. Nandi, C. Qu, P. L. Houston, R. Conte, Q. Yu, and J. M. Bowman, “A CCSD(T)-Based 4-Body Potential for Water,” Journal of Physical Chemistry Letters 12, 10318–10324 (2021). * Daru _et al._ [2022] J. Daru, H. Forbert, J. Behler, and D. Marx, “Coupled cluster molecular dynamics of condensed phase systems enabled by machine learning potentials: Liquid water benchmark,” Phys. Rev. Lett. 129, 226001 (2022). * Yu _et al._ [2022] Q. Yu, C. Qu, P. L. Houston, R. Conte, A. Nandi, and J. M. Bowman, “q-aqua: A many-body ccsd(t) water potential, including four-body interactions, demonstrates the quantum nature of water from clusters to the liquid phase,” The Journal of Physical Chemistry Letters 13, 5068–5074 (2022), pMID: 35652912, https://doi.org/10.1021/acs.jpclett.2c00966 . * Archibald, Krogel, and Kent [2018] R. Archibald, J. T. Krogel, and P. R. Kent, “Gaussian process based optimization of molecular geometries using statistically sampled energy surfaces from quantum Monte Carlo,” Journal of Chemical Physics 149 (2018), 10.1063/1.5040584. * Ryczko, Krogel, and Tamblyn [0] K. Ryczko, J. T. Krogel, and I. Tamblyn, “Machine learning diffusion monte carlo energies,” Journal of Chemical Theory and Computation 0, null (0), pMID: 36317712, https://doi.org/10.1021/acs.jctc.2c00483 . * Huang and Rubenstein [2022] C. Huang and B. M. Rubenstein, “Machine learning diffusion monte carlo forces,” arxiv (2022), 10.48550/arxiv.2211.07103. * Purvis and Bartlett [1982] G. D. Purvis and R. J. Bartlett, “A full coupled‐cluster singles and doubles model: The inclusion of disconnected triples,” The Journal of Chemical Physics 76, 1910–1918 (1982), https://doi.org/10.1063/1.443164 . * Raghavachari _et al._ [1989] K. Raghavachari, G. W. Trucks, J. A. Pople, and M. Head-Gordon, “A fifth-order perturbation comparison of electron correlation theories,” Chemical Physics Letters 157, 479–483 (1989). * Zhang and Krakauer [2003] S. Zhang and H. Krakauer, “Quantum Monte Carlo Method using Phase-Free Random Walks with Slater Determinants,” Phys. Rev. Lett. 90, 136401 (2003). * Seung, Oppert, and Sompolinsky [1992] H. S. Seung, M. Oppert, and H. Sompolinsky, “Query by committee,” in _Proceedings of the fifth annual workshop on Computational learning theory (COLT ’92)_ (Association for Computing Machinery, New York, 1992) pp. 287–294. * Krogh and Vedelsby [1994] A. Krogh and J. Vedelsby, “Neural Network Ensembles, Cross Validation, and Active Learning,” in _Advances in Neural Information Processing Systems_, Vol. 7, edited by G. Tesauro, D. Touretzky, and T. Leen (MIT Press, 1994) pp. 231–238. * Behler and Parrinello [2007] J. Behler and M. Parrinello, “Generalized neural-network representation of high-dimensional potential-energy surfaces,” Physical Review Letters 98, 146401 (2007). * Elstner _et al._ [1998] M. Elstner, D. Porezag, G. Jungnickel, J. Elsner, M. Haugk, T. Frauenheim, S. Suhai, and G. Seifert, “Self-consistent-charge density-functional tight-binding method for simulations of complex materials properties,” Phys. Rev. B 58, 7260–7268 (1998). * Perdew, Burke, and Ernzerhof [1996] J. P. Perdew, K. Burke, and M. Ernzerhof, “Generalized gradient approximation made simple,” Phys. Rev. Lett. 77, 3865–3868 (1996). * Zhang and Yang [1998] Y. Zhang and W. Yang, “Comment on “generalized gradient approximation made simple”,” Phys. Rev. Lett. 80, 890–890 (1998). * Adamo and Barone [1999] C. Adamo and V. Barone, “Toward reliable density functional methods without adjustable parameters: The pbe0 model,” The Journal of Chemical Physics 110, 6158–6170 (1999), https://doi.org/10.1063/1.478522 . * Marsalek and Markland [2017] O. Marsalek and T. E. Markland, “Quantum Dynamics and Spectroscopy of Ab Initio Liquid Water: The Interplay of Nuclear and Electronic Quantum Effects,” Journal of Physical Chemistry Letters 8, 1545–1551 (2017). * Becke [1988] A. D. Becke, “Density-functional exchange-energy approximation with correct asymptotic behavior,” Phys. Rev. A 38, 3098–3100 (1988). * Lee, Yang, and Parr [1988] C. Lee, W. Yang, and R. G. Parr, “Development of the colle-salvetti correlation-energy formula into a functional of the electron density,” Phys. Rev. B 37, 785–789 (1988). * Kühne, Krack, and Parrinello [2009] T. D. Kühne, M. Krack, and M. Parrinello, “Static and dynamical properties of liquid water from first principles by a novel car-parrinello-like approach,” Journal of Chemical Theory and Computation 5, 235–241 (2009), pMID: 26610101, https://doi.org/10.1021/ct800417q . * Pan and Yang [2010] S. J. Pan and Q. Yang, “A survey on transfer learning,” IEEE Transactions on Knowledge and Data Engineering 22, 1345–1359 (2010). * Singraber _et al._ [2019] A. Singraber, T. Morawietz, J. Behler, and C. Dellago, “Parallel multistream training of high-dimensional neural network potentials,” Journal of Chemical Theory and Computation 15, 3075–3092 (2019). * Sun _et al._ [2020] Q. Sun, X. Zhang, S. Banerjee, P. Bao, M. Barbry, N. S. Blunt, N. A. Bogdanov, G. H. Booth, J. Chen, Z.-H. Cui, _et al._ , “Recent developments in the pyscf program package,” J. Chem. Phys. 153, 024109 (2020). * McClain _et al._ [2017] J. McClain, Q. Sun, G. K.-L. Chan, and T. C. Berkelbach, “Gaussian-based coupled-cluster theory for the ground-state and band structure of solids,” Journal of Chemical Theory and Computation 13, 1209–1218 (2017), pMID: 28218843, https://doi.org/10.1021/acs.jctc.7b00049 . * Ye and Berkelbach [2021] H.-Z. Ye and T. C. Berkelbach, “Fast periodic gaussian density fitting by range separation,” J. Chem. Phys. 154, 131104 (2021). * Ye* and Berkelbach [2021] H.-Z. Ye* and T. C. Berkelbach, “Tight distance-dependent estimators for screening two-center and three-center short-range coulomb integrals over gaussian basis functions,” J. Chem. Phys. 155, 124106 (2021). * Kim _et al._ [2018] J. Kim, A. D. Baczewski, T. D. Beaudet, A. Benali, M. C. Bennett, M. A. Berrill, N. S. Blunt, E. J. L. Borda, M. Casula, D. M. Ceperley, _et al._ , “Qmcpack: an open source ab initio quantum monte carlo package for the electronic structure of atoms, molecules and solids,” J. Phys. Cond. Mat. 30, 195901 (2018). * Kent _et al._ [2020] P. R. C. Kent, A. Annaberdiyev, A. Benali, M. C. Bennett, E. J. Landinez Borda, P. Doak, H. Hao, K. D. Jordan, J. T. Krogel, I. Kylänpää, J. Lee, Y. Luo, F. D. Malone, C. A. Melton, L. Mitas, M. A. Morales, E. Neuscamman, F. A. Reboredo, B. Rubenstein, K. Saritas, S. Upadhyay, G. Wang, S. Zhang, and L. Zhao, “QMCPACK: Advances in the development, efficiency, and application of auxiliary field and real-space variational and diffusion quantum Monte Carlo,” J. Chem. Phys. 152, 174105 (2020). * Malone _et al._ [2022] F. D. Malone, A. Mahajan, J. S. Spencer, and J. Lee, “ipie: A Python-based Auxiliary-Field Quantum Monte Carlo Program with Flexibility and Efficiency on CPUs and GPUs,” arXiv (2022), 10.48550/arXiv.2209.04015, 2209.04015 . * Lee, Pham, and Reichman [2022] J. Lee, H. Q. Pham, and D. R. Reichman, “Twenty Years of Auxiliary-Field Quantum Monte Carlo in Quantum Chemistry: An Overview and Assessment on Main Group Chemistry and Bond-Breaking,” J. Chem. Theory Comput. 2022 (2022), 10.1021/acs.jctc.2c00802. * Čížek [1966] J. Čížek, “On the correlation problem in atomic and molecular systems. calculation of wavefunction components in ursell‐type expansion using quantum‐field theoretical methods,” The Journal of Chemical Physics 45, 4256–4266 (1966), https://doi.org/10.1063/1.1727484 . * Taube and Bartlett [2008] A. G. Taube and R. J. Bartlett, “Frozen natural orbital coupled-cluster theory: Forces and application to decomposition of nitroethane,” J. Chem. Phys. 128, 164101 (2008). * Lange and Berkelbach [2020] M. F. Lange and T. C. Berkelbach, “Active space approaches combining coupled-cluster and perturbation theory for ground states and excited states,” Mol. Phys. 118, e1808726 (2020). * Skinner _et al._ [2013] L. B. Skinner, C. Huang, D. Schlesinger, L. G. M. Pettersson, A. Nilsson, and C. J. Benmore, “Benchmark oxygen-oxygen pair-distribution function of ambient water from x-ray diffraction measurements with a wide q-range,” The Journal of Chemical Physics 138, 074506 (2013). * Mariedahl _et al._ [2018] D. Mariedahl, F. Perakis, A. Späh, H. Pathak, K. H. Kim, G. Camisasca, D. Schlesinger, C. Benmore, L. G. M. Pettersson, A. Nilsson, and K. Amann-Winkel, “X-ray Scattering and O-O Pair-Distribution Functions of Amorphous Ices,” Journal of Physical Chemistry B 122, 7616–7624 (2018). * Jeffrey R. Errington and Pablo G. Debenedetti [2001] Jeffrey R. Errington and Pablo G. Debenedetti, “Relationship between structural order and the anomalies of liquid water,” Nature 409, 318 (2001). * Habershon, Markland, and Manolopoulos [2009] S. Habershon, T. E. Markland, and D. E. Manolopoulos, “Competing quantum effects in the dynamics of a flexible water model,” The Journal of Chemical Physics 131, 024501 (2009). * Holz, Heil, and Sacco [2000] M. Holz, S. R. Heil, and A. Sacco, “Temperature-dependent self-diffusion coefficients of water and six selected molecular liquids for calibration in accurate 1h nmr pfg measurements,” Phys. Chem. Chem. Phys. 2, 4740–4742 (2000). * Dietrich [2002] O. Dietrich, “Diffusion Coefficients of Water,” https://dtrx.de/od/diff/ (2002), [Accessed: 30-August-2022]. * Rossi _et al._ [2014] M. Rossi, H. Liu, F. Paesani, J. Bowman, and M. Ceriotti, “Communication: On the consistency of approximate quantum dynamics simulation methods for vibrational spectra in the condensed phase,” The Journal of Chemical Physics 141, 181101 (2014). * Yeh and Hummer [2004] I. C. Yeh and G. Hummer, “System-size dependence of diffusion coefficients and viscosities from molecular dynamics simulations with periodic boundary conditions,” Journal of Physical Chemistry B 108, 15873–15879 (2004). * Kestin, Sokolov, and Wakeham [1978] J. Kestin, M. Sokolov, and W. A. Wakeham, “Viscosity of liquid water in the range -8c to 150c,” Journal of Physical and Chemical Reference Data 7, 941–948 (1978).
# TetraFreeQ: tetrahedra-free quadrature on polyhedral elements Alvise Sommariva<EMAIL_ADDRESS>Marco Vianello<EMAIL_ADDRESS>University of Padova, Italy Member of the INdAM Research group GNCS ###### Abstract In this paper we provide a tetrahedra-free algorithm to compute low- cardinality quadrature rules with a given degree of polynomial exactness, positive weights and interior nodes on a polyhedral element with arbitrary shape. The key tools are the notion of Tchakaloff discretization set and the solution of moment-matching equations by Lawson-Hanson iterations for NonNegative Least-Squares. Several numerical tests are presented. The method is implemented in Matlab as open-source software. ###### keywords: Polyhedral elements, tetrahedra-free algebraic quadrature, Positive Interior rules, Tchakaloff theorem, Tchakaloff sets, Davis-Wilhelmsen theorem, NonNegative Least Squares. ###### MSC: [2020]65D32. ††journal: Applied Numerical Mathematics ## 1 Introduction Let $\Omega\subset{\mathbb{R}}^{3}$ is a polyhedron (either convex or nonconvex or even multiply-connected), and suppose that for a given continuous function $f\in C(\Omega)$ one needs to approximate its integral over $\Omega$ by a quadrature rule with nodes $\\{Q_{j}\\}_{j=1,\ldots,\nu}\subset\Omega$ where $Q_{j}=(x_{j},y_{j},z_{j})$ and positive weights $\\{w_{j}\\}_{j=1,\ldots,\nu}$, $I_{\Omega}(f)=\int_{\Omega}f(x,y,z)\,\,\,dx\,dy\,dz\approx I_{n}(f)=\sum_{j=1}^{\nu}w_{j}f(Q_{j})\;,$ (1) that has Algebraic Degree of Exactness $ADE=n$, i.e. $I_{\Omega}(p)=I_{n}(p)$ for all $p\in{\mathbb{P}}^{3}_{n}$ (the space of trivariate polynomials of total degree not exceeding $n$). Such a kind of rules are often called PI (Positive Interior) in the literature and are optimally stable by the positivity of the weights, being the quadrature conditioning $\sum_{j}|w_{j}|/\left|\sum_{j}w_{j}\right|=1$. As a relevant motivation, we may recall that the algebraic quadrature problem on polyhedra has been extensively studied in the literature, especially during the last decade, being at the core of many discretization methods for PDEs that use polyhedral meshes, such as polyhedral FEM and discontinuous Galerkin, as well as the more recent VEM. We may quote among others, without any pretence of exhaustivity, the papers [1, 4, 15, 23, 24] with the numerous references therein. Despite of such manifest interest in the framework of numerical PDEs, open- source numerical codes to compute PI algebraic quadrature rules on polyhedra, in particular Matlab codes, do not seem to be readily available. On the other hand, differently from other approaches that do not provide an algebraic quadrature formula but a numerical integration algorithm, the availability of low-cardinality quadrature formulas allows to compute efficiently integrals involving forcing terms, without the need of an explicit polynomial approximation of such terms (cf. [1, Rem. 6]). Moreover, since the polyhedral quadrature formulas can be computed once and for all for a fixed polyhedral mesh, even more flexibility is gained when different forcing terms have to be tested, or more generally when parametrized PDEs have to be solved, for example by reduced basis methods or other model order reduction techniques (cf., e.g., [10]). Indeed, the main purpose of the present work is to provide an algorithm implemented by an open-source Matlab package, named TetraFreeQ, for the computation of tetrahedra-free PI algebraic quadrature rules on arbitrary polyhedra (either convex or nonconvex or even multiply-connected), with cardinality $\nu\leq N$ where $N=N_{n}=dim(\mathbb{P}_{n}^{3})=(n+1)(n+2)(n+3)/6\;.$ (2) Existence and computability of such rules are guaranteed by the well-known Tchakaloff theorem [25] and the less known Wilhelmsen theorem on “Tchakaloff sets” [28], the latter being at the core of our approach as described in the nest section. ## 2 Tetrahedra-free quadrature In this section we show how to determine a quadrature rule with $ADE=n$, positive weights and interior nodes, on a general polyhedron $\Omega$, without needing a decomposition of the domain by means of nonoverlapping tetrahedra (tetrahedra-free quadrature, for short). It should be recalled that avoiding “sub-tessellations” for integration on polyhedral elements is a common requirement in the numerical PDEs framework, especially with high-order methods. In the sequel the polyhedron is assumed to be given via its polygonal faces, say $\\{\mathcal{F}_{i}\\}$, each with counterclockwise oriented vertices w.r.t. the outward normal. The procedure works essentially as follows: Algorithm TetraFreeQ (tetrahedra-free quadrature on a general polyhedron $\Omega$) * $(i)$ compute the moment array $\gamma=\\{\gamma_{k}\\}=\left\\{\int_{\Omega}{\phi_{k}(x,y,z)\,dx\,dy\,dz}\right\\}$ (3) of a certain polynomial basis $\\{\phi_{k}\\}$ of ${\mathbb{P}}_{n}^{3}$, $k=1,\ldots,N$, by the divergence theorem and bivariate quadrature rules with ADE equal to $n+1$ on the polyhedron faces $\\{{\mathcal{F}}_{i}\\}_{i=1,\ldots,M}$ * $(ii)$ using an in-polyhedron routine [11], in view of Wilhelmsen theorem [28], determine iteratively a sufficiently dense pointset $\\{P_{l}\\}_{l=1,\ldots,L}$ inside $\Omega$ (a so-called “Tchakaloff set”) such that the underdetermined moment-matching system $V^{t}u=\gamma\;,\;\;\mbox{where}\;\;V=[v_{lk}]=[\phi_{k}(P_{l})]\in\mathbb{R}^{L\times N}$ (4) has a nonnegative solution $u$ with at most $N\leq L$ positive components, computed via Lawson-Hanson NNLS algorithm [13] applied to $\min_{u\geq 0}{\|V^{t}u-\gamma\|_{2}}$ (5) * $(iii)$ once $\\{P_{l}\\}_{l=1,\ldots,L}$ and the weight vector $u$ are determined, the nonzeros of $u$ finally select the nodes of a PI-type quadrature rule with cardinality at most $N$. In spite of the simplicity of this approach, there are many theoretical and practical aspects that deserve explanations. We stress that the key tools are Wilhelmsen theorem on the existence of “Tchakaloff sets” within sufficiently dense discrete subsets (a relevant but somehow overlooked result of multivariate quadrature theory), and the capability of Lawson-Hanson active- set iterative algorithm to compute a sparse solution to the NonNegative Least- Squares problem (made more efficient by the recent Deviation Maximization approach to QR factorization [6, 7]). ### 2.1 Moment computation For an implementation of the algorithm sketched above, we intend to compute first the moments on the polyhedron $\Omega$ of the product Chebyshev basis of total degree $n$, relative to the smallest parallelepiped ${\mathcal{R}}=[a_{1},b_{1}]\times[a_{2},b_{2}]\times[a_{3},b_{3}]$ containing $\Omega$. In other words, defined the scaled Chebyshev polynomial in $[a,b]$ ${\tilde{T}}^{(a,b)}_{m}(t):=T_{m}\left(\frac{2}{b-a}\cdot\left(t-\frac{a+b}{2}\right)\right)$ (6) where $T_{m}$ is the Chebyshev polynomial of first kind, of degree $m$, we compute the value of the moments $\gamma_{\alpha}=\int_{\Omega}{\tilde{T}}^{(a_{1},b_{1})}_{\alpha_{1}}(x){\tilde{T}}^{(a_{2},b_{2})}_{\alpha_{2}}(y){\tilde{T}}^{(a_{3},b_{3})}_{\alpha_{3}}(z){\mbox{ }}dx\,dy\,dz$ for all the triples $\alpha=(\alpha_{1},\alpha_{2},\alpha_{3})$ with $\alpha_{1},\alpha_{2},\alpha_{3}\in\mathbb{N}$ and $0\leq\alpha_{1}+\alpha_{2}+\alpha_{3}\leq n$. To this purpose, we recall that in view of the divergence theorem applied to a continuously differentiable vector field ${\mathbf{F}}=(f,g,h)$ defined on a neighborhood of $\Omega$, we have $\int_{\Omega}\nabla\cdot{\mathbf{F}}{\mbox{ }}dx\,dy\,dz=\int_{\partial\Omega}{\mathbf{F}}\cdot{\mathbf{n}}{\mbox{ }}dS$ (7) where ${\mathbf{n}}=(n_{1},n_{2},n_{3})$ is the outward pointing unit normal at each point on the boundary ${\partial\Omega}$ and $\nabla\cdot{\mathbf{F}}={\frac{\partial f}{\partial x}}+{\frac{\partial g}{\partial y}}+{\frac{\partial h}{\partial z}}$ is the divergence of $\mathbf{F}$. As a particular case of (7), we consider ${\mathbf{F}}=(f,0,0)$ obtaining $\int_{\Omega}{\frac{\partial f(x,y,z)}{\partial x}}dx\,dy\,dz=\int_{\partial\Omega}n_{1}(x,y,z)\,{{f(x,y,z)}}{\mbox{ }}dS\;,$ (8) and we apply (8) to a certain $f$ such that ${\frac{\partial f(x,y,z)}{\partial x}}={\tilde{T}}^{(a_{1},b_{1})}_{\alpha_{1}}(x){\tilde{T}}^{(a_{2},b_{2})}_{\alpha_{2}}(y){\tilde{T}}^{(a_{3},b_{3})}_{\alpha_{3}}(z).$ In this regard, we exploit a fundamental property of Chebyshev polynomials of first kind, namely that for $m\geq 2$ $\int_{0}^{x}T_{m}(t)dt=\frac{T_{m+1}(x)}{2(m+1)}-\frac{T_{m-1}(x)}{2(m-1)}.$ (9) Consequently, setting $s=\frac{2}{b-a}\cdot(t-\frac{a+b}{2})$, by (6) and (9), for $m\geq 2$, $\displaystyle\int_{0}^{x}{\tilde{T}}^{(a,b)}_{m}(t)dt$ $\displaystyle=$ $\displaystyle\int_{0}^{x}T_{m}\left(\frac{2}{b-a}\cdot\left(t-\frac{a+b}{2}\right)\right)dt$ (10) $\displaystyle=$ $\displaystyle\int_{-\frac{a+b}{b-a}}^{\frac{2}{b-a}(x-\frac{a+b}{2})}T_{m}(s)\frac{2}{b-a}ds$ $\displaystyle=$ $\displaystyle\frac{2}{b-a}\left(\int_{-\frac{a+b}{b-a}}^{0}T_{m}(s)ds+\int_{0}^{\frac{2}{b-a}(x-\frac{a+b}{2})}T_{m}(s)ds\right)$ $\displaystyle=$ $\displaystyle\frac{2}{b-a}\left(\int_{-\frac{a+b}{b-a}}^{0}T_{m}(s)ds+\frac{T_{m+1}(\frac{2}{b-a}(x-\frac{a+b}{2}))}{2(m+1)}-\frac{T_{m-1}(\frac{2}{b-a}(x-\frac{a+b}{2}))}{2(m-1)}\right)$ $\displaystyle=$ $\displaystyle\frac{2}{b-a}\left(\int_{-\frac{a+b}{b-a}}^{0}T_{m}(s)ds+\frac{{\tilde{T}}^{(a,b)}_{m+1}(x)}{2(m+1)}-\frac{{\tilde{T}}^{(a,b)}_{m-1}(x)}{2(m-1)}\right).$ Thus, being the term $\frac{2}{b-a}\int_{-\frac{a+b}{b-a}}^{0}T_{m}(s)ds$ costant, setting $\mathcal{Z}^{(a,b)}_{m}(x)=\frac{1}{b-a}\left(\frac{{\tilde{T}}^{(a,b)}_{m+1}(x)}{m+1}-\frac{{\tilde{T}}^{(a,b)}_{m-1}(x)}{m-1}\right)$ we get that for $m\geq 2$ $\displaystyle{\tilde{T}}^{(a,b)}_{m}(x)=\frac{\partial}{\partial x}\mathcal{Z}^{(a,b)}_{m}(x).$ (11) It can be easily checked that (11) holds also in the case $m=0$ and $m=1$, by defining respectively $\mathcal{Z}^{(a,b)}_{0}(x)=x-\frac{a+b}{2},\mbox{ }\mathcal{Z}^{(a,b)}_{1}(x)=\frac{1}{b-a}\left(x-\frac{a+b}{2}\right)^{2},$ and that $\mathcal{Z}^{(a,b)}_{m}\in{\mathbb{P}}_{m+1}$. Now, denoting by $\\{{\mathcal{F}}_{i}\\}_{i=1,\ldots,M}$ the polygonal faces that compose the boundary of the polyhedron $\Omega$, where ${\mathbf{n}}^{(i)}=(n^{(i)}_{1},n^{(i)}_{2},n^{(i)}_{3})$ is the outward normal of ${\mathcal{F}}_{i}$, $i=1,\ldots,M$, by the form of the divergence theorem (8) for $f_{\alpha}(x,y,z)=\mathcal{Z}^{(a_{1},b_{1})}_{\alpha_{1}}(x){\tilde{T}}^{(a_{2},b_{2})}_{\alpha_{2}}(y){\tilde{T}}^{(a_{3},b_{3})}_{\alpha_{3}}(z)$ from the application of (11) with $a=a_{1}$ and $b=b_{1}$, we have $\displaystyle\gamma_{\alpha}$ $\displaystyle=$ $\displaystyle\int_{\Omega}{\tilde{T}}^{(a_{1},b_{1})}_{\alpha_{1}}(x){\tilde{T}}^{(a_{2},b_{2})}_{\alpha_{2}}(y){\tilde{T}}^{(a_{3},b_{3})}_{\alpha_{3}}(z){\mbox{ }}dx\,dy\,dz$ $\displaystyle=$ $\displaystyle\int_{\Omega}\frac{\partial}{\partial x}\mathcal{Z}^{(a_{1},b_{1})}_{\alpha_{1}}(x){\tilde{T}}^{(a_{2},b_{2})}_{\alpha_{2}}(y){\tilde{T}}^{(a_{3},b_{3})}_{\alpha_{3}}(z){\mbox{ }}dx\,dy\,dz$ $\displaystyle=$ $\displaystyle\int_{\Omega}\frac{\partial}{\partial x}\left(\mathcal{Z}^{(a_{1},b_{1})}_{\alpha_{1}}(x){\tilde{T}}^{(a_{2},b_{2})}_{\alpha_{2}}(y){\tilde{T}}^{(a_{3},b_{3})}_{\alpha_{3}}(z)\right){\mbox{ }}dx\,dy\,dz$ $\displaystyle=$ $\displaystyle\int_{\partial\Omega}n_{1}(x,y,z)\mbox{ }\mathcal{Z}^{(a_{1},b_{1})}_{\alpha_{1}}(x){\tilde{T}}^{(a_{2},b_{2})}_{\alpha_{2}}(y){\tilde{T}}^{(a_{3},b_{3})}_{\alpha_{3}}(z){\mbox{ }}dS$ $\displaystyle=$ $\displaystyle\sum_{i=1}^{M}\int_{{\mathcal{F}}_{i}}n^{(i)}_{1}\mbox{ }\mathcal{Z}^{(a_{1},b_{1})}_{\alpha_{1}}(x){\tilde{T}}^{(a_{2},b_{2})}_{\alpha_{2}}(y){\tilde{T}}^{(a_{3},b_{3})}_{\alpha_{3}}(z){\mbox{ }}dS\;.$ Since the outward normal $n_{1}$ is constant on the face ${\mathcal{F}}_{i}$, so is $n^{(i)}_{1}$, and thus the integrand $n^{(i)}_{1}\mbox{ }\mathcal{Z}^{(a_{1},b_{1})}_{\alpha_{1}}(x){\tilde{T}}^{(a_{2},b_{2})}_{\alpha_{2}}(y){\tilde{T}}^{(a_{3},b_{3})}_{\alpha_{3}}(z)$ is a polynomial of degree not exceeding $n+1$ on the $i$-th face, with $i=1,\ldots,M$, and consequently each moment $\gamma_{\alpha}$ can be recovered by a quadrature rule with ADE equal to $n+1$ for all the polygons ${\mathcal{F}}_{1},\ldots,{\mathcal{F}}_{M}$. Having this in mind, let ${\mathcal{F}}\subset\partial\Omega$ be a (polygonal) face of the polyhedron $\Omega$, and suppose that the vertices are oriented counterclockwise w.r.t. its outward normal ${\mathbf{n}}$. Since it is easier to determine a quadrature rule on a planar polygon ${\mathcal{F}}^{*}$, we apply a rototranslation $\Phi$, such that ${\mathcal{F}}^{*}=\Phi({\mathcal{F}})$ is a polygon belonging to the plane $\pi^{*}=\\{(x,y,0),\,\,x,y\in\mathbb{R}\\}$, oriented counterclockwise w.r.t. the vector $(0,0,1)$. A quadrature formula on the planar polygon ${\mathcal{F}}^{*}$ with internal nodes $\\{P_{i}\\}$ and positive weights $\\{w_{i}\\}$, having ADE equal to $n+1$, can be determined as described in [3], by applying a rule with ADE equal to $n+1$ on each element of a minimal triangulation of ${\mathcal{F}}^{*}$ or alternatively by a triangulation free routine as discussed in [21]. Since the rototranslation $\Phi$ is an affine (bijective) map with Jacobian modulus equal to 1, we conclude that $\\{\Phi^{-1}(P_{i})\\}$ and $\\{w_{i}\\}$ are the nodes and the weights of a quadrature rule over the polygonal face ${\mathcal{F}}\subset\partial\Omega$, with ADE equal to $n+1$ and consequently we are able to compute the set of moments $\\{\gamma_{\alpha}\\}$. ###### Remark 2.1. The mapping $\phi$ can be easily obtained by applying a rotation $R$ to the plane $\pi$ containing the vertices of the polygon $\mathcal{F}$, so that $R\pi$ is parallel to the $xy$-plane $\pi^{*}=\\{(x,y,0),\,\,x,y\in\mathbb{R}\\}$ and then by applying a vertical translation $\tau$ so that $\hat{\mathcal{F}}:=R{\mathcal{F}}+\tau\subset\pi^{*}$. If a counterclockwise orientation of $\mathcal{F}$ is given by the user, then it is straightforward to obtain the same for $\hat{\mathcal{F}}$. For the quadrature routines over planar polygons it is a typical requirement that the orientation is counterclockwise. If the planar polygon $\hat{\mathcal{F}}$ is oriented clockwise w.r.t. the z-axis then, applying an additional rotation, we can take as ${\mathcal{F}}^{*}$ the polygon symmetric to $\hat{\mathcal{F}}$ w.r.t. the x-axis, otherwise it is sufficient to set ${\mathcal{F}}^{*}=\hat{\mathcal{F}}$. We recall that a necessary and sufficient condition for a polygon with vertices $(x_{j},y_{j},0)$, $j=1,\ldots,m$ to be oriented counterclockwise is that its signed area $\sum_{j=1}^{m-1}(x_{j+1}-x_{j})(y_{j+1}+y_{j})$ be negative (for further details see [27]). In some instances, instead of the orientation of the vertices of ${\mathcal{F}}$, it is available its outward normal $\mathbf{n}$ (w.r.t. $\Omega$). In this case, the procedure is the similar to that of the previous remark. If in particular, after the application of the rotation $R$ (and the translation $\tau$), the normal to ${\mathcal{\hat{F}}}$ correspond to $\Phi({\mathbf{n}})=(0,0,-1)$ then as ${\mathcal{F}}^{*}$ we can consider the polygon symmetric to $\hat{\mathcal{F}}$ w.r.t. the x-axis, otherwise we set ${\mathcal{F}}^{*}=\hat{\mathcal{F}}$. ### 2.2 Tchakaloff sets and computation of Tchakaloff-like rules In the previous subsection we have implemented step $(i)$ of the Algorithm TetraFreeQ sketched at the beginning of Section 2. With a little abuse of notation, we shall consider an ordering of the multi-indices $\alpha=(\alpha_{1},\alpha_{2},\alpha_{3})$, $0\leq\alpha_{1}+\alpha_{2}+\alpha_{3}\leq n$, e.g. the graded lexicographical ordering, and we shall term $\\{\phi_{k}(x,y,z)\\}$ the product Chebyshev basis and $\\{\gamma_{k}\\}$ the corresponding moments, $1\leq k\leq N$. Now, step $(ii)$ of Algorithm TetraFreeQ rests on the following relevant (though often overlooked) result proved by Wilhelmsen in [28] where the key notion of Tchakaloff set is introduced (extending a result of Davis [8] to rather general functional spaces, including polynomial spaces). ###### Theorem 2.1. (Wilhelmsen, 1976) Let $\Psi$ be the linear span of continuous, real-valued, linearly independent functions $\\{\phi_{k}\\}_{k=1,\ldots,N}$ defined on a compact set $\Omega\subset{\mathbb{R}}^{d}$. Assume that $\Psi$ satisfies the Krein condition (i.e. there is at least one $f\in\Psi$ which does not vanish on $\Omega$) and that $L$ is a positive linear functional, i.e. $Lf>0$ for every $f\geq 0$ not vanishing everywhere in $\Omega$. If $\\{P_{l}\\}_{i=1}^{\infty}$ is an everywhere dense subset of $\Omega$, then for sufficiently large $L$, the set $X=\\{P_{l}\\}_{l=1,\ldots,L}$ is a Tchakaloff set, i.e. there exist weights $w_{j}>0$, $j=1,\ldots,\nu$, and nodes $\\{Q_{j}\\}_{j=1,\dots,\nu}\subset X\subset\Omega$, with $\nu={\mbox{card}}(\\{Q_{j}\\})\leq N$, such that $Lf=\sum_{j=1}^{\nu}w_{j}f(Q_{j})\;,{\mbox{ }}\forall f\in\Psi\;.$ (13) In view of this existence result, we can the implement step $(ii)$ in our framework, that is $Lf=I_{\Omega}(f)$ and $\Psi=\mathbb{P}_{n}^{3}$, by the following iterative substeps: * $(ii1)$ generate a set of uniform random or low-discrepancy points (e.g. Halton points) with cardinality $K$ in the smallest box $[a_{1},b_{1}]\times[a_{2},b_{2}]\times[a_{3},b_{3}]$ containing $\Omega$ * $(ii2)$ determine those points belonging to $\Omega$ by an in-polyhedron routine like [11], say $X=\\{P_{l}\\}_{l=1,\ldots,L}$ with $L\leq K$, and compute the Vandermonde-like matrix $V=[v_{lk}]=[\phi_{k}(P_{l})]\in\mathbb{R}^{L\times N}$ * $(ii3)$ solve the NNLS problem $\min_{u\geq 0}{\|V^{t}u-\gamma\|_{2}}$ by Lawson- Hanson active-set method * $(ii4)$ if the residual $\|V^{t}u-\gamma\|_{2}<\varepsilon$ (where $\varepsilon$ is a given tolerance) then set $\\{w_{j}\\}=\\{u_{l}>0\\}$ and $\\{Q_{j}\\}=\\{P_{l}:\,u_{l}>0\\}$ else goto $(ii1)$ increasing $K$, e.g. by a fixed factor $\theta>1$ We stress that since the set of random points becomes denser along the iterations, in view of Wilhelmsen theorem above termination of the iterative cycle is (at least theoretically) guaranteed. Substep $(ii3)$ is based on the fact that Lawson-Hanson algorithm computes a sparse solution of the NNLS problem, and can be performed in Matlab by resorting to the built-in routine lsqnonneg. On the other hand, a substantial acceleration can be obtained via the recent Deviation-Maximization approach to $QR$ factorizations which are the core of Lawson-Hanson method; cf. [5, 6, 7] for a detailed analysis of this approach. ###### Remark 2.2. (Tetrahedra-based quadrature) As already stressed, the aim of the present work is to provide a completely “tetrahedra-free” approach to quadrature on general polyhedral elements, implemented in Matlab. On the other hand, a classical approach for polyhedral quadrature, whenever a nonoverlapping “tetrahedralization” of the polyhedron is available, consists in collecting and summing up the PI-formulas on the single tetrahedra by integration additivity. A first problem consists in the computation of the tetrahedralization. If $\Omega$ is convex or star-shaped with a known center, this operation is straightforward, but the task becomes not trivial for a general polyhedron that may have a difficult geometry. We recall, incidentally, that differently from the 2D case of polygons, there exist polyhedra that cannot be triangulated using only their vertices (e.g. the well-known Schönardt polyhedron, cf. [17]). Moreover, though Matlab provides a minimal triangulation of a general polygonal region via the environment polyshape, similar codes do not seem available for polyhedra. As an alternative, sometimes useful, the built-in alphashape command creates a bounding volume that envelops a given 3D point cloud pointset. If the polyhedron of interest can be defined in such a way, then alphaTriangulation returns a tetrahedralization that defines the domain of the alpha shape (see [14] for additional details). A second problem concerns the rule to be used in each tetrahedron. For degrees of exactness up to $n=20$, there are in literature several PI-type rules on the reference tetrahedron (the simplex) with vertices $(1,0,0)$, $(0,1,0)$, $(0,0,0)$, $(0,0,1)$, which have near-minimal cardinality (see e.g. [12, 18] and Table 1). For $n>20$, one can use the well-established Stroud rule [22, p.28-32], again of PI-type, that in general is not minimal but still easy to be implemented and with a moderate cardinality equal to $\lceil{\frac{n+1}{2}}\rceil^{3}$. All these rules are typically written in barycentric coordinates $(\lambda_{1},\lambda_{2},\lambda_{3},\lambda_{4})$ and can be applied to a general tetrahedron with vertices $A,B,C,D$ by the standard change of coordinates $\lambda_{1}A+\lambda_{2}B+\lambda_{3}C+\lambda_{4}D$. deg | card | deg | card | deg | card | deg | card ---|---|---|---|---|---|---|--- $1$ | $1$ | $6$ | $23$ | $11$ | $94$ | $16$ | $247$ $2$ | $4$ | $7$ | $31$ | $12$ | $117$ | $17$ | $288$ $3$ | $6$ | $8$ | $44$ | $13$ | $144$ | $18$ | $338$ $4$ | $11$ | $9$ | $57$ | $14$ | $175$ | $19$ | $390$ $5$ | $14$ | $10$ | $74$ | $15$ | $207$ | $20$ | $448$ Table 1: Cardinality of near-minimal rules on the reference tetrahedron. Of course the overall cardinality of the resulting rule is proportional to the number of tetrahedra and can be much larger than $N=(n+1)(n+2)(n+3)/6=dim(\mathbb{P}_{n}^{3})$. In these cases, one can resort to Tchakaloff-like compression of the corresponding discrete measure, obtaining a rule with at most $N$ re-weighted nodes that are a subset of the starting ones. Such a compression consists in computing a sparse positive solution of a moment-matching system like (4), where the points and moments are given by the high-cardinality rule itself. This approach inserts in the more general setting of “Caratheodory-Tchakaloff discrete measure compression”, and can be implemented by both Linear and Quadratic Programming methods; cf. e.g. [9, 16, 20, 26] with the references therein. In our Matlab package we have added, for the purpose of comparison, the routines that provide a compressed Tchakaloff-like rule along the lines sketched above, starting from a tetrahedralization of the polyhedral element. We notice that even this approach does not seem to be yet readily available, at least in Matlab, within existing open-source packages, and thus could be of independent practical interest, especially because it resorts to the best tetrahedral rules available in the literature. ## 3 Numerical tests In this section we test Algorithm TetraFreeQ on several polyhedral domains. All the routines have been implemented in Matlab and are freely available at [19]. The numerical experiments were made using Matlab R2022a, on an Apple MacBook Pro with M1 Chip and 16 GB of RAM. We consider three polyhedra $\Omega_{j}$, $j=1,2,3$ with triangular facets, where $\Omega_{1}$ is nonconvex with 30 facets, $\Omega_{2}$ is convex (an approximation of a sphere with 760 facets) and $\Omega_{3}$ is multiply- connected with 20 facets and a hole (see Figure 1). The domains, as well as the facets and the tetrahedralization (used for comparison), are obtained by Matlab built-in command alphashape, on suitable point clouds. In particular we compute on the three polyhedra: * 1. a reference rule $Q_{T}$ corresponding to a tetrahedralization (cf. Remark 2.2) * 2. the corresponding compressed tetrahedral rule $Q_{TC}$ * 3. the tetrahedra-free rule $Q_{TF}$ produced by Algorithm TetraFreeQ. In our numerical tests, in $(ii1)-(ii4)$ of Algorithm TetraFreeQ we have chosen the parameters $\varepsilon=10^{-14}$, $\theta=4$ and an initial value $K=\lceil 8Nvol(B)/vol(\Omega)\rceil$ where $B$ is the smallest bounding box for $\Omega$. In order to test numerically polynomial exactness of $Q_{TF}$, in Figure 3 we display for each domain the relative errors $E(g_{k})=\frac{|Q_{T}(g_{k})-Q_{TF}(g_{k})|}{|Q_{T}(g_{k})|}$ (14) corresponding to 100 trials of the polynomial $g_{k}(x,y,z)=(a_{k}+b_{k}x+c_{k}y)^{n}$ (15) where $n$ is the algebraic degree of precision of the rule and $(a_{k},b_{k},c_{k})\in[0,1]$ are uniform random coefficients. In the same figures, we have also represented with a circle the logarithmic averages $\sum_{k=1}^{100}{\log(E(g_{k}))/100}$. As additional investigation, we also compute the relative errors of these rules in integrating on $\Omega_{j}$, $j=1,2,3$, the three test functions with different degree of regularity $f_{1}(x,y,z)=\exp(-\|P-P_{0}\|_{2}^{2})\;,\;\;f_{2}(x,y,z)=\|P-P_{0}\|_{2}^{5}\;,\;\;f_{3}(x,y,z)=\|P-P_{0}\|_{2}\;,$ (16) where $P=(x,y,z)$ and $P_{0}(x_{0},y_{0},z_{0})$ was suitably chosen in the domains, more precisely $P_{0}=(1.5,1.5,1.5)$ for $\Omega_{1}$ and $\Omega_{3}$, $P_{0}=(1,1,1)$ for $\Omega_{2}$; cf. Figures 4-6. Moreover, in Table 2 we report the cardinalities of the pointsets from which the tetrahedra-free rules are extracted, varying the degree from 1 to 10. In view of these numerical experiments, we can see that the rules $Q_{TC}$ and $Q_{TF}$ give errors of comparable size, whereas the errors of $Q_{T}$ are smaller. This is not surprising, since the basic tetrahedra-based rule uses a much larger number of nodes than $N$, which turns out to be the number of nodes of both $Q_{TC}$ and $Q_{TF}$, with a ratio increasing with the number of facets (being for example more than 100 with the 760-facets sphere-like polyhedron, cf. Figure 2). We stress once again, however, that $Q_{TF}$ avoids completely any tetrahedralization of the polyhedron, a desiderable feature in many applications (such as polyhedral FEM/VEM methods). Concerning the cputime required by the moment computation in TetraFreeQ, it grows with the algebraic degree of precision and of course with the number of facets, ranging in our tests between $10^{-4}$ and $10^{-3}$ seconds. On the other hand, there is numerical evidence that the bottleneck of TetraFreeQ is due to the in-polyhedron routine and the NNLS algorithm, one dominating on the other depending on the algebraic degree of exactness and on the number of facets, with an overall cputime of the quadrature rule construction varying from $10^{-2}$ seconds for $n=1,2,3$ up to $10^{0}$ seconds for $n=10$. Figure 1: Examples of polyhedral domains. Left: $\Omega_{1}$ (nonconvex, 20 facets); Center: $\Omega_{2}$ (convex, 760 facets); Right: $\Omega_{3}$ (multiply connected, 20 facets). Figure 2: Cardinality versus the exactness degree on the three polyhedra of Figure 1: uncompressed tetrahedra-based rule (blue), compressed tetrahedra- based and tetrahedra-free rule (brown). Figure 3: Relative errors $E(g_{k})$ of the tetrahedra-free rule over 100 polynomial integrands of the form $g_{k}=(a_{k}x+b_{k}y+c_{k}z+d_{k})^{n}$ on the three polyhedra of Figure 1, where $a_{k},b_{k},c_{k},d_{k}$ are uniform random coefficients in $[0,1]$ and $n=1,2,\dots,10$; the circles correspond to the average logarithmic error $\sum_{k=1}^{100}{\log(E(g_{k}))/100}$. Figure 4: Relative errors versus the exactness degree in the integration of the Gaussian $f_{1}$ in (16) on the three polyhedra of Figure 1: uncompressed (blue) and compressed (brown) tetrahedra-based rules, tetrahedra-free rule(yellow). Figure 5: Relative errors versus the exactness degree in the integration of the power function $f_{2}$ in (16) on the three polyhedra of Figure 1: uncompressed (blue) and compressed (brown) tetrahedra-based rules, tetrahedra- free rule(yellow). Figure 6: Relative errors versus the exactness degree in the integration of the distance function $f_{3}$ in (16) on the three polyhedra of Figure 1: uncompressed (blue) and compressed (brown) tetrahedra-based rules, tetrahedra- free rule(yellow). deg | $1$ | $2$ | $3$ | $4$ | $5$ | $6$ | $7$ | $8$ | $9$ | $10$ ---|---|---|---|---|---|---|---|---|---|--- $N$ | $4$ | $10$ | $20$ | $35$ | $56$ | $84$ | $120$ | $165$ | $220$ | $286$ $\Omega_{1}$ | $31$ | $77$ | $160$ | $286$ | $447$ | $676$ | $969$ | $1327$ | $1765$ | $2286$ $\Omega_{2}$ | $28$ | $78$ | $154$ | $273$ | $450$ | $669$ | $956$ | $1313$ | $1757$ | $2287$ $\Omega_{3}$ | $29$ | $77$ | $156$ | $276$ | $446$ | $662$ | $950$ | $1312$ | $1753$ | $2285$ Table 2: Cardinality of the Halton pointsets from which tetrahedra-free rules with $N$ nodes are obtained. ## 4 Acknowledgements Work partially supported by the DOR funds and the biennial project BIRD 192932 of the University of Padova, and by the INdAM-GNCS 2022 Project “Methods and software for multivariate integral models”. This research has been accomplished within the RITA “Research ITalian network on Approximation”, and within the UMI Group TAA “Approximation Theory and Applications” (A. Sommariva). ## References * [1] P.F. Antonietti, P. Houston, G. Pennesi, Fast numerical integration on polytopic meshes with applications to discontinuous Galerkin finite element methods, J. Sci. Comput. 77 (2018) 339–1370. * [2] T.M. Apostol, Calculus, 2nd ed., vol. II, Blaisdell, 1969. * [3] B. Bauman, A. Sommariva, M. Vianello, Compressed cubature over polygons with applications to optical design, J. Comput. Appl. Math., 370, 2020. * [4] L. Beirão da Veiga, F. Dassi, A. Russo, High-order Virtual Element Method on polyhedral meshes, Comput. Math. Appl. 74 (2017) 1110–1122. * [5] M. Dell’Orto, M. Dessole, F. Marcuzzi, The Lawson-Hanson Algorithm with Deviation Maximization: Finite Convergence and Sparse Recovery, Numer. Linear Algebra Appl., accepted upon revision. arXiv:2108.05345v3, 2022. * [6] M. Dessole, F. Marcuzzi, Deviation maximization for rank-revealing QR factorizations, Numer. Algorithms 91 (2022) 1047–1079. * [7] M. Dessole, F. Marcuzzi, M. Vianello, Accelerating the Lawson-Hanson NNLS solver for large-scale Tchakaloff regression designs, Dolomites Res. Notes Approx. DRNA (2020) 20–29. * [8] P.J. Davis, A construction of nonnegative approximate quadratures, Math. Comp. 21 (1967) 578–582. * [9] S. Hayakawa, Monte Carlo cubature construction, Jpn. J. Ind. Appl. Math., 38 (2021) 561–577. * [10] J.S. Hesthaven, G. Rozza, B. Stamm, Certified reduced basis methods for parametrized partial differential equations, Springer, New York, 2016. * [11] S. Holcombe, inpolyhedron \- are points inside a triangulated volume?, 2015, https://www.mathworks.com/matlabcentral/fileexchange/37856-inpolyhedron-are-points-inside-a-triangulated-volume. * [12] J. Jaskowiec, N. Sukumar, High-order cubature rules for tetrahedra, Int. J. Numer. Methods Eng. 121 (2020) 2418–2436. * [13] C.L. Lawson, R.J. Hanson, Solving least squares problems, Classics in Applied Mathematics 15, SIAM, Philadelphia, 1995. * [14] The Mathworks, alphaShape, Polygons and polyhedra from points in 2-D and 3-D, https://www.mathworks.com/help/matlab/ref/alphashape.html. * [15] S.E. Mousavi, N. Sukumar, Numerical integration of polynomials and discontinuous functions on irregular convex polygons and polyhedrons, Comput. Mech. 47 (2011) 535–554. * [16] F. Piazzon, A. Sommariva, M. Vianello, Caratheodory-Tchakaloff Subsampling, Dolomites Res. Notes Approx. DRNA 10 (2017) 5–14. * [17] E. Schönhardt, Über die Zerlegung von Dreieckspolyedern in Tetraeder. Mathematische Annalen, 98:309–312,1928. * [18] L. Shunn, F. Ham, Symmetric quadrature rules for tetrahedra based on a cubic close-packed lattice arrangement, J. Comput. Appl. Math. 236 (2012) 4348–4364. * [19] A. Sommariva, Quadrature and Cubature Codes, available at https://www.math.unipd.it/~alvise/software.html. * [20] A. Sommariva, M. Vianello, Compression of multivariate discrete measures and applications, Numer. Funct. Anal. Optim. 36 (2015) 1198–1223. * [21] A. Sommariva, M. Vianello, Computing Tchakaloff-like cubature rules on spline curvilinear polygons, Dolomites Res. Notes Approx. DRNA 14 (2021) 1–11. * [22] A.H. Stroud, Approximate Calculation of Multiple Integrals, Prentice-Hall, Englewood Cliffs, 1971. * [23] Y. Sudhakar, J.P.M. De Almeida, W.A. Wall, An accurate, robust, and easy-to-implement method for integration over arbitrary polyhedra: application to embedded interface methods, J. Comput. Phys. 273 (2014), 393–415. * [24] Y. Sudhakar, A. Sommariva, M. Vianello, W.A. Wall, On the use of compressed polyhedral quadrature formulas in embedded interface methods, SIAM J. Sci. Comput. 39 (2017) B571–B587. * [25] V. Tchakaloff, Formules de cubatures mécaniques à coefficients non négatifs, (French), Bull. Sci. Math. 81 (1957) 123–134. * [26] M. Tchernychova, Caratheodory cubature measures, Ph.D. dissertation in Mathematics (supervisor: T. Lyons), University of Oxford, 2015. * [27] E.W. Weisstein, Polygon Area, from Wolfram MathWorld, https://mathworld.wolfram.com/PolygonArea.html. * [28] D.R. Wilhelmsen, A Nearest Point Algorithm for Convex Polyhedral Cones and Applications to Positive Linear approximation, Math. Comp. 30 (1976) 48–57.
# Vertex Classification of Planar C-Polygons 111 Illya Ivanov and Cameron Strachan ###### Abstract Given a convex domain $C$, a $C$-polygon is an intersection of $n\geq 2$ homothets of $C$. If the homothets are translates of $C$ then we call the intersection a translative $C$-polygon. This paper proves that if $C$ is a strictly convex domain with $m$ singular boundary points, then the number of singular boundary points a $C$-polygon has is between $n$ and $2(n-1)+m$. For a translative $C$-polygon we show the number of singular boundary points is between $n$ and $n+m$. ## 1 Introduction Let $\mathbb{E}^{2}$ denote the 2-dimensional planar Euclidean space, and $\mathbb{S}^{1}$ the unit sphere inside $\mathbb{E}^{2}$. A set $X\subseteq\mathbb{E}^{2}$ is convex when the line segment connecting any two points in $X$ is contained in $X$. We denote the interior, boundary and convex hull of $X$, as $\text{int}(X)$, $\mathrm{bd}(X)$ and $\rm conv(X)$ respectively. A set $C\subseteq\mathbb{E}^{2}$ is convex domain when it is compact, convex, and has non-empty interior. If the relative interior of the line segment connecting any two boundary points is contained in the interior of $C$, then we call $C$ a strictly convex domain. Let $C$ be a convex domain, the Gauss mapping of $C$, which is denoted as $\Gamma_{C}$, is the mapping between the boundary of $C$ and $\mathbb{S}^{1}$ where each boundary point, $x$, gets mapped to the set of normal unit vectors of the supporting lines of $C$ at the boundary point $x$. The Gauss mapping of any convex domain is always surjective and functional. When the Gauss mapping of $C$ is well defined we call $C$ smooth, and when the Gauss mapping is injective this is equivalent to $C$ being strictly convex. When a convex domain is both strictly convex and smooth we call it a smooth strictly convex domain, which is equivalent to having a bijective function as a Gauss mapping. Let $X_{i}$ be a subset of $\mathbb{E}^{2}$ for each $i\in\\{1,2,\dots,n\\}$, we call $\cap_{i=1}^{n}X_{i}$ a reduced intersection when $\cap_{i=1}^{n}X_{i}\subsetneq\cap_{i=1,i\neq j}^{n}X_{i}$ for every $j\in\\{1,2,\dots,n\\}$. We also denote $X_{1}+X_{2}=\\{x_{1}+x_{2}|x_{1}\in X_{1},x_{2}\in X_{2}\\}$ and $\lambda X_{1}=\\{\lambda x_{1}|x_{1}\in X_{1}\\}$ where $\lambda$ is a positive real. ###### Definition 1. We call a finite intersection of sets, $\cap_{i=1}^{n}X_{i}$ a proper intersection when $\cap_{i=1}^{n}X_{i}$ has non-empty interior and is a reduced intersection. We call $\cap_{i=1}^{n}X_{i}$ an improper intersection otherwise. The main purpose of this paper is to provide a somewhat analogous result to the famous upper bound theorem of McMullen [1], in the planar case. Given a convex polytope in $d$-dimensional space, made by intersecting $n$ halfspaces, the upper bound theorem provides sharp upper bounds for the number of $i$-dimensional faces this polytope can have for $i\in\\{0,1,\dots,d-1\\}$. We investigate the same question one could ask about $C$-polygons which are defined as follows. It should be noted that the results and definitions in this paper are strictly planar, although generalization to higher dimensions can naturally be defined. ###### Definition 2. Given a convex domain, $C\subseteq\mathbb{E}^{2}$, a set of $n$ points $\\{x_{1},x_{2},\dots,x_{n}\\}\subseteq\mathbb{E}^{2}$, and a set of $n$ positive scalars $\\{\lambda_{1},\lambda_{2},\dots,\lambda_{n}\\}\subseteq\mathbb{R}$ where $n\geq 2$, we denote $H_{i}=x_{i}+\lambda_{i}C$ and $T_{i}=x_{i}+C$, and define a $C$-polygon, $H$, and a translative $C$-polygon, $T$, as the following intersections if these intersections are proper. $H:=\cap_{i=1}^{n}H_{i}=\cap_{i=1}^{n}x_{i}+\lambda_{i}C~{}~{}\text{ and }~{}~{}T:=\cap_{i=1}^{n}T_{i}=\cap_{i=1}^{n}x_{i}+C$ We call each $H_{i}$ or $T_{i}$ for $i\in\\{1,2,\dots n\\}$ a generating homothet or generating translate respectively. When $C$ is a ball, then the definitions of $C$-polygons and translative $C$-polygons, define classical notions of generalized ball polygons and ball polygons [3], respectively. We investigate a somewhat dual version of [2] and study the complexity of the boundary structure of $C$-polygons and translative $C$-polygons, and how varied the boundary structure can be if we fix the number of generating translates or homothets. Specifically, we give upper and lower bounds on how many singular boundary points these objects can have when $C$ is a strictly convex domain. We call these singular boundary points the vertices of the $C$-polygon and denote the set of them as $Vert(H)$. The following is the main result established in this paper. ###### Theorem 1. Let $T=\cap_{i=1}^{n}T_{i}$ and $H=\cap_{i=1}^{n}H_{i}$ be a translative $C$-polygon and $C$-polygon respectively. If $C$ is a strictly convex domain with $m$ singular boundary points then $n\leq|Vert(T)|\leq n+m,\text{ and }n\leq|Vert(H)|\leq 2(n-1)+m.$ Our general approach in proving Theorem 1 is to handle the case where $m=0$, or equivalently the case where $C$ is smooth, and then extend this to the case where $m$ is positive. In section 2, we establish some basic properties of $C$-polygons where $C$ is a smooth strictly convex domain. In section 3 we prove Theorem 1 when $m=0$, and in section 4 we prove the case of Theorem 1 where $m$ is positive. Finally in section 5 we explore the sharpness of the bounds established in Theorem 1. ## 2 Basic Properties for C-polygons where C is a smooth strictly convex domain We start with defining a natural face structure for $C$-polygons, where $C$ is a smooth strictly convex domain. It suffices to define them for $C$-polygons as translative $C$-polygons are a particular example of $C$-polygons. Let $H=\cap_{i=1}^{n}H_{i}$ be a $C$-polygon, each generating homothet will be a smooth strictly convex domain, which implies $H$ must also be a strictly convex domain. Since $H$ is a strictly convex domain, its boundary can be separated into singular points and smooth boundary arcs between singular points, we define these to be vertices and edges respectively of our $C$-polygon. However, in order to prove our result we come up with a more useful equivalent characterization of vertices and edges. We can do this because $H$ is not merely a strictly convex domain, it is a proper intersection of $n$ homothets of $C$, which allows for the equivalence of the following definitions with vertices and edges described above. ###### Definition 3. Given a generating homothet $H_{j}$ of a $C$-polygon $H$, the edge family of $H_{j}$ is denoted and defined as $\mathcal{E}_{j}=(\mathrm{bd}(H)\cap\mathrm{bd}(H_{j}))\setminus A.$ The intersection $\mathrm{bd}(H)\cap\mathrm{bd}(H_{j})$ is a union of maximally connected closed boundary arcs of $H_{j}$ and the set $A$ contains all the arcs that are singleton points. We call each maximally connected closed boundary arc of $\mathcal{E}_{j}$ an edge of $H$. We further denote $E_{j}$ as the set whose elements are all the edges in $\mathcal{E}_{j}$. ###### Definition 4. A point $v\in\mathrm{bd}(H)$ is a vertex if it lies in the boundary of at least two generating homothets. (a) (b) (c) Figure 1: Three examples of $C$-polygons. Diagram (c) in Figure 1 shows that there can be cases in which the non- singleton maximally connected component of Definition 3 comes into play. If an edge family has more than one edge in it we call it a multi-edge family, and if it has only one edge we call it a singleton-edge family. In order to show these definitions of vertices and edges align with singular points and smooth boundary arcs of $H$ between singular boundary points, we need the following lemma. ###### Lemma 2. Suppose $H_{1}$ and $H_{2}$ are two homothets of a strictly convex domain, $C$, whose intersection is proper, then $|\mathrm{bd}(H_{1})\cap\mathrm{bd}(H_{2})|=2.$ ###### Proof. Let $H_{1}=x_{1}+\lambda_{1}C$ and $H_{2}=x_{2}+\lambda_{2}C$, first we consider the case when $\lambda_{1}\neq\lambda_{2}$. In this case we can, without loss of generality, choose an origin, $\mathbf{o}\in\mathbb{E}^{2}$, such that $H_{2}=\lambda H_{1}$ for some $\lambda>1$. Since $H_{1}\cap H_{2}$ is a proper intersection, $\mathbf{o}$ must be outside $H_{1}$, and $\mathrm{bd}(H_{1})\cap\mathrm{bd}(H_{2})$ must contain at least two points. If the intersection of the boundary of the two homothets has cardinality greater than two, then we have distinct $q,r,s\in\mathrm{bd}(H_{1})\cap\mathrm{bd}(H_{2})$; notice that these points can’t be colinear as they are boundary points of a strictly convex domain. Naturally, $\lambda q,\lambda r,\lambda s\in\mathrm{bd}(\lambda H_{1})=\mathrm{bd}(H_{2})$. Since $\mathbf{o}\notin H_{1}$, one of the points $q,r,s$ must belong to the positive hull of the other two. Without loss of generality, suppose $s=\mu_{1}q+\mu_{2}r$ where $\mu_{1},\mu_{2}\geq 0$. In fact $\mu_{1},\mu_{2}>0$, suppose for example $\mu_{2}=0$, then we would have the points $q,\mu_{1}q,\lambda q,\lambda\mu_{1}q$ all lie on a straight line and also in $\mathrm{bd}(H_{2})$, a contradiction since at least three out of these four points must be distinct and $H_{1}\cap H_{2}$ is strictly convex. Also note that $\mu_{1}+\mu_{2}\neq 1$, since $q,r,s$ aren’t colinear. Then one can see that if $\mu_{1}+\mu_{2}<1$, then $\lambda s\in\text{int}(\rm conv\left\\{q,r,s,\lambda q,\lambda r\right\\})$ and cannot lie in $\mathrm{bd}(H_{2})$. If in turn $\mu_{1}+\mu_{2}>1$, then $s$ lies in the interior of the convex hull of the other five points, and cannot be in $\mathrm{bd}(H_{2})$. For the other case suppose $\lambda_{1}=\lambda_{2}$, and let $H_{2}=H_{1}+x$. Let $y$ be a normal vector to $x$. The inner product values $\langle q,y\rangle,\langle r,y\rangle,\langle s,y\rangle$ are all distinct, otherwise four of the points $q,r,s,q+x,r+x,s+x$ would lie on a same line and in the boundary of $H_{1}$. Without loss of generality, suppose $\langle q,y\rangle<\langle r,y\rangle<\langle s,y\rangle$. Then $r=\mu_{1}q+\mu_{2}s+kx$ for some $k\neq 0$ and some $\mu_{1},\mu_{2}\in(0,1)$ such that $\mu_{1}+\mu_{2}=1$. If $k<0$, then $r+x\in\text{int}(\rm conv\left\\{r,q,q+x,s,s+x\right\\})$, and if $k>0$, then $r\in\text{int}(\rm conv\left\\{r+x,q,q+x,s,s+x\right\\})$. ∎ We can immediately see that the two points in $\mathrm{bd}(H_{1})\cap\mathrm{bd}(H_{2})$, described in Lemma 2, are singular points in the boundary of $H_{1}\cap H_{2}$. If they weren’t singular then the Gauss image of one intersection point, $x\in\mathrm{bd}(H_{1})\cap\mathrm{bd}(H_{2})$, with respect to the two homothets, will be the same and a singleton unit vector. Since these homothets are strictly convex their Gauss images will be injective, and hence the inverse images of the point $x$ from the two homothets to $C$ will be the same. This implies that the homothets are subsets of each other a contradiction to the fact their intersection was assume to be proper. Turning onto the implications Lemma 2 has to $C$-polygons, first we can immediately see that any $C$-polygon will have a finite amount of vertices. As a vertex is a boundary point that is in the boundary of two generating homothets, and so the set of vertices is a subset of the set of all boundary intersections between each pair of homothets. Which if we have $n$ generating homothets we have at most $2\binom{n}{2}$ many boundary intersections between all the pairs of generating homothets implying $|Vert(H)|\leq 2\binom{n}{2}$. Since we are in the plane, this also implies the number of edges a $C$-polygon has is finite. We can also notice that if we have a boundary point $x\in\mathrm{bd}(H)$, then it must be in the boundary of some generating homothet $H_{j}$. This means we can classify boundary points of $H$ into two cases, the boundary points that are in the boundary of only one generating homothet, and boundary points that are in the boundaries of multiple generating homothets. We will soon see that the latter will be vertices and the former will be elements in the relative interior of edges. The next two lemmas establish that vertices are indeed singular points and edges are smooth boundary arcs of $H$ between vertices. ###### Lemma 3. Let $H$ be a $C$-polygon where $C$ is a smooth strictly convex domain, then a point $v\in\mathrm{bd}(H)$ is a vertex if and only if it’s a singular point of $H$. ###### Proof. Suppose $v\in\mathrm{bd}(H)$ is a vertex, this means it is in the boundary intersection of two generating homothets call them $H_{i}$ and $H_{j}$. By construction we have that $H\subseteq H_{i}\cap H_{j}$ and that $v$ is a boundary element of both sets. From this we obtain that $\Gamma_{H_{i}\cap H_{j}}(v)\subseteq\Gamma_{H}(v)$, as supporting lines of $H_{i}\cap H_{j}$ will support subsets of $H_{i}\cap H_{j}$ that they intersect with. We established $\Gamma_{H_{i}\cap H_{j}}(v)$ is not a singleton point and hence $\Gamma_{H}(v)$ will also not be a singleton, showing that vertices are singular points of $H$. Conversely suppose $v\in\mathrm{bd}(H)$ is a singular point, then as it’s a boundary point it must be either a boundary element of one generating homothet or multiple. If it’s the latter we are done, so suppose this singular boundary point was only in the boundary of one generating homothet $H_{j}$, and in the interior of all the other generating homothets. Then the boundary points of $H$ arbitrarily close to this point must also be only contained in the boundary of $H_{j}$, implying $H_{j}$ is not smooth, a contradiction. ∎ ###### Lemma 4. Edges are smooth boundary arcs between vertices and vice versa. ###### Proof. For the forward implication consider an edge in an edge family, this is a non- singleton maximally connected boundary arc of some homothet that is contain in $H$. Since this arc is in the boundary of one homothet it must be in the boundary of $H$, this arc is also smooth as $C$ is smooth and it’s end point must be vertices otherwise it would not be maximally connected. For the reverse implication consider an arbitrary smooth boundary arc between two vertices $v_{1},v_{2}\in Vert(H)$, and an arbitrary point, $x$, in this arc. It is a smooth boundary point by assumption which implies $x$ must only be contained in the boundary of one homothet, $H_{j}$. Notice that since the arc is smooth, if one point in this smooth arc is in the boundary of only one homothet, the entire smooth arc must be as well. If there was a transition to another generating homothet along this arc, we must have a transition point. This is because boundaries are closed, and so their intersection must have an overlapping point, and that point would be a vertex and thus singular. So the entire arc belongs to only one homothet, this arc is also contain in the boundary of $H$ and is maximally connected and is not a singleton point which means it’s an edge in the edge family $E_{j}$. ∎ ## 3 Proof of Theorem 1, the case when m = 0 There are two statements to prove in Theorem 1, we will first prove the $C$-polygon statement and then the translative $C$-polygon statement. Please note throughout this entire section $C$ will be a smooth strictly convex domain. ### 3.1 The C-polygon case. Let $H=\cap_{i=1}^{n}H_{i}$ be a $C$-polygon, we need to show that $Vert(H)$ has cardinality between $n$ and $2(n-1)$. The lower bound is easily handled with an upcoming observation, so we will focus on the upper bound. We will prove this bound by induction on $n$, the number of generating homothets. We start with the base case where $n=2$, we aim to show that $|Vert(H)|\leq 2=2(n-1)$. We have already seen in Lemma 2, that $|Vert(H)|=|Vert(H_{1}\cap H_{2})|=|\mathrm{bd}(H_{1})\cap\mathrm{bd}(H_{2})|=2$ completing the base case. Our inductive hypothesis is that any $C$-polygon made up of $n$ generating homothets will have at most $2(n-1)$ vertices. We aim to show that any $C$-polygon made up of $n+1$ generating homothets will have at most $2n$ vertices. To that end let $H=\cap_{i=1}^{n+1}H_{i}$ be a $C$-polygon made up of $n+1$ generating homothets. The object we will be applying our inductive hypothesis on is denoted and defined as $W_{j}=\bigcap_{i=1,i\neq j}^{n+1}H_{i}$. The notation $W_{j}$ is meant to invoke that $H$ is without the $j$’th homothet in the intersection creating it. This is in fact a $C$-polygon as $n\geq 2$, it has non-empty interior as it’s a super-set of $H$ and is a reduced intersection by the following lemma. ###### Lemma 5. If $U_{i}\subseteq\mathbb{E}^{d}$ for all $i\in\\{1,2,\dots,n\\}$, and $\cap_{i=1}^{n}U_{i}$ is reduced and $n\geq 2$, then for all $j\in\\{1,2,\dots,n\\}$, $\cap_{i=1,i\neq j}^{n}U_{i}$ is reduced. ###### Proof. Let $U=U_{1}\cap U_{2}\cap\dots\cap U_{n}$, and $S\subseteq\left\\{1,\dots,n\right\\}$. We denote $W_{S}=\bigcap_{i\in\left\\{1,\dots,n\right\\}\setminus S}U_{i},$ as defined before, $W_{\\{j\\}}=W_{j}$. Note that for any $i\in\\{1,2,\dots,n\\}$ we have $U=W_{i}\cap U_{i}$. By assumption, $U$ is reduced, which is equivalent to $U\neq W_{i}$ for any $i\in\\{1,2,\dots,n\\}$. To show that $W_{i}$ is also reduced for any $i\in\\{1,2,\dots,n\\}$, we need to demonstrate that for any distinct $i,j\in\\{1,2,\dots,n\\}$, we have $W_{i}\neq W_{\left\\{i,j\right\\}}$. Indeed if $W_{i}=W_{\left\\{i,j\right\\}}$, we would have $U=W_{i}\cap U_{i}=W_{\left\\{i,j\right\\}}\cap U_{i}=W_{j}$. Thus, we contradict the initial assumption that $U$ is reduced. ∎ An important property of $W_{j}$ is, if $H_{j}$ is a generating homothet of a $C$-polygon $H$, then $\mathrm{bd}(H_{j})\cap\mathrm{int}(W_{j})\neq\emptyset$ (1) This can easily be proven by contradiction due to the fact $H$ is a proper intersection. Another important observation is that the amount of boundary arcs of $H_{j}$ that cross the interior of $W_{j}$ is the amount of edges in the edge family $E_{j}$. With this observation we can immediately prove the lower bound in Theorem 1 as each edge family has cardinality at least 1 by equation (1). Notice that any new vertices found in $H$ that are not in $W_{j}$ must be in the relative boundary of the edges in the edge family $E_{j}$. $W_{j}$$H_{j}$ $W_{j}$$H_{j}$ ‘ Figure 2: Two examples of how $H_{j}$ can intersect $W_{j}$. To complete the inductive step we will show that every $C$-polygon will contain a singleton-edge family. This would complete the induction because if we have this, lets call the singleton-edge family $E_{k}$, then the $\mathrm{bd}(H_{k})$ intersects the interior of $W_{k}$ in one and only one boundary arc. This means that the new vertices added on to $W_{k}$ by the inclusion of $H_{k}$ in the intersection to form $H$ will be at most two, with this we obtain: $|Vert(H)|\leq|Vert(W_{j})|+|Vert(H)\setminus Vert(W_{j})|\leq 2(n-1)+2=2n$ So all we need to show is that every $C$-polygon contains a singleton-edge family. We do this by utilizing a notion of gaps which are defined as follows. ###### Definition 5. Let $H_{j}$ be a generating homothet of $H$, then consider $\mathrm{bd}(H_{j})\setminus\mathcal{E}_{j}$ which is a union of $|E_{j}|$ many maximally connected open boundary arcs of $H_{j}$. Let us denote these open boundary arcs of our generating homothet by $g_{1},g_{2},\dots,g_{|E_{j}|}$. We will call $\rm conv({\rm cl}(g_{k}))$ a gap of $H_{j}$ for $k\in\\{1,2,\dots,|E_{j}|\\}$ and we denote and call $G_{j}=\cup_{i=1}^{|E_{j}|}\\{conv({\rm cl}(g_{i}))\\}$ the gap family of $H_{j}$. Figure 3: Three examples of gap families of a homothet. ###### Lemma 6. For any $j\in\\{1,2,\dots n+1\\}$: $(\mathrm{bd}(H)\setminus E_{j})\subseteq G_{j}$ ###### Proof. What this lemma says is that the boundary arcs of $H$ connecting the edges generated by a particular homothet, must lie in the gaps of that homothet. This can be proven by observing that the line segment connecting the two vertices bounding a gap must be in the interior of $H$ because $H$ is strictly convex, and the boundary of $H$ cannot go outside of the boundary of $H_{j}$ since $H\subseteq H_{j}$. ∎ ###### Lemma 7. If $H_{i}$ and $H_{j}$ are two distinct generating homothets of $H$, then the intersection of $\mathrm{bd}(H_{i})$ with the gap family of $H_{j}$ is non- empty and is contained in a single gap of $H_{j}$. ###### Proof. $H_{i}$ must contain the edge family $E_{j}$, and so when the $\mathrm{bd}(H_{i})$ intersects a gap, its boundary must enter and exit the gap from the boundary of $H_{j}$. The homothets cannot be tangent to each other as that would contradict Lemma 2, so when the $\mathrm{bd}(H_{i})$ intersects a gap it pierces the interior of it, and the $\mathrm{bd}(H_{i})$ must intersect at least one gap in order for the intersection to be reduced. So every gap $\mathrm{bd}(H_{i})$ intersects generates two intersection points in $\mathrm{bd}(H_{i})\cap\mathrm{bd}(H_{j})$, but of course Lemma 2 implies this is at most two proving the lemma. Figure 4: Example of what would happen if a homothet intersected multiple gaps. ∎ As a result of the previous 2 lemmas we see the following. ###### Corollary 8. No two edges in the same edge family may lie in different gaps of a gap family. It is with this final corollary that we are able to show that every $C$-polygon will have a singleton-edge family. First as $n+1\geq 3$ we have at least three edge families in $H$. Consider an arbitrary edge family $E_{k}$, if this is a singleton-edge family we are done so suppose it’s a multi-edge family, consider an arbitrary gap of the homothet $H_{k}$. This gap must have at least one edge in it as the boundary of $H$ connecting the two vertices bounding a gap, is a closed connected curve made out of edges. These edges connecting the vertices must also be contained in the gap of $H_{k}$ according to Lemma 6. Take the edge adjacent to one of the vertices bounding the gap, that is also contained in the gap of $H_{k}$ and let $E_{l}$ be the edge family of this edge. If this is a singleton-edge family we are done, if it’s a multi-edge family then the edges of this edge family must be contained in the gap of $H_{k}$ by Corollary 8. So there is a gap of $H_{l}$ that is contained in the gap of $H_{k}$ which we can repeat this process with. Since we have a finite amount of edges this process must end and so we must obtain a singleton-edge family eventually. $H_{k}$$H_{l}$ Figure 5: Depiction of the process of how to find a singleton- edge family. The following theorem is the most general statement this section has proven for the case where $m=0$. Notice we didn’t use the fact that the generating homothets of our $C$-polygon are homothets of $C$ besides in deriving Lemma 2. So if we assume our intersection satisfies Lemma 2, and remove the restriction that the domains we are intersecting are homothets, our proof will remain identical. ###### Theorem 9. If $X=\cap_{i=1}^{n}X_{i}$ where $X_{i}$ is a smooth strictly convex domain for each $i\in\\{1,2,\dots,n\\}$, where $X$ is a proper intersection, and for every distinct $i,j\in\\{1,2,\dots,n\\}$, $|\mathrm{bd}(X_{i})\cap\mathrm{bd}(X_{j})|=2$, then the cardinality of singular points on $X$ is between $n$ and $2(n-1)$. ### 3.2 The translative C-polygon case: First it should be noted that all the above results established in section 2 and 3.1 apply to translative $C$-polygons as well, this implies that a translative $C$-polygon has at least as many edges as generating translates, since every edge family is non-empty. This implies the desired lower bound for translative $C$-polygons. For the upper bound we prove it by induction in an analogous way to the $C$-polygon case done previously, as a result we omit some details. Our base case is also handled by Lemma 2, so let $T=\cap_{i=1}^{n+1}T_{i}$ be a translative $C$-polygon where $C$ is a smooth strictly convex domain. Our inductive hypothesis is that any translative $C$-polygon made up of $n$ generating translates will have at most $n$ vertices, we wish to show $T$ has at most $n+1$ vertices. We already know there must be a singleton-edge family by section 3.1, let $E_{j}$ be that singleton-edge family. In order to complete the proof we only need to show that the inclusion of $T_{j}$ to $W_{j}$, to form $T$, net total increases the vertex count by at most one. We have seen that $T_{j}$ is an edge family that increases our count of vertices by at most two in section 3.1. However, in our translative case we will show we must exclude at least one of the vertices of $W_{j}$ and hence net total can only increase our vertex count by at most one. We do this with the help of the following lemma. ###### Lemma 10. Let $T_{1}$ and $T_{2}$ be translates of $C$ that produce a proper intersection, then the boundary of $T_{1}$ in the exterior of $T_{2}$ has a Gauss image that contains a hemisphere. ###### Proof. $\mathbb{S}^{1}$$\tau$$\rho$$H_{\rho}$$H_{-\rho}$ Figure 6: Intersection between two translates of $C$. First let $\tau\in\mathbb{S}^{1}$ be the direction of the translation to get from $T_{1}$ to $T_{2}$, and let $\pm\rho\in\mathbb{S}^{1}$ be the two perpendicular vectors of $\tau$. Since $C$ has a bijective function for its Gauss mapping, $T_{1}$ and $T_{2}$ will each have two unique boundary points with Gauss image $\pm\rho$. Notice by the choice of $\rho$ the support lines will be equal, and both translates will be contained in the band between these two support lines, which is depicted in Figure 6. Each open boundary arc of $T_{1}$ and $T_{2}$ between the two points of contact with the support lines, will have a Gauss image of an open hemisphere on $\mathbb{S}^{1}$ centered at $\pm\tau$. Lemma 2 implies only two of these arcs have a boundary intersection with the other translate proving the lemma. ∎ This lemma also implies that the relative interior of any edge has a Gauss image that is contained in an open hemisphere. Since $T_{j}$ has one only edge in its edge family, it has only one maximally connected boundary arc that intersects the interior of $W_{j}$. This boundary arc will enter and exit the boundary of $W_{j}$ in two places and the only way it can include all vertices of $W_{j}$ is if the boundary of $T_{j}$ enters and leaves the boundary of $W_{j}$ in the relative interior of same edge. This would contradict Lemma 10 as the relative interior of edges have a Gauss image that is contained in a hemisphere, but the pairwise intersection between $T_{j}$ and the translate that generates the edge of $W_{j}$ that $\textrm{bd}(T_{j})$ intersects, violates Lemma 10. This proves the induction as we know $W_{j}$ has $n$ vertices by our inductive hypothesis and the inclusion of $T_{j}$ to $W_{j}$ increases our vertex count by at most one implying $|Vert(T)|=|Vert(W_{j}\cap T_{j})|\leq n+1$. $W_{j}$$H_{j}$ Figure 7: Diagram depicting that in order to contain all vertices of $W_{j}$ we need a homothetic copy of $C$ and not a translative one. ## 4 Proof of Theorem 1: the positive m case When we consider a $C$-polygon, $H=\cap_{i=1}^{n}H_{i}$, where $C$ has $m$ singular points, things behave similarly to the smooth case. Notice Lemma 2 holds for strictly convex domains so we know that the boundary intersection between pairs of homothets will have a cardinality of two, and these two points are singular in the intersection body. We define edge families identically although now edges in edge families may not be smooth as boundary arcs of $C$ may not be smooth. This observation lets us classify vertices of $H$ into one of two kinds. Pairwise vertices are singular points of $H$ that are in the boundary of two generating homothets. Inherited vertices are singular points that are in the relative interior of an edge. Notice during the proofs in section 3 our arguments did not use the fact that our edges were smooth. This implies that we have at most $2(n-1)$ many pairwise vertices in the $C$-polygon case, and $n$ many pairwise vertices in the translative $C$-polygon case. Notice that each homothet has a non-empty edge family implying our lower bound for both the $C$-polygon and translative $C$-polygon cases. Edges in an edge family will be non-singleton maximally connected components of a generating homothet. We can view these edges as maximally connected components of the boundary of $C$, by inverting the homothet map. The inherited singular points on an edge will corresponds with a singular point of $C$ since the inverted homothetic image of $C$ preserve the Gauss image. Since $H$ is strictly convex it’s Gauss mapping is injective which implies when we represent the edges of $H$ on $C$, they will not overlap. This implies $H$ has at most $m$ inherited vertices completing the proof. ## 5 Sharpness of Results In this section we discuss the sharpness of the bounds established in Theorem 1. Given a class of convex domains $\mathcal{C}$, and a bound on the number of vertices a $C$-polygon has when $C\in\mathcal{C}$, we say the bound is strongly sharp when for all $C\in\mathcal{C}$ and every natural $n\geq 2$ there is a $C$-polygon with $n$ generating homothets that realizes the bound. If there exists $C\in\mathcal{C}$ such that for every $n\geq 2$ we can find a $C$-polygon with $n$ generating homothets that can realize the bound, we call the bound weakly sharp. ###### Claim 11. The upper bound of $2(n-1)+m$ vertices for $C$-polygons where $C$ is a strictly convex domain is strongly sharp. ###### Proof. Given any strictly convex domain, $C$, with $m$ singular points, we must have a smooth strictly convex boundary arc of $C$. Pick $n-1$ many points on the relative interior of this boundary arc, by expanding $C$ with a homothet centre at each of these boundary points we can create $n-1$ homothets to be arbitrarily close to the supporting lines of the $n-1$ boundary points of $C$. Then we move these $n-1$ homothets arbitrarily inward to pierce the interior of $C$ and create $2$ pairwise vertices with each homothet giving us $2(n-1)+m$ vertices. Figure 8 depicts the construction. Figure 8: Sharpness of upper bound construction for $C$-polygons where $C$ is a strictly convex domain. ∎ ###### Claim 12. The lower bound of $n$ vertices for $C$-polygons and translative $C$-polygons, where $C$ is a strictly convex domain, is weakly sharp. However, the lower bound of $n$ vertices for $C$-polygons and translative $C$-polygons is strongly sharp when $C$ is a smooth strictly convex domain. ###### Proof. The latter part of the claim is trivial as any translative $C$-polygon with $n$ generating translates, where $C$ is a smooth strictly convex domain, will have exactly $n$ vertices according to Theorem 1. For the former part of the claim, the previous sentence shows that the bound is at least weakly sharp. So we need only show this bound is not strongly sharp which means we need to show there exists a strictly convex domain $C$ and a natural $n\geq 2$ where every $C$-polygon with $n$ generating homothets has more than $n$ vertices. We need only choose $n=2$ and $C$ as depicted in Figure 9. We construct $C$ by intersecting three circles, this is a ball-polygon and we choose it such that the antipodal point of the Gauss image of any boundary point of $C$ that is smooth, is contained in the Gauss image of a vertex of $C$. By Lemma 2 we know any intersection of two homothets of $C$ will have two pairwise vertices. So we need to show every $C$-polygon generated by two homothets of $C$ will have at least one inherited vertex. The following statement completes the proof and is left to the reader. Given an intersection of two homothets, $H_{1}\cap H_{2}$, of a strictly convex domain, there exists a point in the relative interior of each of the two edges, call them $x_{1}\in\mathrm{bd}(H_{1})$ and $x_{2}\in\mathrm{bd}(H_{2})$, such that $\Gamma_{H_{1}}(x_{1})$ contains a point that is antipodal to a point in $\Gamma_{H_{2}}(x_{2})$. Figure 9: Construction of $C$. ∎ ###### Claim 13. For every natural $n\geq 2$ there exist a convex domain, $C$, and a $C$-polygon, $H,$ with $n$ generating homothets such that $H$ has zero vertices. ###### Proof. First notice that this condition on the bound of zero vertices is weaker than weak sharpness. Given an arbitrary natural $n\geq 3$ we choose our convex domain $C$ to be a polygon with $n$ vertices that have sufficiently small rounded corners. To construct a $C$-polygon with $n$ generating homothets we take $C$ and enlarge the other $n-1$ copies of $C$ to smoothly transition with $n-1$ many smoothed corners of $C$ as depicted in Figure 10 for the case where $n=3$ and $4$. For the case where $n=2$ two translates of the rounded square can construct a translative $C$-polygon with zero vertices. Figure 10: Depiction of the construction of a smooth $C$-polygon. ∎ We end this paper with an open problem. ###### Problem 14. If $C$ is a convex domain with $m$ singular points, then the number of singular points a $C$-polygon with $n$ generating domains will have is at most $2(n-1)+m$. ## References * [1] P. McMullen, On the upper-bound conjecture for convex polytopes, J. Combinat. Theorey, Ser.B. 10 (1971), 187–200. * [2] P. K. Agarwal, J. Pach, and M. Sharir, State of the Union (of Geometric Objects), Technical report, American Mathematical Society (2008) * [3] K. Bezdek, Z. Langi, M Naszodi, and P. Papez, Ball-Polyhedra, Discrete Comput Geom 38 (2007), 201–230. Illya Ivanov Department of Mathematics and Statistics, University of Calgary, Canada E-mail<EMAIL_ADDRESS> Cameron Strachan Department of Mathematics and Statistics, University of Calgary, Canada E-mail<EMAIL_ADDRESS>
# Values of binary partition function represented by a sum of three squares Bartosz Sobolewski and Maciej Ulas ###### Abstract. Let $m$ be a positive integer and $b_{m}(n)$ be the number of partitions of $n$ with parts being powers of 2, where each part can take $m$ colors. We show that if $m=2^{k}-1$, then there exists the natural density of integers $n$ such that $b_{m}(n)$ can not be represented as a sum of three squares and it is equal to $1/12$ for $k=1,2$ and $1/6$ for $k\geq 3$. In particular, for $m=1$ the equation $b_{1}(n)=x^{2}+y^{2}+z^{2}$ has a solution in integers if and only if $n$ is not of the form $2^{2k+2}(8s+2t_{s}+3)+i$ for $i=0,1$ and $k,s$ are non-negative integers, and where $t_{n}$ is the $n$th term in the Prouhet-Thue-Morse sequence. A similar characterization is obtained for the solutions in $n$ of the equation $b_{2^{k}-1}(n)=x^{2}+y^{2}+z^{2}$. ###### Key words and phrases: binary partitions, recurrence sequence, automatic sequence, sums of squares The research of the authors is supported by the grant of the National Science Centre (NCN), Poland, no. UMO-2019/34/E/ST1/00094 ## 1\. Introduction Let $\mathbb{N}$ be the set of non-negative integers and $\mathbb{N}_{+}$ the set of positive integers. Moreover, for a given $n\in\mathbb{Z}$ we define the 2-adic valuation of $n$ as $\nu_{2}(n)=\max\\{k\in\mathbb{N}:\;2^{k}|n\\},$ with the convention that $\nu_{2}(0)=+\infty$. The problem of representation of integers by quadratic forms or, more generally, by forms or polynomials in many variables, is a classical one. As was proved by Lagrange in 1770, each non-negative integer can be represented as a sum of four squares. On the other hand, there are infinitely many positive integers which cannot be represented by three squares. More precisely, in 1798 Legendre proved that a non-negative integer $N$ can be represented as $N=x^{2}+y^{2}+z^{2}$ for some $x,y,z\in\mathbb{Z}$ if and only if $N$ is not of the form $4^{r}(8s+7)$ for $r,s\in\mathbb{N}$. In particular, the natural density of the set of integers that cannot be represented by a sum of three squares is equal to $1/6$. This raises an interesting question whether for a given sequence of integers $(u_{n})_{n\in\mathbb{N}}$ there exist infinitely many solutions of the Diophantine equation (1) $u_{n}=x^{2}+y^{2}+z^{2}.$ To characterize the solutions of (1) it is necessary to have a good understanding of the 2-adic behavior, or, to be more precise, the 2-adic valuation of the terms of the sequence $(u_{n})_{n\in\mathbb{N}}$. Especially interesting is the case where $u_{n}$ has a combinatorial meaning, i.e., $u_{n}$ counts some discrete objects or structures. Equation (1) with $u_{n}=\binom{2n}{n}$ was investigated by Granville and Zhu in [9]. They characterized those $n\in\mathbb{N}$ such that (1) has no solutions in $x,y,z$. In particular, the set of integers $n$, for which $\binom{2n}{n}$ can be represented as a sum of three squares, has asymptotic density $7/8$ in the set of all natural numbers (for an early study of this problem, see also the paper of Robbins [18]). In the same paper they also obtained a characterization of $n\in\mathbb{N}$ such that $n!$ is not a sum of three squares, given in terms of existence of certain patterns in the binary expansion of $n$. A different approach to this problem (via substitutions) was presented by Deshouillers and Luca [8]. They showed that the natural density of those $n$ such that $n!=x^{2}+y^{2}+z^{2}$ exists and is equal to $7/8$. More precisely, they proved that $\\#\\{n\leq x:\;n!\;\mbox{is a sum of three squares}\\}=\frac{7}{8}x+O(r(x)),$ where $r(x)=x^{2/3}$. The error term was improved by Hajdu and Papp [15] to $r(x)=x^{1/2}\log^{2}x$, and recently by Burns to $r(x)=x^{1/2}$ (see the preprint [6]). On the other hand, Robbins obtained a precise characterzation of the solutions in $n$ of (1) in the case where $u_{n}$ is the $n$-th term of Fibonacci or Lucas sequence [17]. In this paper, we follow the same line of research and consider, in particular, equation (1) with $u_{n}$ being the binary partition function $b(n)$. More precisely, $b(n)$ counts the number of partitions of $n\in\mathbb{N}$ into parts being powers of two. For example, $b(4)=4$ because $4=2^{2}=2+2=1+1+2=1+1+1+1$ are all possible representations of 4 as a sum of powers of $2$. The sequence $(b(n))_{n\in\mathbb{N}}$ was already introduced by Euler. However, it seems that the first serious study of its properties was performed by Churchhouse [7] in 1969. He computed the 2-adic valuation of $b(n)$. Further results, motivated by Churchhouse’s computations, were obtained independently by Gupta [11, 12, 13] and Rødseth [19] (see also recent studies by Rødseth and Sellers [20]). Besides being interesting in its own sake, the study of the equation $b(n)=x^{2}+y^{2}+z^{2}$ connects two different areas: the Diophantine equations and partitions theory. According to our best knowledge, there is no result in the literature providing a characterization of the solutions of equation (1) with $u_{n}$ being a partition function of sub-exponential growth. Recall that $b(n)$ is indeed of sub-exponential growth. More precisely, Mahler [16] proved that $\log_{2}b(n)\sim\frac{1}{2}(\log_{2}n)^{2}$. Our study can be seen as a continuation and extension of recent research concerning solvability of Diophantine equations involving partitions, conducted by Tengely and Ulas [21]. In the same context, we also study the $m$-colored binary partition function $b_{m}(n)$, which counts binary partitions of $n$, where each part can have one of $m\geq 1$ colors. In particular, we have $b(n)=b_{1}(n)$. The study of arithmetic properties of this function was initiated in a paper of Gawron, Miska and Ulas [10], where several useful results were obtained. Let us describe the content of the paper in some detail. In Section 2 we recall some basic properties and results concerning the function $b_{m}(n)$. The main goal of the present investigation, pursued in Sections 3–5, is to obtain an explicit characterization of the set $S_{m}=\\{n\in\mathbb{N}:b_{m}(n)\neq x^{2}+y^{2}+z^{2}\text{ for any }x,y,z\in\mathbb{Z}\\}.$ The reason for considering terms _not_ represented as a sum of three squares is that the description turns out to be more concise (similarly as for non- negative integers). We focus primarily on the case $m=2^{k}-1$ for $k\in\mathbb{N}_{+}$ due to the fact that the $2$-adic valuation $\nu_{2}(b_{2^{k}-1}(n))$ is known to be bounded [7, 10]. This allows us to determine which terms $b_{2^{k}-1}(n)$ are of the form $4^{r}(8s+7)$ through the reduction modulo a suitable power of $2$. The analysis is divided into three cases: $k=1,k=2$, and $k\geq 3$, covered in Section 3, 4, and 5, respectively. Based on these results, in Section 6 we give quite precise bounds for the counting function of the set $S_{2^{k}-1}$, and determine its natural density in the process. In the final section, we state some questions, problems, and conjectures which may serve as a basis for further study. In particular, we discuss the equation $b_{m}(n)=x^{2}+y^{2}+z^{2}$ when $m\neq 2^{k}-1$ as well as some interesting findings concerning the behavior of $b_{m}(n)$ modulo powers of $2$. We also collect results of numerical computations related to the equations $b(n)=x^{2}+y^{2}+z^{4}$ and $b(n)=x^{2}+y^{2}$. ###### Remark 1.1. In this paper, we focus on the problem of representation of $b_{2^{k}-1}(n)$ as a sum of three squares. However, it is possible to use our findings to obtain similar results for other ternary quadratic forms $q$ such that the set of $n\in\mathbb{N}$ represented by $q$ is given in terms of the binary expansion of $n$. Such quadratic forms include, for example, the forms $x^{2}+y^{2}+2z^{2},x^{2}+2y^{2}+2z^{2},x^{2}+2y^{2}+4z^{2}$. ## 2\. Preliminaries In this section we collect known properties and results which will be used throughout the paper. Recall that the ordinary generating function of the sequence $(b(n))_{n\in\mathbb{N}}$ has the form $B(x)=\prod_{n=0}^{\infty}\frac{1}{1-x^{2^{n}}}=\sum_{n=0}^{\infty}b(n)x^{n}.$ As a consequence, we see that $B(x)$ satisfies a Mahler-type functional equation $(1-x)B(x)=B(x^{2})$. Comparing the coefficients on both sides, we see that the sequence $(b(n))_{n\in\mathbb{N}}$ satisfies the recurrence: $b(0)=b(1)=1$ and $b(2n)=b(2n-1)+b(n),\quad b(2n+1)=b(2n).$ Churchouse [7] obtained a characterization of the $2$-adic valuation of the terms $b(n)$. ###### Theorem 2.1. For all $n\geq 2$ we have $\nu_{2}(b(n))=\begin{cases}1&\text{if }\nu_{2}(n)\equiv 0\mkern 4.0mu({\operator@font mod}\mkern 6.0mu2),\\\ 2&\text{if }\nu_{2}(n)\equiv 1\mkern 4.0mu({\operator@font mod}\mkern 6.0mu2).\end{cases}$ Another useful result was obtained independently by Rødseth [19] and Gupta [11], proving a conjecture of Churchhouse. More precisely, we have the following theorem. ###### Theorem 2.2. For all $s\in\mathbb{N}$ and odd $n\in\mathbb{N}$ the following congruence holds: $b(2^{s+2}n)\equiv b(2^{s}n)\mkern 4.0mu({\operator@font mod}\mkern 6.0mu2^{\mu(s)}),$ where $\mu(s)=\left\lfloor\frac{3s+4}{2}\right\rfloor$. For $m\in\mathbb{N}_{+}$ we define the sequence $(b_{m}(n))_{n\in\mathbb{N}}$ as the convolution of $m$ copies of $(b(n))_{n\in\mathbb{N}}$. More precisely, $b_{m}(n)=\sum_{i_{1}+\ldots+i_{m}=n}b(i_{1})\cdots b(i_{m}).$ It is clear that the generating function $B_{m}(x)$ of the sequence $(b_{m}(n))_{n\in\mathbb{N}}$ is the $m$-th power of $B(x)$, and $b_{m}(n)$ also has a combinatorial interpretation. Indeed, $b_{m}(n)$ is the number of binary partitions of $n$, where each part has one of $m$ possible colors. In a recent paper by Gawron, Miska, and Ulas [10], it is proved that for $m=2^{k}-1$ and $n\geq 2^{k}$ the 2-adic valuation of $b_{m}(n)$ belongs to the set $\\{1,2\\}$ . More precisely, they gave the following characterization of the $2$-adic valuation of the terms $b_{2^{k}-1}(n)$. ###### Theorem 2.3 (Theorem 4.6 in [10]). Let $k\in\mathbb{N}_{+}$. For $n,i\in\mathbb{N}$ such that $i<2^{k+2}$ we have $\nu_{2}(b_{2^{k}-1}(2^{k+2}n+i))=\begin{cases}\nu_{2}(b(8n))&\text{if }0\leq i<2^{k},\\\ 1&\text{if }2^{k}\leq i<2^{k+1},\\\ 2&\text{if }2^{k+1}\leq i<3\cdot 2^{k},\\\ 1&\text{if }3\cdot 2^{k+1}\leq i<2^{k+2}.\end{cases}$ In particular, $\nu_{2}(b_{2^{k}-1}(n))\in\\{0,1,2\\}$ and $\nu_{2}(b_{2^{k}-1}(n))=0$ if and only if $n<2^{k}$. The reciprocal of $B(x)$, denoted by $T(x)=\frac{1}{B(x)}=\prod_{n=0}^{\infty}\left(1-x^{2^{n}}\right)=\sum_{n=0}^{\infty}t_{n}x^{n},$ is the ordinary generating function for the famous Prouhet-Thue-Morse sequence $(t_{n})_{n\in\mathbb{N}}$ (PTM sequence for short). Recall that $t_{n}=(-1)^{s_{2}(n)}$, where $s_{2}(n)$ is the number of $1$’s in the unique expansion of $n$ in base 2. Equivalently, we have $t_{0}=1$ and $t_{2n}=t_{n},\quad t_{2n+1}=-t_{n},\quad n\geq 0.$ The formula in Theorem 2.1 can then be written as $\nu_{2}(b(n))=\frac{1}{2}|t_{n}-2t_{n-1}+t_{n-2}|$. We also consider a variant of the PTM sequence, given by $T_{n}=s_{2}(n)\mkern 4.0mu({\operator@font mod}\mkern 6.0mu2),$ i.e., the sequence $(T_{n})_{n\in\mathbb{N}}$ is related to the PTM sequence by $t_{n}=1-2T_{n}$. The PTM sequence is an example of an automatic sequence. More precisely, let $k\geq 2$ be a fixed integer. A sequence $\mathbf{a}=(a_{n})_{n\in\mathbb{N}}$ is called $k$-automatic if its $k$-kernel, namely $K_{k}(\mathbf{a})=\\{(a_{k^{j}n+i})_{n\in\mathbb{N}}:j\in\mathbb{N},0\leq i<k^{j}\\},$ is a finite set. Equivalently, $\mathbf{a}$ is $k$-automatic if there exists a deterministic finite automaton with output (DFAO) that reads the canonical base-$k$ representation of $n$ and the outputs $a_{n}$. In the case of the PTM sequence, it is clear that $K_{k}(\mathbf{t})=\\{\mathbf{t},-\mathbf{t}\\}$. Equivalently, the sequence is generated by the DFAO in Figure 1. To generate $t_{n}$, one moves between the states (symbolized by nodes) according to subsequent digits in the binary representation of $n$. After the final digit has been read, the DFAO returns the output corresponding to the current state. For a detailed treatment of automatic sequences, we refer the reader to the monograph by Allouche and Shallit [3]. $1$start$-1$0110 Figure 1. A DFAO generating the PTM sequence. Finally, let us recall that the sequence ${\bf a}$ with values in $\mathbb{Z}$-module $R$ is $k$-regular if there exist finitely many sequences $\mathbf{a}_{i}=(a_{i}(n))_{n\in\mathbb{N}}$ with values in $R$ such that each sequence in $K_{k}({\bf a})$ is a $\mathbb{Z}$-linear combination of the $\bf{a}_{i}$. In other words, the $\mathbb{Z}$-submodule generated by the $k$-kernel $K_{k}({\bf a})$. In particular, $k$-automatic sequences are precisely $k$-regular sequences taking finitely many values. The class of $k$-regular sequences with values in a ring $R$ has a ring structure itself. A good introduction to the topic of regular sequences are the papers of Allouche and Shallit [4, 5]. ## 3\. The equation $b(n)=x^{2}+y^{2}+z^{2}$ We start with the characterization of the solutions (in variable $n$) of the equation $b(n)=x^{2}+y^{2}+z^{2}.$ Because the values $b(2n)$ and $b(2n+1)$ are equal, we restrict our attention to even indices and consider the set $S_{1}^{\prime}=\\{n\in\mathbb{N}:b(2n)\neq x^{2}+y^{2}+z^{2}\text{ for any }x,y,z\in\mathbb{Z}\\}.$ The first few elements of $S_{1}^{\prime}$ are the following: $10,18,34,40,58,66,72,90,106,114,130,136,154,160,170,178,202,210,226,\ldots.$ Using Theorem 2.1, we get the following characterization of $\nu_{2}(b(2n))$. ###### Proposition 3.1. For all $n\in\mathbb{N}_{+}$ we have $\nu_{2}(b(2n))=\begin{cases}1&\text{if }\nu_{2}(n)\equiv 0\pmod{2},\\\ 2&\text{if }\nu_{2}(n)\equiv 1\pmod{2}.\end{cases}$ We can deduce that if $\nu_{2}(n)\equiv 0\pmod{2}$, then $b(2n)$ is a sum of three squares. Hence, we only need to consider reduction modulo $32$ of $b(4^{k}(8m+4))$, where $k,m\in\mathbb{N}$. More precisely, $b(4^{k}(8m+4))$ is a sum of three squares if and only if (2) $b(4^{k}(8m+4))\equiv 28\pmod{32}.$ From Theorem 2.2 and the main result of Hirschhorn and Loxton [14] one can extract suitable congruence relations, which reduce the general case to $k=0$ and describe the remaining terms $b(8m+4)$. ###### Proposition 3.2. For all $m\in\mathbb{N}$ we have $\displaystyle b(16m)$ $\displaystyle\equiv b(4m)\pmod{32},$ $\displaystyle b(16m+4)$ $\displaystyle\equiv 4t_{m}\pmod{32},$ $\displaystyle b(16m+12)$ $\displaystyle\equiv 20t_{m}\pmod{32}.$ Using these relations, it is straightforward to describe the set consisting of $n\in\mathbb{N}$ such that $b(2n)$ is (not) a sum of three squares. ###### Corollary 3.3. The following conditions are equivalent: 1. (a) The number $b(2n)$ is not a sum of three squares; 2. (b) $n=2^{2k+1}(4s+1)$ for some $k,s\in\mathbb{N}$ such that $t_{s}=-1$; 3. (c) $n=2^{2k+1}(8r+2t_{r}+3)$ for some $k,r\in\mathbb{N}$; 4. (d) $\chi(n)=1$, where $\chi$ is defined by $\chi(0)=0$ and $\chi(2n+1)=0,\quad\chi(4n)=\chi(n),\quad\chi(8n+2)=T_{n},\quad\chi(8n+6)=0.$ ###### Proof. As we have discussed earlier, $b(2n)$ is not a sum of three squares if and only if $2n=4^{k}(8m+4)$ and (2) holds. By Proposition 3.2 this happens if and only if $m$ is even and $t_{m}=t_{m/2}=-1$. Letting $m=2s$, we obtain the equivalence of (a) and (b). To prove that (b) is equivalent to (c), we use the following description from [1]: (3) $\displaystyle\\{n\in\mathbb{N}:\;T_{n}=0\\}$ $\displaystyle=\\{2m+T_{m}:\;m\in\mathbb{N}\\},$ $\displaystyle\\{n\in\mathbb{N}:\;T_{n}=1\\}$ $\displaystyle=\\{2m+1-T_{m}:\;m\in\mathbb{N}\\}.$ Hence, $t_{s}=-1$ if and only if $s=2r+1-T_{r}=2r+(t_{r}+1)/2$ for some $r\in\mathbb{N}$, and our claim follows. Finally, it is simple to check that the set on $n$ of the form given in (b) is precisely $\\{n\in\mathbb{N}:\chi(n)=1\\}$. ∎ From the relation $b(2n+1)=b(2n)$ and part (c) of the corollary we get $S_{1}=2S_{1}^{\prime}\cup(2S_{1}^{\prime}+1)=\\{2^{2k+2}(8r+2t_{r}+3)+i:k,r\in\mathbb{N},i\in\\{0,1\\}\\}.$ Furthermore, part (d) of the corollary directly shows that $S_{1}^{\prime}$ (and thus $S_{1}$) is a $2$-automatic set; i.e., its characteristic sequence $(\chi(n))_{n\in\mathbb{N}}$ is $2$-automatic. A DFAO generating this sequence is shown in Figure 2. $0$start$0$$0$$0$$0$$1$$0$$0$$1$$0,1$$1$$1$$0$$0$$1$$1$$0$ Figure 2. A DFAO generating $(\chi(n))_{n\in\mathbb{N}}$ We now turn to the problem of gaps between consecutive $n$ such that $b(2n)$ is a sum of three squares. More precisely, we define $(f_{n})_{n\in\mathbb{N}}$ to be the increasing sequence consisting of the elements of $S^{\prime}$. Let $(g_{n})_{n\in\mathbb{N}}$ be the sequence of gaps, defined by $g_{n}=f_{n+1}-f_{n}.$ In other words, $g_{n}$ is the distance between $n$th and $(n+1)$th $1$ in the characteristic sequence $(\chi(n))_{n\in\mathbb{N}}$ (counting from $0$). The following proposition shows that the gaps are bounded. Moreover, for each possible gap length $g$ we provide in the proof an infinite set of $n\in\mathbb{N}$ such that $g_{n}=g$. ###### Proposition 3.4. For all $n\in\mathbb{N}$ we have $g_{n}\in\\{6,8,10,16,18,24\\}$ and all possible values are attained infinitely often. ###### Proof. Consider length $16$ subsequences $(\chi(16n+i))_{0\leq i\leq 15}$. A simple case distinction together with the relations in Corollary 3.3(d) show that there are only four possibilities, namely $\begin{cases}1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0&\text{if }\chi(n)=1,\\\ 0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0&\text{if }\chi(n)=0,T_{n}=1,\\\ 0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0&\text{if }\chi(n)=0,T_{n}=0,2\mid n,\\\ 0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0&\text{if }\chi(n)=0,T_{n}=0,2\nmid n.\\\ \end{cases}$ By inspecting the gaps within these subsequences and all their possible concatenations, we can see that the gaps between subsequent $1$’s in $(\chi(n))_{n\in\mathbb{N}}$ can only have lengths $6,8,10,16,18,24$. It remains to show that each of these values is indeed attained infinitely often. In Table 1 for each $g\in\\{6,8,10,16,18,24\\}$ we provide an infinite set $I_{g}$ of indices $n$ such that $\chi(n)=\chi(n+g)=1$ and all the terms inbetween are zero. We note that $I_{g}$ does not necessarily contain all such indices $n$. The verification of each case is straightforward and left to the reader. ∎ $g$ | $I_{g}$ ---|--- $6$ | $\\{32m+2:T_{m}=1\\}$ $8$ | $\\{32m+10:T_{m}=0\\}$ $10$ | $\\{16m:\chi(m)=1\\}$ $16$ | $\\{64m+18:T_{m}=0\\}$ $18$ | $\\{32m+8:T_{m}=1\\}$ $24$ | $\\{256m+178:T_{m}=0\\}$ Table 1. Gaps between $1$’s in $(\chi(n))_{n\in\mathbb{N}}$ ###### Remark 3.5. Let us note that Hajdu and Papp proved that the gap sequence corresponding to those values of $n$ such that $n!$ is a sum of three squares is bounded by 42 [15, Theorem 2.4]. It is also interesting to ask whether the sequence $(f_{n})_{n\in\mathbb{N}}$ itself is $2$-regular (equivalently, $(g_{n})_{n\in\mathbb{N}}$ is $2$ automatic), since its values form a $2$-automatic set. This question seems hard, and we have not been able to give a definitive answer (see also Section 7). The problem comes from the fact that the description of elements of $S_{1}^{\prime}$ in Corollary 3.3 does not give enough information about their ordering. Instead, we consider a simpler version of this question, where we restrict our attention to indices $n\in S_{1}^{\prime}$ with fixed $2$-adic valuation. More precisely, we let $k=0$ in the description of Corollary 3.3, so that $n$ is of the form $4m+2$. Then $b(2n)$ is not a sum of three squares if and only if $b(8m+4)\equiv 28\pmod{32}$ (this is precisely (2) with $k=0$). More generally, put $\beta(m)=\frac{b(8m+4)}{4}\bmod{8},$ and for each $a\in\\{1,3,5,7\\}$ let ${\bf c}_{a}=(c_{a}(l))_{l\in\mathbb{N}}$ be the increasing sequence such that $\\{m\in\mathbb{N}:\;\beta(m)=a\\}=\\{c_{a}(l):\;l\in\mathbb{N}\\}.$ It turns out that these sequences are described by surprisingly simple formulas. ###### Theorem 3.6. For each $a\in\\{1,3,5,7\\}$ the sequence ${\bf c}_{a}$ is 2-regular. More precisely, for $m\in\mathbb{N}$ we have $\displaystyle c_{1}(l)$ $\displaystyle=4l-t_{l}+1,$ $\displaystyle c_{3}(l)$ $\displaystyle=4l+t_{l}+2,$ $\displaystyle c_{5}(l)$ $\displaystyle=4l-t_{l}+2,$ $\displaystyle c_{7}(l)$ $\displaystyle=4l+t_{l}+1.$ ###### Proof. It is easy to see that each of the sequences from the statement is increasing. To prove that $\beta(m)=a$ if and only if $m=c_{a}(l)$ for some $l\in\mathbb{N}$, we restate the second and third relation of Proposition 3.2 in the following way: $\beta(m)=\begin{cases}1&\text{if }2\mid m\text{ and }t_{m}=1,\\\ 3&\text{if }2\nmid m\text{ and }t_{m}=1,\\\ 5&\text{if }2\nmid m\text{ and }t_{m}=-1,\\\ 7&\text{if }2\mid m\text{ and }t_{m}=-1.\\\ \end{cases}$ We now use the relations (3). If $\beta(m)=1$, then $2\mid m$ and $m=2k+T_{k}$ for some $k\in\mathbb{N}$. This implies $T_{k}=0$, and thus $k=2l+T_{l}$ for some $l\in\mathbb{N}$. As a result, we get $m=4l+2T_{l}=4l-t_{l}+1$. Conversely, if $m$ is of this form, then also $\beta(m)=1$, and so we get the claim for $a=1$. The proof for $a=3,5,7$ is similar. ∎ To conclude this section, we point out that similar results can be obtained for other quadratic forms given in Remark 1.1. More precisely, depending on the chosen form, they can be derived from either Proposition 3.2 or the following set of congruence relations (which again follows from Hirschhorn and Loxton’s results). ###### Proposition 3.7. For all $n\in\mathbb{N}$ we have $\displaystyle b(16n+8)$ $\displaystyle\equiv b(4n+2)\pmod{16},$ $\displaystyle b(8n+2)$ $\displaystyle\equiv 2t_{n}\pmod{16},$ $\displaystyle b(8n+6)$ $\displaystyle\equiv 6t_{n}\pmod{16}.$ ## 4\. The equation $b_{3}(n)=x^{2}+y^{2}+z^{2}$ In this section we characterize the elements of the set $S_{3}$ containing those $n$ such that $b_{3}(n)$ is not a sum of three squares. By virtue of Theorem 2.3, to get the required characterization of $S_{3}$, we need to understand of the behaviour of $b_{3}(16n+i)\mod{32}$ for $i=0,1,2,3,8,9,10,11$. Let us recall that the sequence $(b_{3}(n))_{n\in\mathbb{N}}$ satisfies the following recurrence relations: $b_{3}(0)=1,b_{3}(1)=3,b_{3}(2)=9$ and $\displaystyle b_{3}(2n)$ $\displaystyle=3b_{3}(2n-1)-3b_{3}(2n-2)+b_{3}(2n-3)+b_{3}(n),$ $\displaystyle b_{3}(2n+1)$ $\displaystyle=3b_{3}(2n)-3b_{3}(2n-1)+b_{3}(2n-2).$ We start with the following lemma. ###### Lemma 4.1. For all $n\in\mathbb{N}$ the following congruences hold: $\displaystyle b_{3}(8n+i+4)$ $\displaystyle\equiv 2(2i+1+4(-1)^{n})t_{n}\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32),$ $\displaystyle b_{3}(32n+i)$ $\displaystyle\equiv b_{3}(8n+i)\mkern 4.0mu({\operator@font mod}\mkern 6.0mu64),\;i=0,1,2,3,4$ $\displaystyle b_{3}(8(2n+1)+i)$ $\displaystyle\equiv 4(3+3i-i^{2}-2(-1)^{n+i})t_{n}\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32)$ $\displaystyle\equiv\begin{cases}4(3-2(-1)^{n})t_{n}\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32)&\text{if }i=0,\\\ 4(5+2(-1)^{n})t_{n}\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32)&\text{if }i=1,\\\ 4(5-2(-1)^{n})t_{n}\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32)&\text{if }i=2,\\\ 4(3+2(-1)^{n})t_{n}\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32)&\text{if }i=3.\end{cases}$ In particular, for each $k\in\mathbb{N}_{+}$ and $i\in\\{0,1,2,3\\}$, we have $\displaystyle b_{3}(2^{2k}(2n+1)+i)$ $\displaystyle\equiv 2\mkern 4.0mu({\operator@font mod}\mkern 6.0mu4),$ $\displaystyle b_{3}(2^{2k+1}(2n+1)+i)$ $\displaystyle\equiv b_{3}(8(2n+1)+i)\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32),$ ###### Proof. The computation of the values of $b_{3}(8n+i+4)\mod{32}$ and $b_{3}(8(2n+1)+i)\mod{32}$ for $i=0,1,2,3$ is based on a simple induction with the help of recurrence relations satisfied by the PTM sequence $(t_{n})_{n\in\mathbb{N}}$ and the sequence $(b_{3}(n))_{n\in\mathbb{N}}$. Because of this, we omit the simple details. Essentially, the same approach can be used in the case of the congruence $b_{3}(32n+i)\equiv b_{3}(8n+i)\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32)$. However, a more conceptual proof is the following. Invoking [10, Lemma 4.7] we know that for each $a\in\mathbb{N}_{+},b\in\\{0,1,\ldots,2^{a}-1\\}$ there is a polynomial $P_{a,b}\in\mathbb{Z}[x]$ such that $\sum_{n=0}^{\infty}b_{3}(2^{a}n+b)x^{n}=\frac{P_{a,b}(x)}{(1-x)^{3a}}B_{3}(x).$ In particular, in the case we are interested in, we have $\sum_{n=0}^{\infty}(b_{3}(32n+i)-b_{3}(8n+i))x^{n}=\frac{P_{5,i}(x)-(1-x)^{6}P_{3,i}(x)}{(1-x)^{15}}B_{3}(x).$ A quick computation reveals that for each $i=0,1,2,3$, the polynomial $P_{5,i}(x)-(1-x)^{6}P_{3,i}(x)$ is divisible by 64 in the ring $\mathbb{Z}[x]$. Thus, as the function $(1-x)^{-15}B_{3}(x)$ has power series expansion with integer coefficients, then each number $b_{3}(32n+i)-b_{3}(8n+i)$ is divisible by 64 and we are done. To obtain the first congruence from the “in particular” part, we apply induction on $k$ and the congruence $b_{3}(8n+i+4)\equiv 2(2i+1+4(-1)^{n})t_{n}\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32)$. In particular, for $i=0,1,2,3$, the 2-adic valuation of $b_{3}(8n+i+4)$ is equal to 1. To obtain the congruence $b_{3}(2^{2k+1}(2n+1)+i)\equiv b_{3}(8(2n+1)+i)$ we again use induction on $k$ and apply the congruence $b_{3}(32n+i)\equiv b_{3}(8n+i)\mkern 4.0mu({\operator@font mod}\mkern 6.0mu64)$. ∎ We are ready to characterize the set $S_{3}$. ###### Theorem 4.2. We have $n\in S_{3}$ if and only if $n=2^{2k+1}\left(8p+2\left\lfloor\frac{i}{2}\right\rfloor+3+2(-1)^{i}t_{p}\right)+i$ for some $i\in\\{0,1,2,3\\}$ and $k\in\mathbb{N}_{+}$, $p\in\mathbb{N}$. ###### Proof. From the characterization of the 2-adic valuation of $b_{3}(n)$ and Lemma 4.1 we know that if $n\in S_{3}$, then we necessarily have $n\pmod{16}\in\\{0,1,2,3,8,9,10,11\\}$. We perform a case-by-case analysis. Let $i\in\\{0,1,2,3\\}$. If $n\equiv i\mkern 4.0mu({\operator@font mod}\mkern 6.0mu16)$ and $n=2^{2k}(2s+1)+i$, then $\nu_{2}(b_{3}(n))=1$ and hence $n\not\in S_{3}$. If $n=2^{2k+1}(2s+1)+i$, then we have $b_{3}(n)\equiv b_{3}(8(2s+1)+i)\equiv 4(3+3i-i^{2}-2(-1)^{s+i})t_{s}\mkern 4.0mu({\operator@font mod}\mkern 6.0mu32),$ and thus $n\in S_{3}$ if and only if $c(i,s):=(3+3i-i^{2}-2(-1)^{s+i})t_{s}\equiv 7\mkern 4.0mu({\operator@font mod}\mkern 6.0mu8)$. A case-by-case analysis using the characterization (3), reveals the following: 1. (1) If $i=0$, then $c(i,s)\equiv 7\mkern 4.0mu({\operator@font mod}\mkern 6.0mu8)$ if and only if $s$ is even and $t_{s}=-1$. Thus, $s=4p+1+t_{p}$ for some $p\in\mathbb{N}$. 2. (2) If $i=1$, then $c(i,s)\equiv 7\mkern 4.0mu({\operator@font mod}\mkern 6.0mu8)$ if and only if $s$ is even and $t_{s}=1$. Thus $s=4p+1-t_{p}$ for some $p\in\mathbb{N}$. 3. (3) If $i=2$, then $c(i,s)\equiv 7\mkern 4.0mu({\operator@font mod}\mkern 6.0mu8)$ if and only if $s$ is odd, and $t_{s}=1.$ Thus $s=4p+2+t_{p}$ for some $p\in\mathbb{N}$. 4. (4) If $i=3$, then $c(i,s)\equiv 7\mkern 4.0mu({\operator@font mod}\mkern 6.0mu8)$ if and only if $s$ is odd and $t_{s}=-1.$ Thus $s=4p+2-t_{p}$ for some $p\in\mathbb{N}$. Gathering all the obtained characterizations, we get the statement of our theorem. ∎ ## 5\. The equation $b_{2^{k}-1}(n)=x^{2}+y^{2}+z^{2}$ with $k\geq 3$ In this section, we study for $k\geq 3$ representability of $b_{2^{k}-1}(n)$ as a sum of three squares. The main idea is to express $(b_{2^{k}-1}(n))_{n\in\mathbb{N}}$ as the convolution of $(b_{2^{k}}(n))_{n\in\mathbb{N}}$ and the PTM sequence, and apply the following lemma [10, Lemma 4.4(1)] to $m=2^{k}$. ###### Lemma 5.1. Let $m\in\mathbb{N}_{+}$. Then for all $n\in\mathbb{N}$ we have $b_{m}(n)\equiv\binom{m}{n}+2^{\nu_{2}(m)+1}\binom{m-2}{n-2}\pmod{2^{\nu_{2}(m)+2}}.$ We split our reasoning into two parts: $n<2^{k}$ and $n\geq 2^{k}$. Starting with the simpler case $n<2^{k}$, by Theorem 2.3, we have $\nu_{2}(b_{2^{k}-1}(n))=0$. Therefore, it is sufficient for our purposes to describe $b_{2^{k}-1}(n)$ modulo $8$. ###### Proposition 5.2. Let $k\geq 3$ and $n<2^{k}$. Then $b_{2^{k}-1}(n)\equiv t_{n}\cdot\begin{cases}1\pmod{8}&\text{if }0\leq n<2^{k-2},\\\ 5\pmod{8}&\text{if }2^{k-2}\leq n<2^{k-1},\\\ 7\pmod{8}&\text{if }2^{k-1}\leq n<3\cdot 2^{k-2},\\\ 3\pmod{8}&\text{if }3\cdot 2^{k-2}\leq n<2^{k}.\end{cases}$ ###### Proof. By Lemma 5.1 we have $b_{2^{k}-1}(n)\equiv\sum_{l=0}^{n}\binom{2^{k}}{l}t_{n-l}\pmod{8}$ Moreover, [10, Lemma 4.5] says that $\binom{2^{k}}{l}\equiv\begin{cases}1\pmod{8}&\text{if }l=0,2^{k},\\\ 4\pmod{8}&\text{if }l=2^{k-2},3\cdot 2^{k-2},\\\ 6\pmod{8}&\text{if }l=2^{k-1},\\\ 0\pmod{8}&\text{otherwise}.\end{cases}$ From this description we immediately get the claim for the cases $0\leq n<2^{k-2}$ and $2^{k-2}\leq n<2^{k-1}$. If $2^{k-1}\leq n<3\cdot 2^{k-2}$, we get $b_{2^{k}-1}(n)\equiv t_{n}+4t_{n-2^{k-2}}+6t_{n-2^{k-1}}\equiv t_{n}+2t_{n-2^{k-1}}\pmod{8}.$ Since $n$ has $2^{k-1}$ in its binary expansion, we get $t_{n-2^{k-1}}=-t_{n}$, and the required congruence follows. Finally, if $3\cdot 2^{k-2}\leq n<2^{k}$, we again have $t_{n-2^{k-1}}=-t_{n}$ so $b_{2^{k}-1}(n)\equiv t_{n}+4t_{n-2^{k-2}}+6t_{n-2^{k-1}}+4t_{n-3\cdot 2^{k-2}}\equiv-5t_{n}\pmod{8}.\qed$ As an immediate corollary, we can describe $n<2^{k}$ such that $b_{2^{k}-1}(n)$ is (not) a sum of three squares. ###### Corollary 5.3. Let $k\geq 3$ and $n<2^{k}$. Then $b_{2^{k}-1}(n)$ is not a sum of three squares of integers if and only if one of the following cases holds: 1. (1) $0\leq n<2^{k-2}$ and $t_{n}=-1$; 2. (2) $2^{k-1}\leq n<3\cdot 2^{k-2}$ and $t_{n}=1$. We move on to the case $n\geq 2^{k}$. This time we have $\nu_{2}(b_{2^{k}-1}(n))\in\\{1,2\\}$ by Theorem 2.3, which means that it is sufficient to consider $b_{2^{k}-1}(n)$ modulo $32$. To this end, we need a standard lemma concerning the behavior of binomial coefficients modulo powers of $2$ (we provide a proof for completeness). ###### Lemma 5.4. The following statements hold: 1. (a) For all $k,n\in\mathbb{N}$ such that $1\leq n\leq 2^{k}$, we have $\nu_{2}\left(\binom{2^{k}}{n}\right)=k-\nu_{2}(n).$ 2. (b) For all $m,n\in\mathbb{N}$ we have $\binom{2m}{2n}\equiv\binom{m}{n}\pmod{2^{\nu_{2}(m)+1}}.$ ###### Proof. By Legendre’s formula we get $\nu_{2}\left(\binom{2^{k}}{n}\right)=2^{k}-1-(n-s_{2}(n)+2^{k}-n-s_{2}(2^{k}-n))=s_{2}(n)+s_{2}(2^{k}-n)-1.$ We can express $s_{2}(2^{k}-n)$ as $s_{2}(2^{k}-n)=s_{2}((2^{k}-1)-(n-1))=k-s_{2}(n-1).$ Now, write $n=2^{\nu_{2}(n)}l$, which yields $\displaystyle s_{2}(n-1)$ $\displaystyle=s_{2}(2^{\nu_{2}(n)}(l-1)+(2^{\nu_{2}(n)}-1))=s_{2}(l-1)+\nu_{2}(n)$ $\displaystyle=s_{2}(l)-1+\nu_{2}(n)=s_{2}(n)-1+\nu_{2}(n).$ Combining the above equalities, we get (a). Moving on to (b), the claim clearly holds for $n=0$ so we can assume that $n\geq 1$. We have the congruence $\binom{2m}{2n}=\binom{m}{n}\frac{(2m-1)!!}{(2n-1)!!(2m-2n-1)!!}\equiv(-1)^{n}\binom{m}{n}\pmod{2^{\nu_{2}(m)+1}}.$ If $n$ is even, we immediately obtain (b). If $n$ is odd, we use the inequality $\nu_{2}\left(\binom{m}{n}\right)=\nu_{2}\left(\frac{m}{n}\binom{m-1}{n-1}\right)\geq\nu_{2}(m),$ which again leads to the desired result. ∎ We are now ready to describe $b_{2^{k}-1}(n)$ modulo $32$ for $n\geq 2^{k}$. This time, the characterization involves two consecutive terms of the PTM sequence. ###### Theorem 5.5. Fix $k,i,j\in\mathbb{N}$ such that $k\geq 3$, $i<8$, and $j<2^{k-3}$. Then for all $m\geq 1$ we have $b_{2^{k}-1}(2^{k}m+2^{k-3}i+j)\equiv t_{j}(c_{i}t_{m}+d_{i}t_{m-1})\pmod{32},$ where the coefficients $c_{i},d_{i}$ do not depend on $k$ and are given in Table 2. $i$ | $0$ | $1$ | $2$ | $3$ | $4$ | $5$ | $6$ | $7$ ---|---|---|---|---|---|---|---|--- $c_{i}$ | $1$ | $7$ | $3$ | $5$ | $9$ | $-1$ | $3$ | $5$ $d_{i}$ | $-5$ | $-3$ | $1$ | $-9$ | $-5$ | $-3$ | $-7$ | $-1$ Table 2. The coefficients $c_{i}$, $d_{i}$. ###### Proof. Consider first the case $k\geq 4$. By Lemma 5.1 we have $b_{2^{k}-1}(n)=\sum_{l=0}^{n}b_{2^{k}}(l)t_{n-l}\equiv\sum_{l=0}^{n}\binom{2^{k}}{l}t_{n-l}\pmod{32}.$ Now, by Lemma 5.4(a), the binomial coefficients with $v_{2}(l)<k-4$ vanish modulo $32$. Therefore, assuming that $n\geq 2^{k}$, the above sum simplifies to $b_{2^{k}-1}(n)\equiv\sum_{l=0}^{16}\binom{2^{k}}{2^{k-4}l}t_{n-2^{k-4}l}\equiv\sum_{l=0}^{16}\binom{16}{l}t_{n-2^{k-4}l}\pmod{32},$ where the second congruence follows from Lemma 5.4(b). Furthermore, we can eliminate the terms with $l$ odd, since there is an even number of them and they are all congruent with $16$ modulo $32$. Therefore, we get the congruence $b_{2^{k}-1}(n)\equiv\sum_{l=0}^{8}\binom{16}{2l}t_{n-2^{k-3}l}\pmod{32}.$ To simplify the right-hand side, consider $b_{2^{k}-1}$ at indices of the form given in the statement, namely $n=2^{k}m+2^{k-3}i+j$, where $m\geq 1$, $0\leq i<8$, and $0\leq j<2^{k-3}$. For the sake of clarity, we will now momentarily use the notation $t(n)=t_{n}$. By the recurrences that define the Thue–Morse sequence, we get $t(2^{k}m+2^{k-3}i+j-2^{k-3}l)=t_{j}t_{8m+i-l}=t_{j}\cdot\begin{cases}t_{n}t_{i-l}&\text{if }l\leq i,\\\ -t_{n-1}t_{l-i}&\text{if }l>i.\end{cases}$ Therefore, the claimed formula is valid with the coefficients $c_{i}=\sum_{l=0}^{i}\binom{16}{2l}t_{i-l},\qquad d_{i}=-\sum_{l=i+1}^{8}\binom{16}{2l}t_{l-i},$ and a direct calculation (modulo $32$) gives their values as in Table 2. In the case $k=3$, the expression for $b_{2^{k}-1}(n)$ modulo $32$ obtained from Lemma 5.1 also contains the sum $16\sum_{l=0}^{n}\binom{6}{l-2}t_{n-l}.$ If $n\geq 8$, then the whole sum vanishes modulo $32$, so we again arrive at the formula $b_{7}(n)\equiv\sum_{l=0}^{8}\binom{8}{l}t_{8-l}\pmod{32}.$ After a similar calculation as before, we get the result. ∎ Using this result, we can determine the indices $n\geq 2^{k}$ such that $b_{2^{k}-1}(n)$ is not a sum of three squares. The description turns out to be surprisingly simple. ###### Corollary 5.6. Let $k\geq 3$ and $n\geq 2^{k}$. The following conditions are equivalent: 1. (a) $b_{2^{k}-1}(n)$ is not a sum of three squares; 2. (b) $t_{n}=t_{n-2^{k}}=1$; 3. (c) $n=2^{k}m+l,$ where $l,m\in\mathbb{N}$ are such that $l<2^{k}$, $t_{m}=t_{l}$, and $\nu_{2}(m)\equiv 1\pmod{2}$. ###### Proof. Write $n=2^{k}m+2^{k-3}i+j$ as in Theorem 5.5. Observe that $c_{i}+d_{i}=-4t_{i}$, while $c_{i}-d_{i}$ is not divisible by $4$. Hence, the term $b_{2^{k}-1}(2^{k}m+2^{k-3}i+j)$ is not a sum of three squares if and only if $t_{m}=t_{m-1}=t_{i}t_{j},$ which after multiplying both sides by $t_{i}t_{j}$ gives precisely (b). The equivalence with (c) is obtained by writing $l=2^{k-3}i+j$ and observing that $t_{m}=(-1)^{\nu_{2}(m)+1}t_{m-1}$. ∎ ## 6\. Counting the solutions The aim of this section is to provide estimates for the counting functions of the sets $S_{2^{k}-1}$. For real $x\geq 0$ and $m\in\mathbb{N}_{+}$ let $S_{m}(x)=S_{m}\cap[0,x]=\\#\\{n\leq x:b_{m}(n)\text{ is not a sum of three squares}\\}.$ Using the descriptions of the sets $S_{2^{k}-1}$ obtained in the previous sections for various $k$ it is straightforward to check that $S_{2^{k}-1}(x)=\delta_{k}x+O(\log x),$ where $\delta_{1}=\delta_{2}=1/12$ and $\delta_{k}=1/6$ for $k\geq 3$. In the following three results, we provide more precise bounds for $S_{2^{k}-1}(x)-d_{k}x$ in the case $k=1,k=2$ and $k\geq 3$, respectively. In particular, each lower and upper bound is of the form $C_{1}\log_{2}x+C_{2}$, where the constant $C_{1}$ is optimal. ###### Theorem 6.1. For every $x\geq 6$ we have $-\frac{5}{3}<S_{1}(x)-\frac{x}{12}<\frac{1}{2}\log_{2}x-\frac{19}{12}.$ In particular, the density of the set $S_{1}$ in $\mathbb{N}$ exists and is equal to $\lim_{x\rightarrow+\infty}\frac{S_{1}(x)}{x}=\frac{1}{12}.$ Moreover, there exists an increasing sequence $(m_{k})_{k\in\mathbb{N}}\subset\mathbb{N}$ such that $S_{1}(m_{l})-\frac{m_{l}}{12}\sim\frac{1}{2}\log_{2}m_{l}.$ ###### Proof. For real $x\geq 0$ define $\displaystyle P(x)$ $\displaystyle=\\#\\{s\in\mathbb{N}:8s+2t_{s}+3\leq x\\},$ $\displaystyle Q(x)$ $\displaystyle=\sum_{k=0}^{\infty}P\left(\frac{x}{4^{k}}\right).$ By Corollary 3.3(b) and the relation $b(2n+1)=b(2n)$, we get $S_{1}(x)=Q\left(\frac{x}{4}\right)+Q\left(\frac{x-1}{4}\right).$ Hence, it is sufficient to focus on the function $Q$. For $m\in\mathbb{N}$ and $i=0,1,2,3$ we have the recurrence relations $Q(4m+i)=Q(m)+P(4m+i).$ Also, for $i<8$ we have $P(8m+i)=m+\begin{cases}0&\text{if }i=0,\\\ T_{m}&\text{if }i=1,2,3,4,\\\ 1&\text{if }i=5,6,7.\end{cases}$ Put $R(x)=Q(x)-\frac{x}{6},$ so that $S_{1}(x)-\frac{x}{12}=R\left(\left\lfloor\frac{x}{4}\right\rfloor\right)+R\left(\left\lfloor\frac{x-1}{4}\right\rfloor\right)+\frac{\left\lfloor\frac{x}{4}\right\rfloor+\left\lfloor\frac{x-1}{4}\right\rfloor}{6}-\frac{x}{12}.$ It is readily checked that $-\frac{1}{3}<\frac{\left\lfloor\frac{x}{4}\right\rfloor+\left\lfloor\frac{x-1}{4}\right\rfloor}{6}-\frac{x}{12}\leq-\frac{1}{12}$ Therefore, to obtain the estimates for $S_{1}(x)-x/12$ (where $x\geq 9$), it remains to prove that for each integer $m\geq 2$ there holds $-\frac{2}{3}\leq R(m)\leq\frac{1}{4}\lfloor\log_{2}m\rfloor-\frac{1}{4},$ as then $-\frac{4}{3}\leq R\left(\left\lfloor\frac{x}{4}\right\rfloor\right)+R\left(\left\lfloor\frac{x-1}{4}\right\rfloor\right)<2\cdot\frac{1}{4}\left(\log_{2}\frac{x}{4}-1\right)=\frac{1}{2}\log_{2}x-\frac{3}{2}.$ This is done by induction on the length $L(m)$ of the binary expansion of $m$. Direct computation shows that our claim holds when $2\leq L(m)\leq 6$. Hence, let $L(m)\geq 7$. It is sufficient to prove that there exists an integer $n\geq 2$ with $L(n)\leq L(m)-2$ such that $0\leq R(m)-R(n)\leq\frac{1}{2}.$ This is indeed the case, as shown by the following set of identities (ordered according to the residue class modulo $8$): $\displaystyle R(8n)$ $\displaystyle=R(2n),$ $\displaystyle R(16n+1)$ $\displaystyle=R(4n+1),$ $\displaystyle R(16n+9)$ $\displaystyle=R(4n)+\frac{1}{2},$ $\displaystyle R(16n+2)$ $\displaystyle=R(4n+2),$ $\displaystyle R(16n+10)$ $\displaystyle=R(4n)+\frac{1}{3},$ $\displaystyle R(16n+3)$ $\displaystyle=R(4n+3),$ $\displaystyle R(16n+11)$ $\displaystyle=R(4n)+\frac{1}{6},$ $\displaystyle R(8n+4)$ $\displaystyle=R(2n+1)+T_{n}-\frac{1}{2},$ $\displaystyle R(64n+4)$ $\displaystyle=R(16n+4),$ $\displaystyle R(64n+20)$ $\displaystyle=R(16n+2)+1-T_{n},$ $\displaystyle R(64n+36)$ $\displaystyle=R(16n)+1-T_{n},$ $\displaystyle R(64n+52)$ $\displaystyle=R(16n+4),$ $\displaystyle R(16n+12)$ $\displaystyle=R(4n),$ $\displaystyle R(8n+5)$ $\displaystyle=R(2n+1)+\frac{1}{3},$ $\displaystyle R(8n+6)$ $\displaystyle=R(2n+1)+\frac{1}{6},$ $\displaystyle R(8n+7)$ $\displaystyle=R(2n+1).$ We move on to the second part of the statement. Define $m_{0}=0$ and $m_{l+1}=16m_{l}+36$ for $l\in\mathbb{N}$. First, we prove inductively that $R(m_{l})=l$. This is clear for $l=0$. In general, we have $R(m_{l+1})=R(16m_{l}+36)=R(4m_{l})+1-T_{m_{l}}=R(m_{l})+1,$ where we have used $4\mid m_{l}$, the recurrence relations above and $T_{m_{l}}=0$ (easily shown by induction). We thus have $S_{1}(m_{0})=S_{1}(m_{1})=0$ and for $l\geq 1$ the equality $\displaystyle S_{1}(m_{l+1})-\frac{m_{l+1}}{12}$ $\displaystyle=R\left(\left\lfloor\frac{m_{l+1}}{4}\right\rfloor\right)+R\left(\left\lfloor\frac{m_{l+1}-1}{4}\right\rfloor\right)-\frac{1}{6}$ $\displaystyle=R(m_{l})+R(m_{l-1})+1=2l.$ The result follows. ∎ In Figure 3 we show the graph of the function $S_{1}(x)-x/12$ in the range $[1,2^{10}]$ together with the bounds (as in the theorem). Figure 3. The function $S_{1}(x)-x/12$. From the presented graph, it appears it should be possible to obtain an even better additive constant in the upper bound. To do this, one would need to investigate closer the location of the “spikes” on the graph (some of which correspond to $x=m_{l}$). The following two results show that the function $S_{1}(x)$ is exceptional in the sense that $S_{1}(x)-x/12$ is bounded from below by a constant. ###### Theorem 6.2. For all $x\geq 1$ we have $-\frac{1}{6}\log_{2}x-\frac{7}{12}<S_{3}(x)-\frac{x}{12}\leq\frac{1}{6}\log_{2}x-\frac{1}{6}.$ In particular, the density of the set $S_{3}$ in $\mathbb{N}$ exists and is equal to $\lim_{x\rightarrow+\infty}\frac{S_{3}(x)}{x}=\frac{1}{12}.$ Moreover, there exist increasing sequences $(m_{l})_{l\in\mathbb{N}},(n_{l})_{l\in\mathbb{N}}\subset\mathbb{N}$ such that $\displaystyle S_{3}(m_{l})-\frac{m_{l}}{12}$ $\displaystyle\sim\frac{1}{6}\log_{2}m_{l},$ $\displaystyle S_{3}(n_{l})-\frac{n_{l}}{12}$ $\displaystyle\sim-\frac{1}{6}\log_{2}n_{l}.$ ###### Proof. For $i=0,1,2,3$ let $P_{i}(x)=\\#\\{n\in\mathbb{N}:8n+2\left\lfloor\frac{i}{2}\right\rfloor+3+2(-1)^{i}t_{n}\leq x\\},$ so that by Theorem 4.2 we have $S_{3}(x)=\sum_{k=1}^{\infty}\sum_{i=0}^{3}P_{i}\left(\frac{x-i}{2\cdot 4^{k}}\right).$ This time, put $\displaystyle P(x)$ $\displaystyle=\sum_{i=0}^{3}P_{i}(x),$ $\displaystyle Q(x)$ $\displaystyle=\sum_{k=0}^{\infty}P\left(\frac{x}{4^{k}}\right).$ Then for any $x$ we have $Q\left(\frac{x-3}{8}\right)\leq S_{3}(x)\leq Q\left(\frac{x}{8}\right).$ Therefore, we need to bound the function $R(x)=Q(x)-2x/3$. First, for $n\in\mathbb{N}$ we have the easy to check equalities $P(n)=\lceil n/2\rceil$ and $R(4n+i)=R(n)+\begin{cases}0&\text{for }i=0,3,\\\ \frac{1}{3}&\text{for }i=1,\\\ -\frac{1}{3}&\text{for }i=2.\end{cases}$ In a similar fashion as in the previous proof, one can then prove that for $m\in\mathbb{N}$ there holds $-\frac{1}{6}\lfloor\log_{2}m\rfloor-\frac{1}{6}\leq R(m)\leq\frac{1}{6}\lfloor\log_{2}m\rfloor+\frac{1}{3}$ The inequalities for $S_{3}(x)-x/12$ follow shortly by plugging in $m=\lfloor x/8\rfloor$ and $m=\lfloor(x-3)/8\rfloor$. If we define $m_{0}=0$ and $m_{l+1}=4m_{l}+8$, we can inductively compute $R(m_{l}/8)=l/3$, and therefore $S(m_{l})-\frac{1}{12}m_{l}\sim Q\left(\frac{m_{l}}{8}\right)-\frac{1}{12}m_{l}=\frac{l}{3}\sim\frac{1}{6}\log_{2}m_{l}.$ Similarly, for $n_{0}=0$ and $n_{l+1}=4n_{l}+16$, we get $R(n_{l}/8)=-l/3$ so $S(n_{l})-\frac{1}{12}n_{l}\sim-\frac{1}{6}\log_{2}n_{l},$ and the proof is finished. ∎ Figure 4. The function $S_{3}(x)-x/12$. Figure 4 shows the graph of the function $S_{1}(x)-x/12$ in the range $[1,2^{10}]$ together with the proved bounds (in red). Again, the bounds are quite accurate, though the additive constants can probably be improved further. The final result of this section concerns the function $S_{2^{k}-1}(x)$. For the sake of clarity, in the proof we make some rough estimates concerning the additive constant (although the constant near $\log_{2}x$ remains optimal). ###### Theorem 6.3. If $k\geq 3$, then for all $x\geq 2^{k}$ we have $\left|S_{2^{k}-1}(x)-\frac{x}{6}\right|\leq\frac{2^{k-2}}{3}(\log_{2}x-k+26).$ In particular, the density of the set $S_{2^{k}-1}$ in $\mathbb{N}$ exists and is equal to $\lim_{x\rightarrow+\infty}\frac{S_{2^{k}-1}(x)}{x}=\frac{1}{6}.$ Moreover, there exist increasing sequences $(m_{l})_{l\in\mathbb{N}},(n_{l})_{l\in\mathbb{N}}\subset\mathbb{N}$ such that $\displaystyle S_{2^{k}-1}(m_{l})-\frac{m_{l}}{6}$ $\displaystyle\sim\frac{2^{k-2}}{3}\log_{2}m_{l},$ $\displaystyle S_{2^{k}-1}(n_{l})-\frac{n_{l}}{6}$ $\displaystyle\sim-\frac{2^{k-2}}{3}\log_{2}n_{l}.$ ###### Proof. For $\varepsilon\in\\{1,-1\\}$ and non-negative $x\in\mathbb{R}$ we put $\displaystyle P_{\varepsilon}(x)$ $\displaystyle=\\#\\{1\leq m\leq x:t_{m}=\varepsilon\\},$ $\displaystyle Q_{\varepsilon}(x)$ $\displaystyle=\sum_{s=0}^{\infty}(-1)^{s}P_{\varepsilon}\left(\frac{x}{2^{s}}\right)=\\#\\{1\leq m\leq x:t_{m}=\varepsilon\text{ and }\nu_{2}(m)\equiv 0\mkern 4.0mu({\operator@font mod}\mkern 6.0mu2)\\}.$ Then by Corollary 5.6 we get $\displaystyle S_{2^{k}-1}(x)$ $\displaystyle=S_{2^{k}-1}(2^{k}-1)+\sum_{l=0}^{2^{k}-1}\\#\\{1\leq m\leq\frac{x-l}{2^{k}}:t_{m}=t_{l}\text{ and }\nu_{2}(m)\equiv 1\mkern 4.0mu({\operator@font mod}\mkern 6.0mu2)\\}$ $\displaystyle=2^{k-2}+\sum_{l=0}^{2^{k}-1}Q_{t_{l}}\left(\frac{x-l}{2^{k+1}}\right),$ where $S_{2^{k}-1}(2^{k}-1)=2^{k-2}$ follows from Corollary 5.3. Furthermore, we have the obvious inequality $0\leq\sum_{l=0}^{2^{k}-1}Q_{t_{l}}\left(\frac{x}{2^{k+1}}\right)-\sum_{l=0}^{2^{k}-1}Q_{t_{l}}\left(\frac{x-l}{2^{k+1}}\right)\leq 2^{k}.$ Since for each $\varepsilon=\pm 1$ we have $t_{l}=\varepsilon$ for precisely $2^{k-1}$ indices $l$, we obtain (4) $\left|S_{2^{k}-1}(x)-2^{k-1}\left(Q_{1}\left(\frac{x}{2^{k+1}}\right)+Q_{-1}\left(\frac{x}{2^{k+1}}\right)\right)\right|=5\cdot 2^{k-2}.$ Therefore, to bound $S_{2^{k}-1}(x)-x/6$ it remains to estimate for $\varepsilon=\pm 1$ the functions $R_{\varepsilon}(x)=Q_{\varepsilon}(x)-\frac{x}{3}.$ First, note that for $n\in\mathbb{N}$ we have $P_{\varepsilon}(n)=\frac{n-\varepsilon}{2}+\begin{cases}\frac{\varepsilon}{2}t_{n}&\text{if }2\mid n,\\\ 0&\text{if }2\nmid n.\end{cases}$ It follows that $R_{\varepsilon}(4n+i)=R_{\varepsilon}(n)+\begin{cases}0&\text{if }i=0,3,\\\ \frac{1}{2}(1-\varepsilon t_{n})-\frac{1}{3}&\text{if }i=1,\\\ \frac{1}{2}(1-\varepsilon t_{n})-\frac{2}{3}&\text{if }i=2.\end{cases}$ This leads to the relations $\displaystyle R_{\varepsilon}(4n)$ $\displaystyle=R_{\varepsilon}(n),$ $\displaystyle R_{\varepsilon}(8n+1)$ $\displaystyle=R_{\varepsilon}(2n+1),$ $\displaystyle R_{\varepsilon}(16n+5)$ $\displaystyle=R_{\varepsilon}(n)+\frac{1}{3},$ $\displaystyle R_{\varepsilon}(16n+13)$ $\displaystyle=R_{\varepsilon}(4n+1),$ $\displaystyle R_{\varepsilon}(16n+2)$ $\displaystyle=R_{\varepsilon}(4n+2),$ $\displaystyle R_{\varepsilon}(8n+6)$ $\displaystyle=R_{\varepsilon}(2n),$ $\displaystyle R_{\varepsilon}(16n+10)$ $\displaystyle=R_{\varepsilon}(n)-\frac{1}{3},$ $\displaystyle R_{\varepsilon}(4n+3)$ $\displaystyle=R_{\varepsilon}(n).$ By induction we obtain for $\varepsilon=\pm 1$ and all $n\in\mathbb{N}_{+}$ the inequality $|R_{\varepsilon}(n)|\leq\frac{1}{12}\lfloor\log_{2}n\rfloor+\frac{2}{3},$ which implies $\left|Q_{\varepsilon}(x)-\frac{x}{3}\right|\leq\frac{1}{12}\log_{2}x+1$ for all $x\geq 1$. The main part of the result follows shortly. Finally, put $m_{0}=0$ and $m_{l+1}=16m_{l}+5\cdot 2^{k+1}$. Also, let $\alpha=(k+1)\bmod{2}$. Using the fact that $2^{k+1}\mid m_{l}$, from the recurrence relations for $R_{\varepsilon}$ we get $\displaystyle R_{\varepsilon}(2^{\alpha}m_{l+1})$ $\displaystyle=R_{\varepsilon}(2^{4+\alpha}m_{l}+5\cdot 2^{k+1+\alpha})=R_{\varepsilon}(2^{3-k}m_{l}+5)$ $\displaystyle=R_{\varepsilon}(2^{-1-k}m_{l})+\frac{1}{3}=R_{\varepsilon}(2^{\alpha}m_{l})+\frac{1}{3}.$ It follows that $\displaystyle Q_{\varepsilon}\left(\frac{m_{l}}{2^{k+1}}\right)-\frac{m_{l}}{3\cdot 2^{k+1}}=R_{\varepsilon}(2^{\alpha}m_{l})=\frac{l}{3}\sim\frac{1}{12}\log_{2}m_{l},$ and it remains to use (4). Similarly, we can take $n_{0}=0$ and $n_{l+1}=16n_{l}+10\cdot 2^{k+1}$. ∎ ## 7\. Computational results, questions, problems and conjectures In this section, we discuss possible directions for further research and present some conjectures and computational results. To begin, recall that in Section 3 we have defined $(f_{n})_{n\in\mathbb{N}}$ to be the increasing sequence such that $S_{1}^{\prime}=\\{f_{n}:n\in\mathbb{N}\\}$, and asked whether it is regular. We have performed some experimental computations in Mathematica 13 with the help of the IntegerSequences package by Eric Rowland, available at https://ericrowland.github.io/packages.html. More precisely, for each $m\leq 30$ we have used the FindRegularSequenceRecurrence function, which did not find a finite set of (plausible) $\mathbb{Z}$-linear relations between the elements of the $m$-kernel $K_{m}((f_{n})_{n\in\mathbb{N}})$. Hence, we expect that following conjecture holds. ###### Conjecture 7.1. The sequence $(f_{n})_{n\in\mathbb{N}_{+}}$ is not $m$-regular for any $m\geq 2$. On the other hand, note that we if we consider the decomposition $S_{1}^{\prime}=\bigcup_{k=0}^{\infty}U_{k}$ into pairwise disjoint sets $U_{k}=\\{2^{2k+1}(8s+2t_{s}+3):\;s\in\mathbb{N}\\}$, then for each $k\in\mathbb{N}$ the sequence $(2^{2k+1}(8s+2t_{s}+3))_{s\in\mathbb{N}}$ is 2-regular. Next, it is natural to ask whether it is possible to obtain results on the representation of $b_{m}(n)$ as a sum of three squares for any $m\in\mathbb{N}_{+}$. ###### Problem 7.2. Characterize the set $S_{m}$ for $m\in\mathbb{N}_{+}$. If the valuations $\nu_{2}(b_{m}(n))$ are bounded, then the direct approach used in this paper, namely reduction modulo a fixed power of $2$, is sufficient to give a complete description of $S_{m}$. The following proposition implies that in this case $S_{m}$ is a $2$-automatic set (its characteristic sequence is $2$-automatic). ###### Proposition 7.3. For each $m\in\mathbb{N}_{+}$ and $p\in\mathbb{N}$ the sequence $(b_{m}(n)\bmod{2^{p}})_{n\geq 0}$ is $2$-automatic. ###### Proof. Take any $k\geq p-1$ such that $2^{k}\geq m$. Note that $(b_{m}(n))_{n\geq 0}$ is the convolution of the sequence $(b_{2^{k}}(n))_{n\geq 0}$ with $2^{k}-m$ copies of the PTM sequence $(t_{n})_{n\geq 0}$. They are both $2$-regular when treated as sequences over the ring $\mathbb{Z}/2^{p}\mathbb{Z}$ (for $(b_{2^{k}}(n))_{n\geq 0}$ this follows from Lemma 5.1). Hence, $(b_{m}(n)\bmod{2^{p}})_{n\geq 0}$ is $2$-regular as the convolution of $2$-regular sequences. The result follows from the fact that a $2$-regular sequence attaining finitely many values is necessarily $2$-automatic. ∎ Unfortunately, we do not know even for a single value $m\neq 2^{k}-1$, whether or not the valuations $\nu_{2}(b_{m}(n))$ are bounded. It is conjectured that they are unbounded for all $m\neq 2^{k}-1$ (see [10, Conjecture 5.3]). Nevertheless, this does not rule out $2$-automaticity of the set $S_{m}$. Surprisingly, numerical results for $m\leq 30$ (obtained with help of the IntegerSequences package) suggest that the sets $S_{m}$ are $2$-automatic for odd $m$, except for $m=17,21$. It should be possible to get some partial results if we restrict our attention to arithmetic progressions along which $\nu_{2}(b_{m}(n))$ is bounded. For example, [10, Theorem 5.4] provides a collection of suitable arithmetic progressions such that $\nu_{2}(b_{2}(2^{r}n+s))$ is constant. By Proposition 7.3, the set of $n$ such that $b_{2}(2^{r}n+s)$ is a sum of three squares, is $2$-automatic. A related interesting problem concerns the behavior of $b_{m}(n)$ modulo a fixed power of $2$. ###### Problem 7.4. For $m\in\mathbb{N}_{+}$ and $p\in\mathbb{N}$ characterize $b_{m}(n)\bmod{2^{p}}$. We already know that this sequence is $2$-automatic and may ask whether it can be characterized in terms of simpler $2$-automatic sequences. The congruences obtained in the previous sections for subsequences of the form $b_{2^{k}-1}(2^{r}n+s)$ are all “admissible” in the sense of [14], that is, only involve $t_{n}$ and $(-1)^{\nu_{2}(n)}$. In the case $k\geq 3$ Theorem 5.5 provides a congruence in terms of $t_{n},t_{n-1}$ that can be transformed into an admissible one due to the relation $t_{n-1}=(-1)^{\nu_{2}(n)+1}t_{n}$. It turns out that other interesting $2$-automatic sequences already appear if we consider $b_{m}(n)$ modulo suitable powers of $2$. By inspecting modulo $32$ the subsequences described in Proposition 3.7, we have found (without proof) the following set of congruence relations: $\displaystyle b(8n+2)$ $\displaystyle\equiv 2t_{n}+16\sigma_{n}\pmod{32},$ $\displaystyle b(8n+6)$ $\displaystyle\equiv 6t_{n}+16\sigma_{n}+16n\pmod{32},$ $\displaystyle b(16n+8)$ $\displaystyle\equiv(10+8n^{2})t_{n}+16\sigma_{n}\pmod{32},$ where $\sigma_{n}$ counts modulo $2$ the number of blocks of contiguous $1$’s in the binary expansion of $n$. We have later learned that Alkauskas [2, Theorem 2] obtained a set of relations that describe the same sequences and involve the Rudin–Shapiro sequence instead of $(\sigma_{n})_{n\in\mathbb{N}}$. It can be checked that both descriptions are equivalent. Another sequence that arises in this way is the regular paperfolding sequence $(p_{n})_{n\in\mathbb{N}_{+}}$ defined by $p_{2n}=p_{n}$ and $p_{2n+1}=(-1)^{n}$ (see for example [3, Example 5.1.6]). If we let $P(x)=\sum_{n\geq 1}p_{n}x^{n}$, then through manipulation of power series, for $m$ even one can obtain the congruence relation $B_{m}(x)\equiv(1-x)^{m}(1+2mP(x))\mkern 4.0mu({\operator@font mod}\mkern 6.0mu2^{\nu_{2}(m)+3}).$ This is essentially a generalization of Lemma 5.1. We now consider some natural modifications of the original equation $b_{m}(n)=x^{2}+y^{2}+y^{2}$. We have obtained precise characterization of those $n\in\mathbb{N}$ such that $b(2n)$ is a sum of three squares. In particular, the set of such numbers has natural density equal to $11/12$. Analyzing, for a given $n$ not of the form $2^{2k+1}(8s+2t_{s}+3)$, the solution set $(x,y,z)$ of the equation $b(2n)=x^{2}+y^{2}+z^{2}$, we found that in many cases one of the values $x,y,z$ is a square, i.e., the Diophantine equation $b(2n)=x^{2}+y^{2}+z^{4}$ has a solution in non-negative integers. More precisely, for $n\leq 10^{3}$ we know that there are exactly 916 values of $n$ such that $b(2n)$ is a sum of three squares. Among them, there are exactly 831 values of $n$ such that $b(2n)$ is a sum of two squares and a fourth power. This large number of solutions suggest the following conjecture. ###### Conjecture 7.5. Let $Q_{1}:=\\{n\in\mathbb{N}:\;b(2n)=x^{2}+y^{2}+z^{4}\;\mbox{for some}\;x,y,z\in\mathbb{N}\\}$. The set $Q_{1}$ is infinite. Moreover, the set $Q_{1}$ has positive natural density in $\mathbb{N}$. On the other hand, there are exactly seven values of $n\leq 1000$ such that $b(2n)$ is a sum of a square and two fourth powers. This may suggest that the number of solutions of the equation $b(2n)=x^{2}+y^{4}+z^{4}$ is finite. However, due to limited range of our computations we instead formulate the following: ###### Question 7.6. Is the set of $n$ such that $b(2n)=x^{2}+y^{4}+z^{4}$ has a solution in integers $x,y,z$ infinite? An even more interesting and difficult question is whether the set $\mathcal{T}=\\{n\in\mathbb{N}:\;b(2n)=x^{2}+y^{2}\\}$ is infinite or not. Because we know the behaviour of $b(n)\mod{16}$ we can easily prove that the complement of $\mathcal{T}$, i.e., $\mathbb{N}\setminus\mathcal{T}$ is infinite. Indeed, from Proposition 3.2 we have $b(16n+4)\equiv 4t_{n}\mkern 4.0mu({\operator@font mod}\mkern 6.0mu16)$. If $t_{n}=-1$, then $b(16n+4)\equiv 12\mkern 4.0mu({\operator@font mod}\mkern 6.0mu16)$ and thus $b(16n+4)$ is not a sum of two squares. To get a clue what can be expected in the case of the set $\mathcal{T}$, we computed the values of $b(2n)$ for $n\leq 2^{20}$ and check whether $b(2n)$ is a sum of two squares. We put $\mathcal{T}(x)=\\#\\{n\leq x:\;n\in\mathcal{T}\\}$ and in Table 3 we present the values of $\mathcal{T}(2^{n})$ for $n\leq 20$. $n$ | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 ---|---|---|---|---|---|---|---|---|---|--- $\mathcal{T}(2^{n})$ | 2 | 3 | 6 | 8 | 14 | 21 | 37 | 64 | 106 | 174 $n$ | 11 | 12 | 13 | 14 | 15 | 16 | 17 | 18 | 19 | 20 $\mathcal{T}(2^{n})$ | 325 | 617 | 1089 | 2018 | 3699 | 6804 | 12551 | 23624 | 44606 | 84176 Table 3. The number $\mathcal{T}(2^{n})$ for $n\leq 20$. We also define $\mathcal{S}=\\{r_{2}(b(2n)):\;n\in\mathbb{N}\\},$ where $r_{2}(m)$ is the number of representations of $m$ as a sum of two squares. Let us recall that $r_{2}(m)=\sum_{d|m,d\equiv 1\mkern 4.0mu({\operator@font mod}\mkern 6.0mu2)}(-1)^{\frac{d-1}{2}}.$ In the considered range; i.e., $n\leq 2^{20}$ the set $\mathcal{S}$ contains the number 0 and the 35 values $s_{1}\leq\ldots\leq s_{35}$. In Table 4 below, we present the following values: $s_{i}$, $l_{i}$ – the number of times $s_{i}$ is attained, and $n_{i}$ – the smallest value of $n$ such that $r_{2}(b(2n))=s_{i}$. $i$ | $s_{i}$ | $l_{i}$ | $n_{i}$ | $i$ | $s_{i}$ | $l_{i}$ | $n_{i}$ ---|---|---|---|---|---|---|--- 1 | 4 | 4 | 0 | 19 | 224 | 1 | 793875 2 | 8 | 13768 | 4 | 20 | 240 | 1 | 647317 3 | 12 | 2 | 21 | 21 | 256 | 1005 | 15113 4 | 16 | 26411 | 30 | 22 | 288 | 13 | 28561 5 | 24 | 760 | 431 | 23 | 320 | 19 | 113399 6 | 32 | 22889 | 115 | 24 | 384 | 149 | 24877 7 | 40 | 36 | 2522 | 25 | 512 | 202 | 11231 8 | 48 | 1400 | 117 | 26 | 576 | 5 | 420383 9 | 56 | 1 | 27502 | 27 | 640 | 2 | 210415 10 | 64 | 11710 | 482 | 28 | 768 | 23 | 88529 11 | 72 | 9 | 21880 | 29 | 1024 | 27 | 202049 12 | 80 | 46 | 36642 | 30 | 1152 | 1 | 938983 13 | 96 | 1094 | 309 | 31 | 1280 | 1 | 162157 14 | 112 | 2 | 84169 | 32 | 1536 | 5 | 379324 15 | 128 | 4130 | 1036 | 33 | 2048 | 2 | 324442 16 | 144 | 9 | 91925 | 34 | 2560 | 1 | 295411 17 | 160 | 24 | 10785 | 35 | 4096 | 1 | 105400 18 | 192 | 451 | 3085 | | | | Table 4. Values of $s_{i},l_{i}$ and $n_{i}$ for $i\leq 35$. Our numerical computations suggest the following. ###### Conjecture 7.7. The set $\mathcal{T}$ is infinite. The following heuristic reasoning provides further evidence towards our conjecture. More precisely, recall that the counting function of the sums of two squares up to $x$ is $O(x/\sqrt{\log x})$. Thus, one can say that the probability that a random positive integer $n$ can be written as a sum of two squares of integers is $c/\sqrt{\log n}$. Since, $\log_{2}b(n)\approx\frac{1}{2}(\log_{2}n)^{2}$ one could conjecture that the expectation that $b(n)$ is a sum of two squares is $c^{\prime}/\log n$ for some positive constant $c^{\prime}$, provided that $b(n)$ behaves like a random integer of its size. As a consequence, up to $x$, we would have at least $\sum_{n\leq x}\frac{1}{\log n}=\frac{x}{\log x}+O(x/\log^{2}x)$ values of $n$ such that $b(n)$ is a sum of two squares. We dare to formulate the following statement. ###### Conjecture 7.8. There exists a positive real number $c$ such that $\mathcal{T}(x)=c\frac{x}{\log x}+O(x/\log^{2}x)$ as $x\rightarrow+\infty$. Although limited, our computations confirm such an expectation. In Table 5 we give the values $\mathcal{T}(2^{m})\frac{m}{2^{m}}$ for $m=10,\ldots,20$. $m$ | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | 18 | 19 | 20 ---|---|---|---|---|---|---|---|---|---|---|--- $\mathcal{T}(2^{m})\frac{m}{2^{m}}$ | 1.67 | 1.74 | 1.80 | 1.73 | 1.72 | 1.7 | 1.66 | 1.63 | 1.62 | 1.62 | 1.61 Table 5. Values of $\mathcal{T}(2^{m})\frac{m}{2^{m}}$ for $m=10,\ldots,20$. ###### Remark 7.9. The expectation that $b(n)$ behaves like a random integer of its size is very likely. Indeed, numerical computations suggest that for any odd integer $m$ the sequence $(b(n)\pmod{m})_{n\in\mathbb{N}}$ is uniformly distributed; i.e., for any $r\in\\{0,1,\ldots,m-1\\}$ we have $\lim_{N\rightarrow+\infty}\frac{\\#\\{n\leq N:\;b(n)\equiv r\mkern 4.0mu({\operator@font mod}\mkern 6.0mum)\\}}{N}=\frac{1}{m}.$ However, according to the best knowledge of the authors, it is not even known whether the set of prime numbers $p$ such that $p|b(n)$ for some $n$, is infinite. ## References * [1] J.-P. Allouche, B. Cloitre, V. Shevelev, Beyond odious and evil, Aequationes Math. 90(2) (2016), 341–353. * [2] G. Alkauskas, Generalization of the Rødseth–Gupta theorem on binary partitions, Lithuanian Math. Jour. 43 (2) (2003/4), 103–110. * [3] J.-P. Allouche, J. Shallit, Automatic sequences. Theory, applications, generalizations, Cambridge University Press, Cambridge, 2003. * [4] J.-P. Allouche, J. Shallit, The ring of $k$-regular sequences, Theoret. Comput. Sci. 98 (1992), no. 2, 163–197. * [5] J.-P. Allouche, J. Shallit, The ring of $k$-regular sequences II, Theoret. Comput. Sci. 307 (2003), no. 1, 3–29. * [6] R. Burns, Factorials and Legendre’s three-square theorem: II: available at https://arxiv.org/pdf/2203.16469.pdf * [7] R. F. Churchhouse, Congruence properties of the binary partition function, Proc. Cambridge Philos. Soc. 66 (1969), 371–376. * [8] J.-M. Deshouillers, F. Luca, How often is $n!$ a sum of three squares? I: The Legacy of Alladi Ramakrishnan in the Mathematical Sciences, pages 243–251. Springer New York, 2010. * [9] A. Granville, Y. Zhu, Representing binomial coefficients as sums of squares, Amer. Math. Monthly, 97(6) (1990), 486–493. * [10] M. Gawron, P. Miska, M. Ulas, Arithmetic properties of coefficients of power series expansion of $\prod_{n=0}^{\infty}\left(1-x^{2^{n}}\right)^{t}$ (with an Appendix by Andrzej Schinzel), Monatsh. Math. 185 (2018), 307–360. * [11] H. Gupta, Proof of the Churchhouse conjecture concerning binary partitions, Proc. Cambridge Philos. Soc. 70 (1971), 53–56. * [12] H. Gupta, A simple proof of the Churchhouse conjecture concerning binary partitions, Indian J. Pure Appl. Math. 5(3), (1972), 791–794. * [13] H. Gupta, A direct proof of the Churchhouse conjecture concerning binary partitions, Indian J. Math., (1) N.18, (1976), 1–5. * [14] M. D. Hirschhorn, J. H. Loxton, Congruence properties of the binary partition function, Math. Proc. Camb. Phil. Soc., 78(3) (1975), 437–442. * [15] L. Hajdu, A. Papp, On asymptotic density properties of the sequence $n!$, Acta Arith., 184(4) (2018), 317–340. * [16] K. Mahler, On a special functional equation, J. London Math. Soc. 1 (1940), 115–123. * [17] N. Robbins, On Fibonacci and Lucas numbers which are sums of precisely four squares, Fib. Quarterly, 21 (1) (1983), 3–5. * [18] N. Robbins, Representing $\binom{2n}{n}$ as a sum of squares, Fib. Quarterly, 25 (1987), 29–33. * [19] Ø. Rødseth, Some arithmetical properties of $m$-ary partitions, Proc. Cambridge Philos. Soc. 68 (1970), 447–453. * [20] Ø. Rødseth, J. A. Sellers, Binary partitions revisited, J. Comb. Theory, Series A 98 (2002), 33–-45. * [21] Sz. Tengely, M. Ulas, Equal values of certain partition functions via Diophantine equations, Research in Number Theory 7 (4) (2021), Article number: 67. Bartosz Sobolewski, Jagiellonian University, Faculty of Mathematics and Computer Science, Institute of Mathematics, Łojasiewicza 6, 30 - 348 Kraków, Poland e-mail<EMAIL_ADDRESS> Maciej Ulas, Jagiellonian University, Faculty of Mathematics and Computer Science, Institute of Mathematics, Łojasiewicza 6, 30 - 348 Kraków, Poland e-mail<EMAIL_ADDRESS>
# From spin liquid to magnetic ordering in the anisotropic kagome Y-Kapellasite Y3Cu9(OH)19Cl8: a single crystal study Dipranjan Chatterjee Université Paris-Saclay, CNRS, Laboratoire de Physique des Solides, 91405, Orsay, France Pascal Puphal Max-Planck-Institute for Solid State Research, Heisenbergstraße 1, 70569 Stuttgart, Germany Quentin Barthélemy Université Paris-Saclay, CNRS, Laboratoire de Physique des Solides, 91405, Orsay, France Jannis Willwater Institut für Physik der Kondensierten Materie, Technische Universität Braunschweig, Braunschweig, Germany Stefan Süllow Institut für Physik der Kondensierten Materie, Technische Universität Braunschweig, Braunschweig, Germany Christopher Baines Laboratory for Muon-Spin Spectroscopy, Paul Scherrer Institut, 5232 Villigen, Switzerland Sylvain Petit LLB, CEA, CNRS, Université Paris-Saclay, CEA Saclay, Gif-sur-Yvette, France Eric Ressouche Univ. Grenoble Alpes, CEA, IRIG, MEM, MDN, 38000 Grenoble, France Jacques Ollivier Institut Laue- Langevin, 38042 Grenoble, France Katharina M. Zoch Physikalisches Institut, Goethe-Universität Frankfurt, Frankfurt am Main, Germany Cornelius Krellner Physikalisches Institut, Goethe-Universität Frankfurt, Frankfurt am Main, Germany Michael Parzer Institute of Solid State Physics, TU Wien, 1040 Vienna, Austria Alexander Riss Institute of Solid State Physics, TU Wien, 1040 Vienna, Austria Fabian Garmroudi Institute of Solid State Physics, TU Wien, 1040 Vienna, Austria Andrej Pustogow Institute of Solid State Physics, TU Wien, 1040 Vienna, Austria Philippe Mendels Université Paris-Saclay, CNRS, Laboratoire de Physique des Solides, 91405, Orsay, France Edwin Kermarrec Université Paris-Saclay, CNRS, Laboratoire de Physique des Solides, 91405, Orsay, France Fabrice Bert Université Paris-Saclay, CNRS, Laboratoire de Physique des Solides, 91405, Orsay, France ###### Abstract Y3Cu9(OH)19Cl8 realizes an original anisotropic kagome model hosting a rich magnetic phase diagram [M. Hering _et al_ , npj Computational Materials 8, 1 (2022)]. We present an improved synthesis of large phase-pure single crystals via an external gradient method. These crystals were investigated in details by susceptibility, specific heat, thermal expansion, neutron scattering and local $\mu$SR and NMR techniques. At variance with polycristalline samples, the study of single crystals gives evidence for subtle structural instabilities at 33 K and 13 K which preserve the global symmetry of the system and thus the magnetic model. At 2.1 K the compound shows a magnetic transition to a coplanar (1/3,1/3) long range order as predicted theoretically. However our analysis of the spin wave excitations yields magnetic interactions which locate the compound closer to the phase boundary to a classical _jammed_ spin liquid phase. Enhanced quantum fluctuations at this boundary may be responsible for the strongly reduced ordered moment of the Cu2+, estimated to be $\sim$ 0.075 $\mu_{B}$ from $\mu$SR. ## I Introduction Low-dimensional materials with strong magnetic frustration, such as compounds with decoupled antiferromagnetic kagome layers, are prototypical systems to search for an experimental realization of the quantum spin-liquid (QSL) state [1, 2]. This long sought state features no static magnetic order, despite sizeable magnetic interactions, but rather macroscopic entanglement and fractional excitations. Experimental signatures of a QSL state in kagome materials were found first in herbertsmithite, ZnCu3(OH)6Cl2 [3, 4] and in the closely related Zn-doped barlowite [5, 6]. Thus Cu-based systems with a perfect kagome layer have been strongly investigated. Around the A2+Cu3(OH)6Cl2 family [7] a new similar series of compounds has been found with A3+Cu3(OH)6Cl3 [7, 8, 9], which realizes perfect kagome layers in a kapellasite-type structure. However, for Y3+ the compound is not stable in water solutions, leading to a Cl-OH substitution YCu3(OH)6+xCl3-x with $x=1/3$ (or Y3Cu9(OH)19Cl8) [10, 11]. This substitution induces a distortion of the magnetic lattice as the Y atoms are moved out of the kagome plane [10]. The distorted kagome lattice of Cu2+ ions is scarcely studied with the exception of the family around (Cs,Rb)2Cu3(Ti,Sn,Zr,Hf)F12 [12, 13, 14, 15]. However recent structural investigations of low temperature diffraction on various kagome systems show that a structural instability induces distortions in barlowite, claringbullite [16], volborthite [17] and vesigneite [18] making it a typical structural motif at low temperatures. In Y-kapellasite ($x=1/3$) that we investigate in this article, the distortion yields two nonequivalent Cu sites and a unique magnetic model with three different nearest neighbor interactions, while still retaining rotational symmetry around the hexagons of the anisotropic kagome lattice (see inset Fig. 11 and Ref. [11]). The discovery of this compound has triggered a detailed theoretical study [19] which disclosed the full classical phase diagram of the model and its richness. Interestingly, besides two long range ordered phases with propagation vectors $Q=(1/3,1/3)$ and $Q=(0,0)$, a large area in the phase diagram, coined ”classical spin liquid” phase, which encompasses the isotropic kagome model, opens up. It remains so far largely unexplored, although it could realize an unprecedented ”jammed” spin liquid phase, characterized by a discrete ground state degeneracy, in a disorder free model. Besides, first principle calculations confirm the relevance of this anisotropic nearest-neighbor model for Y-kapellasite and locates it in the $(1/3,1/3)$ long range ordered phase [19]. Here, we report a comprehensive study of the physical properties of large phase pure single crystals that can be compared to the theoretical prediction and which show marked deviations with respect to former studies of polycristalline samples [11] or other synthesis route for crystals [20]. The paper is organized as follows. In sections II and III the synthesis of phase-pure large single crystals is described and macroscopic measurements of the susceptibility, heat capacity and thermal expansion are presented, providing evidence for a series of structural and magnetic transitions below 33 K. In sections IV and V, the transitions are investigated in detail by neutron diffraction and by Cl NMR, which shows in particular good agreement with the theoretically predicted spin order in the ground state. In section VI, muon spin spectroscopy is used to prove the bulk nature of the 2.1 K magnetic transition. In section VII, the spin wave excitations are studied by inelastic neutron scattering and give a novel insight into the determination of the magnetic interactions beyond first principle calculations. Finally in section VIII, we summarize and discuss our results in comparison with previous ones on polycristalline samples and in light of theoretical studies. ## II Synthesis and susceptibility The crystal growth of Y-Kapellasite was originally reported in Ref. [10], where 0.59 g Y2O3, 0.82 g CuO and 0.89 g CuCl${}_{2}\cdot$2(H2O) in 10 ml H2O were heated up to the dissolution point of Y2O3, followed by a slow cooling to crystallize. However, these crystals of an average size of 1x1x1 mm3 suffered small CuO inclusions since the growth takes place on the surface of the polycrystalline CuO starting material, as the dissolution point of Y2O3 and the crystallization point of the compound lie above the maximum in solubility of CuO. Reference [20] describes a synthesis of inclusion free crystals using LiOH, Y(NO3)${}_{3}\cdot 6$H2O and CuCl${}_{2}\cdot 2$H2O. However the reported crystals are somewhat disordered as a partial occupation of the Y site is observed, similarly to what was found in the related compound YCu3(OH)6Cl3 [8]. This likely hints at a phase mixture in both cases due to contact with water. Indeed, a low percentage of occupation of Y for YCu3(OH)6Cl3 [8] lies out of the kagome plane, which represents structural parts of Y-Kapellasite Y3Cu9(OH)19Cl8. For the reported inclusion free Y3Cu9(OH)19Cl8 [20], the partial occupation of Y can be explained similarly, since the specific heat shows a mixture of the signal observed for $x=0$ and $x=1/3$ as shown in Fig. 2. We thus improved the synthesis for inclusion-free large bulk single crystals suitable for neutron studies via a horizontal external gradient growth method in a thick-walled quartz ampule as described in detail for herbertsmithite in Ref. [21]. The growth is realized by slowly dissolving CuO in a YCl3-H2O (or YCl3-D2O for deuterated crystals) solution and transporting it to the cold end. The growth is executed in a three zone furnace with a gradient of 25$\degree$C and a temperature of 240$\degree$C at the hot end (note that the elevated pressure at this temperature requires especially thick quartz ampules). The gradient was optimized as too low temperatures yielded a mixture of Y-Kapellasite and Clinoatacamite. Afterwards, the inclusion-free hexagonal single crystals have an average size of 3x3x1 mm3 up to 3x3x3 mm3, if grown over several weeks. The pictures of single crystalline Y-kapellasite are displayed in Fig. 1 a) and b), which show a transparent specimen without any visible impurity inclusions. The dc-susceptibility of a 15.7 mg single crystal of Y-Kapellasite measured in a Quantum Design MPMS XL7 squid system with a 1 T field applied along the $c$ axis is shown in Fig. 1 c). It reproduces published results [10, 22], with an antiferromagnetic Curie-Weiss constant $\theta_{\rm{cw}}=102(1)$ K and $g_{\rm{c}}=2.40(1)$. Measurements in a lower 0.01 T magnetic field, shown in the inset, reveal a bifurcation of the field cooled and zero field cooled curves below $\sim 5$ K, and a maximum around 2.5 K. These low field measurements point at a magnetic transition which is investigated in details with neutron scattering and resonance techniques in the following sections. At variance with reference [20], we observe no magnetic transition at 11 K. Such a transition, reminiscent of the one observed at 12 K in the $x=0$ counterpart, supports a phase mixture scenario for the crystal of reference [20]. Figure 1: a) Y3Cu9(OH)19Cl8 single crystal along the $c$ axis. b) Image in transmission light of an as grown single crystal. c) Susceptibility (M/H) versus temperature with a 1 T field applied along $c$. Bottom left inset: field cooled (FC) and zero field cooled (ZFC) measurements at 0.01 T. Upper right inset: anisotropic kagome lattice with 3 different nearest neighbor Cu- Cu bonds and the related 3 different magnetic interactions. ## III Specific heat and thermal expansion Figure 2: a) Specific heat divided by temperature in the temperature range from 1.8 to 40 K for the $x=1/3$ optimized large crystals, compared to literature results for $x=1/3$ and $x=0$ powder samples [11] and the reported inclusion free $x=1/3$ crystals from Ref. [20]. b) Thermal expansion along the $s$ direction of an optimized large Y-Kapellasite single crystal. c) and d) Insets: enlarged views of the peaks detected in main panel b). Note the different vertical scales in the two insets. Specific heat was measured with the standard option of a Quantum Design Physical Properties Measurement System (PPMS), using a 5.935 mg inclusion-free single crystal formerly described. Its temperature evolution is shown in Fig. 2 a) and compared to the published results for $x=0$ and $x=1/3$ polycrystalline samples [11] and the $x=1/3$ single crystal reported in Ref. [20]. Unexpectedly, our $x=1/3$ single crystal presents a clear double peak at 33 K, which is absent in powder samples of the same compound. The underlying entropy amounts to 0.1537 J/mol K, below 3% of $R$ln2, calculated by using the powder sample as a baseline. There is no associated sharp feature in the susceptibility in this temperature range so that the double peak is likely signaling a structural transition. Note that this 33 K anomaly was not reported in the single crystals of Ref. [20] and that, on the contrary, the large hump around 15 K seen in the latter crystals and in the $x=0$ powder sample, associated to magnetic transitions, is absent from our phase-pure $x=1/3$ single crystal. Only at 2.1 K do we find a sharp peak in specific heat, which was already reported for this compound both in powder and crystals [10, 11], and stems from a magnetic long range ordering transition as shown later in this article. For the disordered powder sample $x=1/3$, the underlying entropy of the transition is reduced and previous muon spin relaxation studies had shown the absence of long range order [11], possibly arising from the slightly altered structure due to the suppression of the 33 K transition. To inspect the 33 K anomaly in more detail, we performed thermal expansion experiments along the crystallographic $c$-axis, i.e. perpendicular to the kagome layers, using a capacitive dilatometer. Figure 2 b) displays the thermal expansion coefficient $\alpha$ on the same temperature scale as the specific heat data in panel a). We find a similar anomaly below 33 K as in $C/T$, consisting of two slightly broadened, and thus overlapping peaks with a ratio between the sizes of the higher-temperature and lower-temperature peaks comparable to a). The relative length change incorporated in these two sharp features amounts to $\Delta L/L=10^{-4}$. In addition, we discover another, significantly smaller anomaly around 13 K, which can be seen in more details in the inset c). Below 5 K an anomaly in $\alpha$ forms as approaching the transition at 2.1 K, signalling the importance of magneto-elastic coupling as antiferromagnetism sets in. Overall, the anomalies at 13 K and 33 K reveal clear structural effects in this compound with a distorted kagome lattice, without noticeable changes upon applying a magnetic field of 9 T along the $c$-axis. ## IV Diffraction T (K) | 0.065 | 8 | 20 | 40 ---|---|---|---|--- | x/a | y/b | z/c | x/a | y/b | z/c | x/a | y/b | z/c | x/a | y/b | z/c Y1 | 0 | 0 | 0.1318(7) | 0 | 0 | 0.1314(6) | 0 | 0 | 0.1324(7) | 0 | 0 | 0.1302(5) Y2 | 0 | 0 | 0.5 | 0 | 0 | 0.5 | 0 | 0 | 0.5 | 0 | 0 | 0.5 Cu1 | 0.6661(10) | 0.8325(11) | 0.5025(4) | 0.6658(10) | 0.8340(11) | 0.5021(4) | 0.6675(11) | 0.8328(12) | 0.5026(4) | 0.6645(7) | 0.8317(7) | 0.5021(3) Cu2 | 0.5 | 0 | 0.5 | 0.5 | 0 | 0.5 | 0.5 | 0 | 0.5 | 0.5 | 0 | 0.5 Cl1 | 0.6745(9) | 0.0033(10) | 0.2845(2) | 0.6743(8) | 0.0037(10) | 0.28435(19) | 0.6748(9) | 0.0031(11) | 0.2845(2) | 0.6732(6) | 0.0027(6) | 0.28395(15) Cl2 | 0 | 0 | 0.3368(5) | 0 | 0 | 0.3368(5) | 0 | 0 | 0.3365(6) | 0 | 0 | 0.3373(4) O1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 D1 | 0.456(9) | 0.776(10) | 0.660(4) | 0.455(9) | 0.783(10) | 0.664(4) | 0.455(9) | 0.776(10) | 0.660(4) | 0.462(7) | 0.777(7) | 0.661(3) O2 | 0.8119(14) | 0.8060(12) | 0.5435(7) | 0.8122(13) | 0.8046(12) | 0.5438(7) | 0.8129(14) | 0.8060(13) | 0.5438(8) | 0.8121(9) | 0.8035(9) | 0.5435(4) D2 | 0.7769(19) | 0.7778(17) | 0.6011(9) | 0.7874(18) | 0.7858(17) | 0.5991(9) | 0.7723(17) | 0.7722(16) | 0.6030(8) | 0.7866(13) | 0.7827(13) | 0.6005(6) O3 | 0.5293(12) | 0.6635(11) | 0.5540(6) | 0.5285(11) | 0.6638(10) | 0.5537(6) | 0.5283(12) | 0.6622(11) | 0.5542(6) | 0.5290(8) | 0.6634(8) | 0.5543(4) D3 | 0.5494(17) | 0.6731(18) | 0.6065(8) | 0.5575(16) | 0.6709(17) | 0.6082(8) | 0.5452(15) | 0.6740(18) | 0.6051(8) | 0.5548(12) | 0.6692(13) | 0.6069(5) O4 | 0.5142(12) | 0.8433(11) | 0.4631(5) | 0.5167(11) | 0.8430(10) | 0.4632(5) | 0.5154(12) | 0.8450(11) | 0.4634(6) | 0.5162(8) | 0.8441(8) | 0.4638(4) D4 | 0.5008(16) | 0.8310(17) | 0.4096(8) | 0.5013(16) | 0.8328(16) | 0.4097(8) | 0.5018(17) | 0.8320(17) | 0.4099(8) | 0.5029(12) | 0.8337(12) | 0.4099(6) Table 1: Rietveld refinement results of neutron diffraction data obtained at 65 mK, 8 K, 20 K and 40 K with R-3 (#148) $a=b=11.539$ Å, and $c=17.1355$ Å. Figure 3: Temperature dependence of the neutron diffraction intensity around the nuclear Bragg peaks a) (6 6 0) and b) (0 0 18). Structure viewed along the $c$-axis obtained from neutron single crystal diffraction measurements at c) 40 K and d) 65 mK. The structure of Y-Kapellasite Y3Cu9(OH)19Cl8 with R-3c was first reported in Ref. [10], where hydrogen was placed by symmetry arguments. Later, from neutron diffraction on deuterated powder samples, we found that the deuterium atom next to the O1 refines to a distance above 1 Å, which is so far unaccounted for an O-H bound. Thus in Ref. [11], we dismissed the D1, leading to the revised chemical formula of Y3Cu9(OD)18OCl8. To tackle further this issue we then measured the stoichiometry including the hydrogen content with a combination of inductively coupled plasma mass spectroscopy (ICP-OES) and gas extraction, the former with a SPECTRO CIROS CCD and the latter with an Eltra ONH-2000 analyzer. For the determination of oxygen and hydrogen content, the powder samples were placed in a Ni crucible and clipped and heated, where the carrier gas takes the oxygen and hydrogen out of the sample. The oxygen reacts with carbon, and CO2 is detected in an infrared cell, while hydrogen is detected by a thermal conductivity cell. Each measurement is repeated three times and compared to a standard. The quoted error bars provide the statistical error. We thus measured three pieces of a hydrated single crystal and a part of a powder sample of Ref. [11]. The results of: Y3.00(3)Cu9.06(9)O19.0(2)H18.7(3)Cl8 for the crystals and Y3.00(3)Cu9.07(9)O19.3(2)H18.8(3)Cl8 for the powder sample show a high enough hydrogen content to assume the ideal stoichiometry of Y3Cu9(OH)19Cl8. Notably we have also seen vibrational O-H modes that exist only for electrical field parallel to the kagome plane solely matching to this hydrogen/deuterium position further confirming the presence of the D1 atom. Therefore, from now on, we use the ideal stoichiometry of Y3Cu9(OH)19Cl8. We performed neutron diffraction on the 2 axis thermal neutron diffractometer D23 instrument at the Institut Laue-Langevin (ILL) using an inclusion-free, deuterated single crystal weighing around 30 mg. Measurements have been performed in the temperature range of 40 K to 0.065 K using constant incident neutron wavelength $\lambda$=1.27 Å. We find the R-3c structure that was reported in Ref. [10] to be stable over the whole temperature range. The previously mentioned D1 position can be reasonably refined in all cases to a similar position and even a release of the occupation is possible. We find in all cases an unusually large distance of d${}_{O-D}=1.3$ Å, which was nevertheless considered as mentioned above. The transition around 33 K observed in specific heat and thermal expansion is reflected here in an increase of the Bragg peak intensity by approximately 30% as shown in Fig. 3 a) and b). Yet, it produces no change in the crystallographic cell, as the crystallographic axes $a,b$ and $c$ remain unaltered in the investigated temperature range. We find in Rietveld refinements a sudden jump around 33 K for certain atomic positions, which in absolute numbers remain however marginal. The Rietveld refinement results are summarized in table LABEL:Rietveld and selected atomic positions are plotted normalized to the 40 K values versus temperature in Fig. 4 to visualize the sudden changes around 30 K (see also 111See Supplemental Material at [URL] for the Crystallographic Information Files corresponding to the refined Y-Kapellasite Y3Cu9(OH)19Cl8 structure at 40 K, 20 K, 8 K and 0.065 K.). These changes lie below 1% of the nominal value, except for deuterium and thus are hardly visible in Fig. 3 c) and d) depicting the crystal structure along the $c$ axis at 40 K and 65 mK, but will be relevant for the discussion of our NMR results. Figure 4: Relative atomic position changes versus temperature of a) Chlorine site 1 and 2, b) Deuterium site 1, 2, 3, 4 and c) Cu site 1. At low temperature $T=65$ mK we have searched for magnetic ordering at different (H,K,L) planes but we found no magnetic Bragg peaks probably due to low magnetic moment of Cu2+ ions in the ground state. Hence extensive studies of the low temperature regime has been done to investigate the magnetic structure using NMR, $\mu$SR and inelastic neutron scattering, as presented in the following sections. ## V Nuclear Magnetic Resonance We further investigated the structural distortions and magnetic transition using 35Cl NMR on a $\sim 30$ mg single crystal of Y-Kapellasite Y3Cu9(OH)19Cl8 with an external magnetic field Bext applied parallel to the $c$-axis, i.e. perpendicular to the large facet of the sample (see Fig. 1). The spectra are measured from room temperature down to 5 K by sweeping the frequency in a fixed field B${}_{ext}=7.553$ T which corresponds to the reference frequency $\nu_{0}=31.510$ MHz. At lower temperatures, the broader spectra were recorded by sweeping the field with a fixed irradiation frequency $\nu_{0}$. The spectrum measured at 100 K, far above the magnetic transition and the structural distortions, is shown in Fig. 5. We observe two distinct central lines (see inset of Fig. 5) corresponding to the two crystallographic Cl sites with intensity ratio close to the expected 3:1. There are indeed 18 Cl1 (resp. 6 Cl2) sites per unit cell located in between the kagome layers, close to the center of each triangle (resp. hexagon) of the kagome structure. For clarity in this section we denote the Cl1 and Cl2 sites as the triangular ($tri$) and hexagonal ($hex$) sites respectively. Figure 5: Full NMR spectrum of 35Cl at 100 K with the configuration $B_{ext}\parallel c$ which shows 2 central lines (C) of Cl1 (triangular site) and Cl2 (hexagonal site) and the 2 corresponding pairs of satellites (Stri and SHex). Inset: enhanced view of the central lines (C). The dashed vertical line indicates the reference frequency $\nu_{0}=31.510$ MHz Since 35Cl possesses a nuclear spin $I=3/2$ with a finite quadrupolar moment, the three allowed transitions between the Zeeman-split adjacent nuclear levels are modified by the quadrupolar interaction with the surrounding electric charges. We therefore observe in the full spectrum three NMR lines per Cl site, one central line ($-1/2\leftrightarrow 1/2$ transition) and two satellite ones ($-3/2\leftrightarrow-1/2$ and $1/2\leftrightarrow 3/2$ transitions). At first order in perturbation, the frequency difference between two satellite lines of a given $\alpha=tri$ or $\alpha=hex$ Cl site arising from the quadrupolar interaction is given by [24] $\Delta\nu_{\alpha}^{(1)}=\nu^{\alpha}_{Q}(3\cos^{2}{\theta_{\alpha}}-1-\eta_{\alpha}\sin^{2}{\theta_{\alpha}}\cos 2\phi_{\alpha})$ (1) where $\theta_{\alpha}$ and $\phi_{\alpha}$ are the polar and azimuthal angles defining Bext in the local frame of the principal axes of the electric field gradient tensor (EFG), $\nu_{Q}^{\alpha}$ is proportional to the quadrupolar moment and the largest eigenvalue of the EFG, therefore reflecting the strength of the quadrupolar interaction, and the asymmetry parameter $\eta_{\alpha}$ reflects the departure of the EFG from cylindrical symmetry. Because of the high symmetry of the hexagonal Cl site, the EFG at this site is axially symmetric ($\eta_{hex}=0$) along the $c$ axis and, under the condition of the experiment $B_{ext}\parallel c$, $\theta_{hex}=0$. For this site, the two satellites are maximally separated by $2\nu_{Q}^{hex}$ and we measure directly from Fig. 5 $\nu_{Q}^{hex}=3640(5)$ kHz. The triangular Cl site is less symmetric. We used the structure determined at 40 K in section IV by neutron diffraction to compute the EFG at this site with a point charge approach. From the calculated parameters $\theta_{tri}=8.3^{\circ}$, $\phi_{tri}=123.0^{\circ}$, $\eta_{tri}=0.33$ and the measured distance between satellites, we obtain with Eq. 1 the value $\nu_{Q}^{tri}=2180(20)$ kHz. Note that if Bext is not strictly applied along the symmetry axis $c$, the triangular sites become non-equivalent because of a different orientation of their EFG with respect to Bext. The minute splitting of the triangular site satellite lines in Fig. 5 reflects such a small misalignment $\pm 1.2^{\circ}$ of our crystal in the experiment. The quadrupolar interaction shifts the position of the central line only at second order in perturbation [25]: $\nu^{(2)}_{\alpha}=-\frac{(\nu^{\alpha}_{Q})^{2}}{2\nu_{0}}f(\theta_{\alpha},\phi_{\alpha},\eta_{\alpha})$ (2) where, for the simplest case of an axially symmetric EFG, $f$ reads $f(\theta,\phi,0)=\frac{3}{8}(1-\cos^{2}{\theta})(9\cos^{2}{\theta}-1).$ (3) In the present case at 100 K, the second order quadrupolar shift is either $\nu^{(2)}_{hex}=0$ since $\theta_{hex}=0$ or negligible $\nu^{(2)}_{tri}=-4(1)$ kHz. The positions of the two central lines $\nu^{C}_{\alpha}=\nu_{0}(1+K_{\alpha}^{0}+K^{s}_{\alpha})+\nu^{(2)}_{\alpha}$ (4) are then rather set by the first term of magnetic origin, where $K_{\alpha}^{0}$ is the temperature-independent orbital shift and $K^{s}_{\alpha}$ the spin shift arising from the polarization of the unpaired electron of the neighbouring Cu2+ ions and therefore proportional to the spin susceptibility. From the plots of $\nu^{C}_{\alpha}$ versus the susceptibility $\chi_{c}$ measured along the $c$ direction in the range 100-300 K (not shown), we extract for the configuration B${}_{ext}\parallel c$ the hyperfine constants $A_{hex}=0.245(3)$ T/$\mu_{B}$ and $A_{tri}=-0.426(16)$ T/$\mu_{B}$ and the orbital terms $K_{hex}^{0}=120(10)$ ppm and $K_{tri}^{0}=52(10)$ ppm. Figure 6: Evolution of the central lines of the 35Cl NMR spectra with temperature. The spectra are shifted vertically for clarity. Note the change of the width and the splitting of the line below 40 K which become prominent below 20 K. The dashed vertical line indicates the reference frequency. Figure 6 shows the evolution of the central lines upon cooling down to 5 K. Marked changes of the spectral shape are observed below 40 K where the structural distortions take place: the triangular site line broadens significantly and the hexagonal one splits. Interestingly, although the changes appear in the range 30-40 K, they really get prominent only below 20 K where a second peak is observed in thermal expansion. From the neutron diffraction study, the structural distortion at 33 K does not affect the crystal symmetry and maintains only two distinct Cl sites. It mainly involves small displacements of the protons, which naturally modify the EFG and therefore the positions of the lines but cannot account for line broadening or line splitting. Therefore, 35Cl NMR reveals more complex distortions with local symmetry breaking but no spatial correlation so that the average symmetry as detected by neutron diffraction is preserved. In particular, the splitting of the hexagonal site line implies the occurrence of at least two locally non-equivalent crystallographic sites and at least one with a finite tilt $\theta_{hex}$ of its EFG so that $\nu^{(2)}_{hex}$ does not vanish in equation 4. From the measured splitting and using Eq. 4, we estimate $\theta_{hex}\sim 5^{\circ}$ at 20 K and $\theta_{hex}\sim 10^{\circ}$ at 10 K. These changes of the local electrostatic environment of Cl affect even more clearly, at first order, the satellite lines. As shown in the inset a) of Fig. 7, the hexagonal satellite acquires a complex structure below 33 K pointing at even more than two nonequivalent sites. In order to determine the onset of the structural transition, we have measured the width of this satellite versus temperature as plotted in Fig. 7. Beside the abrupt change at the $33$ K structural transition evidenced in heat capacity and neutron diffraction experiments, a marked change of the slope is detected at 13(1) K suggesting a secondary transition as detected in thermal expansion and also clearly in NMR relaxation measurements as discussed further in this section. Figure 7: Hexagonal satellite linewidth evolution with temperature which indicates 2 structural changes (black arrows). Inset (a) depicts the line shape changes occurring below 33 K for the hexagonal satellite. Inset (b) shows the hexagonal NMR central lines for 37Cl and 35Cl at 10 K. To confirm the structural nature of the 13 K transition we have compared the 35Cl and 37Cl NMR spectra at 10 K at the same reference frequency of 24.000 MHz. We have fitted the split hexagonal site central lines with gaussian functions and compared the frequency differences $\Delta\nu^{i}$ between the two peak maxima for the two isotopes $i$ (inset (b) Fig. 7). The ratio $\Delta\nu^{35}/\Delta\nu^{37}=1.5(2)$ compares well to the square of the ratio of the quadrupolar moments of the two isotopes 1.61 as expected from Eq. 4. At variance, in case of a magnetic origin with static internal fields, we would have observed a scaling with the gyromagnetic ratios of the two isotopes, i.e. a ratio $\sim 1.20$. At lower temperatures ($T<4.2$ K), the NMR lines broaden significantly, likely as a result of the building up of magnetic correlations, in line with the FC/ZFC divergence observed in this $T$ range in the susceptibility (see Fig. 1). This magnetic broadening blurs the quadrupolar splitting and down to 2.1 K, we observe only two broad central lines, overlapping partially, for the two different, triangular and hexagonal, chlorine sites (see Fig. 8). At lower temperatures an abrupt broadening for the triangular site and more moderately for the hexagonal one occurs, which represents the spectral NMR signature of the magnetic transition detected in the bulk, $T_{1}$ or $\mu$SR measurements. Actually, two triangular site peaks develop on both sides of the hexagonal central line and move away from each other upon cooling down. To quantify these evolutions, we compare in Fig. 9 the change of the linewidth $W(T)$ with temperature for the two sites. Figure 8: Central line spectra at 1.3 K, 2 K, 3.6 K and partial view of the ground state spin texture proposed in Ref. [19] in the top right corner. The spectra are shifted vertically for clarity. The blue line is a gaussian fit of the hexagonal central line. This contribution is subtracted from the experimental spectrum to expose the triangular site one, shown as a red line. The triangle configurations a, b and c yield three different dipolar fields at the triangular Cl site. Configurations a and c yield almost equal but opposite dipolar fields whereas configuration b has almost zero dipolar field along the $z$ axis, i.e. the applied field direction. In the recent theoretical study Ref. [19], a $(1/3,1/3)$ long range magnetic order has been proposed for this system with a coplanar spin configuration reproduced partially in Fig. 8. Remarkably, for this predicted spin configuration, the hyperfine field at the hexagonal sites is zero (hexagons with different colors spins) or nearly zero (hexagons with same color spins). At variance, for each triangle, two spins are nearly opposite leaving the third one uncompensated and thus able to yield a finite field at the triangular sites. There are actually three types of triangle configurations labelled a, b and c in Fig. 8 corresponding to three different orientations of this uncompensated spin. For an anisotropic hyperfine tensor, we thus expect three different lines for the triangular Cl in the magnetically ordered phase. Although we could not determine experimentally all the components of the hyperfine tensor, we note that assuming a purely dipolar coupling, the shown spin configuration yields one very small and two almost equal but opposite $z$-components of the transferred fields at the triangular Cl site. Thus, the low temperature spectrum supports at least qualitatively the theoretically proposed order, both for the triangular and hexagonal Cl sites. Figure 9: Evolution versus temperature of the linewidth $W(T)$ for the triangular (red dots) and hexagonal (blue dots) sites. Figure 10 shows the evolution with temperature of the spin lattice relaxation rate ($T_{1}^{-1}$) measured using the saturation recovery sequence at both Cl sites. Being quadrupolar active, Cl nuclei probe both the magnetic and structural fluctuations. In the temperature evolution of $T_{1}^{-1}$ the structural distortions at $\sim$33 K and $\sim$13 K discussed earlier are clearly observed as a sharp and broad peak respectively, together with the magnetic transition around $\sim$ 2.1 K. To extract the $T_{1}$ values, the recovery curves of the nuclear magnetization $M(t)$ at each temperature were fitted to [26] $M(t)=M_{sat}\left(1-ae^{-\left(\frac{t}{T_{1}}\right)^{\beta}}-be^{-\left(\frac{6t}{T_{1}}\right)^{\beta}}\right)$ (5) where $a+b=1$, $a=0.1$ and $0.23$ respectively for the hexagonal and triangular sites, and $\beta\leq 1$ is a phenomenological stretching parameter used to account for a possible distribution of $T_{1}$ [27]. The evolution of $\beta(T)$ for both sites is shown as insets in Fig. 10. While a single value of $T_{1}$ is measured at high temperature ($\beta=1$), an increasingly broader distribution ($\beta<1$) is needed upon cooling down towards the structural and magnetic transitions, likely pointing at the building up of spatial correlations. During the transition regimes, we could fix the value of $\beta$ with no noticeable reduction of the fit quality, to avoid artefacts in the determination of the peaks of the correlated $T_{1}^{-1}$ quantity. Figure 10: Plots of 35Cl NMR relaxation rate $T_{1}^{-1}$ versus temperature measured at the hexagonal (top) and triangular (bottom) central line positions. The two structural transition regimes (2) and (3) and the magnetic one (1) are highlighted. The $T_{1}$ distribution parameter $\beta$ is plotted in the inset. Below 5 K, when the NMR lines of the two sites overlap, $T_{1}$ was also measured at the isolated hexagonal satellite position (dark cyan dots) and shows the same behaviour as the central line. ## VI Muon spin relaxation To elucidate the nature of the ground state of Y- Kapellasite, we performed a $\mu$SR experiment on the GPS and DOLLY instruments at the PSI muon facility from the paramagnetic regime ($>$20 K) down to 0.28 K. The 100% spin polarized muons are implanted at time zero in the sample, where they stop at the most electronegative positions. The muon spins evolve in the local fields until the muons decay into positrons, emitted preferentially in the last direction of the muon spins after a mean life time of 2.2 $\mu$s. By collecting the positrons emitted forward and backward with respect to the muon beam direction, the muon decay asymmetry, proportional to muon spin polarization can be recorded versus time and allows to characterize the internal fields and their fluctuations [28, 29, 30]. For this $\mu$SR experiment, we used two coaligned phase pure single crystals of sizes $\sim 3\times 3\times 1$ mm3 with their $c$ axis parallel to the initial muon spin direction. The muon decay asymmetry in zero external field (ZF) is depicted at some selected temperatures in Fig. 11. At low temperature, the asymmetry evidently differs from previous measurements [11] on a powder sample shown in the inset, with a much faster relaxation and a characteristic damped oscillation signaling static internal fields. Figure 11: Time evolution of the zero-field (ZF) muon decay asymmetry at some selected temperatures in single crystal samples. Lines are fit to Eq. 9. Inset: powder sample data from Ref. [11]. Above 2.5 K, the relaxation hardly depends on temperature nor on the nature of the sample, powder or single crystal. The electronic spins are fast fluctuating in their paramagnetic regime, and the relaxation is dominated by the quasi static, weak and random, nuclear fields. Following the model used for the powder sample [11], we fitted the muon decay asymmetry at high temperature to $a_{0}P_{para}(t)=a_{0}\left[fP_{OH}(t)+(1-f)KT_{\Delta_{Cl}}(t)\right]\\\ $ (6) where $P_{OH}(t)=e^{-\frac{\Delta^{2}_{OH}t^{2}}{2}}{\bigg{[}}\frac{1}{6}+\frac{1}{3}\cos{{\bigg{(}}\frac{\omega_{OH}t}{2}{\bigg{)}}}+\\\ \frac{1}{3}\cos{{\bigg{(}}\frac{3\hskip 2.84526pt\omega_{OH}t}{2}{\bigg{)}}}+\frac{1}{6}\cos{{\bigg{(}}{\omega_{OH}t}{\bigg{)}}}{\bigg{]}}.\ $ (7) The latter $P_{OH}(t)$ stands for the formation of $\mu$OH complexes with a pulsation $\omega_{OH}=\frac{\hbar\mu_{0}\gamma_{\mu}\gamma_{H}}{4\pi d^{3}}$ which depends on the $\mu$-H distance $d$, the gyromagnetic ratios $\gamma_{H}=267.513$ Mrad/s/T and $\gamma_{\mu}=851.616$ Mrad/s/T respectively for protons and muons, and a gaussian broadening $\Delta_{OH}$ due to the other surrounding nuclear spins. The second term in Eq. 6 is a static Kubo- Toyabe relaxation standing for the minority fraction $1-f$ of muon stopping sites, likely close to the Cl${}^{\textbf{-}}$ ions, where they experience a gaussian distribution of static nuclear fields with a width $\Delta_{Cl}$. The parameter $a_{0}=0.255$ is the initial muon decay asymmetry in our experimental conditions. The nuclear relaxation parameters in Eq. 6 and Eq. 7 evaluated by fitting the ZF muon polarisation at 20 K are presented in table 2. Static parameters | Y3Cu9(OH)19Cl8 ---|--- $f\%$ | 75.00$\pm$1.00 $\omega_{OH}$ (Mrad.s-1) | 0.55$\pm$0.02 d (Å) | 1.63$\pm$0.02 $\Delta_{Cl}~{}(\mu s^{-1})$ | 0.09$\pm$0.02 $\Delta_{OH}$ ($\mu$s-1) | 0.208$\pm$0.02 Table 2: Static nuclear parameters derived from high temperature fit of the ZF asymmetry with Eq. 6. Upon cooling, the relaxation from the electronic spins increases progressively and below 2.1 K, at variance with the powder sample [11], a strongly damped oscillation develops, conclusively indicating a magnetic transition in the single crystal samples (Fig 11). Below 1.5 K, the asymmetry could be fitted accordingly to $a_{0}P_{f}(t)=a_{0}\left[\frac{2}{3}\cos(\omega_{f}t+\phi)e^{-\frac{\sigma^{2}t^{2}}{2}}+\frac{e^{-\lambda_{f}t}}{3}\right]$ (8) which accounts for a magnetically frozen single phase with an average internal field at the muon sites $B_{int}=\omega_{f}/\gamma_{\mu}$. The damping parameter $\sigma$ encodes the width of the distribution of these internal fields. The last exponential term accounts for residual fluctuations in the frozen phase. At base temperature $T=0.28$ K, we obtain $B_{int}=8.6$ mT which is almost twice lower than the lowest internal fields previously reported in the closely related $x=0$ variant of YCu3(OH)6OxCl3-x [11, 31]. Although there is an uncertainty about the exact position of the muons, from this dipolar field value, we calculate a strongly reduced ordered magnetic moment for the Cu2+ $\sim 0.075$ $\mu_{B}$ using the majority $\mu$OH complex forming sites. Besides, the use in Eq. 8 of the coefficient $2/3$ and $1/3$ for the oscillatory and non-oscillatory components in our single crystal, expected for a spin-glass like state, suggests a rather disordered ground state. The strong damping $\sigma/\gamma_{\mu}=3.2$ mT compared to $B_{int}$ also points to a rather short range order and may simply reflect the absence of magnetic correlations in the $c$ direction, as shown by inelastic neutron scattering (see section VII). In order to fit the asymmetry on the whole temperature range, we combined both equations 6 and 8 in $a_{0}P(t)=a_{0}\left[f_{f}P_{f}(t)+(1-f_{f})P_{para}(t)e^{-\lambda_{p}t}\right]$ (9) with a switching parameter $f_{f}$ that tracks the frozen volume fraction from fully ordered for $f_{f}=1$ to fully paramagnetic for $f_{f}=0$. All the nuclear parameters are kept constant, so that the only varying parameters shown in Fig. 12 are the frozen fraction, the oscillation frequency and its damping rate, and the relaxation rate $\lambda_{p}$ or $\lambda_{f}$ for the paramagnetic or the frozen phase. The latter show a clear peak at 2.1 K as expected for the critical slowing down of the fluctuations at the magnetic transition. Surprisingly, there is a singular point at 1.56K in the relaxation rate plot but no significant changes in other variable parameters at the same temperature. Although this feature demands further confirmation, it is reminiscent of the spin dynamics re-entrance reported in clinoatacamite [32] and echoes an anomaly reported in the low temperature magnetization measurements [33]. Figure 12: Temperature evolution of the parameters used in Eq. 9 to fit the ZF $\mu$SR data: a) frequency $\omega_{f}$ reflecting the internal field magnitude, b) damping rate $\sigma$, c) fraction $f_{f}$ of the frozen phase and d) relaxation rates in the paramagnetic ($\lambda_{p}$) and frozen ($\lambda_{f}$) phases. To conclude, from our $\mu$SR measurements, Y-Kapellasite shows a completely frozen ground state in the case of single crystals in contrast to the fluctuating ground state observed in powder sample. The relaxation rate variation with temperature indicates a phase transition at $\sim$2.1K where the internal static fields set in. At 0.28 K, deep in the magnetic phase state, we estimate a strongly reduced frozen moment for the Cu2+ of about 0.075 $\mu_{B}$. ## VII Time-of-flight neutron scattering Figure 13: (a-c) Evolution in temperature of the dynamical structure factor $S(q,E)$ along the $[H,H,0]$ direction (integrated over $L=[-6,6]$) at $T=1.55$ K, $T=5$ K and $T=60$ K . (d) Brillouin zones boundaries (gray lines) with $K$ and $\Gamma$ points and extended Brillouin zone (black dashed line) in the reciprocal space. The red dotted line represents the $[H,H,0]$ direction. (e-g) Evolution in energy of $S(q,E)$ in the $(H,H,L)$ plane at $T=1.55$ K. (h) INS measurement integrated in energy over the elastic resolution showing a weak increase below 5 K at the $K^{\prime}_{1}$ and $K^{\prime\prime}$ positions, characteristic of the $(1/3,1/3)$ magnetic order. With large deuterated single crystals of Y3Cu9(OD)19Cl8 at hand, we also performed inelastic neutron scattering on this anisotropic kagome antiferromagnet in order to probe the magnetic excitations in reciprocal space and to identify the magnetic order [34]. Inelastic neutron scattering measurements were performed on the IN5 disk chopper time-of-flight spectrometer [35] at the Institut Laue-Langevin on a collection of single crystals with a total mass of 0.3g. These crystals were co-aligned in the $(H,H,L)$ horizontal plane and glued on a thin aluminium plate with the hydrogen-free CYTOP solution (CTL-809M). Measurements runs were performed at a constant temperature with an incident neutron wavelength $\lambda=4.8$ Å (incident energy E${}_{i}=3.55$ meV, giving a FWHM energy resolution of 0.08 meV) or $\lambda=2.6$ Å (E${}_{i}=12.1$ meV), while the sample stage was rotated around the direction perpendicular to the beam, to map out the reciprocal space. The detectors efficiency was corrected by a preliminary measurement on a vanadium can. Magnetic scattering could be identified only below 3 meV, an unexpected feature given the large antiferromagnetic Curie-Weiss temperature of $\theta_{\rm cw}=-100$ K [10]. We thus focused on measurements using our low incident energy of 3.55 meV and followed the temperature evolution of the dynamical neutron structure factor $S(q,E)$. Figure 13 shows the intensity contour color plots of $S(q,E)$ along the $[H,H,0]$ direction. At 60 K there is a large phonon contribution that substantially reduces at 20 K and below, where nearly vertical columns of intensity become apparent close to $K^{\prime}_{1}=(2/3,2/3,0)$ and $K^{\prime\prime}=(4/3,4/3,0)$, with a region of broad intensity located around 2 meV and close to $[1,1,0]$. At our base temperature of 1.5 K, sharp dispersive features emerge from these two $Q$ positions which we identify as spin-wave excitations. They show a pronounced two-dimensional character, as revealed by the presence of rods of scattering intensity along the $[0,0,L]$ direction (Fig. 13 e-g), and thus justify the 2D Heisenberg kagome antiferromagnet model applied below. From this time-of- flight measurement, one can isolate the elastic contribution by integrating over twice the energy resolution ($\sim 0.15$ meV) only. Figure 13 h) shows such contribution with the apparition of tiny $q$-resolution limited magnetic Bragg peaks below 5 K located at both $K^{\prime}_{1}$ and $K^{\prime\prime}$ positions. This elastic scattering points to a small ordered fraction of the Cu2+ moment, that could be easily missed in a conventional energy-integrated neutron diffraction experiment, and strongly supports a magnetic non-collinear coplanar phase with a $Q=(1/3,1/3)$ ordering wave vector. The inelastic scattering in the $(H,K,0)$ plane is displayed in Fig. 14. The energy-integrated cuts show clear intensities located at the $K$ positions of the Brillouin zone, mirroring the sharp spin wave dispersions already observed below 1 meV in Fig. 13 a). Above 1 meV, the inelastic scattering broadens significantly and extends toward the Brillouin zone center. Figure 14: Dynamical structure factor $S(Q,E)$ for the $(H,K,0)$ plane measured at $T=1.55$ K. The four color plot maps show its evolution in energy with the corresponding integrated energy range. Gray lines indicate the Brillouin zones boundaries. Bright spots of intensity are observed at the $K$ points, and corresponds to the dispersion of spin-waves. To better isolate the inelastic magnetic contribution present at 1.5 K, we subtracted a background constructed by the replication of the 5 K dataset, integrated in the range $H=[1.6,2]$, along $[H,H,0]$, where the phonon and magnetic contributions appear to be minimal. The result is shown in Fig. 15 a). The dispersive excitations rising from $K^{\prime}_{1}$ and $K^{\prime\prime}$ are clearly singled out, and then merge into a broad arch extending from 1 to 2.5 meV. Guided by the observation of an in-plane distortion of the kagome lattice, and by the recent ab initio DFT results [19], we performed linear spin-wave calculations using the SpinW package [36] of the Heisenberg model with only nearest-neighbors interactions in order to model the excitations on such an anisotropic kagome lattice: $\mathcal{H}=\sum_{\left\langle i,j\right\rangle}J_{ij}\bm{S_{i}}\cdot\bm{S_{j}}\\\ $ (10) where $J_{ij}$ can take three different values $J_{\hexagon}$, $J$ and $J^{\prime}$ depending respectively on the bonds $d_{1},d_{2}$ and $d_{3}$ as defined in Fig. 1. From DFT calculation, Hering et al. [19] confirmed the relevance of such spin models in the context of Y-Kapellasite using three antiferromagnetic couplings $J\simeq J_{\hexagon}\simeq 140$ K and $J^{\prime}\simeq 10$ K. The spin-wave simulations assumed an enlarged unit cell of nine sites, because of the two different copper sublattices, and a coplanar order with a $Q=(1/3,1/3)$ propagation vector, as described in details in Ref. [19]. The $J=J_{\hexagon}=J^{\prime}$ case leads to the well- known $\sqrt{3}\times\sqrt{3}$ phase of the classical kagome antiferromagnetic model which fails to reproduce our data when $J>15$ K, due to the maximum of the band observed around 2.5 meV. Given that $\theta_{\rm cw}=-100$ K imposes a higher exchange value, we discard the isotropic case and instead study the excitations for the parameters $J=J_{\hexagon}$ and $J^{\prime}\neq J$. From the evolution of the spin-wave spectrum with $J$ or $J^{\prime}$, we find our best model using $J\simeq J_{\hexagon}=140$ K and $J^{\prime}=63$ K. However, we also find that a single set of interactions can not account for the spread of intensity observed at $[1,1,0]$, from 1 to 2.5 meV. Instead, we consider a $\sim 10$ % distribution of $J^{\prime}$, tentatively attributed to some local disorder, and that could indeed satisfactorily fit our inelastic data, as shown in Fig. 15 c), d) and e), leading to $J\simeq J_{\hexagon}=140(10)$ K and $J^{\prime}=[56,70]$ K. Finally, we note that the intensity observed at low $Q$, near $Q=(1/3,1/3,0)$, and below 0.5 meV (see Fig. 15 a) is much broader than predicted by the simulation, and may hint at a more complex ground state in Y-kapellasite with the presence of short-range order and fluctuations. We note that this broad intensity extends mostly around the zone boundaries of the first Brillouin zone (Fig. 14) and not close to the boundaries of the _extended_ Brillouin zone, as predicted for the classical spin liquid [19]. In summary, INS gives evidence for elastic scattering below 5 K, in agreement with the static magnetism revealed by $\mu$SR and line broadening in NMR, consistent with a $Q=(1/3,1/3)$ magnetic ground state. Furthermore, we found a set of magnetic interactions in Y-kapellasite which is in very good agreement with the predictions from DFT and variational Monte-Carlo studies [19]. The main difference is our experimental value of $J^{\prime}\simeq 63(7)$ K as compared to the theoretical estimate $J^{\prime}=10.3(7)$ K. This locates Y-kapellasite in the $Q=(1/3,1/3)$ magnetic order of the phase diagram, but very close to the classical spin liquid phase reported in Ref. [19]. Figure 15: Comparison between experimental data (a,b) and spin-waves simulations (c,d) at $T=1.5$ K. The INS data in (a) measured at 1.5 K was subtracted by a 5 K dataset (see text) to highlight the magnetic contribution. (e) Energy-integrated cuts along $[H,H,0]$ of data shown in (a) and (c) for different energies $E$ with an integration range of 0.15 meV. Symbols are for data and lines for the simulation. They are both shifted vertically for clarity. (f) Classical phase diagram reproduced from [19]. The black square denotes the position found from DFT calculation [19] while the red star locates Y-kapellasite using the exchange couplings found in our work. ## VIII Discussion and Conclusion Our experimental investigation of Y3Cu9(OH)19Cl8 reveals an unusual bidimensional coplanar $Q=(1/3,1/3)$ long range ordered ground state, overall in very good agreement with the theoretical study in Ref. [19]. In particular, the proposed spin arrangement in the ground state is found to be compatible with the Cl NMR lineshape at low temperature. However, at variance with first principle calculations where the third interaction of the model $J^{\prime}$ was found to be negligible, we determine experimentally a finite $J^{\prime}\lesssim J\hexagon/2$. This finding of a sizeable antiferromagnetic $J^{\prime}$ seems nonetheless consistent with the Cu-OH-Cu bond angles which vary only by 4.6∘ among the three different bonds. The smallest angle $113.1^{\circ}$ corresponding to $J^{\prime}$ remains significantly larger than 105∘ which is known to yield small $\sim-15$ K ferromagnetic interaction in Zn-kapellasite [37]. Besides, to reproduce the broad magnetic excitations in INS, a distribution $\sim 10\%$ of the $J^{\prime}$ interaction was imposed. This suggests some kind of disorder. Due to the large difference in the ionic radii of Y3+ and Cu2+ we expect no significant intersite mixing at variance with Zn-kapellasite or other Zn and Cu based kagome materials and indeed we found no evidence of such disorder in the magnetic lattice above 33 K, neither in neutron diffraction nor in NMR which on the contrary exposes well resolved narrow lines. Therefore, the distribution of $J^{\prime}$ may arise only at low temperature, as a result of the structural instabilities observed at 33 K and 13 K. The slightly different local environment detected below 33 K by Cl NMR would then reflect in the moderate distribution of the interaction, while preserving on average the high temperature magnetic model, since the symmetry and lattice parameters remain unchanged on the whole temperature range, down to 65 mK. The experimentally determined interactions locate Y-kapellasite in the classical phase diagram close to the boundary between the (1/3,1/3) long range ordered and the ”classical spin liquid” phases (see Fig. 15 f). Although the effect of quantum fluctuations has not been investigated so close to the phase boundary, we may anticipate that they are responsible for the strongly reduced value of the Cu2+ moments $\lesssim 0.1\mu_{B}$, which could not be detected in standard neutron diffraction but yield static internal fields detected both in NMR and $\mu$SR. The proximity to a phase boundary may also help understanding the striking difference in the ground states of the present large single crystal samples and formerly studied polycrystalline ones [11]. Indeed in the latter, the moments were found to remain fluctuating in the ground state and also no structural transition was detected. We may assume that some additional disorder in the polycrystalline samples or slightly different interactions, due to the absence of structural transition, destabilize the fragile ordered state observed in large single crystals. We note that such a sensitivity to the sample crystallinity seems to be a common feature of anisotropic kagome compounds like volborthite [38, 17] or vesigneite [39, 40, 18]. The fragility of the ground state in all these compounds seems to originate from a subtle interplay between structure and magnetic frustration. Beyond the proposed anisotropic nearest neighbor model, and despite its apparent success to reproduce the physics in Y3Cu9(OH)19Cl8, one important perspective to complete our understanding of this material is to quantify and address the role of Dzyaloshinskii-Moriya interaction. Let us indeed recall that this anisotropy of the interaction was pointed out as the main ingredient driving long range order in the sister compound YCu3(OH)6Cl3 [41] with a perfect kagome lattice. Quantum fluctuations and especially their role at the phase boundaries and how they impact the ”jammed” classical spin liquid phase is another key perspective of this work. Finally we note that a Br counterpart [42] was recently synthesized, showing no sign of ordering, and likely allowing to span further the phase diagram of the anisotropic kagome model. ## Acknowledgements The authors acknowledge the support of the French Agence Nationale de la Recherche, under Grant No. ANR-18-CE30-0022 ”LINK”, and funding from Deutsche Forschungsgemeinschaft (DFG) through TRR 288-422213477 (project A03). We thank Samir Hammoud for carrying out the ICP and gas extraction experiments, and A. Razpopov, R. Valenti, H.O. Jeschke and J. Reuther for useful discussions and for sharing INS simulation data. ## References * Balents [2010] L. Balents, Nature 464, 199 (2010). * Chamorro _et al._ [2020] J. R. Chamorro, T. M. McQueen, and T. T. Tran, Chemical Reviews 121, 2898 (2020). * Mendels _et al._ [2007] P. Mendels, F. Bert, M. A. de Vries, A. Olariu, A. Harrison, F. Duc, J. C. Trombe, J. S. Lord, A. Amato, and C. Baines, Phys. Rev. Lett. 98, 077204 (2007). * Han _et al._ [2012] T.-H. Han, J. S. Helton, S. Chu, D. G. Nocera, J. A. Rodriguez-Rivera, C. Broholm, and Y. S. Lee, Nature 492, 406 (2012). * Tustain _et al._ [2020] K. Tustain, B. Ward-O’Brien, F. Bert, T. Han, H. Luetkens, T. Lancaster, B. M. Huddart, P. J. Baker, and L. Clark, npj Quantum Materials 5, 74 (2020). * Fu _et al._ [2021a] Y. Fu, M.-L. Lin, L. Wang, Q. Liu, L. Huang, W. Jiang, Z. Hao, C. Liu, H. Zhang, X. Shi, J. Zhang, J. Dai, D. Yu, F. Ye, P. A. Lee, P.-H. Tan, and J.-W. Mei, Nature Communications 2021 12:1 12, 1 (2021a). * Puphal _et al._ [2018] P. Puphal, K. M. Zoch, J. Désor, M. Bolte, and C. Krellner, Physical Review Materials 2, 063402 (2018). * Sun _et al._ [2016] W. Sun, Y.-X. Huang, S. Nokhrin, Y. Pan, and J.-X. Mi, Journal of Materials Chemistry C 4, 8772 (2016). * Fu _et al._ [2021b] Y. Fu, L. Huang, X. Zhou, J. Chen, X. Zhang, P. Chen, S. Wang, C. Liu, D. Yu, H.-F. Li, L. Wang, and J.-W. Mei, Chinese Physics B 30, 100601 (2021b). * Puphal _et al._ [2017] P. Puphal, M. Bolte, D. Sheptyakov, A. Pustogow, K. Kliemt, M. Dressel, M. Baenitz, and C. Krellner, Journal of Materials Chemistry C 5, 2629 (2017). * Barthélemy _et al._ [2019] Q. Barthélemy, P. Puphal, K. M. Zoch, C. Krellner, H. Luetkens, C. Baines, D. Sheptyakov, E. Kermarrec, P. Mendels, and F. Bert, Physical Review Materials 3, 074401 (2019). * Müller and Müller [1995] M. Müller and B. G. Müller, Zeitschrift für anorganische und allgemeine Chemie 621, 993 (1995). * Matan _et al._ [2019] K. Matan, T. Ono, G. Gitgeatpong, K. de Roos, P. Miao, S. Torii, T. Kamiyama, A. Miyata, A. Matsuo, K. Kindo, S. Takeyama, Y. Nambu, P. Piyawongwatthana, T. J. Sato, and H. Tanaka, Phys. Rev. B 99, 224404 (2019). * Downie _et al._ [2015] L. Downie, E. Ardashnikova, C. Tang, A. Vasiliev, P. Berdonosov, V. Dolgikh, M. de Vries, and P. Lightfoot, Crystals 5, 226 (2015). * Grbić _et al._ [2013] M. S. Grbić, S. Krämer, C. Berthier, F. Trousselet, O. Cépas, H. Tanaka, and M. Horvatić, Phys. Rev. Lett. 110, 247203 (2013). * Henderson _et al._ [2019] A. Henderson, L. Dong, S. Biswas, H. I. Revell, Y. Xin, R. Valenti, J. A. Schlueter, and T. Siegrist, Chemical Communications 55, 11587 (2019). * Ishikawa _et al._ [2015] H. Ishikawa, M. Yoshida, K. Nawa, M. Jeong, S. Krämer, M. Horvatić, C. Berthier, M. Takigawa, M. Akaki, A. Miyake, M. Tokunaga, K. Kindo, J. Yamaura, Y. Okamoto, and Z. Hiroi, Physical Review Letters 114, 227202 (2015). * Boldrin _et al._ [2016] D. Boldrin, K. Knight, and A. S. Wills, (2016), arXiv:1610.01436 [cond-mat.mtrl-sci] . * Hering _et al._ [2022] M. Hering, F. Ferrari, A. Razpopov, I. I. Mazin, R. Valentí, H. O. Jeschke, and J. Reuther, npj Computational Materials 8, 1 (2022). * Sun _et al._ [2021] W. Sun, T. Arh, M. Gomilšek, P. Koželj, S. Vrtnik, M. Herak, J.-X. Mi, and A. Zorko, Physical Review Materials 5, 064401 (2021). * Han _et al._ [2011] T. H. Han, J. S. Helton, S. Chu, A. Prodi, D. K. Singh, C. Mazzoli, P. Müller, D. G. Nocera, and Y. S. Lee, Physical Review B 83, 100402 (2011). * Biesner _et al._ [2022] T. Biesner, S. Roh, A. Razpopov, J. Willwater, S. Süllow, Y. Li, K. M. Zoch, M. Medarde, J. Nuss, D. Gorbunov, Y. Skourski, A. Pustogow, S. E. Brown, C. Krellner, R. Valentí, P. Puphal, and M. Dressel, Advanced Quantum Technologies 5, 2200023 (2022). * Note [1] See Supplemental Material at [URL] for the Crystallographic Information Files corresponding to the refined Y-Kapellasite Y3Cu9(OH)19Cl8 structure at 40 K, 20 K, 8 K and 0.065 K. * Cohen and Reif [1957] M. Cohen and F. Reif, in _Solid state physics_ , Vol. 5 (Elsevier, 1957) pp. 321–438. * Baugher _et al._ [1969] J. F. Baugher, P. C. Taylor, T. Oja, and P. J. Bray, The Journal of Chemical Physics 50, 4914 (1969). * Suter _et al._ [1998] A. Suter, M. Mali, J. Roos, and D. Brinkmann, Journal of Physics: Condensed Matter 10, 5977 (1998). * Choi _et al._ [2021] H. Choi, I. Vinograd, C. Chaffey, and N. Curro, Journal of Magnetic Resonance 331, 107050 (2021). * Bert [2014] F. Bert, École thématique de la Société Française de la Neutronique 13, 03001 (2014). * Blundell [1999] S. Blundell, Contemporary Physics 40, 175 (1999). * Hillier _et al._ [2022] A. D. Hillier, S. J. Blundell, I. McKenzie, I. Umegaki, L. Shu, J. A. Wright, T. Prokscha, F. Bert, K. Shimomura, A. Berlie, _et al._ , Nature Reviews Methods Primers 2, 1 (2022). * Zorko _et al._ [2019] A. Zorko, M. Pregelj, M. Klanjšek, M. Gomilšek, Z. Jagličić, J. S. Lord, J. A. T. Verezhak, T. Shang, W. Sun, and J.-X. Mi, Phys. Rev. B 99, 214441 (2019). * Zheng _et al._ [2005] X. G. Zheng, H. Kubozono, K. Nishiyama, W. Higemoto, T. Kawae, A. Koda, and C. N. Xu, Phys. Rev. Lett. 95, 057201 (2005). * [33] T. Biesner, S. Roh, A. Razpopov, J. Willwater, S. Süllow, Y. Li, K. M. Zoch, M. Medarde, J. Nuss, D. Gorbunov, Y. Skourski, A. Pustogow, S. E. Brown, C. Krellner, R. Valentí, P. Puphal, and M. Dressel, Advanced Quantum Technologies 5, 2200023. * Kermarrec _et al._ [2020] E. Kermarrec, P. Puphal, J. Willwater, Q. Barthélemy, S. Süllow, and J. Ollivier, Institut Laue-Langevin (ILL) (2020). * Ollivier and Mutka [2011] J. Ollivier and H. Mutka, Journal of the Physical Society of Japan 80, SB003 (2011). * [36] S. Toth and B. Lake, Journal of Physics: Condensed Matter 27, 166002. * Fåk _et al._ [2012] B. Fåk, E. Kermarrec, L. Messio, B. Bernu, C. Lhuillier, F. Bert, P. Mendels, B. Koteswararao, F. Bouquet, J. Ollivier, A. D. Hillier, A. Amato, R. H. Colman, and A. S. Wills, Phys. Rev. Lett. 109, 037208 (2012). * Bert _et al._ [2005] F. Bert, D. Bono, P. Mendels, F. Ladieu, F. Duc, J.-C. Trombe, and P. Millet, Phys. Rev. Lett. 95, 087203 (2005). * Quilliam _et al._ [2011] J. A. Quilliam, F. Bert, R. H. Colman, D. Boldrin, A. S. Wills, and P. Mendels, Phys. Rev. B 84, 180401 (2011). * Yoshida _et al._ [2013] M. Yoshida, Y. Okamoto, M. Takigawa, and Z. Hiroi, Journal of the Physical Society of Japan 82, 013702 (2013). * Arh _et al._ [2020] T. Arh, M. Gomilšek, P. Prelovšek, M. Pregelj, M. Klanjšek, A. Ozarowski, S. J. Clark, T. Lancaster, W. Sun, J.-X. Mi, and A. Zorko, Phys. Rev. Lett. 125, 027203 (2020). * Zeng _et al._ [2022] Z. Zeng, X. Ma, S. Wu, H.-F. Li, Z. Tao, X. Lu, X.-h. Chen, J.-X. Mi, S.-J. Song, G.-H. Cao, G. Che, K. Li, G. Li, H. Luo, Z. Y. Meng, and S. Li, Phys. Rev. B 105, L121109 (2022).
Email<EMAIL_ADDRESS> # Electronic transport in copper-graphene composites Kashi N. Subedi Department of Physics and Astronomy, Ohio University, Athens, OH 45701, USA Kishor Nepal Department of Physics and Astronomy, Ohio University, Athens, OH 45701, USA Chinonso Ugwumadu Department of Physics and Astronomy, Ohio University, Athens, OH 45701, USA Keerti Kappagantula Pacific Northwest National Laboratory, Richland, WA 99352, USA D. A. Drabold [ Department of Physics and Astronomy, Ohio University, Athens, OH 45701, USA ###### Abstract We investigate the electronic transport properties of copper-graphene composites using a density-functional framework. Conduction in composites by varying the interface distance of a copper/graphene/copper (Cu/G/Cu) interface models was studied. The electronic density of states reveals increasing contributions from both copper and carbon atoms near the Fermi level with decreasing Cu-G interfacial distance. Electronic conductivity of the models computed using the Kubo-Greenwood formula showed the conductivity increases with decreasing Cu-G distance. We also find that the conductivity saturates below a threshold Cu-G distance. By computing the space-projected conductivity of the Cu/G/Cu models, we show that the graphene forms a bridge to the electronic conduction at small copper-graphene distances, thereby enhancing the conductivity. Metal composites are key industrial materials that may be manufactured using various novel techniques Heringhaus and Raabe (1996); Bouaziz, Kim, and Estrin (2013); Bruschi _et al._ (2021), leading to metastable materials exhibiting desirable mechanical, thermal, electronic, and transport properties without the constraints typically seen in alloys. Copper is used extensively for energy transport and electromagnetic applications such as motors, generators, cable, busbars and transformers. In the last two decades, we have witnessed copper composites with improved transport properties Singh and Gautam (2018), devised as a means to reduce energy loss in applications. One material that has demonstrated improved electrical conductivity and current density is a copper-graphene (Cu-G) composite. Various Cu-G composites ranging from a simple interface laminated to bulk Cu with interconnected graphene flakes have been reported in the literature, with some showing enhanced conductivity Weiping _et al._ (2015); Rho _et al._ (2017); An _et al._ (2019); Yang _et al._ (2021), and more commonly enhancement in mechanical performance Hwang _et al._ (2013); An _et al._ (2019). Conventional wisdom suggests that the introduction of additives to metal would increase scattering that would lead to increased electrical resistivity Ellis _et al._ (1993). This is the reason why addition of silver to copper leads to reduced electrical conductivity of Cu, even though Ag is 8% higher in conductivity than Cu Felicia, Rochiem, and Laia (2018). The additives, on the other hand, also arrest dislocation movement in the metals due to several mechanisms such as precipitate hardening, solid solution strengthening and dispersion strengthening, resulting in enhanced mechanical performance Kuhlmann-Wilsdorf (1989). At this juncture, it is known that the electrical conductivity of the Cu-G composites are dependent on the quantity of graphene used in the composites, the type of graphene and their defect density, and the arrangement of Cu matrix and graphene in the composite microstructure, which eventually dictates the carrier transport in the material during conduction Kappagantula _et al._ (2022); Pan _et al._ (2022). Processing routes adopted to manufacture the composites are also known to influence electrical performance. Processes that can control atomic scale deposition of Cu and graphene, such as vapor deposition and molecular beam epitaxy, have shown promise in manufacturing nano-to-micron scale samples demonstrating enhanced electrical performance Zheng _et al._ (2018). At the bulk scale, solid phase processing techniques have shown promise in enhancing electrical conductivity Kappagantula _et al._ (2022). In all this, however, while there is an evolving understanding of processing approaches and material chemistries that result in Cu-G composites with improved electrical performance, little is known about the nature of charge- carrier transfer between Cu and graphene. Several papers discuss carrier mobility in graphene that is deposited on Cu foils Orofeo _et al._ (2012); Banszerus _et al._ (2015). However, limited experimental and computational research is published on transport phenomena across metal/graphene interfaces. There is minimal understanding on how the nature of the interface (such as atom arrangement, interfacial distance) affects the electron density at the interface. In Cu-G composites demonstrating enhanced conductivity compared to the corresponding Cu substrates, two causes were hypothesized for the enhancement in conductivity. In one case, where the manufacturing conditions allow for it, the Cu grains are surmised to be templated on the adjacent graphene flakes leading to the formation of predominantly 111 textures in the composite during the manufacturing process. Such crystallographic orientations in Cu show higher conductivity compared to other textures such as 200 or 110. If there is a large concentration of 111 grains in the Cu microstructure, it is reasonable to expect an enhanced conductivity overall. In this hypothesis, graphene is thought to only template the grains but not actually participate in the conduction process. In the second case, graphene is assumed to participate during conduction through the exchange of carriers with the surrounding Cu matrix. One key consideration for this hypothesis is the nature of interface between the Cu and graphene that may be suitable for carrier transport across the interface. Some papers indicate that the 111 Cu grains offer the least lattice mismatch with graphene, which may also help with carrier transport. However, it is currently not well known how much conductivity the introduction of graphene brings about for Cu per this hypothesis. In this Letter, we explore the electrical transport properties of the Cu-G composites as a function of the interfacial distance between Cu and graphene, a parameter that can be modulated during Cu-G manufacturing and a simple proxy for a possible strained local configuration. We quantify the electronic transport by computing the electronic conductivity using the Kubo–Greenwood formula (KGF) Kubo (1957); Greenwood (1958), ideally suited for density- functional simulations of materials with its single particle Kohn–Sham orbitals and energies Martin (2008). The KGF method has been utilized in tight-binding and DFT computations of the conductivity of liquids and solids Allen and Broughton (1987); Galli _et al._ (1989). Besides quantifying the conductivity, we determine the conduction-active sites in the Cu-G composites by projecting the KGF conductivity onto real space using the space-projected conductivity (SPC), described elsewhere Prasai _et al._ (2018). The method has been implemented to study conduction processes in crystalline systems with defects, amorphous and semi-conducting systems at atomistic level Subedi _et al._ (2022); Thapa _et al._ (2022); Subedi _et al._ (2019); Subedi, Prasai, and Drabold (2020). We simulated the Cu-G composite by constructing an interface model with a geometry of the form copper/graphene/copper (Cu/G/Cu) as shown in Figure 1a. We adopted one of the low-indexed surfaces, namely 111, to construct the Cu surface. We placed the graphene sheet with one sub-lattice site (A) above the top of the first layer of Cu 111 and the other sub-lattice site (B) placed above the third layer of Cu 111 as shown in Figure 1b. This arrangement is also known as top-fcc, and has shown to represent the low-energy structure for Cu-G surface models Yang _et al._ (2017). We simulated the external pressure on Cu-G composites by varying the copper-graphene distance ($d_{Cu-G}$) of the Cu/G/Cu interface model as shown in Figure 1a. Besides Cu/G/Cu models, we also constructed an orthorhombic supercell of 216 atoms and simulated the external pressure by reducing the vertical dimension of the supercell. A similar approach has been used to simulate copper and aluminum under external pressure in earlier work Lanzillo _et al._ (2014). Figure 1: (a) Geometry of copper/graphene/copper (Cu/G/Cu) model with dCu-G representing copper-graphene distance. b) Top-fcc configuration of graphene above the Cu 111 surface with A and B representing sub-lattice sites above the Cu atom at first and third layer respectively. The pink colored spheres represent Cu atoms and brown colored spheres represent C atoms. Density-functional calculations were carried out using the Vienna Ab Initio Simulation Package Kresse and Hafner (1993) (VASP) code. For atomic structure relaxations of Cu/G/Cu models, we used a plane-wave basis set with a kinetic energy cutoff of 400 eV. For static calculations, we used a larger cutoff of 520 eV. Projected augmented wave (PAW) Blöchl (1994) potentials was used to account for ion-election interactions, and the generalized gradient approximation of Perdew-Burke-Ernzerhof (PBE) Perdew, Burke, and Ernzerhof (1997) as the exchange-correlation functional. All calculations were performed at 4 irreducible k-points, and periodic boundary conditions were implemented throughout. To determine the partial occupancy of electrons near the Fermi level, Fermi-Dirac distribution function with a smearing temperature of 1000 K was considered. $\delta$ function in the expression of KGF by was approximated by a Gaussian distribution function of width 0.05 eV. Factors like thermal broadening, k-point sampling, and finite size effects discussed in references Belmonte _et al._ (2015); Knyazev and levashov (2013); Calder’m, Karaseiv, and Trickey (2017); Bulanchuk (2021) play a role in employing the KGF. To obtain representative low-energy configurations, atomic structure relaxation of the Cu/G/Cu models were performed. The initial and final Cu-G distances after relaxation are tabulated in the first and the second column of table 1 respectively. It is apparent that the extent of repulsion of Cu atoms away from the graphene is higher for models with short Cu-G distances, indicated by larger shift in Cu-G layers (third column of table 1). For long Cu-G distances (up to $d_{Cu-G}$ = 2.61 Å ), most of the Cu-Cu bonds are virtually unaffected and only few Cu-Cu bonds are formed up to $\approx$2.52 Å, whereas for short Cu-G distances, the vertical Cu-Cu bonds are also formed at $\approx$2.49 and $\approx$2.46 Å. Table 1: Variation of Cu-G distance for different Cu/G/Cu models after atomic structure relaxation and corresponding transverse component of Stress Tensor. $d_{Cu-G}$111$d_{Cu-G}$ is copper-graphene distance shown in Figure 1a) (Å) | $d_{Cu-G}$ (Å) | Shift (Å) | Pz222Transverse component of Stress Tensor [GPa] ---|---|---|--- (Initial) | (Relaxed) | | 2.88 | 2.94 | 0.06 | 2.7 2.71 | 2.78 | 0.07 | 3.2 2.54 | 2.61 | 0.07 | 3.1 2.37 | 2.46 | 0.09 | 4.2 2.20 | 2.37 | 0.17 | 6.5 2.02 | 2.29 | 0.27 | 10.3 1.86 | 2.23 | 0.37 | 14.9 1.69 | 2.19 | 0.50 | 20.2 KGF conductivity for the Cu/G/Cu models in the direction normal to the graphene layer (i.e., along c-axis in Figure 1a) was computed. Figure 2a displays conductivities for the models with different Cu-G distances normalized by the KGF conductivity of copper crystal computed at 300 K corresponding to $d_{Cu-Cu}$ = 2.56 Å. From Figure 2a, we see that the conductivity increases with decreasing Cu-G distance in an almost exponential manner up to $d_{Cu-G}$ = 2.37 Å. Beyond 2.37 Å, we find a sharp rise in conductivity followed by a saturation below $\approx$2.23 Å. The increase in conductivity may be attributed to increased constructive overlapping of copper and graphene orbitals with decreasing Cu-G distance. For the short Cu-G distances with $d_{Cu-G}$ = 2.23 Å and beyond, the conductivity is obtained to be $\approx$10 % higher compared to that of copper computed at 300 K. Figure 2: a) KGF conductivities for relaxed Cu/G/Cu models with different Cu-G distance normalized by the KGF conductivity ($\sigma_{0}$) of a copper crystal computed at 300K with $d_{Cu-Cu}$ = 2.56 Å. b) Relative conductivity ($\sigma/\sigma_{0}$) of orthorhombic Cu model computed as a function of vertical Cu-Cu bond length. Figure 3: Electronic density of states for different Cu/G/Cu models. Top subplot corresponds to the total EDOS and the bottom subplot corresponds to the projected density of states (PDOS) onto carbon atoms. The compression of models resulted in an enhanced EDOS near the Fermi level, see inset for details. Figure 4: Space-projected conductivity plotted as a RGB colormap along a plane normal to the graphene sheet for different Cu/G/Cu models corresponding to $d_{Cu-G}$ = 3.25 Å, 2.94 Å, 2.29 Å and 2.23 Å respectively. The colorbar on the left represents the magnitude of SPC values that increases from red towards blue. The SPC values are scaled from the minimum value for $d_{Cu-G}$ = 3.25 Å. Cu and C atoms are represented by pink and brown-colored spheres respectively. To understand such an increasing trend in the conductivity with decreasing Cu-G distance, total electronic density of states (EDOS) and the projected density of states (PDOS) for selected models were computed and are shown in Figure 3. From Figure 3, we find relatively higher EDOS near the Fermi level with decreasing Cu-G distance. Such increase in the EDOS is contributed by both copper and carbon atoms of the Cu-G composites, which are more pronounced for short Cu-G distances. According to Mott and Davis Mott and Davis (1979), the DC conductivity, $\sigma_{dc}\propto(N(\epsilon_{f}))^{2}$, where $N$ represents the density of states. So, one expects increasing conductivity with decreasing Cu-G distance for these composites as a combined effect of enhanced EDOS near the Fermi level. We note that it was not at all obvious that the graphene would enhance the DOS at the Fermi level, and we conjecture that this is what underlies experiments showing improved conductivity of the composites. Further exploration was made by investigating a spatial description of conductivity as a function of Cu-G distance. We computed the SPC for selected Cu/G/Cu models and are shown in Figure 4. Figure 4 displays the SPC projected on 100 plane as a RGB colormap for four such models with different Cu-G distance. The colorbar on the left depicts the magnitude of SPC values, where red color represents low SPC values and blue color represents high SPC values. From Figure 4, it is apparent that the contribution from both Cu and graphene to the conduction increases with decreasing Cu-G distance, and is more significant at short distances. The SPC plots at short Cu-G distances show that the graphene directly participates in conduction and forms a bridge for conduction between Cu atoms on opposite layers. In addition to the graphene, we also see significant contributions from Cu atoms to conduction at short Cu-G distances. This contribution from Cu atoms to conductivity can be associated with different Cu bonding environments in the Cu-G composites. KGF conductivity for the orthorhombic Cu model for different vertical Cu-Cu distances ranging from 2.43 to 2.55 Åwere calculated. Figure 2b shows the relative conductivity of Cu models for different Cu-Cu vertical distances. One observes that the conductivity of Cu increases with reducing the Cu-Cu distance from $\approx$2.55 to $\approx$2.46 Å after which it drops at $\approx$2.44 Å. We observed that for short Cu-G distances ($d_{Cu-G}\leq$ 2.23 Å), the Cu-Cu bond-lengths are $\approx$2.49 and $\approx$2.46 Å. So, the contribution to the conductivity from the Cu atoms with nearest neighbors close to 2.46 and 2.49 Å is higher compared with Cu atoms at other distances and, in agreement with SPC plots in Figure 4. Saturation in conductivity of Cu/G/Cu model for compression below $d_{Cu-G}$ = 2.23 Å (refer Figure 2a) were observed. Such a characteristic is attributed to the contribution of the Cu atoms forming different nearest-neighbor distances. At $d_{Cu-G}$ = 2.23 Å, the repulsion of Cu layers on both sides of graphene leads to formation of Cu-Cu bonds at $\approx$2.49 and $\approx$2.46 Å. With further reduction of Cu-G distance, here $d_{Cu-G}$ = 2.19 Å, the Cu-Cu bonds are formed at $\approx$ 2.46 Å and also at $\approx$2.44 Å. So, the additional disorder in Cu leads to slightly less contribution from the Cu atoms to the conduction (refer Figure 2b) even as the graphene still contributes. In conclusion, copper-graphene composites were modeled by considering the Cu/G/Cu interface model with varying Cu-G distance. We showed that the DC conductivity of the Cu/G/Cu interface models increases with decreasing copper- graphene distance, and such a increase is a combined contributions from both Cu atoms and the graphene sheet. We attribute this to an enhanced EDOS near the Fermi-level. We also provided the spatial view of the electronic conductivity by computing the SPC for varying copper-graphene distance. The SPC calculations showed that graphene forms the bridge to electronic conduction between Cu atoms that lie on the opposite layers. We also showed an interesting characteristic of saturation of the conductivity at a short copper-graphene distance (below $d_{Cu-G}$ = 2.23 Å), and such a characteristic is mostly attributed to slightly reduced contribution to the conduction from Cu atoms that form Cu-Cu bonds at $\approx$2.44 Å even when the graphene showing enhanced contribution. This paper provides one of the first demonstrations of graphene participating in the conduction process along with the copper atoms in the composite, especially at low copper-graphene distances. An implication of this finding is that manufacturing process conditions may influence the electrical performance metal-graphene composites by engineering the interfaces to encourage graphene’s participation in the conduction processes, rather than leaving it as a scattering site. It is important to note that in addition to the interfacial distance, the arrangement of copper-carbon atoms (geometry) can also influence electrical performance of the composite; however, those effects were not explored in this work. Additional factors of importance, such as the number of graphene layers, their dimensions and the defect density of the graphene additives will be explored in future work. Acknowledgements: We acknowledge Extreme Science and Engineering Discovery Environment (XSEDE), supported by NSF Grant No. ACI-1548562 at the Pittsburgh Supercomputing Center, for providing computational resources under allocation DMR-190008P,Department of Energy Advanced Manufacturing Office, and Pacific Northwest National Laboratory operated by the Battelle Memorial Institute for the U.S. Department of Energy under Contract No. DE-AC06-76LO1830. ## References * Heringhaus and Raabe (1996) F. Heringhaus and D. Raabe, “Recent advances in the manufacturing of copper-base composites,” Journal of Materials Processing Technology 59, 367–372 (1996). * Bouaziz, Kim, and Estrin (2013) O. Bouaziz, H. S. Kim, and Y. Estrin, “Architecturing of metal-based composites with concurrent nanostructuring: A new paradigm of materials design,” Advanced Engineering Materials 15, 336–340 (2013). * Bruschi _et al._ (2021) S. Bruschi, J. Cao, M. Merklein, and J. Yanagimoto, “Forming of metal-based composite parts,” CIRP Annals 70, 567–588 (2021). * Singh and Gautam (2018) M. K. Singh and R. K. Gautam, “Mechanical and tribological properties of plastically deformed copper metal matrix nano composite,” Materials Today: Proceedings 5, 5727–5736 (2018). * Weiping _et al._ (2015) L. Weiping, L. Delong, F. Qiang, and P. Chunxu, “Conductive enhancement of copper/graphene composites based on a high-quality graphene,” RSC Advances 5, 80428–80433 (2015). * Rho _et al._ (2017) H. Rho, Y. S. Jang, S. Kim, S. Bae, T.-W. Kim, D. S. Lee, J.-S. Ha, and S. H. Lee, “Porous copper–graphene heterostructures for cooling of electronic devices,” Nanoscale 9, 7565–7569 (2017). * An _et al._ (2019) Z. An, J. Li, A. Kikuchi, Z. Wang, Y. Jiang, and T. Ono, “Mechanically strengthened graphene-cu composite with reduced thermal expansion towards interconnect applications,” Microsystems & Nanoengineering 5, 20 (2019). * Yang _et al._ (2021) J. Yang, Y. He, X. Zhang, W. Yang, Y. Li, X. Li, Q. Chen, X. Chen, K. Du, and Y. Yan, “Improving the electrical conductivity of copper/graphene composites by reducing the interfacial impurities using spark plasma sintering diffusion bonding,” Journal of Materials Research and Technology 15, 3005–3015 (2021). * Hwang _et al._ (2013) J. Hwang, T. Yoon, S. H. Jin, J. Lee, T.-S. Kim, S. H. Hong, and S. Jeon, “Enhanced mechanical properties of graphene/copper nanocomposites using a molecular-level mixing process,” Advanced Materials 25, 6724–6729 (2013). * Ellis _et al._ (1993) T. Ellis, I. Anderson, H. Downing, and J. Verhoeven, “Deformation-processed wire prepared from gas-atomized cu-nb alloy powders,” Metallurgical Transactions A 24, 21–26 (1993). * Felicia, Rochiem, and Laia (2018) D. M. Felicia, R. Rochiem, and S. M. Laia, “The effect of silver (ag) addition to mechanical and electrical properties of copper alloy (cu) casting product,” AIP Conference Proceedings 1945, 020075 (2018). * Kuhlmann-Wilsdorf (1989) D. Kuhlmann-Wilsdorf, “Theory of plastic deformation: - properties of low energy dislocation structures,” Materials Science and Engineering: A 113, 1–41 (1989). * Kappagantula _et al._ (2022) K. S. Kappagantula, J. A. Smith, A. K. Nittala, and F. F. Kraft, “Macro copper-graphene composites with enhanced electrical conductivity,” Journal of Alloys and Compounds 894, 162477 (2022). * Pan _et al._ (2022) C. Pan, A. P. S. Gaur, M. Lynn, M. P. Olson, G. Ouyang, and J. Cui, “Enhanced electrical conductivity in graphene–copper multilayer composite,” AIP Advances 12, 015310 (2022). * Zheng _et al._ (2018) L. Zheng, H. Zheng, D. Huo, F. Wu, L. Shao, P. Zheng, Y. Jiang, X. Zheng, X. Qiu, Y. Liu, _et al._ , “N-doped graphene-based copper nanocomposite with ultralow electrical resistivity and high thermal conductivity,” Scientific reports 8, 1–7 (2018). * Orofeo _et al._ (2012) C. M. Orofeo, H. Hibino, K. Kawahara, Y. Ogawa, M. Tsuji, K.-i. Ikeda, S. Mizuno, and H. Ago, “Influence of cu metal on the domain structure and carrier mobility in single-layer graphene,” Carbon 50, 2189–2196 (2012). * Banszerus _et al._ (2015) L. Banszerus, M. Schmitz, S. Engels, J. Dauber, M. Oellers, F. Haupt, K. Watanabe, T. Taniguchi, B. Beschoten, and C. Stampfer, “Ultrahigh-mobility graphene devices from chemical vapor deposition on reusable copper,” Science advances 1, e1500222 (2015). * Kubo (1957) R. Kubo, “Statistical-mechanical theory of irreversible processes. i. general theory and simple applications to magnetic and conduction problems,” J.Phys.Soc.Jpn 12, 570–586 (1957). * Greenwood (1958) D. A. Greenwood, “The boltzmann equation in the theory of electrical conduction in metals,” Proc.Phys.Soc. 71, 585–596 (1958). * Martin (2008) R. M. Martin, Electronic Structure, Cambridge University Press, Cambridge (2008). * Allen and Broughton (1987) P. B. Allen and J. Q. Broughton, J. Phys. Chem. 91, 4964 (1987). * Galli _et al._ (1989) G. Galli, R. M. Martin, R. Car, and M. Parrinello, Phys. Rev. Lett. 63, 988 (1989). * Prasai _et al._ (2018) K. Prasai, K. N. Subedi, K. Ferris, P. Biswas, and D. A. Drabold, “Spatial projection of electronic conductivity: The example of conducting bridge memory materials,” Phys Status solidi:Rapid Res.Lett. 12 (2018). * Subedi _et al._ (2022) K. N. Subedi, K. Kappagantula, F. Kraft, A. Nittala, and D. A. Drabold, “Electrical conduction processes in aluminum: Defects and phonons,” Phys. Rev. B 105, 104114 (2022). * Thapa _et al._ (2022) R. Thapa, C. Ugwumadu, K. Nepal, J. Trembly, and D. A. Drabold, “Ab initio simulation of amorphous graphite,” Phys. Rev. Lett. 128, 236402 (2022). * Subedi _et al._ (2019) K. N. Subedi, K. Prasai, M. N. Kozicki, and D. A. Drabold, “Structural origins of electronic conduction in amorphous copper-doped alumina,” Phys. Rev. Materials 3, 065605 (2019). * Subedi, Prasai, and Drabold (2020) K. N. Subedi, K. Prasai, and D. A. Drabold, Phys. Status Solidi B 258, 2000438 (2020). * Yang _et al._ (2017) T. Yang, L. Yang, H. Liu, H. Zhou, S. Peng, X. Zhou, F. Gao, and X. Zu, “Ab initio study of stability and migration of point defects in copper-graphene layered composite,” Journal of Alloys and Compounds 692, 49–58 (2017). * Lanzillo _et al._ (2014) N. A. Lanzillo, J. B. Thomas, B. Watson, M. Washington, and S. K. Nayak, “Pressure-enabled phonon engineering in metals,” Proceedings of the National Academy of Sciences 111, 8712–8716 (2014). * Kresse and Hafner (1993) G. Kresse and J. Hafner, “Ab initio molecular dynamics for liquid metals,” Phys. Rev. B 47, 558–561 (1993). * Blöchl (1994) P. E. Blöchl, “Projector augmented-wave method,” Phys. Rev. B 50, 17953–17979 (1994). * Perdew, Burke, and Ernzerhof (1997) J. P. Perdew, K. Burke, and M. Ernzerhof, “Generalized gradient approximation made simple [phys. rev. lett. 77, 3865 (1996)],” Phys. Rev. Lett. 78, 1396–1396 (1997). * Belmonte _et al._ (2015) A. Belmonte, U. Celano, A. Redolfi, A. Fantini, R. Muller, W. Vandervorst, M. Houssa, M. Jurczak, and L. Goux, IEEE Trans.Elec.Devices 62, 4349 (2015). * Knyazev and levashov (2013) D. Knyazev and P. levashov, Computational Materials Science 79, 817 (2013). * Calder’m, Karaseiv, and Trickey (2017) L. Calder’m, V. Karaseiv, and S. Trickey, Computer Physics Communications 221, 118 (2017). * Bulanchuk (2021) P. Bulanchuk, Computational Materials Science 261, 107714 (2021). * Mott and Davis (1979) N. Mott and E. Davis, “Electronic processes in non-crystalline materials, 2nd edn. clarendon,” Oxford (1979).
# The structure of the density-potential mapping Part I: Standard density-functional theory Markus Penz Basic Research Community for Physics, Innsbruck, Austria Erik I. Tellgren Hylleraas Centre for Quantum Molecular Sciences, Department of Chemistry, University of Oslo, Norway Mihály A. Csirik Hylleraas Centre for Quantum Molecular Sciences, Department of Chemistry, University of Oslo, Norway Department of Computer Science, Oslo Metropolitan University, Norway Michael Ruggenthaler Max Planck Institute for the Structure and Dynamics of Matter, Hamburg, Germany Andre Laestadius<EMAIL_ADDRESS>Department of Computer Science, Oslo Metropolitan University, Norway Hylleraas Centre for Quantum Molecular Sciences, Department of Chemistry, University of Oslo, Norway ###### Abstract The Hohenberg–Kohn theorem of density-functional theory (DFT) is broadly considered the conceptual basis for a full characterization of an electronic system in its ground state by just the one-body particle density. Part I of this review aims at clarifying the status of the Hohenberg–Kohn theorem within DFT and Part II at different extensions of the theory that include magnetic fields. We collect evidence that the Hohenberg–Kohn theorem does not so much form the basis of DFT, but is rather the consequence of a more comprehensive mathematical framework. Such results are especially useful when it comes to the construction of generalized DFTs. ###### Contents 1. I Introduction 2. II Preliminaries 3. III Representability of densities 4. IV The Hohenberg–Kohn theorem 5. V The unique-continuation property 6. VI Hierarchy of density functionals 7. VII Density-potential mappings from differentials 8. VIII Linking to a reference system: the Kohn–Sham scheme 9. IX Density-potential mixing and regularized DFT 10. X Abstract density-potential mapping 11. XI Summary 12. XII Outlook ## I Introduction The theorem of Hohenberg and Kohn [1] (HK) is usually presented as the theoretical justification of density-functional theory (DFT). It states that the one-body particle density uniquely (up to an additive constant) determines the scalar potential of a non-relativistic many-electron system in its ground state. The Mathematical analysis of ground-state DFT was pioneered by Lieb [2], using tools from convex analysis. In it, some important problems, especially in relation with differentiability of the involved functionals that map densities to energies, were left unanswered and remained as open questions. Lammert [3] then demonstrated that the key functional of DFT is indeed non-differentiable, but it remained unclear to what extent this threatens the foundations of DFT and its algorithmic realization, the Kohn–Sham scheme employed for practical calculations. Regularization as a means to overcome non-differentiability has been applied to DFT [4] (Section IX) and its extension, current DFT (CDFT) [5; 6]. The existence of functional derivatives through regularization also avoids the problem of $v$-representability that usually haunts DFT, i.e., that not every reasonable density is the solution to a certain potential (Section III). A central result in this work is a very convenient and novel formulation of the HK theorem that restructures it into two sub-theorems, HK1 and HK2 (Section IV): * • (HK1) If two potentials share a common ground-state density then they also share a common ground-state wave function or density matrix. * • (HK2) If two potentials share any common eigenstate and if that eigenstate is non-zero almost everywhere (a property that is guaranteed if the the unique- continuation property (UCP) holds; see Section V) then they are equal up to a constant. Combining HK1 and HK2, one obtains the classical HK theorem and with it a well-defined density-potential mapping. The proof of HK1 will be shown to be immediate from just the formulation of “ground-state energy”. Consequently, it is also easily attainable in an abstract or extended formulation of DFT (Section X). The situation for HK2, on the other hand, is more complicated but, as will be demonstrated, it holds true with certain restrictions in the standard DFT setting. It is known _not_ to hold in paramagnetic CDFT [7] and has, to the best of our knowledge, an unknown status in total CDFT. In Part II of this review, we will exemplify how different DFTs follow this structure and, maybe more importantly, pinpoint why this route might fail. After analyzing its basic structure, the status of the HK theorem within DFT is scrutinized. If only the ground state of a system is the matter of interest, a constrained-search approach seems to be sufficient for the formulation of DFT, and the usual type of constrained-search functional [8; 9] even _implicitly includes_ HK1 (Section VI). Besides being a mathematically more transparent formulation than the HK theorem, the constrained-search formalism is also a better starting point for deriving approximate density functionals. Nonetheless, the full HK theorem remains important for going beyond the bare minimum needed to set up a ground-state theory. For example, the HK theorem implies that the ground-state density determines not only the ground state but, by fixing the scalar potential, also all excited states. This becomes relevant when thermostatistical properties are considered. Furthermore, in order to be able to define the Kohn–Sham scheme (Section VIII), one actually demands more than just the HK result, relying on differentiability of the energy functional that in turn would imply the _whole_ HK result (Section VII). Consequently, in a (Moreau–Yosida) regularized setting, the Kohn–Sham scheme can be rigorously formulated and even proven to converge in finitely many dimensions [10; 11; 12], and HK becomes just a by-product. Although we will do our best to orient the reader within the rich subject that is DFT, the scope of this review is limited. We will mainly focus on, in our opinion, matters closely related to the HK mapping and properties of the exact functional(s). Many excellent reviews and textbooks are available on the subject [13; 14; 15; 16; 17; 18]. For the interested reader, we also point out the very recent article based on a round-table discussion [19]. ## II Preliminaries Density-functional theory is an approach to describe particles that obey the laws of quantum mechanics, but that avoids their full description by a wave function and instead switches to reduced quantities like the one-particle density. In its basic form discussed here, the focus is solely on the ground- state properties of the quantum system. For the configuration space of a single particle we always choose $\mathbb{R}^{3}$ with the additional spin degree-of-freedom for spin-$\frac{1}{2}$ particles. The Hamiltonian comprises three parts, $H[v]=T+W+V[v],$ relating to the kinetic energy, the Coulomb repulsion, and the external scalar potential, respectively. The internal parts will be collected as $H_{0}=T+W$. The kinetic-energy operator is $T=-\frac{1}{2}\sum_{i=1}^{N}\nabla_{i}^{2}$ in standard DFT, where atomic units are employed. Notation-wise, we use small letters for one-body objects. The external potential contribution $V[v]$ is always defined from a one-body potential $v(\mathbf{r})$ and is of an additive form, $V[v](\underline{\mathbf{r}})=\sum_{i=1}^{N}v(\mathbf{r}_{i}),$ where $\underline{\mathbf{r}}=(\mathbf{r}_{1},\dots,\mathbf{r}_{N})$. For later reference we also define $\underline{\sigma}=(\sigma_{1},\dots,\sigma_{N})$ for the spin degrees-of- freedom. The full quantum-mechanical description of a system in its ground state is achieved by determining the eigenstate $\psi_{0}$ of $H[v]$ that has the correct symmetry and the lowest eigenvalue $E_{0}$ (ground-state energy), $H[v]\psi_{0}=E_{0}\psi_{0}.$ (1) If such a lowest eigenstate is not unique, we speak of _degeneracy_ , a case that will often appear in the discussion below and that leads to several complicacies. Then a valid ground state can also be given as a statistical mixture of the pure ground states $\psi_{k}$ in the form of a density matrix $\Gamma=\sum_{k}\lambda_{k}|\psi_{k}\rangle\langle\psi_{k}|$ with $\lambda_{k}\in[0,1]$ and $\sum_{k}\lambda_{k}=1$. It is natural to require states of finite kinetic energy, $\langle\psi|T|\psi\rangle=\frac{1}{2}\sum_{i=1}^{N}\sum_{\underline{\sigma}}\int_{\mathbb{R}^{3N}}|\nabla_{i}\psi|^{2}\,\mathrm{d}\underline{\mathbf{r}}<+\infty,$ and we define the basic set for wave functions $\mathcal{W}=\\{\psi\mid\,\psi\;\text{anti- symmetric},\,\langle\psi|T|\psi\rangle<+\infty\\}.$ In cases where density matrices $\Gamma$ are considered, we require $\psi_{k}\in\mathcal{W}$ for all their components. The one-particle density of a given $\psi$ as the basic variable of standard DFT is $\rho_{\psi}(\mathbf{r}_{1})=N\sum_{\underline{\sigma}}\int_{\mathbb{R}^{3(N-1)}}|\psi|^{2}\,\mathrm{d}\mathbf{r}_{\perp},$ (2) where we used the shorthand notation $\mathbf{r}_{\perp}=(\mathbf{r}_{2},\dots,\mathbf{r}_{N})$, and it is $\rho_{\Gamma}(\mathbf{r})=\sum_{k}\lambda_{k}\rho_{\psi_{k}}(\mathbf{r})$ for a given mixed state $\Gamma$. Since $\Gamma$ already includes the squared wave function from Eq. (2), the mapping $\Gamma\mapsto\rho_{\Gamma}$ is linear. Note that whenever we talk about a “density”, this will be assumed to be a map $\rho:\mathbb{R}^{3}\to\mathbb{R}_{\geq 0}$ that is normalized to the particle number $N$, $\int\rho(\mathbf{r})\,\mathrm{d}\mathbf{r}=N$, like it is automatically the case for $\rho_{\psi}$ and $\Gamma_{\psi}$ if $\psi,\Gamma$ are normalized to 1. The density alone suffices to give an expression for the potential energy contribution. The resulting integral over the single-particle configuration space will be written like an inner product $\langle\cdot,\cdot\rangle$, to wit, $\displaystyle\langle\psi|V[v]|\psi\rangle$ $\displaystyle=\sum_{i=1}^{N}\sum_{\underline{\sigma}}\int_{\mathbb{R}^{3N}}v(\mathbf{r}_{i})|\psi|^{2}\,\mathrm{d}\underline{\mathbf{r}}$ (3) $\displaystyle=N\sum_{\underline{\sigma}}\int_{\mathbb{R}^{3}}v(\mathbf{r}_{1})\int_{\mathbb{R}^{3(N-1)}}|\psi|^{2}\,\mathrm{d}\mathbf{r}_{\perp}\,\mathrm{d}\mathbf{r}_{1}$ $\displaystyle=\int_{\mathbb{R}^{3}}v(\mathbf{r})\rho_{\psi}(\mathbf{r})\,\mathrm{d}\mathbf{r}=\langle v,\rho_{\psi}\rangle.$ The notation $\langle v,\rho\rangle$ thus expresses a dual pairing between two $L^{p}$ spaces or a combination of such, one for densities and the other one for potentials. These density and potential spaces are the topic of the next section. Without going into technicalities, the space $L^{p}(\mathbb{R}^{n})$, $1\leq p\leq\infty$, can be thought of as all functions $f(\mathbf{r})$ that have a finite $L^{p}$ norm $\|f\|_{L^{p}}=\left(\int_{\mathbb{R}^{n}}|f(\mathbf{r})|^{p}\,\mathrm{d}\mathbf{r}\right)^{1/p}<\infty,$ where in the case $p=\infty$ a supremum norm is employed instead. ## III Representability of densities The notion of “representability” is ubiquitous and conceptually important in DFT. It generally refers to the situation that any density of a certain class comes from a well-defined construction. Such a construction can simply be how a density is calculated from an $N$-particle wave function of finite kinetic energy following Eq. (2) and we then call the density “$N$-representable”. Or one demands that the density should be that of an actual ground-state solution of a Schrödinger equation with some given external potential $v$ and one calls it “$v$-representable”. However, this definition of $v$-representability is a bit naive since the set of permitted potentials to choose from was not even specified [3]. One could argue that any potential that can be put into the Schrödinger equation should be considered, but then the dual pairing $\langle v,\rho\rangle$ appearing in Eq. (3) between the spaces of densities and potentials might be “lost”, which has consequences for the density functionals defined later in Section VI. So in order to talk about $v$-representability, we will first have to choose a basic density space that includes the $N$-representable densities. The task of determining $N$-representable density classes was originally tackled by Gilbert [20] and Harriman [21]. In the first work, differentiability of the density was required, whereas in the second work no further conditions on the density were assumed. Here, we rely on the version by Lieb [2, Theorem 1.2] that gives the following class of $N$-representable densities, $N\text{-}\mathsf{rep}=\left\\{\rho\mid\rho(\mathbf{r})\geq 0,\smallint\rho\,\mathrm{d}\mathbf{r}=N,\nabla\sqrt{\rho}\in L^{2}(\mathbb{R}^{3})\right\\}.$ The benefit of the additional constraint $\nabla\sqrt{\rho}\in L^{2}$ is that one can always find a wave-function that not only gives the desired density but also has finite kinetic energy and is thus in $\mathcal{W}$ (and in addition is properly normalized). Lieb [2] further showed that $N\text{-}\mathsf{rep}$ is convex and included in $X=L^{1}(\mathbb{R}^{3})\cap L^{3}(\mathbb{R}^{3})$. This space $X$ is the basic density space in terms of $L^{p}$ spaces, so by Eq. (3) this automatically yields a corresponding potential space that is its dual, $X^{*}=L^{3/2}(\mathbb{R}^{3})+L^{\infty}(\mathbb{R}^{3})$. Any element $v\in X^{*}$ can thus be written as $v=v_{1}+v_{2}$ with $v_{1}\in L^{3/2}(\mathbb{R}^{3})$ and $v_{2}\in L^{\infty}(\mathbb{R}^{3})$. Potentials of Coulomb type, $v(\mathbf{r})=Cr^{-1}$, $r=|\mathbf{r}|$, are for example elements of this $X^{*}$ (by virtue of $\int_{0}^{R}|v(\mathbf{r})|^{3/2}r^{2}\,\mathrm{d}r<\infty$ for any finite $R>0$ and $|v(\mathbf{r})|<\infty$ for $r>R$). The issue of “$v$-representability” is much more profound. To date there is no explicit description for the set of all $v$-representable densities $v\text{-}\mathsf{rep}$. This issue is known as the “$v$-representability problem”. We already noted that $v\text{-}\mathsf{rep}$ should contain all densities that are a ground-state density for some potential $v\in X^{*}$. For a glimpse of what densities have to be included in this set we refer to the illustrative construction of Englisch and Englisch [22]. At this point one has to differentiate between several levels of $v$-representability. We defined $v\text{-}\mathsf{rep}$ as coming from a ground state of a Schrödinger equation with some given external potential $v$. Within DFT we usually consider two settings, the full system that contains a (Coulomb) interaction $W$ and the Kohn–Sham system that does not. So whenever we talk about $v$-representability, this can be amended by the attributes “interacting” or “non-interacting” and it is not obvious at this point if the two classes are equal, overlap, or are even disjoint. After all, the sets are not explicitly known. Within each class we also have the possibility of ground-state degeneracy. Then, instead of ground-state wave functions, the more general concept of density matrices comes into play. The resulting notions are then “pure-state $v$-representability” and “ensemble $v$-representability”. In the second case such a density $\rho$ is then the convex combination of pure-state $v$-representable densities $\rho_{k}$ that come from the degenerate ground- states $\psi_{k}$ of $H[v]$, i.e., $\rho=\sum_{k}\lambda_{k}\rho_{k}$ ($\lambda_{k}\in[0,1],\sum_{k}\lambda_{k}=1$). In the first case only densities from pure states are allowed, but they might still individually come from a set of degenerate ground-state wave functions. It was demonstrated by Englisch and Englisch [22] by giving explicit examples that there are $N$-representable densities that are not ensemble $v$-representable (an obvious example is a density that vanishes on a set of positive measure, however, for more elaborate examples we refer to Section 3.2 in Ref. 22). Levy [23] and Lieb [2] gave arguments that an ensemble $v$-representable density does not have to be pure-state $v$-representable. An explicit example for such a density $\rho\in v\text{-}\mathsf{rep}_{\mathrm{ens}}\setminus v\text{-}\mathsf{rep}_{\mathrm{pure}}$ was found within a finite-lattice system of cuboctahedral symmetry [24]. So we can symbolically note that $v\text{-}\mathsf{rep}_{\mathrm{pure}}\subsetneqq v\text{-}\mathsf{rep}_{\mathrm{ens}}\subsetneqq N\text{-}\mathsf{rep}\subsetneqq X.$ (4) In Garrigue [25] it was demonstrated that the set $v\text{-}\mathsf{rep}_{\mathrm{pure}}$ is path-connected. There are further topological relations between the sets appearing in Eq. (4) that are worth mentioning. Since every $\rho\in v\text{-}\mathsf{rep}_{\mathrm{ens}}$ is a convex combination $\rho=\sum_{k}\lambda_{k}\rho_{k}$ with $\rho_{k}\in v\text{-}\mathsf{rep}_{\mathrm{pure}}$, it holds $v\text{-}\mathsf{rep}_{\mathrm{ens}}\subseteqq\operatorname{conv}v\text{-}\mathsf{rep}_{\mathrm{pure}}\subsetneqq N\text{-}\mathsf{rep}\subsetneqq X,$ where $\operatorname{conv}$ is the convex hull of a set. So while $v\text{-}\mathsf{rep}_{\mathrm{pure}}$ is definitely not convex because of the mentioned counterexamples, $v\text{-}\mathsf{rep}_{\mathrm{ens}}$ might still be (to our understanding this is not known). Lastly, $N\text{-}\mathsf{rep}$ is the closure of $v\text{-}\mathsf{rep}_{\mathrm{ens}}$ within $L^{1}\cap L^{3}$, which means that any $\rho\in N\text{-}\mathsf{rep}$ can be approximated arbitrarily well by densities in $v\text{-}\mathsf{rep}_{\mathrm{ens}}$ when distance is measured in the $L^{1}\cap L^{3}$-norm [2, Theorem 3.14]. With the notion of the “subdifferential” from Section VII, this result can be established as a direct consequence of the Brøndsted–Rockafellar theorem [26, Corollary 2.44]. Still, potentials that lead to densities that are arbitrarily close could be very far apart in the potential space $X^{*}$. On the other hand, it has been suggested that $v\text{-}\mathsf{rep}_{\mathrm{pure}}$ is not dense in $N\text{-}\mathsf{rep}$ (see Conjecture 3.8 in Ref. 27). ## IV The Hohenberg–Kohn theorem The classical HK theorem [1] states the existence of a well-defined density- potential mapping for ground states. For a given potential $v$, $\displaystyle E[v]$ $\displaystyle=\inf\left\\{\langle\psi|H_{0}+V[v]|\psi\rangle\mid\psi\in\mathcal{W},\|\psi\|=1\right\\}$ (5) $\displaystyle=\inf\left\\{\langle\psi|H_{0}|\psi\rangle+\langle v,\rho_{\psi}\rangle\mid\psi\in\mathcal{W},\|\psi\|=1\right\\}$ is the _ground-state energy_ by the Rayleigh–Ritz variation principle. If a minimizer exists then $\psi$ and $\rho_{\psi}$ are the corresponding _ground state_ and _ground-state density_ that might not be unique in the case of degeneracy. If a minimizer does not exist, there is still always a sequence $\psi_{i}$ in $\mathcal{W}$ with $\|\psi_{i}\|=1$ such that $\langle\psi_{i}|H_{0}+V[v]|\psi_{i}\rangle$ converges to $E[v]$. In Eq. (5), $v$ should be selected from a class that makes $E[v]$ bounded below. See Reed and Simon, Section X.2, for an extensive discussion on such potentials [28]. A further demand on $v$ will later be that it guarantees a ground state that is non-zero (almost everywhere), a property needed in the proof of the second part of the HK theorem (HK2) below. In Eq. (5) the problem of solving a partial-differential equation, the stationary Schrödinger equation (1), has been transformed into a variational problem: finding a minimizer for Eq. (5). The route backwards is also feasible and any such minimizer is also a distributional solution to the Schrödinger equation [29, Theorem 11.8]. We will now demonstrate that simply by virtue of the structure of $E[v]$, where density and potential are combined in the term $\langle v,\rho\rangle$ that makes no explicit reference to the wave function while the remaining part $\langle\psi|H_{0}|\psi\rangle$ (or $\operatorname{Tr}(H_{0}\Gamma)$, if density matrices are used to describe the state) does not depend on $v$, we can already define a mapping from ground-state densities to ground-state wave functions or density matrices. This, then, is already half of a HK theorem, that we will already give in a variant for ensemble $v$-representable densities. ###### Theorem 1 (HK1). Let $\Gamma_{1}$ be a ground state of $H[v_{1}]$ and $\Gamma_{2}$ a ground state of $H[v_{2}]$. If $\Gamma_{1},\Gamma_{2}\mapsto\rho$, i.e., if these states share the same density, then $\Gamma_{1}$ is also a ground state of $H[v_{2}]$ and $\Gamma_{2}$ is also a ground state $H[v_{1}]$. ###### Proof 1. Since we assumed the existence of ground states $\Gamma_{1},\Gamma_{2}$ for the potentials $v_{1},v_{2}$, the infimum in Eq. (5), when varied over density matrices, is actually a minimum. Further, the potential-energy contribution $\langle v,\rho\rangle$ is fixed because $\rho$ is given and can be taken out of the minimum, $\displaystyle E[v_{1}]$ $\displaystyle=\min_{\Gamma^{\prime}_{1}\mapsto\rho}\operatorname{Tr}(H_{0}\Gamma^{\prime}_{1})+\langle v_{1},\rho\rangle$ $\displaystyle=\operatorname{Tr}(H_{0}\Gamma_{1})+\langle v_{1},\rho\rangle$ (6a) $\displaystyle E[v_{2}]$ $\displaystyle=\min_{\Gamma^{\prime}_{2}\mapsto\rho}\operatorname{Tr}(H_{0}\Gamma^{\prime}_{2})+\langle v_{2},\rho\rangle$ $\displaystyle=\operatorname{Tr}(H_{0}\Gamma_{2})+\langle v_{2},\rho\rangle$ (6b) For completeness, we also give the same expression for a general $v$ in case the state is pure. $E[v]=\min_{\psi\mapsto\rho}\langle\psi|H_{0}|\psi\rangle+\langle v,\rho\rangle\\\ $ (7) Here, the notation “$\Gamma\mapsto\rho$” and “$\psi\mapsto\rho$” means variation over all states in $\mathcal{W}$ with density $\rho$. But the remaining minima in Eq. (6) are then completely determined by the fixed ground-state density and we can always choose $\Gamma_{1}=\Gamma_{2}$ [primes removed] as a valid ground state. Thus the density alone already defines the ground state, irrespective of the potential $v_{1}$ or $v_{2}$. ∎ As highlighted before, the above proof relies purely on the specific structure of the energy function $E[v]$ that allows the potential part to be taken as a separate, additive contribution that depends solely on the density. This idea is due to Paul E. Lammert (during discussion at the workshop “Do Electron Current Densities Determine All There Is to Know?” in Oslo, 2018). In contrast to this, the usual proofs of this part of the HK theorem additionally depend on the _linear_ structure of the density-potential pairing. Moreover, such proofs are almost always performed indirectly (reductio ad absurdum), with a few notable exceptions [30; 31]. For completeness, we will give an additional, more traditional proof, yet one that is direct and does not work by raising a contradiction. ###### Proof 2. By the variational principle, we have $\displaystyle E[v_{1}]$ $\displaystyle=\operatorname{Tr}(H[v_{1}]\Gamma_{1})\leq\operatorname{Tr}(H[v_{1}]\Gamma_{2}),$ $\displaystyle E[v_{2}]$ $\displaystyle=\operatorname{Tr}(H[v_{2}]\Gamma_{2})\leq\operatorname{Tr}(H[v_{2}]\Gamma_{1}).$ Exploiting the shared density $\rho$, this may be written as $\begin{split}E[v_{1}]&=\operatorname{Tr}(H_{0}\Gamma_{1})+\langle v_{1},\rho\rangle\\\ &\leq\operatorname{Tr}(H_{0}\Gamma_{2})+\langle v_{1}+v_{2}-v_{2},\rho\rangle\\\ &=E[v_{2}]+\langle v_{1}-v_{2},\rho\rangle\end{split}$ and analogously as $E[v_{2}]\leq E[v_{1}]+\langle v_{2}-v_{1},\rho\rangle.$ Combining the inequalities gives $E[v_{1}]-E[v_{2}]=\langle v_{1}-v_{2},\rho\rangle$ and from $\operatorname{Tr}(H[v_{2}]\Gamma_{1})=\operatorname{Tr}(H[v_{1}]\Gamma_{1})-\langle v_{1}-v_{2},\rho\rangle$ that $\operatorname{Tr}(H[v_{2}]\Gamma_{1})=E[v_{1}]$. So $\Gamma_{1}$ is also a ground state of $H[v_{2}]$. Likewise, $\operatorname{Tr}(H[v_{1}]\Gamma_{2})=E[v_{2}]$, so $\Gamma_{2}$ is also a ground state of $H[v_{1}]$, as required. ∎ HK1 holds generally for mixed or pure ground states. The same proofs remain valid when the theorem is specialized to a statement about pure states $\Gamma_{i}=|\psi_{i}\rangle\langle\psi_{i}|$. An immediate but maybe surprising consequence that is often referred to as the basis of DFT is that a ground-state density ${\rho_{\mathrm{gs}}}$ alone already determines a ground state. This result has been coined a _weak HK-like result_ before [32] and it will be used to define the HK1 functionals on $v\text{-}\mathsf{rep}_{\mathrm{pure}}$ and $v\text{-}\mathsf{rep}_{\mathrm{ens}}$ in Eq. (9) below. The ground state (associated with ${\rho_{\mathrm{gs}}}$) is pure if ${\rho_{\mathrm{gs}}}\in v\text{-}\mathsf{rep}_{\mathrm{pure}}$ but has to be an ensemble if ${\rho_{\mathrm{gs}}}\in v\text{-}\mathsf{rep}_{\mathrm{ens}}\setminus v\text{-}\mathsf{rep}_{\mathrm{pure}}$. Any state that is a minimizer in Eq. (6) is really a ground state for _all_ potentials that share the same ground- state density. That all those potentials are in fact equal (up to a constant) is then the statement of HK2, the second part of the HK theorem. It will be formulated for eigenstates, in case of an ensemble we are free to just take any of its components. ###### Theorem 2 (HK2). If two potentials share any common eigenstate and if that eigenstate is non- zero almost everywhere, then the potentials are equal up to a constant. ###### Proof. If $v_{1},v_{2}$ share a common eigenstate $\psi$ it holds $\displaystyle(H_{0}+V[v_{1}])\psi=E[v_{1}]\psi,$ $\displaystyle(H_{0}+V[v_{2}])\psi=E[v_{2}]\psi.$ Subtraction of the two equations and moving all potential parts that do not depend on $\mathbf{r}_{1}$ to the right-hand side gives $\displaystyle(v_{1}(\mathbf{r}_{1})-v_{2}(\mathbf{r}_{1}))\psi=$ $\displaystyle(E[v_{1}]-E[v_{2}])\psi$ (8) $\displaystyle-\sum_{i=2}^{N}(v_{1}(\mathbf{r}_{i})-v_{2}(\mathbf{r}_{i}))\psi.$ Since we assumed $\psi$ non-zero almost everywhere, we can then divide by $\psi$ and get $v_{1}(\mathbf{r}_{1})-v_{2}(\mathbf{r}_{1})=\mathrm{constant}$ (almost everywhere) because the right-hand side does not depend on $\mathbf{r}_{1}$. ∎ Since HK2 states that sharing _any_ common eigenstate for two potentials means that they are equal (up to a constant), this of course implies that the potentials share _all_ eigenstates because they yield exactly the same Hamiltonian (up to an additive constant that just shifts the spectrum). The special requirement that the wave function is non-zero (almost everywhere) is guaranteed for a large class of potentials by the _unique-continuation property (UCP) from sets of positive measure_. This property will be further discussed in Section V. That zeroes (nodes) in the wave function _are_ still allowed on a set of measure zero is important here, since the fermionic many- particle wave functions will exhibit nodal surfaces when particle positions agree. Outside of the continuum setting, for example in finite-lattice systems, such a UCP is _not_ at hand and there are actual counterexamples to HK2, were two different potentials share a common eigenstate [24]. The complete HK result is then obtained by combining the two theorems above. We will assume here that the potential is from the mentioned class that guarantees a non-zero ground state. We should remember that such or similar restrictions will always come into play if we want to show validity of a density-potential mapping in other settings. The statement will be formulated for densities in $v\text{-}\mathsf{rep}_{\mathrm{ens}}$, so it automatically holds for $v\text{-}\mathsf{rep}_{\mathrm{pure}}$ as well. ###### Corollary 3 (HK). If two potentials share a common ensemble $v$-representable ground-state density, then they are equal up to a constant. ###### Proof. By HK1 there is a density matrix $\Gamma=\sum_{k}\lambda_{k}|\psi_{k}\rangle\langle\psi_{k}|$ that is a ground state for both potentials. Since at least one $\lambda_{k}\neq 0$ and since the corresponding $\psi_{k}$ is a ground-state wave function for both Hamiltonians, the proof can be completed by HK2. ∎ This structuring into two separate theorems was already used in Kohn, Savin, and Ullrich [33], just in the reverse order, for a brief argument about DFT with magnetization. Historically, the HK theorem was first given only for the non-degenerate case and was only later extended to include degeneracy [22; 34]. The proof presented here does not suffer from any limitation to non- degenerate ground states. A final note is directed towards more general DFTs that will be briefly discussed in Section X and especially in the forthcoming Part II of this review. For other types of potentials, like vector potentials, the statement in the HK theorem would not necessarily be that the potentials are equal “up to a constant”, but for example “up to a gauge transformation”. The set of gauge transformations that are possible without affecting the physical properties of the system then have to be specified within the respective theory. ## V The unique-continuation property In this section, we summarize some important results on the unique- continuation property (UCP) of solutions to the Schrödinger equation that is heavily used in the context of (mathematical formulation of) HK-type theorems. The current understanding is that the UCP cannot be avoided in a rigorous proof of a HK-type theorem. The setting will be slightly more general than before and allow for dimensionality $d$ of the spatial part of the single- particle configuration space $\mathbb{R}^{d}$. The $N$-particle configuration space is then $\mathbb{R}^{n}$ with $n=dN$. Roughly speaking, the desired UCP result states that under certain conditions on the potentials building up the operators $V$ and $W$ and if a solution $\psi$ to the (distributional) equation $H[v]\psi=0$ vanishes on a set of positive measure, then $\psi$ vanishes everywhere. That the right hand side is zero comes as no restriction here, since the energy $E$ can always be absorbed into the scalar potential $v$. The usual literature on the UCP shows _strong_ UCP, which means that $\psi$ is assumed to _vanish to infinite order_ at a point $\underline{\mathbf{r}}_{0}\in\mathbb{R}^{n}$ and then the statement follows. A function $f(\mathbf{r})$ is said to vanish to infinite order at $\underline{\mathbf{r}}_{0}\in\mathbb{R}^{n}$ if for all $k\geq 1$ there is a $c_{k}$ such that $\int_{|\underline{\mathbf{r}}-\underline{\mathbf{r}}_{0}|<\epsilon}|f(\underline{\mathbf{r}})|^{2}\,\mathrm{d}\underline{\mathbf{r}}<c_{k}\epsilon^{k}$ for every $0<\epsilon<1$. Now a very convenient result by Regbaoui [35] shows that the UCP on sets of positive measure actually follows from such a strong UCP if the potentials are in $L^{n/2}_{\mathrm{loc}}$. This work apparently built on de Figueiredo and Gossez [36] that again rests on an early estimate for general Sobolev spaces by Ladyzenskaya and Ural’tzeva [37, Lemma 3.4]. The result and its proof have been repeated in Lammert [38]. For us that means that any strong UCP can also be used as a UCP on sets of positive measure which is the one needed for the proof of HK2. Yet the traditional strong UCP results, like most notably in Jerison and Kenig [39], also give dimension- dependent constraints on the potentials like $L^{n/2}_{\mathrm{loc}}$, which approaches $L^{\infty}$ for growing particle number and is thus too restrictive for our use where singular potentials need to be considered. The saving idea recently came from Garrigue [31] and was also extended to more complex systems [40; 41]: To take the special $N$-body structure of the potentials into account and thus avoid any dependence of the constraints on the particle number $N$. ###### Theorem 4 (Garrigue’s UCP). Suppose that the potentials are in $L_{\mathrm{loc}}^{p}(\mathbb{R}^{d})$ with $p>2$ for $d=3$ and $p=\max(2d/3,2)$ else. If a solution $\psi$ to the Schrödinger equation vanishes on a set of positive measure or if it vanishes to infinite order at any point, then $\psi=0$. The most relevant case here is obviously $d=3$ which means that the potentials need to be in $L_{\mathrm{loc}}^{p}(\mathbb{R}^{3})$ with $p>2$ but exactly $p=2$ is not enough yet. This clearly does not fit our potential space $X^{*}=L^{3/2}(\mathbb{R}^{3})+L^{\infty}(\mathbb{R}^{3})$, so while this UCP result is the best one available, it cannot be used for a HK2 theorem that covers the whole potential space of DFT in the formulation discussed here. Lieb [2] also remarked on the UCP in the context of the HK theorem, which “is believed to hold” for potentials in $X^{*}$, however in a weaker form that is not sufficient for the current purpose. So whenever we state that the HK holds in standard DFT, we actually mean under the given restrictions on the potentials. ## VI Hierarchy of density functionals The first part of the HK theorem, HK1, analogously holds in many different varieties of DFT (that will be explored in Part II), simply because its validity just depends on the form of the energy functional. HK1 then ensures that we can map from pure-state $v$-representable ground-state densities ${\rho_{\mathrm{gs,pure}}}$ to ground-state wave functions $\psi[{\rho_{\mathrm{gs,pure}}}]$ and from ensemble $v$-representable ground- state densities ${\rho_{\mathrm{gs,ens}}}$ to ground-state density matrices $\Gamma[{\rho_{\mathrm{gs,ens}}}]$. This makes it possible to define the HK1 functionals $\displaystyle F_{\mathrm{HK1,pure}}[{\rho_{\mathrm{gs}}}]=\langle\psi[{\rho_{\mathrm{gs}}}]|H_{0}|\psi[{\rho_{\mathrm{gs}}}]\rangle$ $\displaystyle\,\,\,\,\text{on}\;v\text{-}\mathsf{rep}_{\mathrm{pure}}$ (9a) and $\displaystyle F_{\mathrm{HK1,ens}}[{\rho_{\mathrm{gs}}}]=\operatorname{Tr}(H_{0}\Gamma[{\rho_{\mathrm{gs}}}])$ $\displaystyle\,\,\,\,\text{on}\;v\text{-}\mathsf{rep}_{\mathrm{ens}}$ (9b) as the energy contribution only from the internal parts $H_{0}$ of the Hamiltonian. The universal nature of such functionals, being independent of any external $v$, justifies the usual attribution as universal functionals. It is then possible to determine also the internal energy contributions for any state with density ${\rho_{\mathrm{gs}}}$ just from ${\rho_{\mathrm{gs}}}$. To get the total ground-state energy (5) with the help of the functional above, it is enough to vary over $v$-representable densities alone, instead of the much larger set of wave functions. We can write $\displaystyle E[v]$ $\displaystyle=\inf\\{\langle\psi|H_{0}|\psi\rangle+\langle v,\rho_{\psi}\rangle\mid\psi\in\mathcal{W},\|\psi\|=1\\}$ (10) $\displaystyle=\inf_{{\rho^{\prime}_{\mathrm{gs}}}}\\{\langle\psi[{\rho^{\prime}_{\mathrm{gs}}}],H_{0}\psi[{\rho^{\prime}_{\mathrm{gs}}}]\rangle+\langle v,{\rho^{\prime}_{\mathrm{gs}}}\rangle\\}$ $\displaystyle=\inf_{\rho^{\prime}_{\mathrm{gs}}}\\{F_{\mathrm{HK1,pure}}[{\rho^{\prime}_{\mathrm{gs}}}]+\langle v,{\rho^{\prime}_{\mathrm{gs}}}\rangle\\}\quad\text{on}\;X^{*},$ or equivalently with $F_{\mathrm{HK1,ens}}$. We see already that there is a certain ambiguity in which density functional to use in the definition of $E[v]$. The other density functionals presented here will all have the property that they give the correct ground-state energy when applied in Eq. (10) which makes them all _admissible_ functionals [42]. Yet, they will differ with respect to their mathematical properties and we thus aim for the one with the best features. The first problem here is that the densities to be considered in the variational problem are limited to those that are actual ground-state densities ($v\text{-}\mathsf{rep}$), because else $F_{\mathrm{HK1}}[{\rho_{\mathrm{gs}}}]$ is left undefined, and we already learned in Section III that $v\text{-}\mathsf{rep}$ is not an explicitly characterized set. Apart from that, HK1 just states the existence of a map ${\rho_{\mathrm{gs}}}\mapsto\psi$ or $\Gamma$ without giving any hints towards a constructive scheme. A first step to overcome these problems is to inspect Eq. (7). This suggests the definition of another pair of density functionals that goes under the name of “constrained search”, $\displaystyle F_{\mathrm{CS,pure}}[\rho]=\inf_{\psi\mapsto\rho}\langle\psi|H_{0}|\psi\rangle$ $\displaystyle\quad\text{on}\;N\text{-}\mathsf{rep}\;\text{and}$ (11a) $\displaystyle F_{\mathrm{CS,ens}}[\rho]=\inf_{\Gamma\mapsto\rho}\operatorname{Tr}(H_{0}\Gamma)$ $\displaystyle\quad\text{on}\;N\text{-}\mathsf{rep}.$ (11b) The domain is now the larger, convex, and explicitly defined $N\text{-}\mathsf{rep}$ in both cases. Note that the literature mostly denotes those functionals as $F_{\mathrm{CS,pure}}=F_{\mathrm{LL}}$ (“Levy–Lieb” [9; 2]) and $F_{\mathrm{CS,ens}}=F_{\mathrm{DM}}$ (from “density matrix” [2]). A recent, comprehensive study of these functionals can be found in Lewin, Lieb, and Seiringer [43]. Since the density is limited to the set $N\text{-}\mathsf{rep}$ that guarantees finite kinetic energy, the infima in Eq. (11) are always attained, though not necessarily by a possible ground state (if $\rho$ is not $v$-representable), and can thus be replaced by minima in both cases [2, Theorem 3.3]. The convex combination of pure-state projections into density matrices translates to the functionals, so that $F_{\mathrm{CS,ens}}$ is the convex envelope of $F_{\mathrm{CS,pure}}$ [24, Proposition 18, the article treats DFT on a lattice but the statement and the proof of proposition remains exactly the same in the continuum case.]. This automatically ensures that $F_{\mathrm{CS,ens}}$ is convex, a fact that can also be concluded from observing that $\Gamma\mapsto\rho$ linear [2, Section 4.B]. Since these density functionals appear in the optimization problem that determines the ground-state energy and density, like in Eq. (10), convexity is of great importance because only for a convex functional can we be sure that identifying any _local_ minimum also means that a _global_ minimum has been found. So while we now know that $F_{\mathrm{CS,ens}}$ is convex, the previous functional $F_{\mathrm{HK1,pure}}$ does not even have a convex domain and therefore cannot be convex. Levy [23] and Lieb [2] also gave arguments for the non-convexity of $F_{\mathrm{CS,pure}}$. Since $F_{\mathrm{CS,ens}}=\operatorname{conv}F_{\mathrm{CS,pure}}$, any density where $F_{\mathrm{CS,ens}}[\rho]\neq F_{\mathrm{CS,pure}}[\rho]$ already shows non-convexity of $F_{\mathrm{CS,pure}}$. But this is equivalent to saying that $\rho$ is ensemble $v$-representable while it is _not_ pure-state $v$-representable, so $\rho\in v\text{-}\mathsf{rep}_{\mathrm{ens}}\setminus v\text{-}\mathsf{rep}_{\mathrm{pure}}$ [24, Proposition 21]. Note especially, that HK1 was necessary to define $F_{\mathrm{HK1}}$, but is not needed any more for the constrained-search functional $F_{\mathrm{CS}}$. Being able to define a universal constrained-search functional, one that is independent of the potential like in Eq. (11), already fully facilitates the proof of HK1 and thus implies this result. A potential-independent constrained-search functional _already implicitly includes HK1_. This implication was proven by Levy [9] along the lines of the usual HK proof and is mentioned in textbooks like Parr and Yang [18, after their Eq. (3.4.4)] and Tsuneda [44, after Eq. (4.5)]. Speaking generally though, a constrained search is just as feasible if the constrained-search functional also depends on the external potential $v$ (although it would not be universal), so indeed this approach is more general than relying on HK1. Such a case turns up in CDFT when the current variable is the total current that itself depends on the vector potential (see Part II of this review for more on this). By employing the constrained-search functional, the ground-state energy from Eq. (5) can now be rewritten again as $\displaystyle E[v]$ $\displaystyle=\inf\\{\langle\psi|H_{0}|\psi\rangle+\langle v,\rho_{\psi}\rangle\mid\psi\in\mathcal{W},\|\psi\|=1\\}$ $\displaystyle=\inf\\{F_{\mathrm{CS,pure}}[\rho_{\psi}]+\langle v,\rho_{\psi}\rangle\mid\psi\in\mathcal{W},\|\psi\|=1\\}$ $\displaystyle=\inf_{\rho}\\{F_{\mathrm{CS,pure}}[\rho]+\langle v,\rho\rangle\\}\quad\text{on}\;X^{*},$ or equivalently with $F_{\mathrm{CS,ens}}$, where minimization is now performed over $N\text{-}\mathsf{rep}$. When looking at non-interacting systems, the definitions of $F_{\mathrm{HK1}}$, Eq. (9), and $F_{\mathrm{CS}}$, Eq. (11), involve only the kinetic-energy operator $T$ instead of $H_{0}$. We will then denote these functionals with a zero superscript, $F_{\mathrm{HK1}}^{0},F_{\mathrm{CS}}^{0}$, etc., that indicates that non- interacting systems are considered. A further functional then comes into play that is defined like $F_{\mathrm{CS,pure}}$, but where only Slater determinants are considered as wave functions. We define on $N\text{-}\mathsf{rep}$, $F^{0}_{\mathrm{SD}}[\rho]=\inf_{\phi\mapsto\rho}\left\\{\langle\phi|T|\phi\rangle\mid\text{{\hbox{\phi}} is a Slater determinant}\right\\}.$ The usual name in the literature is $F^{0}_{\mathrm{SD}}=T_{S}$. This functional is of importance because it is the one used in Kohn–Sham theory which will be discussed in Section VIII. In their original article, Kohn and Sham [45] implicitly set $F^{0}_{\mathrm{SD}}=F_{\mathrm{HK1,pure}}^{0}$ for all non-interacting pure-state $v$-representable densities, which has been noted to be wrong because of possible degeneracy [2, Section 4.C]. On the other hand, for non-degenerate ground states $\phi$, which by necessity are always determinants in non-interacting systems, it holds that $F^{0}_{\mathrm{SD}}[\rho_{\phi}]=F_{\mathrm{CS,pure}}^{0}[\rho_{\phi}]=F_{\mathrm{HK1,pure}}^{0}[\rho_{\phi}]$, and else $F^{0}_{\mathrm{SD}}\geq F_{\mathrm{CS,pure}}^{0}$. Nevertheless, for practical purposes, $F^{0}_{\mathrm{SD}}$ usually takes up the role of the density functional when defining the energy functional in a non-interacting setting. The transformation from any density functional $F_{\bullet}$ for an interacting system from above to the energy functional, $E[v]=\inf_{\rho}\\{F_{\bullet}[\rho]+\langle v,\rho\rangle\\}\quad\text{on}\;X^{*},$ (12) is called the _convex conjugate_ or Legendre–Fenchel transformation [26, Section 2.1.4]. There is also a way to reverse the transformation and we define $F[\rho]=\sup_{v}\\{E[v]-\langle v,\rho\rangle\\}\quad\text{on}\;X.$ (13) This $F$ is the famous _Lieb functional_ [2], yet another density functional, but this time the last one to be defined in standard DFT. It is the _biconjugate_ of any $F_{\bullet}$ considered before. Defined this way, both $E$ and $F$ are lower-semicontinuous and $E$ is concave while $F$ is convex and has the property $F\leq F_{\bullet}$ [26, Proposition 2.19]. Actually, as a biconjugate, $F$ is the largest convex and lower semicontinuous functional that fulfills $F\leq F_{\bullet}$ which makes it the _convex envelope_ of $F_{\bullet}$. The domain is now the whole $X=L^{1}(\mathbb{R}^{3})\cap L^{3}(\mathbb{R}^{3})$, but automatically $F[\rho]=\infty$ for all densities that are not in $N\text{-}\mathsf{rep}$ [2, Theorem 3.8], while at the same time $F[\rho]<\infty$ if $\rho\in N\text{-}\mathsf{rep}$ [2, Theorem 3.9 and the following Remark]. Let the _effective domain_ ‘$\operatorname{dom}$’ of a convex functional be the elements from its domain where it is finite, then this means that $\operatorname{dom}F=N\text{-}\mathsf{rep}$. Having reached $F$, it does not matter any more which (admissible) functional has been used in Eq. (12), which means the convex envelopes of all the functionals above agree. Conversely, the Legendre–Fenchel transformation can also be utilized to go back from $F$ to $E$ [26, Theorem 2.22], $E[v]=\inf_{\rho}\\{F[\rho]+\langle v,\rho\rangle\\}.$ (14) We already noted that $F$ is convex and lower-semicontinuous, which are both important properties if we want to use the variational problem $E[v]=\inf_{\rho}\\{F[\rho]+\langle v,\rho\rangle\\}$ to find a minimizing density. The same properties come into play when defining the minimizers by differentiation in Section VII. From the definition of $F$ it follows directly that $E[v]\leq F[\rho]+\langle v,\rho\rangle,$ a version of the Young inequality. Equality in the above estimate holds if the density _is_ the ground-state density ${\rho_{\mathrm{gs}}}$ for the potential $v$, $E[v]=F[{\rho_{\mathrm{gs}}}]+\langle v,{\rho_{\mathrm{gs}}}\rangle.$ For $F_{\mathrm{CS,pure}}$ the converse holds too: If $E[v]=F_{\mathrm{CS,pure}}[\rho]+\langle v,\rho\rangle$ then $\rho$ is a ground-state density within $v\text{-}\mathsf{rep}_{\mathrm{pure}}$ for the potential $v$ and further $F_{\mathrm{HK1,pure}}[\rho]=F_{\mathrm{HK1,ens}}[\rho]=F_{\mathrm{CS,pure}}[\rho]=F_{\mathrm{CS,ens}}[\rho]=F[\rho]$ [2, Theorem 3.10]. But what about using the more general functional $F$ for the variational principle like in Eq. (14)? Can we find a real ground state like this or will this variational principle yield additional artificial solutions because it is too general? Because it is the convex envelope of the other functionals, it cannot produce a functional value below the ground-state energy, but it could produce a minimizing density where there are no $v$-representable ground-state densities! The problem is solved if we allow for ensembles of ground states: An “amusing fact” in Lieb [2, Eq. (4.5)] gives $F=F_{\mathrm{CS,ens}}$ on $N\text{-}\mathsf{rep}$, which effectively means $F=F_{\mathrm{CS,ens}}$ since we can just set $F_{\mathrm{CS,ens}}[\rho]=\infty$ outside of its domain $N\text{-}\mathsf{rep}$ to achieve equality globally on $X$. So any minimizer of $F+\langle v,\cdot\rangle$ is also one of $F_{\mathrm{CS,ens}}+\langle v,\cdot\rangle$ and it is further the convex combination of ground states for the potential $v$. Consequently, when talking about ground states in the context of the functional $F$, we will always actually mean _ensembles_ of possibly degenerate ground states. When comparing the functionals on $X$, we just set them to $\infty$ whenever we are outside their domains. The following hierarchy can be set up and is further laid out in Table 1. $F=F_{\mathrm{CS,ens}}\leq\begin{array}[]{c}F_{\mathrm{CS,pure}}(\leq F^{0}_{\mathrm{SD}})\\\ F_{\mathrm{HK1,ens}}\end{array}\leq F_{\mathrm{HK1,pure}}.$ Here, $F^{0}_{\mathrm{SD}}$ appears in parentheses since it only comes into play in the non-interacting setting where we can perform the same type of transformations and have $F^{0}[\rho]$ and $E^{0}[v]$. $F_{\bullet}$ | convex | domain | convex ---|---|---|--- $F_{\mathrm{HK1,pure}}$ | no | | $v\text{-}\mathsf{rep}_{\mathrm{pure}}$ | no | $F_{\mathrm{HK1,ens}}$ | ? | | $v\text{-}\mathsf{rep}_{\mathrm{ens}}$ | ? | $\Downarrow$ cl $F_{\mathrm{CS,pure}}$ | no | $\Downarrow$ conv | $N\text{-}\mathsf{rep}$ | yes $F_{\mathrm{CS,ens}}$ | yes | $N\text{-}\mathsf{rep}$ | yes | $F$ | yes | | $L^{1}\cap L^{3}$ | yes | Table 1: The table shows the relations between the functionals discussed in Section VI. From $F_{\mathrm{HK1,pure}}$ to $F_{\mathrm{HK1,ens}}$ the domain gets extended to $v\text{-}\mathsf{rep}_{\mathrm{ens}}$ while they agree on $v\text{-}\mathsf{rep}_{\mathrm{pure}}$. From $F_{\mathrm{HK1,ens}}$ to $F_{\mathrm{CS,pure}}$ the domain gets closed (cl) within $L^{1}\cap L^{3}$ and from $F_{\mathrm{CS,pure}}$ to $F_{\mathrm{CS,ens}}$ the functional itself gets convexified (conv) while the domain remains the same. Finally, $F$ is just equal to $F_{\mathrm{CS,ens}}$ on $N\text{-}\mathsf{rep}$. ## VII Density-potential mappings from differentials In the previous section it was stated that in order to get the ground-state density of any system we have to find a solution to the variational problem $E[v]=\inf_{\rho}\\{F[\rho]+\langle v,\rho\rangle\\},$ (15) now relying on the density functional $F$ from Eq. (13). To find the global minimum of a _convex and lower-semicontinuous_ functional we can perform differentiation, i.e., demand that the differential of $F[\rho]+\langle v,\rho\rangle$ with respect to $\rho$ must equal zero at the position of a ground-state density ${\rho_{\mathrm{gs}}}$. $\rho$$\rho_{0}$$\infty$ Figure 1: Example of a convex and lower- semicontinuous function with a discontinuity at $\rho_{0}$ and some elements from the subdifferential displayed as linear continuous tangent functionals at $\rho_{0}$, represented by dashed lines. The suitable notion of differentiation here is the subdifferential $\underline{\partial}F$ that gives the _set_ of all linear continuous tangent functionals to a convex functional $F$ at a given density $\rho$, $\underline{\partial}F[\rho]=\\{v\in X^{*}\mid\forall\rho^{\prime}\in X:F(\rho)\leq F(\rho^{\prime})+\langle v,\rho-\rho^{\prime}\rangle\\}.$ It is always well-defined, since the set $\underline{\partial}F[\rho]$ can contain many elements, in case the functional $F$ has a kink (like the example shown in Fig. 1), or can even be empty. Finally, if it contains exactly one element, we found a _unique_ potential yielding that ground-state density. In any case, the variational problem (15) has a minimizer ${\rho_{\mathrm{gs}}}$ if and only if the following condition is fulfilled [26, Proposition 2.33], $\left.\underline{\partial}(F[\rho]+\langle v,\rho\rangle)\right|_{\rho={\rho_{\mathrm{gs}}}}\ni 0\Longleftrightarrow\langle v,\cdot\rangle\in-\underline{\partial}F[{\rho_{\mathrm{gs}}}].$ (16) In what follows, we identify $v$ with the functional $\langle v,\cdot\rangle$ whenever the context implies a functional on density space instead of a potential on configuration space, so Eq. (16) can be written $v\in-\underline{\partial}F[{\rho_{\mathrm{gs}}}]$. The potential as the subdifferential of the density functional means that potentials $v$ are from the dual of the space of densities like already noted in Section III. This general principle is not always respected in more complex versions of DFT, as we will see in Section X and discuss further in Part II of this review. If the set $\underline{\partial}F[{\rho_{\mathrm{gs}}}]$ is non-empty then there is at least one potential $v\in X^{*}$ that yields the given ground- state density. The set of all densities where $\underline{\partial}F[{\rho_{\mathrm{gs}}}]\neq\emptyset$ is called the domain of the subdifferential, so it follows that $\operatorname{dom}\underline{\partial}F=v\text{-}\mathsf{rep}_{\mathrm{ens}}$. Note that by a theorem of convex analysis [26, Corollary 2.44], $\operatorname{dom}\underline{\partial}F$ is dense in $\operatorname{dom}F$, so $v\text{-}\mathsf{rep}_{\mathrm{ens}}$ is dense in $N\text{-}\mathsf{rep}$, a fact already expressed with $N\text{-}\mathsf{rep}$ being the closure of $v\text{-}\mathsf{rep}_{\mathrm{ens}}$ in Section III. The meaning of a valid HK theorem for a class of densities is that they can all be mapped as ground-state densities back to a unique potential (modulo a constant) and consequently $-\underline{\partial}F[{\rho_{\mathrm{gs}}}]=\\{v+c\mid c\in\mathbb{R}\\}$. By eliminating the (physically unimportant) constant potentials from the potential space, the subdifferential of a $v$-representable density is precisely $-\underline{\partial}F[{\rho_{\mathrm{gs}}}]=\\{v\\}$ if the HK theorem holds. If, on the other hand, $F$ is assumed differentiable, then the directional derivative $-\frac{\delta}{\delta\rho}F[{\rho_{\mathrm{gs}}}]=v$ anyway always maps to a unique potential. One thus has a well-defined map from densities in $v\text{-}\mathsf{rep}$ to the corresponding potentials, exactly the content of the HK theorem! But where did it enter? The HK theorem is here a _consequence_ from the assumption of differentiability of $F[\rho]$ at $v$-representable densities. The situation will be summarized diagrammatically in Section XI. Because any potential $v$ that we determine by Eq. (16) will also be the maximizer in the conjugate variational problem $F[{\rho_{\mathrm{gs}}}]=\sup_{v}\\{E[v]-\langle v,{\rho_{\mathrm{gs}}}\rangle\\},$ we can just as well say the same with the superdifferential of the concave functional $E$, $\left.\overline{\partial}(E[v^{\prime}]-\langle v^{\prime},{\rho_{\mathrm{gs}}}\rangle)\right|_{v^{\prime}=v}=0\Longleftrightarrow{\rho_{\mathrm{gs}}}\in\overline{\partial}E[v].$ (17) The right hand side, ${\rho_{\mathrm{gs}}}\in\overline{\partial}E[v]$, means to find a density (or possibly many) that comes from a wave-function that minimizes the total energy including $v$. It is thus a conceptual shortcut to map from potentials to ground-state densities without any reference to an underlying wave function or Schrödinger equation. The situation of a set $\overline{\partial}E[v]$ with more than one element is known from degeneracies of the Hamiltonian $H_{0}+V[v]$, where different linearly independent ground states with eventually different densities all have the same eigenvalue. We showed in this section the important role of the generalized concepts of sub/superdifferentials in the context of DFT, because indeed the functionals from Section VI can _not_ be assumed differentiable as van Leeuwen [46] has demonstrated for the $F_{\mathrm{HK1}}$ functionals and Lammert [3] for $F_{\mathrm{CS}}$. The reason for non-differentiability even of $F_{\mathrm{CS}}$ is that at any $\rho$ the functional $F[\rho+\delta\rho]$ is infinite for various, arbitrarily small shifts $\delta\rho$ that lead out of $N\text{-}\mathsf{rep}$, even if the normalization of the density is kept constant. This happens by infinitely increasing the internal energy through tiny oscillations of the density. A possible way to prevent that is to limit the density space $X$ so that such shifts $\delta\rho$ are not possible any more and Lammert [3] actually shows this for the Sobolev space $H^{2}(\mathbb{R}^{3})$ when $\rho$ is also assumed $v$-representable. Another way is to establish a coarse-grained model for DFT in which $F$ really becomes differentiable and every density is ensemble $v$-representable [47]. In the following section, in accordance with the vast majority of the literature, we will assume functional differentiability of $F$ and consequently $v$-representability. This strong assumption can be justified a posteriori, as discussed later in Section IX, when a regularization procedure is applied. ## VIII Linking to a reference system: the Kohn–Sham scheme In Section VI it was noted that a functional might be introduced for an interacting or a non-interacting system. This means the respective Hamiltonian has the internal part $T+\lambda W$ with $\lambda\in\\{0,1\\}$. We will now write $F^{1}$ and $F^{0}$ to differentiate clearly between those two situations. We then introduce the difference functional $F_{\mathrm{Hxc}}=F^{1}-F^{0}$, which just corresponds to the internal-energy difference between the interacting and the non-interacting system and that will later be linked to the Hartree-exchange-correlation potential $v_{\mathrm{Hxc}}$. This potential effectively compensates for the Hartree- mean-field interaction as well as ‘exchange’ and ‘correlation’ effects. The idea behind introducing this auxiliary non-interacting system is that the energy difference between the (numerically tractable) non-interacting system and the (numerically unfeasible) interacting system is small and can be efficiently approximated. Since the reference system is non-interacting, $F^{0}_{\mathrm{SD}}$ can be employed for $F^{0}$ if degeneracy for the ground state does not have to be taken into account, like it was mentioned in Section VI, and this switchover is performed in most practical situations. Then the energy functional for the full system is $\displaystyle E^{1}[v]$ $\displaystyle=\inf_{\rho}\\{F^{1}[\rho]+\langle v,\rho\rangle\\}$ $\displaystyle=\inf_{\rho}\\{F^{0}[\rho]+F_{\mathrm{Hxc}}[\rho]+\langle v,\rho\rangle\\}$ $\displaystyle=\inf_{\phi}\\{\langle\phi|T|\phi\rangle+F_{\mathrm{Hxc}}[\rho_{\phi}]+\langle v,\rho_{\phi}\rangle\\}.$ In the last step the variation is changed from $N\text{-}\mathsf{rep}$ densities to single Slater determinants $\phi$, the minimizer – if it exists – is then the Kohn–Sham Slater determinant. In order to link this to a partial differential equation for the orbitals $\varphi_{i}$ constituting $\phi$, the Kohn–Sham equation, variation of the energy expression above with respect to $\varphi_{i}$ is performed under the constraint that all the $\varphi_{i}$ stay normalized. This means $\rho_{\phi}(\mathbf{r})=\sum_{i=1}^{N}\sum_{\sigma}|\varphi_{i}(\mathbf{r}\sigma)|^{2}$ always stays in $N\text{-}\mathsf{rep}$, but generally the issue of non- differentiability from Section VII remains. The resulting equation is a one- particle Schrödinger equation with effective potential $v_{s}$ and eigenstates $\varphi_{i}$, $\left(-\tfrac{1}{2}\nabla^{2}+v_{s}(\mathbf{r})\right)\varphi_{i}(\mathbf{r}\sigma)=\varepsilon_{i}\varphi_{i}(\mathbf{r}\sigma).$ (18) On the other hand this approach does not lead to the effective potential $v_{s}$ for the Kohn–Sham equation right away, but requires the additional, computationally challenging step of extracting the effective potential from the variation of $F_{\mathrm{Hxc}}$ with respect to the orbitals (OEP integral equation[48]). To have a well defined $F_{\mathrm{Hxc}}[\rho]=F^{1}[\rho]-F^{0}[\rho]$, the $\rho$ must be both, interacting and non-interacting $v$-representable. Both systems then share the same ground-state density $\rho$ when the different external potentials $v\in-\underline{\partial}F^{1}[\rho]\quad\text{and}\quad v_{s}\in-\underline{\partial}F^{0}[\rho]$ (19) are assigned to them. That the density $\rho$ is simultaneously interacting _and_ non-interacting $v$-representable is tacitly assumed here, else one of the subdifferentials above is empty. This means that actually the $v$-representability problem from Section III shows up at this point. A purported solution [49; 50; 51] rests on an ill-founded notion of differentiability where the functionals are extended to distributions, but with an incorrect application of the calculus of distributions (see, e.g., Eq. (0.24) in Gonis [49]). The usual rationale of DFT is to assume that the potentials from Eq. (19) exist and are unique (modulo a constant; after all the latter is the content of the HK theorem). The difference $v_{\mathrm{Hxc}}=v_{s}-v$ is then known as the Hartree-exchange-correlation potential: what needs to be added to the fixed external potential $v$ in order to simulate all interactions in an non- interacting system. Note that such missing effects from interactions do not stem exclusively from the $W$-term in $F^{1}$, but also from the different kinetic energy contributions between the interacting and non-interacting system. Nevertheless, the usual understanding is that most of the kinetic energy contributions can already be captured by a non-interacting system (with an uncorrelated wave function) and that they thus practically cancel between $F^{1}$ and $F^{0}$ when we calculate $v_{\mathrm{Hxc}}[\rho]=v_{s}-v\in\underline{\partial}F^{1}[\rho]-\underline{\partial}F^{0}[\rho].$ (20) At this point a problematic discrepancy is introduced, since the subdifferential is not linear and thus $v_{\mathrm{Hxc}}$ and $\underline{\partial}F_{\mathrm{Hxc}}$ need not match. If $v_{\mathrm{Hxc}}$ cannot be determined as $\underline{\partial}F_{\mathrm{Hxc}}$ we are left with the necessity of individually solving the inverse problems $\rho\mapsto v$ and $\rho\mapsto v_{s}$ in Eq. (20) for both systems, interacting and non- interacting. In practice this means one cannot benefit from finding good approximations to $F_{\mathrm{Hxc}}$ which are the most important elements of applied DFT. A possible circumvention lies in a conceptual shift from describing a system in terms of energies to forces. The ground state is then characterized by a certain force-balance equation that can be equally found in non-equilibrium settings, just with an additional dynamical term [52; 53]. At a density that is simultaneously interacting and non-interacting $v$-representable and where the wave function has a sufficient regularity, the force-balance equation can be employed to derive $v_{\mathrm{Hxc}}$ as the solution of a Poisson equation instead of a functional derivative [54]. An alternative derivation for this was already given earlier using line integrals describing the work it takes to move an electron from infinity against the force field of the exchange- correlation hole charge [55; 56]. Yet, we will proceed here for the sake of argument by _assuming_ differentiability for now. Since the functional derivative $\frac{\delta}{\delta\rho}$ is linear and it holds $\displaystyle v_{\mathrm{Hxc}}[\rho]$ $\displaystyle=v_{s}-v=-\frac{\delta}{\delta\rho}F^{0}[\rho]+\frac{\delta}{\delta\rho}F^{1}[\rho]$ (21) $\displaystyle=\frac{\delta}{\delta\rho}(F^{1}[\rho]-F^{0}[\rho])=\frac{\delta}{\delta\rho}F_{\mathrm{Hxc}}[\rho].$ Also, several important properties that the Hxc potential needs to have are automatically fulfilled when they are functional derivatives [57], which is especially relevant for functional approximations to $v_{\mathrm{Hxc}}$. The Kohn–Sham scheme is now introduced in order to find an unknown ground- state density ${\rho_{\mathrm{gs}}}$ of an interacting system by starting from an initial guess $\rho_{0}$ and by using $v_{\mathrm{Hxc}}$ (in practice a suitable approximation to it) as the connection between the interacting system and a non-interacting reference system. To this end, rewrite Eq. (19) with assumed differentiability as $\frac{\delta}{\delta\rho}F^{1}[{\rho_{\mathrm{gs}}}]+v=0$ and $\frac{\delta}{\delta\rho}F^{0}[{\rho_{\mathrm{gs}}}]+v_{s}=0$ and set the two equations equal, $\frac{\delta}{\delta\rho}F^{1}[{\rho_{\mathrm{gs}}}]+v=\frac{\delta}{\delta\rho}F^{0}[{\rho_{\mathrm{gs}}}]+v_{s}.$ Now, apart from the fixed external potential $v$ of the interacting system, all variables in this equation still remain generally unknown: the effective potential of the non-interacting system $v_{s}$ and, especially, the density $\rho$ of both systems that we would like to determine. The trick lies in introducing sequences $\rho_{i}\to{\rho_{\mathrm{gs}}}$, $v_{i}\to v_{s}$ and define an update rule, $v_{i+1}=v+\frac{\delta}{\delta\rho}F^{1}[\rho_{i}]-\frac{\delta}{\delta\rho}F^{0}[\rho_{i}]=v+v_{\mathrm{Hxc}}[\rho_{i}].$ (22) We see immediately that if $\rho_{i}$ has converged to the correct ground- state density ${\rho_{\mathrm{gs}}}$ of the interacting system, then $v+\frac{\delta}{\delta\rho}F^{1}[\rho_{i}]=0$ and the remaining equation tells us that indeed $v_{i+1}$ is the potential that gives the same density ${\rho_{\mathrm{gs}}}$ in the non-interacting system. The next step after Eq. (22) in the Kohn–Sham iteration lies in determining the density $\rho_{i+1}$ that comes from $v_{i+1}$ in the non-interacting system (which is comparably easy achieved by solving the corresponding Kohn–Sham equation (18)) and then iterate. Convergence problems are a big issue within this iteration scheme and have been dealt with by either damping the iteration step from $\rho_{i}\to\rho_{i+1}$ to $\rho_{i}\to\rho_{i}+\mu(\rho_{i+1}-\rho_{i})$, $\mu\in(0,1)$, or mixing several of the previous steps $\\{\rho_{i}\\}$ into the result $\rho_{i+1}$ [58; 59; 60]. Guaranteed convergence has been studied and proven for the finite-lattice case [10; 11; 12] by combining an optimal damping step and a regularization technique [4; 12], the latter truly making $F$ differentiable and $E$ a strictly concave functional. This solves the problem of defining $v_{\mathrm{Hxc}}$ in Eq. (21) and yields a curvature bound on $F$ that is needed for guaranteed convergence. The regularization method is briefly explained in Section IX below. For the Kohn–Sham iteration in continuum DFT the convergence is still an open problem, a direct generalization of the finite-lattice case has been found to be insufficient [61]. In practical applications that suffer from convergence issues, imaginary-time propagation in time-dependent DFT has recently been found as a viable alternative to find a Kohn–Sham ground state [62]. ## IX Density-potential mixing and regularized DFT The full HK theorem guarantees a unique inversion from densities to potentials, but the whole discussion, especially regarding the necessary conditions for showing HK2, probably already made us a little bit sceptical about its validity in different settings. We will thus introduce a method that always guarantees a bijective mapping, not between densities and potentials, but between _quasidensities_ (called _pseudo-densities_ in the original work on regularization [4]) and potentials. The basic idea is simple: If for some reason we cannot guarantee a unique (injective) mapping from potentials to ground-state densities $v\mapsto\rho[v]$, meaning that different $v\neq v^{\prime}$ map to the same $\rho[v]=\rho[v^{\prime}]$, then let us try it for $v\mapsto\rho_{\varepsilon}[v]=\rho[v]-\varepsilon v$, where at least in the previous example we would have $\rho_{\varepsilon}[v]\neq\rho_{\varepsilon}[v^{\prime}]$ for sure. One could argue that this could just as easily introduce new problems for injectivity, like having $v\neq v^{\prime}$ such that $\rho_{\varepsilon}[v]=\rho_{\varepsilon}[v^{\prime}]$, but we will show in the following that this cannot be the case for the functionals considered here. Remember that the mapping $v\mapsto\rho[v]$ can be defined by the superdifferential of $E$, $\rho[v]=\overline{\partial}E[v]$, as explained in Eq. (17). So what is the corresponding functional $E_{\varepsilon}$ such that $\rho_{\varepsilon}[v]=\rho[v]-\varepsilon v=\overline{\partial}E_{\varepsilon}[v]$? The superdifferential retains the linear nature of a derivative if only concave functionals are added, so we can look for a convex functional $\phi$ such that $\overline{\partial}(-\phi)[v]=-\underline{\partial}\phi[v]=-v$. In a general space, such a question proves hard [61], but it is easy to see that in the usual space $L^{2}$ of square-integrable functions the norm square gives exactly what we need, $\phi[v]=\tfrac{1}{2}\|v\|^{2}=\tfrac{1}{2}\langle v,v\rangle$. In any case, we have established $E_{\varepsilon}=E-\varepsilon\phi$ and $\rho_{\varepsilon}[v]=\overline{\partial}E_{\varepsilon}[v]$ with such a convex $\phi$. But in many cases, not only for the mentioned $L^{2}$ space, the functional $\phi$ is not only convex, but _strictly convex_ , meaning that any local minimizer is not only global but even unique. But this feature transfers to $E_{\varepsilon}$ if $-\varepsilon\phi$, as a _strictly concave_ functional, is added to $E$. Consequently, $E_{\varepsilon}$ is also strictly concave and any maximizing potential in $F_{\varepsilon}[x]=\sup_{v}\\{E_{\varepsilon}[v]-\langle v,x\rangle\\}$ (23) is necessarily unique (not just up to a constant). This means we can always uniquely map $v\mapsto x=\rho[v]-\varepsilon v$ and back. We wrote $x$ now to make clear that this is a quasidensity, a mixture between a density and its associated potential. As such it is neither necessarily normalized nor positive, just a general element of the density space, $x\in X$. By what we learned in Section VII, the _quasidensity-potential mapping_ can also be directly defined by $-\underline{\partial}F_{\varepsilon}[x]=\\{v\\}$ for all $x$ without any “$v$-representability” restriction for $x$. Consequently, the mapping is defined for all $x$ in the density space $X$ and thus bijective. The whole maneuver of passing from $F$ to $F_{\varepsilon}$ corresponds to a regularization strategy called Moreau–Yosida regularization [4; 12] by which not only the concave $E$ transforms into a strictly concave $E_{\varepsilon}$, but also the $F_{\varepsilon}$ defined by Eq. (23) is finally differentiable if the spaces $X,X^{*}$ have some additional properties [5, Theorem 9]. The only problem is that this requires the space $X$ to be reflexive, which it is not in our current formulation as introduced in Section III, since it includes the non-reflexive $L^{1}$ in its definition. So a different choice for the basic spaces, like $X=L^{2}$ on a bounded domain [4] or $X=L^{3}$ as a larger alternative to our space[5] has to be taken. This section demonstrated how such a regularization that facilitates a unique (quasi)density-potential mapping can be used to fully circumvent any reference to the HK theorem. But to avoid confusion we will _not_ say that in a regularized setting the HK theorem “holds” even though a unique and well- defined (quasi)density-potential mapping exists. It is interesting to note that the popular Zhao–Morrison–Parr method for density-potential inversion already implicitly employs Moreau–Yosida regularization and a limit procedure $\varepsilon\to 0$ [63]. ## X Abstract density-potential mapping The presented form of HK1 allows for an abstraction and thereby for generalizations. Therein, the density is generalized to any system-inherent quantity that seems suitable to describe other system parameters that we are interested in. This could be the density together with the spin density, a current-quantity etc. On the other side, we select a generalized form of the potential that enters the Hamiltonian and that is able to steer the “density- quantity” by coupling to it. Such a framework was developed in Laestadius _et al._ [5], building on Banach spaces and their duals for density and potential quantities. This enables us to employ the regularization technique from Section IX to obtain a well-defined Kohn–Sham iteration scheme. In order to be more concrete, let $\mathbf{x}$ be the density quantity describing a state that will in general include many components, like different densities, currents etc., and $\mathbf{v}$ the collection of external potentials acting on them. At this point we do not even assume that $\mathbf{x}$ and $\mathbf{v}$ have the same number or type of components like a dual structure between densities and potentials would impose. Instead of a linear pairing $\langle\mathbf{v},\mathbf{x}\rangle$ for the coupling to the external potential we can introduce an arbitrary functional $f[\mathbf{v},\mathbf{x}]$. Then the _only_ necessary condition left for an abstract HK1 is that the ground-state energy expression has the form $\displaystyle\tilde{F}[\mathbf{x}]=\inf_{\psi\mapsto\mathbf{x}}\left\\{\langle\psi|H_{0}|\psi\rangle\right\\},$ $\displaystyle E[\mathbf{v}]=\inf_{\mathbf{x}}\\{\tilde{F}[\mathbf{x}]+f[\mathbf{v},\mathbf{x}]\\}.$ (24) Since $\tilde{F}[\mathbf{x}]$ is independent of $\mathbf{v}$, the critical argument in the first proof of HK1 still holds and thus two potentials that share a common $\mathbf{x}$ in the ground state will also share a common ground-state wave function or density matrix. Consequently, HK1 is secured in any such formulation of DFT, while the situation for HK2 quite generally is more problematic. Even if the coupling between $\mathbf{v}$ and $\mathbf{x}$ that enters the energy functional in Eq. (24) is linear like in $f[\mathbf{v},\mathbf{x}]=\langle\mathbf{v},\mathbf{x}\rangle$, the critical step (8) in the proof of HK2 will involve more degrees-of-freedom on the potential side and the argument may fail. In the literature, the presented situation with linear coupling corresponds to what Schönhammer, Gunnarsson, and Noack [64] call $\\{a\\}$-functional theory. Similarly, Higuchi and Higuchi [65; 66] allow for a more general choice of basic variables in DFT next to the usual density. Xu _et al._ [67] derived conditions that need to be fulfilled to also have a HK2 in such a general setting. One can then try and extend DFT and the Kohn–Sham scheme systematically to predict further system parameters, if good approximative functionals can be found. A first example would be the spin-resolved functional that has the usual one- particle density $\rho=\rho_{\uparrow}+\rho_{\downarrow}$ and the spin-density $\rho_{\uparrow}-\rho_{\downarrow}$ as basic variables, $\mathbf{x}=(\rho_{\uparrow}+\rho_{\downarrow},\rho_{\uparrow}-\rho_{\downarrow})$. An alternative possible choice would clearly be $\mathbf{x}=(\rho_{\uparrow},\rho_{\downarrow})$ [18, Section 8.1]. The energy functional is $E[v]=\inf_{\mathbf{x}}\\{\tilde{F}[\mathbf{x}]+\langle v,\rho\rangle\\}$, with $v$ just the usual scalar potential that couples to the one-particle density $\rho$. The involved spaces for densities and potentials are not dual in this example, since they involve a different number of components. But by choosing an $F_{\mathrm{Hxc}}[\mathbf{x}]$ that depends on the spin-resolved density, the Hxc-potential as its derivative (and with it the effective potential of the Kohn–Sham system) must be from the dual space of $\mathbf{x}$ and thus include components that act on the different spin- components individually. A second example is CDFT and its variants that will be thoroughly discussed in Part II of this review. The paramagnetic current density of a given state $\psi\in\mathcal{W}$ is defined as $\mathbf{j}^{\mathrm{p}}_{\psi}(\mathbf{r}_{1})=N\sum_{\underline{\sigma}}\int_{\mathbb{R}^{3(N-1)}}\mathrm{Im}\left\\{\psi^{*}\nabla_{1}\psi\right\\}\,\mathrm{d}\mathbf{r}_{\perp}.$ Then the amended density quantity is $\mathbf{x}=(\rho,\mathbf{j}^{\mathrm{p}})$ which couples linearly to $\mathbf{v}=(v+\frac{1}{2}|\mathbf{A}|^{2},\mathbf{A})$ [68]. Since by this the potential-energy contribution amounts exactly to the linear pairing $f[\mathbf{v},\mathbf{x}]=\langle\mathbf{v},\mathbf{x}\rangle$ that allows to define a potential-independent constrained-search functional, HK1 holds. This means one can continue along the lines started in this work and try to generalize many concepts and results from above to such extended DFTs. This includes the definition of representable densities (Section III), different functionals (Section VI), functional differentiability (Section VII), setting up a Kohn–Sham scheme (Section VIII), as well as regularisation (Section IX), since also there the existence of a full HK theorem was hardly ever assumed. $\begin{array}[]{ccccccc}&\lx@intercol\hfil\text{$F$ differentiable}\hfil\lx@intercol&&&\text{$F_{\varepsilon}$ from}\\\ &\lx@intercol\hfil\text{at $\rho\in v\text{-}\mathsf{rep}$}\hfil\lx@intercol&&&\text{regularization}\\\ &&\Downarrow&&&&\Downarrow\\\ -\underline{\partial}F[\rho]=\\{v+c\\}&\hbox{\multirowsetup$\Longleftrightarrow$}&\hbox{\multirowsetup HK}&\hbox{\multirowsetup$\Longleftrightarrow$}&\text{unique density-}&\hbox{\multirowsetup$\Longrightarrow$}&\text{unique \emph{quasi}density-}\\\ \text{at $\rho\in v\text{-}\mathsf{rep}$}&&&&\text{potential mapping}&&\text{potential mapping}\\\ &&\Updownarrow\\\ &\lx@intercol\hfil$\downbracefill$\hfil\lx@intercol\\\ &\text{HK1}&&\text{HK2}\\\ &\Uparrow&&\Uparrow\\\ \lx@intercol\hfil$\text{potential- independent}$\hfil\lx@intercol&\lx@intercol$\text{UCP result}$\hfil\lx@intercol\\\ \lx@intercol\hfil$\text{constrained-search fctl.}$\hfil\lx@intercol\end{array}$ Figure 2: Logical implications between the different statements relating to a “unique density-potential mapping” and the HK theorem in standard DFT. ## XI Summary We will give a brief summary of the structure of the density-potential mapping and its relation to the HK theorem. Following the last section on abstract DFTs, at least the HK1 result does not only hold for standard DFT (that maps one-particle densities to scalar potentials), but it holds for all variants of DFTs that offer the required structures. This will be especially useful with foresight towards CDFT, the topic of the second part of this review. In standard DFT, with a setting that yields the unique-continuation property that in turn prevents the ground-state density from being zero on a set of non-zero measure (Section V) and due to the simple relation (8) in the proof of HK2, a full HK result can be established. In any higher DFT this proof strategy potentially fails. The status of HK1, on the other hand, is much less critical, since this result holds automatically whenever a _potential- independent_ (“universal”) constrained-search functional can be set up. But also in cases where the constrained-search functional depends on the external potential, a valid statement like in the HK theorem, that two potentials that share a common ground-state density are equal up to gauge changes, is still possible in general. The more general way how to think and talk about a HK result is by calling it a “unique density-potential mapping” and we explained how such a mapping can be established as the subdifferential of the density functional $F$ at $v$-representable densities. If the potentials in the resulting subdifferential are equal up to a gauge transformation, then this is just the HK result again. Assuming full differentiability of $F$ implies a one-element subdifferential, so there would not even be any room for gauge changes, and a unique density-potential mapping would be the result once more. This property of differentiability of the density functional $F$ is desirable also in the context of Kohn–Sham theory in order to be able to link the functional $F_{\mathrm{Hxc}}$ to the Hxc potential like in Eq. (21). But since differentiability is _not_ a property of the usual DFTs, a regularization strategy was devised and briefly explained in Section IX. This yields a unique _quasi_ density-potential mapping, where quasidensities are actually mixtures between ground-state densities and their potentials. The mixing parameter $\varepsilon$ could be set to zero to retrieve the unregularized theory together with the problem of non-differentiability. The whole structure is laid out diagrammatically in Figure 2. ## XII Outlook In this outlook, we first want to collect the problems that still remain open within the foundations of standard DFT and that will surely be the topic in upcoming works. Considering Lieb’s mathematical formulation of DFT, summarized above in Section III and VI, there are two main issues. Firstly, HK2 is guaranteed only for eigenstates that are non-zero almost everywhere, a property that is secured by the UCP explained in Section V. But the potential space required for this does not cover all potentials from the Lieb setting and a sufficiently general UCP result is not available to date. Secondly, the issue of $v$-representability, explained in Section III, still remains open. While regularization as described in Section IX formally allows us to circumvent this problem, it has not yet been put to practical use. Since the overlap between interacting and non-interacting $v$-representability is poorly understood, this has direct implications for Kohn–Sham theory. But even with $v$-representability assumed, convergence of the Kohn–Sham self-consistent field iterations in the standard setting is still an open problem. Both issues, availability of UCP and $v$-representability, relate to the function spaces for densities and potentials. Possibly, with a more refined choice of these spaces, full $v$-representability or even differentiability of $F$ might be achievable. However, it also cannot be ruled out that non-differentiability is fundamental to DFT. This non-differentiability of $F$, that has been repeatedly stressed in this work, implies that the exchange-correlation potential cannot be found as a functional derivative with respect to the density, as it is usually assumed in standard DFT. Orbital-dependent functionals [69] can be formally viewed as relying on the HK1 map $\rho\mapsto\phi$ to obtain the Kohn–Sham wave function from a density. Non-differentiability of $F[\rho]$ might then be represented in the noninteracting wave function $\phi[\rho]$, which may benefit the functional approximations if they rely directly on the Kohn–Sham orbitals. The lack of differentiability also favors approaches based on forces instead of energies, as mentioned in Section VIII. However, practical functionals that are derived from this approach remain unexplored and there is still a dependence on $v$-representability. It is interesting to note which useful structures of DFT carry over to “higher” density-functional theories, and in Part II we will discuss density- functional theory for systems involving magnetic fields. While one of its flavours, paramagnetic CDFT, already briefly discussed in Section X, still allows for a constrained-search functional (HK1), the realization of a full density-potential mapping is highly problematic. For this reason, in the classical formulation of paramagnetic CDFT [70] the HK2 result that different potentials lead to different ground states was just tacitly assumed with the words: “Let $\psi$ and $\psi^{\prime}$ be the two different ground states corresponding to the two sets of fields [$(v,\mathbf{A})$ and $(v^{\prime},\mathbf{A}^{\prime})$].” Later, Capelle and Vignale [7] even found counterexamples to HK2 which shows that a density-potential mapping cannot be constructed in paramagnetic CDFT. But this clearly does not mean that in different versions of CDFT the density-potential mapping is impossible to achieve in general. A formulation utilizing the total current will be studied as well, but here the constrained-search functional would depend on $\mathbf{A}$ and thus HK1 is not available in the fashion as it was presented here. So while for paramagnetic CDFT the HK2 fails, for total (physical) CDFT already HK1 does not hold. Overall, the existence of a well-defined density- potential mapping in CDFT is still an open issue that will be considered in the second part of this review. ## Acknowledgement EIT, MAC and AL thank the Research Council of Norway (RCN) under CoE (Hylleraas Centre) Grant No. 262695, for AL and MAC also CCerror Grant No. 287906 and for EIT also “Magnetic Chemistry” Grant No. 287950, and MR acknowledges the Cluster of Excellence “CUI: Advanced Imaging of Matter” of the Deutsche Forschungsgemeinschaft (DFG), EXC 2056, project ID 390715994. AL and MAC was also supported by the ERC through StG REGAL under agreement No. 101041487. The authors thank Centre for Advanced Studies (CAS) in Oslo, since this work includes insights gathered at the YoungCAS workshop “Do Electron Current Densities Determine All There Is to Know?”, held July 9-13, 2018, in Oslo, Norway. ## Bibliography ## References * Hohenberg and Kohn [1964] P. Hohenberg and W. Kohn, “Inhomogeneous electron gas,” Phys. Rev. 136, B864–B871 (1964). * Lieb [1983] E. H. Lieb, “Density functionals for Coulomb-systems,” Int. J. Quantum Chem. 24, 243–277 (1983). * Lammert [2007] P. E. Lammert, “Differentiability of Lieb functional in electronic density functional theory,” Int. J. Quantum Chem. 107, 1943–1953 (2007). * Kvaal _et al._ [2014] S. Kvaal, U. Ekström, A. M. Teale, and T. Helgaker, “Differentiable but exact formulation of density-functional theory,” J. Chem. Phys. 140, 18A518 (2014). * Laestadius _et al._ [2018] A. Laestadius, M. Penz, E. I. Tellgren, M. Ruggenthaler, S. Kvaal, and T. Helgaker, “Generalized Kohn–Sham iteration on Banach spaces,” J. Chem. Phys. 149, 164103 (2018). * Laestadius _et al._ [2019] A. Laestadius, M. Penz, E. I. Tellgren, M. Ruggenthaler, S. Kvaal, and T. Helgaker, “Kohn–Sham theory with paramagnetic currents: Compatibility and functional differentiability,” J. Chem. Theory Comput. 15, 4003–4020 (2019). * Capelle and Vignale [2002] K. Capelle and G. Vignale, “Nonuniqueness and derivative discontinuities in density-functional theories for current-carrying and superconducting systems,” Phys. Rev. B 65, 113106 (2002). * Percus [1978] J. Percus, “The role of model systems in the few-body reduction of the N-fermion problem,” Int. J. Quant. Chem. 13, 89–124 (1978). * Levy [1979] M. Levy, “Universal variational functionals of electron densities, first-order density matrices, and natural spin-orbitals and solution of the v-representability problem,” Proc. Natl. Acad. Sci. USA 76, 6062–6065 (1979). * Penz _et al._ [2019] M. Penz, A. Laestadius, E. I. Tellgren, and M. Ruggenthaler, “Guaranteed convergence of a regularized Kohn–Sham iteration in finite dimensions,” Phys. Rev. Lett. 123, 037401 (2019). * Penz _et al._ [2020] M. Penz, A. Laestadius, E. I. Tellgren, M. Ruggenthaler, and P. E. Lammert, “Erratum: Guaranteed convergence of a regularized Kohn–Sham iteration in finite dimensions,” Phys. Rev. Lett. 125, 249902–249902 (2020). * [12] S. Kvaal, “Moreau–Yosida regularization in DFT,” (10 Aug 2022) arXiv e-prints [math.NA] 2208.05268, accessed 2023-01-31. * von Barth [2004] U. von Barth, “Basic density-functional theory—an overview,” Phys. Scr. 2004, 9 (2004). * Burke and friends [2007] K. Burke and friends, “The ABC of DFT,” (2007), accessed 2023-01-31. * Burke [2012] K. Burke, “Perspective on density functional theory,” J. Chem. Phys. 136, 150901 (2012). * Dreizler and Gross [2012] R. M. Dreizler and E. K. Gross, _Density Functional Theory: An Approach to the Quantum Many-body Problem_ (Springer, 2012). * Eschrig [2003] H. Eschrig, _The Fundamentals of Density Functional Theory_ , 2nd ed. (Springer, 2003). * Parr and Yang [1989] R. Parr and W. Yang, _Density Functional Theory of Atoms and Molecules_ (Oxford University Press, 1989). * Teale _et al._ [2022] A. M. Teale, T. Helgaker, A. Savin, C. Adamo, B. Aradi, A. V. Arbuznikov, P. W. Ayers, E. J. Baerends, V. Barone, and P. Calaminici, “DFT exchange: Sharing perspectives on the workhorse of quantum chemistry and materials science,” Phys. Chem. Chem. Phys. 24, 28700–28781 (2022). * Gilbert [1975] T. L. Gilbert, “Hohenberg–Kohn theorem for nonlocal external potentials,” Phys. Rev. B 12, 2111 (1975). * Harriman [1981] J. E. Harriman, “Orthonormal orbitals for the representation of an arbitrary density,” Phys. Rev. A 24, 680 (1981). * Englisch and Englisch [1983] H. Englisch and R. Englisch, “Hohenberg–Kohn theorem and non-V-representable densities,” Physica A Stat. Mech. Appl. 121, 253 – 268 (1983). * Levy [1982] M. Levy, “Electron densities in search of Hamiltonians,” Phys. Rev. A 26, 1200 (1982). * Penz and van Leeuwen [2021] M. Penz and R. van Leeuwen, “Density-functional theory on graphs,” J. Chem. Phys. 155, 244111 (2021). * Garrigue [2021] L. Garrigue, “Some properties of the potential-to-ground state map in quantum mechanics,” Comm. Math. Phys. 386, 1803–1844 (2021). * Barbu and Precupanu [2012] V. Barbu and T. Precupanu, _Convexity and Optimization in Banach Spaces_ , 4th ed. (Springer, 2012). * Garrigue [2022] L. Garrigue, “Building Kohn–Sham potentials for ground and excited states,” Arch. Ration. Mech. Anal. 245, 949–1003 (2022). * Reed and Simon [1975] M. Reed and B. Simon, _II: Fourier Analysis, Self-Adjointness_ , Vol. 2 (Elsevier, 1975). * Lieb and Loss [2001] E. H. Lieb and M. Loss, _Analysis_ (American Mathematical Society, Providence, Rhode Island, 2001). * Pino _et al._ [2007] R. Pino, O. Bokanowski, E. V. Ludeña, and R. L. Boada, “A re-statement of the Hohenberg–Kohn theorem and its extension to finite subspaces,” Theor. Chem. Accounts 118, 557–561 (2007). * Garrigue [2018] L. Garrigue, “Unique continuation for many-body Schrödinger operators and the Hohenberg–Kohn theorem,” Math. Phys. Anal. Geom. 21, 27 (2018). * Tellgren _et al._ [2018] E. I. Tellgren, A. Laestadius, T. Helgaker, S. Kvaal, and A. M. Teale, “Uniform magnetic fields in density-functional theory,” J. Chem. Phys. 148, 024101 (2018). * Kohn, Savin, and Ullrich [2004] W. Kohn, A. Savin, and C. A. Ullrich, “Hohenberg–Kohn theory including spin magnetism and magnetic fields,” Int. J. Quantum Chem. 100, 20–21 (2004). * Kohn [1985] W. Kohn, “Density functional theory: Fundamentals and applications,” in _Highlights of condensed-matter theory_ , Proceedings of the international school of physics “Enrico Fermi”, Course LXXXIX at Varenna on Lake Como, 28 June-16 July 1983, edited by F. Bassani, F. Fumi, and M. P. Tosi (North-Holland, Amsterdam, 1985) pp. 1–15. * Regbaoui [2001] R. Regbaoui, “Unique continuation from sets of positive measure,” in _Carleman Estimates and Applications to Uniqueness and Control Theory_, Progress in Nonlinear Differential Equations and Their Applications, Vol. 46, edited by F. Colombini and C. Zuily (Birkhäuser, Basel, 2001) pp. 179–190. * de Figueiredo and Gossez [1992] D. G. de Figueiredo and J.-P. Gossez, “Strict monotonicity of eigenvalues and unique continuation,” Comm. Par. Diff. Eq. , 339–346 (1992). * Ladyzenskaya and Ural’tzeva [1968] O. A. Ladyzenskaya and N. N. Ural’tzeva, _Linear and Quasilinear Elliptic Equations_ (Academic Press, New York, 1968). * Lammert [2018] P. E. Lammert, “In search of the Hohenberg–Kohn theorem,” J. Math. Phys. 59, 042110 (2018). * Jerison and Kenig [1985] D. Jerison and C. E. Kenig, “Unique continuation and absence of positive eigenvalues for Schrödinger operators,” Ann. Math. 121, 463–488 (1985). * Garrigue [2020] L. Garrigue, “Unique continuation for many-body Schrödinger operators and the Hohenberg-Kohn theorem. II. The Pauli Hamiltonian,” Doc. Math. 25, 869-898 (2020). * Garrigue [2019] L. Garrigue, “Hohenberg–Kohn theorems for interactions, spin and temperature,” J. Stat. Phys. 177, 415–437 (2019). * Kvaal and Helgaker [2015] S. Kvaal and T. Helgaker, “Ground-state densities from the Rayleigh–Ritz variation principle and from density-functional theory,” J. Chem. Phys. 143, 184106 (2015). * [43] M. Lewin, E. H. Lieb, and R. Seiringer, “Universal functionals in density functional theory,” (13 Sep 2022) arXiv e-prints [math-ph] 1912.10424, accessed 2023-01-31. * Tsuneda [2014] T. Tsuneda, _Density functional theory in quantum chemistry_ (Springer, 2014). * Kohn and Sham [1965] W. Kohn and L. J. Sham, “Self-consistent equations including exchange and correlation effects,” Phys. Rev. 140, A1133 (1965). * van Leeuwen [2003] R. van Leeuwen, “Density functional approach to the many-body problem: key concepts and exact functionals,” Adv. Quantum Chem. 43, 43002–5 (2003). * Lammert [2010] P. E. Lammert, “Well-behaved coarse-grained model of density-functional theory,” Phys. Rev. A 82, 012109 (2010). * Kümmel and Perdew [2003] S. Kümmel and J. P. Perdew, “Optimized effective potential made simple: Orbital functionals, orbital shifts, and the exact Kohn–Sham exchange potential,” Phys. Rev. B 68, 035103 (2003). * Gonis [2014] A. Gonis, “Functionals and functional derivatives of wave functions and densities,” World J. Condens. Matter Phys. 2014 (2014), 10.4236/wjcmp.2014.43022. * Gonis _et al._ [2016] A. Gonis, X.-G. Zhang, M. Däne, G. M. Stocks, and D. M. Nicholson, “Reformulation of density functional theory for n-representable densities and the resolution of the v-representability problem,” J. Phys. Chem. Solids 89, 23–31 (2016). * Gonis [2019] A. Gonis, “Is an interacting ground state (pure state) v-representable density also non-interacting ground state v-representable by a Slater determinant? in the absence of degeneracy, yes!” Phys. Lett. A 383, 2772–2776 (2019). * Tchenkoue _et al._ [2019] M.-L. M. Tchenkoue, M. Penz, I. Theophilou, M. Ruggenthaler, and A. Rubio, “Force balance approach for advanced approximations in density functional theories,” J. Chem. Phys. 151, 154107 (2019). * Stefanucci and van Leeuwen [2013] G. Stefanucci and R. van Leeuwen, _Nonequilibrium Many-Body Theory of Quantum Systems: A Modern Introduction_ (Cambridge University Press, 2013). * [54] M. Ruggenthaler, N. Tancogne-Dejean, A. Laestadius, M. Penz, M. A. Csirik, and A. Rubio, “Exchange energies with forces in density-functional theory,” (31 Mar 2022) arXiv e-prints [physics.chem-ph] 2203.16980, accessed 2023-01-31. * Harbola, Slamet, and Sahni [1991] M. K. Harbola, M. Slamet, and V. Sahni, “Local exchange-correlation potential from the force field of the Fermi–Coulomb hole charge for non-symmetric systems,” Phys. Lett. A 157, 60–64 (1991). * Slamet, Sahni, and Harbola [1994] M. Slamet, V. Sahni, and M. K. Harbola, “Force field and potential due to the fermi-coulomb hole charge for nonspherical-density atoms,” Phys. Rev. A 49, 809 (1994). * Gaiduk and Staroverov [2009] A. P. Gaiduk and V. N. Staroverov, “How to tell when a model Kohn–Sham potential is not a functional derivative,” J. Chem. Phys. 131, 044107 (2009). * Pulay [1980] P. Pulay, “Convergence acceleration of iterative sequences. the case of scf iteration,” Chem. Phys. Lett. 73, 393–398 (1980). * Cancès and Le Bris [2000] E. Cancès and C. Le Bris, “Can we outperform the DIIS approach for electronic structure calculations?” Int. J. Quantum Chem. 79, 82–90 (2000). * Cancès, Kemlin, and Levitt [2021] E. Cancès, G. Kemlin, and A. Levitt, “Convergence analysis of direct minimization and self-consistent iterations,” SIAM Journal on Matrix Analysis and Applications 42, 243–274 (2021). * [61] M. Penz and A. Laestadius, “Convergence of the regularized Kohn–Sham iteration in Banach spaces,” (1 Oct 2020) arXiv e-prints [math-ph] 2003.05389, accessed 2023-01-31. * Flamant _et al._ [2019] C. Flamant, G. Kolesov, E. Manousakis, and E. Kaxiras, “Imaginary-time time-dependent density functional theory and its application for robust convergence of electronic states,” J. Chem. Theory Comput. 15, 6036–6045 (2019). * [63] M. Penz, M. A. Csirik, and A. Laestadius, “Density-potential inversion from Moreau–Yosida regularization,” (27 Dec 2022) arXiv e-prints [physics.chem-ph] 2212.12727, accessed 2023-01-31. * Schönhammer, Gunnarsson, and Noack [1995] K. Schönhammer, O. Gunnarsson, and R. M. Noack, “Density-functional theory on a lattice: Comparison with exact numerical results for a model with strongly correlated electrons,” Phys. Rev. B 52, 2504–2510 (1995). * Higuchi and Higuchi [2004a] M. Higuchi and K. Higuchi, “Arbitrary choice of basic variables in density functional theory: Formalism,” Phys. Rev. B 69, 035113 (2004a). * Higuchi and Higuchi [2004b] K. Higuchi and M. Higuchi, “Arbitrary choice of basic variables in density functional theory. II. Illustrative applications,” Phys. Rev. B 69, 165118 (2004b). * Xu _et al._ [2022] L. Xu, J. Mao, X. Gao, and Z. Liu, “Extensibility of Hohenberg–Kohn theorem to general quantum systems,” Adv. Quantum Techn. 5, 2200041 (2022). * Vignale and Rasolt [1987] G. Vignale and M. Rasolt, “Density-functional theory in strong magnetic fields,” Phys. Rev. Lett. 59, 2360–2363 (1987). * Kümmel and Kronik [2008] S. Kümmel and L. Kronik, “Orbital-dependent density functionals: Theory and applications,” Rev. Mod. Phys. 80, 3 (2008). * Vignale, Rasolt, and Geldart [1990] G. Vignale, M. Rasolt, and D. Geldart, “Magnetic fields and density functional theory,” Adv. Quantum Chem. 21, 235–253 (1990).
longtable # Understanding Complex Patterns in Social, Geographic, and Economic Inequities in COVID-19 Mortality at the County Level in the US Using Generalized Additive Models Christian Testa1,* (November 29, 2022) ###### Abstract I present three types of applications of generalized additive models (GAMs) to COVID-19 mortality rates in the US for the purpose of advancing methods to document inequities with respect to which communities suffered disproportionate COVID-19 mortality rates at specific times during the first three years of the pandemic. First, GAMs can be used to describe the changing relationship between COVID-19 mortality and county-level covariates (sociodemographic, economic, and political metrics) over time. Second, GAMs can be used to perform spatiotemporal smoothing that pools information over time and space to address statistical instability due to small population counts or stochasticity resulting in a smooth, dynamic latent risk surface summarizing the mortality risk associated with geographic locations over time. Third, estimation of COVID-19 mortality associations with county-level covariates conditional on a smooth spatiotemporal risk surface allows for more rigorous consideration of how socio-environmental contexts and policies may have impacted COVID-19 mortality. Each of these approaches provides a valuable perspective to documenting inequities in COVID-19 mortality by addressing the question of which populations have suffered the worst burden of COVID-19 mortality taking into account the nonlinear spatial, temporal, and social patterning of disease. Abbreviations used: United States (US), Coronavirus Disease 2019 (COVID-19), Generalized Additive Model (GAM), Centers for Disease Control and Prevention (CDC), Index of Concentration at the Extremes (ICE), Confidence Interval (CI) 1 Department of Social and Behavioral Sciences, Harvard T.H. Chan School of Public Health * Correspondence: Christian Testa<EMAIL_ADDRESS> ## Introduction As we enter the third winter with the novel coronavirus disease COVID-19 in the United States, evidence documenting the intense disparities in COVID-19 mortality rates comparing socially advantaged and disadvantaged populations continues to mount. Eliminating inequities in health outcomes has been stated as a major policy goal of the Biden administration (The White House, 2021, 2022), representing a revitalized commitment to health equity and underlining the importance of adequate data reporting systems that report timely estimates of prevalent health inequities. To this end, I use generalized additive models (GAMs) as a flexible regression framework to illustrate the evolving roles and relationships sociodemographic, geographic, and economic conditions have with respect to trends in COVID-19 mortality. The code, data, and documentation necessary to reproduce the analyses contained in this paper are online and free to access at https://github.com/ctesta01/covid.gradient.estimation. #### Background Having passed over 1 million COVID-19 deaths in the United States in 2022 (Donovan, 2022), and facing uncertain prospects for the third COVID-19 winter looming even as new iterations on the COVID-19 vaccines become available, it remains critical that inequities in COVID-19 outcomes are documented and analyzed to reckon with the unjust and unfair burden of preventable illness. Even though the first vaccines were granted emergency use authorization by the U.S. Food and Drug Administration in 2020 (Mayo Clinic, 2022), with the first shots going in arms in December 2020, COVID-19 is still continuing to cause hundreds of deaths a day in the US in the fall of 2022 (“United States COVID - Coronavirus Statistics - Worldometer,” 2022). The new bivalent vaccines released at the end of August 2022 contain mRNA sequences from both the original strain as well as the recently emergent BA.4 and BA.5 lineages in an effort to make the nation’s immunity more up-to-date and robust against the myriad of phylogenetic directions the COVID-19 virus is evolving to explore (Office of the Commissioner, 2022). Despite the updated bivalent boosters representing a significant step forward in prevention strategy, less than 4% of eligible Americans had taken the booster in the first month after it became available (Bendix, 2022; Lambert, 2022). As such, and with an enduring history of inequities in health care access in the US (Bailey et al., 2021; Blendon et al., 2002; Carpenter, 2021; Chrisler et al., 2016; Feldman et al., 2021; Okonkwo et al., 2021; Ortega & Roby, 2021; Rapp et al., 2022; Whitehead et al., 2016), it is clear that without further intervention not all communities will be equally able to benefit from the updated vaccines and inequities in COVID-19 illness and mortality may persist despite the technological innovations in vaccine technology. #### The Role of Geography in COVID-19 Figure 1: Estimates of monthly COVID-19 mortality rates per 100,000 person- years by county organized by Census Division. For each division, the median trendline and quantile ranges are shown weighted by county population size. Prior literature has demonstrated that the spread and impact of COVID-19 has varied geography over time reflecting dynamics in the timing of introduction and transmission events. Methods employed to highlight the geographic patterning in COVID-19 outcomes have included quantile regression (Sigler et al., 2021), Besag-York-Molli'e mixed models (Whittle & Diaz-Artiles, 2020), spatial cluster analysis (Sugg et al., 2021), geographically weighted regression (Mollalo et al., 2020; Park et al., 2021), and others. Figure 1 shows the monthly COVID-19 mortality rates for the counties grouped within each of the nine U.S. Census Divisions (U.S. Department of Commerce Economics and Statistics Administration & U.S. Census Bureau, 2000). The figure summarizes each division’s median mortality rates weighted by county population size. Notably, the mortality associated with the early surge of cases starting in New York City and spreading through New York, New Jersey, and Massachusetts is visible in the Middle Atlantic and New England division figures. The figure also illustrates how the first peak in the mortality time- series for states in the Midwest (West North Central, East North Central) occurred later, in late 2020 and going into early 2021. In the US context, one of the key aspects to the geographic story of COVID-19’s spread and diffusion was the early surge of cases and epicenter in New York City during March 2020 (Thompson, 2020) followed by subsequent waves of cases in the South and Midwest (Glenza, 2020; Scott, 2020; Shumaker & Wu, 2020). As Park et al. stated summarized the trends in the US from March 2020 to May 2021, “hot spots have shifted from densely populated cities and the states with a high percentage of socially vulnerable individuals to the states with relatively relaxed social distancing requirements, and then to the states with low vaccination rates” (2021). When considering the drivers of the COVID-19 pandemic, it’s necessary to note that geography and social conditions are inextricably linked. In July 2021, the CDC reported that “the COVID-19 cumulative death rate in non-metropolitan areas has exceeded that of metropolitan areas since December 2020,” noting that of the approximately 1/5th of Americans who live in rural areas, many “are considered highly vulnerable according to CDC’s Social Vulnerability Index (SVI), which includes factors such as housing, transportation, socioeconomic status, race, and ethnicity” (CDC, 2021). Moreover, rural communities often have lower health insurance rates, higher disability rates, older populations, and limited access to health care. One of CDC’s Morbidity and Mortality Weekly Reports found that vaccination against COVID-19 was lower in rural communities than in urban communities between December 2020 and April 2021 (Murthy, 2021). #### The Social Determinants of COVID-19 Mortality Even since the beginning of the COVID-19 outbreak in the US, data reflected sharp inequities in mortality rates. During January 22nd to May 5th 2020, county COVID-19 mortality rates were 4.94 (95% CI 4.78, 5.09) times higher in counties in the highest quintile of percent People of Color (61%-100%) compared to counties in the lowest quintile of percent People of Color (0%-17.2%) (Chen & Krieger, 2021). This was not wholly unanticipated: as COVID-19 was beginning to take off in the US, some were already calling attention to the fact that COVID-19 threatened to exacerbate existing disparities (Kim et al., 2020). Kim, Marrast, and Conigliaro noted at least three structural barriers in COVID-19 prevention and care: 1) originally requiring residents to have a doctor’s prescription for a COVID-19 test reduced the opportunity for healthcare for People of Color as they are less likely to have a primary care provider; 2) drive-through testing made testing disadvantaged those who rely on public transportation; and 3) quarantining at home while waiting the 7-10 days originally required for COVID-19 test results to come back posed an economic and social challenge that many in already financially difficult situations may not have been able to take on (2020). Others noted yet more reasons why COVID-19 threatened to worsen an already inequitable healthcare landscape in the US: in particular, those who reside in prisons and jails, immigrants and undocumented people, people with disabilities, and people experiencing homeless all face additional challenges in seeking and getting the healthcare they deserve (Okonkwo et al., 2021). Even though stay-at-home orders designed to mitigate spread that were prevalent in many states (Moreland et al., 2020), workers functioning in capacities essential to the functioning of critical infrastructure operations (later termed “essential workers”) were exposed to heightened risk of COVID-19 transmission (Hanage et al., 2020; National Bureau of Economic Research, 2021; National Conference of State Legislatures, 2021; The Lancet, 2020; Wei et al., 2022). It’s clear that those with more structurally enfranchised privileges have been more able to mitigate their risk of negative health outcomes associated with COVID-19 — during February 1 to April 1, 2020, New York City residents who lived in more affluent neighborhoods were more likely to have left the city, while New York City residents from more Black and Hispanic neighborhoods were more likely to continue working (Coven & Gupta, 2020). > Contrary to the oft used phrase that the ‘virus does not discriminate’, the > data presented here suggest that this virus, as many other infectious > diseases, has the greatest implications for the most vulnerable people. The > intersections between health and human rights are clear—the health of a > society and vulnerability to a pandemic are directly related to its human > rights track record for those who are marginalised. (Okonkwo et al., 2021) When vaccines became available, vaccine appointments were often only available to be scheduled through online web-portals contributing to the inequities between those who had internet access and technological literacy and those who didn’t (Press et al., 2021). Vaccination sites have not been equally distributed and areas determined to be vaccine deserts have been found to have disproportionately more Black and Hispanic residents (Rader et al., 2021). In fact, healthcare facilities in counties with higher Black composition had 32% (95% CI 14%-47%) lower odds of serving as vaccine sites (Hernandez et al., 2022). What vaccination has been administered hasn’t suddenly erased the unequal burden of COVID-19 either; in August 2022 the New York Times was reporting “Black death rates at this winter’s peak were greater than those of white people by 34 percent in rural areas, 40 percent in small or medium cities and 57 percent in big cities and their suburbs” (Goldstein, 2022). As COVID-19 case and mortality rates have waxed and waned, the inequities have widen and shrunken, often with racial/ethnic inequities growing during times when COVID-19 rates have surged (Hill & Artiga, 2022). During 2022, the age- standardized COVID-19 mortality rates for white people have, at times, been slightly higher than those of Black and Hispanic people, predominantly because the mortality rate among white people has increased (Johnson & Keating, 2022). It’s important to note that white COVID-19 mortality rates overtaking the Black mortality rates does not imply that equity has been established: neither does this undo the cumulative impact of mortality (which has been twice as high for Black people compared to white people (Hill & Artiga, 2022)), nor does it imply the underlying systemic barriers to equity have been overturned (Del Rios et al., 2022). As Del Rios, Chomilo, and Lewis note, instead, COVID-19 leaves in its wake more years of life expectancy lost, wages lost, and degradation of mental health in Communities of Color. ## Methods ### Data Sources The following variables were retrieved at the county level: * • Counts of COVID-19 deaths (The New York Times, 2021). * • Population size estimates for 2020 from the U.S. Census (US Census Bureau, 2021). * • Median age, median household income, racial/ethnic composition, population density, percent below the federal poverty line, and number of households with high ($100k+)/low (<$25k) household income by racial/ethnic group from the 2014-2019 5-year American Community Survey (US Census Bureau, 2020) through the tidyverse R package (Walker & Herman, 2022). * • Votes cast in the 2020 presidential election (MIT Election Data and Science Lab, 2022) ### Generalized Additive Models Generalized additive models (GAMs) improve upon generalized linear models by allowing for the fitting of smooth functions that transform right-hand-side variables. This is a convenient means to account for nonlinear relationships between the outcome and predictor variables. Whereas a generalized linear model may look like $g(\mu_{i})=\beta_{0}+\beta_{1}x_{i1}+\beta_{2}x_{i2}+\beta_{3}x_{i3}...$ a generalized additive model could look like $g(\mu_{i})=\mathbf{A}_{i}\mathbf{\theta}+f_{1}(x_{i1})+f_{2}(x_{i2})+f_{3}(x_{i3},x_{i4})+\dots$ where the expected value of the outcome is given as $\mu_{i}\equiv\mathbb{E}(Y_{i})$, $\mathbf{A}_{i}$ is a row of the model matrix for any strictly non-parametric model components, $\mathbf{\theta}$ is the corresponding parameter vector, and the $f_{j}$ are smooth functions of the covariates $x_{k}$ (S. N. Wood, 2017). Figure 2: A demonstration of fitting regression models using B-spline basis functions. Part A. shows the simulated dataset. Part B. shows the B-spline basis functions. Part C. shows the weighting of the B-spline basis functions and their weighted sum with an intercept added to create the regression model fit to the data. The smooth functions estimated as part of fitting a GAM are constructed using spline basis functions. These spline basis functions allow for the smooth interpolation of trends in the data allowing for the incorporation of nonlinearity. To avoid overfitting the data, a penalty term is introduced that controls the degree of “wiggliness” or smoothness and this penalty term is fit using generalized cross validation. B-splines are one kind of spline basis function that is commonly used and are especially popular because of their property that they are non-negative on only a finite interval. While a variety of spline-based approaches exist (cubic splines, B- and P-splines, thin-plate splines, etc.), our particular application setting warrants the use of tensor-product smooths because we require anisotropic penalization. That is to say, when we seek to create spatiotemporally smoothed model estimates, it’s inappropriate to assume that the amount of smoothing across space should be the same as the amount of smoothing across time because they’re in fundamentally different units (space being measured in Cartesian coordinates and time being measured in years, months, days, etc.). More details about tensor-product smooths are available in Simon Wood’s _Generalized Additive Models - An Introduction with R_ (2017). ### Overdispersion and Event Modeling Above and beyond using the GAM framework to allow for flexible, nonlinear relationships between our observed county-level variables and COVID-19 mortality, we must have a model specification that agrees with the data generating process. In our case, the data generated are counts of deaths per population, and the class of models most suited to represent counts of events are Poisson, quasi-Poisson, and negative binomial models. Here we’ve chosen to use the negative binomial model as it accounts for overdispersion and has the intuitive interpretation of a Poisson model with gamma distributed underlying rate parameter (Gelman et al., 2013). A negative binomial model in the context of modeling count data can be written $y_{i}\sim\text{NegativeBinomial}(u_{i}\theta_{i},\phi),$ where $y_{i}$ are the counts observed, $u_{i}$ is the “exposure”, $\theta_{i}$ are the rates, and $\phi$ determines the amount of overdispersion. The rates are modeled as $\theta_{i}=e^{X_{i}\beta}$ where $X_{i}$ are the observed covariates and $\beta$ are the coefficients on the covariates corresponding to log rate ratios. The logarithm of the exposure, $\log(u_{i})$ is often called (and later herein referenced as) the offset. In epidemiological contexts, the offset is often representative of the amount of person-time during which observations were recorded. Whereas the Poisson model holds that $\text{var}(y)=\mu$ where $\mu$ is the average rate, the negative binomial model instead assumes that $\text{var}(y)=\mu+\mu^{2}/\phi$. The Poisson model is a special case of the negative binomial model when $\phi\to\infty$. An alternative and equivalent formulation of the negative binomial that is commonly used makes the connection to the Poisson model even more clear: instead of $\theta_{i}$ and $\phi$, using $\alpha$ and $\beta$, $y\sim\text{NegativeBinomial}(\alpha,\beta),\text{ and}$ $\text{NegativeBinomial}(y|\alpha,\beta)=\int\text{Poisson}(y|\theta)\text{Gamma}(\theta|\alpha,\beta)d\theta.$ Introductions to and additional exposition on the negative binomial model, especially in the context of modeling counts of events outcome data, are available in _Bayesian Data Analysis_ and _Regression and Other Stories_ (Gelman et al., 2020, 2013). ### Variables of Interest The following variables have been included as covariates of interest: * • Median Age * • Population Density per Square Mile * • Median Household Income * • Proportion in Poverty * • the Index of Concentration at the Extremes for Racialized Economic Segregation * • Political Lean in the 2020 Election (1 = 100% Republican votes, -1 = 100% Democratic Votes) ##### Index of Concentration at the Extremes for Racialized Economic Segregation The Index of Concentration at the Extremes (ICE) is a measure which describes how concentrated a given area’s population is in terms of the extreme ends of privilege and marginalization (Krieger et al., 2016). In general, the ICE measure is formulated as $\text{ICE}=\frac{\text{Number of People in Most Privileged Category}-\text{Number of People in Least Privileged Category}}{\text{Total Population}}$ Applying the ICE approach to a specific context involves defining the axes of privilege of interest. In this case, data on racialized economic segregation are used from the US Census American Community Survey on white households earning more than $100,000 a year (the most privileged group) or households of People of Color earning less than $25,000 a year (the least privileged group). This variable is referred to throughout as the ICE for racialized economic segregation, or ICEraceinc in the code. Compared with the Gini coefficient which is one of the most popular methods for summarizing area-level rates of inequities, the ICE has the advantage that it is suitable for describing inequities at smaller area levels (Krieger et al., 2016). While the Gini coefficient measures within-area dissimilarity (as in, for example, how unequal wealth is distributed within a county), the ICE measure establishes where on a spectrum a given county’s population falls allowing for comparison across counties. The Gini coefficient suffers from the fact that areas which are made up of relatively homogeneous populations will appear as having low within-area inequality (and therefore low Gini coefficient). Instead, the ICE measures how much of the population is either privileged or not. The Gini coefficient remains useful for reporting on the degree of inequity in larger areal units (like countries, states, and regions), but at smaller areal units (like counties, ZIP codes, census tracts) can be more difficult to interpret and compare. Maps of ICE measure can elucidate what spatial social segregation and polarization exist, and the ICE for racialized economic segregation has been repeatedly and significantly associated with COVID-19 outcomes (Brown et al., 2021; Chen & Krieger, 2021; Eichenbaum & Tate, 2022; Hanage et al., 2020; Krieger et al., 2022; Saha & Feldman, 2020; Sonderlund et al., 2022). ##### Political Lean Political lean has been associated with COVID-19 mortality in numerous studies, with plausible mechanisms explaining the association including differences in non-pharmaceutical intervention uptake (mask usage, social distancing, quarantining), differences in rates of vaccination, differences in political leadership’s messaging, resource allocation, and the adoption of policy interventions (Gonzalez et al., 2021; Grossman et al., 2020; Kaashoek et al., 2021; Krieger et al., 2022; Leonhardt, 2021). ## Results ### Application 1: Non-Spatial Covariate Effects Over Time The GAM models shown in Figure 3 are fit with the following structure using the gam function from the mgcv package (S. Wood, 2022): gam( formula = deaths ~ s(median_age) + te(covariate, date, d=c(1,1)), # regression formula offset = log(popsize/1e5/12), # our offset represents the person-time data = df, # our data-set of county-level observations family = nb(link='log') # indicates negative binomial family and a log-link function ) The formula used puts a one-dimensional smoothing spline on median age to represent a nonlinear age-effect and a two-dimensional tensor-product smooth on the interaction between the given covariate and the date. The d=c(1,1) argument provides the instruction necessary to consider covariate and date as being on separate scales and therefore to fit the tensor-product smooth with anisotropy — that is, to allow for independent amounts of scaling in the dimensions of the covariate and time. The offset used structures the regression to estimate rates in units of person-time per 100,000 person-years. Since the death counts are aggregated to the monthly level, the person-time in units of 100,000 person-years are calculated by taking each county’s population size, dividing by 100,000, and dividing by 12 (for the 12 months in a year). The above structure is used to estimate models for our different covariate variables of interest: median income, percent in poverty, the ICE for racialized economic segregation, political lean. The model presenting median age treats median age as the main covariate of interest including it as the te(covariate, date, d=c(1,1)) and dropping the s(median_age) term which otherwise becomes redundant. Results of these models are summarized in Figure 3 where the additional COVID-19 mortality associated with each covariate is visualized. Figure 3: Additional COVID-19 mortality associated with covariates over time. A) age and time, B) log (base 10) population density and time, C) median income and time, D) proportion in poverty and time, E) ICEraceinc and time, F) political lean and time. In panels B-F the effect of age is marginalized out using the median age in the US, 38.8 (US Census Bureau, 2022). Likelihood ratio tests confirmed that models with covariates interacted with time had significantly lower residual deviance ($p\leq 2.2e^{-16}$ for all models) compared to models only including a spline term on median age and the given covariate not interacted with time. The model interacting median age and time was compared to a model with a spline for median age alone. Akaike Information Criteria values were also lower for all models compared to models that did not interact the covariates with time. A more complex non-spatial application of GAMs to describe the distribution of COVID-19 mortality in the US over time is to consider the associations of mortality with three-way interactions of time and two covariates taken together. In the following example, the formula used is deaths ~ te(date, ICEraceinc, median_age, d=c(1,1,1)). Again, the d argument specifying the marginal basis dimensions is used indicate that each of the date, ICEraceinc, and median_age measures are in different units and should not be smoothed assuming that a one unit difference in one variable is comparable to a one unit difference in another variable. This approach is represented in Figure 4. Figure 4: The changing interacted effect of the ICE for racialized economic segregation and age over time. A) May 2020, B) November 2020, C) May 2021, D) October 2021, E) February 2022, F) August 2022. An animated version is available online at https://github.com/ctesta01/covid.gradient.estimation/blob/main/analysis/05_two_variables_at_a_time/animation_ICEraceinc_age/readme.md ### Application 2: Spatiotemporal Smoothing By fitting GAMs with a tensor-product smoothing term on latitude, longitude, and time we can estimate a spatiotemporally smoothed trend in COVID-19 mortality. To do this, the GAM is constructed similarly as in Application 1 but with the smoothing term specified as te(latitude, longitude, time, d=c(2,1)) where d=c(2,1) indicates that latitude and longitude share the same dimensions (i.e., both are spatial and in units of degrees) while the time data are in separate units. Note that results presented in Application 2 and 3 are based on county-level data from the contiguous US excluding Alaska and Hawaii. Using GAMs to present spatiotemporally smoothed estimates of COVID-19 mortality allows for the synthesis of trends over time. Whereas the raw rates of COVID-19 mortality are noisy due to 72% of the land area in the US being classified as rural and low population density (Health Resources & Services Administration, 2022), the spatiotemporally smoothed GAM estimates use population weighting via the offset specified to estimate a latent surface that represents localized mortality rate averages in spatial coordinates and in time. Figure 5 shows the difference between crude mortality rates and spatially smoothed mortality rates in January 2022 to illustrate the level of noise present in raw rates and how spatially smoothing synthesizes local geographic patterns into trends that can be meaningfully interpreted as local area mortality risk levels with information information pooled across nearby county rates. Figure 6 shows the results from a spatiotemporally smoothed model in select months to highlight how spatiotemporal smoothing can yield results that synthesize trends in space and time. Figure 5: Comparison of crude vs. smoothed mortality rates. A) Crude mortality rates per 100,000 person-years in January 2022. B) Smoothed mortality rates from a GAM applied to mortality rates from January 2022. Figure 6: Spatiotemporally smoothed COVID-19 mortality estimates from a GAM fit to data from March 2020-August 2022. Panels show months selected to highlight the changing spatial patterns of COVID-19 mortality risk over time. An animated version is available online at https://github.com/ctesta01/covid.gradient.estimation/blob/main/analysis/09_spatiotemporal_models/animation/readme.md ### Application 3: Estimating Covariate Effects Adjusted for Spatiotemporal Autocorrelation The final application of GAMs presented here is to estimate the effects of covariates after adjusting for a spatiotemporally smoothed latent risk surface. This has the interpretation of asking what are the changes to COVID-19 mortality rates associated with each covariate after taking into account regional geographic patterns in COVID-19 mortality over time. In particular, this is relevant for understanding what county-level measures are associated with elevated COVID-19 mortality even after adjusting for where COVID-19 rates were locally elevated or depressed due to variation in the timing of local introduction, transmission, and diffusion. Figure 7: The spatial autocorrelation of COVID-19 rates in January 2022. Panel A shows the graph of US counties connected according to which are neighbors. Panel B shows a scatterplot depicting each US counties’ COVID-19 mortality rate (x-axis) compared to the average of its neighbors’ COVID-19 mortality rates (y-axis). A regression line shows the association between counties’ COVID-19 rates and their neighboring counties’ COVID-19 mortality rates. Dotted lines indicated the average for US county-level crude COVID-19 mortality rates (vertical) and for the average of each counties’ neighboring counties’ COVID-19 mortality rates (horizontal). Figure 8: The median (black) and quantiles (shaded bands) of temporal autocorrelation of counties’ COVID-19 mortality rates over time during April 2020 — August 2022. To motivate the need for and application of spatiotemporal smoothing in assessing the associated changes in COVID-19 mortality with county-level covariates, Figure 7 presents the Moran’s I diagnostic plot for January 2022 which summarizes the amount of autocorrelation between counties’ COVID-19 mortality rates and the average of each counties’ neighboring counties’ COVID-19 mortality rates. The observed autocorrelation indicates that there is an association between counties’ COVID-19 rates and their neighboring counties’ COVID-19 mortality rates, implying that without taking this patterning into account, regression models that do not explicitly model the effect of spatial autocorrelation may be biased due to inappropriately assuming that county-level data are independent of one another. Second, while the Moran’s I plot demonstrates the spatial-correlation present in the data at a single time-point, the following Figure 8 shows the tendency of county COVID-19 observations to be autocorrelated between subsequent months. Figure 9: Estimates of increases in COVID-19 mortality associated with county- level covariates without (first) and with (second) adjustment for spatiotemporal autocorrelation. Panels A and B show COVID-19 mortality increases associated with varying levels of median age over time; C and D show associations with log population density; E and F median income; G and H for proportion in poverty; I and J the ICE for racialized economic segregation; K and L political lean. Figure 9 shows the additional COVID-19 mortality associated with each covariate over time without and with adjustment for spatiotemporal trends. The models for population density and median income show particularly marked changes with the spike at high population density and high median incomes disappearing after adjusting for a spatiotemporally smoothed term. This likely reflects the early emergence and surge of COVID-19 in the greater New York City area as being predominantly a feature of the combined effect of human geography and infectious disease dynamics where the earliest introductions of COVID-19 were made to dense social networks in the US. Other effects remain qualitatively changed little, suggesting that the trends in COVID-19 mortality associated with these county-level measures remains durable when taking into account spatiotemporal autocorrelation. ## Discussion This paper presents three applications of GAMs to county-level COVID-19 mortality data from the US during March 2020 to August 2022. This paper addresses an important need, which is to assess what measures of the socioeconomic and geopolitical landscape are associated with increases in COVID-19 mortality in a way that takes into account how infectious disease dynamics are locally correlated in space and time. However, the approaches outlined in this paper are not without several limitations which include (at least) the following: the potential for cross-level confounding, lack of individual level data, an inability to include all possible metrics, further improvements to the methods used may be warranted, and an inability to make claims about the causal relationships between the variables investigated. In analyses that use area-level aggregated data, the ecologic fallacy refers to the possibility that associations between area-level measures and outcomes may not reflect the associations that would be observed if analyzing individual level data on the same measures and outcomes as a result of cross- level confounding (Greenland, 2001). It’s important to note that some of the variables in this analysis are included to contextualize the places in which people live, reside, and work and area-based social metrics should be included even in individual-level analyses so as to accurately capture the effects of both individual-level risk factors and risk factors that stem from the contexts in which people exist (Testa et al., 2022). One particular improvement that could be made in future work is to address age adjustment by using direct or indirect age standardization (Anderson et al., 1998) or to create age-group stratified models. In the applications presented here, county-level median age is adjusted for in order to control for age effects as age-specific data were not publicly available without substantial levels of suppression. Additional variables like vaccination rates, mobility data, further variables that relate to risk factors for COVID-19, and other COVID-19 related outcomes like case trends and hospitalizations may be worthwhile to investigate under a similar framework. The emphasis on trends in COVID-19 mortality in the applications in this paper was done to prioritize presenting data that are less subject to the reporting inconsistencies in COVID-19 cases and hospitalizations (Galaitsi et al., 2021). Further improvements to the methods employed here could include use of soap film smoothing splines which are more appropriate in settings where non-convex geographic boundaries are common (as in with peninsulas and bays) (S. N. Wood et al., 2008). Taking a critical eye to the assumptions employed reveals yet another area of potential improvement, which is the nature of how smoothing splines are penalized in the GAM framework. It is conceivable that geographic changes in COVID-19 mortality rates are not equally smooth across all regions of the US, potentially warranting rapid changes in some areas and more smooth changes in other areas in a risk surface model. The Bayesian wombling literature provides the means to identify areas of rapid change in latent surface models and may pose a fruitful direction of exploration for future efforts to understand trends in COVID-19 mortality rates in the US (Gelfand & Banerjee, 2015). ## Conclusions This work introduces generalized additive models (GAMs) applied to COVID-19 mortality data to document the disparities and inequities in which US counties suffered the worst mortality outcomes and at what times. The effort to document inequities and explain the driving mechanisms behind them is crucial in building up the evidence necessary to enact policy changes that can mitigate unjust and preventable harm in the future. To this end, it is critical that efforts to model infectious disease dynamics take into account how disease transmission and outcomes are spatiotemporally patterned, violating any misplaced assumptions that area-level disease rates can be treated as independent observations. By presenting novel applications of GAMs to trends in COVID-19 mortality, this paper illustrates new ways of understanding and visualizing the disproportionate burden some communities have suffered by incorporating complex patterns of interaction across socioeconomic and geopolitical variables and spatiotemporal trends. ## References reAnderson, R. N., Rosenberg, H. M., & National Center for Health Statistics. (1998). Age standardization of death rates; implementation of the year 2000 standard. _National Vital Statistics Report_. https://stacks.cdc.gov/view/cdc/13357 preBailey, Z. D., Feldman, J. M., & Bassett, M. T. (2021). How Structural Racism Works — Racist Policies as a Root Cause of U.S. Racial Health Inequities. _New England Journal of Medicine_ , _384_(8), 768–773. https://doi.org/10.1056/NEJMms2025396 preBendix, A. (2022). Less than 4% of eligible people have gotten updated Covid booster shots, one month into the rollout. In _NBC News_. https://www.nbcnews.com/health/health-news/updated-covid-booster-shots-doses- administered-cdc-rcna48960 preBlendon, R. J., Schoen, C., DesRoches, C. M., Osborn, R., Scoles, K. L., & Zapert, K. (2002). Inequities In Health Care: A Five-Country Survey. _Health Affairs_ , _21_(3), 182–191. https://doi.org/10.1377/hlthaff.21.3.182 preBrown, K. M., Lewis, J. Y., & Davis, S. K. (2021). An ecological study of the association between neighborhood racial and economic residential segregation with COVID-19 vulnerability in the United States’ capital city. _Annals of Epidemiology_ , _59_ , 33–36. https://doi.org/10.1016/j.annepidem.2021.04.003 preCarpenter, E. (2021). “The Health System Just Wasn’t Built for Us”: Queer Cisgender Women and Gender Expansive Individuals’ Strategies for Navigating Reproductive Health Care. _Women’s Health Issues_ , _31_(5), 478–484. https://doi.org/10.1016/j.whi.2021.06.004 preCDC. (2021). Location, location, location – COVID Data Tracker Weekly Review. In _Centers for Disease Control and Prevention_. https://www.cdc.gov/coronavirus/2019-ncov/covid-data/covidview/past- reports/07162021.html preChen, J. T., & Krieger, N. (2021). Revealing the Unequal Burden of COVID-19 by Income, Race/Ethnicity, and Household Crowding: US County Versus Zip Code Analyses. _Journal of Public Health Management and Practice_ , _27_(1), S43–S56. https://doi.org/10.1097/PHH.0000000000001263 preChrisler, J. C., Barney, A., & Palatino, B. (2016). Ageism can be Hazardous to Women’s Health: Ageism, Sexism, and Stereotypes of Older Women in the Healthcare System. _Journal of Social Issues_ , _72_(1), 86–104. https://doi.org/10.1111/josi.12157 preCoven, J., & Gupta, A. (2020). _Disparities in Mobility Responses to COVID-19_. https://arpitgupta.info/s/DemographicCovid.pdf preDel Rios, M., Chomilo, N. T., & Lewis, N. A. (2022). Covid-19 is an inverse equity story, not a racial equity success story. In _STAT_. https://www.statnews.com/2022/10/25/covid-19-inverse-equity-story-not-racial- equity-success-story/ preDonovan, D. (2022). U.S. Officially Surpasses 1 Million COVID-19 Deaths. In _Johns Hopkins Coronavirus Resource Center_. https://coronavirus.jhu.edu/from- our-experts/u-s-officially-surpasses-1-million-covid-19-deaths preEichenbaum, A., & Tate, A. D. (2022). Health Inequity in Georgia During the COVID-19 Pandemic: An Ecological Analysis Assessing the Relationship Between County-Level Racial/Ethnic and Economic Polarization Using the ICE and SARS- CoV-2 Cases, Hospitalizations, and Deaths in Georgia as of October 2020. _Health Equity_ , _6_(1), 230–239. https://doi.org/10.1089/heq.2021.0118 preFeldman, J. L., Luhur, W. E., Herman, J. L., Poteat, T., & Meyer, I. H. (2021). Health and health care access in the US transgender population health (TransPop) survey. _Andrology_ , _9_(6), 1707–1718. https://doi.org/10.1111/andr.13052 preGalaitsi, S. E., Cegan, J. C., Volk, K., Joyner, M., Trump, B. D., & Linkov, I. (2021). The challenges of data usage for the United States’ COVID-19 response. _International Journal of Information Management_ , _59_ , 102352. https://doi.org/10.1016/j.ijinfomgt.2021.102352 preGelfand, A. E., & Banerjee, S. (2015). Bayesian wombling: Finding rapid change in spatial maps. _WIREs Computational Statistics_ , _7_(5), 307–315. https://doi.org/10.1002/wics.1360 preGelman, A., Carlin, J. B., Stern, H. S., Dunson, D. B., Vehtari, A., & Rubin, D. B. (2013). Bayesian Data Analysis. In _Routledge & CRC Press_. https://www.routledge.com/Bayesian-Data-Analysis/Gelman-Carlin-Stern-Dunson- Vehtari-Rubin/p/book/9781439840955 preGelman, A., Hill, J., & Vehtari, A. (2020). _Regression and Other Stories_ (1st ed.). Cambridge University Press. https://doi.org/10.1017/9781139161879 preGlenza, J. (2020). Covid cases increase across US as upper midwest sees rapid rise. _The Guardian_. https://www.theguardian.com/world/2020/oct/22/covid-cases-coronavirus-us- midwest preGoldstein, J. (2022). During the Omicron surge, Black New Yorkers were hospitalized at a rate more than twice that of white residents. _The New York Times_. https://www.nytimes.com/2022/03/03/nyregion/omicron-hospitalizations- new-york-city.html preGonzalez, K. E., James, R., Bjorklund, E. T., & Hill, T. D. (2021). Conservatism and infrequent mask usage: A study of US counties during the novel coronavirus (COVID-19) pandemic. _Social Science Quarterly_ , _102_(5), 2368–2382. https://doi.org/10.1111/ssqu.13025 preGreenland, S. (2001). Ecologic versus individual-level sources of bias in ecologic estimates of contextual health effects. _International Journal of Epidemiology_ , _30_(6), 1343–1350. https://doi.org/10.1093/ije/30.6.1343 preGrossman, G., Kim, S., Rexer, J. M., & Thirumurthy, H. (2020). Political partisanship influences behavioral responses to governors’ recommendations for COVID-19 prevention in the United States. _Proceedings of the National Academy of Sciences_ , _117_(39), 24144–24153. https://doi.org/10.1073/pnas.2007835117 preHanage, W. P., Testa, C., Chen, J. T., Davis, L., Pechter, E., Seminario, P., Santillana, M., & Krieger, N. (2020). COVID-19: US federal accountability for entry, spread, and inequities—lessons for the future. _European Journal of Epidemiology_ , _35_(11), 995–1006. https://doi.org/10.1007/s10654-020-00689-2 preHealth Resources & Services Administration. (2022). _Defining Rural Population_. https://www.hrsa.gov/rural-health/about-us/what-is-rural preHernandez, I., Dickson, S., Tang, S., Gabriel, N., Berenbrok, L. A., & Guo, J. (2022). Disparities in distribution of COVID-19 vaccines across US counties: A geographic information system–based cross-sectional study. _PLOS Medicine_ , _19_(7), e1004069. https://doi.org/10.1371/journal.pmed.1004069 preHill, L., & Artiga, S. (2022). COVID-19 Cases and Deaths by Race/Ethnicity: Current Data and Changes Over Time. In _KFF_. https://www.kff.org/coronavirus- covid-19/issue-brief/covid-19-cases-and-deaths-by-race-ethnicity-current-data- and-changes-over-time/ preJohnson, A., & Keating, D. (2022). Whites now more likely to die from covid than Blacks: Why the pandemic shifted. In _Washington Post_. https://www.washingtonpost.com/health/2022/10/19/covid-deaths-us-race/ preKaashoek, J., Testa, C., Chen, J., Stolerman, L., Krieger, N., Hanage, W. P., & Santillana, M. (2021). _The Evolving Roles of US Political Partisanship and Social Vulnerability in the COVID-19 Pandemic from February 2020 - February 2021_ [{SSRN} {Scholarly} {Paper}]. https://doi.org/10.2139/ssrn.3933453 preKim, E. J., Marrast, L., & Conigliaro, J. (2020). COVID-19: Magnifying the Effect of Health Disparities. _Journal of General Internal Medicine_ , _35_(8), 2441–2442. https://doi.org/10.1007/s11606-020-05881-4 preKrieger, N., Testa, C., Chen, J. T., Hanage, W. P., & McGregor, A. J. (2022). Relationship of political ideology of US federal and state elected officials and key COVID pandemic outcomes following vaccine rollout to adults: April 2021–March 2022. _The Lancet Regional Health – Americas_ , _16_. https://doi.org/10.1016/j.lana.2022.100384 preKrieger, N., Waterman, P. D., Spasojevic, J., Li, W., Maduro, G., & Van Wye, G. (2016). Public Health Monitoring of Privilege and Deprivation With the Index of Concentration at the Extremes. _American Journal of Public Health_ , _106_(2), 256–263. https://doi.org/10.2105/AJPH.2015.302955 preLambert, J. (2022). Most Americans Don’t Know About the Omicron Covid Boosters. In _Grid News_. https://www.grid.news/story/science/2022/10/04/most- americans-dont-know-about-the-omicron-covid-boosters-that-spells-trouble-for- the-coming-winter/ preLeonhardt, D. (2021). Red Covid. _The New York Times_. https://www.nytimes.com/2021/09/27/briefing/covid-red-states-vaccinations.html preMayo Clinic. (2022). History of COVID-19: Outbreaks and Vaccine Timeline. In _MayoClinic.org_. https://www.mayoclinic.org/coronavirus-covid-19/history- disease-outbreaks-vaccine-timeline/covid-19 preMIT Election Data and Science Lab. (2022). _County Presidential Election Returns 2000-2020_. Harvard Dataverse. https://doi.org/10.7910/DVN/VOQCHQ preMollalo, A., Vahedi, B., & Rivera, K. M. (2020). GIS-based spatial modeling of COVID-19 incidence rate in the continental United States. _Science of The Total Environment_ , _728_ , 138884. https://doi.org/10.1016/j.scitotenv.2020.138884 preMoreland, A., Herlihy, C., Tynan, M. A., Sunshine, G., McCord, R. F., Hilton, C., Poovey, J., Werner, A. K., Jones, C. D., Fulmer, E. B., Gundlapalli, A. V., Strosnider, H., Potvien, A., García, M. C., Honeycutt, S., Baldwin, G., CDC Public Health Law Program, CDC COVID-19 Response Team, & Mitigation Policy Analysis Unit. (2020). Timing of State and Territorial COVID-19 Stay-at-Home Orders and Changes in Population Movement — United States, March 1–May 31, 2020. _MMWR. Morbidity and Mortality Weekly Report_ , _69_. https://doi.org/10.15585/mmwr.mm6935a2 preMurthy, B. P. (2021). Disparities in COVID-19 Vaccination Coverage Between Urban and Rural Counties — United States, December 14, 2020–April 10, 2021. _MMWR. Morbidity and Mortality Weekly Report_ , _70_. https://doi.org/10.15585/mmwr.mm7020e3 preNational Bureau of Economic Research. (2021). Measuring the Virus Risk of Essential Workers and Dependents. In _NBER_. https://www.nber.org/digest/202103/measuring-virus-risk-essential-workers-and- dependents preNational Conference of State Legislatures. (2021). COVID-19: Essential Workers in the States. In _COVID-19: Essential Workers in the States_. https://www.ncsl.org/research/labor-and-employment/covid-19-essential-workers- in-the-states.aspx preOffice of the Commissioner. (2022). Coronavirus (COVID-19) Update: FDA Authorizes Moderna, Pfizer-BioNTech Bivalent COVID-19 Vaccines for Use as a Booster Dose. In _FDA_. https://www.fda.gov/news-events/press- announcements/coronavirus-covid-19-update-fda-authorizes-moderna-pfizer- biontech-bivalent-covid-19-vaccines-use preOkonkwo, N. E., Aguwa, U. T., Jang, M., Barré, I. A., Page, K. R., Sullivan, P. S., Beyrer, C., & Baral, S. (2021). COVID-19 and the US response: Accelerating health inequities. _BMJ Evidence-Based Medicine_ , _26_(4), 176–179. https://doi.org/10.1136/bmjebm-2020-111426 preOrtega, A. N., & Roby, D. H. (2021). Ending Structural Racism in the US Health Care System to Eliminate Health Care Inequities. _JAMA_ , _326_(7), 613–615. https://doi.org/10.1001/jama.2021.11160 prePark, Y. M., Kearney, G. D., Wall, B., Jones, K., Howard, R. J., & Hylock, R. H. (2021). COVID-19 Deaths in the United States: Shifts in Hot Spots over the Three Phases of the Pandemic and the Spatiotemporally Varying Impact of Pandemic Vulnerability. _International Journal of Environmental Research and Public Health_ , _18_(17), 8987. https://doi.org/10.3390/ijerph18178987 prePress, V. G., Huisingh-Scheetz, M., & Arora, V. M. (2021). Inequities in Technology Contribute to Disparities in COVID-19 Vaccine Distribution. _JAMA Health Forum_ , _2_(3), e210264. https://doi.org/10.1001/jamahealthforum.2021.0264 preRader, B., Astley, C. M., Sewalk, K., Delamater, P. L., Cordiano, K., Wronski, L., Rivera, J. M., Hallberg, K., Pera, M. F., Cantor, J., Whaley, C. M., Bravata, D. M., & Brownstein, J. S. (2021). _Spatial Accessibility Modeling of Vaccine Deserts as Barriers to Controlling SARS-CoV-2_. medRxiv. https://doi.org/10.1101/2021.06.09.21252858 preRapp, K. S., Volpe, V. V., Hale, T. L., & Quartararo, D. F. (2022). State–Level Sexism and Gender Disparities in Health Care Access and Quality in the United States. _Journal of Health and Social Behavior_ , _63_(1), 2–18. https://doi.org/10.1177/00221465211058153 preSaha, S., & Feldman, J. M. (2020). _Neighborhood-level Racial/Ethnic and Economic Inequities in COVID-19 Burden Within Urban Areas in the US and Canada_. medRxiv. https://doi.org/10.1101/2020.12.07.20241018 preScott, D. (2020). These 4 Midwestern states are seeing worrying Covid-19 spikes. In _Vox_. https://www.vox.com/2020/9/2/21418812/covid-19-coronavirus- us-cases-midwest-surge preShumaker, L., & Wu, T. (2020). COVID-19 cases surge in U.S. Midwest and Northeast. _Reuters_. https://www.reuters.com/article/us-health-coronavirus- usa-trends-idUSKBN26Q30B preSigler, T., Mahmuda, S., Kimpton, A., Loginova, J., Wohland, P., Charles- Edwards, E., & Corcoran, J. (2021). The socio-spatial determinants of COVID-19 diffusion: The impact of globalisation, settlement characteristics and population. _Globalization and Health_ , _17_(1), 56. https://doi.org/10.1186/s12992-021-00707-2 preSonderlund, A. L., Charifson, M., Schoenthaler, A., Carson, T., & Williams, N. J. (2022). Racialized economic segregation and health outcomes: A systematic review of studies that use the Index of Concentration at the Extremes for race, income, and their interaction. _PLOS ONE_ , _17_(1), e0262962. https://doi.org/10.1371/journal.pone.0262962 preSugg, M. M., Spaulding, T. J., Lane, S. J., Runkle, J. D., Harden, S. R., Hege, A., & Iyer, L. S. (2021). Mapping community-level determinants of COVID-19 transmission in nursing homes: A multi-scale approach. _Science of The Total Environment_ , _752_ , 141946. https://doi.org/10.1016/j.scitotenv.2020.141946 preTesta, C., Chen, J., Hall, E., Javadi, D., Morgan, J., Rushovich, T., Saha, S., Waterman, P., & Krieger, N. (2022). _Public Health Disparities Geocoding Project 2.0 Training Manual_. https://phdgp.github.io/PHDGP2.0/index.html preThe Lancet. (2020). The plight of essential workers during the COVID-19 pandemic. _The Lancet_ , _395_(10237), 1587. https://doi.org/10.1016/S0140-6736(20)31200-9 preThe New York Times. (2021). _Coronavirus (Covid-19) Data in the United States_. The New York Times. https://github.com/nytimes/covid-19-data preThe White House. (2021). Executive Order On Advancing Racial Equity and Support for Underserved Communities Through the Federal Government. In _The White House_. https://www.whitehouse.gov/briefing-room/presidential- actions/2021/01/20/executive-order-advancing-racial-equity-and-support-for- underserved-communities-through-the-federal-government/ preThe White House. (2022). Advancing Equity and Racial Justice Through the Federal Government. In _The White House_. https://www.whitehouse.gov/equity/ preThompson, C. N. (2020). COVID-19 Outbreak — New York City, February 29–June 1, 2020. _MMWR. Morbidity and Mortality Weekly Report_ , _69_. https://doi.org/10.15585/mmwr.mm6946a2 preUnited States COVID - Coronavirus Statistics - Worldometer. (2022). In _Worldometers_. https://www.worldometers.info/coronavirus/country/us/ preU.S. Department of Commerce Economics and Statistics Administration, & U.S. Census Bureau. (2000). _Census Regions and Divisions of the United States_. https://www2.census.gov/geo/pdfs/maps-data/maps/reference/us_regdiv.pdf preUS Census Bureau. (2022). Nation Continues to Age as It Becomes More Diverse. In _Census.gov_. https://www.census.gov/newsroom/press- releases/2022/population-estimates-characteristics.html preUS Census Bureau. (2021). 2020 Census: Redistricting File (Public Law 94-171) Dataset. In _Census.gov_. https://www.census.gov/data/datasets/2020/dec/2020-census-redistricting- summary-file-dataset.html preUS Census Bureau. (2020). American Community Survey 2015-2019 5-Year Data Release. In _Census.gov_. https://www.census.gov/newsroom/press- kits/2020/acs-5-year.html preWalker, K., & Herman, M. (2022). _Tidycensus: Load US Census Boundary and Attribute Data as ’tidyverse’ and ’sf’-Ready Data Frames_. preWei, C.-F., Lan, F.-Y., Hsu, Y.-T., Lowery, N., Dibona, L., Akkeh, R., Kales, S. N., & Yang, J. (2022). Risk of SARS-CoV-2 Infection Among Essential Workers in a Community-Based Cohort in the United States. _Frontiers in Public Health_ , _10_. https://www.frontiersin.org/articles/10.3389/fpubh.2022.878208 preWhitehead, J., Shaver, J., & Stephenson, R. (2016). Outness, Stigma, and Primary Health Care Utilization among Rural LGBT Populations. _PLOS ONE_ , _11_(1), e0146139. https://doi.org/10.1371/journal.pone.0146139 preWhittle, R. S., & Diaz-Artiles, A. (2020). An ecological study of socioeconomic predictors in detection of COVID-19 cases across neighborhoods in New York City. _BMC Medicine_ , _18_(1), 271. https://doi.org/10.1186/s12916-020-01731-6 preWood, S. (2022). _Mgcv: Mixed GAM Computation Vehicle with Automatic Smoothness Estimation_. https://CRAN.R-project.org/package=mgcv preWood, S. N. (2017). _Generalized Additive Models: An Introduction with R_ (2nd ed.). Chapman & Hall/CRC Texts in Statistical Science. https://www.routledge.com/Generalized-Additive-Models-An-Introduction-with-R- Second-Edition/Wood/p/book/9781498728331 preWood, S. N., Bravington, M. V., & Hedley, S. L. (2008). Soap film smoothing. _Journal of the Royal Statistical Society: Series B (Statistical Methodology)_ , _70_(5), 931–955. https://doi.org/10.1111/j.1467-9868.2008.00665.x p
# Hierarchical Transformer for Survival Prediction Using Multimodality Whole Slide Images and Genomics Chunyuan Li, Xinliang Zhu, Jiawen Yao, Junzhou Huang Department of Computer Science and Engineering University of Texas at Arlington Arlington, Texas 76013 Email: {chunyuan.li, xinliang.zhu, jiawen.yao, <EMAIL_ADDRESS> ###### Abstract Learning good representation of giga-pixel level whole slide pathology images (WSI) for downstream tasks is critical. Previous studies employ multiple instance learning (MIL) to represent WSIs as bags of sampled patches because, for most occasions, only slide-level labels are available, and only a tiny region of the WSI is disease-positive area. However, WSI representation learning still remains an open problem due to: (1) patch sampling on a higher resolution may be incapable of depicting microenvironment information such as the relative position between the tumor cells and surrounding tissues, while patches at lower resolution lose the fine-grained detail; (2) extracting patches from giant WSI results in large bag size, which tremendously increases the computational cost. To solve the problems, this paper proposes a hierarchical-based multimodal transformer framework that learns a hierarchical mapping between pathology images and corresponding genes. Precisely, we randomly extract instant-level patch features from WSIs with different magnification. Then a co-attention mapping between imaging and genomics is learned to uncover the pairwise interaction and reduce the space complexity of imaging features. Such early fusion makes it computationally feasible to use MIL Transformer for the survival prediction task. Our architecture requires fewer GPU resources compared with benchmark methods while maintaining better WSI representation ability. We evaluate our approach on five cancer types from the Cancer Genome Atlas database and achieved an average c-index of $0.673$, outperforming the state-of-the-art multimodality methods. ## I Introduction Survival analysis is an essential task in cancer study and diagnosis. Instead of observing the biopsies under the microscope, whole slide pathology images (WSI) are widely used to present tumor growth and morphology information. Automatically analyzing histology provides valuable help for histologists on precision medicine and reduces diagnosis bias. However, WSIs are typically in multi-gigapixel level with high morphological variance, which brings a tremendous challenge to the feature representation of WSIs. Many solutions have been proposed to solve the representation problem of WSIs [1, 2]. A widely used strategy is patch-based processing, where thousands of patches are extracted from a WSI using either a non-overlap sliding window or a random sample method [3, 4, 5]. In the clinics, the tumor area usually only takes a tiny portion of the WSI, which means most patches sampled from the gigapixel WSI are unrelated to survival prediction. Unfortunately, the patch- wise annotation for WSIs is laborious and even infeasible in most situations. Many researchers consider weakly supervised learning approaches to solve the problem with a lack of annotations. Zhu et al. [3] proposed a two-stage framework for survival prediction on WSIs without annotations. They applied K-means clustering to group the patches according to visual appearance and then aggregated patches from the selected clusters for survival prediction. Recent approaches [6, 7, 8, 9] use multiple instance learning (MIL) [10] to formulate the WSI representation, where each patient is considered as a bag containing a set of instances of patches. A bag is annotated as disease- positive if there is any disease-positive patch in the bag. Patch features are integrated with fully connected (FC) layers for specific tasks. On the basis of attention-based MIL, Chen et al., [8] first proposed a Multimodal Co- Attention Transformer (MCAT) architecture that investigates early fusion mechanisms for identifying informative patches. MCAT learns the long-range relationship between image and gene using attention mechanism and visually shows the interpretability of multimodal interaction. However, MCAT uses non- overlapping patches sampling and maintains a bag size of over ten thousand, which essentially increases the computation requirement during training and is time-consuming. Two major limitations exist in the above approaches for WSI representation. Firstly, current studies use a fixed scale for patch extraction. The widely used patch sizes, either $[256\times 256]$ or $[512\times 512]$, have limited ability to capture coarser level tissue morphology characteristics, such as tumor shape, size, and circularity that are essential to determine grades of glioma [11, 12, 13, 14]. As shown in Figure 1, WSIs are scanned at $20\times$ with a resolution of $0.5\mu p$ per pixel. The $256\times 256$ patches sampled from WSI at $20\times$ magnification can capture a small set of cell characteristics and interactions, while patches extracted from downsampled WSIs at $10\times$ or $5\times$ present rich microenvironmental information between different sets of cells or tissue. The multi-scale patches depict significant inter- and intra-instance differences in shape, size, and pattern. Thus they better represent the overall heterogeneity of the tumor microenvironment [15]. Secondly, existing approaches use VGG and Resnet models pre-trained on ImageNet to obtain patch features. Compared with vision Transformer (ViT), CNN models do not incorporate spatial location information and maintain less global information. The different structures result in quantitative variability in out-of-distribution generalization ability [16, 17]. Figure 1: Hierarchy structure and example patches. Left: WSIs at $5x$, $10x$, and $20x$ magnification. Right: Patches extracted from WSIs at different magnifications. The blue box indicated the corresponding region in upper magnification. To address these challenges, we present a weakly supervised, Transformer-based hierarchical architecture named Hierarchical MIL Transformer (HiMT) that learns a mapping between genetics and WSIs at different scales for survival prediction. Specifically, HiMT randomly samples a few patches from different scales, then constructs bag representation using processed images and gene categories. HiMT learns a co-attention imaging-genomic mapping to reduce the space complexity of patches feature. The gene-guided visual concept can be used on various tasks with proper FC layers. Here we demonstrate the power with the survival prediction task. During the patch sampling stage, our hierarchy strategy ensures the patient bag has a reasonable size of approximately 3000, while our benchmark architecture [8] collected an average bag size of 15,000, which requires a plenty of GPU resources during training. Our main contributions can be summarized as follows. * • We introduce a hierarchical mechanism to identify the informative instance across different scales using genetic features as queries. * • We use a pre-trained ViT for feature extraction and showed in ablation studies that even with a single WSI scale, features extracted from $80\%$ fewer patches than MCAT could achieve competitive performance. * • Instead of using a non-overlapping patch sampling procedure, we randomly sampled fewer patches and essentially reduced the computation stress while maintaining outperforming results. We evaluate HiMT on five popular TCGA cancer datasets. Our results outperform the state-of-the-art MIL models on imaging-genetic data mapping tasks, which show that our architecture can capture representative features with vastly reduced feature space. Experiments further show that although the number of patches dramatically reduced compared to previous work [8] which collects on average over ten thousand patches from each WSI, our architecture performs better in survival prediction. In addition, we conduct an ablation study on the five TCGA datasets and demonstrate that HiMT improves with patches from different magnification. Our code is publicly available at https://github.com/chunyuan1/HiMT. ## II Related Work ### II-A Weakly Supervised Learning for WSI Analysis Recent works have developed many weakly supervised models for medical image diagnosis to overcome the unavailability of manual annotation in clinical data [18, 7, 19]. Edwards and Storkey and Zaheer et al. first proposed network architecture on set-based data [20, 21]. Ilse et al. extend the set-based concept by employing an attention mechanism and applying it on WSIs [6]. Yao et al. incorporated attention-based MIL on clustered phenotype and achieved promising results [7]. Instead of only focusing on the instance-level feature, Chen et al. proposed co-attention MIL Transformer that learns an interpretable mapping and visually demonstrates the relationship between imaging and genetics [8]. However, the approaches at a single scale cannot maintain high- level morphological features. We believe that fewer patches with the multi- scale patches feature can achieve promising results compared to tens of thousands of patches. ### II-B Self-Attention and Vision Transformer Solving computer vision tasks using non-convolutional neural networks has been an active research area [22, 23, 24]. The original ViT takes $256\times 256$ images as input and processes it as a sequence of $[16\times 16]$ tokens with positional information. Since ViT was first proposed by Dosovitskiy et al., plenty of recent work is developed to analyze aspects of ViT, such as robustness [25], effects of self-attention [26], designing improved ViT models [27, 28] and comparison with CNN [16, 17]. ViT has been proved for its out-of- distribution generalization and outstanding feature representation due to the accessibility to global information at an early stage [16, 17]. In the research of medical images, pre-trained Resnet is widely used for feature extraction, while ViT is rarely explored. The reason may be that training a transformer model from scratch requires an immense amount of data and intensive computational resources. We highlight that the out-of-distribution generalization ability of ViT makes the ImageNet-pretrained ViT model been able to extract high-quality representative features from the medical image. Based on the fact that ViT model relationships between the tokens, we design a hierarchy model that allows ViT to learn interaction at different scales of the WSIs. ## III Method Figure 2: Framework overview of HiMT contains three stages. Bag representation: 1) Patches sampled from WSIs with different magnification are sent into a pre-trained ViT model to extract instance-level patch feature; 2) DNA microarray are categorized according to functional expression. Early feature fusion: This stage learns a gene-guided co-attention mapping that directly models the pairwise relationship between pathological images and genomics. The GCA is used to reduce the sequence length of WSI bags, leading to more flexibility in choosing aggregation strategies in the following stage. MIL transformer and survival prediction: applies set-based MIL transformer on gene-guided imaging feature and genomics for survival prediction. Figure 2 depicts an overview of HiMT architecture. Given a set of $\Omega$ patients ${X_{i}},i=1,...,\Omega$, each patient has its feature vector $(Surv_{i},\delta_{i})$. The $Surv_{i}$ indicates the time until the event of interest occurs, and vital status $\delta_{i}\in\\{0,1\\}$ indicates whether the event is censored or uncensored (death occurs). The objective is to predict a corresponding target variable $o_{i}$ for patient $X_{i}$ from the imaging data. ### III-A Patient Bag Construction This stage aims to formulate the problem by constructing bags using pathology images and genomics. Multiple Instance Learning (MIL) uses bags as the data samples, each containing a set of unordered instances. In our work, patient $X_{i}$ is a bag of instance, which contains feature vectors $\\{x_{1},\ldots,x_{M}\\}\in\mathbb{R}^{M\times d}$ from all WSIs of $X_{i}$ and functional gene sets $\\{G\\}_{i}$, where the number of instances $M$ various for each image bag. Hierarchy Feature Embedding using ViT. Due to the considerable size, it is infeasible to use WSIs as input. We extract image features by sampling patches at different magnifications from all WSIs belonging to the same patients. Instead of sampling non-overlapping patches, we randomly generate patches from multi-scale WSIs. A thousand patches are extracted randomly from WSIs at $5\times$, $10\times$, and $20\times$ magnification, respectively, with a fixed size of $256\times 256\times 3$. The patches that contain background are excluded. The above procedure results in three thousand patches for each WSI. The number of patches is about eight times less than the non-overlapping sampling strategy, while it contains more morphology information at different resolutions. Then a ViT-L16 model [23] (pre-trained on ImageNet) is used to extracts $d_{k}$-dim feature embeddings $h\in\mathbb{R}^{d_{k}\times 1}$. The $M$ extracted patch embeddings from all WSIs of patient $X_{i}$ are packed into a bag $H_{i}$, where $H_{bag}\in\mathbb{R}^{M\times d_{k}}$ and $M$ is approximately $3000$. Genomic Feature Embedding. The heterogeneity within a gene set limited the utility of the gene database. A Gene alone from the gene set cannot coherently describe the biological impact. To use the semantic information within the gene set, genes that contribute to the same biological function are categorized together and end up with $N$ gene sets. Let $D_{i}=\\{d_{j}\\}$ be the gene expression array of patient $X_{i}$ and $\\{S_{n}\\}^{N}_{n=1}$ represents the functional categories according to [29, 30]. For each gene expression in $\\{d_{j}\\}$, if its attribute $attr_{j}\in S_{n}$, $d_{j}$ is assigned to gene set $g_{n}$. In the resulting gene sets $\\{g_{n}\\}_{n=1}^{N}$, each set $g_{n}$ has various size. We then apply a FC layer over all the gene sets to obtain the final genomic embeddings $\\{G_{n}\in\mathbb{R}^{d_{k}\times 1}\\}^{N}_{n=1}$ and pack it into a bag $G_{bag}\in\mathbb{R}^{N\times d_{k}}$. In our experiments, $N$ is set to $6$, indicating the following functional categories: Tumor Suppression, Oncogenesis, Protein Kinases, Cellular Differentiation, Transcription, and Cytokines and Growth. ### III-B Multimodality Feature Aggregation To capture interpretable interactions between genes and pathological images, we use an early fusion mechanism, Genomic-Guided Co-Attention (GCA) [8], to discover the genotype-phenotype interactions from the tumor microenvironment. Inspired by Transformer attention [22], GCA directly model the pairwise relationship by learning a co-attention matrix $A_{coa}$ between patch embeddings in $H_{bag}\in\mathbb{R}^{M\times d_{k}}$ and gene embeddings $G_{bag}\in\mathbb{R}^{N\times d_{k}}$. Then $A_{coa}$ is used to map the $H_{bag}$ to a set of gene-guided visual concept $\widehat{H}_{bag}$. The mapping can be expressed as $\displaystyle CoA(G,H)=softmax(\frac{QK^{\top}}{\sqrt{d_{k}}})$ (1) $\displaystyle=softmax(\frac{W_{q}GH^{\top}W_{k}^{\top}}{\sqrt{d_{k}}})W_{v}H$ $\displaystyle\to A_{coa}W_{v}H\to\widehat{H}$ where weight matrices $W_{q},W_{k},W_{v}\in\mathbb{R}^{d_{k}\times d_{k}}$ are trainable parameters to map the query $G_{bag}$ and key-value pairs $(H_{bag},H_{bag})$ to an output. In the experiment, bag size of $Q$ is much smaller than that of $K$ and $V$. Consequently, GCA can largely reduce the complexity of WSI bags. ### III-C MIL Transformers with Survival Prediction With image embeddings $\widehat{H}_{coa}\in\mathbb{R}^{N\times d_{v}}$ and genomic embeddings $G_{bag}\in\mathbb{R}^{N\times d_{k}}$ as inputs, we use two set-based MIL Transformers $\mathcal{T}_{H}$ and $\mathcal{T}_{G}$ to aggregate image and genomic features, respectively [21, 22, 31, 8]. This step can be written as $\begin{split}&\mathcal{E}^{(l)}(H^{(l)})=\zeta^{(l)}(\psi^{(l)}(\\{\phi^{(l)}(x_{i}):h_{i}^{(l)}\in H^{(l)}\\}))\\\ &\mathcal{F}^{(L)}(H^{(L)})=\zeta^{(L)}(\rho^{(L)}(\\{\phi^{(L)}(x_{i}):h_{i}^{(L)}\in H^{(L)}\\}))\\\ &\mathcal{T}(X)=\mathcal{F}^{(L)}(\mathcal{E}^{(L-1)}(...\mathcal{E}^{(1)}\\{(x_{i}):x_{i}\in X\\}))\end{split}$ (2) Among the above equations: $h_{i}^{(l)}$ is an arbitrary instance in the set $H^{(l)}$ at hidden layer $l$. $\phi:\mathbb{R}^{d_{in}}\to\mathbb{R}^{d_{out}}$ is an instance-level functions. $\psi^{(l)}$ is the self-attention layer. $\rho:\mathbb{R}^{m\times d_{out}}\to\mathbb{R}^{d_{out}}$ is a permutation-invariant function that aggregate instances to bag-level feature. $\zeta:\mathbb{R}^{d_{out}}\to\mathbb{R}^{\text{\\# class}}$ is a bag-level classifier that undertake target-specific task. In our work, $\zeta$ is a position-wise FC layer and is used to estimate risk score for survival analysis. $\mathcal{E}^{(l)}$ is an encoder block. $\mathcal{F}^{(L)}$ is the MIL network that perform global pooling at the last layer $L$. Precisely, $\psi^{(l)}$ can be written as the set function which is permutation- equivariant: $\psi^{(l)}(\\{h_{i}^{(l)}\\}^{M}_{i=1})=\\{\sum^{M}_{i=1}\frac{\exp{(h_{i}^{(l)}h_{j}^{(l)\top})}}{d_{k}\sum_{j}\exp{(h_{i}^{(l)}h_{j}^{(l)\top})}}\cdot h_{i}^{(l)}\to h_{i}^{(l+1)}\\}$ (3) The expression indicates that Transformer is a generalization of shallow set- based data structure. $\rho_{H}$ and $\rho_{G}$ are implemented following [6]: $\begin{split}&\phi^{(L)}(h_{i}^{(L)})=W_{\phi}h_{i}^{(l)}\\\ &\rho^{(L)}(\\{h_{i}^{(L)}\\}^{M}_{i=1})=\sum^{M}_{i=1}a_{i}\phi^{(L)}(h_{i}^{(L)})\to h^{(l)}\text{ where}\\\ &a_{i}=\frac{\exp\\{W_{\rho}(\tanh(V_{\rho}h_{i}^{(L)\top})\odot\text{sigm}(U_{\rho}h_{i}^{(L)\top}))\\}}{\sum^{M}_{j=1}\exp\\{W_{\rho}(\tanh(V_{\rho}h_{j}^{(L)\top})\odot\text{sigm}(U_{\rho}h_{j}^{(L)\top}))\\}}\\\ &\zeta^{(L)}(h^{(L)})=W_{\zeta}h^{(L)}\end{split}$ (4) in which $W_{\phi}$, $W_{\rho}$, $V_{\rho}$, $U_{\rho}$, $W_{\zeta}\in\mathbb{R}^{d_{v}\times d_{v}}$ are trainable parameters, $\phi^{(L)}$ is a FC layer processing instance-level feature. $\rho^{(L)}$ is a bag-level attention pooling layer and $a_{i}$ controls weight of embedding $h_{i}^{(L)}$ contributing to bag-level $h^{(L)}$. Finally, HiMT integrates output of $\mathcal{T}_{H}$ and $\mathcal{T}_{G}$ to bag-level features by concatenating, denoted as $[\zeta_{h}^{(L)}(h^{(L)}),\zeta_{g}^{(L)}(g^{(L)})]$. Several FC layers are added in the end to obtain the final risk score $o_{risk}$. ### III-D Loss Function For $i$-th patient, the output hazard risk is denoted as $o_{risk,i}$. Let continuous random variable $T$ represent overall survival time. Following the discovery from chen et al. [8] that loss function is mini-batch dependent, we model the survival prediction problem using discrete time intervals. Given right-censored survival data, we partition the time scale into non-overlapping intervals: $[t_{0},t_{1}),[t_{1},t_{2}),[t_{2},t_{3}),[t_{3},t_{4})$ depending on the quantiles of survival time. For each patient, the discrete event time with the continuous event time $T_{i,cont}$ can be defined as $T_{i}=r\text{ if }T_{i,cont}\in[t_{r},t_{r+1})\text{ for }r\in\\{0,1,2,3\\}$ (5) Denote the discrete time ground truth of $i^{th}$ patient as $Y_{i}$. For a patient with bag-level risk $o_{risk,i}$, the hazard function that measures the probability of patient die at time $r$ can be expressed as $f_{hazard}(r|o_{risk,i})=P(T_{i}=r|T_{i}\geq r,o_{risk,i})$ (6) while the survival function that estimate the probability of a patient live longer than time point $R$ can then be defined as $\begin{split}f_{surv}(r|o_{risk,i})=P(T_{i}>r|o_{risk,i})\\\ =\prod^{r}_{u=1}(1-f_{hazard}(u|o_{risk,i}))\end{split}$ (7) When updating the parameters of discrete survival model, we use log likelihood [32] that consider the vital status of patients: $\begin{split}L=-c_{i}\cdot\log(f_{surv}(Y_{i}|o_{risk,i}))\\\ -(1-c_{i})\cdot\log(f_{surv}(Y_{i}-1|o_{risk,i}))\\\ -(1-c_{i})\cdot\log(f_{surv}(Y_{i}|o_{risk,i}))\\\ \end{split}$ (8) Weighted sum is applied on $L$ and $L_{uncensored}$ that up-weight the contribution of uncensored patients. $L_{surv}=(1-\beta)\cdot L+\beta\cdot L_{uncensored}$ (9) where the uncensored loss is computed as $\begin{split}L_{uncensored}=-(1-c_{i})\cdot\log(f_{surv}(Y_{i}-1|o_{risk,i}))\\\ -(1-c_{i})\cdot\log(f_{hazard}(Y_{i}|o_{risk,i}))\end{split}$ (10) ## IV Experiments ### IV-A Dataset Description To show the performance of our architecture, we conduct experiments on five largest cancer datasets from The Cancer Genome Atlas (TCGA): Bladder Urothelial Carcinoma (BLCA), Breast Invasive Carcinoma (BRCA), Glioblastoma Multiforme (GBM) $\&$ Brain Lower Grade Glioma (LGG), Lung Adenocarcinoma (LUAD), Uterine Corpus Endometrial Carcinoma (UCEC). The numbers of WSIs and patients in each dataset are shown in Table I. For each WSI, $1000$ patches are randomly sampled form each hierarchy layer. 5-fold cross-validation is performed on each dataset. TABLE I: The number of WSIs and patients in each TCGA dataset. Dataset | BLCA | BRCA | GBMLGG | LUAD | UCEC ---|---|---|---|---|--- # patients | 373 | 957 | 569 | 453 | 480 # WSIs | 437 | 1022 | 1011 | 515 | 538 TABLE II: Performance comparison with different methods using C-index values. Methods | BLCA | BRCA | GBMLGG | LUAD | UCEC | Overall ---|---|---|---|---|---|--- DeepMISL (WSI Only) [6] | 0.536$\pm$0.038 | 0.564$\pm$0.050 | 0.787$\pm$0.028 | 0.559$\pm$0..060 | 0.625$\pm$0.057 | 0.614 DeepMISL (Concat) | 0.605$\pm$0.045 | 0.551$\pm$0.077 | 0.816$\pm$0.011 | 0.563$\pm$0.050 | 0.614$\pm$0.052 | 0.630 DeepMISL (Bilinear Pooling) | 0.567$\pm$0.034 | 0.536$\pm$0.074 | 0.812$\pm$0.005 | 0.578$\pm$0.036 | 0.562$\pm$0.058 | 0.611 DeepAttnMISL (WSI Only) [7] | 0.504$\pm$0.042 | 0.524$\pm$0.043 | 0.734$\pm$0.029 | 0.548$\pm$0.050 | 0.597$\pm$0.059 | 0.581 DeepAttnMISL (Concat) | 0.611$\pm$0.049 | 0.545$\pm$0.071 | 0.805$\pm$0.014 | 0.595$\pm$0.061 | 0.615$\pm$0.020 | 0.634 DeepAttnMISL (Bilinear Pooling) | 0.575$\pm$0.032 | 0.577$\pm$0.063 | 0.813$\pm$0.022 | 0.551$\pm$0.038 | 0.586$\pm$0.036 | 0.621 MCAT [8] | 0.624$\pm$0.034 | 0.580$\pm$0.069 | 0.817$\pm$0.021 | 0.620$\pm$0.032 | 0.622$\pm$0.019 | 0.653 ours | 0.660$\pm$0.021 | 0.606$\pm$0.028 | 0.823$\pm$0.019 | 0.616$\pm$0.016 | 0.658$\pm$0.047 | 0.673 ### IV-B Implementation Details HiMT is trained with one Geforce GTX 1080 Ti GPU. The functional signatures to category the gene embeddings are obtained from [29]. After collecting the six gene categories, several FC layers are applied to convert the gene categories into similar lengths, followed by one encoder layer. Patch features are extracted by ViT-Large model pre-trained on ImageNet-21k with $16\times 16$ token embedding strategy [23], followed by a FC layer that obtains a feature vector of length $1024$ for each patch. The third phase of HiMT uses two Vision Transformer from [23] with an attention pooling layer to deal with WSI and genomic features, respectively. The processed logits are then concatenated together and go through several FC layer with sigmoid function to obtain the final risk score $o_{risk}$. During training, we follow the experimental settings in [8] that use Adam optimizer [33] with a learning rate of $2\times 10^{-4}$ and weight decay of $1\times 10^{-5}$. Batch size is set to $1$ due to the various bag size. ### IV-C Evaluation metrics Concordance index (C-index) is a popular metric in evaluating survival prediction models [4, 3, 7]. We also use it as evaluation metric in our experiments. The C-index quantifies the correlation of risk score and survival time, calculated as follows: $c=\frac{1}{n}\sum_{i\in\\{1...N|\sigma_{i}=1\\}}\sum_{s_{j}>s_{i}}I[X_{i}\hat{\beta}>X_{j}\hat{\beta}]$ (11) where $n$ is the number of comparable pairs, $I[.]$ is the indicator function and $s.$ is the actual observation. The value of C-index ranges from 0 to 1. The larger CI value means the better prediction performance of the model and vice versa. ### IV-D Experimental Results We compare our architecture with several state-of-the-art methods for survival prediction using the same 5-fold cross-validation splits, training hyper- parameters, and loss function. * • DeepMISL [6]: Deep multiple survival learning is a set-based network that first applies global attention into MIL algorithm. * • DeepAttnMISL [7]: DeepAttnMISL first applies K-Mean clustering to group instant-level feature into phenotypes. Then an attention-based global pooling is used to aggregate the risk score of each patient. * • MCAT [8]: A state-of-the-art multimodal survival architecture that incorporate self-attention mechanism to learn an interpretable mapping between imaging and genomics in early fusion manner. Table II presents the results using the above methods on five cancer datasets. Compared with WSI-based MIL methods (DeepMISL, DeepAttnMISL), HiMT increases the overall c-index by $9.6\%$ and $15.8\%$. Against the similar methods with multimodal approaches, HiMT achieves $6.8\%$, $6.2\%$, and $3.1\%$ advancement over DeepMISL, DeepAttnMISL, and MCAT, respectively. DeepMISL and DeepAttnMISL both attain more reasonable results than baseline models that only use WSIs. However, DeepMISL shows a considerable variance among all experiments with $p>0.1$, indicating the result is unreliable. DeepMISL exhibits a marginal performance boost over the baseline method with statistically significant for the BLCA ($p<0.01$). Unlike the late fusion strategy used in DeepMISL and DeepAttnMISL, MCAT proposed an early fusion mechanism called co-attention and performs notable progress over previous work with $p<0.01$ for BLCA, GBMLGG, and UCEC. Our method achieves the best average c-index among all the approaches with $p<0.01$ for all five cancer datasets. Specifically, HiMT achieves the highest c-index in 4 out of 5 cancer datasets in comparison concerning each cancer. This indicates that the hierarchy architecture in HiMT can capture representative features from medical images while also requiring less computational resources. Figure 3: Boxplots of C-index values of five cancer datasets compared with MCAT. The red dot indicates the mean C-index of the 5-fold cross-validation. The boxplot of C-index in Figure 3 depicts the variation of C-index values among the 5-fold and further demonstrate that our algorithm not only performs better in terms of mean C-index but also is more robust than previous work. TABLE III: Ablation study assessing C-index performance with respect to different magnification settings. 1000 $256\times 256$ patches are sampled from each magnification for experiments shown below. Methods | BLCA | BRCA | GBMLGG | LUAD | UCEC | Overall (C-index) ---|---|---|---|---|---|--- Patches at $5\times$ | 0.636$\pm$0.022 | 0.612$\pm$0.050 | 0.806$\pm$0.006 | 0.593$\pm$0.061 | 0.556$\pm$0.045 | 0.641 Patches at $10\times$ | 0.634$\pm$0.010 | 0.586$\pm$0.047 | 0.804$\pm$0.009 | 0.598$\pm$0.045 | 0.635$\pm$0.071 | 0.651 Patches at $20\times$ | 0.638$\pm$0.028 | 0.609$\pm$0.046 | 0.821$\pm$0.027 | 0.609$\pm$0.030 | 0.638$\pm$0.031 | 0.663 Patches at $20\times$ and $10\times$ | 0.626$\pm$0.018 | 0.592$\pm$0.029 | 0.805$\pm$0.008 | 0.617$\pm$0.048 | 0.645$\pm$0.078 | 0.657 Patches at $20\times$, $10\times$ and $5\times$ | 0.660$\pm$0.021 | 0.606$\pm$0.028 | 0.823$\pm$0.019 | 0.616$\pm$0.016 | 0.658$\pm$0.047 | 0.673 TABLE IV: Ablation study assessing AUC values with respect to different magnification settings. 1000 $256\times 256$ patches are sampled from each magnification for experiments shown below. Methods | BLCA | BRCA | GBMLGG | LUAD | UCEC | Overall (AUC) ---|---|---|---|---|---|--- Patches at $5\times$ | 0.664$\pm$0.051 | 0.648$\pm$0.055 | 0.843$\pm$0.016 | 0.612$\pm$0.073 | 0.583$\pm$0.047 | 0.670 Patches at $10\times$ | 0.659$\pm$0.031 | 0.611$\pm$0.045 | 0.844$\pm$0.016 | 0.616$\pm$0.050 | 0.651$\pm$0.082 | 0.676 Patches at $20\times$ | 0.663$\pm$0.055 | 0.626$\pm$0.040 | 0.862$\pm$0.028 | 0.624$\pm$0.047 | 0.655$\pm$0.028 | 0.686 Patches at $20\times$ and $10\times$ | 0.654$\pm$0.048 | 0.615$\pm$0.009 | 0.841$\pm$0.007 | 0.633$\pm$0.067 | 0.662$\pm$0.075 | 0.681 Patches at $20\times$, $10\times$ and $5\times$ | 0.697$\pm$0.040 | 0.632$\pm$0.035 | 0.856$\pm$0.037 | 0.631$\pm$0.022 | 0.678$\pm$0.067 | 0.699 ### IV-E Ablation Studies To evaluate the impact of hierarchical architecture in solving MIL tasks, we conduct an ablation study to evaluate patches sampled on fixed magnification with pretrained ViT for feature extraction. Table III and table IV shows the C-index and AUC results, respectively, for HiMT models using: 1) patches sampled from single magnification, 2) patches sampled from multi-scale WSIs. The results illustrate that both feature extraction using pre-trained ViT model and hierarchical architecture contribute to the overall c-index. Specifically, from Table III and Table IV, the models use patches from all three resolutions have the best overall performance, with an average C-index of $0.673$ and average AUC of $0.699$. In terms of each cancer dataset, although it beat almost all the benchmark methods, the performance of LUAD and BRCA do not show significant improvement with different experimental settings. LUAD stands for lung cancer, which is well-known for the hybrid entities, high genetic instability, high malignancy, and mortality [34]. Even in the clinic, the diversity causes problems for experts to diagnose. Surprisingly, both the mean C-index and mean AUC of BRCA dataset with $5\times$ magnification achieve the highest score among multi-scale settings. BRCA is a heterogeneous complex of diseases with many subtypes that has distinct biological features [35]. The result might suggest that tissue-level information is more important to HiMT on survival prediction. Noticeably, in the third row of Table III and IV where the model only takes $1000$ $256\times 256$ patches sampled from each $20\times$ WSI as input, HiMT outperforms the MCAT[8], which uses over $15,000$ patches per WSI with a similar patch size for feature representation. The result further suggests that pre-trained ViT has out-of-distribution generalization ability and could be used for image feature extraction in a wide variety of research. ## V Conclusion This paper presents a Transformer-based hierarchical architecture for weakly supervised survival prediction on multimodality data. Compared with the state- of-the-art methods, our work requires remarkably less computational resources by reducing the number of patches from WSIs without losing performance. Obtaining features from different magnification capture representative information from the microenvironment. Our results have quantitatively shown significant improvement over the state-of-the-art methods on survival prediction. Future work. The relative position of different types of cells is vital to identifying the cancer stage, which is a critical measurement during diagnosis. Future research could include integrating spatial relations to capture relative position between different cell types, potentially improving further. Besides, the ViT models used for feature extraction are pre-trained on ImageNet-1k, while the pathology image is distinct from the nature image. A ViT model pretrained on an ensemble of cancers might be able to extract more representative patch features. ## Acknowledgment This work was partially supported by the NSF CAREER grant IIS-1553687 and Cancer Prevention and Research Institute of Texas (CPRIT) award (RP190107). The results in this paper are in whole based upon data generated by the TCGA Research Network: https://www.cancer.gov/tcga. ## References * [1] J. Yao, S. Wang, X. Zhu, and J. Huang, “Imaging biomarker discovery for lung cancer survival prediction,” in _International Conference on Medical Image Computing and Computer-Assisted Intervention_. Springer, 2016, pp. 649–657. * [2] X. Zhu, J. Yao, X. Luo, G. Xiao, Y. Xie, A. Gazdar, and J. Huang, “Lung cancer survival prediction from pathological images and genetic data—an integration study,” in _2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI)_. IEEE, 2016, pp. 1173–1176. * [3] X. Zhu, J. Yao, F. Zhu, and J. Huang, “Wsisa: Making survival prediction from whole slide histopathological images,” in _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 2017, pp. 7234–7242. * [4] X. Zhu, J. Yao, and J. Huang, “Deep convolutional neural network for survival analysis with pathological images,” in _2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)_. IEEE, 2016, pp. 544–547. * [5] J. Yao, X. Zhu, F. Zhu, and J. Huang, “Deep correlational learning for survival prediction from multi-modality data,” in _International Conference on Medical Image Computing and Computer-Assisted Intervention_. Springer, 2017, pp. 406–414. * [6] M. Ilse, J. Tomczak, and M. Welling, “Attention-based deep multiple instance learning,” in _International conference on machine learning_. PMLR, 2018, pp. 2127–2136. * [7] J. Yao, X. Zhu, J. Jonnagaddala, N. Hawkins, and J. Huang, “Whole slide images based cancer survival prediction using attention guided deep multiple instance learning networks,” _Medical Image Analysis_ , vol. 65, p. 101789, 2020. * [8] R. J. Chen, M. Y. Lu, W.-H. Weng, T. Y. Chen, D. F. Williamson, T. Manz, M. Shady, and F. Mahmood, “Multimodal co-attention transformer for survival prediction in gigapixel whole slide images,” in _Proceedings of the International Conference on Computer Vision_ , 2021, pp. 4015–4025. * [9] B. Li, Y. Li, and K. W. Eliceiri, “Dual-stream multiple instance learning network for whole slide image classification with self-supervised contrastive learning,” in _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2021, pp. 14 318–14 328. * [10] O. Maron and T. Lozano-Pérez, “A framework for multiple-instance learning,” _Advances in neural information processing systems_ , pp. 570–576, 1998. * [11] X. Wang, D. Wang, Z. Yao, B. Xin, B. Wang, C. Lan, Y. Qin, S. Xu, D. He, and Y. Liu, “Machine learning models for multiparametric glioma grading with quantitative result interpretations,” _Frontiers in neuroscience_ , vol. 12, p. 1046, 2019. * [12] A. H. Beck, A. R. Sangoi, S. Leung, R. J. Marinelli, T. O. Nielsen, M. J. Van De Vijver, R. B. West, M. Van De Rijn, and D. Koller, “Systematic analysis of breast cancer morphology uncovers stromal features associated with survival,” _Science translational medicine_ , vol. 3, no. 108, pp. 108ra113–108ra113, 2011. * [13] S. Kothari, J. H. Phan, A. O. Osunkoya, and M. D. Wang, “Biological interpretation of morphological patterns in histopathological whole-slide images,” in _Proceedings of the ACM Conference on Bioinformatics, Computational Biology and Biomedicine_ , 2012, pp. 218–225. * [14] L. Chan, M. S. Hosseini, C. Rowsell, K. N. Plataniotis, and S. Damaskinos, “Histosegnet: Semantic segmentation of histological tissue type in whole slide images,” in _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , 2019, pp. 10 662–10 671. * [15] S. Graham, Q. D. Vu, S. E. A. Raza, A. Azam, Y. W. Tsang, J. T. Kwak, and N. Rajpoot, “Hover-net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images,” _Medical Image Analysis_ , vol. 58, p. 101563, 2019. * [16] M. Raghu, T. Unterthiner, S. Kornblith, C. Zhang, and A. Dosovitskiy, “Do vision transformers see like convolutional neural networks?” _Advances in Neural Information Processing Systems_ , vol. 34, 2021. * [17] C. Zhang, M. Zhang, S. Zhang, D. Jin, Q. Zhou, Z. Cai, H. Zhao, S. Yi, X. Liu, and Z. Liu, “Delving deep into the generalization of vision transformers under distribution shifts,” _arXiv preprint arXiv:2106.07617_ , 2021. * [18] L. Hou, D. Samaras, T. M. Kurc, Y. Gao, J. E. Davis, and J. H. Saltz, “Patch-based convolutional neural network for whole slide tissue image classification,” in _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 2016, pp. 2424–2433. * [19] M. Y. Lu, T. Y. Chen, D. F. Williamson, M. Zhao, M. Shady, J. Lipkova, and F. Mahmood, “Ai-based pathology predicts origins for cancers of unknown primary,” _Nature_ , vol. 594, no. 7861, pp. 106–110, 2021. * [20] H. Edwards and A. Storkey, “Towards a neural statistician,” _arXiv preprint arXiv:1606.02185_ , 2016. * [21] M. Zaheer, S. Kottur, S. Ravanbakhsh, B. Poczos, R. R. Salakhutdinov, and A. J. Smola, “Deep sets,” _Advances in Neural Information Processing Systems_ , vol. 30, 2017. * [22] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” in _Advances in neural information processing systems_ , 2017, pp. 5998–6008. * [23] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly _et al._ , “An image is worth 16x16 words: Transformers for image recognition at scale,” _arXiv preprint arXiv:2010.11929_ , 2020. * [24] H. Touvron, M. Cord, M. Douze, F. Massa, A. Sablayrolles, and H. Jégou, “Training data-efficient image transformers & distillation through attention,” in _International Conference on Machine Learning_. PMLR, 2021, pp. 10 347–10 357. * [25] S. Bhojanapalli, A. Chakrabarti, D. Glasner, D. Li, T. Unterthiner, and A. Veit, “Understanding robustness of transformers for image classification,” _arXiv preprint arXiv:2103.14586_ , 2021. * [26] M. Caron, H. Touvron, I. Misra, H. Jégou, J. Mairal, P. Bojanowski, and A. Joulin, “Emerging properties in self-supervised vision transformers,” _arXiv preprint arXiv:2104.14294_ , 2021. * [27] Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” _arXiv preprint arXiv:2103.14030_ , 2021. * [28] C.-F. Chen, Q. Fan, and R. Panda, “Crossvit: Cross-attention multi-scale vision transformer for image classification,” _arXiv preprint arXiv:2103.14899_ , 2021. * [29] A. Liberzon, C. Birger, H. Thorvaldsdóttir, M. Ghandi, J. P. Mesirov, and P. Tamayo, “The molecular signatures database hallmark gene set collection,” _Cell systems_ , vol. 1, no. 6, pp. 417–425, 2015. * [30] A. Subramanian, P. Tamayo, V. K. Mootha, S. Mukherjee, B. L. Ebert, M. A. Gillette, A. Paulovich, S. L. Pomeroy, T. R. Golub, E. S. Lander _et al._ , “Gene set enrichment analysis: a knowledge-based approach for interpreting genome-wide expression profiles,” _Proceedings of the National Academy of Sciences_ , vol. 102, no. 43, pp. 15 545–15 550, 2005. * [31] J. Lee, Y. Lee, J. Kim, A. Kosiorek, S. Choi, and Y. W. Teh, “Set transformer: A framework for attention-based permutation-invariant neural networks,” in _International Conference on Machine Learning_. PMLR, 2019, pp. 3744–3753. * [32] S. G. Zadeh and M. Schmid, “Bias in cross-entropy-based training of deep survival networks,” _IEEE transactions on pattern analysis and machine intelligence_ , 2020. * [33] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” _3rd International Conference on Learning Representations, ICLR_ , 2014\. * [34] I. Petersen, “The morphological and molecular diagnosis of lung cancer,” _Deutsches Ärzteblatt International_ , vol. 108, no. 31-32, p. 525, 2011\. * [35] O. Yersal and S. Barutca, “Biological subtypes of breast cancer: Prognostic and therapeutic implications,” _World journal of clinical oncology_ , vol. 5, no. 3, p. 412, 2014.
# On the dynamics of mortality and the ephemeral nature of mammalian megafauna Taran Rallings1, Christopher P Kempes2, Justin D. Yeakel1,∗ 1School of Natural Sciences, University of California Merced 2Santa Fe Institute ∗Corresponding author<EMAIL_ADDRESS> ###### Abstract Energy flow through consumer-resource interactions is largely determined by body size. Allometric relationships govern the dynamics of populations by impacting rates of reproduction, as well as alternative sources of mortality, which have differential impacts on smaller to larger organisms. Here we derive and investigate the timescales associated with four alternative sources of mortality for terrestrial mammals: mortality from starvation, mortality associated with aging, mortality from consumption by predators, and mortality introduced by anthropogenic subsidized harvest. The incorporation of these allometric relationships into a minimal consumer-resource model illuminates central constraints that may contribute to the structure of mammalian communities. Our framework reveals that while starvation largely impacts smaller-bodied species, the allometry of senescence is expected to be more difficult to observe. In contrast, external predation and subsidized harvest have greater impacts on the populations of larger-bodied species. Moreover, the inclusion of predation mortality reveals mass thresholds for mammalian herbivores, where dynamic instabilities may limit the feasibility of megafaunal populations. We show how these thresholds vary with alternative predator-prey mass relationships, which are not well understood within terrestrial systems. Finally, we use our framework to predict the harvest pressure required to induce mass-specific extinctions, which closely align with previous estimates of anthropogenic megafaunal exploitation in both paleontological and historical contexts. Together our results underscore the tenuous nature of megafaunal populations, and how different sources of mortality may contribute to their ephemeral nature over evolutionary time. ## Introduction Consumer-resource interactions are the fundamental unit from which complex food webs arise [deangelis1980energy]. In such dynamics, the rates governing transitions of biomass and energy from one species to another are largely determined by body size [Yodzis:1992hg]. Specifically, the allometric relationships between consumer body mass and metabolic rate constrain energetic assimilation [hou2008], storage [Lindstedt:2002td], and growth [West:2001bv], all of which govern the dynamics of populations [hennemann1983relationship, West:2002it, Kempes:2012hy, yeakel2018dynamics]. Because allometrically-constrained models of population dynamics apply generally across large taxonomic clades, they are useful for examining dynamic constraints that may contribute to community structure across macroevolutionary timescales [DeLong:2012fjb, DeLong2012carnivores, Pawar2012, yeakel2018dynamics, bhat2020scaling]. Furthermore, examination of community dynamics at these scales enables the investigation of extinct communities where body size distributions were different than those in contemporary ecosystems [alroy2001multispecies, brook2008synergies, bradshaw2021relative]. The dynamics of populations represent an energetic balance between reproduction and mortality [murdoch:2003]. Across Mammalia, the average rate of reproduction can be predicted from allometric scaling relationships [CalderIII:1983jd], though individual clades demonstrate a large variety of reproductive strategies – from changing reproductive cycles, litter sizes, and dynamic responses to changing resource conditions to name a few [roff1993evolution]. As these strategies are typically evolved responses to particular conditions and clade-specific, they are not universally experienced. On the other hand, mortality has a variety of forms that nearly all species must deal with to a greater or lesser extent, and do not all scale similarly with body size [weitz2006size]. Mortality originates from both internal and external drivers, where the former depends on an organism’s internal state to initiate death. For example, senescence and starvation involve physiological states that change with respect to clock time, metabolic rate, and resource depletion [yeakel2018dynamics, robert2015actuarial]. In contrast, external drivers of mortality consist of an outside force that induces death more independently of an organism’s internal state, such as mortality due to natural predation or subsidized anthropogenic harvest. Often, mortality occurs through correlations between internal and external drivers, where for example, the starvation state of prey may alter the success rates of predators [Alonzo:2002ub]. While virtually all primary consumer populations must deal with the effects of resource limitation, aging, and predation, the effects of anthropogenic harvesting (the subsidized extraction of prey) are uniquely limited to those species serving as resources for human populations [Dunne:2016jp]. How do different sources of mortality impact the dynamics of mammalian populations? Here we construct a general consumer-resource framework to examine mammalian herbivore populations as a function of consumer body size $M_{C}$, as well as size-dependent vulnerability to different internal and external pressures. Our approach integrates relationships governing specific timescales of physiology and assimilation from a process-based energetic perspective [West:2001bv]. Our model is low-dimensional and compact [cf. yeakel2018dynamics], but due to its close connection to fundamental energetic mechanism, it is also capable of reproducing observed large-scale empirical patterns of mammalian communities. We begin by describing our approach, reproducing key macroecological relationships such as Damuth’s law [Damuth1987], and examine how changes to energetic parameters impact these predictions. We then derive timescales associated with four sources of mortality experienced by mammalian consumers: _i_) natural mortality, _ii_) starvation mortality, _iii_) natural predation, and _iv_) subsidized anthropogenic harvest. By examining each source of mortality in turn, our framework illuminates central constraints governing mass-specific behaviors, strategies, and risks experienced by mammalian consumers. Our results reveal four key insights into the constraints structuring mammalian communities. First, our allometric consumer-resource system accurately captures both the central tendency and variability of Damuth’s law, suggesting that the included vital rates capture mass-specific dynamics. Second, our results demonstrate that natural and starvation mortality differentially impact small mammals, confirming expectations, and point to why the allometric effects of senescence are difficult to observe in nature. Third, we detail the differences in how mortality under different levels of predation intensity induce dynamic instabilities for large-bodied megaherbivores. We also show that the body size at which these instabilities occur is dependent on the prevailing predator-prey mass relationship (PPMR). Finally, we evaluate the harvest pressure required to induce mass-specific extinction, and show that our predictions are comparable to estimates of both paleontological and historical exploitation of mammalian megafauna. ## Methods We model a consumer-resource interaction, where the resource $R$ (g/m2) grows logistically with intrinsic growth rate $\alpha$ to a carrying capacity $k$, and declines due to consumption by an herbivore consumer population $C$ (g/m2) (Eq. 3). Consumed resources govern both consumer somatic maintenance and reproduction. The rate of consumption to fuel somatic maintenance is given by $\rho$, and is independent of resource density, as these are invariant requirements of the consumer population [yeakel2018dynamics]. In contrast, the rate of consumption to fuel reproduction is proportional to resource density and is given by $\lambda_{C}(R)/Y_{C}$, where $\lambda_{\rm C}(R)$ is the consumer growth rate and $Y_{\rm C}$ is the consumer yield coefficient, or the grams of consumer produced per gram of resource consumed. As in DeLong & Vasseur [DeLong:2012fjb], the consumer’s growth rate $\lambda_{C}(R)$ follows a Type II (saturating) functional response given the resource density $R$, where the maximum growth is $\lambda^{\rm max}_{\rm C}$ and the resource half- saturation density is $\hat{k}=k/2$, such that $\lambda_{\rm C}(R)=\lambda^{\rm max}_{\rm C}\left(\frac{R}{\hat{k}+R}\right).$ (1) While the consumer population density grows at rate $\lambda_{\rm C}(R)$, we assume for now that consumer mortality is a function of both natural mortality $\mu$ and starvation $\sigma(R)$, where the rate of starvation, $\sigma(R)=\sigma^{\rm max}\left(1-\frac{R}{k}\right),$ (2) increases as resources become scarce. In this context, $\sigma^{\rm max}$ is the maximal rate of starvation that occurs when the environment is devoid of resources. The full system describing resource and consumer dynamics is given by $\displaystyle\frac{{\rm d}}{{\rm dt}}C$ $\displaystyle=\lambda_{\rm C}(R)C-\left(\mu+\sigma(R)+...\right)C,$ $\displaystyle\frac{{\rm d}}{{\rm dt}}R$ $\displaystyle=\alpha R\left(1-\frac{R}{k}\right)-\left(\frac{\lambda_{\rm C}(R)}{Y_{\rm C}}+\rho\right)C,$ (3) where the ‘$...$’ denotes where additional mortality terms, described later, will be included. The dynamic outcomes of this system of equations include two trivial steady states at $(R^{*}=0,C^{*}=0)$ and $(R^{*}=k,C^{*}=0)$, and one internal steady state where both the consumer and resource population coexist. See tab. 1 for a description of parameters. The rate laws describing resource consumption as well as consumer growth and mortality all vary as a function of consumer body mass $M_{C}$, where the consumer is assumed to be a mammalian herbivore, and the resource is an unspecified primary producer with characteristic growth rate, carrying capacity, and energy density $E_{d}$. We approach the derivation of vital rates with respect to consumer mass by solving for multiple timescales associated with ontogenetic growth, maintenance, and expenditure. The growth of an individual consumer from birth mass $m=m_{0}$ to its reproductive size $m=0.95M_{C}$ is given by the solution to the general balance condition $B_{0}m^{\eta}=E_{m}\frac{\rm d}{\rm dt}m+B_{m}m$, where $E_{m}$ is the energy needed to synthesize a unit of biomass, $B_{m}$ is the metabolic rate to support an existing unit of biomass (tab. 1), and the metabolic exponent $\eta=3/4$ [Pirt1965, West:2001bv, hou2008]. From this balance condition, the time required for an organism starting from mass $m_{1}$ to reach mass $m_{2}$ follows $\tau(m_{1},m_{2})=\ln\left(\frac{1-(m_{1}/M_{C})^{1-\eta}}{1-(m_{2}/M_{C})^{1-\eta}}\right)\frac{M_{C}^{1-\eta}}{a(1-\eta)}$ (4) where $a=B_{0}/E_{m}$ [West:2001bv]. We use this general timescale equation to calculate maximal rates of growth and starvation as a function of organismal body size $M_{C}$, which are then modified by resource density $R$ to provide realized timescales (eqs. 1,2). We note that a more complex framework could include the effects of changing resource densities on timescales directly, where individual growth is itself variable, effectively introducing dynamic population structure [deRoos2020]. From this general equation, we calculate the timescale of reproduction for an herbivore consumer of mass $M_{C}$ as $t_{\lambda_{C}}=\tau(m_{0},0.95M_{C})$, such that the maximal reproductive rate is $\lambda^{\rm max}_{C}=\ln(\nu)/t_{\lambda_{C}}$, where $\nu=2$ is the set number of offspring per reproductive cycle [Savage2004, yeakel2018dynamics]. The consumer yield coefficient is given by $Y_{C}=M_{C}E_{d}/B_{\lambda_{C}}$ (g consumer per g resource), where $B_{\lambda_{C}}$ is the lifetime energy use required by the herbivore to reach maturity $B_{\lambda_{C}}=\int_{0}^{t_{\lambda_{C}}}B_{0}m(t)^{\eta}{\rm d}t$, and the maintenance rate is given by $\rho=B_{0}M_{C}^{\eta}/M_{C}E_{d}$ [yeakel2018dynamics]. To determine the rate of mortality from starvation, we calculate the time required for an organism to metabolize its endogenous energetic stores, estimated from its cumulative fat and muscle mass, where the remaining mass is given by $M_{C}^{\rm starve}=M_{C}-(M_{C}^{\rm fat}+M_{C}^{\rm musc})$ (see app. C). During starvation, we assume that an organism burns its existing endogenous stores as its sole energy source, such that the balance condition is altered to $\frac{\rm d}{\rm dt}mE^{\prime}_{m}=-B_{m}m$, where $E_{m}^{\prime}$ is the amount of energy stored in a unit of biomass (differing from the amount of energy used to synthesize a unit of biomass $E_{m}$; tab. 1) [moses2008rmo, hou2008]. The starvation timescale is then given by $t_{\sigma}=-\frac{M_{C}^{1-\eta}}{a^{\prime}}\ln(M_{C}^{\rm starve}/M_{C}),$ (5) where $a^{\prime}=B_{0}/E^{\prime}_{m}$, such that the starvation rate is the $\sigma^{\rm max}=1/t_{\sigma}$. Importantly, the starvation mortality expressed here is specifically that experienced by adult organisms (as the timescale of metabolizing fat stores is conditioned on adult mass), and does not capture potential starvation mortality of juveniles. To determine the rate of mortality from aging, we note that population cohorts experience two primary sources of natural mortality: the initial cohort mortality rate $q_{0}$ and the annual rate of increase in mortality as the cohort ages, or the actuarial aging rate, $q_{a}$ over lifetime $t_{\ell}$. We begin by assuming that the number of survivors over time follows a Gompertz relationship [CalderIII:1983jd] from which we derive the average rate of natural mortality $\mu=\frac{q_{0}}{q_{a}t_{\ell}}\big{(}{\rm exp}(q_{a}t_{\ell})-1\big{)}.$ (6) The three parameters $(q_{0},q_{a},t_{\ell})$ each have well-documented allometric relationships for terrestrial mammals, such that natural mortality can be written as a function of consumer mass $\mu(M_{C})$ (see app. A). Because both cohort and actuarial mortality are not subdivided into specific categories, we may assume that this rate is capturing the combined effects of all sources of mortality, particularly during early development when starvation and predation risks are highest. As the sizes of physiological biomass compartments are obtained from empirical observations, the rates determining biomass flux are derived from process- based energetic relationships (eq. 4). Together, the allometric rate laws and the dynamic system presented in Eq. 3 allow us to assess the dynamics of consumer-resource systems for mammalian herbivores spanning the observed range of terrestrial body sizes, from the smallest (the Etruscan shrew at roughly $1$ g) to the largest (the Oligocene paraceratheres and Miocene deinotheres at ca. $1.5-1.74\times 10^{7}$ g) [Smith:2010p3442]. We next examine how this minimal framework is well-suited to provide general insight into several key allometric constraints that contribute to the functioning and limitations of terrestrial mammalian communities. ## Results & Discussion ### Recovering Damuth’s mass-density relationship Our consumer-resource system is related to the nutritional state model (NSM) proposed in yeakel2018dynamics, where an explicit starvation dynamic was incorporated by separating the consumer population density into ‘full’ and ‘hungry’ states. Here we eliminate the transition between these states, and because the timescales of transitioning between full and hungry states are short relative to those of reproduction, have sacrificed only a modest degree of physiological realism to enable analytical expression of steady states with additional sources of mortality. If we ignore the negligible effects of $\rho$ (see app. B), analytical expression of the consumer steady state as a function of mass – or the mass-density relationship – follows $C^{*}(M_{C})\approx\alpha kY_{C}\frac{\sigma^{\rm max}-2\lambda_{C}^{\rm max}+A}{4{\sigma^{\rm max}}^{2}},$ (7) where $A=\sqrt{8{\sigma^{\rm max}}^{2}+(\sigma^{\rm max}-2\lambda_{C}^{\rm max})}$, where $\lambda_{C}^{\rm max}$, $\sigma^{\rm max}$ and $Y_{C}$ are functions of mass $M_{C}$. Figure 1: Model predictions of mammalian steady states ($\rm inds\cdot m^{-2}$) as a function of herbivore consumer body mass $M_{C}$ (thick blue line) compared to observational data from Damuth [Damuth1987] (black points). The black line denotes the best-fit linear regression on observed densities based on ordinary least squares. Variation in steady state densities is captured by allowing the plant resource growth rate to vary as $\alpha=2.81\times 10^{-10}:2.19\times 10^{-8}\leavevmode\nobreak\ {\rm s}^{-1}$ (dark blue shaded region), and both $\alpha$ and the plant resource carrying capacity to vary as $k=2.3:34\leavevmode\nobreak\ {\rm kg/m}{}^{2}$ (light blue shaded region). The scaling of mammalian population densities was originally observed by Damuth1981, Damuth1981b as the reciprocal of energy use requirements with an exponent of ca. -3/4. Consumer-resource models parameterized using allometric relationships can effectively predict this mass-density relationship [Yodzis:1992hg, DeLong:2012fjb, yeakel2018dynamics], while the addition of predator-prey size ratios and consumer capture relationships enable similar predictions at higher trophic levels [weitz2006size, DeLong2012carnivores]. By integrating dimensional scaling into search and consumption rates, Pawar2012 captured the mass-density relationship while highlighting potential instabilities arising in 3-Dimensional (aquatic) environments. Our approach differs from most prior efforts by deriving timescales associated with reproduction and mortality directly from the energetic trade-offs associated with somatic growth and maintenance. After substituting allometric relationships into the rate laws in Eq. 3, we observe that the internal steady state of consumer densities in our framework is very close, though slightly elevated, to observed mammalian densities, similarly approximating Damuth’s Law (blue line, fig. 1). Our predicted mass-density relationship is premised on the assumption of resource growth rates and carrying capacities characteristic of grasses (tab. 1), contributing to the slightly elevated mass-density relationship compared to the observed best-fit (black line, fig. 1). As the resource growth rate and carrying capacity are in the numerator of eq. 7, they determine the intercept of the relationship such that lower values will more closely match observed densities. Along these lines, incorporating observed ranges of $\alpha$ and $k$ reveal strong alignment between model predictions and the variability of empirical mammalian densities (fig. 1; see app. B for details). Compared to the NSM [yeakel2018dynamics], and similar to DeLong2012carnivores, our prediction reveals exaggerated densities for smaller-bodied consumers, though within the observed range of variation, resulting in a predicted mass-density relationship with a steeper slope than expected. An elevated mass-density slope is not observed when explicit starvation and recovery are included [yeakel2018dynamics], suggesting these dynamics play an important role in depressing the populations of smaller-bodied species, in particular. While eq. 7 cannot be readily expressed when allometric relationships are included, for larger body sizes the maximal starvation rate $\sigma^{\rm max}\propto M_{C}^{-0.3}$, the yield coefficient $Y_{C}\propto M_{C}^{-1/4}$, the maximal consumer growth rate $\lambda_{C}^{\rm max}\propto M_{C}^{-1/4}$, and the quantity $A\propto M_{C}^{-0.37}$. For larger body masses, this results in a predicted mass-density relationship $\propto M_{C}^{-0.82}$ inds/m2, only slightly steeper than Damuth’s mass-density relationship $\propto M_{C}^{-0.77}$ inds/m2. At unrealistically large body sizes, the consumer steady state encounters a vertical asymptote [also noted in yeakel2018dynamics]. In this region, the superlinear body fat allometry (tab. 1) predicts the organism to be 100% fat, such that the starvation timescale is infinite. While this is mathematically entertaining, we restrict our interpretations to realistic body size ranges, thereby avoiding this particular physiological singularity. We examine additional effects of altered vital rates on the slope and intercept of the mass-density relationship in app. B. ### Senescence and starvation have a larger impact on smaller consumers We first consider two internal sources of mortality: that due to the effects of aging, where mortality changes with an organism’s temporal state, and that due to starvation, where mortality scales with an organism’s energetic state. To understand the effect of changes to $\mu(M_{C})$ on consumer steady states, we examine variations in the principle components of $\mu$: initial cohort mortality $q_{0}$ and actuarial mortality $q_{a}$. The initial cohort mortality represents the mortality experienced by a cohort prior to accruing effects of age. We observe that the mortality rate changes proportionally with $q_{0}$ independent of consumer mass, where the ratio $\mu/\lambda_{C}^{\rm max}<1$ even with respect to large increases in $q_{0}$, unless $q_{a}$ is similarly magnified (fig. 2A,B). For survivorship mortality to approach the rate of reproduction ($\mu/\lambda_{C}^{\rm max}=1$), where perceptible declines in population densities result, the initial cohort mortality must increase by roughly an order of magnitude (shaded region in fig. 2C). Due to the steepness of the scaling of $\mu$ relative to $\lambda_{C}^{\rm max}$, this effect is felt exclusively by small-bodied organisms. Figure 2: Changes in natural mortality as a function of initial cohort mortality $q_{0}$, and actuarial aging, $q_{a}$ for two different consumer body masses, $M_{C}$. A,B. The ratio reproduction $\lambda_{C}^{\rm max}$ to natural mortality $\mu$ for a mammalian herbivore of A. $M_{C}=10^{2}$ g and B. $M_{C}=10^{6}$ g, across proportional changes to the initial cohort mortality rate $q_{0}$ and the actuarial aging rate $q_{a}$. The black contour denotes $\mu/\lambda=1$. C,D. Natural mortality $\mu$ (green) relative to reproduction $\lambda_{C}^{\rm max}$ (orange) as a function of consumer body mass $M_{C}$. The range of variation (light green shaded region) shows proportional changes to the C. cohort mortality rate $q_{0}$ and the D. actuarial aging rate $q_{a}$ from -0.99 to 10. Actuarial mortality represents the cumulative effects of aging, or senescence, across the organism’s expected lifetime. We observe that as $q_{a}$ increases, the magnitude of mortality increases disproportionately (fig. 2A,B), while the slope of $\mu(M_{C})$ becomes more shallow (fig. 2D), primarily due to the cumulative nature of senescence magnifying its effects across the longer lifetimes of larger mammals. As such, an increase in $q_{a}$ overwhelms reproduction such that $\mu/\lambda_{C}^{\rm max}>1$, resulting in population instability (fig. 2A,B). The extinction risk imposed by senescence has been explored across mammalian taxa, and while some life-history characteristics such as the inter-birth interval appear to correlate strongly with these risks, the role of body size is notably ambiguous [robert2015actuarial]. Though our model – which considers averaged effects across terrestrial mammals – predicts that the risks of increased actuarial mortality are disproportionately felt by smaller size-classes, we also show that $\mu(M_{C})$ increasingly resembles $\lambda_{C}^{\rm max}(M_{C})$ with increasing $q_{a}$ (the top border of the shaded region in fig. 2D). This increased similarity implies that relatively small variations in other demographic processes or interactions may have potentially large and destabilizing effects on population size that cannot be predicted from body mass, a potential source for the noted ambiguity between size and actuarial extinction risk [robert2015actuarial]. Figure 3: The relative change in consumer steady state $\Delta C^{*}_{s}$ as a function of consumer body mass $M_{C}$ given an altered rate of starvation $\sigma(R)\cdot(1+\chi_{s})$ across the proportional change $\chi_{s}\in(-0.99,1)$. While the temporal state of an organism is unidirectional and linear, other internal states, such as an organism’s energetic state, fluctuate nonlinearly over time. In this case, the rate of starvation is low when resources become plentiful ($R\rightarrow k$) and increases to $\sigma^{\rm max}$ as resources become scarce ($R\rightarrow 0$). Because organisms metabolize their fat and muscle tissue during starvation, and die from starvation when these energetic stores are metabolized, the timescale of starvation varies with the amount of endogenous energetic stores an organism carries. Larger organisms carry a larger proportion of body mass as fat [Lindstedt:2002td], such that they are more protected from the effects of short-term resource scarcity [Millar:1990p923]. We observe this effect by modifying the starvation rate and examining how the steady state population size is altered. We introduce variation to the rate of starvation as $\sigma(R)\cdot(1+\chi_{s})$, from which the altered steady state $C^{*}_{s}$ is calculated. The relative change in steady states introduced by the altered starvation rate is then given by $\Delta C^{*}_{s}=(C^{*}_{s}-C^{*})/C^{*}$, where positive values indicate a relative gain in steady state densities from the proportional change $\chi_{s}$, and negative values indicate a relative loss (fig. 3). We observe that, while all mammals benefit from reduced starvation rates ($\chi_{s}<0$), smaller-bodied mammals benefit to a much greater extent, and this effect tapers off with increasing body mass. Because fat biomass scales super- linearly with body mass, the populations of larger consumers are more resilient to the effects of starvation, whereas those of smaller consumers are more prone. An organism’s rate of starvation emerges from two governing forces – the amount of energy storage and the rate of its use – and as such can be be manipulated both physiologically and behaviorally. For instance, behaviorally supplementing endogenous fat stores with exogenous caches magnifies an individual’s energetic stores [Lucas1991, yeakel2020caching], whereas physiologically-mediated responses to starvation risk such as torpor can introduce significant temporal delays to the effects of resource scarcity [schubert2010daily]. In both cases the time required to pass from a replenished to a starved state is effectively increased, lowering the rate of starvation. The predicted benefits of such adaptations to mammalian steady state densities will be realized primarily by smaller mammals (fig. 3, app. B), and it is the smaller body size range where traits such as caching and torpor are most commonly observed [geiser1998evolution, smith1984evolution, yeakel2020caching]. ### Predation mortality and the feasibility of megatrophic interactions Predators introduce an external source of mortality on prey populations, fueling their own population growth in whole (trophic specialists) or in part (trophic generalists), by the rate at which prey are consumed. We account for the effects of an implicit predator density $P$ with body size $M_{P}$ on the herbivore consumer density $C$ with body size $M_{C}$, where we assume the predator population to exist at a fixed density $P\equiv P^{*}$. The mortality rate of the herbivore consumer from an external predator is given by $\beta(C,P)=w\frac{\lambda_{P}(C)P}{CY_{P}},$ (8) where $\lambda_{P}(C)$ is the growth rate of the predator and $Y_{P}$ is the predator yield coefficient, describing the grams of predator produced per gram of prey consumed, and $w$ is the predation intensity. Mirroring the calculation of the consumer yield coefficient, $Y_{P}=M_{C}E_{C}/B_{\lambda_{P}}$, where $E_{C}$ is the energy density of consumable biomass carried by herbivore prey, and the $B_{\lambda_{P}}$ is the lifetime energy requirement of the predator (app. C). Assuming a linear functional response for predation mortality, $\lambda_{P}(C)$ is maximized when the consumer reaches its theoretical maximum population density, which we calculate by converting the resource carrying capacity directly to grams of consumer produced, or $C^{\rm max}=Y_{C}k$. While this is an ultimately unattainable theoretical bound, it allows for a direct calculation of the predator growth rate as a function of $C$, written as $\lambda_{P}(C)=\lambda_{P}^{\rm max}\frac{C}{C^{\rm max}}=\lambda_{P}^{\rm max}\frac{C}{Y_{C}k},$ (9) where $\lambda_{P}^{\rm max}=\ln(\nu)/t_{\lambda_{P}}$ is the maximum predator growth rate, given $\nu=2$, and $t_{\lambda_{P}}$ is the time required for the predator to reach maturity (following eq. 4). The theoretical boundary density for herbivore consumers $C^{\rm max}$ can similarly be used to calculate the boundary density for predators, $P^{\rm max}=Y_{P}C^{\rm max}$, both of which accurately capture the upper-bounds of herbivore and carnivore mass-density observations (dashed lines in fig. 4A). Because the effects of the predator are implicit, we assume that the predator population remains at empirically measured steady state densities for mammalian carnivores, where $P^{*}=p_{0}M_{P}^{p_{1}}$ given $p_{0}=8.62\times 10^{-4}\leavevmode\nobreak\ {\rm inds}^{1-p_{1}}/{\rm m^{2}}$ and $p_{1}=-0.88$ [carbone2002common]. As we are employing this framework to evaluate longer-term evolutionary consequences, this condition assumes that predator densities do not have long- term feasibility if they stray far from $P^{*}$. The predation mortality rate depends on both the body size of the herbivore consumer and its respective predator. Trophic interactions are constrained by body size [Sinclair2003, Brose2005, Hatton:2015fk], though the nature of the predator-prey mass relationship (PPMR) varies across communities [barnes2010global] and size classes [Carbone:1999ju, Brose2005, Rohr2010, riede2011stepping, pires2015pleistocene, nakazawa2017individual]. Compellingly, PPMRs for many clades can be predicted from the scaling of handling time [delong2020], suggesting that the signatures of body size evolution has cascading effects on community structure and function. While prior work has largely focused on the expected prey mass for a given predator mass, because our framework is prey-centric we require a prediction of the expected predator mass $M_{P}$ given an herbivore of body size $M_{C}$. For larger predators and prey ($>10^{5}$ g), the expected predator mass given a particular herbivore mass follows roughly ${\rm E}\\{M_{P}\\}=v_{0}M_{C}^{v_{1}}$, where $v_{0}=9.76\times 10^{3}$ g${}^{1-v_{1}}$ and $v_{1}=0.21$ [fig. 4B; see app. C; hayward2005lion, Hayward2006hyena, hayward2006leopard, hayward2006lycaon, hayward2006cheetah, Hayward2008]. [Here and throughout the prefix ‘mega’ is used to signify size classes $>5\times 10^{5}$ g; hayward2005lion]. Accordingly, larger terrestrial herbivores tend to suffer mortality from proportionately smaller predators, an asymmetry that becomes more pronounced with increasing size [cf. Sinclair2003]. We note that smaller terrestrial predator/prey size classes tend to be much larger than prey [e.g. rodent- or insect-specialist mesocarnivores; cruz2022geography, Cruz2022], also captured by the ${\rm E}\\{M_{P}\\}$ scaling. Figure 4: The feasibility of predation. A. Empirical mammalian herbivore (blue points) and predator (red points) mass-densities shown alongside the theoretical maximum herbivore $C^{\rm max}$ and predator $P^{\rm max}$ densities across body size (dashed lines). The solid blue curves denote the predicted herbivore consumer steady state $C^{*}(M_{C})$ with predation mortality given high ($w=1$) and low ($w=0.37$) predation intensities. B. Predator body mass as a function of prey body mass observed among contemporary mammalian fauna provides the allometric predator-prey mass relationship (PPMR). The green line denotes the best-fit, where $M_{P}=9.76\times 10^{3}M_{C}^{0.21}$ g. C. Threshold herbivore mass $M_{C}^{\dagger}$ given changes to the PPMR intercept $\chi_{\rm int}$ and slope $\chi_{\rm slope}$ (see Eq. A25 and legend in B.). White shaded region highlights the mass range enabling feasible megatrophic interactions, shown in D. for herbivore mass thresholds $M_{C}^{\dagger}$ with the expected mass of associated carnivores ${\rm E}\\{M_{P}\\}$, assuming alternative predation intensities. Integrating the large-bodied PPMR into the predation mortality rate reveals the emergence of a dynamic instability at megaherbivore size classes (fig. 4A,B), the product of a transcritical bifurcation at consumer mass $M_{C}^{\dagger}$ (app. C), and similar to the more general instability documented by weitz2006size. An implicit predator population with body size ${\rm E}\\{M_{P}\\}$ is thus able to withdraw sufficient biomass from an herbivore population – without crashing the herbivore population – below a threshold herbivore size of $M_{C}^{\dagger}=2.58\times 10^{6}$ g (fig. 4A). Above this critical size threshold, the herbivore population has such low densities that it is unable to sustain a specialist predator species large enough to consume it, introducing a strong upper-bound to mammalian carnivore body size driven by a trophic cascade. This boundary matches the herbivore maximum size limit observed in contemporary terrestrial systems [Sinclair2003], at roughly the size of an elephant (fig. 4B; app. C), though the exact placement of $M_{C}^{\dagger}$ varies with the resource growth rate and carrying capacity. While we have assumed values representative of grass resources, decreasing $\alpha$ and/or $k$ lowers the steady state mass-density intercept (eq. 7), setting the mass threshold at a lower body size (app. C). This means that lower-productivity environments, or environments subject to large and long-term oscillations in productivity, may be expected to have more severe limitations on feasible megaherbivore sizes. And while we do not consider stochastic or transient effects directly, we may also assume actualized extinction risk to emerge at smaller-than-predicted masses where transient or stochastic effects may push populations below the point of recovery. $M_{C}^{\dagger}$ marks the threshold herbivore mass above which predation is unsustainable, though Sinclair2003 revealed contemporary herbivores to escape predation at ca. $4.22\times 10^{5}$ g. This change-point reflects the limitations of contemporary carnivores, which reach a maximum body size of $1.15$ to $2.60\times 10^{5}$ g [Sinclair2003], and have preferences for prey up to $5.50\times 10^{5}$ g [hayward2005lion]. Importantly, the sole predators of contemporary giants are not megaherbivore specialists, instead opportunistically subsidizing their preferred prey with larger taxa. While we have so far assumed a predator-prey interaction where the entirety of predator growth is fueled by the focal herbivore, the largest predators in natural systems tend be dietary generalists [Sinclair2003, gross2009generalized]. We observe that reducing the predation intensity (such that $w<1$) increases $M_{C}^{\dagger}$ to a larger threshold mass (app. C). For example, $w=0.37$ increases the herbivore body mass boundary to $M_{C}^{\dagger}=1.75\times 10^{7}$ g (fig. 4B; app. C), roughly the body mass attained by the largest terrestrial herbivores, the Oligocene paraceratheres and Miocene deinotheres [Smith:2010p3442]. That the threshold herbivore mass decreases with increasing predation intensity suggests that larger predators are dynamically constrained to be dietary generalists [Sinclair2003], while also pointing to an amplifying feedback mechanism [brook2008synergies] that may operate in diverse communities undergoing megafaunal extinctions. As megaherbivore species are lost, the largest predators must respond by increasing the intensity of predation on those remaining. Our results suggest that this energetic redirection reduces the threshold herbivore mass $M_{C}^{\dagger}$ to lower size classes, increasing the likelihood of additional extinctions and attendant increases in predation intensity on survivors. Together, this demonstrates a dynamic mechanism for the previously proposed influence of top- down dietary ratcheting hypothesized for the Pleistocene extinctions, in particular [kay2002false, ripple2010linking]. Adaptations to hypercarnivorous strategies are difficult to reverse on macroevolutionary time scales, resulting in the so-called ‘hypercarnivore ratchet’ [VanValkenburgh1991, Holliday2004]. Moreover, there is a strong correlation between hypercarnivorous adaptations and body size among terrestrial carnivores, the combination of which may promote vulnerability to extinction [VanValkenburgh:2004p2451]. While deinotheres and paraceratheres top the megaherbivore scale, the Eocene artiodactyl _Andrewsarchus_ may have been the largest terrestrial mammalian predator at ca. $1\times 10^{6}$ g [burness2001dinosaurs], while the Miocene Hyaenodontid _Megistotherium osteothlastes_ ranged between $5$ to $8\times 10^{5}$ g and the early Eocene Oxyaenodont _Sarkastodon mongoliensis_ weighed ca. $8\times 10^{5}$ g [sorkin2008biomechanical]. A theoretical maximum mammalian carnivore size of $1.1\times 10^{6}$ g has been proposed based on the intersection of daily energetic uptake requirements against metabolic expenditures [Carbone:2007dz], closely aligning with the largest known megapredators. While our consumer- resource framework provides a range of predicted megaherbivore body mass thresholds depending on the fraction of predator growth it fuels, we next ask under what conditions megatrophic relationships between megaherbivores and megapredators are dynamically feasible. A central relationship in our framework is the allometric PPMR observed for the largest contemporary herbivores and carnivores, however empirical observations of PPMRs reveal tremendous variability both across and within clades [delong2020], and are completely unknown for megatrophic interactions in terrestrial systems (fig. 4B; app. C). While it is unknown whether these super-sized carnivores were specialists on deinothere size-classes, our framework allows us to investigate whether and to what extent changes to the contemporary PPMR enable megatrophic interactions (fig. 4C,D). To examine this, we allow the expected predator mass given a particular prey mass to vary as ${\rm E}\\{M_{P}\\}=v_{0}(1+\chi_{\rm int})M_{C}^{v_{1}(1+\chi_{\rm slope})},$ (10) where the proportional changes in the PPMR intercept and slope are given by $\chi_{\rm int}$ and $\chi_{\rm slope}\in(-0.99,2)$ (see the legend in fig. 4B). We note that while $\chi_{\rm slope}$ explores changes to the inferred steepness of the PPMR, $\chi_{\rm int}$ explores variation around the scaling relationship for a particular prey mass. We observe that only a small range of values for PPMR intercepts and slopes permit the existence of dynamically feasible megatrophic interactions, where megaherbivores serve as prey for specialized megapredators (white band in fig. 4C,D; app. C). Such interactions could be realized if the PPMR in fig. 4B had an increased intercept or alternatively both a lower intercept and higher slope for mega size-classes. That alternative PPMRs could characterize different size-classes across foraging guilds has been previously examined [vezina1985empirical, Carbone:1999ju], and the clear disconnect between that shown for contemporary large-bodied mammals and those in the megatrophic range (fig. 4B) supports this notion. When predation intensity is high ($w=1$), the PPMR enabling feasible megatrophic interactions lowers the herbivore mass threshold $M_{C}^{\dagger}$, resulting in megapredators consuming relatively smaller prey. However, if predation intensity is lowered ($w=0.37$), we observe feasible megatrophic interactions for size-classes capturing megaherbivores and megapredators at their largest documented sizes in the fossil record (fig. 4D; app. C). That decreased predation intensity enables feasibility of the largest mammalian size-classes agrees with previous conjectures that the largest mammalian terrestrial predators were likely dietary generalists [farlow1993rareness]. Our framework thus highlights dynamic constraints existing between predators and prey that may serve to structure mammalian communities over evolutionary time, in particular revealing the susceptibility of megaherbivores to perturbations. As carnivorous clades evolved body sizes enabling megaherbivore predation, their super-sized appetites may have suppressed megaherbivores to unsustainable densities where the risk of extinction became overwhelming – an evolutionary trap marking the final tooth in the hypercarnivore ratchet [cf. VanValkenburgh:2004p2451]. ### Harvesting to extinction We last consider the effects of anthropogenic harvest-induced mortality on herbivore populations. While the predation rate is naturally limited by the energetic needs of the predator, we consider harvest to be a comparatively unconstrained source of mortality. This may be the case if the human population(s) engaged in harvesting are subsidized by alternative resources [Brook2005]. Harvest pressure has potentially varying relationships with consumer (prey) body mass, a complex product of environment, climate, culture, and technology [churchill1993weapon]. For example, hunting traditions specializing in mass-collecting, by way of trapping or netting [churchill1993weapon, ugan2005does] are expected to exhibit harvest allometries biased towards smaller species, whereas a purely opportunistic strategy may be expected to have very little allometric dependence. While smaller mammals do not appear to offer a significant return on investment, the mass-collecting of invertebrates, such as grasshoppers, and fish can offer significant returns [ugan2005does]. In contrast, the innovation of advanced projectiles is thought to have enabled harvest of terrestrial megafauna [churchill1993weapon, prates2022changes], and archeological evidence points to many Pleistocene human populations as potential megafaunal specialists [Smith:2018gm]. Figure 5: The effects of harvest mortality on herbivore consumers. A. Proportion mortality due to an extinction-inducing harvest rate $h^{\dagger}$ without predation ($w=0$; blue line), and with low ($w=0.37$; orange line) or high predation intensity ($w=1$; red line), as a function of consumer body mass $M_{C}$. B. Harvest pressure $\psi^{\dagger}$ resulting from extinction- inducing harvest (inds/year/$A_{\rm CA}$) without predation ($w=0$; blue line), and with low ($w=0.37$; orange line) or high predation intensity ($w=1$; red line), as a function of consumer body mass $M_{C}$. Black point and line: median and range of estimated harvest rates for woolly mammoths [_Mammuthus primigenius_ ; fordham2022process]; Green point: estimated harvest pressure for the Australian _Diprotodon_ [bradshaw2021relative]; lower and higher yellow point: estimated harvest rates for contemporary _Loxodonta_ during the early 1800s and just prior to 1987, respectively [milner1993exploitation]. Because harvest scaling may be difficult to measure and idiosyncratic, we instead calculate the harvest rate required to induce extinction, $h^{\dagger}$, as a function of body size $M_{C}$, and find a scaling relationship proportional to the rate of reproduction where $h^{\dagger}\propto M_{C}^{-1/4}$. This is a natural result, as the effort required to suppress a population is expected to be proportional to its reproductive rate, reflecting the increased susceptibility of large-bodied organisms to extinction [enquist2020megabiota]. As a proportion of the other sources of consumer mortality that we have considered (excluding predation; $w=0$), extinction-level harvesting is lower for smaller consumers, saturating at close to unity at large size classes, reflecting the elevated role of starvation mortality among smaller-sized organisms (Fig 5A). With predation mortality included at both low ($w=0.37$) and high ($w=1$) intensities, extinction-level harvesting accounts for an increasingly smaller proportion of mortality for larger organisms (orange and red lines, fig. 5A). This highlights the delicate nature of the megafaunal niche, where smaller changes in mortality rates can induce population collapse [enquist2020megabiota]. To examine how our estimate of extinction-level harvesting rates $h^{\dagger}$ compare to those estimated for human hunting of paleontological and historical mammalian populations, we converted $h^{\dagger}$ to harvest pressure $\psi^{\dagger}$, or the number of individuals harvested per year to reduce the population to a fraction of its steady state $\epsilon C^{*}$ where we set $\epsilon=0.01$. We calculate $\psi^{\dagger}$ for an arbitrary area (see app. D), which we standardize to the area of California ($A_{\rm CA}=4.24\times 10^{5}\leavevmode\nobreak\ {\rm km}^{2}$), such that $\psi^{\dagger}\propto-h^{\dagger}\frac{C^{*}(1-\epsilon)}{M_{C}\log(\epsilon)}.$ (11) Though the annual harvesting pressure is unrealistically high for smaller organisms, we observe that it is ca. $4.3\times 10^{3}\leavevmode\nobreak\ {\rm inds/yr/}A_{\rm CA}$ for elephant-sized mammals (ca. $2.5\times 10^{6}$ g) in the absence of predation mortality ($w=0$). With increasing predation intensity, the harvest pressure required to induce extinction is much less for these larger consumers (orange and red lines in fig. 5B). We note that this calculation of harvest pressure should be viewed as a minimum estimate given that we do not account for demographic rebound. As such, this measure is appropriate only if the timescale of harvest is less than the generational timescale, which is the case for the megafauna considered here. Our predictions of extinction-inducing harvest pressure compare well with paleontological and historical estimates of harvest pressure on mammalian megafauna (fig. 5B; see app. D). For example, using a formulation similar to that of alroy2001multispecies, fordham2022process estimate the harvest pressure required to collapse mammoth (_Mammuthus primigenius_) populations, revealing a range of values consistent with our expectation for similar size- classes (est. $\psi^{\dagger}=$ ca. $1.24\times 10^{4}\leavevmode\nobreak\ {\rm inds/yr/}A_{\rm CA}$), as did estimates of extinction-inducing harvest of the Australian _Diprotodon_ [est. $\psi^{\dagger}=$ ca. 763 ${\rm inds/yr/}A_{\rm CA}$; bradshaw2021relative]. Within the historical record, elephant (_Loxodonta_) populations experienced comparatively lower harvest pressure through 1850 [ca. $466\leavevmode\nobreak\ {\rm inds/yr/}A_{\rm CA}$, derived from the volume of ivory exports; milner1993exploitation]. While fluctuating over the next century, harvest pressure increased to a maximum of ca. $13.3\times 10^{5}\leavevmode\nobreak\ {\rm inds/yr/}A_{\rm CA}$ just prior to 1987 (fig. 5B). This level of harvest was not sustained, as ivory export volume plummeted following the implementation of trade restrictions in 1989 [milner1993exploitation]. Both the fordham2022process estimate for Pleistocene mammoths and the short-lived harvest maximum for African elephants in 1987 [milner1993exploitation] achieved pressures greater than $\psi^{\dagger}$ under the conservative assumption of no natural predation (fig. 5B). While estimates for Diprotodon harvest are considerably lower [bradshaw2021relative], it is important to note that our framework is parameterized for eutherian rather than marsupial mammals. Nevertheless, the estimated _Diprotodon_ $\psi^{\dagger}$ is well within range of extinction- inducing harvest rates if natural predation pressures are also included, and there is evidence to suggest that _Diprotodon_ likely served as prey for marsupial lions [horton1981cuts, wroe1999estimating], and both giant crocodylians (_Pallimnarchus_ spp.) and varanid lizards [_Megalania_ spp.; webb2009late]. ## Conclusion We have shown that the inclusion of mass-specific energetic transfer between resources and consumers, combined with the unique timescales governing consumer mortality, both predict Damuth’s Law [Damuth1987] and provide insight into dynamic thresholds constraining populations. While natural and starvation mortality primarily impact small-bodied species, trophic mortality primarily impacts large-bodied species with longer generational timescales. Moreover, while mass-specific predation gives rise to dynamic thresholds for herbivore populations, these effects are sensitive to both predation intesnity as well as the associated predator-prey mass relationship, which isn’t well understood in terrestrial ecosystems [nakazawa2017individual]. While assessment of particular communities and/or species requires more detailed approaches – integrating, for example, life history dynamics as in bradshaw2021relative – we suggest that a lower-dimensional framework is useful for extracting general, first-order energetic constraints that both shape and potentially limit the nature of mammalian communities. That extinction risk appears to increase with body size [Cardillo:2005et] is integral to our understanding of the Pleistocene extinctions [alroy2001multispecies, johnson2002determinants, brook2008synergies, Smith:2018gm, bradshaw2021relative] and anthropogenic effects throughout the Holocene [Estes:2011eo]. Because megafaunal loss may have disproportionately large impacts on ecosystem functioning [enquist2020megabiota], understanding the mechanistic drivers that may lead these species to the brink is of paramount importance. Assessing which energetic walls close in and why as body size increases, is a fundamental aspect of reconciling the nature of extinction [brook2008synergies], particularly when there is size-selectivity [Smith:2018gm]. That we observe dynamically-feasible megatrophic interactions to occupy a narrow band of predator-prey mass relationships points to a broader range of interaction structures than are realized in contemporary communities. As the threshold consumer mass decreases with increased predation intensity, how megafaunal trophic structure changes during extinction cascades may be central for understanding the dynamics of community disassembly [Yeakel2014]. And while these dynamics may arise naturally from the energetic limitations of mammalian interactions, it may be that the added pressure of subsidized harvest, particularly on megafauna, inevitably leads to collapse. Table 1: Model parameters and values/units Definition | Parameter | Value/Units ---|---|--- Resource | | density | $R$ | ${\rm g/m^{2}}$ reproduction rate | $\alpha$ | $9.49\times 10^{-9}\leavevmode\nobreak\ ({\rm 1/s})$ carrying capacity | $k$ | $23\times 10^{3}\leavevmode\nobreak\ ({\rm g/m^{2}})$ energy density | $E_{d}$ | $1.82\times 10^{4}$ (J/g) Consumer | | density | $C$ | ${\rm g/m^{2}}$ body mass | $M_{C}$ | g timescale of growth from $m_{1}$ to $m_{2}$ | $\tau(m_{1},m_{2})$ | s reproduction rate | $\lambda_{C}^{\rm max}$ | ${\rm 1/s}$ yield coefficient | $Y_{C}$ | $({\rm g/m^{2}\leavevmode\nobreak\ }C)/({\rm g/m^{2}\leavevmode\nobreak\ }R)$ maintenance rate | $\rho$ | ${\rm 1/s}$ natural mortality rate | $\mu$ | ${\rm 1/s}$ starvation rate | $\sigma^{\rm max}$ | ${\rm 1/s}$ harvest rate | $h$ | ${\rm 1/s}$ Predator | | steady state density1 | $P^{*}$ | $P_{0}M_{P}^{-0.88}\leavevmode\nobreak\ {\rm inds/m^{2}}$ body mass | $M_{P}$ | g growth rate | $\lambda_{P}^{\rm max}$ | ${\rm 1/s}$ yield coefficient | $Y_{P}$ | $({\rm g/m^{2}\leavevmode\nobreak\ }P)/({\rm g/m^{2}\leavevmode\nobreak\ }C)$ predation intensity | $w$ | (0,1) Metabolic normalization constant | $B_{0}$ | 0.047 (W g-3/4) Energy to synthesize a unit of mass | $E_{m}$ | 5774 (J g-1) Energy stored in a unit of mass | $E_{m}^{\prime}$ | 7000 (J g-1) Prop. change PPMR intercept1 | $\chi_{\rm int}$ | (-0.99,2) Prop. change PPMR slope1 | $\chi_{\rm slope}$ | (-0.99,2) Extinction-inducing harvest rate | $h^{\dagger}$ | ${\rm 1/s}$ Extinction-inducing harvest pressure | $\psi^{\dagger}$ | inds/yr/$A_{\rm CA}$ 1PPMR: Predator-Prey Mass Relationship | ## Acknowledgments We would like to thank Irina Birskis-Barros, Uttam Bhat, Jessica Blois, Nathaniel Fox, Jacquelyn Gill, Paulo Guimarães Jr., Emily Lindsey, Mathias Pires, Megha Suswaram, and Ritwika VPS for insightful comments and discussions that greatly improved the ideas and concepts that contributed to this manuscript. These ideas benefited greatly from travel funds provided to JDY from the Santa Fe Institute. This project was supported by National Science Foundation grant EAR-1623852 to JDY. ## Statement of Authorship JDY, CPK, and TR conceived of the model. JDY and TR developed the code and oversaw model analysis. All authors reviewed and edited the writing at all stages of composition. ## Data and Code Availability Code and data archived on Zenodo: https://doi.org/10.5281/zenodo.8213158 ## References * Alonzo [2002] Alonzo, S. H. 2002. State-dependent habitat selection games between predators and prey: the importance of behavioural interactions and expected lifetime reproductive success. Evol. Ecol. Res. 4:759–778. * Alroy [2001] Alroy, J. 2001. A multispecies overkill simulation of the end-pleistocene megafaunal mass extinction. Science 292:1893–1896. * Barnes et al. [2010] Barnes, C., D. Maxwell, D. C. Reuman, and S. Jennings. 2010. Global patterns in predator–prey size relationships reveal size dependency of trophic transfer efficiency. Ecology 91:222–232. * Bhat et al. [2020] Bhat, U., C. P. Kempes, and J. D. Yeakel. 2020. Scaling the risk landscape drives optimal life-history strategies and the evolution of grazing. Proceedings of the National Academy of Sciences 117:1580–1586. * Bradshaw et al. [2021] Bradshaw, C. J., C. N. Johnson, J. Llewelyn, V. Weisbecker, G. Strona, and F. Saltré. 2021. Relative demographic susceptibility does not explain the extinction chronology of sahul’s megafauna. Elife 10:e63870. * Brook and Bowman [2005] Brook, B. W., and D. M. J. S. Bowman. 2005. One equation fits overkill: why allometry underpins both prehistoric and modern body size-biased extinctions. Population Ecology 47:137–141. * Brook et al. [2008] Brook, B. W., N. S. Sodhi, and C. J. Bradshaw. 2008. Synergies among extinction drivers under global change. Trends in ecology & evolution 23:453–460. * Brose et al. [2005] Brose, U., L. Cushing, E. L. Berlow, and T. Jonsson. 2005. Body sizes of consumers and their resources. Ecology . * Burness et al. [2001] Burness, G. P., J. Diamond, and T. Flannery. 2001. Dinosaurs, dragons, and dwarfs: the evolution of maximal body size. Proceedings of the National Academy of Sciences 98:14518–14523. * Calder III [1983] Calder III, W. A. 1983. An allometric approach to population cycles of mammals. J. Theor. Biol. 100:275–282. * Carbone and Gittleman [2002] Carbone, C., and J. L. Gittleman. 2002. A common rule for the scaling of carnivore density. Science 295:2273–2276. * Carbone et al. [1999] Carbone, C., G. M. Mace, S. C. Roberts, and D. W. Macdonald. 1999. Energetic constraints on the diet of terrestrial carnivores. Nature 402:286–288. * Carbone et al. [2007] Carbone, C., A. Teacher, and J. M. Rowcliffe. 2007. The Costs of Carnivory. PLoS Biol 5:e22. * Cardillo et al. [2005] Cardillo, M., G. M. Mace, K. E. Jones, J. Bielby, O. R. P. Bininda-Emonds, W. Sechrest, C. D. L. Orme, and A. Purvis. 2005. Multiple causes of high extinction risk in large mammal species. Science 309:1239–1241. * Churchill [1993] Churchill, S. E. 1993. Weapon technology, prey size selection, and hunting methods in modern hunter-gatherers: implications for hunting in the palaeolithic and mesolithic. Archeological Papers of the American Anthropological Association 4:11–24. * Cruz et al. [2022] Cruz, L. R., R. L. Muylaert, M. Galetti, and M. M. Pires. 2022. The geography of diet variation in neotropical carnivora. Mammal Review 52:112–128. * Cruz and Pires [2022] Cruz, L. R., and M. M. Pires. 2022. Body mass ratios determine dietary patterns and help predicting predator–prey interactions of neotropical carnivora. Mammal Research . * Damuth [1981 _a_] Damuth, J. 1981 _a_. Home range, home range overlap, and species energy use among herbivorous mammals. Biol. J. Linn. Soc. 15:185–193. * Damuth [1981 _b_] ———. 1981 _b_. Population density and body size in mammals. Nature 290:699–700. * Damuth [1987] ———. 1987. Interspecific allometry of population density in mammals and other animals: the independence of body mass and population energy-use. Biol. J. Linn. Soc. 31:193–246. * de Roos [2020] de Roos, A. M. 2020. 53The impact of population structure on population and community dynamics. _In_ Theoretical Ecology: concepts and applications. Oxford University Press. * DeAngelis [1980] DeAngelis, D. 1980. Energy flow, nutrient cycling, and ecosystem resilience. Ecology 61:764–771. * DeLong [2020] DeLong, J. P. 2020. Detecting the signature of body mass evolution in the broad-scale architecture of food webs. The American Naturalist 196:443–453. PMID: 32970468. * DeLong and Vasseur [2012 _a_] DeLong, J. P., and D. A. Vasseur. 2012 _a_. A dynamic explanation of size–density scaling in carnivores. Ecology 93:470–476. * DeLong and Vasseur [2012 _b_] ———. 2012 _b_. Size-density scaling in protists and the links between consumer–resource interaction parameters. J. Anim. Ecol. 81:1193–1201. * Dunne et al. [2016] Dunne, J. A., H. Maschner, M. W. Betts, N. Huntly, R. Russell, R. J. Williams, and S. A. Wood. 2016. The roles and impacts of human hunter-gatherers in North Pacific marine food webs. Sci. Rep. 6:1–9. * Enquist et al. [2020] Enquist, B. J., A. J. Abraham, M. B. Harfoot, Y. Malhi, and C. E. Doughty. 2020\. The megabiota are disproportionately important for biosphere functioning. Nature communications 11:1–11. * Estes et al. [2011] Estes, J. A., J. Terborgh, J. S. Brashares, M. E. Power, J. Berger, W. J. Bond, S. R. Carpenter, T. E. Essington, R. D. Holt, J. B. C. Jackson, R. J. Marquis, L. Oksanen, T. Oksanen, R. T. Paine, E. K. Pikitch, W. J. Ripple, S. A. Sandin, M. Scheffer, T. W. Schoener, J. B. Shurin, A. R. E. Sinclair, M. E. Soulé, R. Virtanen, and D. A. Wardle. 2011. Trophic downgrading of planet Earth. Science 333:301–306. * Farlow [1993] Farlow, J. O. 1993. On the rareness of big, fierce animals; speculations about the body sizes, population densities, and geographic ranges of predatory mammals and large carnivorous dinosaurs. American Journal of Science 293:167. * Fordham et al. [2022] Fordham, D. A., S. C. Brown, H. R. Akçakaya, B. W. Brook, S. Haythorne, A. Manica, K. T. Shoemaker, J. J. Austin, B. Blonder, J. Pilowsky, et al. 2022\. Process-explicit models reveal pathway to extinction for woolly mammoth using pattern-oriented validation. Ecology letters 25:125–137. * Geiser [1998] Geiser, F. 1998. Evolution of daily torpor and hibernation in birds and mammals: importance of body size. Clinical and experimental pharmacology and physiology 25:736–740. * Gross et al. [2009] Gross, T., L. Rudolf, S. A. Levin, and U. Dieckmann. 2009. Generalized models reveal stabilizing factors in food webs. Science 325:747–750. * Hatton et al. [2015] Hatton, I. A., K. S. McCann, J. M. Fryxell, T. J. Davies, M. Smerlak, A. R. E. Sinclair, and M. Loreau. 2015. The predator-prey power law: Biomass scaling across terrestrial and aquatic biomes. Science 349:aac6284–aac6284. * Hayward [2006] Hayward, M. 2006. Prey preferences of the spotted hyaena (Crocuta crocuta) and degree of dietary overlap with the lion (Panthera leo). J. Zoology 270:606–614. * Hayward et al. [2006 _a_] Hayward, M., P. Henschel, J. O’Brien, M. Hofmeyr, G. Balme, and G. I. Kerley. 2006 _a_. Prey preferences of the leopard (panthera pardus). Journal of Zoology 270:298–313. * Hayward et al. [2006 _b_] Hayward, M., M. Hofmeyr, J. O’brien, and G. I. Kerley. 2006 _b_. Prey preferences of the cheetah (acinonyx jubatus)(felidae: Carnivora): morphological limitations or the need to capture rapidly consumable prey before kleptoparasites arrive? Journal of Zoology 270:615–627. * Hayward and Kerley [2008] Hayward, M. W., and G. Kerley. 2008. Prey preferences and dietary overlap amongst Africa’s large predators. S. African J. Wild. Res. 38:93–108. * Hayward and Kerley [2005] Hayward, M. W., and G. I. Kerley. 2005. Prey preferences of the lion (panthera leo). Journal of zoology 267:309–322. * Hayward et al. [2006 _c_] Hayward, M. W., J. O’Brien, M. Hofmeyr, and G. I. Kerley. 2006 _c_. Prey preferences of the african wild dog lycaon pictus (canidae: Carnivora): ecological requirements for conservation. Journal of Mammalogy 87:1122–1131. * Hennemann [1983] Hennemann, W. W. 1983. Relationship among body mass, metabolic rate and the intrinsic rate of natural increase in mammals. Oecologia 56:104–108. * Holliday and Steppan [2004] Holliday, J., and S. Steppan. 2004. Evolution of hypercarnivory: the effect of specialization on morphological and taxonomic diversity. Paleobiology 30:108. * Horton and Wright [1981] Horton, D. R., and R. V. Wright. 1981. Cuts on lancefield bones: carnivorous thylacoleo, not humans, the cause. Archaeology in Oceania 16:73–80. * Hou et al. [2008] Hou, C., W. Zuo, M. E. Moses, W. H. Woodruff, J. H. Brown, and G. B. West. 2008\. Energy Uptake and Allocation During Ontogeny. Science 322:736–739. * Johnson [2002] Johnson, C. N. 2002. Determinants of loss of mammal species during the late quaternary ’megafauna’ extinctions: life history and ecology, but not body size. Proceedings of the Royal Society of London. Series B: Biological Sciences 269:2221–2227. * Kay [2002] Kay, C. E. 2002. False gods, ecological myths, and biological reality. Wilderness and political ecology: aboriginal influences and the original state of nature. University of Utah Press, Salt Lake City pages 238–261. * Kempes et al. [2012] Kempes, C. P., S. Dutkiewicz, and M. J. Follows. 2012. Growth, metabolic partitioning, and the size of microorganisms. PNAS 109:495–500. * Lindstedt and Schaeffer [2002] Lindstedt, S. L., and P. J. Schaeffer. 2002. Use of allometry in predicting anatomical and physiological parameters of mammals. Lab. Anim. 36:1–19. * Lucas and Walter [1991] Lucas, J. R., and L. R. Walter. 1991. When should chickadees hoard food? theory and experimental results. Animal Behaviour 41:579–601. * Millar and Hickling [1990] Millar, J., and G. Hickling. 1990. Fasting Endurance and the Evolution of Mammalian Body Size. Funct. Ecol. 4:5–12. * Milner-Gulland and Beddington [1993] Milner-Gulland, E., and J. Beddington. 1993. The exploitation of elephants for the ivory trade: an historical perspective. Proceedings of the Royal Society of London. Series B: Biological Sciences 252:29–37. * Moses et al. [2008] Moses, M. E., C. Hou, W. H. Woodruff, G. B. West, J. C. Nekola, W. Zuo, and J. H. Brown. 2008. Revisiting a Model of Ontogenetic Growth: Estimating Model Parameters from Theory and Data. http://dx.doi.org.proxy.lib.sfu.ca/10.1086/679735 171:632–645. * Murdoch et al. [2003] Murdoch, W. W., C. J. Briggs, and R. M. Nisbet. 2003. Consumer-resource Dynamics. Monographs in population biology. Princeton University Press. * Nakazawa [2017] Nakazawa, T. 2017. Individual interaction data are required in community ecology: a conceptual review of the predator–prey mass ratio and more. Ecological Research 32:5–12. * Pawar et al. [2012] Pawar, S., A. I. Dell, and V. M. Savage. 2012. Dimensionality of consumer search space drives trophic interaction strengths. Nature . * Pires et al. [2015] Pires, M. M., P. L. Koch, R. A. Fariña, M. A. de Aguiar, S. F. dos Reis, and P. R. Guimarães Jr. 2015. Pleistocene megafaunal interaction networks became more vulnerable after human arrival. Proceedings of the Royal Society B: Biological Sciences 282:20151367. * Pirt [1965] Pirt, S. 1965. The maintenance energy of bacteria in growing cultures. Proc. Roy. Soc. B 163:224. * Prates et al. [2022] Prates, L., D. Rivero, and S. I. Perez. 2022. Changes in projectile design and size of prey reveals the role of fishtail points in megafauna hunting in south america . * Riede et al. [2011] Riede, J. O., U. Brose, B. Ebenman, U. Jacob, R. Thompson, C. R. Townsend, and T. Jonsson. 2011. Stepping in elton’s footprints: a general scaling model for body masses and trophic levels across ecosystems. Ecology letters 14:169–178. * Ripple and Van Valkenburgh [2010] Ripple, W. J., and B. Van Valkenburgh. 2010. Linking top-down forces to the pleistocene megafaunal extinctions. BioScience 60:516–526. * Robert et al. [2015] Robert, A., S. Chantepie, S. Pavard, F. Sarrazin, and C. Teplitsky. 2015. Actuarial senescence can increase the risk of extinction of mammal populations. Ecological Applications 25:116–124. * Roff [1993] Roff, D. 1993. Evolution Of Life Histories: Theory and Analysis. Springer US. * Rohr et al. [2010] Rohr, R. P., H. Scherer, P. Kehrli, C. Mazza, and L.-F. Bersier. 2010. Modeling food webs: Exploring unexplained structure using latent traits. Am. Nat. 176:170–177. * Savage et al. [2004] Savage, V. M., J. F. Gillooly, J. H. Brown, G. B. West, and E. L. Charnov. 2004\. Effects of Body Size and Temperature on Population Growth. http://dx.doi.org.proxy.lib.sfu.ca/10.1086/679735 163:429–441. * Schubert et al. [2010] Schubert, K. A., A. S. Boerema, L. M. Vaanholt, S. F. de Boer, A. M. Strijkstra, and S. Daan. 2010. Daily torpor in mice: high foraging costs trigger energy-saving hypothermia. Biology letters 6:132–135. * Sinclair et al. [2003] Sinclair, A. R. E., S. Mduma, and J. S. Brashares. 2003. Patterns of predation in a diverse predator–prey system. Nature 425:288–290. * Smith and Reichman [1984] Smith, C., and O. Reichman. 1984. The evolution of food caching by birds and mammals. Annual Review of Ecology and Systematics 15:329–351. * Smith et al. [2010] Smith, F., A. Boyer, J. Brown, and D. Costa. 2010. The Evolution of Maximum Body Size of Terrestrial Mammals. Science . * Smith et al. [2018] Smith, F. A., R. E. Elliott Smith, S. K. Lyons, and J. L. Payne. 2018. Body size downgrading of mammals over the late Quaternary. Science 360:310–313. * Sorkin [2008] Sorkin, B. 2008. A biomechanical constraint on body mass in terrestrial mammalian predators. Lethaia 41:333–347. * Ugan [2005] Ugan, A. 2005. Does size matter? body size, mass collecting, and their implications for understanding prehistoric foraging behavior. American Antiquity 70:75–89. * Van Valkenburgh [1991] Van Valkenburgh, B. 1991. Iterative evolution of hypercarnivory in canids (mammalia: Carnivora): evolutionary interactions among sympatric predators. Paleobiology 17:340–362. * van Valkenburgh et al. [2004] van Valkenburgh, B., X. Wang, and J. Damuth. 2004. Cope’s rule, hypercarnivory, and extinction in North American canids. Science 306:101. * Vézina [1985] Vézina, A. F. 1985. Empirical relationships between predator and prey size among terrestrial vertebrate predators. Oecologia 67:555–565. * Webb [2009] Webb, S. 2009. Late quaternary distribution and biogeography of the southern lake eyre basin (sleb) megafauna, south australia. Boreas 38:25–38. * Weitz and Levin [2006] Weitz, J. S., and S. A. Levin. 2006. Size and scaling of predator–prey dynamics. Ecology letters 9:548–557. * West et al. [2001] West, G. B., J. H. Brown, and B. J. Enquist. 2001. A general model for ontogenetic growth. Nature 413:628–631. * West et al. [2002] West, G. B., W. H. Woodruff, and J. H. Brown. 2002. Allometric scaling of metabolic rate from molecules and mitochondria to cells and mammals. Proc. Natl. Acad. Sci. USA 99 Suppl 1:2473–2478. * Wroe et al. [1999] Wroe, S., T. Myers, R. Wells, and A. Gillespie. 1999. Estimating the weight of the pleistocene marsupial lion, thylacoleo carnifex (thylacoleonidae: Marsupialia): implications for the ecomorphology of a marsupial super-predator and hypotheses of impoverishment of australian marsupial carnivore faunas. Australian Journal of Zoology 47:489–498. * Yeakel et al. [2020] Yeakel, J. D., U. Bhat, and S. D. Newsome. 2020. Caching in or falling back at the sevilleta: the effects of body size and seasonal uncertainty on desert rodent foraging. The American Naturalist 196:241–256. * Yeakel et al. [2018] Yeakel, J. D., C. P. Kempes, and S. Redner. 2018. Dynamics of starvation and recovery predict extinction risk and both damuth’s law and cope’s rule. Nature communications 9:1–10. * Yeakel et al. [2014] Yeakel, J. D., M. M. Pires, L. Rudolf, N. J. Dominy, P. L. Koch, P. R. Guimarães Jr, and T. Gross. 2014. Collapse of an ecological network in Ancient Egypt. Proceedings of the National Academy of Sciences 111:14472–14477. * Yodzis and Innes [1992] Yodzis, P., and S. Innes. 1992. Body Size and Consumer-Resource Dynamics. Am. Nat. 139:1151–1175. ## Appendix A: Natural mortality The natural mortality rate is obtained by first assuming that the number of surviving individuals in a cohort $N$ follows a Gompertz relationship [CalderIII:1983jd], where $N=N_{0}{\rm exp}\left(\frac{q_{0}}{q_{a}}\Big{(}1-{\rm exp}({-q_{a}t})\Big{)}\right),$ (A12) given that $q_{0}$ is the initial cohort mortality rate, and $q_{a}$ is the annual rate of increase in mortality, or the actuarial mortality rate. The change in the cohort’s population over time then follows $\frac{\rm d}{\rm dt}N=-dN,$ (A13) such that $d=-\frac{1}{N}\frac{\rm d}{\rm dt}N.$ (A14) If $t_{\rm\ell}$ is the expected lifetime of the organism, then the average rate of mortality over a lifetime $t_{\ell}$ is $\displaystyle\mu$ $\displaystyle=\frac{1}{t_{\rm\ell}}\int_{0}^{t_{\rm\ell}}q_{0}{\rm exp}(q_{a}t_{\ell})$ $\displaystyle=\frac{q_{0}}{q_{a}t_{\ell}}\Big{(}{\rm exp}(q_{a}t_{\ell})-1\Big{)}.$ (A15) The cohort mortality rate $q_{0}$, the actuarial mortality rate $q_{a}$ and the expected lifetime $t_{\ell}$ of a mammal with mass $M_{C}$ all follow allometric relationships, where $q_{0}=1.88\times 10^{-8}M_{C}^{-0.56}$ (1/s) and $q_{a}=1.45\times 10^{-7}M_{C}^{-0.27}$ (1/s) where $M_{C}$ is in grams. Together, we obtain the allometric relationship $\mu(M_{C})=\frac{3.21\times 10^{-8}\left({\rm exp}({0.586M_{C}^{0.03}})-1\right)}{M_{C}^{0.59}}.$ (A16) ## Appendix B: Variations in model parameters and allometric rates While our framework dictates that plant growth rates and carrying capacities are directly proportional to consumer steady states, we can gain insight into what drives the very large range of observed consumer densities by exploring the observed ranges of $\alpha$ and $k$ in terrestrial systems. We assume an intrinsic growth rate roughly that of grass where $\alpha=9.45\times 10^{-9}$ (s-1), whereas observations among terrestrial plants reveal a range in growth rates from $2.81\times 10^{-10}$ to $2.19\times 10^{-8}$ [michaletz2014convergence], according with a change in $\alpha$ of roughly 97% lower and 130% higher than the set value. By incorporating this range into the estimated resource growth rate, we observe that we can account for a large portion of consumer steady state densities around the mean density (inner shaded region, Fig. 1, main text). If we additionally adjust the carrying capacity $k$ of the resource to 90% less-than and 150% more-than the assumed value of $23\times 10^{3}$ g/m2, our framework accounts for nearly the full range of mammalian steady state densities (outer shaded region, Fig. 1, main text). In this context, the upper-boundary of $k$ observed to capture most higher herbivore densities is ca. 34 kg/m2, which is on the higher end of estimated live above-ground biomass densities in terrestrial forests such as in Isle Royal and the Allegheny National Forest [de2017simulating]. Our model’s ability to capture the bounds of mammalian densities at low and high productivity invites some speculation into the actual steepness of the mass-density relationship. While the best-fit slope to Damuth’s Law is -0.77 we also observe that the steeper relationship given by our framework better captures the boundaries of mass-density data, whereas varying the intercept of the statistical best-fit would not capture the lower-density outer-boundary of larger species. While within-clade mass-density relationships often reveal a shallower slope than if measured across clades [Pedersen:2017he], it is possible that the absence of data for larger mammals may bias estimates of the slope towards smaller (shallower) values. Mammalian communities have undergone significant anthropogenic restructuring throughout the Holocene such that many larger species are excluded from the mass-density relationship by way of extinction [Koch:2006vt], and the greater prevalence of smaller species may introduce size-dependent biases. For example, if species $<100$ g are excluded, the empirical mass-density slope steepens from $-0.77$ to $-0.85$. Considering how variations to the underlying energetic parameters driving consumer-resource dynamics alters the expected mass-density relationship may shed light on key constraints shaping mammalian communities. We next explore how variations in the vital rates included in the consumer-resource model modify the expected intercept and slope of the mammalian mass-density relationship. Different vital rates impact the mass-density relationship in three distinct ways, by either _i_) influencing only the mass-density slope, _ii_) influencing only the mass-density intercept, or _iii_) influencing both. Aside from the resource growth rate and carrying capacity, our framework also includes the intrinsic consumer reproductive rate $\lambda^{\rm max}_{C}$, the consumer yield coefficient $Y_{C}$, and the maximum rate of starvation $\sigma^{\rm max}$. We introduce changes to these rates as, for example, $\lambda^{\rm max\prime}_{C}=\lambda^{\rm max}_{C}(1+\chi)$, where $\chi\in(-1,2)$ represents the proportion increase or decrease of the altered parameter denoted by ′. We note that the recovery rate $\rho$ is sufficiently small that alterations do not have an influence on either the consumer mass- density intercept or slope (fig. B1). Figure B1: The predicted mass-density relationship with $\rho$ set to its allometric quantity versus $\rho=0$. The two relationships cannot be distinguished, allowing us to extract insight from the consumer steady state (eq. 7, main text) without its inclusion. Ignoring the effects of $\rho$, we can more easily intuit analytical expressions of the steady state conditions for both the consumer $C$ and resoure $R$, where $\displaystyle C^{*}$ $\displaystyle=\alpha kY_{C}\frac{\sigma^{\rm max}-2\lambda_{C}^{\rm max}+\sqrt{8{\sigma^{\rm max}}^{2}+(\sigma^{\rm max}-2\lambda_{C}^{\rm max})^{2}}}{4{\sigma^{\rm max}}^{2}}$ $\displaystyle R^{*}$ $\displaystyle=k\frac{\sigma^{\rm max}-2\lambda_{C}^{\rm max}+\sqrt{8{\sigma^{\rm max}}^{2}+(\sigma^{\rm max}-2\lambda_{C}^{\rm max})^{2}}}{4\sigma^{\rm max}}.$ (A17) We thus observe that the consumer steady state can also be expressed as $C^{*}=\frac{\alpha}{\sigma^{\rm max}}Y_{C}R^{*}.$ (A18) We discuss how the specific mass-scalings of the relationships impacting the steady states provide more intuition into Damuth’s mass-density relationship in the main text. We can gain additional insight into the role of each vital rate by exploring their quantitative effects on the mass-density relationship directly. Changes to the starvation rate have a large effect on both the consumer-density intercept and slope (Figs. B2,B3). We observe that decreasing $\sigma^{\rm max}$ from the expected value ($\chi<0$) serves to increase the steady state intercept, while decreasing the mass-density slope. By comparison, increasing $\sigma$ from the expected value ($\chi>0$) has less effect on the mass- density relationship. In the consumer-resource model described in Eq. 2.3 (main text), starvation is the primary source of consumer mortality, and therefore plays an out-sized role in determining consumer steady states. As this mortality is reduced, consumer densities increase, raising the intercept. However, as consumer starvation rates decline we observe a steeper mass- density slope. Reduced starvation rates therefore principally benefit the steady state densities of smaller species, with reduced effects observed for larger-bodied mammals. Because fat biomass scales super-linearly with body mass (see Table 1, main text), the populations of larger consumers are more resilient to the effects of starvation, whereas those of smaller consumers are more prone. Figure B2: The effects of changes to metabolic parameters on the prediction of the mass-density relationship. Figure B3: The effects of changes to metabolic parameters on the prediction of the mass-density relationship. The consumer’s maximal rate of reproduction $\lambda^{\rm max}_{C}$ influences only the mass-density slope except for the case $\chi\rightarrow-1$, where growth becomes zero. Above this trivial limit, we observe the consumer growth rate to have a negative effect on the mass-density slope, such that as the growth rate increases, the mass-density relationship becomes steeper (Figs. B2,B3). As the intercept does not change, this means that the steady states of larger bodied consumers decline with increasing $\lambda_{C}^{\rm max}$, while those of smaller-bodied consumers remain unaltered, though the effect is slight. Of more interest is the effect of the yield coefficient $Y_{C}$ and starvation rate $\sigma^{\rm max}$ (Figs. B2,B3). The yield coefficient represents the conversion of resources to consumer biomass, where an increase in $\chi$ correlates to large increases in consumer steady state without altering the mass-density slope. Here we observe that increased efficiency in converting resource to consumer biomass will have an effect similar to increasing resource productivity, as the effective abundance of the resource is greater when relatively fewer resources fuel a given unit of consumer biomass. Because $Y_{C}\propto E_{d}$, where $E_{d}$ is the energy density of the resource (see methods), resource quality is therefore expected to translate directly to higher consumer steady state densities. ## Appendix C: Mortality from predation Per-capita mortality rate from predation The per-capita mortality rate from predation of the herbivore consumer with mass $M_{C}$ and population density $C$ by a mammalian predator with body mass $M_{P}$ and population density $P$ is given by $\beta(C,P)=w\frac{\lambda_{P}(C)P}{CY_{P}},$ (A19) where $\lambda_{P}(C)$ is the growth rate of the predator, $Y_{P}$ is the predator yield coefficient, describing the grams of predator produced per gram of prey consumed, and $w$ is the degree of predation intensity ($w=1$ denotes high predation intensity, whereas $w<1$ denotes lower predation intensity). Assuming a linear functional response for predation mortality, $\lambda_{P}(C)$ is maximized to $\lambda_{P}^{\rm max}$ when the consumer reaches its theoretical maximum population density, which we calculate by converting the resource carrying capacity directly to grams of consumer produced, or $C^{\rm max}=Y_{C}k$. The growth rate of the predator is then given by $\lambda_{P}(C)=\lambda_{P}^{\rm max}\frac{C}{C^{\rm max}}=\lambda_{P}^{\rm max}\frac{C}{Y_{C}k}.$ (A20) Together, we observe the per-capita mortality rate to be (as expected) independent of the consumer density $C$, and is simplified to $\beta(P)=w\frac{\lambda_{P}^{\rm max}P}{Y_{P}Y_{C}k},$ (A21) where we assume that the predator population remains at empirically measured steady state densities for mammalian carnivores, where $P\equiv P^{*}=P_{0}M_{P}^{-0.88}$ [carbone2002common]. This assumption is required because the effects of predation are implicit rather than explicit, and effectively assumes that predator populations operating far below this relationship are not viable. While there is bound to be a range of viable densities for a predator of a given body size, that mass-density relationships exist at all indicates that population densities are highly constrained over evolutionary time and therefore represent a predator energetic demand as a function of body size. Accordingly, if the predator mass-density relationship $P^{*}$ represents an expected energetic requirement for a functioning predator population, our assumption of predation as a constant, rather than dynamic, influence on herbivore mortality reveals the dynamic consequence of such energetic relationships. We suggest that it is these energetic mismatches that may constrain longer-timescale macroevolutionary forces, even if the shorter-timescale ecological dynamics may be more idiosyncratic and complex than our minimal model captures. Herbivore and predator yields As described in the main text, consumer yield is calculated $Y_{C}=\frac{M_{C}E_{d}}{\int_{0}^{t_{\lambda_{C}}}B_{0}m(t)^{\eta}{\rm dt}},$ (A22) where $E_{d}$ is the energy density of the plant resource $R$ (Joules/g) and the denominator is the lifetime energy use required by the herbivore consumer to reach maturity (Joules). The parameters $t_{\lambda_{C}}$ and $B_{0}$ are the timescale associated with reaching reproductive maturity and the metabolic coefficient for herbivorous mammals, respectively, and $\eta=-3/4$ is the metabolic exponent (see Table 1, main text). The predator yield is calculated similarly, where $Y_{P}=\frac{M_{C}E_{C}}{\int_{0}^{t_{\lambda_{P}}}{B_{0}}_{P}m(t)^{\eta}{\rm dt}},$ (A23) where $E_{C}$ is the energy density of the herbivore being consumed, and the denominator is the lifetime energy use required by the predator to reach maturity. The parameters $t_{\lambda_{P}}$ and ${B_{0}}_{P}$ are the timescale associated with reaching reproductive maturity and the metabolic coefficient for predatory mammals, respectively, and $\eta=-3/4$ is the metabolic exponent. We note that the metabolic coefficient for predators is different than that for mammals [MunozGarcia2005]. The energy density of herbivore consumers changes with body mass $M_{C}$. For example, small mammals have very low percent body fat, whereas very large mammals have high percent body fat. We assume that predators consume all non- skeletal mass of prey. Because the amount of consumable tissues with different energy densities within an herbivore varies allometrically, so too should the energy density $E_{C}$. We consider four primary tissue groups: a consumable set composed of muscle, fat, and _other_ tissues, and an non-consumable set composed only of skeletal tissues. If the scalings associated with fat, muscle, and skeletal tissues are $M_{C}^{\rm fat}=f_{0}M_{C}^{1.19}$, $M_{C}^{\rm musc}=g_{0}M_{C}^{1.00}$, and $M_{C}^{\rm skel}=h_{0}M_{C}^{1.09}$ [with normalization constants $f_{0}=0.02$, $g_{0}=0.38$, and $h_{0}=0.0335$; prange1979scaling], the scaling of the _other_ tissue (gut tissue, organ tissue, etc) is given by $M_{C}^{\rm other}=M_{C}-(M_{C}^{\rm fat}+M_{C}^{\rm musc}+M_{C}^{\rm skel})$. The energy density of fat is $E_{\rm fat}=37700$ J/g, whereas the energy density of muscle is $E_{\rm musc}=17900$ J/g [merrill1973part]. If we assume that gut and organ tissues have roughly the same energy density as muscle, the attainable energy density for an herbivore of size $M_{C}$ is given by $E_{C}(M_{C})=E_{\rm fat}\frac{M_{C}^{\rm fat}}{M_{C}}+E_{\rm musc}\left(\frac{M_{C}^{\rm musc}}{M_{C}}+\frac{M_{C}^{\rm other}}{M_{C}}\right).$ (A24) Large-bodied Predator-Prey Mass Relationship (PPMR) The predator growth rate $\lambda_{P}^{\rm max}$, the time required for the predator to reach reproductive maturity $t_{\lambda_{P}}$, and the predator’s steady state population density $P^{*}$ are allometric relationships that depend on predator body mass $M_{P}$. Accordingly, for an herbivore of a given mass $M_{C}$, we must anticipate the size of its likely predator $M_{P}$. This is very different than the predator-centric perspective of anticipating the average prey size for a given predator [Carbone:1999ju]. For example, the most preferred prey mass for an African lion is ca. $350$ kg [hayward2005lion], where the inclusion of megaherbivores to diet is comparatively low. However from a megaherbivore’s perspective, lions may represent the only potential predator. In other words, because the range of prey body mass increases for predators of larger body mass [Sinclair2003], it is the upper limit of the range that impacts the populations of larger herbivores. To obtain an herbivore-centric measure of the expected predator mass given a particular herbivore mass ${\rm E}\\{M_{P}|M_{C}\\}$, we first compiled the known diets of large-bodied predators, including tigers, lions, hyenas, leopards, dhole, wild dogs, and cheetahs [hayward2005lion, Hayward2006hyena, hayward2006leopard, hayward2006lycaon, hayward2006cheetah, Hayward2008]. Because smaller mammalian predators and prey have very different PPMRs than larger-bodied mammalian predators and prey, we here focus exclusively on the predators of large-bodied herbivore prey $>10^{5}$ g. From the mean proportional reliance of predators on large-bodied prey [hayward2005lion, Hayward2006hyena, hayward2006leopard, hayward2006lycaon, hayward2006cheetah, Hayward2008], we repeatedly sampled predator dietary distributions to reflect each predator’s reliance as a function of prey mass. We introduced variability in predator and prey masses by assuming that body sizes were normally distributed about the expected value with a standard deviation of $\pm 25\%$, allowing us to obtain a distribution of expected predator diets as a function of prey mass. From this relationship, we then evaluated the expected predator mass for a given prey mass range to obtain ${\rm E}\\{M_{P}\\}$ (Fig. 4b, main text), demonstrating the allometric relationship of ${\rm E}\\{M_{P}\\}=9.76\times 10^{3}M_{C}^{0.21}$, where we used the output of 100 independent replicates to robustly estimate the best fit. We emphasize that this relationship only pertains to large-bodied predators and prey $>10^{5}$ g. Alterations to and variations from this relationship are explored in the main text. Including the empirically-measured PPMR (or a variant of the empirically measured PPMR – see below) results in the appearance of a transcritical bifurcation at consumer mass $M_{C}^{\dagger}$. We observe that this critical mass threshold results in the extinction of the consumer population characterized by body sizes $M_{C}\geq M_{C}^{\dagger}$ (fig. C1A). At this body mass, the Determinant of the Jacobian matrix characterizing the system presented in eq. 3 (main text) with predation mortality included (eq. 8, main text) is zero (fig. C1B), aligning with the real component of a single eigenvalue crossing zero and becoming positive (fig. C1C). While we do not derive a normal form for this bifurcation, these features strongly suggest the observed bifurcation is transcritical in nature [Kuznetsov1995]. Figure C1: A. Consumer population density as a function of mass $M_{C}$. B. The determinant of the Jacobian matrix for the system presented in eq. 3 (main text) with predation mortality included (eq. XX, main text) as a function of mass $M_{C}$. C. The Real component of the two eigenvalues of the Jacobian assessed in panel B. as a function of mass $M_{C}$. The dashed line in each panel is the measured consumer mass threshold $M_{C}^{\dagger}=2.5*10^{6}$ g. The Gray horizontal line in panels B,C denotes zero on the y-axis. As explored in the main text, the empirically-measured PPMR for large-bodied mammals results in a threshold body size for herbivore consumers $M_{C}^{\dagger}$. This size marks the point where the predator population, with a body mass derived from the PPMR, cannot sustain its own growth from the predated herbivore population, thereby driving the herbivore population to extinction. The size at which $M_{C}^{\dagger}$ occurs is both dependent on the nature of the PPMR, as well as predation intensity $w$. As predation intensity $w$ decreases, $M_{C}^{\dagger}$ increases (Fig. C2). Figure C2: The effect of changing the predation intensity $w$ on the single herbivore consumer population. If $w=1$, predation intensity is maximized, is lower if $0<w<1$, and is above the maximal level required to support a predator population if $w>1$. The blue region denotes herbivore threshold mass range characterizing $w=1\pm 0.1$. The yellow line denotes the mass range of contemporary elephantids. Vertical dashed line denotes the size of the largest terrestrial mammal (_Deinotherium_ at ca. $1.74\times 10^{7}$, corresponding to $w=0.37$, such that predation intensity is moderate. By allowing the PPMR to vary as ${\rm E}\\{M_{P}\\}=v_{0}(1+\chi_{\rm int})M_{C}^{v_{1}(1+\chi_{\rm slope})},$ (A25) where the proportional changes in the PPMR intercept and slope are given by $\chi_{\rm int}$ and $\chi_{\rm slope}\in(-0.99,2)$, so does the threshold herbivore body mass $M_{C}^{\dagger}$ and, by extension, the related threshold predator body mass $M_{P}^{\dagger}$. From Fig. 4D (main text), we observe that changing the intercept and slope of the PPMR has a large influence on $M_{C}^{\dagger}$ and $M_{P}^{\dagger}$. Across this range of potential PPMRs, we highlight those values for the intercept and slope of the PPMR that permit megatrophic interactions, where both megapredators subsist on megaherbivores at the threshold body mass (highlighted region in Fig. 4D, main text). Fig. C3 shows the relationship between megapredator and megaherbivore body masses highlighted within this region. Allowing both the PPMR to vary and assuming lower predation intensity ($w=0.37$) enables much larger body sizes for megaherbivores and their associated megapredators (Fig. C4). Figure C3: Mass ranges corresponding to feasible megatrophic interactions (where herbivore and predator threshold masses are $>5\times 10^{5}$ g) across variations to the assumed PPMR, demarcated by the white bands in Fig. 4C,D (main text), and assuming high predation intensity ($w=1$). Figure C4: The effects of lower predation intensity ($w=0.37$) on A. threshold herbivore mass $M_{C}^{\dagger}$ and B. threshold predator mass $M_{P}^{\dagger}$ across variable PPMRs, where ${\rm E}\\{M_{P}\\}=v_{0}(1+\chi_{\rm int})M_{C}^{v_{1}(1+\chi_{\rm slope})}$ and both $\chi_{\rm int}$ and $\chi_{\rm slope}\in(-0.99,2)$, and $v_{0}=9.76\times 10^{3}$ and $v_{1}=0.21$ are set as in the main text. White bands denote regions of $\chi_{\rm int}$ and $\chi_{\rm slope}$ where megatrophic interactions are feasible (i.e. both predator and herbivore threshold masses are $>5\times 10^{5}$ g). C. Mass ranges corresponding to feasible megatrophic interactions in the white bands in A. and B. Figure C5: The effect of an altered carrying capacity $k$ on the consumer mass threshold $M_{C}^{\dagger}$. Carrying capacity values that vary from -20% to +20% give rise to an $M_{C}^{\dagger}$ roughly 65% and 140% the original estimated mass of $2.55\times 10^{6}$ g. Equivalent variations in $\alpha$ will result in the same changes to $M_{C}^{\dagger}$ given its role in eq. 7 (main text). The shaded region highlights the mass range captured by the altered carrying capacity. Finally, we note that changes to both the resource growth rate $\alpha$ as well as to the resource carrying capacity $k$ can impact the consumer size at which populations become infeasible $M_{C}^{\dagger}$. Because the consumer steady state is directly proportional to both of these parameters (see eq. 7, main text), a lower growth rate and/or carrying capacity lowers the intercept of the steady state mass-density relationship $C^{*}(M_{C})$. Analysis of the effect of changes to $k$ (and this will be the same for $\alpha$) reveals that while it has influence on $M_{C}^{\dagger}$, it is not incredibly large (Fig. C5. However this relationship carries with it an important message: in environments with lower carrying capacities and/or plant growth rates, we would expect a lower mass threshold bounding feasible megaherbivore populations. ## Appendix D: Derivation of harvesting mortality We first determined the harvest rate $h=h^{\dagger}$ required to drive an herbivore population to extinction, thereby satisfying the condition $C^{*}(M_{C}|h)=0$ as a function of herbivore body mass $M_{C}$. This extinction-inducing harvest rate, itself now a function of consumer body mass $h^{\dagger}(M_{C})$, defines the rate at which the population must be harvested to drive the steady state to zero. To compare this rate against measures of harvest both in nature and predicted from other mathematical or computational treatments of harvest-induced extinction, we calculated the harvest pressure $\psi^{\dagger}$, which we defined as the number of herbivore individuals per area harvested at this rate to reduce the population to some proportion $\epsilon$ of its steady state. This harvest pressure is thus defined by some number of individuals harvested per year over a certain number of years to reduce the population from $C^{*}$ to its post-harvest density $\epsilon C^{*}$. To calculate harvest pressure, we first assume that at the steady state, harvest is occurring on a shorter-than-generational timescale. For megaherbivores such as elephants, a generation is roughly 25 years [wittemyer2013comparative], and for harvest pressures that must be applied beyond this period of time, we would expect population growth to counter the negative effects of harvest. Assuming harvest-only change, we simplify the dynamics to $\frac{\rm d}{\rm dt}C=-h^{\dagger}(M_{C})C,$ (A26) where the time to reduce $C^{*}$ to $\epsilon C^{*}$ is $\displaystyle C(t)$ $\displaystyle=C_{0}{\rm e}^{-h^{\dagger}(M_{C})t},$ $\displaystyle\epsilon C^{*}$ $\displaystyle=C^{*}{\rm e}^{-h^{\dagger}(M_{C})t}$ $\displaystyle t_{\epsilon}$ $\displaystyle=-\frac{\log(\epsilon)}{h^{\dagger}(M_{C})}.$ (A27) We note that for elephant-sized herbivores and larger, $t_{\epsilon}\leq 23$ years. While the time required to harvest the population to $\epsilon C^{*}$ is only just approaching generational timescales, it should be treated as a minimum $t_{\epsilon}$ given the effects of population growth will prolong the imposed harvest effort. Harvest pressure is then calculated as $\psi^{\dagger}=\frac{C^{*}(1-\epsilon)}{M_{C}t_{\epsilon}}c_{0}=-h^{\dagger}(M_{C})\frac{C^{*}(1-\epsilon)}{M_{C}\log(\epsilon)}c_{0}$ (A28) where the constant $c_{0}$ denotes the conversion from inds/m2/second to inds/$A_{\rm CA}$/year, where $A_{\rm CA}=4.24\times 10^{11}\leavevmode\nobreak\ {\rm m}^{2}$ is the arbitrarily-chosen area of California. This conversion is particularly important for evaluating other harvest measures from the historical record and estimates from independent models and simulations for extinct species. As described in the main text, the extinction-inducing harvest pressure is calculated to be $4.3\times 10^{3}$ inds/yr/$A_{\rm CA}$ for an elephant-sized organism of $M_{C}=2.5\times 10^{6}$ g (see Fig 5, main text). Harvest pressure on Pleistocene mammoths We compare our measure of harvest pressure to that calculated for mammoths (_Mammuthus primigenius_) by fordham2022process. Because fordham2022process employ a much more specific and detailed assessment of the effects of harvest specifically for mammoths over a spatially explicit landscape, we must make a few simplifications in order to derive a comparable estimate. First, the harvest interaction between mammoth populations and humans is modeled as a Type II functional response, where, again isolating population-level effects to that of harvest we obtain $\frac{\rm d}{\rm dt}C=-\frac{sNFC}{G+\frac{C}{C_{\rm max}M_{C}}},$ (A29) where $N$ is the normalized human population density maximized at unity, the constant $s=7.884\times 10^{-8}$ generations/second (where a generation is 25 years), $F$ represents the effectiveness of human hunting, ranging from $(0.01,0.34)$, $C_{\rm max}=1.875\times 10^{-6}$ g/${\rm m^{2}}$ is the maximum mammoth population density (converted from the average degree-by- degree grid cells in Siberia), $G=0.4$ is the half-saturation constant, and $M_{C}=2.5\times 10^{6}$ grams. Solving for the time required to reduce the population to $\epsilon C^{*}$, we obtain $t_{\epsilon}^{\rm mammoth}=\frac{C^{*}-C_{\rm max}GM_{C}\log\left[C^{*}\exp(\frac{C^{*}\epsilon}{C_{\rm max}GM_{C}})\epsilon\right]}{sC_{\rm max}FM_{C}N}.$ (A30) We then calculate the harvest pressure as $\psi^{\rm mammoth}=c_{0}\frac{sC_{\rm max}FN(C^{*}-C^{*}\epsilon)}{C^{*}-C_{\rm max}GM_{C}\log\left[C^{*}\exp(\frac{C^{*}\epsilon}{C_{\rm max}GM_{C}})\epsilon\right]},$ (A31) where the constant $c_{0}$ again denotes the conversion from inds/m2/second to inds/$A_{\rm CA}$/year, where $A_{\rm CA}$ is the arbitrarily-chosen area of California. Given a range in $F\in(0.01,0.35)$ and $N\in(0.01,1)$, we obtain a distribution of values for mammoth harvest pressure with a median value of $1.24\times 10^{4}$ inds/yr/$A_{\rm CA}$ over the course of 9.8 years. The bounds of the estimated range from $5\times 10^{4}$ inds/yr/$A_{\rm CA}$ over the course of 2 years to $5\times 10^{2}$ inds/yr/$A_{\rm CA}$ over the course of ca. 200 years (the range is plotted as the vertical black line in Fig. 5, main text). Harvest pressure on Pleistocene _Diprotodon_ The harvest rate needed to collapse _Diprotodon_ populations was calculated by Bradshaw et al. [bradshaw2021relative], where a harvest pressure of between 400-500 inds/year/area of Australia was sufficient to collapse the population. Translating this to the area of California, we obtain between 678 to 848 inds/yr/$A_{\rm CA}$, with a mean of $763.2$ inds/yr/$A_{\rm CA}$. Harvest pressure on historical elephants _Loxodonta africana_ Elephant harvest rates are estimated from historical documentation of the ivory trade detailed in milner1993exploitation. While the trade volume oscillates with changes in technology, access to habitats within Africa, and the feedbacks of trade on elephant population size, we compare our results against estimates taken at two points in time: early in the ivory trade (1810), and late in the ivory trade (1987). From milner1993exploitation we assume that each elephant killed contributes 1.88 tusks, and that tusk mass begins at 15 kg per tusk early in trade to 5 kg per tusk in later years. While the area from which elephants were harvested is largely unknown, we assume the area harvested is that assessed to be suitable elephant habitat in sub-Saharan Africa, estimated at $3.22\times 10^{12}\leavevmode\nobreak\ {\rm m}^{2}$ [thouless2016african]. From rates of ca. $1\times 10^{5}$ kg/yr of ivory harvested in 1810 to ca. $9.7\times 10^{5}$ kg/yr of ivory harvested in 1987, normalized to habitat area and converted to the area of California, we obtain estimates of ca. $467$ inds/yr/$A_{\rm CA}$ in 1810 to ca. $1.33\times 10^{4}$ inds/yr/$A_{\rm CA}$ in 1987 (see Fig. 5, main text). ## References * Bradshaw et al. [2021] Bradshaw, C. J., C. N. Johnson, J. Llewelyn, V. Weisbecker, G. Strona, and F. Saltré. 2021. Relative demographic susceptibility does not explain the extinction chronology of sahul’s megafauna. Elife 10:e63870. * Calder III [1983] Calder III, W. A. 1983. An allometric approach to population cycles of mammals. J. Theor. Biol. 100:275–282. * Carbone and Gittleman [2002] Carbone, C., and J. L. Gittleman. 2002. A common rule for the scaling of carnivore density. Science 295:2273–2276. * Carbone et al. [1999] Carbone, C., G. M. Mace, S. C. Roberts, and D. W. Macdonald. 1999. Energetic constraints on the diet of terrestrial carnivores. Nature 402:286–288. * De Jager et al. [2017] De Jager, N. R., P. J. Drohan, B. M. Miranda, B. R. Sturtevant, S. L. Stout, A. A. Royo, E. J. Gustafson, and M. C. Romanski. 2017. Simulating ungulate herbivory across forest landscapes: A browsing extension for landis-ii. Ecological Modelling 350:11–29. * Fordham et al. [2022] Fordham, D. A., S. C. Brown, H. R. Akçakaya, B. W. Brook, S. Haythorne, A. Manica, K. T. Shoemaker, J. J. Austin, B. Blonder, J. Pilowsky, et al. 2022\. Process-explicit models reveal pathway to extinction for woolly mammoth using pattern-oriented validation. Ecology letters 25:125–137. * Hayward [2006] Hayward, M. 2006. Prey preferences of the spotted hyaena (Crocuta crocuta) and degree of dietary overlap with the lion (Panthera leo). J. Zoology 270:606–614. * Hayward et al. [2006 _a_] Hayward, M., P. Henschel, J. O’Brien, M. Hofmeyr, G. Balme, and G. I. Kerley. 2006 _a_. Prey preferences of the leopard (panthera pardus). Journal of Zoology 270:298–313. * Hayward et al. [2006 _b_] Hayward, M., M. Hofmeyr, J. O’brien, and G. I. Kerley. 2006 _b_. Prey preferences of the cheetah (acinonyx jubatus)(felidae: Carnivora): morphological limitations or the need to capture rapidly consumable prey before kleptoparasites arrive? Journal of Zoology 270:615–627. * Hayward and Kerley [2008] Hayward, M. W., and G. Kerley. 2008. Prey preferences and dietary overlap amongst Africa’s large predators. S. African J. Wild. Res. 38:93–108. * Hayward and Kerley [2005] Hayward, M. W., and G. I. Kerley. 2005. Prey preferences of the lion (panthera leo). Journal of zoology 267:309–322. * Hayward et al. [2006 _c_] Hayward, M. W., J. O’Brien, M. Hofmeyr, and G. I. Kerley. 2006 _c_. Prey preferences of the african wild dog lycaon pictus (canidae: Carnivora): ecological requirements for conservation. Journal of Mammalogy 87:1122–1131. * Koch and Barnosky [2006] Koch, P. L., and A. Barnosky. 2006. Late Quaternary extinctions: state of the debate. Annu. Rev. Ecol. Evol. Syst. 37:215–250. * Kuznetsov [1998] Kuznetsov, Y. 1998. Elements of Applied Bifurcation Theory. Springer, New York. * Merrill and Watt [1973] Merrill, A., and B. Watt. 1973. Part 2: digestibility and available energy of foods. Energy value of foods: basis and derivation. Agriculture handbook No 74:8–24. * Michaletz et al. [2014] Michaletz, S. T., D. Cheng, A. J. Kerkhoff, and B. J. Enquist. 2014. Convergence of terrestrial plant production across global climate gradients. Nature 512:39–43. * Milner-Gulland and Beddington [1993] Milner-Gulland, E., and J. Beddington. 1993. The exploitation of elephants for the ivory trade: an historical perspective. Proceedings of the Royal Society of London. Series B: Biological Sciences 252:29–37. * Muñoz-Garcia, Agustí and Williams, Joseph B. [2005] Muñoz-Garcia, Agustí and Williams, Joseph B. 2005. Basal metabolic rate in carnivores is associated with diet after controlling for phylogeny. Physiological and Biochemical Zoology 78:1039–1056. PMID: 16228943. * Pedersen et al. [2017] Pedersen, R. Ø., S. Faurby, and J.-C. Svenning. 2017. Shallow size–density relations within mammal clades suggest greater intra-guild ecological impact of large-bodied species. J. Anim. Ecol. 86:1205–1213. * Prange et al. [1979] Prange, H. D., J. F. Anderson, and H. Rahn. 1979. Scaling of skeletal mass to body mass in birds and mammals. The American Naturalist 113:103–122. * Sinclair et al. [2003] Sinclair, A. R. E., S. Mduma, and J. S. Brashares. 2003. Patterns of predation in a diverse predator–prey system. Nature 425:288–290. * Thouless et al. [2016] Thouless, C., H. T. Dublin, J. Blanc, D. Skinner, T. Daniel, R. Taylor, F. Maisels, H. Frederick, and P. Bouché. 2016. African elephant status report 2016. An update from the African Elephant Database . * Wittemyer et al. [2013] Wittemyer, G., D. Daballen, and I. Douglas-Hamilton. 2013. Comparative demography of an at-risk african elephant population. PloS one 8:e53726.
# Unitary Howe dualities for fermionic and bosonic algebras and related Dirac operators Guner Muarem Campus Middelheim Middelheimlaan 1 M.G.221 2020 Antwerpen België<EMAIL_ADDRESS> ###### Abstract. In this paper we use the canonical complex structure $\mathbb{J}$ on $\mathbb{R}^{2n}$ to introduce a twist of the symplectic Dirac operator. As a matter of fact, these operators can be interpreted as the bosonic analogues of the Dirac operators on a Hermitian manifold. Moreover, we prove that the algebra of these symplectic Dirac operators is isomorphic to the Lie algebra $\mathfrak{su}(1,2)$ which leads to the Howe dual pair $(\mathsf{U}(n),\mathfrak{su}(1,2))$. ## 1\. Introduction The CCR (canonical commuting relation) and CAR algebras (canonical anticommuting relation) are fundamental algebras in theoretical physics used for the study of bosons and fermions. From a mathematical viewpoint, these algebras are named the Weyl algebra (or symplectic Clifford algebra) and Clifford algebra. These algebras can be constructed in a very analogous way. The Clifford algebra is constructed on a vector space $V$ equipped with a symmetric bilinear form $B$, whereas the Weyl algebra requires an even dimensional vector space equipped with a skew-symmetric bilinear form (or symplectic form) $\omega$. In both cases, one then constructs the tensor algebra $T(V)$ where an ideal $I(V)$ is divided out. In the orthogonal setting, this is the ideal $I_{B}(V)$ with elements subject to the relation $\\{u,v\\}=2B(u,v)$. In the symplectic setting, this is the ideal $I_{\omega}(V)$ generated by $[u,v]=-\omega(u,v)$. There is, however, a fundamental difference: the Clifford algebra is finite- dimensional, whereas the Weyl algebra is infinite-dimensional. For the spinors (orthogonal versus symplectic) the same infinite-dimensional principle holds as for the Clifford algebras. As a matter of fact, the symplectic spinors are the smooth vectors in the metaplectic representation [2]. Using the generators of the Clifford (resp. Weyl) algebra, one can associate a natural first order spin (resp. metaplectic) invariant differential operator by contracting the Clifford algebra elements using the bilinear form $B$ (resp. the symplectic form $\omega)$ with derivatives. This gives rise to the Dirac operator $\partial_{x}=\sum_{k=1}^{n}e_{k}\partial_{x_{k}}$ where $\\{e_{j},e_{k}\\}=-2\delta_{ij}$ and the symplectic Dirac operator $\sum_{k=1}^{n}\left(iq_{k}\partial_{y_{k}}-\partial_{q_{k}}\partial_{x_{k}}\right)$ where $[\partial_{q_{j}},iq_{k}]=i\delta_{jk}$ are the Heisenberg relations. The theory which studies the solutions of the Dirac operator is known as Clifford analysis and can be seen as a hypercomplex function theory. Moreover, quite some generalisations have occurred in the last two decades. This involves e.g. Clifford analysis on superspace and Clifford analysis on (hyper)Kähler spaces (see for instance [3]). It is in the latter framework in which this paper is situated, but then from a symplectic point of view. More precisely, we provide the foundations of what we will call a hermitian variant of symplectic Clifford analysis, where we incorporate the additional datum of a compatible complex structure $\mathbb{J}$ on the flat symplectic space $\mathbb{R}^{2n}$. This leads to the study of the symplectic Dirac operator on a Kähler manifold as was already initiated in [2, 4]. However, the underlying invariance symmetry and the algebra generated by these type of operators (and their duals) was never investigated. ## 2\. Rudiments of symplectic Clifford analysis Let us consider the canonical symplectic space $\mathbb{R}^{2n}$ with coordinates $({x},{y})$ and the usual symplectic form $\omega_{0}=\sum_{j=1}^{n}dx_{j}\wedge dy_{j}$ which has the matrix representation $\Omega_{0}=\begin{pmatrix}0&I_{n}\\\ -I_{n}&0\end{pmatrix}.$ Recall that the symplectic group $\mathsf{Sp}(2n,\mathbb{R})$ is the group given by invertible linear transformations preserving the non-degenerate skew- symmetric bilinear form from above and is given (in terms of matrices) by $\mathsf{Sp}(2n,\mathbb{R})=\\{M\in\mathsf{GL}(2n,\mathbb{R})\mid M^{T}\Omega_{0}M=\Omega_{0}\\}.$ The group is non-compact and has dimension $2n^{2}+n$. Moreover, the corresponding Lie algebra is denoted by $\mathfrak{sp}(2n,\mathbb{R})$. The main difference with the orthogonal case, lies in the fact that the metaplectic group (the double cover of the symplectic group) does not admit a finite dimensional representation (it is not a matrix group). This is a strong contrast with the spin representation in the orthogonal case. Moreover, the orthogonal spinors $\mathbb{S}$ are realised as a idempotent left ideal in the Clifford algebra, which is not the case for the symplectic spinors. As mentioned, the symplectic equivalent of the spin representations are infinite dimensional, which means that one needs to work with the theory of unitary representations. ### 2.1. The Schwartz space and metaplectic representation For further convenience, we fix notation and define the Schwartz space, which plays a crucial role in the construction of the metaplectic representation. On the space $\mathcal{C}^{\infty}(\mathbb{R}^{n},\mathbb{C})$ we define (using the multi-index notation) the norm $||f||_{\alpha,\beta}:=\sup_{{q}\in\mathbb{R}^{n}}|{q}^{\alpha}(D^{\beta}f)({q})|$ for all $\alpha,\beta\in\mathbb{N}^{n}$. The Schwartz space $\mathcal{S}(\mathbb{R}^{n})$ is the subspace of $L^{p}(\mathbb{R}^{n})$ (for $1\leq p\leq\infty$) consisting of rapidly decreasing functions and is given by $\mathcal{S}(\mathbb{R}^{n},\mathbb{C}):=\\{f\in\mathcal{C}^{\infty}(\mathbb{R}^{n},\mathbb{C}):||f||_{\alpha,\beta}<\infty\text{ for all }\alpha,\beta\in\mathbb{N}^{n}\\}.$ We now describe (following [2]) the infinite-dimensional Segal-Shale-Weil representation (also oscillator or metaplectic representation) of the metaplectic group. The smooth vectors of the unitary representation $\mathfrak{m}:\mathsf{Mp}(2n)\to\mathsf{U}(L^{2}(\mathbb{R}^{n}))$ coincide with the Schwartz space $\mathcal{S}(\mathbb{R}^{n})$ and are a model for the symplectic spinors $\mathbb{S}^{\infty}$. Due to Stone-Von Neumann theorem the representation is unique (up to unitary equivalence). ### 2.2. The symplectic Clifford algebra and the related Dirac operator Let $(V,\omega)$ be a symplectic vector space. The symplectic Clifford algebra $\mathsf{Cl}_{s}(V,\omega)$ is defined as the quotient algebra of the tensor algebra $T(V)$ of $V$, by the two-sided ideal $\mathcal{I}_{\omega}:=\\{v\otimes u-u\otimes v+\omega(v,u):u,v\in V\\}.$ In other words $\mathsf{Cl}_{s}(V,\omega):=T(V)/\mathcal{I}_{\omega}$ is the algebra generated by $V$ in terms of the relation $[v,u]=-\omega(v,u)$, where we have omitted the tensor product symbols. We refer to the symplectic Clifford algebra on $\mathbb{R}^{2n}$ as the $n$th Weyl algebra $\mathcal{W}_{n}$ with generators $iq_{1},\dots,iq_{n},\partial_{q_{1}},\dots,\partial_{q_{n}}$ satisfying the commutation relations $[q_{j},q_{k}]=0$ and $[\partial_{q_{j}},q_{k}]=\delta_{jk}$. Denote by $\mathcal{F}$ a suitable function space (e.g. the space of polynomials, or smooth funtions). ###### Definition 2.1. The symplectic Dirac operator on $(\mathbb{R}^{2n},\omega_{0})$ is the first- order (in the base variables ${x}$ and ${y}$) differential operator acting on a symplectic spinor-valued functions space $\mathcal{F}\otimes\mathbb{S}^{\infty}$ given by $D_{s}=\sum_{j=1}^{n}(iq_{j}\partial_{y_{j}}-\partial_{q_{j}}\partial_{x_{j}}).$ With respect to the symplectic Fischer inner product (see [5]), we obtain the dual operator $X_{s}=\sum_{j=1}^{n}(iq_{j}{x_{j}}+\partial_{q_{j}}{y_{j}}).$ These operators satisfy the relations: $\displaystyle[\mathbb{E}+n,X_{s}]=X_{s},$ $\displaystyle[\mathbb{E}+n,D_{s}]=-D_{s},$ $\displaystyle[D_{s},X_{s}]=-i(\mathbb{E}+n)$ where $\mathbb{E}=\sum_{j=1}^{n}(x_{j}\partial_{x_{j}}+y_{j}\partial_{y_{j}})$ is the Euler operator. In other words, the three operators give rise to a copy of the Lie algebra $\mathfrak{sl}(2)$. ## 3\. The interaction with a complex structure ### 3.1. Definition of the twisted symplectic Dirac operators We will now introduce a complex structure $\mathbb{J}$ on the symplectic manifold $(\mathbb{R}^{2n},\omega_{0})$ which is compatible with the symplectic form $\omega_{0}$. This means that $\omega_{0}(x,\mathbb{J}y)$ defines a Riemannian metric $g$. Otherwise said, we will be working with the canonical Kähler manifold $(\mathbb{R}^{2n},\omega_{0},g,\mathbb{J})$. By Darboux’s theorem, we obtain, with respect to the canonical symplectic basis $\\{e_{j}\\}_{j=1}^{2n}$ the following complex structure $\mathbb{J}=\left(\begin{smallmatrix}0&-I_{n}\\\ I_{n}&0\end{smallmatrix}\right).$ The action of the complex structure $\mathbb{J}$ on $\mathbb{R}^{2n}$ is given by $\displaystyle(x_{1},\dots,x_{n},y_{1},\dots,y_{n})\mapsto(y_{1},\dots,y_{n},-x_{1},\dots,-x_{n}).$ ###### Definition 3.1. The new differential operators acting on symplectic spinor-valued functions $\displaystyle\tilde{D}_{s}$ $\displaystyle=\sum_{j=1}^{n}iq_{j}\partial_{x_{j}}+\partial_{y_{j}}\partial_{q_{j}}$ $\displaystyle\tilde{X}_{s}$ $\displaystyle=\sum_{j=1}^{n}x_{j}\partial_{q_{j}}-iy_{j}q_{j}$ $\displaystyle\mathbb{E}$ $\displaystyle=\sum_{j=1}^{n}x_{j}\partial_{x_{j}}+y_{j}\partial_{y_{j}}$ also give rise to a copy of the Lie algebra $\mathfrak{sl}(2)$. We call these first two operators the twists of $D_{s}$ and $X_{s}$. Both sets of operators, i.e. $(D_{s},X_{s})$ and $(\tilde{D}_{s},\tilde{X}_{s})$, are symplectic invariant, albeit under the following two different realisations of the symplectic Lie algebra given by: $\displaystyle\begin{cases}X_{jk}=x_{j}\partial_{x_{k}}-y_{k}\partial_{y_{j}}-(q_{k}\partial_{q_{j}}+\frac{1}{2}\delta_{jk})&1\leq j\leq k\leq n\\\ Y_{jk}=x_{j}\partial_{y_{k}}+x_{k}\partial_{y_{j}}+i\partial_{q_{j}}\partial_{q_{k}}&j<k=1,\dots,n\\\ Z_{jk}=y_{j}\partial_{x_{k}}+y_{k}\partial_{x_{j}}+iq_{j}q_{k}&j<k=1,\dots,n\\\ Y_{jj}=x_{j}\partial_{y_{j}}+\frac{i}{2}\partial_{q_{j}}^{2}&j=1,\dots,n\\\ Z_{jj}=y_{j}\partial_{x_{j}}+\frac{i}{2}q_{j}^{2}&j=1,\dots,n\end{cases}$ and $\displaystyle\begin{cases}\tilde{X}_{jk}=x_{j}\partial_{x_{k}}-y_{k}\partial_{y_{j}}+q_{k}\partial_{q_{j}}+\frac{1}{2}\delta_{jk}&1\leq j\leq k\leq n\\\ \tilde{Y}_{jk}=x_{j}\partial_{y_{k}}+x_{k}\partial_{y_{j}}-iq_{j}q_{k}&j<k=1,\dots,n\\\ \tilde{Z}_{jk}=y_{j}\partial_{x_{k}}+y_{k}\partial_{x_{j}}-i\partial_{q_{j}}\partial_{q_{k}}&j<k=1,\dots,n\\\ \tilde{Y}_{jj}=x_{j}\partial_{y_{j}}-\frac{i}{2}q_{j}^{2}&j=1,\dots,n\\\ \tilde{Z}_{jj}=y_{j}\partial_{x_{j}}-\frac{i}{2}\partial_{q_{j}}^{2}&j=1,\dots,n\end{cases}$ Of course, it is not very useful that $D_{s}$ and $\widetilde{D}_{s}$ are invariant under different (yet isomorphic) $\mathfrak{sp}(2n,\mathbb{R})$-realisations. Therefore, we will perform a symmetry reduction so that both operators become invariant under one and the same Lie algebra. To that end, we need to find the symplectic matrices which commute with the complex structure. ###### Lemma 3.2. We have that $\mathsf{Sp}_{\mathbb{J}}(2n,\mathbb{R}):=\\{M\in\mathsf{Sp}(2n,\mathbb{R})\mid M\mathbb{J}=\mathbb{J}M\\}$ defines a realisation for the unitary Lie group. ###### Proof. In order to see this, assume that $M$ is of the block-form: $\left(\begin{smallmatrix}A&B\\\ C&D\end{smallmatrix}\right),$ where $A,B,C$ and $D$ are $(n\times n)-$matrices. The condition that $M$ is symplectic is equivalent to the one of the following conditions: the matrices $A^{T}C$ and $B^{T}D$ are symmetric and $A^{T}D-C^{T}B=I$. So, in order to determine $\operatorname{Sp}_{\mathbb{J}}(2n,\mathbb{R})$ we need to determine the symplectic matrices $M$ which commute with the complex structure $\mathbb{J}$. The latter conditions means that $\displaystyle M\mathbb{J}=\mathbb{J}M$ $\displaystyle\iff\mathbb{J}^{-1}M\mathbb{J}=M$ $\displaystyle\iff\begin{pmatrix}0&I\\\ -I&0\end{pmatrix}\begin{pmatrix}A&B\\\ C&D\end{pmatrix}\begin{pmatrix}0&-I\\\ I&0\end{pmatrix}=\begin{pmatrix}A&B\\\ C&D\end{pmatrix}$ $\displaystyle\iff\begin{pmatrix}D&-C\\\ -B&A\end{pmatrix}=\begin{pmatrix}A&B\\\ C&D\end{pmatrix}$ This implies $A=D$ and $B=-C$. In other words the matrix $M$ is of the form: $M=\left(\begin{smallmatrix}A&B\\\ -B&A\end{smallmatrix}\right).$ Next, we still have the condition that $M$ is symplectic, i.e. $\displaystyle M^{T}\Omega M=\Omega$ $\displaystyle\iff\begin{pmatrix}A^{T}&C^{T}\\\ B^{T}&D^{T}\end{pmatrix}\begin{pmatrix}0&I\\\ -I&0\end{pmatrix}\begin{pmatrix}A&B\\\ C&D\end{pmatrix}=\begin{pmatrix}0&I\\\ -I&0\end{pmatrix}$ $\displaystyle\iff\begin{pmatrix}-C^{T}&A^{T}\\\ -D^{T}&B^{T}\end{pmatrix}\begin{pmatrix}A&B\\\ C&D\end{pmatrix}=\begin{pmatrix}-C^{T}A+A^{T}C&-C^{T}B+A^{T}D\\\ -D^{T}A+B^{T}C&-D^{T}B+B^{T}D\end{pmatrix}=\begin{pmatrix}0&I\\\ -I&0\end{pmatrix}$ $\displaystyle\iff\begin{cases}A^{T}C=C^{T}A\\\ A^{T}D-C^{T}B=I\\\ B^{T}C-D^{T}A=-I\\\ B^{T}D=D^{T}B\end{cases}$ This means that $A^{T}C$ and $B^{T}D$ should be symmetric matrices and $A^{T}D-C^{T}B=I$. But now, due to the first condition this reduces to $B^{T}A=A^{T}B$ and $A^{T}A+B^{T}B=I$. In other words, the matrices we are looking for must be of the form $M=\left(\begin{smallmatrix}A&B\\\ -B&A\end{smallmatrix}\right)$ with $B^{T}A=A^{T}B$ and $A^{T}A+B^{T}B=I$, i.e. $\displaystyle\begin{pmatrix}A^{T}&-B^{T}\\\ B^{T}&A^{T}\end{pmatrix}\begin{pmatrix}A&B\\\ -B&A\end{pmatrix}=\begin{pmatrix}A^{T}A+B^{T}B&A^{T}B-B^{T}A\\\ B^{T}A-A^{T}B&B^{T}B+A^{T}A\end{pmatrix}=\begin{pmatrix}I&0\\\ 0&I\end{pmatrix}$ Which is exactly the condition for a unitary matrix. The map $\Phi:\mathsf{Sp}_{\mathbb{J}}(2n,\mathbb{R})\to\mathsf{U}(n):M\mapsto A+iB$ gives the wanted isomorphism. ∎ ### 3.2. Unitary invariant symplectic Dirac operators One can now check that the symplectic Dirac operator and its twist, are unitary invariant differential operators. This can be done by verifying that the operators commute with the following realisation of unitary Lie algebra $\mathfrak{u}(n)$: ###### Lemma 3.3. We have the following operator realisation of the Lie algebra $\mathfrak{u}(n)$: $\displaystyle\begin{cases}A_{jk}=y_{j}\partial_{x_{k}}+y_{k}\partial_{x_{j}}-x_{j}\partial_{y_{k}}-x_{k}\partial_{y_{j}}+i(q_{j}q_{k}-\partial_{q_{j}}\partial_{q_{k}})&\quad 1\leq j<k\leq n\\\ B_{jj}=y_{j}\partial_{x_{j}}-x_{j}\partial_{y_{j}}+\frac{i}{2}\left(q_{j}^{2}-\partial_{q_{j}}^{2}\right)&\quad 1\leq j\leq n\\\ C_{jk}=x_{j}\partial_{x_{k}}-x_{k}\partial_{x_{j}}+y_{j}\partial_{y_{k}}-y_{k}\partial_{y_{j}}+q_{j}\partial_{q_{k}}-q_{k}\partial_{q_{j}}&\quad 1\leq j<k\leq n\end{cases}$ This means that we can refine the $\mathfrak{sp}(2n)$-invariant PDE $D_{s}f=0$ into two $\mathfrak{u}(n)$-invariant PDEs given by $D_{s}f=0$ and $\widetilde{D}_{s}f=0$, for a symplectic spinor valued polynomial $f\in\mathcal{P}(\mathbb{R}^{2n},\mathbb{C})\otimes\mathcal{S}(\mathbb{R}^{n})$. In analogy with the orthogonal case, we call the solutions hermitian symplectic monogenics (or $h$-symplectic monogenics in short). ### 3.3. Symplectic Dolbeault operators Moreover, there is a second way of introducing the twist of the symplectic Dirac operator. Let us define the following operators which are known in the literature as the symplectic Dolbeault operators [4] defined by means of $D_{z}=\frac{D_{s}+i\widetilde{D}_{s}}{2}$ and $D_{z}^{\dagger}:=\frac{D_{s}-i\widetilde{D}_{s}}{2}.$ One easily verifies that $\displaystyle\frac{1}{2}(D_{s}+i\widetilde{D}_{s})$ $\displaystyle=-\sum_{j=1}^{n}\mathfrak{F}_{j}\partial_{z_{j}}\quad$ $\displaystyle\ \frac{1}{2}(D_{s}-i\widetilde{D}_{s})$ $\displaystyle=\sum_{j=1}^{n}\mathfrak{F}_{j}^{\dagger}\partial_{\overline{z}_{j}}.$ where we have introduced the symbols $\mathfrak{F}_{j}=(q_{j}+\partial_{q_{j}})$ and $\mathfrak{F}_{j}^{\dagger}=(q_{j}-\partial_{q_{j}})$. The structure of the operators $D_{z}$ and $D_{z}^{\dagger}$ is similar to the orthogonal case. However, the raising/lowering operators $\mathfrak{F}_{j}$ and $\mathfrak{F}_{j}^{\dagger}$ are used instead of isotropic Witt vectors $\mathfrak{f}_{j}$ and $\mathfrak{f}_{j}^{\dagger}$ (see for instance [6] and the references therein). ### 3.4. Class of simultaneous solutions of $D_{s}$ and $\widetilde{D}_{s}$ We will now describe a wide class of examples of $h$-symplectic monogenics, by making the link with holomorphic functions in several variables. Let $f:\Omega\subset\mathbb{C}^{n}\to\mathbb{C}$ be a complex-valued function in several complex variables which is of the class $\mathcal{C}^{1}(\Omega)$ (i.e. continuously differentiable). We say that $f$ is holomorphic (in several variables) if $\partial_{\overline{z}_{j}}f(z_{1},\dots,z_{n})=0$ for all $1\leq j\leq n$, where $\partial_{\overline{z}_{j}}:=\frac{1}{2}(\partial_{x_{j}}+\partial_{y_{j}})$ is the Cauchy-Riemann operator in the relevant variable. Moreover, we denote the set of holomorphic functions in $\Omega$ by $\operatorname{Hol}(\Omega)$. In order not to overload notations, we use the summation convention. Suppose that we have a function of the form $F({x},{y},{q})=e^{-\frac{1}{2}|{q}|^{2}}H({x},{y})$. Letting the symplectic Dirac operator act on $F$ gives: $\displaystyle D_{s}\left(e^{-\frac{1}{2}|{q}|^{2}}H({x},{y})\right)$ $\displaystyle=(iq_{k}\partial_{y_{k}}-\partial_{x_{k}}\partial_{q_{k}})\left(e^{-\frac{1}{2}|{q}|^{2}}H({x},{y})\right)$ $\displaystyle=iq_{k}e^{-\frac{1}{2}|{q}|^{2}}\partial_{y_{k}}H({x},{y})+e^{-\frac{1}{2}|{q}|^{2}}q_{k}\partial_{x_{k}}H({x},{y})$ $\displaystyle=e^{-\frac{1}{2}|{q}|^{2}}q_{k}(\partial_{x_{k}}+i\partial_{y_{k}})H({x},{y})$ We note that this equals zero if $(\partial_{x_{k}}+i\partial_{y_{k}})H({x},{y})=0$ for all $k=1,\dots,n$, i.e. if $H({x},{y})$ is a holomorphic function in several variables. Completely similar, $\displaystyle\widetilde{D_{s}}\left(e^{-\frac{1}{2}|{q}|^{2}}H({x},{y})\right)$ $\displaystyle=(iq_{k}\partial_{x_{k}}+\partial_{y_{k}}\partial_{q_{k}})\left(e^{-\frac{1}{2}|{q}|^{2}}H({x},{y})\right)$ $\displaystyle=e^{-\frac{1}{2}|{q}|^{2}}q_{k}(i\partial_{x_{k}}-\partial_{y_{k}})H({x},{y})$ $\displaystyle=ie^{-\frac{1}{2}|{q}|^{2}}q_{k}(\partial_{x_{k}}+i\partial_{y_{k}})H({x},{y}),$ which is zero for holomorphic $H({x},{y})$. This means that every function of the form $e^{-\frac{1}{2}(q_{1}^{2}+\dots+q_{n}^{2})}H({x},{y})$ with $H$ an holomorphic function is several variables is a solution of both $D_{s}$ and $\widetilde{D}_{s}$. This observation generalises the class of solutions obtained by Habermann in [2] for $n=1$. As a matter of fact, it turned out that there are much more solutions than the ones of this form. In order to describe these systematically, we will need the notion of Howe dualities and corresponding Fischer decompositions. This will be done in full detail in our upcoming paper [7]. In the following section we will reveal the algebraic structures required for this approach. ## 4\. A unitary Howe duality associated with $D_{s}$ and $\widetilde{D}_{s}$ Recall that the Lie algebra $\mathfrak{su}(1,2)$ is a quasi-split real form of the complex Lie algebra $\mathfrak{sl}(3)$ and is defined in terms of matrices as $\displaystyle\mathfrak{su}(1,2)=\left\\{\begin{pmatrix}\alpha&\beta&ic\\\ \gamma&\overline{\alpha}-\alpha&-\overline{\beta}\\\ id&-\overline{\gamma}&-\overline{\alpha}.\end{pmatrix}\mid c,d\in\mathbb{R}\ \&\ \alpha,\beta,\gamma\in\mathbb{C}\right\\}$ Some calculations now lead to observation that the Lie algebra generated by the symplectic Dirac operators $D_{s},\widetilde{D}_{s}$ and their duals $X_{s},\widetilde{X_{s}}$ gives rise to a copy of the Lie algebra $\mathfrak{su}(1,2)$. In order to close the algebra, we introduce the following differential operators: $\displaystyle\mathcal{O}:=\sum_{j=1}^{n}i(x_{j}\partial_{y_{j}}-y_{j}\partial_{x_{j}})+\partial_{q_{j}}^{2}-q_{j}^{2},$ $\displaystyle\Delta:=\sum_{j=1}^{n}\partial_{x_{j}}^{2}+\partial_{y_{j}}^{2}$ and $\displaystyle r^{2}:=\sum_{j=1}^{n}{x_{j}}^{2}+{y_{j}}^{2}.$ Together with symplectic Dirac operators $D_{s},\widetilde{D}_{s}$ and their duals $X_{s},\widetilde{X}_{s}$ these operators satisfy the commutator relations of $\mathfrak{su}(1,2)$. Moreover, their commutators are given by: $[\cdot,\cdot]$ | $D_{s}$ | $\widetilde{D}_{s}$ | $\Delta$ | $X_{s}$ | $\mathbb{E}$ | $\mathcal{O}$ | $\widetilde{X}_{s}$ | $r^{2}$ ---|---|---|---|---|---|---|---|--- $D_{s}$ | 0 | $\Delta$ | 0 | -$\mathbb{E}$ | $D_{s}$ | $-3\widetilde{D}_{s}$ | $-\mathcal{O}$ | $-2\widetilde{X}_{s}$ $\widetilde{D}_{s}$ | $-\Delta$ | 0 | 0 | $\mathcal{O}$ | $\widetilde{D}_{s}$ | $3D_{s}$ | $-\mathbb{E}$ | $2X_{s}$ $\Delta$ | 0 | 0 | 0 | $\widetilde{D}_{s}$ | $\Delta$ | 0 | $-2{D}_{s}$ | $\mathbb{E}$ $X_{s}$ | $\mathbb{E}$ | $-\mathcal{O}$ | $-\widetilde{D}_{s}$ | 0 | $X_{s}$ | $-3\widetilde{X}_{s}$ | $-r^{2}$ | 0 $\mathbb{E}$ | $-D_{s}$ | $-\widetilde{D}_{s}$ | $-\Delta$ | $-X_{s}$ | 0 | 0 | $-\widetilde{X}_{s}$ | 0 $\mathcal{O}$ | $3\widetilde{D}_{s}$ | $-3D_{s}$ | 0 | $3\widetilde{X}_{s}$ | 0 | 0 | $-3X_{s}$ | 0 $\widetilde{X}_{s}$ | $\mathcal{O}$ | $\mathbb{E}$ | $2D_{s}$ | $r^{2}$ | $\widetilde{X}_{s}$ | $-3X_{s}$ | 0 | 0 $r^{2}$ | $2\widetilde{X}_{s}$ | $-2X_{s}$ | $-\mathbb{E}$ | 0 | $-2r^{2}$ | 0 | 0 | 0 1. (1) We have two copies of the Heisenberg algebra: $\operatorname{Alg}\\{D_{s},\tilde{D}_{s},\Delta\\}\cong\operatorname{Alg}\\{X_{s},\widetilde{X}_{s},r^{2}\\}\cong\mathfrak{h}_{3}.$ 2. (2) We have two copies of the Lie algebra $\mathfrak{sl}(2)$ $\operatorname{Alg}\\{D,D^{\dagger},\mathbb{E}\\}\cong\operatorname{Alg}\\{\tilde{D},\tilde{D}^{\dagger},\mathbb{E}\\}\cong\operatorname{Alg}\\{\Delta,N,\mathbb{E}\\}\cong\mathfrak{sl}(2).$ This means that there is a canonical $\mathfrak{su}(1,2)$-action on the space of spinor valued polynomials $\mathcal{P}(\mathbb{R}^{2n},\mathbb{C})\otimes\mathcal{S}(\mathbb{R}^{n})$ where restricting to the subalgebra $\operatorname{Alg}(D_{s},X_{s})$ corresponds to $\mathfrak{sl}(2)$-copy obtained in the Howe duality for symplectic Clifford analysis (see [5] for more details). Now, taking into account the symplectic Dirac operators and its twists, we obtain the dual pair $\mathsf{U}(n)\times\mathfrak{su}(1,2)$ (i.e. the underlying group of invariance, together with the algebra generated by the operators and their duals). We now focus on the reduction of the symplectic spinor space. In the orthogonal case, the spinor space $\mathbb{S}$ decomposes under the action of the unitary group $\mathsf{U}(n)$ as $\mathbb{S}=\bigoplus_{r}\mathbb{S}_{(r)},$ with $\mathbb{S}_{(r)}$ inequivalent irreducible pieces, which are eigenspaces of the fermionic quantum oscillator (also called spin-Euler operator, see for instance [6]). In the symplectic case, the relevant operator for decomposing the infinite dimensional spinor space is the bosonic quantum oscillator. The hamiltonian of the quantum oscillator, the so-called Hermite operator, is given by $\displaystyle\mathcal{H}:\mathcal{S}(\mathbb{R}^{n})\to\mathcal{S}(\mathbb{R}^{n}),\quad f({q})\mapsto\frac{1}{2}\sum_{j=1}^{n}(\partial_{q_{j}}^{2}-q_{j}^{2})f({q}).$ Note that we can write $\mathcal{O}=\sum_{j=1}^{n}i(x_{j}\partial_{y_{j}}-y_{j}\partial_{x_{j}})+2\mathcal{H},$ so that the Hermite operator is in fact the spinor-valued part of the operator $\mathcal{O}$, i.e. the differential operator in $\partial_{q_{j}}$ and the variables $q_{j}$. Moreover, the eigenspaces can be identified with the irreducible decomposition of $\mathbb{S}^{\infty}$ into $\mathfrak{u}(n)$-irreducible representations. This means that the symplectic spinor space $\mathbb{S}^{\infty}$ decomposes into $\mathfrak{u}(n)$-irreducible representations ${\widetilde{S}}^{\infty}_{(k)}$ of dimension ${n+k-1\choose k}$ which can be thought of as $k$-homogeneous polynomials or the eigenspaces of the Hermite operator $\mathcal{H}$. Moreover, the solutions of the corresponding Dirac operators, called monogenics, can be introduced from a purely representation theoretical viewpoint. In general, this boils down to determining the decomposition (this is called a Fischer decomposition) $\mathcal{P}_{k}(\mathbb{R}^{m},\mathbb{C})\otimes\mathbf{S}$ where $\mathbf{S}$ is the spinor space, which is $\mathbb{S}$ in the orthogonal case and $\mathbb{S}^{\infty}$ in the symplectic case, where we take $m=2n$ in particular. Moreover, the space of $k$-homogeneous polynomials $\mathcal{P}_{k}$ coincides with the $k$-symmetric power of the fundamental representation of resp. the orthogonal or symplectic algebra. We denote by $\mathcal{M}_{k}$ the $k$-homogeneous solutions of the Dirac operator $\partial_{x}$, these are called monogenics. They can be defined as follows: $\displaystyle\mathcal{M}_{k}\leftrightarrow(k,0,\dots,0)\boxtimes\mathbb{S}=(k)\boxtimes\left(\frac{1}{2},\dots,\frac{1}{2}\right)\cong\left(k+\frac{1}{2},\dots,\frac{1}{2}\right),$ where $\boxtimes$ denotes the Cartan product of the $\mathfrak{so}(m)$-representations. In the symplectic case, we analoguously obtain: $\displaystyle\mathcal{M}_{k}^{s}\leftrightarrow(k,0,\dots,0)_{s}\boxtimes\mathbb{S}^{\infty}$ $\displaystyle=(k)_{s}\boxtimes\left(\left(-\frac{1}{2},\dots,-\frac{1}{2}\right)\oplus\left(-\frac{1}{2},\dots,-\frac{3}{2}\right)\right)$ $\displaystyle\cong\left(k-\frac{1}{2},\dots,-\frac{1}{2}\right)\oplus\left(k-\frac{1}{2},\dots,-\frac{3}{2}\right).$ In order to obtain an algebraic characterisation of the space of $h$-symplectic monogenics, one proceeds as follows. First of all, we note that we need to consider the symplectic spinors $\mathbb{S}^{\infty}$ from an unitary viewpoint. We saw that $\mathbb{S}^{\infty}$ decomposes as an infinite direct sum of finite dimensional $\mathfrak{u}(n)$-modules ${\widetilde{S}}^{\infty}_{(k)}$ which are in fact eigenspaces of the Hermite operator. We denote the branched spinor space (which is in fact a direct sum of $\mathfrak{u}(n)$-irreps) by $\widetilde{\mathbb{S}^{\infty}}$. However, the space of $k$-homogeneous polynomials is not irreducible as a $\mathfrak{u}(n)$-module and we denote the branched module by $\widetilde{(k)}$. This means that we are left with the following Cartan product $\mathcal{M}_{k}^{hs}\leftrightarrow\widetilde{(k)}\boxtimes\widetilde{\mathbb{S}^{\infty}}$ as a representation theoretical definition of the $h$-symplectic monogenics. Recall that this are the symplectic spinor-valued polynomial functions $f\in\mathcal{P}(\mathbb{R}^{2n},\mathbb{C})\otimes\mathcal{S}(\mathbb{R}^{n})$ that satisfy the system of unitary unitary-invariant partial differential equations $\begin{cases}D_{s}f&=0\\\ \widetilde{D}_{s}f&=0\end{cases}$ The explicit calculation of the Cartan product (and more generally the tensor product) will be done in [7]. Moreover, as an application we will prove a Fischer decomposition for the Howe dual pair we obtained in this paper. ## 5\. Conclusion In this paper we investigated a new Howe dual pair occurring in symplectic Clifford analysis by allowing a compatible complex structure. This Howe duality is of the form $(G,\mathfrak{g}^{\prime})$ where $G$ is the underlying invariance group for which the relevant Dirac operators are invariant and $\mathfrak{g}^{\prime}$ is the algebra generated by the Dirac operators and their duals. Depending on the orthogonal or symplectic framework, we have the following ‘types’ of Clifford analysis and refinements thereof: 1. (1) Orthogonal geometry (giving rise to a Clifford algebra) 1. (a) Clifford analysis: $\mathsf{SO}(m)\times\mathfrak{osp}(1|2)$ 2. (b) Hermitian Clifford analysis $\mathsf{U}(m)\times\mathfrak{osp}(2|2)$ 3. (c) Quaternionic Clifford analysis $\mathsf{USp}(m)\times\mathfrak{osp}(4|2)$ 2. (2) Symplectic geometry (giving rise to a Weyl algebra) 1. (a) Symplectic Clifford analysis: $\mathsf{Sp}(2m)\times\mathfrak{sl}(2)$ 2. (b) Hermitian symplectic Clifford analysis: $\mathsf{U}(m)\times\mathfrak{su}(1,2)$ 3. (c) Quaternionic symplectic Clifford analysis: $\mathsf{USp}(m)\times$? Thus far, we extended the framework of hermitian Clifford analysis in the presence of a symplectic structure in the case of the (flat) Kähler manifold $\mathbb{R}^{2n}$. It is an interesting question to further reduce the symmetry to the compact symplectic group $\mathsf{USp(n)}$ so that we have the chain $\mathsf{Sp}(2n)\supset\mathsf{U}(n)\supset\mathsf{USp}(n)$. In our furture work [7], we will describe the Fischer decomposition accompanying this new Howe dual pair. ### Funding information The author is supported by the FWO-EoS project ‘Symplectic Techniques in Differential Geometry’ G0H4518N. ## References * [1] D. Eelbode, G. Muarem, The orthogonal branching problem for symplectic monogenics, Advances in Applied Clifford Algebras 32, (2022) * [2] K. Habermann, L. Habermann, Introduction to symplectic Dirac operators, In Lecture Notes in Mathematics. Springer Berlin Heidelberg (2006) * [3] F. Brackx, H. De Schepper, D. Eelbode, R. Lávička, V. Soucek, $\mathfrak{osp}(4|2)$–monogenicity in Clifford analysis, In 15th International Conference on Computational and Mathematical Methods in Science and Engineering (CMMSE 2015) (Vol. 1, pp. 240-243) (2015) * [4] E. Korman, Symplectic Dolbeault operators on Kähler manifolds, Annals of Global Analysis and Geometry 44(3), 339-358 (2013) * [5] H. De Bie, P. Somberg, V. Soucek, The metaplectic Howe duality and polynomial solutions for the symplectic Dirac operator, Journal of Geometry and Physics 75, 120-128 (2014) * [6] D. Eelbode, Stirling numbers and spin-Euler polynomials, Experimental Mathematics, 16(1), 55-66 (2007) * [7] Eelbode, D., Muarem G. (2022). A unitary refinement of the symplectic Fischer decomposition. In preparation.
# Security Issuance, Institutional Investors and Quid Pro Quo: Insights from SPACs††thanks: We would like to thank Christian Opp, Joanna Wu, Michael Gofman, and Bill Wilhelm for their helpful comments and suggestions. We thank the 3Cavaliers at the University of Virginia for their generous research support and Lam Bui and Hengxiang Cao for their excellent research assistance. Gaurab Aryal Zhaohui Chen Yuchi Yao Chris Yung Department of Economics, Washington University in St. Louis<EMAIL_ADDRESS>School of Commerce, University of Virginia<EMAIL_ADDRESS>School of Business, University of Rochester<EMAIL_ADDRESS>School of Commerce, University of Virginia<EMAIL_ADDRESS> ###### Abstract Security issuance is subject to informational and agency-related frictions. However, it is difficult to separate their effects. We consider SPACs and assess those effects separately. To this end, we identify premium investors who produce value-relevant information: their participation is associated with higher SPAC success and announcement-day returns. Non-premium investors, however, engage only in quid pro quo arrangements. Their high returns today from issuers (quid) mean they are more likely to participate in low-returns future deals initiated by those issuers (quo). Thus quid pro quo is not pure agency cost but partly a transfer that enables more firms to go public. Keywords: Security Issuance, Institutional Investors, SPAC, Underpricing JEL classification: G14, G32, G34. ## 1 Introduction Firms issuing securities through intermediaries face two types of friction. One, there may be an informational gap between them (issuers) and the investors, which may prevent securities from being sold at full value. Second, there is an agency problem if intermediaries do not maximize sale proceeds but instead engage in quid pro quo arrangements with investors. Specifically, intermediaries may favor certain buyers (investors) at the expense of the issuers, expecting their favoritism to be returned through profitable future transactions. Ritter and Welch (2002) point out that both frictions coexist and emphasize the need and difficulty in separating those frictions. In a survey paper, Lowry et al. (2017) similarly pose this separation as an unresolved question, using IPO underpricing as an example: > Does [IPO underpricing] reflect a reward to institutions for sharing value- > relevant information with underwriters, or does it reflect a quid pro quo > arrangement where the bank is rewarding its best clients? [..] This is an > immensely important issue and one that we hope future research will be able > to address. We revisit this problem in the context of special purpose acquisition companies (SPACs) and assess and separate the effects of these frictions on security issuance.111We focus on broader issues in security issuance rather than questions specific to SPACs and their regulatory oversight. Thus we sidestep questions about the issuer’s choice between traditional IPO and SPAC, whether SPACs have misled public investors, and how SPACs should be regulated. See Gahng, Ritter and Zhang (Forthcoming) and Klausner, Ohlrogge and Ruan (2022) for comprehensive reviews. To this end, we identify two types of institutional investors: “premium” and “non-premium” (defined shortly below). We find evidence consistent with premium investors acting in the role of information production and non-premium investors engaging in only quid pro quo arrangements. However, unlike the extant literature that identifies any quid pro quo as purely value-destroying, our estimates suggest that in SPACs, quid pro quo can create value by enabling marginal businesses to access public markets. We can separate the effects of information and quid pro quo by relying on the institutional features of SPACs. A SPAC is a publicly traded shell (i.e., blank check) corporation formed by a sponsor and public investors with the sole purpose of effecting a merger with a private firm to enable the latter to go public. As such, SPACs offer four research design advantages over a traditional IPO. First, unlike traditional IPOs, a SPAC divides the process of going public into two distinct phases that are affected by either informational or agency- related friction, but not both. In the initial stage, the _sponsor_ of the SPAC raises capital from investors. The SPAC starts as a shell company without an operating business, which limits the amount of information asymmetry compared to a traditional IPO.222Consistent with this, in our sample, we find a small (1.3%) underpricing at SPAC IPO. The next phase consists of a “business combination” in which the SPAC acquires or merges with a private business. Second, during the business combination phase, SPACs report the allocations to each institutional investor or the _private investment in public equity_ (acronym, “PIPE”) investor. These data enable us to measure the investment return history for every sponsor-PIPE pair and create in-sample measures of past “relationships” based on their track records. In traditional IPOs, allocations to institutional investors are not reported, so one cannot credibly measure past relationships between underwriters and investors.333In a traditional IPO, although one can read quarterly 13F filings after the IPO, it may be insufficient to quantify favoritism if preferred clients are allowed to “flip” shares in secondary market trading. Furthermore, 13F filings do not distinguish between shares allocated at the IPO and purchased in the secondary market. Third, SPACs are publicly traded firms. Consequently, we observe the effect of major events (e.g., securing new external capital or identifying a target) on share prices. Fourth, the quid pro quo channel is easier to observe in SPACs than in traditional IPOs. Unlike the investment banks in traditional IPOs, SPAC sponsors typically do not operate in other lines of business (e.g., trading and wealth management). So, to the extent that sponsors earn quid pro quo from investors in future SPACs deals, we can measure those gains. Once the SPAC finds a target and negotiates a “business combination” deal, the existing SPAC investors have to approve the deal. Those who do not like the deal can redeem their investments. Such redemptions reduce the money available for the post-merger firm. As a safeguard, once the target is identified but before the target firm begins public trading, the SPAC sponsor brings in PIPE investors. These investors that participate in the target announcements are the analogs to the institutional buyers in the traditional IPO bookbuilding models. Specifically, the sponsor finds a target and approaches PIPE (at) investors to solicit their interest. Then the sponsor builds a book of PIPE demand, sets prices, and allocates shares. Thus, when it comes to the business combination phase, SPAC and traditional IPO are similar in that they both are subject to information and agency frictions–the subject of this paper. Much of our analysis focuses on the role of such PIPE investors. We classify them into two types of investors$\\--$premium and non-premium$\\--$based on their participation frequency and investments in SPACs, using k-means clustering, which is an objective approach to classification. The underlying assumption is that the PIPE investors who participate frequently and invest large amounts are likely more reputable than others (Benveniste and Spindt, 1989), which behooves them to conduct due diligence and evaluate the proposed merger deal. Using data on all 1,072 SPACs from January 2010 to December 2021, we find evidence consistent with premium investors offering value-relevant information and playing certification roles. In particular, the participation of premium investors is associated with significantly less share redemption by public SPAC investors and a large positive announcement effect on the SPAC stock price. The estimates suggest that one standard deviation (32.85 pps) increase in premium PIPE investors’ participation is associated with a seven pps decrease in the redemption rate (which is 12.4% of the sample mean) and a four pps increase in the SPAC’s announcement-day return (which is 108.8% of the sample mean). Creating a SPAC and finding a suitable target falls on the sponsors, who act as intermediaries between the target (who issues shares) and PIPE investors (who buy shares). Sponsors profit an average of $126.9 million when the merger is successful: otherwise, the SPAC is liquidated, their securities expire worthlessly, and they lose their initial investment, which, on average, is $8.2 million. Therefore, the sponsors benefit from participation by PIPE investors which improves the chances of a successful merger. So, the sponsors may have incentives to offer discounted shares to PIPE investors in strong deals in exchange for a tacit promise to help out in (future) weaker deals. Under such a quid pro quo arrangement, high-quality firms (inadvertently) transfer to (in part) low-quality targets by allowing the latter to go public.444This transfer between issuers is reminiscent of the mechanism described in Benveniste, Busaba and Wilhelm (2002), who argue that underwriters cross-subsidize issuers depending on their position in a wave. There, the leader is subsidized, whereas, in SPACs, the follow-on SPACs (by the same sponsor) are subsidized. Such a quid pro quo arrangement suggests the following testable predictions. First, for each sponsor-PIPE pair, money left on the table (acronym, “MLOT”) should be mean-reverting, i.e., high profits in a current deal predict low profits in a future deal, and vice versa. As is standard in the literature (e.g., Lowry and Schwert, 2004), we measure MLOT using the first-day closing price, taken as a proxy for the true value. Second, when the deal is weak, the sponsor allocates shares disproportionately to those who earned strong returns from that sponsor in the past. Third, when a sponsor with strong relationships with investors, accumulated through past deals, leads a SPAC, that SPAC takes fewer days to find a target and is less likely to be liquidated. These predictions have no analog in the information production paradigm: investors who did well in the past receive new signals of the current deal’s quality and are under no obligation to support future weak deals. If anything, the MLOTs of the premium investors should exhibit positive dependence if investor skill is persistent. We find evidence consistent with all three quid pro quo predictions, but _only_ when considering non-premium investors. Specifically, non-premium investors profit less in the current deal when they have high returns on past deals. In particular, one dollar _above_ average MLOT in the last deal leads to a $2.72 _below_ average MLOT from the current deal. We interpret this _mean-reversion_ as the sponsor building a relationship and drawing it down, likely for weaker deals. We refer to weaker deals as “cold” deals and use the difference between the offer share price of the SPAC, i.e., the share price at SPAC IPO, and the share price on the closing day of the merger to measure it. Indeed, we find that investors with a stronger relationship are more inclined to participate in weaker deals. In particular, non-premium investors with a strong relationship with a sponsor (with a total MLOT from this sponsor that is one standard deviation above the sample mean) are 3.54 pps more likely to participate in deals with the cold measure that are in the 90th percentile of sample than deals with the cold measure that are at 10th percentile. Conversely, non-premium investors without a prior relationship with the sponsor are less likely to participate in cold(er) deals. Moreover, we observe that when a non-premium investor with a strong relationship does participate, the sponsor allocates a larger portion of securities to them in weaker deals than in stronger deals. We contribute to the literature by establishing a new institutional theory in security issuance based on a sponsor’s relationship with institutional investors. We provide evidence that the quid pro quo arrangement enables cross-deal subsidies that enable weaker firms to go public by reducing the SPACs’ liquidation risk. In contrast, the existing literature treats quid pro quo as pure agency costs. More precisely, we find that if sponsors allocate one standard deviation more MLOT above average ($300 million) to non-premium investors, the liquidation probability of the current deal falls by 48.93 pps. As a reference, the average liquidation rate in our sample is 17.82%, so this decrease is economically meaningful. However, the MLOT allocated to the premium investors is uncorrelated with the liquidation risk.555In Appendix A, we verify that our main findings are robust to the presence of outlier (large and small) investors and placebo assignments. We also contribute to the information production mechanism by measuring the information content of institutional investors. The traditional IPO literature argues that price-limited bids are likely more informative than market-order bids (Cornelli and Goldreich, 2001; Jenkinson, Jones and Suntheim, 2018). SPACs are publicly traded and thus afford us the advantage of quantifying price reactions to information conveyed by PIPE investments. Overall, premium and non-premium investors play different but complementary roles in SPACs and are paid accordingly. In our sample, each premium investor, on average, is paid $6.96 million, and a non-premium investor $3.17 million. Moreover, their observable differences can only explain $0.04 million of the $4.79 million gap between their pay. We interpret this result as evidence of their different “business models.” That is, a premium investor is paid $6.96 million to produce information for the current target, and a non-premium investor is paid $3.17 million to reduce liquidation risk for the future SPAC. Related Literature. Our research contributes to a long line of research that uses underpricing to infer the importance of agency and informational problems. For example, Benveniste and Spindt (1989) predict positive average underpricing because investors are rewarded for value-relevant information. The implication is that in a traditional IPO, when prices rise from the midpoint of the filing range until the offer price (i.e., when demand is high), there should be further price increases when the stock starts trading. In other words, underwriters partially adjust to good information revealed by investors but fully adjust to bad information.666See Hanley (1993) for evidence of partial adjustment. Bradley and Jordan (2002) and İnce (2014) argue that prices respond to public information, which is inconsistent with bookbuilding models, whereas Lowry and Schwert (2004) argue that the IPO price is “almost efficient” with respect to public information. In contrast, Beatty and Ritter (1986); Loughran and Ritter (2002) and Loughran and Ritter (2004) argue that underpricing is pure unproductive rent. We complement this literature by looking at a broader range of outcomes: redemption, announcement price reactions, liquidation, and allocations, and show that both information and quid pro quo arrangements matter. We also contribute to a nascent literature that seeks to understand the role and functioning of SPACs. For instance, Gahng, Ritter and Zhang (Forthcoming) study the cost and benefit of using SPACs to go public; Gofman and Yao (2022) study the conflict of interest and associated harm from directors serving on the boards of multiple SPACs; and Klausner, Ohlrogge and Ruan (2022) analyze the structure of SPACs and the associated costs. Our findings about the differential effects of premium and non-premium investors complement the findings in Wang and Yung (2011) that partial adjustment is present only in the subset of highly reputable underwriters. In particular, they show that low-reputation underwriters cluster heavily on exactly zero-dollar price revision. This behavior is consistent with no information being produced or revealed. We also complement Chemmanur, Hu and Huang (2010), who study the role of institutional investors in IPOs. Our finding that information production coexists with quid pro quo is reminiscent of the findings in Jenkinson, Jones and Suntheim (2018). Using data on the initial allocations of investment banks participating in traditional IPOs, they find evidence that investors are rewarded more when they submit price-limited bids and participate more actively in the meetings before or during bookbuilding. Both behaviors are consistent with an informational role. However, they also find that allocations are affected by how much revenue the investor generates for the bank, which is consistent with the quid pro quo argument. As previously mentioned, SPACs offer an opportunity to examine this issue in a context where we can directly observe the effect of PIPE investor participation on redemption decisions, announcement period returns, and the likelihood of liquidation, none of which are observable in a traditional IPO. The notion of quid pro quo employed in our papers is slightly different. Jenkinson, Jones and Suntheim (2018) document that investors build a relationship with a bank to get underpriced shares. We instead document that _sponsors_ build relationships with non-premium PIPE investors to use them to complete future deals. ## 2 Institutional Details This section introduces important institutional details about SPACs. In particular, these features of SPACs are important for understanding how they differ from traditional IPOs and the measurement of underpricing in the data that follow in the next section. ### 2.1 A Primer on SPACs A SPAC is formed when a sponsor begins the SEC-mandated initial public offering process. A SPAC is, first and foremost, an IPO and must comply with all SEC IPO regulations. In addition, SPACs are subject to additional requirements as “blank check company”, including how they use their escrow fund (described below). The SPAC’s purpose is to identify and acquire a private company, thereby bringing it public. During its IPO process, the SPAC usually announces its intended target industry. This industry is meant to reflect the expertise of its management team and founders and to inform potential investors of its acquisition strategy. More often than not, these industries are broad, e.g., financial services, healthcare, and technology, so as not to constrain the search. A SPAC, however, is not obligated to pursue a target in the announced industry, although most do.777For example, _Hunter Maritime Acquisition Corp._ announced it would target the international maritime shipping industry but merged with a fintech company named _NCF Wealth Holdings_. During their IPOs, SPACs offer securities to public investors called _units_ , typically priced at $10 per unit.888One exception is _Periphas Capital Partnering Corporation_ , which priced its unit at $25. A unit consists of one share of Class A common stock and a proportion (e.g., 1, 1/2, 1/3) of one redeemable warrant. Initially, these units trade as a whole. However, usually around the 52nd calendar day following the IPO, owners have the option to either continue to hold units as a whole or trade shares and warrants separately.999After the separation, SPACs do not issue fractional warrants and only whole warrants trade. In other words, the number of warrants issued after separation equals the fraction rounded down to the nearest whole number. For example, if a unit contains one share and 1/3 of a warrant, then an investor with 1 unit receives no warrant, but an investor with 30 units receives ten warrants. Once the merger deal between the SPAC and a target is closed, a warrant gives an investor the right to purchase additional shares of common stock. Typically, the strike price is set well above the original SPAC price; a common choice is $11.50 per share. Any shares created in this manner automatically convert into ordinary shares after a merger. Sponsors purchase shares before SPACs go public. These shares are called the “founder shares” or the sponsor’s “promote”. These founder shares are identical to the Class A common shares included in the units owned by public investors, except that only the founder shares (i) contain the voting rights to elect directors, (ii) are subject to transfer restrictions (“lock-up”), and (iii) are irredeemable and become worthless if the SPAC liquidates. A SPAC generally has 24 months from its IPO to identify a target company, negotiate a deal, and complete the merger. Failing to complete the merger on time leads to its liquidation, and the SPAC has to redeem all public shares and pay its investors with the fund held in the trust account, including any interest earned. Once the SPAC locates a target, it uses proceeds from the trust fund for the merger. Until then, the trust fund cannot be touched except to pay public shareholders who redeem. To meet operating expenses, the sponsor creates a “risk capital” by selling private placement warrants. These warrants are irredeemable when the SPAC liquidates and are subject to transfer restrictions. Next, SPAC shareholders vote on the deal, and the transaction occurs only if a majority votes in favor. Shareholders also can redeem their shares, typically having at least 20 business days to redeem their shares before the merger is complete. The merger agreement with the target company requires that the SPAC have a minimum net worth or a certain amount of cash (as per the negotiation) as a closing condition. Thus, if too many public shareholders redeem, the SPAC may not meet the closing conditions of the merger. These financing shortfalls create a need for _private investments in public equities_ (PIPE) investors. ### 2.2 PIPE Investment Sometimes the funds raised in the SPAC IPO are sufficient to purchase the target located by the sponsor. More typically, the sponsor needs (or chooses) to raise additional funds. Because the SPAC is a publicly traded firm, these funds are referred to as PIPEs. PIPE investors are accredited institutional investors, e.g., hedge funds and mutual funds, that agree to purchase securities issued by a company. Such an agreement requires the company to file a resale registration statement (in the future) to allow investors to resell their securities. Compared to raising capital in the public market, PIPE investments tend to have lower transaction costs, and companies can disclose transaction details _after_ receiving commitments from PIPE investors. We categorize PIPE transactions based on the timing of their commitment: before, at, and after the business combination. In particular, the sponsor can invite PIPE investors before the SPAC IPO to commit to making certain investments in connection with the business combination. The PIPE transactions formed at the time of the IPO but before a target is identified are called forward purchase agreements; henceforth, PIPE (pre). The second type of PIPE transaction is the one announced along with the news about the business combination; henceforth, PIPE (at). The third type of PIPE transaction occurs after the target is announced but before the merger is completed; henceforth, PIPE (post). Table 1: Example of Key Events and Timeline for a SPAC Key Events | Dates ---|--- S-1 Filed and PIPE (pre) Announced | 08-Jan-2021 IPO | 29-Jan-2021 Deal Announced, PIPE (at) Announced | 06-Jul-2021 Proxy Filed | 12-Aug-2021 PIPE (post) Announced | 18-Jan-2021 Shareholder Vote | 24-Jan-2022 Deal Closed | 25-Jan-2022 Liquidation Deadline | 02-Feb-2023 Note: Key events and the dates for the SPAC _CF Acquisition Corp V_. In principle and practice, a SPAC can have any combination of these three types of PIPEs. For example, 1 shows an example SPAC, CF Capital Acquisition Corp V, which had all three types of PIPEs. The PIPE (at) investors are similar to the institutional investors in traditional IPO. After finding a target, sponsors engage in an “roadshow” to raise PIPE investments. During these roadshows, sponsors share material and proprietary information about the target, creating a scope for information production. ### 2.3 SPAC vs. Traditional IPO SPAC IPOs are different from traditional IPO in that SPAC is a blank-check company without a business model. However, the merger between a SPAC and a target, or de-SPAC transaction, is analogous to the traditional IPO because, during the de-SPAC process, a private company transitions to public trading and receives new financial capital in exchange for newly created securities.101010For more about differences and similarities see the SEC report “ _Special Purpose Acquisition Companies, Shell Companies, and Projections_ ” available at https://www.sec.gov/rules/proposed/2022/33-11048.pdf. In Figure 1, we compare the timeline of a SPAC and a traditional IPO. Figure 1: Events timing for Traditional IPO and SPAC Note: Illustration of the timelines of a traditional IPO (top) and a SPAC (bottom). Here, the acronym “BC" stands for a business combination. A private operating company’s path to becoming a publicly listed company through a SPAC begins when the SPAC considers it a candidate for the business combination. The sponsors perform due diligence and negotiate a letter of intent with the target company. Once the letter is signed, the second type of PIPE fundraising process begins. In particular, sponsors and the target company management consult with potential PIPE investors, pitch the company’s business plan, and take subscriptions for shares from PIPE investors. Thus, PIPE investors are similar to institutional investors in traditional IPOs. After PIPE investments are secured, the deal is announced to the public investors, and the de-SPAC process starts. Private companies become publicly listed on the secondary market when the business combination is closed. Finally, the _Private Securities Litigation Reform Act_ provides a safe harbor for forward-looking statements made by SPACs in their de-SPAC transactions. There is no such safe harbor for the traditional IPO. Thus, the private company’s performance projections can be reported to investors during the de- SPAC announcement. ## 3 Testable Hypotheses We develop testable hypotheses based on two competing paradigms – the information production and quid pro quo arrangements – that we assess in our data. Information Production. Under the first paradigm, investors have or obtain private information about the issuer and earn rents according to the value of their information. We refer to this paradigm as “information production” because it views the central goal of the “going public” process as producing or revealing value-relevant information about the issuer. The process of going public via SPACs has three instances that involve investors (see Figure 1). So, information production will have different implications for those steps. In particular, information production predicts zero underpricing during the SPAC IPO because there is yet to be a target identified and, thus, no due diligence to perform. There is virtually no scope for investors to have private information; they exchange cash for a proportion of a trust fund of a fixed value. In the business combination or de-SPAC phase, investors that participate in the target announcement (PIPE (at)) are the analogs to the buyers in the IPO bookbuilding models. Specifically, the sponsor finds a target and approaches PIPE (at) investors to solicit their interest. Then the sponsor builds a book of PIPE demand, sets prices, and allocates shares. However, these investors may differ in their abilities to produce information. For instance, Benveniste and Spindt (1989) suggest that premium investors who frequently access the market are likely more informed than others. In our context, such investors have private information about the quality of the issuing (target) firm and overall conditions affecting the market valuation. Conversely, non-premium investors are less informed and consequently have less impact on prices. Under these assumptions, the securities of the target firm may be underpriced to compensate premium investors for truthfully disclosing their private information about the target. Furthermore, a premium PIPE (at) investor is incentivized to purchase more shares when its private information indicates the de-SPAC deal is profitable. Additionally, sponsors prioritize allocating to premium investors to encourage them to disclose their information. Consequently, a larger allocation to premium investors signals the good quality of the deal, which reduces redemption by public investors and leads to a positive market reaction. However, as non-premium PIPEs have no certification role, their participation should neither affect redemption nor lead to higher announcement-day returns. Conversely, we posit that PIPE (pre) and PIPE (post) investors do not play certification roles. In the case of PIPE (pre), there has yet to be any information about the target, so there is nothing to certify. PIPE (post) investors are used less often than PIPE (at) investors and can be used to fill the financing gap caused by redemption. If the sponsor intended a PIPE investor to be used for certification purposes (as opposed to “gap filling”), the sponsor would have announced the participation of the PIPE contemporaneously when the deal was announced rather than waiting until after public investors reacted. Finally, suppose a premium PIPE (at) investor acts purely as an information producer and is under no obligation to the sponsor to maintain a relationship. In this case, the PIPE will participate in deals it likes and be absent from deals otherwise. Such an arrangement should not generate any mean reversion when the unit of observation is sponsor-PIPE. ###### Hypothesis 1. Under the information production paradigm, the following should hold: 1. 1. Allocations to premium PIPE (at) investors are associated with lower redemption and positive announcement-day returns. 2. 2. For a sponsor-premium PIPE (at) pair, MLOT is not mean-reverting. Quid Pro Quo. We now discuss the implications of the quid pro quo paradigm. Under this paradigm, the seller of securities acts through a self-interested intermediary. Rather than selling securities at full price, say, to the premium investors, the intermediary gives discounts to some buyers, with an implicit understanding that this favor is rewarded through profits on future SPAC transactions. The SPAC IPO presents the underwriter with such an opportunity. Underwriters likely deal with institutional investors more frequently (and in more varied capacities) than sponsors. As an intermediary, the underwriter may choose to favor institution investors (the buyers) over sponsors (the sellers), just as in a traditional IPO. If so, the SPAC offering will be underpriced. The business combination phase presents another opportunity for us to observe favoritism, except the actors are swapped. The issuer is the private firm rather than the SPAC because it is the private firm, not the SPAC, that issues new securities. The buyers are the PIPE investors who infuse new capital in exchange for securities. The intermediary is the sponsor that negotiates the terms under which PIPE investors buy these securities and profits only if the deal is completed. Suppose that the underpricing at the business combination phase is not because of information production but because the sponsor allocates rents to PIPE (at) investors. Then, the sponsor should receive returns from these investors in the future. If so, then including PIPE investors in a SPAC would not benefit the SPAC. Therefore, higher allocations to any PIPE (at) investors should _not_ be associated with either lower redemption rates or higher announcement- day returns. In any given deal, sponsors can favor particular PIPE investors and insist on favorable terms from the target firm. Favoritism may manifest differently at this stage than in traditional IPOs. In a traditional IPO, underwriters operate in many lines of business, including trading and wealth management. An institutional investor receiving underpriced shares in a traditional IPO may repay the underwriter by providing business through other channels. In contrast, the sponsors operate only in the SPAC industry, and the primary objective of a SPAC sponsor is to close the existing fund and start future SPACs We hypothesize that quid pro quo arrangements are employed only with non- premium PIPE investors. A sponsor showing favoritism to a particular non- premium PIPE investor expects reciprocity in future deals. Specifically, non- premium PIPE investors participate in an underpriced deal (acting as a tax on the current private firm) only in exchange for implied participation in a future weak deal. If a PIPE is involved only in “strong” deals, then the relationship is one-sided and involves “quid” with no “quo.” While a long-term relationship (e.g., Levin, 2003) between a non-premium PIPE investor and a sponsor can provide necessary incentives to the investors to participate in weaker future deals, the observed interactions in any data are likely finite. Then the question remains if non-premium investors have incentives to participate in weaker future deals. In our context, it is appropriate to view non-premium investors are long-lived players who interact with sponsors who are short-lived players. In such a setting, results from Fudenberg and Levine (1994) indicate that an investor does not need to have repeated interactions with a particular sponsor for the sponsor to have the incentive to honor their commitment to participate in weak future deals. That is because an investor’s participation is publicly observed. Suppose they refrain from participating in a weak deal led by a sponsor from whom they have received favorable returns. In that case, they risk losing their market-wide reputation and opportunities to participate in future SPACs. Furthermore, under quid pro quo arrangements, the sponsor benefits in creating goodwill among investors, hoping to “draw down” on this goodwill in the future when the deals are weak. So, we hypothesize that the liquidation risk of a SPAC is negatively correlated with the amount of MLOT investors have received from this SPAC’s sponsor in the past. ###### Hypothesis 2. Under the quid pro quo paradigm, the following should hold: 1. 1. Allocations to non-premium PIPE (at) investors are uncorrelated with redemption and announcement-day returns. 2. 2. For a given sponsor -non-premium PIPE (at) pair, MLOT is mean-reverting. 3. 3. When a deal is weak, the sponsor is likelier to allocate shares to non-premium PIPE investors who earned large past MLOT. 4. 4. SPACs are less likely to liquidate when the sponsor has earned goodwill through high past MLOT. This connection occurs through only non-premium and not premium PIPEs. ## 4 Data We consider all 1,072 SPACs traded in major exchanges between January 2010 and December 2021. See Figure A4 for the monthly scatter plot in Figure A4, which includes the so-called boom and bust cycle for SPACs. We construct our sample by merging five datasets: (i) the _SPAC Research_ ,111111https://www.spacresearch.com/. (ii) the _PrivateRaise_ –SPACs Platform,121212https://www.privateraise.com/about/platform_spac.php. (iii) the _PrivateRaise_ –PIPEs Platform (iv) _Bloomberg_ data on SPAC prices, and (v) Super 8-K filings from SEC.131313https://www.sec.gov/edgar/search-and- access.,141414Thus, we exclude SPACs traded in the over-the-counter markets. After a SPAC’s merger with a target is complete, the SPAC is delisted. The new (merged) company may be assigned a new ticker symbol. We use Bloomberg- assigned unique identifiers for matching across samples. The SPAC Research database and the PrivateRaise’s SPACs platform database contain information on SPACs, including dates for important events, and the PrivateRaise’s PIPEs platform database contains comprehensive information on PIPE transactions. PrivateRaise’s PIPEs platform covers more than 20,000 placement profiles of PIPEs transactions. It also includes contract terms, investment amounts, types of transactions, prices, identities, and allocations of the PIPE investors. For these 1,072 SPACs, we manually merge the first two databases. The issuer’s name recorded in the PIPE transactions can be either that of SPAC or that of the post-merger company. In those cases, we use both names to match the SPAC sample with the PrivateRaise database. To ensure match accuracy, we also use the total PIPE amount from PrivateRaise to cross-check the match. If there are inconsistencies, we use the SPAC’s SEC filings documents to record the correct amount. We then merge our sample with Bloomberg’s pricing data; warrant prices are unavailable in CRSP. Our matching process yields detailed information on transactions for all three types of PIPE investments for each SPAC in our sample. In summary, we observe each SPAC’s life-cycle status, IPO structure, IPO outcomes, sponsor and management team, redemption, merger details, merger outcome, PIPE transactions, and price movement across events. Out of these 1,072 SPACs, as of February 17, 2023, 473 have successfully merged, 128 have announced that they have identified a target, 280 are in the process of identifying targets, and 191 have failed and liquidated. The number of SPAC shares changes over time on account of i) investors exercising warrants, ii) redemption of shares, iii) securities purchased by PIPES, and iv) the fact that founders sometimes voluntarily forfeit securities. Consequently, the number of shares listed in the IPO prospectus is no longer current as of the business combination date. We use Super 8-K (Super 20-F for foreign issuer) filings to collect information on the number and types of outstanding securities on the first listing day of the post-merger company. Information about securities at the time of the SPAC IPO may not be comprehensive either because the sponsors have forfeited their founder shares and private placement securities or the units that contain rights would have become common shares or both. To accurately measure the number of outstanding securities post-merger, we manually collect information from the Super 8-K (or Super 20-F) filings. SEC mandates that a SPAC file Super 8-K within four business days since the merger is complete, and it records the shares and warrants held by sponsors and public investors. Table 2 provides summary data. On average, SPACs raise $272.42 million proceeds during their IPO, with a median of $240 million and 90th percentile at $440 million. In terms of securities, typically, a SPAC unit contains 0.44 warrants and 0.01 rights. During the IPO process, SPACs’ underwriters choose, on average, to exercise the 11.59% over-allotment option to meet the additional demand. SPACs take, on average, 361 days to find a target and an additional 168 days to complete the merger process. After targets’ announcements, 56.54% of SPAC’s public shareholders redeem their shares on average. Table 2 also displays information about the participation of PIPE investors for 473 SPACs that completed the merger process during our sample period. At the time of their IPOs, 10% of SPACs have pre-committed investment from PIPE investors, i.e., PIPE (pre). When SPACs announce their targets, 66% of them also announce investments from PIPE (at), and 12% of SPACs obtain additional investments from PIPE (post) after the target announcement. Returns. We measure SPAC returns at its IPO and when it announces a target. Following the literature on IPO underpricing, we measure SPAC’s IPO return as $\text{Listing-day return}=\frac{\text{First-Trading Day Closing Price}}{\text{Offer Price}}-1,$ and announcement-day return as $\text{Announcement-day return}=\frac{\text{Announcement Day Closing Price}}{\text{One-day-before-announcement Day Closing Price}}-1.$ As seen in Table 2, the average return for SPAC IPO investors on the first trading day is 1.3%, and when the target is announced, SPAC investors receive an average return of 3.74%. Money Left on the Table. To measure underpricing, i.e., the return earned by investors, we use _money left on the table_ (“MLOT”), which is the difference between the value of securities on the first day of listing the post-merger company and their initial investment. We begin with the total investments made by public investors, which we denote using superscript $\mathfrak{pu}$. Their MLOT is the sum of the value of their shares and the value of their warrants minus their leftover investments during IPO, i.e., $\text{MLOT}^{\mathfrak{pu}}=\text{Shares}^{\mathfrak{pu}}\times\text{P}_{share}+\text{Warrants}^{\mathfrak{pu}}\times\text{P}_{warrant}-\text{IPO Proceeds}\times(1-\text{Redemption Rate}).$ Table 2: Summary Statistics | Mean | Std.Dev. | P10 | P50 | P90 | Obs. ---|---|---|---|---|---|--- SPAC Measures | | | | | | IPO proceeds (mm USD) | 272.42 | 218.22 | 100.00 | 230.00 | 440.00 | 1,072 No. of Warrant | 0.44 | 0.24 | 0.20 | 0.50 | 0.75 | 1,072 No. of Right | 0.01 | 0.03 | 0.00 | 0.00 | 0.00 | 1,072 Overallotment (%) | 11.59 | 5.37 | 0.00 | 15.00 | 15.00 | 1,072 Listing-day return (%) | 1.30 | 3.02 | -0.70 | 0.40 | 4.00 | 857 Days searching | 361.71 | 215.79 | 114.00 | 329.50 | 677.00 | 614 Days remaining | 340.98 | 189.87 | 120.00 | 303.00 | 611.00 | 614 Announcement-day return (%) | 3.74 | 12.56 | -0.68 | 0.34 | 9.95 | 586 Redeemption(%) | 56.54 | 37.49 | 0.00 | 70.00 | 97.00 | 445 Days deSPAC | 167.99 | 67.06 | 100.00 | 160.00 | 241.00 | 474 PIPE Participation | | | | | | $\mathbbm{1}${PIPE (pre)} | 0.10 | 0.30 | 0.00 | 0.00 | 0.00 | 473 $\mathbbm{1}${PIPE (at)} | 0.66 | 0.47 | 0.00 | 1.00 | 1.00 | 473 $\mathbbm{1}${PIPE (post)} | 0.12 | 0.33 | 0.00 | 0.00 | 1.00 | 473 $\mathbbm{1}${Premium PIPE (at)} | 0.50 | 0.50 | 0.00 | 1.00 | 1.00 | 473 $\mathbbm{1}${Non-Premium PIPE (at)} | 0.65 | 0.48 | 0.00 | 1.00 | 1.00 | 473 No. of PIPE (at) | 10.76 | 11.83 | 0.00 | 8.00 | 26.00 | 472 No. of Premium PIPE (at) | 3.02 | 4.36 | 0.00 | 1.00 | 10.00 | 472 No. of Non-Premium PIPE (at) | 7.75 | 8.64 | 0.00 | 6.00 | 18.00 | 472 PIPE Investment/IPO Proceeds | | | | | | PIPE (pre) (%) | 2.93 | 11.22 | 0.00 | 0.00 | 0.00 | 473 Premium PIPE (at) (%) | 17.84 | 32.85 | 0.00 | 0.00 | 52.39 | 473 Non-Premium PIPE (at) (%) | 41.94 | 58.64 | 0.00 | 28.57 | 101.16 | 473 PIPE (post) (%) | 7.45 | 30.39 | 0.00 | 0.00 | 22.68 | 473 MLOT (mm USD) | | | | | | Public Investors | 113.22 | 433.26 | -3.76 | 11.24 | 226.41 | 433 Sponsors | 126.93 | 324.28 | 15.66 | 64.32 | 195.48 | 425 PIPE | 201.22 | 1,200.54 | -25.60 | 3.75 | 248.17 | 358 PIPE (pre) | 29.37 | 120.76 | -10.52 | 3.17 | 35.70 | 45 PIPE (at) | 193.07 | 1,135.58 | -25.36 | 4.77 | 248.17 | 303 PIPE (post)∗ | 8.07 | 42.62 | -15.74 | 0.25 | 31.76 | 56 Premium PIPE (at) | 76.63 | 374.15 | -10.22 | 1.68 | 129.46 | 231 Non-Premium PIPE (at) | 134.65 | 838.67 | -17.73 | 3.40 | 138.48 | 303 MLOT/Investment (%) | | | | | | Public Investors | 182.10 | 898.42 | -7.89 | 24.87 | 179.08 | 433 Sponsors | 1,400.46 | 2,570.29 | 370.04 | 861.15 | 2,220.70 | 324 PIPE | 91.26 | 472.71 | -16.50 | 4.47 | 85.80 | 358 PIPE (pre) | 70.33 | 289.37 | -8.05 | 4.36 | 91.80 | 45 PIPE (at) | 63.74 | 307.47 | -16.00 | 4.20 | 80.90 | 303 PIPE (post)∗ | 70.74 | 285.16 | -36.02 | 0.80 | 110.60 | 56 Premium PIPE (at) | 64.05 | 291.75 | -13.00 | 5.30 | 94.90 | 225 Non-Premium PIPE (at) | 64.10 | 308.46 | -16.00 | 4.20 | 80.90 | 301 Note: The sample includes 1,072 SPACs that went through the IPO process between January 2010 and December 2021. All variables are defined in Table B1. $(^{*})$ We exclude a PIPE (post) investor named _Quinpario Acquisition Corporation 2_ , with unusually large returns. Here Sharespu is the total number of outstanding shares held by public investors on the first-listing day of the post-merger company. Thus, Sharespu includes non-redeemed public shares, and shares are automatically converted from rights included in the SPAC unit. Warrantspu is the total number of warrants held by public investors on the first-listing day of the post-merger company. Similarly, $\text{P}_{share}$ and $\text{P}_{warrant}$ denote the price for shares and warrants of the post-merger company at the end of the first trading day. For the sponsor’s MLOT, we can replace the IPO proceeds that were not redeemed in the above definition of $\text{MLOT}^{\mathfrak{p}}$ with risk capital, i.e., $\text{MLOT}^{\mathfrak{s}}=\text{Shares}^{\mathfrak{s}}\times\text{P}_{share}+\text{Warrants}^{\mathfrak{s}}\times\text{P}_{warrant}-\text{Risk Capital},$ where Sharess measures the total number of shares held by the sponsor on the first-listing day of the post-merger company. Thus Sharess includes the founder shares and private placement shares purchased using the sponsor’s “risk capital” and excludes shares forfeited by the sponsor during the de-SPAC process. Warrantss measures the private placement warrants purchased by the sponsor and excludes warrants forfeited by the sponsor. For a PIPE investor’s MLOT, a similar definition applies. However, instead of risk capital, we subtract the issuance amount, i.e., $\text{MLOT}^{\mathfrak{pi}}=\text{Shares}^{\mathfrak{pi}}\times\text{P}_{share}+\text{Warrants}^{\mathfrak{pi}}\times\text{P}_{warrant}-\text{Issuance Amount}^{\mathfrak{pi}},$ where Sharespi, Warrantspi are the number of shares and warrants purchased by PIPE investors, and Issuance Amountpi is the dollar amount of the issuance. We calculate MLOTpi for all three types of (pre, at, and post) PIPE transactions. Table 2 highlights the MLOT for various investor types. Investing in a SPAC generates an average of $113.22 million for public investors, $126.93 million for sponsors, and $201.22 million for PIPE investors. Out of the 45 SPAC deals in which they participate, PIPE (pre) investors make $29.37 million on average. From the 303 SPAC deals in which they take part, PIPE (at) investors earn $193.07 million per deal. PIPE (post) investors participate in 56 deals and earn $8.07 million per deal on average. On average, a SPAC deal generates 182.1% return for public investors and 1,400% for the sponsors. PIPE investors get an average return of 91.26%, where PIPE (pre) investors earn 70.33%, PIPE (at) investors earn 63.74%, and PIPE (post) investors earn 70.74%. Premium PIPE Investors. In total, there are 2,005 PIPE investors in our sample. Out of them, 76 invested in PIPE (pre) transaction, 249 in PIPE (post), and 1,826 invested in at least one PIPE (at) transaction, but only a few large institutional investors participated frequently and invested significantly more than others. To capture this distinction, we use k-means clustering on the total investments in all SPACs and the number of SPACs to classify PIPE investors into two groups. We refer to a subset of PIPE investors who frequently participate as “premium” PIPE investors and the rest as non-premium PIPE investors. There are 34 premium PIPE investors in our sample (Table 3).151515As a way to corroborate this classification, we counted the number of articles written about each of these 34 premium investors in _The New York Times_ and found that these premium investors have the highest counts. For example, excluding research universities, e.g., Columbia University, and non-financial companies, e.g., General Motors and Microsoft Inc., top ten investors include Morgan Stanley, Fidelity, USB, Wells Fargo, Blackstone, Citadel, Bain Capital, BlackRock, T. Rowe Price, and SoftBank. Table 3: Premium PIPE Investors Investor | #SPAC | Volume (million USD) ---|---|--- Fidelity Management & Research Company | 69 | 4,342 BlackRock, Inc. | 57 | 2,317 Capital Research and Management Company | 29 | 2,231 Alyeska Investment Group L.P. | 106 | 1,704 Morgan Stanley | 14 | 1,238 Millennium Management, LLC | 77 | 1,059 Luxor Capital Group | 52 | 972 Blackstone Inc. (f/k/a The Blackstone Group Inc.) | 34 | 965 Koch Industries Inc. | 21 | 926 Third Point, LLC | 12 | 910 Hedosophia Services Limited | 23 | 910 T. Rowe Price Group, Inc. | 27 | 899 SoftBank Group Corp. | 16 | 795 Franklin Resources, Inc. (d/b/a Franklin Templeton Investments) | 24 | 741 Senator Investment Group LP | 36 | 713 Moore Capital Management, Inc. | 66 | 629 Neuberger Berman Group LLC | 20 | 626 Baron Capital Group, Inc. | 25 | 621 Citadel LLC | 51 | 604 Hudson Bay Capital Management LP | 56 | 577 Suvretta Capital Management, LLC | 31 | 576 UBS AG | 57 | 494 Wellington Management Company, LLP | 26 | 437 Heights Capital Management, Inc. | 65 | 431 Magnetar Capital LLC | 39 | 424 D.E. Shaw & Co., L.P. | 38 | 411 Monashee Investment Management LLC | 61 | 408 Ghisallo Capital Management LLC | 62 | 389 Jane Street Group, LLC | 49 | 368 Park West Asset Management, LLC | 27 | 342 Kepos Capital LP | 51 | 286 Linden Advisors LP | 43 | 282 Schonfeld Strategic Advisors LLC | 32 | 224 BlueCrest Capital Management Ltd. | 31 | 181 Note: This table lists the name of premium PIPE investors identified using k-means clustering based on the number of SPAC deals and total investment volumes using data on SPACs that went through an IPO from January 2010 to December 2021. Table 2 provides summary statistics for premium and non-premium PIPE investors. A typical SPAC deal includes 10 PIPE (at) investors, with around three premium and seven non-premium investors. In the 231 SPAC deals in which they participate, premium investors earn an average of $76.63 million (MLOT) with an average return of 64.05%. Non-premium investors participate in 303 SPACs and make an average of $134.65 million (MLOT) and earn an average return of 64.10%. As the standard deviations suggest, the high average returns for the non-premium PIPEs are driven by five outlier SPACs with extreme returns. We revisit this difference in premium and non-premium investors’ MLOTs in Section 6. ## 5 Empirical Results In this section, we investigate whether the data are consistent with the hypotheses stated in Section 3. We begin measuring underpricing at the SPAC IPO level and compare it with the traditional IPO underpricing. Then, we show that the premium PIPE investors who participate in business combinations generate value-relevant information, whereas non-premium PIPE investors engage in a quid pro quo relationship. We end by quantifying the benefit of this quid pro quo relationship in terms of lowering the liquidation risk of future SPACs. ### 5.1 Underpricing #### 5.1.1 Underpricing during SPAC IPO We first examine underpricing during the SPAC IPO. Table 2 reports the average equal-weighted first-day return of SPACs is 1.3%. Although our data start from 2010, most SPAC IPOs occurred between 2020 and 2021. In 2020, 248 SPACs went public, with an average equal-weighted first-day return of 1.3%; in 2021, 613 SPACs went public, and their average equal-weighted first-day return was 1.51%. Although these measures are statistically different from zero (p-values $\approx 0$ in both cases), they are small.161616For comparison, the average equal-weighted first-day return for traditional IPOs was 41.6% in 2020 and 32% in 2021. For more on the IPO underpricing, see https://site.warrington.ufl.edu/ritter/files/IPOs-Underpricing.pdf, accessed 09/22/2022. In addition, SPACs almost always trade at $10 per share until the target announcement date. This uniformity is inconsistent with any information about sponsor quality having been revealed during the SPAC IPO. #### 5.1.2 Underpricing during de-SPAC This section focuses on the underpricing associated with the de-SPAC stage. We say that a private company’s shares are underpriced if the first-day trading value of the post-merger company’s securities held by all SPAC investors (including sponsors, PIPEs, and original SPAC investors who did not redeem) exceeds the proceeds invested in the private firm. There are at least two sources of asymmetric information at this stage. First, the target company may be better informed about its business prospect than anyone else (Smith Jr, 1986; Booth and Smith II, 1986). Second, the PIPE investors may also have better information about factors affecting the target firm than the public investors. In response to PIPE investors’ information, sponsors may underprice the securities issued to them as an incentive payment for their value-relevant information. We focus on the second type of information. From Table 2, we see that, on average, a private company that uses the SPAC route to access the public market leaves $126.93 million MLOT to the sponsors and $201.22 million to the PIPE investors. Furthermore, we also see that PIPE investors who participate in different stages of SPAC are compensated differently. During the SPAC IPO, PIPE (pre) investors earn an average of $29.37 million through forward-purchase agreements, but PIPE (at) investors get, on average, a total of $193.07 million per deal. After the target is disclosed, PIPE (post) investors who participate receive an additional $8.07 million MLOT. Even among PIPE (at) investors, approximately one-third of the MLOT accrues to 34 premium investors. This difference suggests that the two types of investors have different business models, which we return to later in Section 6. Overall, the SPAC shares offered to PIPE (at) investors, especially the premium PIPE investors, during the de-SPAC phase are underpriced. These results are consistent with Hypothesis 1 that PIPE (at) investors participating in the de-SPAC stage are privately informed about the quality of the target company. ### 5.2 Information Production #### 5.2.1 Redemption Rates To test the first part of Hypothesis 1-1, we use the following regression: $\displaystyle\text{Redemption rate}_{i}$ $\displaystyle=$ $\displaystyle\alpha_{0}+\alpha_{1}\times\%\text{Premium PIPE (at)}_{i}+\alpha_{2}\times\%\text{Non-Premium PIPE (at)}_{i}$ (1) $\displaystyle+\alpha_{3}\times\%\text{PIPE (pre)}_{i}+\alpha_{4}\times\%\text{PIPE (post)}_{i}+{\bf X}^{\top}\alpha_{5}+\varepsilon_{i},$ where $\text{Non-Premium PIPE (at)}_{i}$ denotes the investment allocated to premium PIPE (at) investors as a fraction of the total IPO proceeds in SPAC $i$, similarly for other allocations, and ${\bf X}$ is a vector of covariates including total IPO proceeds and fixed effects that may vary across specifications. If premium investors generate value-relevant information, the higher the share allocated to premium investors lower the redemption rate, i.e., $\alpha_{1}<0$. Table 4 provides evidence consistent with the hypothesis. In column (1), we include only the allocations to four types of PIPE investors. In column (2), we control for the size of the SPAC IPO and the liquidation risk of the SPAC when the target is announced using Days Remaining until the expiration. In column (3), we control for the quarter-fixed effects. In column (4), which is also our preferred specification, we additionally control for the target company’s sector fixed effects to alleviate the problem that investors may have varying investment preferences across industries. In all these specifications, we cluster the robust standard errors at the sponsor level to capture correlations within a sponsor but across investors and SPACs. We find that the allocations to premium PIPE (at) investors negatively correlate with the redemption rates, and this result is robust across all these specifications. The estimates from Table 4 column (4) suggest that one standard deviation (32.85 pps) increase in premium PIPE investors’ participation, is associated with a 7.03 pps decrease in the redemption rate, which equals 12.4% of the sample mean. Furthermore, we find no association between non-premium PIPE (at) or PIPE (pre) allocations and redemption rates, as in Hypothesis and Hypothesis 2-1. These results are consistent with the hypothesis that premium PIPE (at) investors are better informed about the deal quality than either the non- premium PIPE (at) investors or PIPE (pre). Table 4: Redemption Rates | (1) | (2) | (3) | (4) ---|---|---|---|--- PIPE (pre) (%) | 0.125 | 0.192 | 0.148 | 0.156* | (0.123) | (0.119) | (0.116) | (0.095) Premium PIPE (at) (%) | -0.419*** | -0.380*** | -0.366*** | -0.214*** | (0.069) | (0.064) | (0.059) | (0.050) Non-Premium PIPE (at) (%) | -0.013 | -0.009 | -0.002 | -0.037 | (0.034) | (0.035) | (0.034) | (0.030) PIPE (post) (%) | 0.120*** | 0.103*** | 0.114*** | 0.148*** | (0.035) | (0.031) | (0.033) | (0.035) Log IPO proceeds | ✗ | ✓ | ✓ | ✓ Days Remaining | ✗ | ✓ | ✓ | ✓ IPO Month FEs | ✗ | ✗ | ✓ | ✓ Target Sector FEs | ✗ | ✗ | ✗ | ✓ Adj. $R^{2}$ | 0.161 | 0.182 | 0.183 | 0.400 Obs. | 444 | 444 | 444 | 430 Note: This table shows the results from estimating (1) using SPACs that went through an IPO from January 2010 to December 2021. The independent variable is the investment amount as a percentage of the SPAC’s IPO proceeds for all PIPE (pre) investors, premium PIPE (at) investors, non-premium PIPE (at) investors, and all PIPE (post) investors. Premium PIPE (at) investors are listed in Table 3. All other variables are defined in Table B1. All unbounded continuous variables are winsorized at their $1^{st}$ and $99^{th}$ percentile values. Robust standard errors are clustered at the sponsor level and are reported in parentheses. ${}^{*},^{**}$, and ∗∗∗ denote p-values less than 0.1, 0.05, and 0.01, respectively. The negative association between premium PIPE (at) allocation and redemption rates is consistent with premium investors producing information. Suppose not, and instead, suppose that sponsors underprice and distribute them to the premium investors in exchange for a promise to help them in the future on an unrelated SPAC. In that case, because the timing of the investors should be immaterial, we should also expect the allocations to PIPE (post) investors to correlate negatively with the redemption rate. However, as we see from Table 4, the two are positively correlated, further strengthening our finding. Lastly, we note that almost all (231 out of 25) PIPE (post) investors are non- premium investors. In fact, among the top 20 PIPE (post) investors by investment size, only Luxor Capital Group is a premium investor. Thus, it is reasonable to assume that sponsors include premium investors at the time of the announcement as certification and use non-premium investors later as needed to backfill the trust fund. #### 5.2.2 Announcement-day Return To test the second part of Hypothesis 1-1 about the announcement-day return, we use the following regression: $\displaystyle\text{Announcement-day return}_{i}$ $\displaystyle=$ $\displaystyle\beta_{0}+\beta_{1}\times\%\text{Premium PIPE (at)}_{i}$ (2) $\displaystyle+\beta_{2}\times\%\text{Non-Premium PIPE (at)}_{i}$ $\displaystyle+\beta_{3}\times\%\text{PIPE (pre)}_{i}+\varpi_{i}.\qquad$ The estimation results are shown in Table 5. There is a strong positive association between announcement-day returns (%) and allocations to premium PIPE (at) investors. In particular, the estimates in Table 5 column (4) suggest that one standard deviation (32.85 pps) increase in premium PIPE (at) investors’ participation is associated with a 4.07 pps increase in the SPAC’s announcement-day return, which is 108.8% of the sample mean of the announcement-day returns. Thus the effect of premium investors on announcement-day return is economically substantial and statistically significant. Furthermore, the announcement-day returns are uncorrelated with the allocations to non-premium PIPE (at) investors. Again, this finding is consistent with the hypothesis that larger investment by premium PIPE (at) investors acts as a certification of the quality of the deal. In contrast, investment by non-premium PIPE (at) is non-informative. We also observe a negative correlation between announcement-day returns and allocations to PIPE (pre) investors. Specifically, the estimates in column (4) suggest that one standard deviation (11.22 pps) increase in PIPE (pre) investors’ participation is associated with 0.81 pps lower in announcement-day return, which equals 21.7% of the sample mean. Table 5: Announcement-day Return | (1) | (2) | (3) | (4) ---|---|---|---|--- PIPE (pre) (%) | -0.118*** | -0.119*** | -0.099*** | -0.072 | (0.042) | (0.042) | (0.037) | (0.044) Premium PIPE (at) (%) | 0.148*** | 0.140*** | 0.136*** | 0.124*** | (0.041) | (0.041) | (0.038) | (0.040) Non-Premium PIPE (at) (%) | -0.003 | -0.007 | -0.010 | -0.002 | (0.014) | (0.014) | (0.013) | (0.015) Log IPO proceeds | ✗ | ✓ | ✓ | ✓ Days Remaining | ✗ | ✓ | ✓ | ✓ IPO Month FEs | ✗ | ✗ | ✓ | ✓ Target Sector FEs | ✗ | ✗ | ✗ | ✓ Adj. $R^{2}$ | 0.099 | 0.104 | 0.098 | 0.095 Obs. | 457 | 457 | 431 | 417 Note: This table shows the results from estimating (2) using SPACs that went through an IPO from January 2010 to December 2021. The key independent variable is the investment amount as a percentage of the SPAC’s IPO proceeds for all PIPE (pre) investors, premium PIPE (at) investors, non-premium PIPE (at) investors, and all PIPE (post) investors. Premium PIPE (at) investors are listed in Table 3. All other variables are defined in Table B1. All unbounded continuous variables are winsorized at their $1^{st}$ and $99^{th}$ percentile values. Robust standard errors are clustered at the sponsor level and are reported in parentheses. ${}^{*},^{**}$, and ∗∗∗ denote p-values less than 0.1, 0.05, and 0.01, respectively. Likewise, the hypothesis predicts a non-positive market reaction to the announcement of PIPE (post) investments. To assess this claim, we use an event study approach. In particular, we use different asset-pricing factor models to estimate the cumulative abnormal returns (CAR) (Campbell, Lo and MacKinlay, 1997) associated with announcements about PIPE investors in the process of announcing the target. Specifically, we estimate the following linear regression model: $\displaystyle r_{i,t}=\kappa_{i}+\bm{\rho}_{i}\texttt{W}_{t}+\upsilon_{i,t},$ (3) where $r_{i,t}$ is the share return of SPAC $i$, at date $t$, X includes portfolio returns of different factor models. For each SPAC, we estimate (3) using an event window of $[t-\kappa,t+\kappa]$, which is $\kappa$ days around each PIPE (post) transaction’s announcement date. The estimation period is chosen to start from the SPAC’s unit separation date and ends before the event window. Then we obtain the predicted return $\hat{r}_{i,t}$, which gives us the “abnormal” return as $\hat{\upsilon}_{i,t}\equiv r_{s,t}-\hat{r}_{i,t}$. The CAR associated with target announcement by $i$ is given by accumulating $\hat{\upsilon}_{i,t}$ during the event window, i.e., $\savestack{\tmpbox}{\stretchto{\scaleto{\scalerel*[\widthof{\texttt{CAR}}]{\kern-0.6pt\bigwedge\kern-0.6pt}{\rule[-505.89pt]{4.30554pt}{505.89pt}}}{}}{0.5ex}}\stackon[1pt]{\texttt{CAR}}{\tmpbox}_{i,t}=\sum_{t=t_{i}-\kappa}^{t_{i}+\kappa}\hat{\upsilon}_{i,t}.$ Table 6 reports the average of the $\savestack{\tmpbox}{\stretchto{\scaleto{\scalerel*[\widthof{\texttt{CAR}}]{\kern-0.6pt\bigwedge\kern-0.6pt}{\rule[-505.89pt]{4.30554pt}{505.89pt}}}{}}{0.5ex}}\stackon[1pt]{\texttt{CAR}}{\tmpbox}$ and their respective standard deviation across all SPACs using a three-day, five-day, and 10-day window around the event date. For each event window, we estimate CAPM, the Fama-French three-factor model, and Fama-French five-factor model. In all cases, the average $\savestack{\tmpbox}{\stretchto{\scaleto{\scalerel*[\widthof{\texttt{CAR}}]{\kern-0.6pt\bigwedge\kern-0.6pt}{\rule[-505.89pt]{4.30554pt}{505.89pt}}}{}}{0.5ex}}\stackon[1pt]{\texttt{CAR}}{\tmpbox}$ estimates are not statistically different from zero. This result suggests that the investors do not respond to the announcement of PIPE (post), likely because almost all of them are non-premium and unlikely to add information. Table 6: Cumulative (average) Abnormal Returns Event Window | CAPM | FF3 | FF5 ---|---|---|--- [-3, 3] | 0.71 | 0.74 | 0.75 | (0.80) | (0.79) | (0.78) [-5, 5] | -1.04 | -1.04 | -1.06 | (1.14) | (1.13) | (1.12) [-10, 10] | -2.37 | -2.30 | -2.28 | (1.92) | (1.92) | (1.91) Note: This table reports the SPAC share’s cumulative average abnormal return in percentage for three different event windows around announcements of 81 PIPE (post) transactions. Here, CAPM stands for capital asset pricing model, and FF3 and FF5 stand for Fama-French three- and five-factor models, respectively. The cross-sectional standard deviations of the cumulative abnormal returns are reported in parentheses. ${}^{*},^{**}$, and ∗∗∗ denote p-values less than 0.1, 0.05, and 0.01, respectively. ### 5.3 Quid Pro Quo Next, we test Hypothesis 2 that there is a quid pro quo arrangement between sponsors and non-premium PIPE (at) investors and that it helps a weaker firm go public. Recall that under favoritism, the current MLOT should be negatively correlated with the past MLOT. In particular, if the sponsor allocates securities to non-premium investors as a favor, then if the MLOT in the past is high, the sponsor is likely to ask this investor to finance a weak deal today. We verify that the probability that a sponsor allocates securities to a “relationship” investor increases if the investor has higher MLOT from the past deals _and_ the SPAC deal is cold. Lastly, we show that quid pro quo lowers the liquidation risks. #### 5.3.1 Mean-reverting MLOT As a first step, we assess if the MLOT exhibits mean reversion for the non- premium PIPE (at) investors, i.e., MLOT is stable around a mean. In contrast, under the information production hypothesis, the correlation between the current MLOT and past MLOT should be zero if the information production across deals is independent and, as (Benveniste et al., 2003) show, they can be positively correlated if there is information spillover. A necessary condition to test favoritism is the presence of repeated deals between a sponsor and a PIPE investor. In Table 7, we list the top 20 sponsors in our sample in terms of the total number of initialized and completed SPAC deals. In our sample, 139 sponsors have initialized more than one SPAC, totaling 439 SPACs. Furthermore, non-premium PIPE investors are more likely to participate in repeat deals (with the same sponsor) than premium investors. To gauge the difference, consider the following. On average, the probability that a non-premium investor from the previous deal participates in the current deal is 17%. As there are around eight non-premium investors in a SPAC, one out of these eight non-premium investors in the current deal is from the previous deal. If these investors were randomly (uniformly) participating in a SPAC deal, the probability would be much smaller because there are 1,792 non- premium investors in our sample. If non-premium investors participate randomly in a SPAC, then the probability that at least one non-premium investor participated in the previous deal is negligible. Table 7: Top 20 Sponsors Initialized | Completed ---|--- Sponsor | #SPAC | Sponsor | #SPAC The Gores Group | 13 | The Gores Group | 9 Cohen & Company | 11 | Chardan | 7 Social Capital | 10 | Social Capital | 6 Hennessy Capital Group | 8 | Cohen & Company | 6 Cantor Fitzgerald | 8 | TPG | 5 TPG | 8 | Cantor Fitzgerald | 5 Michael Klein | 8 | Riverstone Investment Group | 5 Jonathan J. Ledecky | 7 | Hennessy Capital Group | 5 Chardan | 7 | Harry L. You | 5 Hedosophia | 6 | Eagle Equity Partners | 5 Harry L. You | 6 | Michael Klein | 4 Riverstone Investment Group | 6 | Jonathan S. Huberman | 4 Fortress Investment Group | 6 | Hedosophia | 4 GigCapital Global | 6 | GigCapital Global | 4 Apollo Global Management | 6 | Chinh Chu | 4 Bill Foley | 6 | Niccolo de Masi | 4 Chinh Chu | 5 | Bill Foley | 4 Jaws Estates Capital | 5 | Apollo Global Management | 4 Craig-Hallum Capital Group | 5 | Union Acquisition Group | 3 Niccolo de Masi | 5 | Casdin Capital | 3 Note: This table presents the top 20 sponsors in terms of the total number of SPACs initialized and completed. The probability gap reverses for the premium investors. The probability that a premium investor from the previous deal participates in the current deal is 36.60%. There are, on average, three premium investors in a SPAC, so a sponsor includes one premium investor from the previous deal. There are 34 premium investors in our sample, so if they participate randomly in a SPAC, then the probability that at least one premium investor in the current deal has participated in the previous deal by the same sponsor is 90.31%. Let $\mu$ be the average MLOT for all PIPE (at) investors in our sample. Table 8 is the summary statistics. As we see, on average, 28% of SPACs have at least one premium PIPE investor. Then we estimate the following linear model, $\displaystyle\text{MLOT}_{p,i,s}$ $\displaystyle=$ $\displaystyle\theta_{0}+\theta_{1}\times(\text{lag-MLOT}_{p,i,s}-\mu)+{\bf X}^{\top}\theta_{2}+\eta_{p,i,s},$ (4) where MLOTp,i,s is the MLOT that PIPE (at) investor $p$ gets in the current SPAC deal $i$ initialized by the sponsor $s$, and $\text{lag-MLOT}_{p,i,s}$ is the MLOT from the most recent deal in the past. As before, ${\bf X}$ includes several covariates, including fixed effects (see Table 9). Table 8: Summary Statistics of PIPE Investors | Mean | Std.Dev. | P10 | P50 | P90 | Obs. ---|---|---|---|---|---|--- At Least 1 Premium | 0.28 | 0.45 | 0.00 | 0.00 | 1.00 | 5,812 All PIPE Investors | | | | | | Investment Amount ($mm) | 16.67 | 37.62 | 1.00 | 7.00 | 35.00 | 5,535 MLOT ($mm) | 9.76 | 75.33 | -0.98 | 0.28 | 13.05 | 4,816 No. of Past SPACs | 0.13 | 0.47 | 0.00 | 0.00 | 0.00 | 5,812 MLOT (past) | 2.48 | 23.66 | 0.00 | 0.00 | 0.00 | 5,812 Premium PIPE Investors | | | | | | Investment Amount ($mm) | 20.50 | 38.79 | 3.00 | 10.00 | 45.00 | 1,609 MLOT ($mm) | 11.01 | 45.60 | -1.17 | 0.58 | 23.97 | 1,462 No. of Past SPACs | 0.20 | 0.53 | 0.00 | 0.00 | 1.00 | 1,647 MLOT (past) | 3.98 | 19.82 | 0.00 | 0.00 | 2.55 | 1,647 Non-premium PIPE Investors | | | | | | Investment Amount ($mm) | 15.10 | 37.01 | 0.75 | 5.00 | 30.00 | 3,926 MLOT ($mm) | 9.22 | 85.10 | -0.96 | 0.17 | 10.37 | 3,354 No. of Past SPACs | 0.10 | 0.43 | 0.00 | 0.00 | 0.00 | 4,165 MLOT(past) | 1.88 | 25.00 | 0.00 | 0.00 | 0.00 | 4,165 Note: This table contains summary statistics for SPACs that went through an IPO from January 2010 to December 2021. Premium PIPE investors are identified using k-means clustering and listed in Table 3. All variables are defined in Table B1. Under information production, the deviation of lagged MLOT from its mean is either uncorrelated or positively correlated with the current MLOT; hence $\theta_{1}\geq 0$ in (4). Under quid pro quo arrangements, however, the MLOT should increase if the past deal was lower than the average and vice versa; hence $\theta_{1}<0$. The results are in column (1) of Table 9. As expected under favoritism, MLOT exhibits mean reversion. Table 9: Mean-Reverting MLOT | MLOT ---|--- | (1) | (2) | (3) All PIPE Investors | -0.099* | | | (0.052) | | Premium PIPE Investors | | -0.025 | 0.235** | | (0.062) | (0.109) Non-premium PIPE Investors | | -0.218*** | -2.729*** | | (0.076) | (0.756) Instrumental Variables | ✗ | ✗ | ✓ F-stat (first-stage) | | | 92.82 Adjusted $R^{2}$ | 0.418 | 0.419 | Observations | 3,790 | 3,790 | 3,790 | | Mean | Std.Dev. $\text{Non-Premium}\times$(lag-MLOT-$\mu$) | | -8.12 | 5.41 (lag-MLOT-$\mu$)$\times$(lag2-MLOT-$\mu$) | | 121.96 | 85.54 (lag2-MLOT-$\mu$)2 | | 135.78 | 5.77 (lag-MLOT-$\mu$)2 | | 166.93 | 297.54 Note: This table shows results from estimating (5) using SPACs that went through IPO between January 2010 to December 2021. Each observation is at (SPAC, sponsor, PIPEat investor) level. The dependent variable, MLOT, measures the money left on the table the PIPEat investor makes from the SPAC deal. The key independent variables are the MLOT earned with the same sponsor in the last deal for all PIPEat investors, premium PIPEat investors, and non-premium PIPEat investors. Additional controls include the log of IPO proceeds, the number of PIPE investors, the number of premium PIPE investors, days remaining, and separate fixed effects for the month, sponsor, and PIPE. All continuous variables are winsorized at the $1^{st}$ and $99^{th}$ percentile values. All other variables are defined in Table B1. Robust standard errors are reported in parentheses. ${}^{*},^{**}$, and ∗∗∗ denote p-values less than 0.1, 0.05, and 0.01, respectively. Next, we differentiate between premium and non-premium investors and estimate $\displaystyle\text{MLOT}_{p,i,s}$ $\displaystyle=$ $\displaystyle\delta_{0}+\delta_{1}\times(\text{lag-MLOT}_{p,i,s}-\mu)$ (5) $\displaystyle+\delta_{2}\times\text{Non-Premium}_{p}\times(\text{lag- MLOT}_{p,i,s}-\mu)+{\bf X}^{\top}\delta_{3}+\tilde{\eta}_{p,i,s},$ where $\text{Non-Premium}_{p}$ is a binary variable equal to one if the PIPE (at) investor $p$ is a non-premium investor. The results from the estimation are in column (2) of Table 9. Comparing the second and third rows, we find that MLOT exhibits mean reversion for non-premium PIPE investors but not for premium investors. This result is consistent with the previous results that only the premium PIPE investors generate information, and Hypothesis 1-2. A potential concern with the estimate of $\delta_{2}$ in column (2) of Table 9 is that sponsors’ decisions to include non-premium investors may be correlated with unobserved deal quality. Indeed, favoritism implies that the sponsor is more likely to involve non-premium investor(s) for a weak deal. If so, the sample used to identify the parameter is selected and not representative of the population. Because the sponsor is likely to allocate higher MLOT to non- premium investors in weak deals, the above estimate can be biased upwards. We use the instrumental variable (IV) approach to address this problem. The idea behind our choice of IV is based on the following observation: suppose $v=v_{1}\times v_{2}$ is endogenous because $v_{1}$ is endogenous, then $v_{2}$ can be an IV for $v$ under appropriate additional assumptions. For a recent application of this intuition, see Fabra and Reguant (2014) and Aryal, Ciliberto and Leyden (2022). In our setting, $v$ is $\text{Non- Premium}\times(\text{lag-MLOT}-\mu)$ and $v_{1}$ is the Non-Premium dummy, and given the specification in (5) $(\text{lag-MLOT}-\mu)$ is uncorrelated with the quality of the current deal. However, $(\text{lag-MLOT}-\mu)$ may not satisfy the exclusion restriction because it directly affects the outcome variable. Based on the intuition above, we propose to use mean deviation from the second-to-last deal, i.e., $(\text{lag${}^{2}$-MLOT}-\mu)$ as an excluded IV. Using lagged variables as IVs is widely used in the dynamic panel data under the Arellano-Bond estimator. Furthermore, to capture the nonlinearity of the interaction terms, we include $\\{(\text{lag- MLOT}-\mu)\times(\text{lag${}^{2}$-MLOT}-\mu),(\text{lag${}^{2}$-MLOT}-\mu)^{2},(\text{lag- MLOT}-\mu)^{2}\\}$ as additional excluded variables. The results from this estimation are in column (3) of Table 9. The first-stage Cragg-Donald Wald F-statistic of 92.82 suggests that our excluded variables are not weak instruments. Comparing these estimates with the ones in column (2), we see that, as expected, the fixed-effect estimate for the non-premium investor is biased upwards. More importantly, and consistent with the quid pro quo hypothesis, we find that one dollar higher mean deviation of MLOT in the last deal leads to a $2.729 reduction in the mean deviation from the current deal and vice versa presumably because now the non-premium investor is “asked” to participate in tepid deals that have smaller underpricing. In contrast, for the premium PIPE investors, the effect is the opposite: if the investor made one more dollar in MLOT above the mean in the last deal, on average, the MLOT in the next deal with the same sponsor increases by 23 cents consistent with the hypothesis that premium PIPE investors generate value-relevant information.171717Some deals have more than one sponsor. In such cases, we treated each sponsor separately. Although we do not show this here, we verify that the results are qualitatively the same if we exclude any SPAC with more than one sponsor or consider only the sponsor with the maximum number of past SPAC deals. #### 5.3.2 Allocations and Past Relationship Next, we test the Hypothesis 2-3. A sponsor builds a relationship with a non- premium PIPE investor by providing MLOT, so when the deal is tepid, it expects the _relationship investors_ with high past MLOT to return the favor by participating in the deal. To this end, we evaluate if a sponsor is more likely to use non-premium PIPE (at) investors if the investor has made larger returns from the sponsor in the past and the current target is cold. Although we do not observe the quality of the target, we propose to proxy it by the difference between the offer share price of the SPAC, i.e., the share price at SPAC IPO, and the share price on the closing day of the merger. We denote this gap as “Cold.” Then, we estimate the probability that a sponsor allocates securities to a non-premium investor as a function of the past MLOT and a measure of cold, among other variables. For a sponsor $s$, let ${\mathcal{R}}_{s}$ denote the set of relationship investors as the set of all non-premium investors that invested in at least one SPAC with sponsor $s$. Let $\text{Participation}_{p,i,s}\in\\{0,1\\}$ be a binary variable equal to one if investor $p\in{\mathcal{R}}_{s}$ is allocated securities from SPAC $i$ started by sponsor $s$, and zero otherwise. Then we estimate the participation probability under the assumption that the probability takes a logit form, i.e., $\displaystyle\Pr(\text{Participation}_{p,i,s}=1\mid\mathbf{Z}_{p,i,s}^{\top}\bm{\delta})$ $\displaystyle=$ $\displaystyle\frac{\exp(\mathbf{Z}_{p,i,s}^{\top}\bm{\delta})}{1+\exp(\mathbf{Z}_{p,i,s}^{\top}\bm{\delta})},$ (6) $\displaystyle\mathbf{Z}_{p,i,s}^{\top}\bm{\delta}$ $\displaystyle=$ $\displaystyle\mathbf{X}^{\top}\bm{\delta_{0}}+\delta_{1}\times\log(1+\text{MLOT(past)}_{p,s})+\delta_{2}\times\text{Cold}_{i}$ $\displaystyle+\delta_{3}\times\log(1+\text{MLOT(past)}_{p,s})\times\text{Cold}_{i},$ where ${\bf X}$ is the vector of controls shown in Table 9; $\log(1+\text{MLOT(past)}_{p,s})$ is the log of the total MLOT $p$ made in past deals with $s$ and measures the past relationship between $s$ and $p$; and, and as we defined above, $\text{Cold}_{i}$ is the measure of the quality of the deal $i$. We estimate (6) using the method of maximum likelihood and present the results in column (1) of Table 10. The estimated coefficient of the interaction term (third row) in column (1) is positive (p-value $<0.01$), supporting the quid pro quo hypothesis that the sponsor first builds a relationship and draws it down as and when needed. Table 10: Sponsor and Non-Premium PIPE Investors | Participation | Allocation ---|---|--- | (1) | (2) | (3) | (4) | (5) $\log(\text{1+MLOT(past)})$ | 0.085 | -0.004 | -0.088 | -0.335∗∗∗ | 0.019 | (0.136) | (0.123) | (0.098) | (0.081) | (0.010) | | | | | [0.013] Cold | 0.006∗∗∗ | 0.004∗∗∗ | -0.003 | 0.003∗∗ | -0.00003 | (0.001) | (0.001) | (0.002) | (0.002) | (.00006) | | | | | [0.0006] $\log(\text{1+MLOT(past)})\times\text{Cold}$ | 0.008∗∗∗ | 0.011∗∗∗ | 0.011∗∗∗ | 0.008∗∗∗ | 0.0003∗∗∗ | (0.002) | (0.003) | (0.003) | (0.002) | (0.0001) | | | | | [0.00012] Outside Option | | -0.017∗∗∗ | -0.007∗∗∗ | -0.002 | | | (0.003) | (0.002) | (0.002) | $\widehat{\mathbb{E}(\text{Participation}\mid{\mathbf{Z}})}$ | | | | | 0.239∗∗∗ | | | | | (0.018) | | | | | [0.024] Log IPO proceeds | ✗ | ✓ | ✓ | ✓ | ✓ Days Remaining | ✗ | ✓ | ✓ | ✓ | ✓ No. Premium Investors | ✗ | ✓ | ✓ | ✓ | ✗ Sponsor Dummies | ✗ | ✗ | ✓ | ✓ | ✗ Month Dummies | ✗ | ✗ | ✗ | ✓ | ✓ Pseudo/Adj. $R^{2}$ | 0.0098 | 0.0644 | 0.2763 | 0.4061 | 0.8114 Obs. | 10,104 | 10,044 | 7,917 | 7,913 | 7,838 Note: Columns (1) to (4) of this table show the MLE of participation (6), and column (5) shows the estimates of the allocation (5.3.2). Each observation is at (SPAC, sponsor, non-premium PIPE investor) level. For each SPAC, non- premium investors include all relationship investors that participate in at least one SPAC deal initialized by the sponsor during the sample period. For the first four columns, the dependent variable, $Participation$, equals one if the non-premium investor participates in the focal SPAC deal. The dependent variable for the last column, $Allocation$, is the investor’s allocation as a fraction of all relationship investors’ allocations. The key independent variables include the non-premium investor’s MLOT earned with this sponsor in the past in log form, $\log(1+\text{MLOT ({past})})$, and a measure of the (reciprocal) of the deal quality, Cold. For the allocation regression, we also include the estimated probability of participation using estimates from column (4) as an additional variable. All continuous variables at winsorized at $99^{th}$ percentile. Robust standard errors are clustered at the sponsor-PIPE pair level and reported in parentheses. For column (5), bootstrapped standard errors (500 replications) are reported in the square brackets. ${}^{*},^{**}$, and ∗∗∗ denote p-values less than 0.1, 0.05, and 0.01, respectively. One possible concern with the estimation is that it ignores the outside option of the sponsor, which may be correlated with the unobserved error. A better outside option lowers the probability of issuing securities to a non-premium investor and vice versa. Ignoring outside options could introduce attenuation bias if cold deals have profitable outside options. To alleviate this concern, we use the highest MLOT among all premium PIPE investors _not included_ in the current deal as an additional control variable to capture the strength of the outside option. We also include the log of IPO proceeds, days remaining, and the number of premium investors as additional covariates. The estimates from this new model are in column (2) of Table 10. A better outside option lowers the probability of selecting the non-premium investor, and our previous finding becomes slightly stronger. The results are robust when we include dummy variables for the sponsor and the month when the deal is announced. In all cases, the coefficient for the interaction term is positive and estimated precisely (p-value $<0.01$). To help interpret these estimates, we evaluate the probability in (6) at a different relationship and tepidness while holding other variables at their average value using the estimates from column (4) of Table 10. We find that investors with a stronger relationship have higher participation rates in deals with lower demand. In SPAC deals with Cold at its $90^{th}$-percentile, non-premium investors who have a strong relationship with the sponsor (one standard deviation above the mean) are 3.54 pps more likely to participate than in SPAC deals with Cold at its $10^{th}$-percentile. Conversely, non- premium investors without a prior relationship with the sponsor are less likely to participate in cold deals. Next, we examine whether, conditional on participation, the sponsor allocates a higher portion of securities to non-premium investors with a stronger relationship when a deal is cold and vice versa. To this end, we determine the allocation of SPAC $i$’s securities from sponsor $s$ to PIPE investor $p$ among all relationship investors ${\mathcal{R}}_{s}$, $\text{Allocation}_{p,i,s}\in[0,1]$, as the ratio of $p$’s shares to the total shares allocated to ${\mathcal{R}}_{s}$. We set the share to zero if a non- premium investor does not participate in a deal. We find that the mean allocation is $2.93$, and the standard deviation of the allocation is $8.77$. Furthermore, the allocation is skewed towards zero, with the median of $0$ and the $99^{th}$-percentile at $42.37$. Given that there are 1,792 non-premium investors, and so the allocation in our sample is skewed to the left with a mass at zero, we estimate the following Tobit model: $\displaystyle\text{Allocation}_{p,i,s}^{*}$ $\displaystyle=$ $\displaystyle\mathbf{W}_{p,i,s}^{\top}\bm{\gamma^{*}}+u_{s},$ $\displaystyle\text{Allocation}_{p,i,s}$ $\displaystyle=$ $\displaystyle\min\\{\max\\{0,\mathbf{W}_{p,i,s}^{\top}\bm{\gamma}+\nu_{1}\\},1\\}$ (7) $\displaystyle\mathbf{W}_{p,i,s}^{\top}\bm{\gamma}$ $\displaystyle=$ $\displaystyle\mathbf{X}^{\top}\bm{\gamma_{0}}+\gamma_{1}\times\log(1+\text{MLOT(past)}_{p,s})+\gamma_{2}\times\text{Cold}_{i}$ $\displaystyle+\gamma_{3}\times\log(\text{MLOT(past)}_{p,s}+1)\times\text{Cold}_{i}+\gamma_{4}\times\widehat{\mathbb{E}(\text{Participation}|{\mathbf{Z}})}_{p,i,s},$ where $(\mathbf{W},\text{Allocation}_{p,i,s})$ is always observed but the latent variable $\text{Allocation}_{p,i,s}^{*}$ is observed only when $\text{Allocation}_{p,i,s}\in[0,1]$. We further assume that $\mathbf{W}$ is independent of $(u_{1},]\nu_{1})$ and that $\nu_{1}$ is a mean-zero Normal random variable with unknown variabce and $\mathbb{E}(u_{1}\mid\nu_{1})=\rho\nu_{1}$. Furthermore, $\widehat{\mathbb{E}(\text{Participation}\mid{\mathbf{Z}})}_{p,i,s}$ is the estimated participation probability of allocation given ${\bf Z}$ and is determined by the estimates in column (4) of Table 10. Thus, the outside option serves as our excluded variable, with the underlying exclusion restriction being that the outside option should not directly affect a non- premium investor’s allocation. In Column (5) of Table 10, the estimates indicate a positive association between a higher probability of participation and an increased allocation of shares. Moreover, non-premium investors who have previously profited from deals with the sponsor receive a larger share of allocation than their counterparts when a deal is considered cold. In economic terms, a one standard deviation increase in the interaction term leads to a 5.8% increase in allocation relative to its standard deviation, amounting to 14.80% of the sample mean. In our sample, an average non-premium investor is allocated 2.68% of the total non-premium investors’ allocations. Furthermore, as the non-premium investor’s MLOT (past) times the deal’s Cold measure increase together by one standard deviation, the allocation fraction rises by 0.40 pps to 3.08%. In terms of dollar amount, the allocation increases from $0.40 million to $0.46 million. These findings support the hypothesis that non-premium investors and sponsors engage in quid pro quo agreements. #### 5.3.3 Relationship Capital and Liquidation Risk We have shown that the relationship investors support the sponsor by participating in a tepid deal and purchasing a larger fraction of shares. These non-premium investors received larger MLOTs from the same sponsor during the previous deals. Therefore, it stands to reason that a sponsor builds these relationships with an eye toward potential tepid deals in the future. If so, some of the underpricing driven by quid pro quo could be advantageous to private companies which want to go public. Next, we verify Hypothesis 2-4 by estimating the liquidation risk and liquidation probability of SPACs as a function of the total MLOT paid by the sponsor to non-premium investors. Here, we use the number of days from the SPAC’s IPO until a target is announced or liquidation is announced to denote liquidation risk because a SPAC has two years to find a target, or else it will have to liquidate. That is, the sooner a SPAC can find a deal, the lower the liquidation risk. A sponsor with more relationship capital can afford to take more risk with a potential tepid target because it can draw down the relationship capital to close the deal and move on to a new SPAC If the risk and the probability decrease with the MLOT, then we consider that as evidence of a benefit of quid pro quo. Table 11: Liquidation and Relationship | Liquidation Risk | Liquidation ---|---|--- | (1) | (2) Premium | 0.092 | 0.363 | (0.061) | (0.222) Non-premium | -0.138∗∗ | -0.406∗ | (0.066) | (0.246) Log IPO proceeds | ✓ | ✓ No. Past Successful Deals | ✓ | ✓ Obs. | 679 | 680 Note: This table demonstrates that SPACs initiated by sponsors with a history of stronger relationships with non-premium investors are less likely to liquidate and face reduced liquidation risk when a target is announced or liquidated. Column (1) uses a Poisson regression on the dependent variable _Liquidation Risk_ , which represents the number of days from the SPAC’s IPO until a target or liquidation is announced. Column (2) uses a Logit regression on the dependent variable _Liquidation_ , which is equal to one if the SPAC liquidates and zero if the SPAC successfully completes its merger. The primary independent variable is the total MLOT earned by premium and non-premium investors in previous SPACs initiated by the same sponsor. All continuous variables are winsorized at the $1^{st}$ and $99^{th}$ percentiles. Robust standard errors, clustered at the sponsor level, are reported in parentheses. ${}^{*},^{**}$, and ∗∗∗ denote p-values less than 0.1, 0.05, and 0.01, respectively. We begin with a count (Poisson) regression of the number of days from the SPAC’s IPO until a target is announced or liquidation is announced on the MLOT earned by premium and non-premium investors in previous SPACs initiated by the same sponsor. We also include the log of the IPO proceeds and the number of past successful SPACs by the same sponsor as additional controls. The results from this estimation exercise are in column (1) of Table 11, which suggests that the days it takes for a SPAC to find a target decreases with the MLOT earned by non-premium investors in the past. In particular, a one standard deviation increase in the log of MLOT decreases the number of days to find a target by 64 days, which is 15.17% of the average 422 days it takes to either find a target or liquidate in the sample, holding all other variables constant at their respective average values. However, consistent with the information creation hypothesis, we find that the MLOT earned by premium investors in the past is uncorrelated with the liquidation risk. Next, we consider the probability of liquidation. In our sample, we observe 191 liquidations and 473 successful mergers. We verify the extent to which the difference between these two groups is attributable to the relationship payments the SPAC sponsor made to non-premium investors. Let $\text{Liquidation}_{i}\in\\{0,1\\}$ be a binary variable equal to one if the SPAC $i$ has liquidated and zero otherwise. Under the assumption that we can use the Logit model to specify the conditional distribution of liquidation, we estimate the probability of liquidation and present it in column (2) of Table 11. The estimates suggest that SPACs initialized by sponsors who have built a stronger relationship with non-premium investors are less likely to liquidate. However, the coefficient is imprecisely estimated (p-value $<0.1$). In contrast, but consistent with the information production hypothesis, the MLOT earned by premium PIPE investors from the same sponsor’s deals in the past is uncorrelated with the success of the current deal. In terms of economic magnitude, we find that if sponsors had allocated one standard deviation more MLOT ($300.53 million) to non-premium investors above average, the liquidation probability of the current deal would fall by 48.93 pps. As a reference, the average liquidation rate in our sample is 17.82%. In contrast, MLOT allocated to the premium investors is uncorrelated with the liquidation risk. In summary, we find evidence consistent with these conditions: (i) premium PIPE investors generate value-relevant information, (ii) non-premium PIPE investors are paid agency fees, and (iii) the non-premium investors help close weaker deals in the future. In Appendix A, we show that these results are robust to removing outlier investors, placebo treatments, and the definition of premium PIPE investors. ## 6 Observed and Unobserved Heterogeneity In this section, we determine how much of the difference in average earnings between premium and non-premium investors can be explained by differences in their observable characteristics. If they have different “business models” (information production vs. agency service), we expect observed heterogeneity to explain a smaller portion of the difference. To minimize the effect of outliers in this exercise, we exclude seven out of 473 successful SPACs with closing share prices higher than $100.181818These SPACs are: Queen’s Gambit Growth Capital, Insurance Acquisition Corp., Matlin & Partners Acquisition Corporation, Northern Genesis Acquisition Corp. II, VPC Impact Acquisition Holdings III, Inc., Reinvent Technology Partners Z (f/k/a Reinvent Technology Partners B), and Ajax I. In Figure 3 (left panel), we display the histogram of total MLOT paid to all PIPE (at) investors across SPACs that have successfully merged with a private target firm. The average MLOT per deal is $62.38 million, with a large standard deviation of $199.07 million. These earnings are divided among, on average, 10.76 PIPE investors (see Table 2), out of which 3.02 are premium investors. In Figure 3 (right panel), we show the shares of premium PIPE investors (left vertical line) and the total MLOT (right vertical line). Averaging these shares, we find that premium PIPE investors get (around) 26% of the MLOT. In other words, 26% of the underpricing at the time of the business closing announcement is for information production, and 74% is for quid pro quo. Figure 2: Money Left on the Table Note: The left panel is the histogram of total MLOT for all $\texttt{PIPE}_{\texttt{at}}$ investors. The right panel shows the premium PIPE investors’ share of MLOT (vertical axis on the left) and the total MLOT in that deal (vertical axis on the right). Considering each investor separately, the MLOT of a premium investor, on average, is $6.96 million, and that of a non-premium investor is $3.17 million. Next, we determine what part of this $3.79 million difference in average MLOT can be explained by differences in the observable characteristics of these two types of investors. First, we introduce new variables capturing some of the observed differences between these two groups. In particular, for each PIPE investor and each SPAC, we determine the number of SPACs they have participated in before the current one; how much they have invested; the number of past SPACs that involved a sponsor different from the current one; total MLOT from other sponsors; and the number of SPACs with sponsors different from the current one, and whether the target is in the same industry as the current target. These variables capture market-wide experience and successes. We have 4,152 deals at the SPAC and PIPE (at) investor level, of which 1,245 are with premium investors and the rest with non-premium investors. Table 12 presents the descriptive statistics and the results from the tests of mean differences between premium and non-premium investors. Table 12: Difference in Characteristics of Premium and Non-premium Investors | Premium PIPE Sample | Non-Premium PIPE Sample | Difference ---|---|---|--- MLOT ($mm) | 6.95 | 3.17 | 3.78∗∗∗ SPAC Characteristics | | | IPO Proceeds ($mm) | 397.25 | 363.87 | 33.38∗∗∗ Days Remaining | 492.98 | 467.59 | 25.38∗∗∗ PIPE Investor Characteristics | | | No. of past SPACs | 23.99 | 5.51 | 18.48∗∗∗ Past Investment Amount ($mm) | 460.09 | 50.77 | 409.32∗∗∗ No. of SPACs with other sponsors | 24.40 | 5.44 | 18.95∗∗∗ Past MLOT with other sponsors ($mm) | 292.85 | 24.61 | 268.25∗∗∗ No. of SPACs with other sponsors, same industry | 3.59 | 0.80 | 2.79∗∗∗ Past MLOT with other sponsors, same industry ($mm) | 45.83 | 3.65 | 42.18∗∗∗ Note: This table reports means and differences in our sample’s average characteristics for premium and non-premium investors. ${}^{*},^{**}$, and ∗∗∗ denote p-values less than 0.1, 0.05, and 0.01, respectively. The two groups display distinct MLOTs and characteristics. Specifically, premium investors are involved in larger SPACs that take fewer days to identify a target. Moreover, these investors possess greater experience in the SPAC market, as evidenced by a higher number of previous SPAC participations (23.99 vs. 5.51) and a larger investment amount ($460.09 million vs. $50.77 million). They also work with more SPACs (24.40 vs. 5.44) and obtain a higher MLOT ($292.85 million vs. $24.61 million) from other sponsors. In addition, premium investors operate more in the same industry with different sponsors (3.59 vs. 0.80). Figure 3: Kitagawa-Blinder-Oaxaca Decomposition Note: The figure displays the explained components of a twofold Kitagawa- Blinder-Oaxaca decomposition of the premium vs. non-premium MLOT. The x-axis is the mean difference in MLOT (in millions), and the y-axis lists observable characteristics. Next, we use the Kitagawa-Blinder-Oaxaca decomposition to document the extent and drivers of the MLOT differences between premium and non-premium investors. In particular, we begin with the following: $\text{MLOT}_{i,g}={\mathbf{W}}_{g}^{\top}\omega_{g}+\iota_{i,g},\quad g\in\\{\text{premium, non-premium}\\},$ where ${\bf W}$ is the vector of observed characteristics in Table 12, and $\iota$ is a mean zero random variable. Let $\Delta\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{MLOT}=\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{W}_{premium}^{\top}\hat{\omega}_{\text{premium}}-\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{W}_{non-premium}^{\top}\hat{\omega}_{\text{non-premium}}$, where $\overline{W}_{g}$ denotes the sample average of the group $W_{g}$. We can then re-write $\Delta\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{MLOT}$ as $\displaystyle\Delta\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{MLOT}=\underbrace{(\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{\bf W}_{premium}-\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{\bf W}_{non-premium})^{\top}\hat{\omega}_{non-premium}}_{\text{explained portion}}+\underbrace{\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{\bf W}^{\top}_{premium}(\hat{\omega}_{premium}-\hat{\omega}_{non- premium})}_{\text{unexplained portion}}.$ (8) The explained portion is the portion of the MLOT difference explained by differences in observed characteristics between premium and non-premium investors. In other words, if we change $\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{\bf W}_{non-premium}$ to $\macc@depth\char 1\relax\frozen@everymath{\macc@group}\macc@set@skewchar\macc@nested@a 111{\bf W}_{premium}$, it measures how much additional MLOT a non-premium investor gets. The unexplained portion is the part driven by differences in the coefficients. We find that the mean increase in non-premium investors’ MLOT is $0.04 million (s.e. $1.38 million) if they had the same characteristics as premium investors. So the differences in observed characteristics do not account for the MLOT gap. Figure 3 displays how much each of the eight components in ${\bf W}$ explains, along with their confidence intervals. Insofar as ${\bf W}$ contains all characteristics that affect MLOT, we find that most of the difference in average MLOT cannot be explained by differences in average ${\bf W}$, suggesting that the difference in the business models of these two groups is important. ## 7 Conclusion Understanding reasons for underpricing hinges on the relative importance of information production and quid pro quo. In this paper, we determine the relative importance within the context of SPACs. Distinguishing between premium and non-premium PIPE (institutional) investors, we show that the former produce information and, on average, are paid 26% of the total underpricing. The latter extracts the remainder 74% as quid pro quo payments. While we have relied on some institutional features specific to SPACs, understanding how these features arise to alleviate information and agency frictions have implications for security issuance under asymmetric information in general and IPOs in particular. After all, the SPAC business combination process is also a case of security issuance through sponsors as an intermediary with PIPE investors as buyers and the target company as the seller. There are two main takeaways for IPOs. First, information production and quid pro quo can play important roles in IPO underpricing. In other words, they are not necessarily mutually exclusive. Second, non-premium PIPE investors help sponsors with weak deals, suggesting that, on average, they are value- enhancing for targets and issuers. This role is unlike the traditional IPO, where some investment bank clients get favorable allocations of “hot” IPO shares in exchange for paying high trading fees to the bank. The benefit of quid pro quo payment is a function of its positive effect on successful business combinations. However, to determine if quid pro quo payments enhance welfare, we have to show that the total welfare is higher when the sponsor is allowed to build and deplete relationships than if not. These are interesting questions for future research. ## References * (1) * Aryal, Ciliberto and Leyden (2022) Aryal, Gaurab, Federico Ciliberto, and Benjamin T. Leyden. 2022\. “Coordinated Capacity Reductions and Public Communication in the Airline Industry.” Review of Economic Studies, 89(6): 3055–3084. * Beatty and Ritter (1986) Beatty, Randolph P., and Jay R. Ritter. 1986\. “Investment Banking, Reputation, and the Underpricing of Initial Public Offerings.” Journal of Financial Economics, 15(1-2): 213–232. * Benveniste et al. (2003) Benveniste, Lawrence M., Alexander Ljungqvist, William J. Wilhelm Jr, and Xiaoyun Yu. 2003\. “Evidence of Information Spillovers in the Production of Investment Banking Services.” Journal of Finance, 58(2): 577–608. * Benveniste and Spindt (1989) Benveniste, Lawrence M, and Paul A Spindt. 1989\. “How Investment Bankers Determine the Offer Price and Allocation of New Issues.” Journal of Financial Economics, 24(2): 343–361. * Benveniste, Busaba and Wilhelm (2002) Benveniste, Lawrence M., Walid Y. Busaba, and William Wilhelm. 2002\. “Information Externalities and the Role of Underwriters in Primary Equity Markets.” Journal of Financial Intermediation, 11(1): 61–86. * Booth and Smith II (1986) Booth, James R, and Richard L Smith II. 1986\. “Capital Raising, Underwriting and the Certification Hypothesis.” Journal of Financial Economics, 15(1-2): 261–281. * Bradley and Jordan (2002) Bradley, Daniel J., and Bradford Jordan. 2002\. “Partial Adjustment to Public Information and IPO Underpricing.” Journal of Financial and Quantitative Analysis, 37(4): 595–616. * Campbell, Lo and MacKinlay (1997) Campbell, John Y., Andrew W. Lo, and A. Craig MacKinlay. 1997\. The Econometrics of Financial Markets. Princeton University Press. * Chemmanur, Hu and Huang (2010) Chemmanur, Thomas J., Gang Hu, and Jiekun Huang. 2010\. “The Role of Institutional Investors in Initial Public Offerings.” Review of Financial Studies, 23(12): 4496–4540. * Cornelli and Goldreich (2001) Cornelli, Francesca, and David Goldreich. 2001\. “Bookbuilding and strategic allocation.” Journal of Finance, 56(6): 2337–2369. * Fabra and Reguant (2014) Fabra, Natalia, and Mar Reguant. 2014\. “Pass-Through of Emissions Costs in Electricity Markets.” American Economic Review, 104(9): 2872–2899. * Fudenberg and Levine (1994) Fudenberg, Drew, and David K. Levine. 1994\. “Efficiency and Observability with Long-Run and Short-Run Players.” Journal of Economic Theory, 62(1): 103–135. * Gahng, Ritter and Zhang (Forthcoming) Gahng, Minmo, Jay R. Ritter, and Donghang Zhang. Forthcoming. “SPACs.” Review of Financial Studies. * Gofman and Yao (2022) Gofman, Michael, and Yuchi Yao. 2022\. “SPACs’ Directors Network: Conflicts of Interest, Compensation, and Competition.” Available at SSRN 4148668. * Hanley (1993) Hanley, Kathleen Weiss. 1993\. “The Underpricing of Initial Public Offerings and the Partial Adjustment Phenomenon.” Journal of Financial Economics, 34(2): 231–250. * İnce (2014) İnce, Özgür Ş. 2014\. “Why Do IPO Offer Prices Only Partially Adjust?” The Quarterly Journal of Finance, 04(03): 1450009. * Jenkinson, Jones and Suntheim (2018) Jenkinson, Tim, Howard Jones, and Felix Suntheim. 2018\. “Quid Pro Quo? What Factors Influence IPO Allocations to Investors?” Journal of Finance, 73(5): 2303–2341. * Klausner, Ohlrogge and Ruan (2022) Klausner, Michael, Michael Ohlrogge, and Emily Ruan. 2022\. “A Sober Look at SPACs.” Yale Journal of Regulation, 39(1): 228–303. * Levin (2003) Levin, Jonathan. 2003\. “Relational Incentive Contracts.” American Economic Review, 93(3): 835–857. * Loughran and Ritter (2002) Loughran, Tim, and Jay R. Ritter. 2002\. “Why Don’t Issuers Get Upset about Leaving Money on the Table in IPOs?” Review of Financial Studies, 15(2): 413–444. * Loughran and Ritter (2004) Loughran, Tim, and Jay R. Ritter. 2004\. “Why Has IPO Underpricing Changed over Time?” Financial Management, 33(3): 5–37. * Lowry and Schwert (2004) Lowry, Michelle, and G. Schwert. 2004\. “Is the IPO Pricing Process Efficient?” Journal of Financial Economics, 71(1): 3–26. * Lowry et al. (2017) Lowry, Michelle, Roni Michaely, Ekaterina Volkova, et al. 2017\. “Initial public offerings: A synthesis of the literature and directions for future research.” Foundations and Trends® in Finance, 11(3-4): 154–320. * Ritter and Welch (2002) Ritter, Jay R., and Ivo Welch. 2002\. “A Review of IPO Activity, Pricing, and Allocations.” Journal of Finance, 57(4): 1795–1828. * Smith Jr (1986) Smith Jr, Clifford W. 1986\. “Investment Banking and the Capital Acquisition Process.” Journal of Financial Economics, 15(1-2): 3–29. * Wang and Yung (2011) Wang, Wei, and Chris Yung. 2011\. “IPO Information Aggregation and Underwriter Quality.” Review of Finance, 15(2): 301–325. APPENDIX ## Appendix A Robustness In Section 5, we found evidence consistent with the hypotheses that premium investors produce value-relevant information and non-premium investors engage in quid pro quo arrangements. We also found evidence that the quid pro quo relationship between a sponsor and non-premium investors enables weaker private firms to go public. In this section, we perform a series of robustness exercises. ### A.1 Possible Outliers. Figure A4: Scatter Plots of SPACs Note: The figure shows the scatter plot of SPACs by month and year. The top figure shows the number of SPAC mergers, and the bottom figure shows the deal sizes (total IPO proceeds) in billions of dollars. The size of each dot is proportional to the number of deals in a month (top) and the total IPO proceeds (bottom) in that month. As we can see from Figure A4, although our sample starts in 2010, most of the variation comes from SPAC from 2020 to 2022. Although our sample includes the so-called boom and bust cycle for SPAC, we may worry about the effects of outliers in our analysis. One such concern is that not all 34 premium PIPE investors (in Table 3) produce value-relevant information and that our results are driven by either the large premium PIPE investors or that the small premium PIPE investors are, in fact, non-premium investors. To address this concern, we exclude six (large) premium investors with at least $1 billion invested in SPACs and re-estimate all our models. Then, we repeat the same exercise, but this time we exclude seven (small) premium investors with at most $400 million invested in SPACs and re-estimate all models. Table A1: Robustness: Excluding Outliers | Redemption | Announcement-day | Mean Reversion | Liquidation Risk | Liquidation ---|---|---|---|---|--- | Rate (%) | Return (%) | MLOT | | | | | (1) | (2) | (3) | (4) | (5) | (6) | (7) | (8) | (9) | (10) All PIPE Investors (pre) | 0.180* | 0.159* | -0.085* | -0.074* | | | | | | | (0.099) | (0.095) | (0.045) | (0.044) | | | | | | Premium PIPE Investors (at) | -0.241*** | -0.214*** | 0.143** | 0.126*** | 0.153 | 0.149 | 0.114** | 0.080 | 0.423* | 0.356 | (0.082) | (0.051) | (0.060) | (0.041) | (0.104) | (0.099) | (0.058) | (0.061) | (0.226) | (0.222) Non-premium PIPE Investors (at) | -0.042 | -0.038 | 0.001 | -0.001 | -1.633*** | -2.075*** | -0.155** | -0.126* | -0.448* | -0.391 | (0.030) | (0.030) | (0.014) | (0.015) | (0.531) | (0.615) | (0.062) | (0.065) | (0.235) | (0.241) All PIPE Investors (post) | 0.152*** | 0.148*** | | | | | | | | | (0.036) | (0.036) | | | | | | | | Exclude Large Premium Investors | ✓ | ✗ | ✓ | ✗ | ✓ | ✗ | ✓ | ✗ | ✓ | ✗ Exclude Small Premium Investors | ✗ | ✓ | ✗ | ✓ | ✗ | ✓ | ✗ | ✓ | ✗ | ✓ Pseudo/Adj. $R^{2}$ | 0.389 | 0.399 | 0.074 | 0.094 | | | 0.0248 | 0.0211 | 0.0153 | 0.0133 F-stat (first-stage) | | | | | 175.54 | 140.73 | | | | Observations | 430 | 430 | 417 | 417 | 3,487 | 3,578 | 680 | 680 | 681 | 680 Note: This table presents estimates after excluding outliers in premium investors. Large premium investors are PIPE investors that have invested a total of more than $1 billion in more than 50 SPACs. They include Fidelity Management & Research Company, BlackRock, Inc., Capital Research and Management Company, Alyeska Investment Group L.P., Morgan Stanley, and Millennium Management, LLC. Small premium investors are those that invested less than $0.4 billion and participated in at most 50 SPAC deals. They include Ghisallo Capital Management LLC, Jane Street Group, LLC, Park West Asset Management LLC, Kepos Capital LP, Linden Advisors LP, Schonfeld Strategic Advisors LLC, and BlueCrest Capital Management Ltd. Columns (1) and (2) have the same specification as Table 4, column (4). Columns (3) and (4) have the same specification as Table 5, column (4). Columns (5) and (6) have the same specification as Table 9, column (3). Columns (7) and (8) have the same specification as Table 11, column (1). Columns (9) and (10) have the same specification as Table 11, column (2). ${}^{*},^{**}$, and ∗∗∗ denote p-values less than 0.1, 0.05, and 0.01, respectively. The results are presented in Table A1. In particular, the estimates from excluding large premium investors are in columns (1), (3), (5), and (7) of Table A1, and the estimates from excluding small premium investors are in the remaining columns. Comparing these estimates with those in columns (4), (4), (3), and (2) of Tables 4, 5, 9, and 11, respectively, we see that our main findings are neither due only to the large premium investors nor the small premium investors. The only exception is the liquidation result when we exclude the large investors (in column (9)), which is less stable. ### A.2 Placebo Exercises. Next, we verify that our results are not purely noise and that they capture the effects of premium and non-premium investors. To this end, we explore the effect of choosing an arbitrary set of PIPE investors and designating them as premium PIPE investors on the estimates. If our results were only noise, then a random group of premium PIPE investors should also give similar estimates. To implement this “placebo” exercise, we randomly picked 34 out of all 1,826 PIPE investors and labeled them premium investors. Then we estimate all the models using this new set of premium investors. We repeat this exercise 500 times and present only the coefficients for premium and non-premium investors and the 95% confidence intervals in Figure A5, in the first and the second column, respectively. The five rows are redemption rate, announcement day return, mean-reversion, liquidation risk, and probability of liquidation.191919For the mean-reversion exercise, we present the first 500 cases with the first-stage F-stat $\geq 50$. Now, let us consider each row separately. For the redemption rates, comparing the first row in Figure A5 with the estimates in Table 4 column (4), we see that the randomly chosen premium PIPE investors do not affect redemption rates. However, because the non-premium PIPE investors include some (actual) premium PIPE investors, the redemption rate is low but negatively correlated with allocations for non-premium PIPE investors. Similarly, comparing the second row in Figure A5 with the estimates in Table 5 column (4), we find no effect of premium PIPE investors on the announcement-day return but a small and positive effect of non-premium PIPE investors on first-day return because the latter set now contains the premium investors identified in Table 3. These effects are consistent with only the premium investors identified in Table 3 producing information. For the mean-reversion (in the third row), some estimates are negative, which is consistent with the new premium PIPE investor containing non-premium PIPE investors and vice versa. Finally, for liquidation risk and probability of liquidation, comparing rows four and five in Figure A5 with Table 11 columns (1) and (2), we also find that the new estimates are mixed, which is again consistent with the new premium investors containing non-premium investors, although the estimates have wider confidence intervals. Figure A5: “Placebo” Exercises Note: This figure shows estimation results using randomly generated premium PIPE identities. We randomly pick 34 PIPE investors out of 1,826, label them as premium PIPE investors, and re-estimate our models. We repeat this exercise so that we have 500 sets of randomly chosen premium PIPE samples. The first column presents the coefficients corresponding to (the new) premium PIPE investors and the second column corresponds to the (new) non-premium PIPE investors. The five rows correspond to the (i) redemption rates, (ii) announcement-day return, (iii) mean reversion, (iv) liquidation risk, and (v) probability of liquidation. In all these five cases, the specifications are the same as in Tables 4 and 5, column (4), Table 9 column (3), and Table 11 columns (1) and (2), respectively. Table B1: Variable Definition Variable | Unit | Definition ---|---|--- Panel A: SPAC-Level Characteristics SPAC Measures | IPO proceeds | MM USD | Gross proceeds raised in the IPO, including any full or partial exercise of the Greenshoe. No. of Warrant | | Number of warrants included in the unit issued at the SPAC’s IPO. No. of Right | | Number of rights included in the unit issued at the SPAC’s IPO. Overallotment | % | Amount of the overallotment option exercised by the underwriter(s) as a fraction of total IPO proceeds. Listing-day return | % | SPAC IPO investors’ first-day return. Days searching | Day | The number of days between the SPAC’s IPO date and the target announcement date. Days remaining | Day | The number of days between the SPAC’s target announcement date and the liquidation deadline. Announcement-day return | % | One-day return on the SPAC’s shares on the target’s announcement date. Redemption | % | Redeemed SPAC common shares as a percentage of total shares issued at IPO. Days deSPAC | Day | The number of days between the SPAC’s target announcement date and the merger closing date. Cold | pps | The difference between the SPAC’s offer price and the SPAC’s share price on the closing day of the merger. Liquidation Risk | Day | The number of days between the SPAC’s IPO date, the target announcement date, or the liquidation date. Liquidation | binary | One if the SPAC liquidated during our sample period, zero otherwise. PIPE Participation | $\mathbbm{1}${PIPE (pre)} | Binary | One if PIPE (pre) investors participate in the deal. $\mathbbm{1}${PIPE (at)} | Binary | One if PIPE (at) investors participate in the deal. $\mathbbm{1}${PIPE (post)} | Binary | One if PIPE (post) investors participate in the deal. $\mathbbm{1}${Premium PIPE (at)} | Binary | One if premium PIPE (at) investors participate in the deal. $\mathbbm{1}${Non-Premium PIPE (at)} | Binary | One if non-premium PIPE (at) investors participate in the deal. No. of PIPE (at) | | Number of PIPE (at) investors. No. Premium PIPE (at) | | Number of premium PIPE (at) investors. No. of Non-Premium PIPE (at) | | Number of non-premium PIPE (at) investors. PIPE Investment/IPO Proceeds | %PIPE (pre) | % | PIPE (pre) investors’ investment as a percentage of total IPO proceeds. %Premium PIPE (at) | % | Premium PIPE (at) investors’ investment as a percentage of total IPO proceeds. %Non-Premium PIPE (at) | % | Non-premium PIPE (at) investors’ investment as a percentage of total IPO proceeds. %PIPE (post) | % | PIPE (post) investors’ investment as a percentage of total IPO proceeds. Panel B: PIPE-Investor-Level Characteristics Non-Premium | Binary | One if either the PIPE (at) investor’s number of participated PIPE deals or the total commitment amount is not in the top 5th percentile. Investment Amount | MM USD | Dollar amount of the PIPE investor’s investment in the SPAC deal. MLOT | MM USD | The difference between the value of securities held by the PIPE investor on the first day of listing of the post-merger company and the PIPE investor’s initial investment. lag-MLOT | MM USD | The PIPE investor’s MLOT from the previous SPAC deal with the same sponsor. lag2-MLOT | MM USD | The PIPE investor’s MLOT from the one before the previous SPAC deal with the same sponsor. MLOT(past) | MM USD | Total MLOT a PIPE investor made in past deals with the sponsor. No. of past SPACs | | Number of SPAC deals participated by the PIPE investor before the focal SPAC deal. Past Investment Amount | MM USD | The PIPE investor’s total amount of investments in SPAC deals before the focal SPAC deal. No. of SPACs with other sponsors | | Number of other sponsors’ SPAC deals participated in by the PIPE investor before the focal SPAC deal. Past MLOT with other sponsors | MM USD | The PIPE investor’s total MLOT made in other sponsors’ SPACs before the focal SPAC deal. No. of SPACs with other sponsors, same industry | | Number of other sponsors’ SPAC deals that found targets in the same industry as the focal SPAC and participated by the PIPE investor before the focal SPAC deal. Past MLOT with other sponsors, same industry | MM USD | The PIPE investor’s total MLOT made in other sponsors’ SPACs that found targets in the same industry as the focal SPAC. Outside Option | MM USD | Highest MLOT among all premium PIPE investors not included in the current deal. Participation | Binary | One if the sponsor allocates securities to the non-premium PIPE investor, and zero otherwise. Allocation | | Ratio of the non-premium investor’s shares to the total shares allocated to the sponsor’s relationship investors.
# Asymmetric Dependence Measurement and Testing H. D. Vinod address: H. D. Vinod, Professor of Economics, Fordham University, Bronx, New York, USA 10458. E-mail<EMAIL_ADDRESS>JEL codes C30, C51. Keywords: Kernel regression, Standardized beta coefficients, Partial Correlation. ###### Abstract Measuring the (causal) direction and strength of dependence between two variables (events), $X_{i}$ and $X_{j}$, is fundamental for all science. Our survey of decades-long literature on statistical dependence reveals that most assume symmetry in the sense that the strength of dependence of $X_{i}$ on $X_{j}$ exactly equals the strength of dependence of $X_{j}$ on $X_{i}$. However, we show that such symmetry is often untrue in many real-world examples, being neither necessary nor sufficient. Vinod’s (2014) asymmetric matrix $R^{*}\in[-1,1]$ of generalized correlation coefficients provides intuitively appealing, readily interpretable, and superior measures of dependence. This paper proposes statistical inference for $R^{*}$ using Taraldsen’s (2021) exact sampling distribution of correlation coefficients and the bootstrap. When the direction is known, proposed asymmetric (one-tail) tests have greater power. ## 1 Introduction A great deal of science focuses on understanding the dependence between variables. Its quantification has a long history starting with Galton-Pearson correlation coefficient $r_{ij}$ from the 1890s and its cousins, including Spearman’s $\rho$, Kendall’s $\tau$, and Hoeffding’s $D$. Let $dep(X_{i}|X_{j})$ measure the strength of dependence of $X_{i}$ on $X_{j}$ given a measurement $X_{j}$. Many measures of dependence try to satisfy the symmetry postulate by Renyi (1959), which posits that the two strengths based on opposite conditioning are identical: $dep(X_{i}|X_{j})\equiv dep(X_{j}|X_{i}).$ (1) We regard the symmetry postulate akin to an avoidable dogma. The following subsection explains why attempting to satisfy this symmetry equation (1) provides misleading measures of dependence in practice. ### 1.1 Four Examples of Asymmetric Dependence A correct notion of dependence in nature or data is rarely (if ever) symmetric. * • A newborn baby boy depends on his mother for his survival, but it is ludicrous to expect that his mother must exactly equally depend on the boy for her survival, as implied by (1). * • Meteorologists know that the average daily high of December temperatures in New York city is 44 degrees Fahrenheit and that this number depends on New York’s latitude (40.7). The latitude is a geographical given and does not depend on anything like city temperatures. Symmetric dependence by (1) between temperature and latitude implies the ludicrous claim that latitude depends on temperature with equal strength. * • For a third example, imagine a business person B with several shops. B’s 30% earnings depend on the hours worked by a key employee in one shop. Now the symmetry by (1) means that hours worked by the key employee always depend on B’s earnings, exactly 30%. * • Our fourth example assumes $Y$ as complete data, but a subset of $Y$ is unavailable. The available subset $X$ is a proxy that depends on $Y$, but the complete set $Y$ does not equally depend on its subset $X$. These four examples are enough to convince the reader that the symmetry postulate is neither necessary nor sufficient for real-world dependence. However, it is interesting that the unrealistic property (1) is an old established sacrosanct postulate from the 1950s, Renyi (1959). Even in 2022, Geenens and de Micheaux (2022)(“GM22”), still adhere to the symmetry postulate (dogma) by proposing an ingenious new definition of dependence to fit the model in (1). Actually, measure of dependence satisfying (1) can be ludicrous some contexts analogous to the four examples above. ### 1.2 Sources of the Symmetry Dogma What is the origin of symmetry dogma? (i) The definitional and numerical equality of covariances, $Cov(X_{i},X_{j})=Cov(X_{j},X_{i})$ may have been the initial reason for the symmetry result. (ii) In a bivariate linear regression $X_{1}=a+bX_{2}+\epsilon$, the strength of dependence of $X_{1}$ on $X_{2}$ is clearly measured by the coefficient of determination $R^{2}_{1|2}$. If we consider a flipped linear regression, $X_{2}=a^{\prime}+b^{\prime}X_{1}+\epsilon^{\prime}$, the strength of dependence is $R^{2}_{2|1}$. The assumption of linearity makes the two strengths equal to each other $R^{2}_{2|1}=R^{2}_{1|2}$. The equality of two $R^{2}$ strengths supports the symmetry dogma. When we consider the signed square roots of the two $R^{2}$ values, we have a symmetric matrix of correlation coefficients, $r_{ij}=r_{ji}$. These signed measures of dependence further support the dogma. The symmetry dogma depends on the harmless-looking linearity assumption. Back in 1784, the German philosopher Kant said: “Out of the crooked timber of humanity, no straight thing was ever made.” Since social sciences and medicine deal with human subjects, evidence supporting linearity and the implied symmetry dogma is missing. (iii) Since all distances satisfy symmetry, it may have been another reason behind Renyi’s postulate. (iv) The concept of statistical independence in probability theory is symmetric. It can be formulated in terms of the absence of any divergence between a joint density and a product of two marginal densities, $f(X_{i}X_{j})=f(X_{i})\,f(X_{j}).$ (2) Since dependence is the opposite of independence, it is tempting (but unhelpful) to impose symmetry on dependence as well. ### 1.3 Statistical Independence in Contingency Tables Two-way contingency tables refer to tabulated data on a grand total of $T$ observations distributed over an $r\times c$ matrix. There are two categorical variables represented by $r$ manifestations of row characteristics $R_{i}$ along ($i=1,2,\ldots r$) rows, and $c$ column characteristics $C_{j}$ along ($j=1,2,\dots c$) columns. The body of the ($r\times c$) contingency table has observed values $O_{ij}$ in a matrix cell located at row number $i$ and column number $j$. The joint probability $P(R_{i},C_{j})$ is simply $O_{ij}/GT$, where GT denotes the grand total of the tabulated numbers. The row margins of the contingency table have row totals, $R_{i}=\Sigma_{j}O_{ij}$. The column margin has column totals, $C_{j}=\Sigma_{i}O_{ij}$. The marginal probabilities are $P(R_{i})=R_{i}/GT$ and $P(C_{j})=C_{j}/GT$, which are also called unconditional probabilities. A conditional probability restricts the sample space to the part of the Table which satisfies the specified condition, referring to a particular row $R_{i}$ or column $C_{j}$. The direct computation of conditional probability has the respective row or column sums in its denominator instead of the grand total $GT$. An equivalent calculation of conditional probability defines $P(R_{i}|C_{j})=P(R_{i},C_{j})/P(C_{j})$, as a ratio of the joint probability to the marginal probability of the conditioning characteristic. Analogous conditional probability conditioning on row characteristic is a ratio of the same joint probability to the marginal probability of the conditioning row, $P(C_{j}|R_{i})=P(R_{i},C_{j})/P(R_{i})$. In probability theory based on contingency tables, the notion of statistical independence is studied by considering the following three criteria. (a) $P(R_{i},C_{j})=P(R_{i})P(C_{j})$, joint probability equals the product of marginals. (b) $P(R_{i}|C_{j})=P(R_{i})$, conditional probability equals unconditional or marginal probability. (c) $P(C_{j}|R_{i})=P(C_{j})$, the other conditional probability equals unconditional or marginal probability. Note that criterion (a) is both necessary and sufficient for independence. It is symmetric in that the joint probability is the same even if we interchange the order and write it as $P(C_{j},R_{i})$. However, data can satisfy (b) without satisfying (c), and vice versa. Hence tests of independence typically rely on the symmetric criterion (a). However, dependence is the opposite of independence and is generally asymmetric. We find that using (b) and (c) helps avoid the misleading symmetry postulate in the context of dependence. It is customary to imagine a population of thousands of contingency tables; the observed table is one realization from that population. The null hypothesis ($H_{0}$) is that row and column characteristics are statistically independent. The sample table may not exactly satisfy independence in the sense of (a) to (c) above. The testing problem is whether the observed table of $O_{ij}$ values could have arisen from a population where conditions (a) to (c) are satisfied. That is, $O_{ij}$ are numerically close enough to the expected values $E_{ij}=R_{i}C_{j}$ obtained from the cross-product of relevant marginal totals. Pearson’s Chi-square test statistic for ($H_{0}$) or independence of row effect and column effect in a contingency table is $\chi^{2}=(O_{ij}-E_{ij})^{2}/E_{ij},\quad df=(r-1)(c-1),$ (3) where $df$ denotes the degrees of freedom. Note that $\chi^{2}\in[0,\infty)]$ of (3) cannot be computed unless we have contingency tables. Statisticians have long recognized that the magnitude of $\chi^{2}$ cannot reliably measure the direction and strength of dependence. This paper assumes that a practitioner would want to know both the general direction and strength of dependence. ## 2 Symmetric Measures of Dependence Granger et al. (2004) (“Gr04”) is an important paper on formal testing for statistical independence, especially for time series data. They cite a survey by Tjostheim (1996) on the topic. The novelty in Gr04 is in using nonparametric nonlinear kernel densities in testing the equality (2) in their test of independence. Unfortunately, Gr04 authors adhere to the symmetry dogma by insisting that, similar to independence, a measure of dependence should be a symmetric distance-type ‘metric.’ ### 2.1 Dependence Measures and Entropy Shannon defined information content in 1948 as the amount of surprise in a piece of information. His information is inversely proportional to the probability of occurrence and applies to both discrete and continuous random variables with probabilities defined by a probability distribution $f(y)$. In the context of entropy, let us use the fourth example of Section 1.1, where $Y$ is the complete data and $X$ is a subset with some missing observations. How does $X$ depend on $Y$? We develop a measure of dependence using information theory, especially entropy. Intuitively, entropy is our ignorance or the extent of disorder in a system. The entropy $H(Y)$ is defined by the mathematical expectation of the Shannon information or $E(-logf(y))$. And the conditional entropy of Y given X averaged over $X$is $H(Y|X)=-E[E[log(f_{Y|X}(Y|X))|X]].$ (4) The reduction in our ignorance $H(Y)$ by knowing the proxy $X$ is $H(Y)-H(Y|X)$. Mutual information $I_{mu}(X,Y)$ is defined as $H(x)+H(Y)-H(X,Y).$ It is symmetric since $I_{mu}(X,Y)=I_{mu}(Y,X)$. The entropy-based measure of dependence is $D(X;Y)=\frac{H(Y)-H(Y|X)}{H(Y)},$ (5) or proportional reduction in entropy of $Y$ by knowing $X$. Reimherr and Nicolae (2013) complain that (5) is not symmetric. By contrast, we view asymmetry as a desirable property. Neyman-Pearson showed that a way to distinguish between two distributions $f(X)$ and $f(Y)$ for parameter $\theta$ is the difference between logs of their likelihood functions. Shannon’s relative entropy, also known as Kullback–Leibler (KL) divergence, is the expected value of that difference, $KLD=E(logf(\theta|X)-logf(\theta|Y)).$ (6) It is easy to verify that KLD or relative entropy is not symmetric. Gr04 authors state on page 650 that “Shannon’s relative entropy and almost all other entropies fail to be ‘metric’, as they violate either symmetry, or the triangularity rule, or both.” We argue that asymmetry is an asset, not a liability, in light of four examples in Section 1.1. Hence, we regard (5) or (6) as superior measures compared to the symmetric measure by Gr04. D(X;Y) of (5) and KLD of (6) cannot be used directly on data vectors. They need frequency distribution counts as input based on the grouping of data into bins (histogram class intervals). The choice of the number of bins is arbitrary, and D(X;Y) and KLD are sensitive to that choice. Hence, we do not recommend D(X;Y) or KLD as a general-purpose measure of dependence. ### 2.2 Dependence Measures and Fisher Information Fisher information measures the expected amount of information given by a random variable $Y$ about a parameter $\theta$ of interest. Under Gaussian assumptions, the Fisher information is inversely proportional to the variance. Reimherr and Nicolae (2013) use the Fisher information to define a measure of dependence. Consider the estimation of a model parameter $\theta$ using $X$ as a proxy for unavailable $Y$. That is, $X$ is a subset of $Y$ with missing observations, as in the fourth example of Section 1.1. If the Fisher information for $\theta$ based on proxy $X$ is denoted by $\mathcal{I}_{X}(\theta)$, they define a measure of dependence as: $D(X;Y)=\frac{\mathcal{I}_{X}(\theta)}{\mathcal{I}_{Y}(\theta)},$ (7) where $\mathcal{I}_{X}(\theta)\leq\mathcal{I}_{Y}(\theta)$. Consider the special case where a proportion $p$ of the $Y$ data are missing in $X$ at completely random locations. Then, the measure of dependence (7) equals $p$. This measure of dependence is almost acceptable because it is asymmetric, where subset $X$ being a proxy for $Y$ cannot be interchanged with $Y$, except that $D(X;Y)$ of (7) cannot be negative. Later, we recommend in Section 3 a more generally applicable and intuitive measure of dependence. ### 2.3 Regression Dependence from Copulas Consider a two-dimensional joint (cumulative) distribution function $F(X,Y)$ and marginal densities $U=F_{1}(X)$ and $V=F_{2}(Y)$ obtained by probability integral transformations. Sklar proved in 1959 that a copula function $C(F_{1},F_{2})=F$ is unique if the components are continuous. The copula function $C:[0,1]^{2}\to[0,1]$ is subject to certain conditions forcing it to be a bivariate uniform distribution function. It is extended to the multivariate case to describe the dependence structure of the joint density. We have noted in section 1.3 that a contingency table represents the joint dependence structure of row and column characteristics. Copulas represent similar joint dependence when row and column characteristics are continuous variables rather than simple categories. Dette et al. (2013) (“DSS13”) define joint density as $F_{X,Y}$, and conditional density of $Y$ given $X$ as $F_{Y|X=x}$. They use uniform random variables $U$ and $V$ to construct copula $C$ as a joint distribution function. The copula serves as their measure of dependence based on the quality of regression-based prediction of $Y$ from $X$. The flipped prediction of $X$ from $Y$ ignored by DSS13 is considered in Section 3 in the sequel. DSS13 assume Lipschitz continuity, which implies that a copula is absolutely continuous in each argument, so that it can be recovered from any of its partial derivatives by integration. The conditional distribution $F_{V|U=u}$ is related to the corresponding copula $C(X,Y)$ by $F_{V|U=u}(v)=\partial_{1}C_{X,Y}(u,v)$. A symmetric measure of dependence proposed by DSS13 is denoted here as $r_{D}(X,Y)=6\int_{0}^{1}\int_{0}^{1}F_{V|U=u}(v)^{2}dvdu,$ (8) where $r_{D}=0$ represents independence, and $r_{D}=1$ represents almost sure functional dependence. DSS13 focus on $r_{D}$ filling the intermediate range of the closed interval $[0,1]$ while ignoring the negative range $[-1,0)$. Section 3 covers $[-1,1]$, including the negative range. DSS13 rely on parametric copulas, making them subject to identification problems, as explained by Allen (2022). The numerical computation of (8) is involved since it requires the estimation of the copula’s partial derivative. DSS13 authors propose a kernel-based estimation method without providing any ready-to-use computational tools for $r_{D}$. Remark 3.7 in Beare (2010) states that symmetric copulas imply time reversibility, which is unrealistic for economic and financial data. Bouri et al. (2020) reject the symmetry dogma and note that their parametric copula can capture tail dependence, which is important in a study of financial markets. Allen (2022) uses nonparametric copula construction and asymmetric $R^{*}$ from Vinod (2014). Allen’s application to financial data shows that cryptocurrencies do not help portfolio diversification. ### 2.4 Hellinger Correlation $\eta$ as a Dependence Measure Now we turn to the recent GM22 paper mentioned earlier, which proposes Hellinger correlation $\eta$ as a new symmetric measure of the strength of dependence. They need to normalize to ensure that $\eta\in[0,1]$. GM22 denote the normalized version as $\hat{\eta}$. GM22 authors explain why dependence axioms by Renyi (1959) need updating, while claiming that their $\eta$ satisfies all updated axioms. Unfortunately, GM22 retain the symmetry axiom criticized in Section 1.1 above. An advantage of $\eta$ over Pearson’s $r_{ij}$ is that it incorporates some nonlinearities. Let $F_{1}$ and $F_{2}$ denote the known marginal distributions of random variables $X_{1}$ and $X_{2}$, and let $F_{12}$ denote their joint distribution. Now, GM22 authors ask readers to imagine reconstructing the joint distribution from the two marginals. The un-intuitive (convoluted?) definition of the strength of dependence by GM22 is the size of the “missing link” in reconstructing the joint from marginals. This definition allows GM22 to claim that symmetry is “unquestionable.” GM22 authors define squared Hellinger distance $\mathcal{H}^{2}(X_{1},X_{2})$ as the missing link between $F_{12}$ and $F_{1}F_{2}$. They approximate a copula formulation of $\mathcal{H}^{2}$ using the Bhattacharyya (1943) affinity coefficient $\mathcal{B}$. Let $C_{12}$ denote the copula of $(X_{1},X_{2})$, and $c_{12}$ denote its density. The computation of $\hat{\eta}$ in the R package HellCor uses numerical integrals $\mathcal{B}=\int\int\surd c_{12}$. Hellinger correlation $\eta$ is $\eta=\frac{2}{\mathcal{B}^{2}}\\{\mathcal{B}^{4}+(4-3\mathcal{B}^{4})^{1/2}-2\\}^{1/2}.$ (9) The Hellinger correlation is symmetric, $\eta(X_{1},X_{2})=\eta(X_{2},X_{1})$. GM22 provide an R package HellCor to compute $\hat{\eta}$ from data as a measure of dependence, and test the null hypothesis of independence of two variables. A direct and intuitive measure of dependence in a regression framework is the multiple correlation coefficient (of determination) $R_{1|2}^{2}$. It is symmetric because even if we flip $X_{1}$ and $X_{2}$, the $R_{2|1}^{2}$ from linear regressions is exactly the same. The reason for the equality of two flipped $R^{2}$ values, $R^{2}_{(2|1)}=R^{2}_{(1|2)}$, is the assumption of linearity of the two regressions. When we relax linearity, the two $R^{2}$ values generally differ, $R_{2|1}^{2}\neq R^{2}_{1|2}$. We argue that quantitative researchers should reject the unrealistic linearity assumption in the presence of ready-to-use kernel regression (np package) software. Kernel-based $R^{2}$ values of flipped regressions are rarely equal. GM22 cite Janzing et al. (2013) only to reject such asymmetric dependence suggested by nonparametric regressions. ## 3 Recommended Measures of Dependence We have noted earlier that covariances satisfy symmetry $Cov(X_{i},X_{j})=Cov(X_{j},X_{i})$. However, the sign of symmetric covariances suggests the overall direction of the dependence between the two variables. For example, $Cov(X_{i},X_{j})<0$ means when $X_{i}$ goes up $X_{j}$ goes down, by and large. Most of the symmetric measures of dependence discussed above fail to provide this type of useful directional information except Pearson’s correlation coefficients $r_{ij}$. Hence, $r_{ij}$ has retained its popularity as a valuable measure of dependence for over a century, despite assuming unrealistic linearity. Zheng et al. (2012) reject the dogma by introducing nonsymmetric generalized measures of correlation ($GMC\in[0,1]$), proving that $GMC(Y|X)\neq GMC(X|Y).$ (10) Since GMCs fail to provide directional information in covariances needed by practitioners, Vinod (2014) and Vinod (2017) extend Zheng et al. (2012) to develop a non-symmetric correlation matrix $R^{*}=\\{r^{*}_{ij}\\}$, where $r^{*}_{ij}\neq r^{*}_{ji}$, while providing an R package. The R package generalCorr uses kernel regressions to overcome the linearity of $r_{ij}$ from the np package by Hayfield and Racine, which can handle kernel regressions among both continuous and discrete variables. Sometimes the research interest is focused on the strength of dependence, while the direction is ignored, perhaps because it is already established. In that case, one can use the R package generalCorr and the R function depMeas(,). It is defined as appropriately signed larger of the two generalized correlations, or $depMeas(X_{i},X_{j})=sgn*max(|r^{*}(i|j)|,|r^{*}(j|i)|),$ (11) where $sgn$ is the sign of the covariance between the two variables. In general, both the strength and general direction of quantitative dependence matter. Hence, we recommend two asymmetric measures $r^{*}(X_{i}|X_{j})$ and $r^{*}(X_{j}|X_{i})$. The generalCorr package functions for computing $R^{*}$ elements are rstar(x,y) and gmcmtx0(mtx). The latter converts a data matrix argument (mtx) with $p$ columns, into a $p\times p$ asymmetric matrix $R^{*}$ of generalized correlation coefficients. Regarding the direction of dependence, the convention is that the variable named in the column is the “cause” or the right-hand regressor, and the variable named along the row is the response. Thus the recommended measures from $R^{*}$ are easy to compute. See an application to forecasting the stock market index of fear (VIX) and causal path determination in Allen and Hooper (2018). ### 3.1 Statistical Inference for Recommended Measures We recommend the signed generalized correlation coefficients $-1\leq r^{*}_{ij}\neq r^{*}_{ji}\leq 1$ from the $R^{*}$ matrix as the best dependence measure. This is because they do not adhere to the potentially misleading symmetry dogma while measuring arbitrary nonlinear dependence dictated by the data. An additional reason is its potential for more powerful (one-tail) inference, discussed in this section. The sign of each element of the $R^{*}$ matrix is based on the sign of the covariance $Cov_{ij}=Cov(X_{i},X_{j})$. A two-tail test of significance is appropriate only when $Cov_{ij}\approx 0$. Otherwise, a one-tail test is appropriate. Any one-tailed test provides greater power to detect an effect in one direction by not testing the effect in the other direction, Kendall and Stuart (1977), sections 22.24 and 22.28. Since sample correlation coefficient $r_{ij}$ from a bivariate normal parent has a non-normal distribution, Fisher developed his famous z-transformation in the 1920s. He proved that the following transformed statistic $r^{T}_{ij}$ is approximately normal with a stable variance, $r^{T}_{ij}=(1/2)\quad log\frac{(1+r_{ij})}{(1-r_{ij})}\sim N(0,1/n),$ (12) provided $r_{ij}\neq 1$. Recent work has developed the exact distribution of a correlation coefficient. It is now possible to directly compute a confidence interval for any hypothesized value $\rho$ of the population correlation coefficient. Let $r$ be the empirical correlation of a random sample of size $n$ from a bivariate normal parent. Theorem 1 of Taraldsen (2021) generalized Fisher’s famous z-transformation extended by C. R. Rao. The exact density with $v=(n-1)>1$ is $\displaystyle f(\rho|r,v)$ $\displaystyle=\frac{v(v-1)\Gamma(v-1))}{\surd(2\pi)\Gamma(v+0.5)}(1-r^{2})^{\frac{v-1}{2}}\,(1-\rho^{2})^{\frac{v-2}{2}}(1-r\rho)^{\frac{1-2v}{2}}$ $\displaystyle\quad F(\frac{3}{2};-0.5;v+0.5;\frac{1+r\rho}{2}),$ where F(.;.;.;.) denotes the Gaussian hypergeometric function, available in the R package hypergeom by R.K.S Hankin. The following R code readily computes (3.1) over a grid of 2001 $r$ values. library(hypergeo); r=seq(-1,1,by=0.001) Tarald=function(r,v,rho,cum){ #find quantile r given cum Trm1=(v*(v-1)*gamma(v-1))/((sqrt(2*pi)*gamma(v+0.5))) Trm2=(1-r^2)^((v-1)/2) Trm2b=((1-rho^2)^((v-2)/2))*((1-rho*r)^((1-2*v)/2)) Trm3b=hypergeo(3/2,-1/2,(v+0.5),(1+r*rho)/2) y0=Re(Trm1*Trm2*Trm2b*Trm3b) p=y0/sum(y0) cup=cumsum(p) loc=max(which(cup<cum))+1 return(r[loc])} Tarald(r=seq(-1,1,by=0.001),v=11,rho=0,cum=0.05) #example Assuming that the data come from a bivariate normal parent, the sampling distribution of any correlation coefficient is (3.1). Hence, the sampling distribution of unequal off-diagonal elements of the matrix of generalized correlations $R^{*}$ also follows (3.1). When we test the null hypothesis $H_{0}:\rho=0$, the relevant sampling distribution is obtained by plugging $\rho=0$ in (3.1) depicted in Figure 1 for two selected sample sizes. Both distributions are centered at the null value $\rho=0$. A two-tail (95%, say) confidence interval is obtained by using the 2.5% and 97.5% quantiles of the density. If the observed correlation coefficient $r$ is inside the confidence interval, we say that the observed $r$ is statistically insignificant, as it could have arisen from a population where the null value $\rho=0$ holds. Figure 1: Taraldsen’s exact sampling density of a correlation coefficient under the null of $\rho=0$, solid line n=50, dashed line n=15 Similarly, one can test the nonzero null hypothesis $H_{0}:\rho=0.5$ using the equation obtained by plugging $\rho=0.5$ in (3.1) depicted in Figure 2 . Figure 2: Taraldsen’s exact sampling density of correlation coefficient under the null of $\rho=0.5$, solid line n=50, dashed line n=15 Figures 1 and 2 show that the formula (3.1) and our numerical implementation are ready for practical use. These exact densities depend on the sample size and on the value of the population correlation coefficient, $-1\leq\rho\leq 1$. Given any hypothesized $\rho$ and sample size, a computer algorithm readily computes the exact density, similar to Figures 1 and 2. Suppose we want to help typical practitioners who want the tail areas useful for testing the null hypothesis $\rho=0$. Then, we need to create a table of a set of typical quantiles evaluated at certain cumulative probabilities and a corresponding selected set of common sample sizes with a fixed $\rho=0$. Because of the complicated form of the density (3.1), it is not surprising that its (cumulative) distribution function $\int_{-1}^{r}f(\rho|r,v)$ by analytical methods is not available in the literature. Hence, let us compute cumulative probabilities by numerical integration defined as the rescaled area under the curve $f(r,v)$ for $\rho=0$. See Figure 1 for two choices of $v(=n-1)$ for sample sizes (n=50, 15). The cumulative probability becomes a sum of rescaled areas of small-width rectangles whose heights are determined by the curve tracing $f(r,v)$. The accuracy of numerical approximation to the area is obviously better, the larger the number of rectangles. We use a sequence of $r\in[-1,1]$ created by the R command r=seq(-1,1, by =0.001), yielding 2001 rectangles. Denote the height of $f(r,v)$ by $H_{f}=H_{f(r,v)}$. Now, the area between any two $r\in[-1,1]$ limits, say $r_{Lo}$ and $r_{Up}$ is a summation of areas (height times width=0.001) of all rectangles. Now, the cumulative probabilities in the range are $\Sigma_{r_{Lo}}^{r_{Up}}H_{f}/\Sigma_{-1}^{1}H_{f},$ (14) where the common width cancels, and where the denominator $\Sigma_{-1}^{1}H_{f}$ converts the rectangle areas into probabilities. More generally, we can use $f(\rho,r,v)$ for any $\rho\in[-1,1]$. Thus we have a numerical approximation to the exact (cumulative) distribution function under the bivariate normality of the parent, $F(\rho,r,v)=\int_{-1}^{r}f(\rho|r,v).$ The transform from $f(.)$ to $F(.)$ is called the probability integral transform, and its inverse $F^{-1}(c|\rho,v)$ gives relevant correlation coefficients $r$ as quantiles for specified cumulative probability $c$ as the argument. A computer algorithm can readily find such quantiles. The exact $F^{-1}(c|\rho,v)$ allows the construction of confidence intervals based on quantiles for each $\rho$ and sample size. For example, a 95% two- tail confidence interval uses the 2.5% quantile $F^{-1}(c=0.025)$ as the lower limit, and 97.5% quantile $F^{-1}(c=0.975)$ as the upper limit. These limits depend on hypothesized $\rho$ and sample size. Since $\rho=0$ is a common null hypothesis for correlation coefficients, let us provide a table of $F^{-1}(c)$ quantiles for eleven sample sizes (listed in row names) and eight cumulative probabilities listed in column titles of Table 1. The p-values in statistical inference are defined as the probability of observing the random variable (correlation coefficient) as extreme or more extreme than the observed value of the correlation coefficient $r$ for a given null value $\rho=0$. Any one-tail p-values based on $f(\rho|r,v)$ of (3.1) for arbitrary nonzero “null” values of $\rho$ can be similarly computed by numerical integration defined as the area under the curve. Some code for R functions Tarald(.), and pTarald(.) is included in Sections 3.1 and 4, respectively. Table 1: Correlation coefficients as quantiles evaluated at specified cumulative probabilities (c=.) using Taraldsen’s exact sampling distribution for various sample sizes assuming $\rho=0$ | | c= | | | | | c= | ---|---|---|---|---|---|---|---|--- | c=0.01 | 0.025 | c=0.05 | c=0.1 | c=0.9 | c=0.95 | 0.975 | c=0.99 n=5 | -0.83 | -0.75 | -0.67 | -0.55 | 0.55 | 0.67 | 0.75 | 0.83 n=10 | -0.66 | -0.58 | -0.50 | -0.40 | 0.40 | 0.50 | 0.58 | 0.66 n=15 | -0.56 | -0.48 | -0.41 | -0.33 | 0.33 | 0.41 | 0.48 | 0.56 n=20 | -0.49 | -0.42 | -0.36 | -0.28 | 0.28 | 0.36 | 0.42 | 0.49 n=25 | -0.44 | -0.38 | -0.32 | -0.26 | 0.26 | 0.32 | 0.38 | 0.44 n=30 | -0.41 | -0.35 | -0.30 | -0.23 | 0.23 | 0.30 | 0.35 | 0.41 n=40 | -0.36 | -0.30 | -0.26 | -0.20 | 0.20 | 0.26 | 0.30 | 0.36 n=70 | -0.27 | -0.23 | -0.20 | -0.15 | 0.15 | 0.20 | 0.23 | 0.27 n=90 | -0.24 | -0.20 | -0.17 | -0.14 | 0.14 | 0.17 | 0.20 | 0.24 n=100 | -0.23 | -0.20 | -0.16 | -0.13 | 0.13 | 0.16 | 0.20 | 0.23 n=150 | -0.19 | -0.16 | -0.13 | -0.10 | 0.10 | 0.13 | 0.16 | 0.19 For the convenience of practitioners, we explain how to use the cumulative probabilities in Table 1 in the context of testing the null hypothesis $\rho=0$. The Table confirms that the distribution is symmetric around $\rho=0$ as in Figure 1. Let us consider some examples. If n=100, the critical value from Table 1 for a one-tail 95% test is 0.16 (line n=100, column c=0.95). Let the observed positive r be 0.3. Since $r$ exceeds the critical value, $(r>0.16$), we reject $\rho=0$. If n=25, the critical value for a 5% left tail in Table 1 is $-0.32$. If the observed $r=-0.44$, is less than the critical value $-0.32$ it falls in the left tail, and we reject $\rho=0$ to conclude that it is significantly negative. Table 1 can be used for constructing a few two-tail 95% confidence intervals as follows. If the sample size is 30, we see along the row n=30, and column c=0.025 gives $-0.35$ as the lower limit, and column c=0.975 gives $0.35$ as the upper limit. In other words, for n=30, any correlation coefficient smaller than 0.35 in absolute value is statistically insignificant. If the standard bivariate normality assumption is not believed, one can use the maximum entropy bootstrap (R package meboot) designed for dependent data. A bootstrap application creates a large number $J=999$, say, versions of data ($X_{i\ell},X_{j\ell}$) for $\ell=1,\ldots J$. Each version yields $r^{*}(i|j;\ell),r^{*}(j|i;\ell)$ values. A large set of $J$ replicates of these correlations give a numerical approximation to the sampling distribution of these correlations. Note that such a bootstrap sampling distribution is data-driven. It does not assume bivariate normality needed for the construction of Table 1 based on (3.1). Sorting the replicated $r^{*}(i|j;\ell),r^{*}(j|i;\ell)$ values from the smallest to the largest, one gets their “order statistics” denoted upon inserting parentheses by replacing $\ell$ by $(\ell)$. Now a left-tail 95% confidence interval for $r^{*}(i|j)$ leaves a 5% probability mass in the left tail. The interval is approximated by the order statistics as $[r^{*}(i|j;(50)),1]$. If the hypothesized $\rho=0$ is inside the one-tail interval, one fails to reject (accepts) the null hypothesis $H_{0}:\rho=0$. We conclude this section by noting that recommended measures of dependence based on the $R^{*}$ matrix and their formal inference are easy to implement. The tabulation of Taraldsen’s exact sampling distribution of correlation coefficients in Table 1 is new and should be of broader applicability. It is an improvement over standard significance tests of correlation coefficients based on Fisher’s z-transform. The next section illustrates with examples the use of Table 1, newer dependence measures, and other inference tools. ## 4 Dependence Measure Examples & Tests This section considers some examples of dependence measures. Our first example deals with fuel economy in automobile design. R software comes with ‘mtcars’ data on ten aspects of automobile design and performance for 32 automobiles. We consider two design features for illustration, miles per gallon $mpg$ and horsepower $hp$. Vinod (2014) reports the Pearson correlation coefficient $r(mpg,hp)=-0.78$ in his Figure 2. The negative sign correctly shows that one gets reduced $mpg$ when a car has larger horsepower $hp.$ Table 2 in Vinod (2014) reports two generalized correlation coefficients obtained by using kernel regressions as $r^{*}(mpg|hp)=-0.938$ and $r^{*}(hp|mpg)=-0.853$. One can interpret these $r^{*}(X_{i}|X_{j})$ values as signed strengths of dependence of $X_{i}$ on the conditioning variable $X_{j}$. The strengths are asymmetric, $|r^{*}(mpg|hp)|>|r^{*}(hp|mpg)|$, the absolute value of a dependence strength using both generalized correlation coefficients is larger than the dependence strength suggested under linearity. Thus Pearson’s correlation coefficient can underestimate dependence by assuming linearity. For the ‘mtcars’ data depMeas based on (11) is $-0.938$. Now consider Table 1 for n=30 row and c=0.05 for a one-tail critical value $-0.30$. The observed correlation $-0.938$ is obviously in the left tail (rejection region) of the exact sampling distribution of the correlation coefficient. Thus the negative dependence of fuel economy (mpg) on the car’s horsepower is statistically significant. We re-confirm the significance by computing the one-tail p-value (= 1e-16) using an R function pTarald(.). Our R code for p-values from Taraldsen’s exact density of correlation coefficients is given next. pTarald=function(r,n,rho,obsr){ v=n-1 if(v<=164) Trm1=(v*(v-1)*gamma(v-1))/((sqrt(2*pi)*gamma(v+0.5))) if(v>164) Trm1=(164*(163)*gamma(163))/((sqrt(2*pi)*gamma(163.5))) Trm2=(1-r^2)^((v-1)/2) if(rho!=0) Trm2b=((1-rho^2)^((v-2)/2))*((1-rho*r)^((1-2*v)/2)) if(rho==0) Trm2b=1 Trm3b=Re(hypergeo(3/2,-1/2,(v+0.5),(1+r*rho)/2)) y0=Re(Trm1*Trm2*Trm2b*Trm3b) p=y0/sum(y0) cup=cumsum(p) loc=max(which(r<obsr))+1 if(obsr<0) ans=cup[loc] if(obsr>=0) ans=1-cup[loc] return(ans)} pTarald(r=seq(-1,1,by=0.001),n=32,rho=0,obsr=-0.938) The first term (Trm1) in the R function computing the p-values involves a ratio of two gamma (factorial) functions appearing in (3.1). For $n>164$ each gamma becomes infinitely large, and Trm1 becomes ‘NaN’ or not a number. Our code winsorizes large $n$ values. Since mtcars data has n=32, and the observed generalized correlation is $r^{*}=0.938$, we use the command on the last line of the code to get the p-value of 1e-16, or extremely small, suggesting statistical significance. If we use the same automobile data using GM22’s R package called HellCor we find that $\eta=0.845>0$, giving no hint that mpg and hp are negatively related. If we compare numerical magnitudes, we have $\eta>|r(mpg,hp)|=0.78$. Since $\eta$ exceeds Pearson’s correlation in absolute value, $\eta$ is seen to incorporate nonlinear dependence. However, $\eta=0.78$ may be an underestimation of the absolute value of depMeas=$-0.938$. We fear that either $\eta$ may be failing to incorporate some nonlinear dependence or is paying an unknown penalty for adhering to the symmetry dogma. ### 4.1 Further Real-Data Applications in GM22 GM22 illustrate the Hellinger correlation measure of dependence using two sets of data where the Pearson correlation is statistically insignificant, yet their Hellinger correlation is significant. Their first data set refers to the population of seabirds and coral reef fish residing around n = 12 islands in the British Indian Ocean Territory of Chagos Archipelago. Ecologists and other scientists cited by GM22 have determined that fish and seabirds have an ecologically symbiotic relationship. The seabirds create an extra nutrient supply to help algae. Since fish primarily feed on those algae, the two variables should have a significantly positive dependence. GM22 begin with the low Pearson correlation $r(fish,seabirds)=0.374$ and a 95% confidence interval $[-0.2548,0.7803]$ that contains a zero, suggesting no significant dependence. The p-value using pTarald(..,obsr=0.374) is 0.0935, which exceeds the benchmark of 0.05, confirming statistical insignificance. The wide confidence interval, which includes zero, is partly due to the small sample size (n=12). Our Table 1 with the exact distribution of correlations suggests that when $n=10$, more conservative than the correct n=12, the exact two-tail 95% confidence interval (leaving 2.5% probability mass in both tails) also has a wide range $[-0.58,0.58]$, which includes zero. Assuming the direction is known, a one-tail interval with 5% in the right tail (n=10) value is 0.50. That is, only when the observed correlation is larger than 0.50 it is significantly positive (assuming a bivariate normal parent density). Figure 3: Marginal densities of fish and seabirds data are skewed, not Normal GM22 find that their Hellinger correlation $\eta$ needs to be normalized to ensure that $\eta\in[0,1]$, because their estimate of $\mathcal{B}$ can exceed unity. They denote the normalized version as $\hat{\eta}$ and claim an easier interpretation of $\hat{\eta}$ on the “familiar Pearson scale,” though Pearson’s $r_{ij}\in[-1,1]$ scale admits negative values. GM22 employ considerable ingenuity to achieve the positive range [0, 1] described in their Section 5.3. They state on page 650 that their range normalization “comes at the price of a lower power when it comes to test for independence.” Using the population of seabirds and coral reef fish residing around n = 12 islands, GM22 report the estimate $\hat{\eta}$(fish, seabirds)=0.744. If one assumes bivariate normal parent distribution and uses Taraldsen’s exact density from Table 1, $\hat{\eta}$(fish, seabirds)$=0.744>0.50$ suggests statistical significance. The p-value using pTarald(..,obsr=0.744) is 0.0027, which is smaller than the benchmark 0.05, confirming significance. In light of Figure 3, it is unrealistic to assume that the data come from a bivariate normal parent distribution. Hence the evidence showing a significantly positive correlation between fish and seabirds based on Taraldsen’s exact density is suspect. Accordingly, GM22 report a bootstrap p-value of $0.045<0.05$ as their evidence. Since this p-value is too close to 0.05, we check for unintended p-hacking. When one runs their HellCor(.) function with set.seed(99) and default settings, the bootstrap p-value becomes $0.0513>0.05$, which exceeds the benchmark suggesting insignificant $\hat{\eta}(fish,seabirds).$ Then, GM22’s positive Hellinger correlation estimate of $\hat{\eta}$= 0.744 is not statistically significant at the usual 95% level. Thus, the Hellinger correlation fails to be strongly superior to Pearson’s correlation $r$ because $\hat{\eta}$ is also insignificantly positive. Now, let us compare $\hat{\eta}$ with off-diagonal elements of the generalized correlation matrix $R^{*}$ recommended here. Our gmcmtx0(cbind(fish,seabirds)) suggests the “causal” direction (seabirds$\to$ fish), to be also positive, $r^{*}(fish|\break seabirds)=0.6687$. The p-value using pTarald(..,obsr=0.6687) is 0.0086, which is smaller than the benchmark 0.05, confirming significance. There is no suspicion of p-hacking here. A 95% bootstrap two-tail confidence interval using the meboot R package is [0.3898, 0.9373]. A one-tail interval is [0.4394, 1], which includes the observed 0.6687 with a p-value of zero. See Figure 4, where almost the entire density has positive support. Note that the interval does not include a zero, suggesting significant positive dependence consistent with what the ecologists expect. The lower limit of our meboot confidence interval is not close to zero. More importantly, our $R^{*}$ generalized correlation coefficients do not impose symmetric dependence, revealing sign information borrowed from the covariance, absent in the Hellinger correlation. Figure 4: Bootstrap density of generalized correlation coefficient r*(seabirds, fish). The second example in GM22 has the number of births ($X_{1}$) and deaths ($X_{2}$) per year per 1000 individuals in n=229 countries in 2020. A data scatterplot in their Figure 7 displays a C-shaped nonlinear relation. Pearson’s correlation $r_{12}=-0.13$ is negative and insignificant at level $\alpha=0.05$. This is based on a two-tail traditional Fisher approximation to the sampling distribution of a correlation coefficient. It is reversed by our more powerful one-tail p-value using Taraldsen’s exact sampling distribution. Our pTarald(..,n=229, obsr=-0.13) is $(0.0246<0.05)$, implying a statistically significant negative correlation. On the other hand, GM22 estimate $\hat{\eta}=0.69$ with a two-tail 95% bootstrap confidence interval [0.474, 0.746], hiding important information about the negative direction of dependence. Since zero is outside the confidence interval, GM22 claim that they have correctly overcome an apparently incorrect inference based on traditional methods. We have shown that traditional inference was incorrect because more accurate Taraldsen distribution was not used. Our gmcmtx0(cbind(birth,death)) estimates that $r^{*}(death|birth)$ is $=-0.6083$. A one-tail 95% confidence interval using the maximum entropy bootstrap (R package meboot) is [-1, -0.5693]. A somewhat less powerful two- tail interval [$-0.6251,-0.5641$] is also entirely negative. The null hypothesis states that the true unknown $r^{*}$ is zero. Since our random interval excludes zero, the dependence is significantly negative. The p-value is zero in Figure 5, since almost the entire density has negative support. A larger birth rate significantly leads to a lower death rate in 229 countries in 2020. Figure 5: Bootstrap density of generalized correlation coefficient r*(birth, death). In summary, the two examples used by GM22 to sell their Hellinger correlation have a discernible advantage over Pearson’s $r_{ij}$, but not over our generalized correlation $R^{*}$. The examples confirm four shortcomings of Hellinger’s correlation $\hat{\eta}$ over $R^{*}$. (a) It imposes an unrealistic symmetry assumption. (b) It provides no information about the direction of dependence. (c) It forces the use of less powerful two-tail confidence intervals. (d) It is currently not implemented for discrete variables. ## 5 Final Remarks Many scientists are interested in measuring the directions and strengths of dependence between variables. This paper surveys quantitative measures of dependence between two variables. We use four real-world examples in Section 1.1 to show that any symmetric measure of dependence alleging equal strength in both directions is unacceptable. Yet, the majority of statistical dependence extant in the literature adheres to the symmetry dogma. A 2022 paper (GM22) develops an intrinsically flawed symmetric measure of dependence while proposing Hellinger correlations. We show that off-diagonal elements of the asymmetric $R^{*}$ matrix of generalized correlation coefficients provide an intuitively sensible measure of dependence after incorporating nonlinear and nonparametric relations among the variables involved. The R package generalCorr makes it easy to implement our proposal. Its six vignettes provide ample illustrations of the theory and applications of $R^{*}$. We discuss statistical inference for elements of the $R^{*}$ matrix, providing a new Table 1 of quantiles of Taraldsen (2021) exact density of a correlation coefficient for eleven typical sample sizes and eight cumulative probabilities. We illustrate with two data sets used by GM22 to support their Hellinger correlation. Directional information is uniquely provided by our asymmetric measure of dependence in the form of generalized correlation coefficients $\\{r^{*}(i|j)\\}$. It allows the researcher to achieve somewhat better qualitative results and more powerful one-tail tests compared to symmetric measures of dependence in the literature. We claim that one-tail p-values of the Taraldsen’s density can overcome the inaccuracy of the traditional Pearson correlation inference based on Fisher’s z-transform. We illustrate the claim using GM22’s second example where Pearson correlation $r(birth,death)$ is shown to be significantly negative using Taraldsen’s density. Hence, the complicated Hellinger correlation inference is not really needed to achieve correct significance. Interestingly, both hand- picked examples designed to show the superiority of GM22’s $\hat{\eta}$ over $r_{ij}$ show the merit of our proposal based on $R^{*}$ over $\hat{\eta}$. Almost every issue of every quantitative journal refers to correlation coefficients at least once, underlining its importance in measuring dependence. We hope that $R^{*}$ and our implementation of Taraldsen’s exact sampling distribution of correlation coefficients receive further attention and development. ## References * Allen (2022) Allen, D. E. (2022), “Cryptocurrencies, Diversification and the COVID-19 Pandemic,” Journal of Risk and Financial Management, 15, URL https://www.mdpi.com/1911-8074/15/3/103. * Allen and Hooper (2018) Allen, D. E. and Hooper, V. (2018), “Generalized Correlation Measures of Causality and Forecasts of the VIX Using Non-Linear Models,” Sustainability, 10, 1–15, URL https://www.mdpi.com/2071-1050/10/8/2695. * Beare (2010) Beare, B. K. (2010), “Copulas and Temporal Dependence,” Econometrica, 78(1), 395–410. * Bhattacharyya (1943) Bhattacharyya, A. (1943), “On a Measure of Divergence Between Two Statistical Populations Defined by Their Probability Distributions,” Bulletin of the Calcutta Mathematical Society, 35, 99–109. * Bouri et al. (2020) Bouri, E., Shahzad, S. J. H., Roubaud, D., Kristoufek, L., and Lucey, B. (2020), “Bitcoin, gold, and commodities as safe havens for stocks: New insight through wavelet analysis,” The Quarterly Review of Economics and Finance, 77, 156–164, URL https://www.sciencedirect.com/science/article/pii/S1062976920300326. * Dette et al. (2013) Dette, H., Siburg, K. F., and Stoimenov, P. A. (2013), “A Copula-Based NonParametric Measure of Regression Dependence,” Scandinavian Journal of Statistics, 40, 21–41. * Geenens and de Micheaux (2022) Geenens, G. and de Micheaux, P. L. (2022), “The Hellinger Correlation,” Journal of the American Statistical Association, 117 (538), 639–653, URL 10.1080/01621459.2020.1791132. * Granger et al. (2004) Granger, C. W. J., Maasoumi, E., and Racine, J. (2004), “A Dependence Metric for Possibly Nonlinear Processes,” Journal of Time Series Analysis, 25, 649–669. * Janzing et al. (2013) Janzing, D., Balduzzi, D., Grosse-Wentrup, M., and Schölkopf, B. (2013), “Quantifying Causal Influences,” The Annals of Statistics, 41, 2324–2358. * Kendall and Stuart (1977) Kendall, M. and Stuart, A. (1977), The Advanced Theory of Statistics, vol. 2, New York: Macmillan Publishing Co., 4th ed. * Reimherr and Nicolae (2013) Reimherr, M. and Nicolae, D. L. (2013), “On Quantifying Dependence: A Framework for Developing Interpretable Measures,” Statistical Science, 28, 116–130, URL https://doi.org/10.1214/12-STS405. * Renyi (1959) Renyi, A. (1959), “On Measures of Dependence,” Acta Mathematica Academiae Scientiarum Hungarica, 10, 441–4510. * Taraldsen (2021) Taraldsen, G. (2021), “Confidence in Correlation,” arxiv, 1–7, URL 10.13140/RG.2.2.23673.49769. * Tjostheim (1996) Tjostheim, D. (1996), “Measures and tests of independence: a survey,” Statistics, 28, 249–284, URL https://arxiv.org/pdf/1809.10455.pdf. * Vinod (2014) Vinod, H. D. (2014), “Matrix Algebra Topics in Statistics and Economics Using R,” in “Handbook of Statistics: Computational Statistics with R,” , eds. Rao, M. B. and Rao, C. R., New York: North-Holland, Elsevier Science, vol. 34, chap. 4, pp. 143–176. * Vinod (2017) — (2017), “Generalized correlation and kernel causality with applications in development economics,” Communications in Statistics \- Simulation and Computation, 46, 4513–4534, available online: 29 Dec 2015, URL https://doi.org/10.1080/03610918.2015.1122048. * Zheng et al. (2012) Zheng, S., Shi, N.-Z., and Zhang, Z. (2012), “Generalized Measures of Correlation for Asymmetry, Nonlinearity, and Beyond,” Journal of the American Statistical Association, 107, 1239–1252.
# Ultrafast learning of 4-node hybridization cycles in phylogenetic networks using algebraic invariants Zhaoxing Wu Wisconsin Institute for Discovery Department of Statistics University of Wisconsin-Madison Madison, WI 53706 &Claudia Solís-Lemus Wisconsin Institute for Discovery Department of Plant Pathology University of Wisconsin-Madison Madison, WI 53706 Corresponding author<EMAIL_ADDRESS> ###### Abstract Motivation: The abundance of gene flow in the Tree of Life challenges the notion that evolution can be represented with a fully bifurcating process, as this process cannot capture important biological realities like hybridization, introgression, or horizontal gene transfer. Coalescent-based network methods are increasingly popular, yet not scalable for big data, because they need to perform a heuristic search in the space of networks as well as numerical optimization that can be NP-hard. Results: Here, we introduce a novel method to reconstruct phylogenetic networks based on algebraic invariants. While there is a long tradition of using algebraic invariants in phylogenetics, our work is the first to define phylogenetic invariants on concordance factors (frequencies of 4-taxon splits in the input gene trees) to identify level-1 phylogenetic networks under the multispecies coalescent model. Our novel inference methodology is optimization-free as it only requires the evaluation of polynomial equations, and as such, it bypasses the traversal of network space, yielding a computational speed at least 10 times faster than the fastest-to-date network methods. We illustrate the accuracy and speed of our new method on a variety of simulated scenarios as well as in the estimation of a phylogenetic network for the genus Canis. Availability and Implementation: We implement our novel theory on an open- source publicly available Julia package PhyloDiamond.jl available at https://github.com/solislemuslab/PhyloDiamond.jl with broad applicability within the evolutionary biology community. Contact<EMAIL_ADDRESS> _K_ eywords Hybridization $\cdot$ Polynomials $\cdot$ Reticulate evolution $\cdot$ Coalescent model ## 1 Introduction The Tree of Life is the graphical structure that represents the evolutionary process from single-cell organisms at the origin of life to present day biodiversity. Mathematically, a phylogenetic tree is a fully bifurcating graph in which internal nodes represent ancestral species that over time differentiate into two separate species giving rise to its two children nodes. Recent evidence [28, 11, 1, 33, 35, 44] has challenged the notion that evolution across the Tree of Life can be represented with a fully bifurcating process, as this process cannot capture important biological realities like hybridization, introgression, or horizontal gene transfer, that require two fully separated branches to join again. These processes of reticulate evolution are more prevalent in certain groups like plants [9], fungi [14, 43] and prokaryotes [13]. To accurately include these groups in the Tree of Life, recent years have seen an explosion of methods to reconstruct phylogenetic networks, which naturally account for reticulate evolution [12, 15, 7, 26]. Existing methods, however, are still not scalable enough to tackle the complexities of the present day’s biological big data. The most scalable alternatives infer split (or implicit) networks [8, 24, 21] which are not biologically interpretable as internal nodes no longer represent biological events (speciation or hybridization), and the resulting network is completely unrooted, even with the inclusion of an outgroup. Among methods to infer explicit networks (those whose internal nodes do represent speciation or hybridization events), likelihood-based approaches are among the most popular. PhyloNet [47, 48] has been the pioneer of likelihood inference of phylogenetic networks, later expanding to Bayesian [46] and pseudolikelihood alternatives [49]. However, even in its most scalable option (pseudolikelihood), the method is still not suitable beyond dozens of taxa and hundreds of genes. SNaQ [38] within the PhyloNetworks Julia package [39] has proven a scalable alternative to infer large phylogenetic networks from multilocus alignments. The method is based on a pseudolikelihood model under the coalescent, and running time does not increase as a function of number of genes because the input data for SNaQ are split frequencies on subsets of $4$ taxa (concordance factors). That is, after the estimation of the concordance factors, the running time is the same for data with 10 genes or 10,000 genes. Despite its popularity, SNaQ still lacks scalability as the number of taxa increases and remains a suitable alternative only for $50$ or fewer taxa. Last, BEAST2 [51] provides the only co-estimation method that infers simultaneously the phylogenetic networks and the gene trees, and it allows users to infer a variety of relevant biological parameters (such as divergence times), yet the complexity of this method renders it unsuitable for anything beyond a handful of taxa. Because of the complexity in inferring a phylogenetic network, the evolutionary biology community has embraced the use of hybrid detection methods such as ABBA-BABA test [32], MSCquartets [30, 36], and HyDe [27] to identify hybridization events on a fixed phylogeny. These methods take subsets of taxa and test whether the current subset is well-explained with a tree-like pattern or if there is a hybridization event in this subset. While fast, these methods can be inaccurate in the presence of multiple hybridization events affecting the same taxa [6] or by ghost lineages [31, 45, 6]. Furthermore, the process of first reconstructing a phylogenetic tree and then adding hybridization events can be flawed given the bias that gene flow causes on the inference of the backbone phylogenetic tree [41]. Here, we introduce a novel method to infer hybridization events using phylogenetic invariants. Because our method only requires evaluation of polynomial equations, it bypasses optimization in the space of networks and it is at least $10$ times faster than current inference network methodologies. Our method uses gene trees as input and it exploits the signal in the frequencies of splits in the data, also well-known as concordance factors. Indeed, under the coalescent model, splits of taxa display certain probabilities of appearing in the sample of gene trees. Under the true network, these frequencies need to satisfy certain polynomial equations, denoted phylogenetic invariants. By plugging in the observed frequencies on the phylogenetic invariants of every possible placement of the hybridization cycle, we can identify the hybridization cycle that better agrees with the data as the one whose evaluated invariants are closest to zero. Algebraic invariants have been widely used in phylogenetics to identify trees under a variety of models of evolution (Jukes-Cantor [16, 42], GTR [2]), and more recently, to identify phylogenetic networks [19, 20, 3, 10, 4]. Our work, however, is the first to define phylogenetic invariants on concordance factors under the multispecies coalescent model on networks as existing work on phylogenetic invariants on networks has been restricted to models of evolution. This work first introduces the phylogenetic invariants that concordance factors need to satisfy under level-1 phylogenetic networks with a 4-node hybridization cycle under the multispecies coalescent model. Next, we describe an inference methodology to use the phylogenetic invariants to infer the correct 4-node hybridization cycle among $n$ taxa. Moreover, we demonstrate the performance of our method with simulated data, and we revisit the inference of the Canis phylogenetic network that had already been published in [17]. We prove that our method is faster than all existing network inference methods, and it accurately identifies the correct hybridization cycle in the Canis genus. Additionally, in all simulating scenarios, our method identifies the correct placement of hybridization events with high accuracy. Even in the few cases when our method identifies a wrong network (usually cases when there are not sufficient sampling of taxa that descends from the hybrid node), the correct network is still within the top 5 ranked networks given the invariant score (evaluated phylogenetic invariants that need to vanish to zero). This means that, regardless of sampling, our method is always able to reduce the space of candidate networks that can later be tested with a likelihood-based approach. We highlight that our method has two main limitations: 1) it is only suitable to identify level-1 hybridization cycles with 4 nodes, and 2) it sometimes fails to identify the correct network as the one with the top 1 ranked invariant score when there is only one taxon sampled that is descendant of the hybrid node. Despite these limitations, we think that our method is a valuable innovation in the landscape of phylogenetic network methods, if anything, to reduce the space of candidate networks to be used in likelihood-based methodologies. Indeed, we do not see our method as a replacement for other network inference approaches. On the contrary, we believe that our ultrafast learning methodology can serve to identify the true phylogenetic network when it involves a simple hybridization event, or it can serve to provide several candidate networks to be later tested with a likelihood-based approach, effectively bypassing heuristic optimization in the space of networks. We further highlight that our method is open source, publicly available as the new PhyloDiamond.jl Julia package in https://github.com/solislemuslab/PhyloDiamond.jl. The structure of the paper is as follows. In Section 2, we introduce the phylogenetic invariants under the multispecies coalescent model on networks, and we describe an inference methodology that uses the phylogenetic invariants to identify the 4-node hybridization cycle that generates the data. In this Section, we also describe the simulation study and the re-analysis of the Canis data from [17]. In Section 3, we describe the results on the simulated data and the Canis phylogeny. Last, in Section 4, we describe the main limitations of our method and potential future work. ## 2 Materials and Methods ### 2.1 Phylogenetic invariants for 4-node hybridization cycles in level-1 phylogenetic networks Under the coalescent model, the distribution of gene trees estimated from multilocus sequence alignments provides information on the true network that generated the data [47]. In [38], it was shown that the split frequencies on subsets of 4 taxa, namely the concordance factors, also provide information on the true network. Namely, a concordance factor (CF) of a given quartet (or split) is the proportion of genes whose true tree displays that quartet (or split) [5]. For example, for a taxon set $s=\\{a,b,c,d\\}$, there are only three possible quartets, represented by the splits $q_{1}=ab|cd$, $q_{2}=ac|bd$ and $q_{3}=ad|bc$. The concordance factor for the split $ab|cd$ is the proportion of gene trees that display this split. As in [38], our method uses the CFs as input data, and we focus on the case of 4-node hybridization cycles on level-1 phylogenetic networks as shown on Figure 1 (left). A level-1 phylogenetic networks is a network with no vertex belonging to more than one hybridization cycle. The semi-directed network in Figure 1 (left) is unrooted, yet the direction of the hybrid edges (in blue) is known, so the root placement of the network is constrained. That is, the root cannot be anywhere below the hybrid node (in blue). The hybridization cycle partitions the taxa into 4 subsets with $n_{0},n_{1},n_{2}$ and $n_{3}$ as the number of taxa in each subset. For example, the clade below the hybrid node in blue in Figure 1 (left), also known as the hybrid clade, has $n_{0}$ taxa, and the two sister clades to the hybrid clade have $n_{1}$ and $n_{2}$ taxa. Even when $n_{i}$ represents the number of taxa in a specific clade, we use this terminology to refer to the clade. That is, we refer to the hybrid clade as the $n_{0}$ clade. Last, we highlight that each clade could be a subtree or a subnetwork as long as the overall network is level-1 [23]. Figure 1: Left: 4-node hybridization cycle on a level-1 phylogenetic network where $t_{i}$ represents the branch length in coalescent units. The hybrid node and hybrid edges are in blue. The minor hybrid edge is labelled with $\gamma$ that represents the inheritance probability (or proportion of genes transferred through the minor edge). Top right: Two examples of level-1 networks, each with one 4-node hybridization cycle, and its corresponding notation (N2222 for the 8-taxon network and N2112 for the 6-taxon network). Bottom right: Examples of quartets drawn from each network in top right along with their vector notations. For example, the quartet $AB|CD$ drawn from the N2222 network corresponds to the 4-taxon subset $(0,2,0,2)$ where each entry in this vector corresponds to the number of taxa drawn from each of the four clades: $n_{0},n_{1},n_{2},n_{3}$. In this work, we introduce a method to infer 4-node hybridization cycles in level-1 semi-directed networks. Since we focus on level-1 4-node hybridization cycles (as in Figure 1 (left)), we utilize a simplifying notation to represent each network. Namely, let $N$represent an $n-$taxon semi-directed level-1 phylogenetic network. We represent this network by the number of taxa in its four clades. For example, the network in the top right of Figure 1 is denoted N2222 as there are two taxa in each of the four clades, while the network N2112 contains two taxa in clades $n_{0}$ and $n_{3}$, and one taxon in clades $n_{1}$ and $n_{2}$. Another relevant notation is the vector representation of every 4-taxon subset. For example, in the bottom right of Figure 1, the 4-taxon subset $(0,2,0,2)$ drawn from network N2222 corresponds to the case when two taxa are taken from $n_{1}$ and $n_{3}$ and no taxon is taken from $n_{0}$ and $n_{2}$. For this specific network (N2222), this 4-taxon subset corresponds to the quartet $AB|CD$. Another example is the 4-taxon subset $(0,1,1,2)$ drawn from network N2112. Note that since this network only has one taxon in clades $n_{1}$ and $n_{2}$, any 4-taxon subset drawn from this network must have at most one taxon in the second and third value in the 4-taxon vector notation. In general, a 4-taxon subset is represented by a 4-dimensional vector so that each element in this vector corresponds to the number of taxa drawn from each of the four clades $n_{0},n_{1},n_{2},n_{3}$. Next, we describe the formulas for the expected CFs under the coalescent model for every 4-taxon subset on a level-1 semi-directed network with branch lengths in coalescent units ($t_{i}$) and inheritance probability ($\gamma$). These formulas have already been derived in [38]. Let $N$ be the network on Figure 1 (left), and let $(0,0,2,2)$ be the 4-taxon subset for which we want to compute the CF formulas under the coalescent model. Let taxa $k_{1},k_{2}\in n_{2}$ and taxa $l_{1},l_{2}\in n_{3}$. The probability of taxa $k_{1}$ and $k_{2}$ coalescing in a branch with length $t_{2}+t_{23}+t_{3}$ is given by $1-\dfrac{2}{3}z_{2}z_{23}z_{3}$ where $z_{i}=\exp(-t_{i})$. Then, the formulas for the three CF under the coalescent model are defined as: $\displaystyle P(k_{1},k_{2}|l_{1},l_{2})$ $\displaystyle=1-\dfrac{2}{3}z_{2}z_{23}z_{3}=a_{1}$ $\displaystyle P(k_{1},l_{1}|k_{2},l_{2})$ $\displaystyle=\dfrac{1}{3}z_{2}z_{23}z_{3}=a_{2}$ $\displaystyle P(k_{1},l_{2}|k_{2},l_{1})$ $\displaystyle=\dfrac{1}{3}z_{2}z_{23}z_{3}=a_{3}$ where $a_{1},a_{2}$ and $a_{3}$ denote the values for the true concordance factors corresponding to the splits $k_{1},k_{2}|l_{1},l_{2}$, $k_{1},l_{1}|k_{2},l_{2}$ and $k_{1},l_{2}|k_{2},l_{1}$ respectively. We thus note that for the 4-taxon subset $(0,0,2,2)$, there are three polynomial equations that represent the probabilities for each quartet under the coalescent model. These three formulas are equal to the true CF values ($a_{1},a_{2},a_{3}$) which will later be replaced by the observed CF estimated from the sample of gene trees. In the Appendix, we list of all the CF formulas for all 4-taxon subsets on a level-1 semi-directed network with one 4-node hybridization cycle (Figure 1 left). Because of the theoretical work in [38, 40], we know that we only need two taxa per clade to represent all CF formulas involving the hybridization cycle, and thus, we can restrict to the case of $n=8$ taxa for the definition of all CF formulas (yet our method is not restricted to 8 taxa, see Section 2.2). Furthermore, not all ${8\choose 4}=70$ 4-taxon subsets are required. Indeed, we ignore 4-taxon subsets of the form $(3,1,0,0)$ since three or more taxa from one of the clades (in this case, $3$ taxa from clade $n_{0}$) do not provide information about the hybridization cycle. Recall that only the length of internal branches appears in the CF formulas but not the length of external branches. Therefore, if there are three or more taxa on one clade (e.g. $3$ taxa in clade $n_{0}$), then branches involved in the hybridization cycle will correspond to an external branch and thus, will not be included in any of the CF formulas. The same is true if all four taxa in the 4-taxon subset all come from the same clade. Thus, only 4-taxon subsets where at most $2$ taxa come from a given clade contain information about the hybridization cycle. There are $19$ of such 4-taxon subsets in which $0\leq i,j,k,l\leq 2$ for $(i,j,k,l)$, and thus, there are $19\times 3=57$ concordance factor polynomial equations along with $57$ true CF values ($a_{i}$). As just described, the semi-directed network (denoted $N$) in Figure 1 (left) can be represented by a set of $57$ polynomial equations, denoted $CF(N)$ in $66$ variables: $9$ variables corresponding to branch lengths and inheritance probability ($z_{0},z_{1},z_{2},z_{3},z_{01},z_{02},z_{13},z_{23},\gamma$ for $z_{i}=\exp(-t_{i})$) and $57$ variables corresponding to the true CF values ($a_{i}$ for $i=1,\dots,57$). Here, we identify the relationships that the $a_{i}$ values need to satisfy for the polynomial system $CF(N)$ to be consistent. These relationships are denoted phylogenetic invariants. For example, because the three concordance factors corresponding to a given 4-taxon subset (say $(0,0,2,2)$ as described above) need to sum up to one, we have one phylogenetic invariant for these three numbers: $a_{1}+a_{2}+a_{3}=1$, or similarly $i_{1}(a_{1},a_{2},a_{3})=a_{1}+a_{2}+a_{3}-1$ with $i_{1}(a_{1},a_{2},a_{3})=0$. In addition, since the two minor concordance factors are equal, we have a second phylogenetic invariant defined for these numbers: $a_{2}=a_{3}$, or similarly $i_{2}(a_{2},a_{3})=a_{2}-a_{3}$ with $i_{2}(a_{2},a_{3})=0$. It turns out that the set of all concordance factor values for all 4-taxon subsets (the 57 polynomial equations denoted $CF(N)$) define a set of phylogenetic invariants (equations only in the $a_{i}$ variables) that need to vanish to zero whenever the concordance factors come from the true network. Note that not all networks will have the same number of CF polynomial equations. For example, for the 8-taxon network $N=2222$, all 19 4-taxon subsets can be extracted, and thus, all 57 CF equations are defined in $CF(N)$. For the 5-taxon network $N=1112$, the 4-taxon subset $(2,1,1,0)$ cannot be considered since this subset requires two taxa from the clade $n_{0}$ and the 5-taxon network only has one taxon in this clade. Therefore, the 5-taxon network $N=1112$ defines a set of fewer polynomial equations (12 to be exact) in $CF(N)$ involving only the CF values: $a_{7},a_{8},a_{9},a_{22},a_{23},a_{24},a_{28},a_{29},a_{30},a_{31},a_{32}$ and $a_{33}$. To obtain the phylogenetic invariants defined by $N$, denoted $\mathcal{I}(N)$, we get the Gröbner basis of $CF(N)$ on the $a_{i}$ variables using any elimination method in Macaulay2 [18]. All Macaulay2 scripts (and output files) are publicly available in https://github.com/solislemuslab/PhyloDiamond.jl. For example, for the network $N=1112$, there are 10 phylogenetic invariants in the Gröbner basis of $CF(N)$ on the $a_{i}$ variables: 1. 1. $a_{32}-a_{33}$ 2. 2. $a_{31}+2a_{33}-1$ 3. 3. $a_{28}+a_{29}+a_{30}-1$ 4. 4. $a_{23}-a_{24}$ 5. 5. $a_{22}+2a_{24}-1$ 6. 6. $a_{8}-a_{9}$ 7. 7. $a_{7}+2a_{9}-1$ 8. 8. $3a_{9}*a_{30}+a_{9}-a_{24}-a_{33}$ 9. 9. $a_{24}*a_{29}+2a_{24}*a_{30}+a_{29}*a_{33}-a_{30}*a_{33}-a_{33}$ 10. 10. $3a_{9}*a_{29}-2a_{9}+2a_{24}-a_{33}$ We note that the first 7 invariants correspond to the "trivial" invariants related to the sum-to-one property and the equality of the minor concordance factors. The last three invariants, however, identify relationships that the true concordance factors need to satisfy if they originated from the $N=1112$ network. We present the phylogenetic invariants corresponding to each $n-$taxon network ($5\leq n\leq 8$) in the Appendix. ### 2.2 Inference of n-taxon phylogenetic networks with one 4-node hybridization cycle using phylogenetic invariants The procedure to infer a phylogenetic network with one 4-node hybridization cycle using phylogenetic invariants starts with a table of estimated concordance factors. Each value in this table is mapped to the vector of $\begin{bmatrix}a_{1}&\dots&a_{57}\\\ \end{bmatrix}$ and this vector is plugged in the invariants for each candidate network. For example, as mentioned, the candidate network $N=1112$ has 10 phylogenetic invariants, and thus, after evaluation on the observed CFs, we have a $10$-dimensional vector. The invariant score is defined as the $L_{2}$ norm of the vector of evaluated invariants and the candidate network with the smallest invariant score is identified as the network that agrees best with the observed CFs. See Figure 2 for a graphical abstract of the procedure and Algorithm 1 for a more detailed description of the inference steps. Figure 2: Graphical abstract of the inference methodology using phylogenetic invariants to identify the best phylogenetic network to fit the table of observed concordance factors. In this case, there are 8 taxa shown as colored circles. The invariant score is defined as the $L_{2}$ norm of the vector of evaluated invariants. Input : Table of estimated concordance factors; optional: the number of optimal networks ($m$) to return (default $m=5$) Output : Top $m$ optimal networks with smallest invariant score scores $\leftarrow$ an empty array; nets $\leftarrow$ an empty array; for _$N_{i}$ , a possible partition of taxa_ do append(nets,$N_{i}$); map observed CF values to $a_{i}$ values; score $\leftarrow$ L2norm($\mathcal{I}(N_{i})$); append(scores, score); rank(scores, nets); return scores[1:m], nets[1:m]; Algorithm 1 Inference of n-taxon phylogenetic network ($5\leq n\leq 8$) with phylogenetic invariants Given that we only have phylogenetic invariants for the cases of $5\leq n\leq 8$ taxa, if the true network has $5\leq n\leq 8$ taxa, the procedure is just as described (Figure 2 and Algorithm 1). If the true network has more than 8 taxa, then the procedure is slightly different. Namely, our method iterates through all possible subsets of 8 taxa, and fit the phylogenetic invariants on each subset to identify the top subnetwork with $n=8$ taxa. Let $N^{*}_{8}$ be the best 8-taxon network to fit the observed CFs. The missing taxa are added to $N^{*}_{8}$ according their placement in the second best network, or if the missing taxa are not in the second best network, the algorithm adds them according to their placement in the third best or the fourth best and so on. For example, assume that we have $n=9$ taxa (A,B,C,D,E,F,G,H,I) and the best network with $n=8$ taxa ($N^{*}_{8}$) contains A,B,C,D,E,F,G,H in the following partition: A,B in $n_{0}$; C,D in $n_{1}$; E,F in $n_{2}$, and G,H in $n_{3}$. We then find the best network containing taxon I. Assume that this network puts taxon I in the clade $n_{2}$, then the resulting best network would have E,F, and I in $n_{2}$. Note that our algorithm is unable to resolve the bifurcation structure within a given clade (e.g. the $n_{2}$ clade here with E,F, and I), but it is able to 1) correctly identify the partition of taxa among the four clades defined by the 4-node hybridization cycle and 2) correctly identify who are in the hybrid clade ($n_{0}$), who are in the sister clades to the hybrid clade ($n_{1}$ and $n_{2}$), and who are in the remaining fourth clade ($n_{3}$). See Algorithm 2 for the steps in the procedure for $n>8$ taxa. Input : Table of estimated concordance factors; optional: the number of optimal networks ($m$) to return (default $m=5$ Output : Top $m$ optimal networks with smallest invariant score scores $\leftarrow$ an empty array; subnets $\leftarrow$ an empty array; for _$P_{i}$ , a subset of 8 taxa_ do $\text{scores}_{i}$,$\text{subnets}_{i}$ $\leftarrow$ Algorithm 1(CF, $P_{i}$, m=2520); /* Note total number of subnets = 2520 */ append(scores,$\text{scores}_{i}$); append(subnets, $\text{subnets}_{i}$); subnets_sorted $\leftarrow$ sort subnets in descending order according to its score; result $\leftarrow$ an empty array; for _i in 1:length(subnets_sorted)_ do miss_taxa $\leftarrow$ array of missing taxa in subnets_sorted[i]; for _t in miss_taxa_ do for _j in (i+1):length(subnets_sorted)_ do if _t in $n_{0}$ of subnets_sorted[j]_ then add t to $n_{0}$ of subnets_sorted[i]; break; else if _t in $n_{3}$ of subnets_sorted[j]_ then add t to $n_{3}$ of subnets_sorted[i]; break; else if _t in $n_{1}$ of subnets_sorted[j]_ then if _( $n_{2}$ of subnets_sorted[j] == $n_{1}$ of subnets_sorted[i]) & ($n_{1}$ of subnets_sorted[j] without t is in $n_{2}$ of subnets_sorted[i])_ then add t to $n_{2}$ of subnets_sorted[i]; break; else add t to $n_{1}$ of subnets_sorted[i]; break; else if _t in $n_{2}$ of subnets_sorted[j]_ then if _( $n_{1}$ of subnets_sorted[j] == $n_{2}$ of subnets_sorted[i]) & ($n_{2}$ of subnets_sorted[j] without t is in $n_{1}$ of subnets_sorted[i])_ then add t to $n_{1}$ of subnets_sorted[i]; break; else add t to $n_{2}$ of subnets_sorted[i]; break; if _subnets_sorted[i] not in result_ then append(result, subnets_sorted[i]) if _length(result)==m_ then break; return result; Algorithm 2 Inference of n-taxon phylogenetic network ($n>8$) with phylogenetic invariants ### 2.3 Simulation study #### 2.3.1 Networks with eight or fewer taxa In this section, we focus on the case of six taxa ($N=2211,2121,2112,1122,1212,1221$), seven taxa ($N=2221,2212,2122,1222$) and eight taxa ($N=2222$). Later, Section 2.3.2 describes the simulations with more than eight taxa. For a given network $N$, we generate data under four settings: 1) true concordance factors (proof of concept); 2) true concordance factors perturbed by Gaussian error; 3) simulated gene trees, and 4) estimated gene trees. To generate the true concordance factors on a given network in the first scenario, we use the Julia package PhyloNetworks.jl [39], and this scenario is designed to prove that the invariants method works under ideal conditions. For the second setting, we add Gaussian noise to the true concordance factors with zero mean and standard deviation of $\sigma=0.0005,0.00005,0.000005$. We choose Gaussian noise as is standard practice. In the third setting, we use ms-converter [25] that runs ms [22] to simulate $k=100,1000,10000$ gene trees under the coalescent model on a given network $N$. All internal branches in the network are set to 1.0 coalescent unit so that the amount of incomplete lineage sorting is controlled, but not trivial, and we set an inheritance probability parameter of $\gamma=0.3$. For the fourth setting, we simulate $L=500,2000$ sequences on each simulated gene tree with seq-gen [34] under the HKY model, scale the branch lengths by $0.036$, and set nucleotide frequencies as $0.300414$, $0.191363$, $0.196748$, and $0.311475$. Then, we estimate gene trees with IQ-Tree [29] with ModelFinder Plus for model selection [37]. We repeat each simulation setting $30$ times. All simulation scripts are in the GitHub repository https://github.com/solislemuslab/PhyloDiamond.jl. #### 2.3.2 Networks with more than eight taxa We also analyze cases of nine taxa ($N=2223,2232,2322,3222$) and ten taxa ($N=3322,3232,3223,2233,2323,2332$). Given that the algorithm to infer a network with more than 8 taxa relies entirely on the algorithm on 8 taxa, we focus on extensive simulations for the algorithm on 8 taxa in the previous section, and simply show proof-of-concept simulations for the case of more than 8 taxa in this section. In particular, we focus on testing the following scenarios: 1) true concordance factors, and 2) true concordance factors perturbed by Gaussian error as described above ($\sigma=0.0005$). ### 2.4 Reticulate evolution in the Genus Canis We further test our method on the Canis dataset from [17]. The original genomic dataset contained $12$ gray wolves, $14$ dogs, five coyotes, one Ethiopian wolf, three golden jackals, six African golden wolves, two dholes, four African hunting dogs, and one Andean fox. Given the widespread gene flow reported in the original study [17], we need to subsample the taxa so that there is one 4-node hybridization event among the selected taxa: one African hunting dog, one coyote, one dhole, one dog, one golden jackal, and one grey wolf. We use the estimated gene trees from [50] available in https://github.com/chaoszhang/Weighted-ASTRAL_data to infer the concordance factors with the Julia package PhyloNetworks [39]. To better evaluate our method performance, we also run other methods on this dataset for comparison. In addition to our method, we also run SNaQ [38] on $h=1$ hybridization event, 10 independent runs, and using one randomly selected gene tree as the starting tree. We use all 449,450 gene trees to infer the table of CFs and then map the allele names to species names. We also run PhyloNet on two options: 1) maximum likelihood (ML) and 2) maximum pseudolikelihood (MPL). We choose $h=1$ hybridization event on both cases, with 10 independent runs for each. We choose one representative individual per species in the gene trees and removed gene trees that contained fewer than 5 taxa. In total, we use 448,758 gene trees in the PhyloNet analyses. We root the gene trees on the known outgroup (African hunting dog) or on dhole if African hunting dog is not in the gene tree. ## 3 Results ### 3.1 Simulation study on eight or fewer taxa #### 3.1.1 Identifying the correct network as the top 1 ranked network with the smallest invariant score Figure 3 shows the proportion of times (out of the 30 replicates) that our invariants method identifies the true network as the top 1 ranked network with the smallest invariant score for the cases of true concordance factors and the Gaussian-perturbed CFs with increasing standard deviation (from left to right). We also quantify the proportion of times that our invariants method identifies the symmetric network (when clades $n_{1}$ and $n_{2}$ in Figure 1 are switched) because in the 4-node hybridization cycle it is difficult for the method to distinguish clades $n_{1}$ and $n_{2}$ given that they both contribute to the hybridization node. Even for the case of true CFs, our method identifies the symmetric network as the top 1 network, instead of the true network, in two cases ($N=1221$ and $N=2121$), yet the difference in invariance score ($L_{2}$-norm of evaluated polynomials) of the true and symmetric networks in these cases is of the order of $10^{-16}$. For the noisiest setting ($\sigma=0.0005$, far right), it is evident that some networks are difficult to be detected by our method, but it is worth highlighting that whenever there are two taxa sampled on each of the four clades ($n_{0},n_{1},n_{2},n_{3}$) as in $N=2222$, our method always identifies the true (or symmetric) network as the top 1. Figure 3: Proportion of times that the true network or the symmetric network (with clades $n_{1}$ and $n_{2}$ inverted) are the top 1 ranked networks identified by the phylogenetic invariants on the cases of true and Gaussian- perturbed CFs. Y-axis is between 0 and 100. Each panel corresponds to a type of simulation: using true concordance factors (left) and using concordance factors with added Gaussian noise (with increasing standard deviation for noise from left to right). Whenever there are at least two taxa in each of the four clades (network $2222$), our method is very accurate in detecting the true (or its symmetric) network. Our method is also accurate to identify the true network in the top 5 ranked networks (see Figure 6) which will greatly reduce the space of candidate networks to be compared with a likelihood approach. Figure 4 shows the proportion of times (out of the 30 replicates) that our invariants method identifies the true network as the top 1 ranked network with smallest invariant score for the case of true simulated gene trees with increasing number of gene trees (from left to right), where “g.t." is abbreviation for “gene trees". Unlike the case of true or Gaussian-perturbed CFs, our method only identifies the true (or symmetric) network when there are 2 taxa sampled per clade ($N=2222$) or at least 2 taxa sampled on the hybrid clade and sister clades ($N=2221$). With 1,000 gene trees or more, the networks $N=2121$ and $N=2211$ are also accurately identified by our method. Figure 4: Proportion of times that the true network or the symmetric network (with clades $n_{1}$ and $n_{2}$ inverted) are the top 1 ranked networks identified by the phylogenetic invariants for the case of true simulated gene trees (“g.t."). Y-axis is between 0 and 100. Whenever there are at least two taxa in each of the four clades (network $2222$), our method is very accurate in detecting the true (or its symmetric) network. Our method is also accurate to identify the true network in the top 5 ranked networks (see Figure 7) which will greatly reduce the space of candidate networks to be compared with a likelihood approach. Figure 5 shows the proportion of times (out of the 30 replicates) that our invariants method identifies the true network as the top 1 ranked network with smallest invariant score for the case of estimated gene trees with increasing number of gene trees and sequence length (from left to right). Again, our method only identifies the true (or symmetric) network when there are 2 taxa sampled per clade ($N=2222$) or at least 2 taxa sampled on the hybrid clade and sister clades ($N=2221$). As before, with 1,000 gene trees, the networks $N=2121$ and $N=2211$ are also accurately identified by our method, and with 10,000 gene trees, the networks $N=1221$ and $N=1222$ are also accurately identified. Summarizing all figures related to our method’s ability to identify the true (or symmetric) network as top 1 ranked network by its invariant score (Figures 3, 4, 5), we can highlight that networks with only one sampled taxon in some of the clades are harder to be identified. As long as two taxa are sampled from each of the clades, our method is able to identify the network with high accuracy on all simulation settings. The number of genes is more important than the sequence length, and with at least 100 genes, our method is able to correctly identify networks with fewer than two taxa sampled from some clades (like $2221$). We highlight that we are equally interested in the true and symmetric networks since a follow-up optimization of branch lengths and inheritance probability ($\gamma$) on the fixed network will estimate the correct $\gamma$ and identify which of the sister clades is the major ($\gamma>0.5$) and which is the minor ($\gamma<0.5$). Figure 5: Proportion of times that the true network or the symmetric network (with clades $n_{1}$ and $n_{2}$ inverted) are the top 1 ranked networks identified by the phylogenetic invariants for the case of estimated gene trees. Y-axis is between 0 and 100. Each panel corresponds to a number of gene trees (g.t. from 1000 to 10,000) and sequence length ($L$ from 500 to 2000). Whenever there are at least two taxa in each of the four clades (network $2222$) or only clade $n_{3}$ having one taxon ($N=2221$), our method is very accurate in detecting the true (or its symmetric) network. Our method is also accurate to identify the true network in the top 5 ranked networks (see Figure 8) which will greatly reduce the space of candidate networks to be compared with a likelihood approach. #### 3.1.2 Reducing the space of candidate networks using the invariant score Figure 6 shows the proportion of times (out of the 30 replicates) that the true (or symmetric) network are within the top 5 ranked networks based on smallest invariant score for the cases of true concordance factors and the Gaussian-perturbed CFs with increasing standard deviation (from left to right). In all cases, our method includes the true (and symmetric) networks within the top 5 which means that our method accurately and in a fast manner reduces the space of candidate networks to just 5 candidates that can later be tested with another accurate methodology like likelihood. Figure 6: Proportion of times that the true network or the symmetric network (with clades $n_{1}$ and $n_{2}$ inverted) are in the top 5 ranked networks identified by the phylogenetic invariants on the cases of true and Gaussian- perturbed CFs. Y-axis is between 0 and 100. Each panel corresponds to a type of simulation: using true concordance factors (left) and using concordance factors with added Gaussian noise (with increasing standard deviation for noise from left to right). All networks are accurately identified in the top 5 ranked based on invariant score which provides evidence that our method fast and accurately reduce the space of candidate networks to just 5 alternatives. Figure 7 shows the proportion of times (out of the 30 replicates) that the true (or symmetric) network are within the top 5 ranked networks based on smallest invariant score for the case of true simulated gene trees. Only the networks with one sampled taxon on the hybrid clade ($n_{0}$) cannot be recovered with 100 or 1000 gene trees. Whenever there are 10,000 gene trees in the sample, the networks $N=1221$ and $N=1222$ can now be recovered as part of the top 5, yet the networks $N=1122$ or $N=1212$ (when there is one sampled taxon in the hybrid clade and one of the sister clades) still refuse detection. The same conclusions are true for the case of estimated gene trees (Figure 8) with number of gene trees having more weight on the accuracy than sequence length. Figure 7: Proportion of times that the true network or the symmetric network (with clades $n_{1}$ and $n_{2}$ inverted) are in the top 5 ranked networks identified by the phylogenetic invariants for the case of true simulated gene trees (“g.t."). Y-axis is between 0 and 100. Only the networks that have one taxon below the hybrid node (1122, 1212, 1221, 1222) do not allow accurate reconstruction with fewer than 10,000 gene trees which brings into attention the importance of taxon sampling for this method. Summarizing all figures related to our method’s ability to identify the true (or symmetric) network within the top 5 ranked networks by its invariant score (Figures 6, 7, 8), we can highlight that as long as there are at least 2 taxa sampled from the hybrid clade, then our method is able to place the true (and symmetric) network within the top 5 networks. This effectively reduces the space of candidate networks to compare with a likelihood approach. As mentioned, we do not view our method as a substitute to existing network inference methods. On the contrary, we believe that our invariants method will serve as a complement to identify a small subset of network possibilities that will help bypass the optimization on network space. Figure 8: Proportion of times that the true network or the symmetric network (with clades $n_{1}$ and $n_{2}$ inverted) are in the top 5 ranked networks identified by the phylogenetic invariants for the case of estimated gene trees. Y-axis is between 0 and 100. Each panel corresponds to a number of gene trees (g.t. from 1000 to 10,000) and sequence length ($L$ from 500 to 2000). Only the networks that have one taxon below the hybrid node (1122, 1212, 1221, 1222) do not allow accurate reconstruction with fewer than 10,000 gene trees which brings into attention the importance of taxon sampling for this method. Figure 9 shows the rank in the invariant scores of the true and symmetric networks on the cases of true and Gaussian-perturbed CFs. Ideally, the true network (or at least its symmetric version) should be ranked as 1 by our invariants method. While the true and symmetric networks are not always ranked in number 1, they are ranked within the top 5 for most of the cases. This further confirms the advantage of our current method. It is a fast way to reduce the space of candidate networks (to 5) that can later be tested using a likelihood approach. Figure 9: Rank (y-axis) in the invariant score for each network (x-axis) for the true network (red) and its symmetric network (inverted clades $n_{1}$ and $n_{2}$, in blue) on the cases of true and Gaussian-perturbed CFs. Dashed line corresponds to rank 5. Each panel corresponds to a type of simulation: using true concordance factors (left) and using concordance factors with added Gaussian noise (with increasing standard deviation for noise from left to right). Both the true and symmetric networks are within the top 5 ranked networks by the method in all cases, and are thus, easy to distinguish from wrong networks. Figure 10 shows the rank in the invariant scores of the true and symmetric networks for the case of true simulated gene trees (increased number of gene trees from left to right). The true (and symmetric) networks are not within the top 5 networks only for the cases of one sampled taxon from the hybrid clade ($N=1122,1212,1221,1222$). This same behavior is evident in Figure 11 for estimated gene trees. Again, these figures show our method’s ability to reduce the possible networks that fit the data. Figure 10: Rank (y-axis) in the invariant score for each network (x-axis) for the true network (red) and its symmetric network (inverted clades $n_{1}$ and $n_{2}$, in blue) in the case of true simulated gene trees (“g.t"). Each panel corresponds to a number of simulated gene trees from 100 (left) to 10,000 (right). Both the true and symmetric networks are within the top 5 ranked networks by the method as the number of genes increases, and are thus, easy to distinguish from wrong networks. Only the networks that have one taxon below the hybrid node (1122, 1212, 1221, 1222) do not allow accurate reconstruction which brings into attention the importance of taxon sampling for this method. Figure 11: Rank (y-axis) in the invariant score for each network (x-axis) for the true network (red) and its symmetric network (inverted clades $n_{1}$ and $n_{2}$, in blue) in the case of estimated gene trees. Each panel corresponds to a number of gene trees (g.t. from 1000 to 10,000) and sequence length ($L$ from 500 to 2000). Only the networks that have one taxon below the hybrid node (1122, 1212, 1221, 1222) do not allow accurate reconstruction which brings into attention the importance of taxon sampling for this method. In the Appendix, we also show plots with the values of the invariant scores for the true and symmetric networks. Last, we present a comparison of the running times of four network methods in Table 1. While all methods are able to identify the correct network, our phylogenetic invariants method (top row) only takes 7.17 seconds to correctly infer the network $N=2222$ which is over twice as fast as the second fastest (PhyloNet MPL [49]). The difference in running times (and even accuracy) is more evident in the results on the Canis dataset (Section 3.3). Method | Time (seconds) ---|--- Phylogenetic invariants (our method) | 7.17 SNaQ | 80.23 PhyloNet ML | 84.34 PhyloNet MPL | 17.99 Table 1: Running times (in seconds) on the inference of network $N=2222$ from 100 true simulated gene trees by four network methods: 1) our phylogenetic invariants (top row); 2) SNaQ [38]; 3) PhyloNet ML [48], and 4) PhyloNet MPL [49]. ### 3.2 Simulation study on more than eight taxa Table 2 shows the proof-of-concept results on simulated data with more than 8 taxa. Since the algorithm for more than 8 taxa (Algorithm 2) relies entirely on the algorithm for 8 or fewer taxa (Algorithm 1), we simply show here that our steps to append taxon to the optimal subset of 8 taxa identified by the method indeed yields the desired network with more than 8 taxa. Network | True CF | Noisy CF ---|---|--- 2223 | 1 | 1 2232 | 1 | 1 (symmetric) 2322 | 1 | 2 3222 | 1 | 1 2233 | 1 | 1 (symmetric) 2323 | 1 | 2 (symmetric) 3223 | 1 | 1 2332 | 1 | 2 (symmetric) 3232 | 1 | 2 (symmetric) 3322 | 1 | 1 (symmetric) Table 2: Rank of the true (or symmetric) network under the two types of simulations: true CFs (left column) and Gaussian-perturbed CFs (right column) for a level of noise of $\sigma=0.0005$. In all cases, either the true network or the symmetric networks are within the top 2 of networks identified by the method which proves that our algorithm for more than 8 taxa that builds on Algorithm 1 works appropriately. ### 3.3 Phylogenetic network for the Canis genus The original analysis of the Canis genus identified nine gene flow events using the D statistics [32] (Figure 3a in [17]). Here, we are able to replicate the hybridization event involving the ancestor of dog and grey wolf into golden jackal (Figure 12 top). The same network is correctly inferred by SNaQ and by PhyloNet ML, but both methods take much longer to run (Table 3). PhyloNet MPL identifies a different hybridization event that has not been reported in previous studies, so it is impossible to validate it at this point. This hybridization involves an unsampled or extinct taxon (Figure 12 bottom). Figure 12: Left top: Semi-directed phylogenetic network estimated by SNaQ [38], PhyloNet ML [48], and our method based on phylogenetic invariants. Left bottom: Semi-directed phylogenetic network estimated by PhyloNet MPL [49]. The star marks the place where the root should be. Right: Rooted version of the semi-directed networks. The network estimated with PhyloNet MPL (bottom) is different from the network inferred by the other methods, and it involves an unsampled (or extinct) taxon in the hybridization. SNaQ, PhyloNet ML and our phylogenetic invariants method (top) identify the same hybridization event as the original publication [17] between the ancestral species to dog and grey wolf and the ancestral species of golden jackal, yet PhyloNet ML takes 389 times longer and SNaQ takes 20 times longer than our invariants method (see running times in Table 3). In terms of running time, our invariants method is able to identify the correct hybridization in under 7 seconds while SNaQ (which also identifies the correct hybridization) takes over 140 seconds. PhyloNet MPL takes 40 times longer than our invariants method (with running time of over 281 seconds) and identifies a different hybridization event to those originally published (so, not possible to validate at this point). PhyloNet ML is the slowest method with over 45 minutes of running time on this small network of only 6 taxa. The inferred network by PhyloNet ML also agrees with the one originally published in [17]. Method | Time (seconds) ---|--- Phylogenetic invariants (our method) | 6.78 SNaQ | 140.58 PhyloNet ML | 2723.99 PhyloNet MPL | 281.25 Table 3: Running times (in seconds) on the Canis dataset. Our method based on evaluation of phylogenetic invariants is 20 times faster than the second fastest method (SNaQ). ## 4 Conclusion Here, we introduce a novel method to infer 4-node hybridization cycles on level-1 phylogenetic networks of over five taxa based on ultrafast evaluations of phylogenetic invariants. Our method bypasses optimization on network space and is able to accurately detect the hybridization cycles based on simulated and real data, especially when there are at least 2 taxa sampled from each of the four clades defined by the hybridization cycle (Figure 1 left). While our method is at least $\sim 10$ times faster than other existing network methods, it is not meant to replace existing methods. On the contrary, we believe that our ultrafast algorithm can work in conjunction with existing methodologies by reducing the space of candidate networks that can later be evaluated based on likelihood or Bayesian approaches. Nevertheless, there are limitations in our current method. Specifically, we are only able to detect hybridization cycles with 4 nodes on level-1 networks. For the case of three nodes in the hybridization cycle, it has been investigated already in [38] that the set of phylogenetic invariants is empty. However, there is still room to extend the current methodology to hybridization cycles of five nodes or more. In addition, even when our method can infer networks with any number of taxa, it is unable to resolve the topology inside any of the four clades defined by the hybridization cycle (Figure 1 left). That is, our method is only able to identify the placement of the 4-node hybridization cycle. For this reason, we view our method as in line with other hybrid detection methods such as ABBA-BABA test [32], MSCquartets [30, 36], and HyDe [27] that are also only able to detect specific hybridization patterns in subsets of taxa. Future work will involve the generation of the phylogenetic invariants related to hybridization cycles of five nodes or more, and the development of a merging algorithm that can produced a fully resolved $n$-taxon phylogenetic network from the 8-taxon estimated networks inferred by our invariants method. In addition, our proposed method is unable to infer biological parameters like inheritance probability or branch length, but future work could exploit the CF formulas (which depend on branch lengths and inheritance probabilities) to provide estimated values to these parameters from the observed CFs. ## 5 Data and code availability The Canis dataset was made publicly available by the original publication [17] and can be accessed through the GitHub repository of [50] here https://github.com/chaoszhang/Weighted-ASTRAL_data. All the scripts for our work are publicly available in the GitHub repository https://github.com/solislemuslab/PhyloDiamond.jl. The new Julia package PhyloDiamond.jl is open-source, publicly available in https://github.com/solislemuslab/PhyloDiamond.jl. ## 6 Acknowledgements This work was supported by the National Science Foundation [DEB-2144367 to CSL]. This was also funded by the UW-Madison Fall Competition [to CSL]. The authors thank Marianne Bjørner and Nathan Kolbow for help with phylogenetics code. ## References * [1] Roya Adavoudi and Małgorzata Pilot. Consequences of hybridization in mammals: A systematic review. Genes, 13(1):50, 2021. * [2] Elizabeth S Allman and John A Rhodes. Phylogenetic invariants for the general markov model of sequence mutation. Mathematical biosciences, 186(2):113–144, 2003. * [3] Muhammad Ardiyansyah. Distinguishing level-2 phylogenetic networks using phylogenetic invariants. arXiv preprint arXiv:2104.12479, 2021. * [4] Travis Barton, Elizabeth Gross, Colby Long, and Joseph Rusinko. Statistical learning with phylogenetic network invariants. arXiv, 2022. * [5] David A Baum. Concordance trees, concordance factors, and the exploration of reticulate genealogy. Taxon, 56(May):417–426, 2007. * [6] Marianne Bjorner, Erin K Molloy, Colin N Dewey, and Claudia Solis-Lemus. Detectability of varied hybridization scenarios using Genome-Scale hybrid detection methods. November 2022. * [7] Christopher Blair and Cécile Ané. Phylogenetic trees and networks can serve as powerful and complementary approaches for analysis of genomic data. Systematic biology, 69(3):593–601, 2020. * [8] David Bryant and Vincent Moulton. Neighbor-net: an agglomerative method for the construction of phylogenetic networks. Molecular biology and evolution, 21(2):255–265, 2004. * [9] Kimberly M Charles and Ivana Stehlik. Assisted species migration and hybridization to conserve cold-adapted plants under climate change. Conservation Biology, 35(2):559–566, 2021. * [10] Joseph Cummings, Benjamin Hollering, and Christopher Manon. Invariants for level-1 phylogenetic networks under the cavendar-farris-neyman model. arXiv preprint arXiv:2102.03431, 2021. * [11] Vanessa De Santis, Silvia Quadroni, Robert J Britton, Antonella Carosi, Catherine Gutmann Roberts, Massimo Lorenzoni, Giuseppe Crosa, and Serena Zaccara. Biological and trophic consequences of genetic introgression between endemic and invasive barbus fishes. Biological invasions, 23(11):3351–3368, 2021. * [12] James H Degnan. Modeling Hybridization Under the Network Multispecies Coalescent. Systematic Biology, 67(5):786–799, 05 2018. * [13] Awa Diop, Ellis L Torrance, Caroline M Stott, and Louis-Marie Bobay. Gene flow and introgression are pervasive forces shaping the evolution of bacterial species. Genome Biology, 23(1):1–19, 2022. * [14] Melania D’Angiolo, Matteo De Chiara, Jia-Xing Yue, Agurtzane Irizar, Simon Stenberg, Karl Persson, Agnès Llored, Benjamin Barré, Joseph Schacherer, Roberto Marangoni, et al. A yeast living ancestor reveals the origin of genomic introgressions. Nature, 587(7834):420–425, 2020. * [15] RA Elworth, Huw A Ogilvie, Jiafan Zhu, and Luay Nakhleh. Advances in computational methods for phylogenetic networks in the presence of hybridization. In Bioinformatics and phylogenetics, pages 317–360. Springer, 2019\. * [16] Joseph Felsenstein. Counting phylogenetic invariants in some simple cases. Journal of Theoretical Biology, 152(3):357–376, 1991. * [17] Shyam Gopalakrishnan, Mikkel-Holger S. Sinding, Jazmín Ramos-Madrigal, Jonas Niemann, Jose A. Samaniego Castruita, Filipe G. Vieira, Christian Carøe, Marc de Manuel Montero, Lukas Kuderna, Aitor Serres, Víctor Manuel González-Basallote, Yan-Hu Liu, Guo-Dong Wang, Tomas Marques-Bonet, Siavash Mirarab, Carlos Fernandes, Philippe Gaubert, Klaus-Peter Koepfli, Jane Budd, Eli Knispel Rueness, Claudio Sillero, Mads Peter Heide-Jørgensen, Bent Petersen, Thomas Sicheritz-Ponten, Lutz Bachmann, Øystein Wiig, Anders J. Hansen, and M. Thomas P. Gilbert. Interspecific gene flow shaped the evolution of the genus canis. Current Biology, 28(3), 2018. * [18] DR Grayson and ME Stillman. Macaulay2, a software system for research in algebraic geometry. Available at http://www.math.uiuc.edu/Macaulay2/. * [19] Elizabeth Gross and Colby Long. Distinguishing phylogenetic networks. SIAM Journal on Applied Algebra and Geometry, 2(1):72–93, 2018\. * [20] Elizabeth Gross, Leo van Iersel, Remie Janssen, Mark Jones, Colby Long, and Yukihiro Murakami. Distinguishing level-1 phylogenetic networks on the basis of data generated by markov processes. Journal of Mathematical Biology, 83(3):1–24, 2021. * [21] Stefan Grünewald, Kristoffer Forslund, Andreas Dress, and Vincent Moulton. Qnet: an agglomerative method for the construction of phylogenetic networks from weighted quartets. Molecular Biology and Evolution, 24(2):532–538, 2007. * [22] Richard R. Hudson. Generating samples under a wright–fisher neutral model of genetic variation. Bioinformatics, 18(2):337–338, February 2002. * [23] Daniel Huson, Regula Rupp, and Celine Scornavacca. Phylogenetic Networks. Cambridge University Press, New York, NY, first edition, 2010. * [24] Daniel H Huson and David Bryant. Application of phylogenetic networks in evolutionary studies. Molecular biology and evolution, 23(2):254–267, 2006. * [25] Nathan Kolbow. Ms-converter. https://github.com/NathanKolbow/msConverter, 2022. * [26] Sungsik Kong, Joan Carles Pons, Laura Kubatko, and Kristina Wicke. Classes of explicit phylogenetic networks and their biological and mathematical significance. Journal of Mathematical Biology, 84(6):1–44, 2022. * [27] Laura S. Kubatko and Julia Chifman. An invariants-based method for efficient identification of hybrid species from large-scale genomic data. BMC Evolutionary Biology, 19(1):112, 2019. * [28] James Mcinerney, Eric Bapteste, Leo van Iersel, Axel Janke, Scot Kelchner, Steven Kelk, James O McInerney, David A Morrison, Luay Nakhleh, Mike Steel, et al. Networks: expanding evolutionary thinking. Trends in genetics: TIG, 29(8), 2013. * [29] Bui Quang Minh, Heiko A Schmidt, Olga Chernomor, Dominik Schrempf, Michael D Woodhams, Arndt von Haeseler, and Robert Lanfear. Iq-tree 2: New models and efficient methods for phylogenetic inference in the genomic era. Molecular Biology and Evolution, 37(5):1530–1534, May 2020. * [30] Jonathan D Mitchell, Elizabeth S Allman, and John A Rhodes. Hypothesis testing near singularities and boundaries. Electron J Stat, 13(1):2150–2193, 2019. * [31] Xiao-Xu Pang and Da-Yong Zhang. Impact of ghost introgression on coalescent-based species tree inference and estimation of divergence time. January 2022. * [32] Nick Patterson, Priya Moorjani, Yontao Luo, Swapan Mallick, Nadin Rohland, Yiping Zhan, Teri Genschoreck, Teresa Webster, and David Reich. Ancient admixture in human history. Genetics, 192(3):1065–1093, 11 2012. * [33] Oscar A Pérez-Escobar, Sidonie Bellot, Natalia AS Przelomska, Jonathan M Flowers, Mark Nesbitt, Philippa Ryan, Rafal M Gutaker, Muriel Gros-Balthazard, Tom Wells, Benedikt G Kuhnhäuser, et al. Molecular clocks and archeogenomics of a late period egyptian date palm leaf reveal introgression from wild relatives and add timestamps on the domestication. Molecular biology and evolution, 38(10):4475–4492, 2021. * [34] Andrew Rambaut and Nicholas C. Grass. Seq-gen: an application for the monte carlo simulation of dna sequence evolution along phylogenetic trees. Bioinformatics, 13(3):235–238, June 1997. * [35] Olivier Rey, Eve Toulza, Cristian Chaparro, Jean-François Allienne, Julien Kincaid-Smith, Eglantine Mathieu-Begné, Fiona Allan, David Rollinson, Bonnie L Webster, and Jérôme Boissier. Diverging patterns of introgression from schistosoma bovis across s. haematobium african lineages. PLoS pathogens, 17(2):e1009313, 2021. * [36] John A Rhodes, Hector Baños, Jonathan D Mitchell, and Elizabeth S Allman. Mscquartets 1.0: quartet methods for species trees and networks under the multispecies coalescent model in r. Bioinformatics, 37(12):1766–1768, Jul 2021. * [37] Kalyaanamoorthy S, Minh BQ, Wong TKF, von Haeseler A, and Jermiin LS. Modelfinder: fast model selection for accurate phylogenetic estimates. Nat Methods, 14(6):587–589, June 2017. * [38] Claudia Solís-Lemus and Cécile Ané. Inferring phylogenetic networks with maximum pseudolikelihood under incomplete lineage sorting. PLoS Genet., 12(3):e1005896, March 2016. * [39] Claudia Solís-Lemus, Paul Bastide, and Cécile Ané. PhyloNetworks: A package for phylogenetic networks. 34(12):3292–3298, December 2017. * [40] Claudia Solis-Lemus, Arrigo Coen, and Cecile Ane. On the identifiability of phylogenetic networks under a pseudolikelihood model. October 2020. * [41] Claudia Solís-Lemus, Mengyao Yang, and Cécile Ané. Inconsistency of species tree methods under gene flow. Syst. Biol., 65(5):843–851, September 2016. * [42] Mike A Steel and YX Fu. Classifying and counting linear phylogenetic invariants for the jukes–cantor model. Journal of computational biology, 2(1):39–47, 1995. * [43] Jan Steensels, Brigida Gallone, and Kevin J Verstrepen. Interspecific hybridization as a driver of fungal evolution and adaptation. Nature Reviews Microbiology, 19(8):485–500, 2021. * [44] Anton Suvorov, Celine Scornavacca, M Stanley Fujimoto, Paul Bodily, Mark Clement, Keith A Crandall, Michael F Whiting, Daniel R Schrider, and Seth M Bybee. Deep ancestral introgression shapes evolutionary history of dragonflies and damselflies. Systematic Biology, 71(3):526–546, 2022. * [45] Théo Tricou, Eric Tannier, and Damien M de Vienne. Ghost lineages highly influence the interpretation of introgression tests. Syst. Biol., February 2022. * [46] Dingqiao Wen, Yun Yu, and Luay Nakhleh. Bayesian inference of reticulate phylogenies under the multispecies network coalescent. PLOS Genetics, 12(5):e1006006–, 05 2016. * [47] Yun Yu, James H. Degnan, and Luay Nakhleh. The probability of a gene tree topology within a phylogenetic network with applications to hybridization detection. PLoS genetics, 8(4):e1002660, jan 2012. * [48] Yun Yu, Jianrong Dong, Kevin J Liu, and Luay Nakhleh. Maximum Likelihood Inference of Reticulate Evolutionary Histories. PNAS, 111(46):16448–16453, 2014. * [49] Yun Yu and Luay Nakhleh. A maximum pseudo-likelihood approach for phylogenetic networks. BMC genomics, 16(10):1–10, 2015. * [50] Chao Zhang and Siavash Mirarab. Weighting by Gene Tree Uncertainty Improves Accuracy of Quartet-based Species Trees. Molecular Biology and Evolution, 10 2022. msac215. * [51] Chao Zhang, Maryam Rabiee, Erfan Sayyari, and Siavash Mirarab. Astral-iii: polynomial time species tree reconstruction from partially resolved gene trees. BMC Bioinformatics, 19(6):153, 2018. ## Appendix A Representation of a level-1 semi-directed phylogenetic network with a set of polynomial equations corresponding to the expected concordance factors under the coalescent model Notation: * • $n$ corresponds to the number of individuals from each of the 4 clades: $(n_{0},n_{1},n_{2},n_{3})$. For example, $n=(0,0,2,2)$ means that you have the quartet with 2 individuals in $n_{2}$ and 2 individuals in $n_{3}$ * • "Individuals" list the individuals taken from each of the four clades. For example, $i_{1},i_{2}\in n_{0}$ means that we took two species from clade $n_{0}$ denoted $i_{1}$ and $i_{2}$. We use these species names to define the splits for the formulas: e.g. $P(i_{1},i_{2}|j_{1},j_{2})$ represents the probability that $i_{1}$ and $i_{2}$ are together in one side of the split (and $j_{1},j_{2}$ together) * • "Type" corresponds to the type of quartet which matches the types in [38] * • "CF Formula" corresponds to the formula of the expected CF for that given quartet under the multispecies coalescent model * • "CF value" is the variable we give to the observed CF we will read from the data table By [40], we know that we only need to select at most two individuals per subgraph $(n_{0},n_{1},n_{2},n_{3})$ to define all the CF formulas that involve the hybridization cycle. We list all the CF formulas in the table below assuming that we do have 2 individuals per clade $(n_{0},n_{1},n_{2},n_{3})$ (that is, we have at least eight species). Table 4: Concordance factor equations for all 4-taxon subsets that involve the hybridization cycle. $n$ | Type | CF Formula | CF value ---|---|---|--- $(0,0,2,2)$ | $5$ | $1-\dfrac{2}{3}z_{2}z_{2,3}z_{3}$ | $a1$ | | $\dfrac{1}{3}z_{2}z_{2,3}z_{3}$ | $a2$ | | $\dfrac{1}{3}z_{2}z_{2,3}z_{3}$ | $a3$ $(0,1,2,1)$ | $5$ | $1-\dfrac{2}{3}z_{2,3}z_{2}$ | $a4$ | | $\dfrac{1}{3}z_{2,3}z_{2}$ | $a5$ | | $\dfrac{1}{3}z_{2,3}z_{2}$ | $a6$ $(0,1,1,2)$ | $5$ | $1-\dfrac{2}{3}z_{3}$ | $a7$ | | $\dfrac{1}{3}z_{3}$ | $a8$ | | $\dfrac{1}{3}z_{3}$ | $a9$ $(0,2,2,0)$ | $5$ | $1-\dfrac{2}{3}z_{2}z_{2,3}z_{1,3}z_{1}$ | $a10$ | | $\dfrac{1}{3}z_{2}z_{2,3}z_{1,3}z_{1}$ | $a11$ | | $\dfrac{1}{3}z_{2}z_{2,3}z_{1,3}z_{1}$ | $a12$ $(0,2,1,1)$ | $5$ | $1-\dfrac{2}{3}z_{1,3}z_{1}$ | $a13$ | | $\dfrac{1}{3}z_{1,3}z_{1}$ | $a14$ | | $\dfrac{1}{3}z_{1,3}z_{1}$ | $a15$ $(0,2,0,2)$ | $5$ | $1-\dfrac{2}{3}z_{3}z_{1,3}z_{1}$ | $a16$ | | $\dfrac{1}{3}z_{3}z_{1,3}z_{1}$ | $a17$ | | $\dfrac{1}{3}z_{3}z_{1,3}z_{1}$ | $a18$ $(1,0,2,1)$ | $2$ | $(1-\gamma)\left(1-\dfrac{2}{3}z_{2,3}z_{2}\right)+\gamma\left(1-\dfrac{2}{3}z_{2}\right)$ | $a19$ | | $(1-\gamma)\dfrac{1}{3}z_{2,3}z_{2}+\gamma\dfrac{1}{3}z_{2}$ | $a20$ | | $(1-\gamma)\dfrac{1}{3}z_{2,3}z_{2}+\gamma\dfrac{1}{3}z_{2}$ | $a21$ $(1,0,1,2)$ | $2$ | $(1-\gamma)\left(1-\dfrac{2}{3}z_{3}\right)+\gamma\left(1-\dfrac{2}{3}z_{2,3}z_{3}\right)$ | $a22$ | | $(1-\gamma)\dfrac{1}{3}z_{3}+\gamma\dfrac{1}{3}z_{2,3}z_{3}$ | $a23$ | | $(1-\gamma)\dfrac{1}{3}z_{3}+\gamma\dfrac{1}{3}z_{2,3}z_{3}$ | $a24$ $(1,1,2,0)$ | $2$ | $(1-\gamma)\left(1-\dfrac{2}{3}z_{1,3}z_{2,3}z_{2}\right)+\gamma\left(1-\dfrac{2}{3}z_{2}\right)$ | $a25$ | | $(1-\gamma)\dfrac{1}{3}z_{1,3}z_{2,3}z_{2}+\gamma\dfrac{1}{3}z_{2}$ | $a26$ | | $(1-\gamma)\dfrac{1}{3}z_{1,3}z_{2,3}z_{2}+\gamma\dfrac{1}{3}z_{2}$ | $a27$ $(1,1,1,1)$ | $3$ | $(1-\gamma)\left(1-\dfrac{2}{3}z_{1,3}\right)+\gamma\dfrac{1}{3}z_{2,3}$ | $a28$ | | $(1-\gamma)\dfrac{1}{3}z_{1,3}+\gamma\left(1-\dfrac{2}{3}z_{2,3}\right)$ | $a29$ | | $(1-\gamma)\dfrac{1}{3}z_{1,3}+\gamma\dfrac{1}{3}z_{2,3}$ | $a30$ $(1,1,0,2)$ | $2$ | $(1-\gamma)\left(1-\dfrac{2}{3}z_{1,3}z_{3}\right)+\gamma\left(1-\dfrac{2}{3}z_{3}\right)$ | $a31$ | | $(1-\gamma)\dfrac{1}{3}z_{1,3}z_{3}+\gamma\dfrac{1}{3}z_{3}$ | $a32$ | | $(1-\gamma)\dfrac{1}{3}z_{1,3}z_{3}+\gamma\dfrac{1}{3}z_{3}$ | $a33$ $(1,2,1,0)$ | $2$ | $(1-\gamma)\left(1-\dfrac{2}{3}z_{1}\right)+\gamma\left(1-\dfrac{2}{3}z_{2,3}z_{1,3}z_{1}\right)$ | $a34$ | | $(1-\gamma)\dfrac{1}{3}z_{1}+\gamma\dfrac{1}{3}z_{2,3}z_{1,3}z_{1}$ | $a35$ | | $(1-\gamma)\dfrac{1}{3}z_{1}+\gamma\dfrac{1}{3}z_{2,3}z_{1,3}z_{1}$ | $a36$ $(1,2,0,1)$ | $2$ | $(1-\gamma)\left(1-\dfrac{2}{3}z_{1}\right)+\gamma\left(1-\dfrac{2}{3}z_{1,3}z_{1}\right)$ | $a37$ | | $(1-\gamma)\dfrac{1}{3}z_{1}+\gamma\dfrac{1}{3}z_{1,3}z_{1}$ | $a38$ | | $(1-\gamma)\dfrac{1}{3}z_{1}+\gamma\dfrac{1}{3}z_{1,3}z_{1}$ | $a39$ $(2,0,2,0)$ | $4$ | $(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{2}z_{0}z_{0,1}z_{1,3}z_{2,3}\right)+2\gamma(1-\gamma)\left(1-\dfrac{2}{3}z_{2}z_{0}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{2}z_{0}z_{0,2}\right)$ | $a40$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{2}z_{0}z_{0,1}z_{1,3}z_{2,3}+2\gamma(1-\gamma)\dfrac{1}{3}z_{2}z_{0}+\gamma^{2}\dfrac{1}{3}z_{2}z_{0}z_{0,2}$ | $a41$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{2}z_{0}z_{0,1}z_{1,3}z_{2,3}+2\gamma(1-\gamma)\dfrac{1}{3}z_{2}z_{0}+\gamma^{2}\dfrac{1}{3}z_{2}z_{0}z_{0,2}$ | $a42$ $(2,0,1,1)$ | $1$ | $(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{0}z_{1,3}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-z_{0}+\dfrac{1}{3}z_{0}z_{2,3}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,2}\right)$ | $a43$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{1,3}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{2,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}$ | $a44$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{1,3}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{2,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}$ | $a45$ $(2,0,0,2)$ | $4$ | $(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{3}z_{0}z_{1,3}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-\dfrac{2}{3}z_{3}z_{0}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{3}z_{0}z_{2,3}z_{0,2}\right)$ | $a46$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{3}z_{0}z_{1,3}z_{0,1}+2\gamma(1-\gamma)\dfrac{1}{3}z_{3}z_{0}+\gamma^{2}\dfrac{1}{3}z_{3}z_{0}z_{2,3}z_{0,2}$ | $a47$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{3}z_{0}z_{1,3}z_{0,1}+2\gamma(1-\gamma)\dfrac{1}{3}z_{3}z_{0}+\gamma^{2}\dfrac{1}{3}z_{3}z_{0}z_{2,3}z_{0,2}$ | $a48$ $(2,1,1,0)$ | $1$ | $(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-z_{0}+\dfrac{1}{3}z_{0}z_{2,3}z_{1,3}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,2}\right)$ | $a49$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{2,3}z_{1,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}$ | $a50$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{2,3}z_{1,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}$ | $a51$ $(2,1,0,1)$ | $1$ | $(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-z_{0}+\dfrac{1}{3}z_{0}z_{1,3}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,2}z_{2,3}\right)$ | $a52$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{1,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}z_{2,3}$ | $a53$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{1,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}z_{2,3}$ | $a54$ $(2,2,0,0)$ | $4$ | $(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{1}z_{0}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-\dfrac{2}{3}z_{1}z_{0}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{1}z_{0}z_{0,2}z_{2,3}z_{1,3}\right)$ | $a55$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{1}z_{0}z_{0,1}+2\gamma(1-\gamma)\dfrac{1}{3}z_{1}z_{0}+\gamma^{2}\dfrac{1}{3}z_{1}z_{0}z_{0,2}z_{2,3}z_{1,3}$ | $a56$ | | $(1-\gamma)^{2}\dfrac{1}{3}z_{1}z_{0}z_{0,1}+2\gamma(1-\gamma)\dfrac{1}{3}z_{1}z_{0}+\gamma^{2}\dfrac{1}{3}z_{1}z_{0}z_{0,2}z_{2,3}z_{1,3}$ | $a57$ For the mapping of observed CFs to $a_{i}$ values, we need to know to which of the three splits each of the $a_{i}$’s corresponds to. So, next, we elaborate on which specific split each of the $a_{i}$ corresponds to: #### $n=(0,0,2,2)$ * • Individuals $k_{1},k_{2}\in n_{2}$ and $l_{1},l_{2}\in n_{3}$ * • $P(k_{1},k_{2}|l_{1},l_{2})=1-\dfrac{2}{3}z_{2}z_{2,3}z_{3}=a_{1}$ * • $P(k_{1},l_{1}|k_{2},l_{2})=\dfrac{1}{3}z_{2}z_{2,3}z_{3}=a_{2}$ * • $P(k_{1},l_{2}|k_{2},l_{1})=\dfrac{1}{3}z_{2}z_{2,3}z_{3}=a_{3}$ #### $n=(0,1,2,1)$ * • Individuals $j_{1}\in n_{1}$; $k_{1},k_{2}\in n_{2}$ and $l_{1}\in n_{3}$ * • $P(k_{1},k_{2}|j_{1},l_{1})=1-\dfrac{2}{3}z_{2}z_{2,3}=a_{4}$ * • $P(k_{1},j_{1}|k_{2},l_{1})=\dfrac{1}{3}z_{2}z_{2,3}=a_{5}$ * • $P(k_{1},l_{1}|k_{2},j_{1})=\dfrac{1}{3}z_{2}z_{2,3}=a_{6}$ #### $n=(0,1,1,2)$ * • Individuals $j_{1}\in n_{1}$; $k_{1}\in n_{2}$ and $l_{1},l_{2}\in n_{3}$ * • $P(l_{1},l_{2}|j_{1},k_{1})=1-\dfrac{2}{3}z_{3}=a_{7}$ * • $P(l_{1},j_{1}|l_{2},k_{1})=\dfrac{1}{3}z_{3}=a_{8}$ * • $P(l_{1},k_{1}|l_{2},j_{1})=\dfrac{1}{3}z_{3}=a_{9}$ #### $n=(0,2,2,0)$ * • Individuals $j_{1},j_{2}\in n_{1}$; $k_{1},k_{2}\in n_{2}$ * • $P(j_{1},j_{2}|k_{1},k_{2})=1-\dfrac{2}{3}z_{2}z_{2,3}z_{1,3}z_{1}=a_{10}$ * • $P(j_{1},k_{1}|j_{2},k_{2})=\dfrac{1}{3}z_{2}z_{2,3}z_{1,3}z_{1}=a_{11}$ * • $P(j_{1},k_{2}|j_{2},k_{1})=\dfrac{1}{3}z_{2}z_{2,3}z_{1,3}z_{1}=a_{12}$ #### $n=(0,2,1,1)$ * • Individuals $j_{1},j_{2}\in n_{1}$; $k_{1}\in n_{2}$; $l_{1}\in n_{3}$ * • $P(j_{1},j_{2}|k_{1},l_{1})=1-\dfrac{2}{3}z_{1,3}z_{1}=a_{13}$ * • $P(j_{1},k_{1}|j_{2},l_{1})=\dfrac{1}{3}z_{1,3}z_{1}=a_{14}$ * • $P(j_{1},l_{1}|j_{2},k_{1})=\dfrac{1}{3}z_{1,3}z_{1}=a_{15}$ #### $n=(0,2,0,2)$ * • Individuals $j_{1},j_{2}\in n_{1}$; $l_{1},l_{2}\in n_{3}$ * • $P(j_{1},j_{2}|l_{1},l_{2})=1-\dfrac{2}{3}z_{3}z_{1,3}z_{1}=a_{16}$ * • $P(j_{1},l_{1}|j_{2},l_{2})=\dfrac{1}{3}z_{3}z_{1,3}z_{1}=a_{17}$ * • $P(j_{1},l_{2}|j_{2},l_{1})=\dfrac{1}{3}z_{3}z_{1,3}z_{1}=a_{18}$ #### $n=(1,0,2,1)$ * • Individuals $i_{1}\in n_{0}$; $k_{1},k_{2}\in n_{2}$; $l_{1}\in n_{3}$ * • $P(k_{1},k_{2}|i_{1},l_{1})=(1-\gamma)\left(1-\dfrac{2}{3}z_{2,3}z_{2}\right)+\gamma\left(1-\dfrac{2}{3}z_{2}\right)=a_{19}$ * • $P(k_{1},i_{1}|k_{2},l_{1})=(1-\gamma)\dfrac{1}{3}z_{2,3}z_{2}+\gamma\dfrac{1}{3}z_{2}=a_{20}$ * • $P(k_{1},l_{1}|k_{2},i_{1})=(1-\gamma)\dfrac{1}{3}z_{2,3}z_{2}+\gamma\dfrac{1}{3}z_{2}=a_{21}$ #### $n=(1,0,1,2)$ * • Individuals $i_{1}\in n_{0}$; $k_{1}\in n_{2}$; $l_{1},l_{2}\in n_{3}$ * • $P(l_{1},l_{2}|i_{1},k_{1})=(1-\gamma)\left(1-\dfrac{2}{3}z_{3}\right)+\gamma\left(1-\dfrac{2}{3}z_{2,3}z_{3}\right)=a_{22}$ * • $P(l_{1},i_{1}|l_{2},k_{1})=(1-\gamma)\dfrac{1}{3}z_{3}+\gamma\dfrac{1}{3}z_{2,3}z_{3}=a_{23}$ * • $P(l_{1},k_{1}|l_{2},i_{1})=(1-\gamma)\dfrac{1}{3}z_{3}+\gamma\dfrac{1}{3}z_{2,3}z_{3}=a_{24}$ #### $n=(1,1,2,0)$ * • Individuals $i_{1}\in n_{0}$; $j_{1}\in n_{1}$; $k_{1},k_{2}\in n_{2}$ * • $P(k_{1},k_{2}|i_{1},j_{1})=(1-\gamma)\left(1-\dfrac{2}{3}z_{1,3}z_{2,3}z_{2}\right)+\gamma\left(1-\dfrac{2}{3}z_{2}\right)=a_{25}$ * • $P(k_{1},i_{1}|k_{2},j_{1})=(1-\gamma)\dfrac{1}{3}z_{1,3}z_{2,3}z_{2}+\gamma\dfrac{1}{3}z_{2}=a_{26}$ * • $P(k_{1},j_{1}|k_{2},i_{1})=(1-\gamma)\dfrac{1}{3}z_{1,3}z_{2,3}z_{2}+\gamma\dfrac{1}{3}z_{2}=a_{27}$ #### $n=(1,1,1,1)$ * • Individuals $i_{1}\in n_{0}$; $j_{1}\in n_{1}$; $k_{1}\in n_{2}$; $l_{1}\in n_{3}$ * • $P(i_{1},j_{1}|k_{1},l_{1})=(1-\gamma)\left(1-\dfrac{2}{3}z_{1,3}\right)+\gamma\dfrac{2}{3}z_{2,3}=a_{28}$ * • $P(i_{1},k_{1}|j_{1},l_{1})=(1-\gamma)\dfrac{1}{3}z_{1,3}+\gamma\left(1-\dfrac{2}{3}z_{2,3}\right)=a_{29}$ * • $P(i_{1},l_{1}|j_{1},k_{1})=(1-\gamma)\dfrac{1}{3}z_{1,3}+\gamma\dfrac{1}{3}z_{2,3}=a_{30}$ #### $n=(1,1,0,2)$ * • Individuals $i_{1}\in n_{0}$; $j_{1}\in n_{1}$; $l_{1},l_{2}\in n_{3}$ * • $P(l_{1},l_{2}|i_{1},j_{1})=(1-\gamma)\left(1-\dfrac{2}{3}z_{1,3}z_{3}\right)+\gamma\left(1-\dfrac{2}{3}z_{3}\right)=a_{31}$ * • $P(l_{1},i_{1}|l_{2},j_{1})=(1-\gamma)\dfrac{1}{3}z_{1,3}z_{3}+\gamma\dfrac{1}{3}z_{3}=a_{32}$ * • $P(l_{1},j_{1}|l_{2},i_{1})=(1-\gamma)\dfrac{1}{3}z_{1,3}z_{3}+\gamma\dfrac{1}{3}z_{3}=a_{33}$ #### $n=(1,2,1,0)$ * • Individuals $i_{1}\in n_{0}$; $j_{1},j_{2}\in n_{1}$; $k_{1}\in n_{2}$ * • $P(j_{1},j_{2}|i_{1},k_{1})=(1-\gamma)\left(1-\dfrac{2}{3}z_{1}\right)+\gamma\left(1-\dfrac{2}{3}z_{2,3}z_{1,3}z_{1}\right)=a_{34}$ * • $P(j_{1},i_{1}|j_{2},k_{1})=(1-\gamma)\dfrac{1}{3}z_{1}+\gamma\dfrac{1}{3}z_{2,3}z_{1,3}z_{1}=a_{35}$ * • $P(j_{1},k_{1}|j_{2},i_{1})=(1-\gamma)\dfrac{1}{3}z_{1}+\gamma\dfrac{1}{3}z_{2,3}z_{1,3}z_{1}=a_{36}$ #### $n=(1,2,0,1)$ * • Individuals $i_{1}\in n_{0}$; $j_{1},j_{2}\in n_{1}$; $l_{1}\in n_{3}$ * • $P(j_{1},j_{2}|i_{1},l_{1})=(1-\gamma)\left(1-\dfrac{2}{3}z_{1}\right)+\gamma\left(1-\dfrac{2}{3}z_{1,3}z_{1}\right)=a_{37}$ * • $P(j_{1},l_{1}|j_{2},i_{1})=(1-\gamma)\dfrac{1}{3}z_{1}+\gamma\dfrac{1}{3}z_{1,3}z_{1}=a_{38}$ * • $P(j_{1},i_{1}|j_{2},l_{1})=(1-\gamma)\dfrac{1}{3}z_{1}+\gamma\dfrac{1}{3}z_{1,3}z_{1}=a_{39}$ #### $n=(2,0,2,0)$ * • Individuals $i_{1},i_{2}\in n_{0}$; $k_{1},k_{2}\in n_{2}$ * • $P(i_{1},i_{2}|k_{1},k_{2})=(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{2}z_{0}z_{0,1}z_{1,3}z_{2,3}\right)+2\gamma(1-\gamma)\left(1-\dfrac{2}{3}z_{2}z_{0}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{2}z_{0}z_{0,2}\right)=a_{40}$ * • $P(i_{1},k_{1}|i_{2},k_{2})=(1-\gamma)^{2}\dfrac{1}{3}z_{2}z_{0}z_{0,1}z_{1,3}z_{2,3}+2\gamma(1-\gamma)\dfrac{1}{3}z_{2}z_{0}+\gamma^{2}\dfrac{1}{3}z_{2}z_{0}z_{0,2}=a_{41}$ * • $P(i_{1},k_{2}|k_{1},i_{2})=(1-\gamma)^{2}\dfrac{1}{3}z_{2}z_{0}z_{0,1}z_{1,3}z_{2,3}+2\gamma(1-\gamma)\dfrac{1}{3}z_{2}z_{0}+\gamma^{2}\dfrac{1}{3}z_{2}z_{0}z_{0,2}=a_{42}$ #### $n=(2,0,1,1)$ * • Individuals $i_{1},i_{2}\in n_{0}$; $k_{1}\in n_{2}$; $l_{1}\in n_{3}$ * • $P(i_{1},i_{2}|k_{1},l_{1})=(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{0}z_{1,3}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-z_{0}+\dfrac{1}{3}z_{0}z_{2,3}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,2}\right)=a_{43}$ * • $P(i_{1},k_{1}|l_{1},i_{2})=(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{1,3}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{2,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}=a_{44}$ * • $P(i_{1},l_{1}|i_{2},k_{1})=(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{1,3}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{2,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}=a_{45}$ #### $n=(2,0,0,2)$ * • Individuals $i_{1},i_{2}\in n_{0}$; $l_{1},l_{2}\in n_{3}$ * • $P(i_{1},i_{2}|l_{1},l_{2})=(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{3}z_{0}z_{1,3}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-\dfrac{2}{3}z_{3}z_{0}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{3}z_{0}z_{2,3}z_{0,2}\right)=a_{46}$ * • $P(i_{1},l_{1}|i_{2},l_{2})=(1-\gamma)^{2}\dfrac{1}{3}z_{3}z_{0}z_{1,3}z_{0,1}+2\gamma(1-\gamma)\dfrac{1}{3}z_{3}z_{0}+\gamma^{2}\dfrac{1}{3}z_{3}z_{0}z_{2,3}z_{0,2}=a_{47}$ * • $P(i_{1},l_{2}|i_{2},l_{1})=(1-\gamma)^{2}\dfrac{1}{3}z_{3}z_{0}z_{1,3}z_{0,1}+2\gamma(1-\gamma)\dfrac{1}{3}z_{3}z_{0}+\gamma^{2}\dfrac{1}{3}z_{3}z_{0}z_{2,3}z_{0,2}=a_{48}$ #### $n=(2,1,1,0)$ * • Individuals $i_{1},i_{2}\in n_{0}$; $j_{1}\in n_{1}$; $k_{1}\in n_{2}$ * • $P(i_{1},i_{2}|j_{1},k_{1})=(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-z_{0}+\dfrac{1}{3}z_{0}z_{2,3}z_{1,3}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,2}\right)=a_{49}$ * • $P(i_{1},j_{1}|i_{2},k_{1})=1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{2,3}z_{1,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}=a_{50}$ * • $P(i_{1},k_{1}|i_{2},j_{1})=1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{2,3}z_{1,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}=a_{51}$ #### $n=(2,1,0,1)$ * • Individuals $i_{1},i_{2}\in n_{0}$; $j_{1}\in n_{1}$; $l_{1}\in n_{3}$ * • $P(i_{1},i_{2}|j_{1},l_{1})=(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-z_{0}+\dfrac{1}{3}z_{0}z_{1,3}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{0}z_{0,2}z_{2,3}\right)=a_{52}$ * • $P(i_{1},j_{1}|i_{2},l_{1})=(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{1,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}z_{2,3}=a_{53}$ * • $P(i_{1},l_{1}|i_{2},j_{1})=(1-\gamma)^{2}\dfrac{1}{3}z_{0}z_{0,1}+\gamma(1-\gamma)z_{0}\left(1-\dfrac{1}{3}z_{1,3}\right)+\gamma^{2}\dfrac{1}{3}z_{0}z_{0,2}z_{2,3}=a_{54}$ #### $n=(2,2,0,0)$ * • Individuals $i_{1},i_{2}\in n_{0}$; $j_{1},j_{2}\in n_{1}$; * • $P(i_{1},i_{2}|j_{1},j_{2})=(1-\gamma)^{2}\left(1-\dfrac{2}{3}z_{1}z_{0}z_{0,1}\right)+2\gamma(1-\gamma)\left(1-\dfrac{2}{3}z_{1}z_{0}\right)+\gamma^{2}\left(1-\dfrac{2}{3}z_{1}z_{0}z_{0,2}z_{2,3}z_{1,3}\right)=a_{55}$ * • $P(i_{1},j_{1}|i_{2},j_{2})=(1-\gamma)^{2}\dfrac{1}{3}z_{1}z_{0}z_{0,1}+2\gamma(1-\gamma)\dfrac{1}{3}z_{1}z_{0}+\gamma^{2}\dfrac{1}{3}z_{1}z_{0}z_{0,2}z_{2,3}z_{1,3}=a_{56}$ * • $P(i_{1},j_{2}|i_{2},j_{1})=(1-\gamma)^{2}\dfrac{1}{3}z_{1}z_{0}z_{0,1}+2\gamma(1-\gamma)\dfrac{1}{3}z_{1}z_{0}+\gamma^{2}\dfrac{1}{3}z_{1}z_{0}z_{0,2}z_{2,3}z_{1,3}=a_{57}$ ## Appendix B Phylogenetic invariants for $n$-taxon phylogenetic networks with one 4-node hybridization cycle Below, we present the invariants for different networks $N$ all with one 4-cycle, but with different number of species on the clades $n_{0},n_{1},n_{2},n_{3}$. For example, the network $N=1112$ corresponds to a network with 5 species: one in $n_{0}$, one in $n_{1}$, one in $n_{2}$ and two in $n_{3}$. The number of species defines the number of CF formulas. For example, for 6 species, there are ${6\choose 4}=15$ 4-taxon subsets, each with 3 CF formulas. So, for 6 species, we have 45 CF formulas and thus, 45 CF values. However, we only want to focus on the 4-taxon subsets that involve the hybridization cycle. For the case of $N=1112$, they are only 4: $(0,1,1,2),(1,0,1,2),(1,1,1,1),(1,1,0,2)$. Note that our table of CF formulas has 57 different CF formulas (and therefore, values). This discrepancy is due to the fact that the table is listing all possible CF formulas and we will have fewer formulas if we have less than 8 species. For some examples, for computational restrictions, we had to include just a subset of the original CF equations to obtain the Gröbner basis in $a_{i}$. We denote these cases with "subset". All Macaulay2 scripts (and output) can be found in the GitHub repository: https://github.com/solislemuslab/phylo- diamond.jl. * • $N=1112$ 1. 1. $a_{32}-a_{33}$ 2. 2. $a_{31}+2a_{33}-1$ 3. 3. $a_{28}+a_{29}+a_{30}-1$ 4. 4. $a_{23}-a_{24}$ 5. 5. $a_{22}+2a_{24}-1$ 6. 6. $a_{8}-a_{9}$ 7. 7. $a_{7}+2a_{9}-1$ 8. 8. $3a_{9}*a_{30}+a_{9}-a_{24}-a_{33}$ 9. 9. $a_{24}*a_{29}+2a_{24}*a_{30}+a_{29}*a_{33}-a_{30}*a_{33}-a_{33}$ 10. 10. $3a_{9}*a_{29}-2a_{9}+2a_{24}-a_{33}$ * • $N=1121$ 1. 1. $a_{28}+a_{29}+a_{30}-1$ 2. 2. $a_{26}-a_{27}$ 3. 3. $a_{25}+2a_{27}-1$ 4. 4. $a_{20}-a_{21}$ 5. 5. $a_{19}+2a_{21}-1$ 6. 6. $a_{5}-a_{6}$ 7. 7. $a_{4}+2a_{6}-1$ 8. 8. $a_{6}*a_{29}+2a_{6}*a_{30}-a_{6}+a_{21}-a_{27}$ * • $N=1122$ 1. 1. $a_{32}-a_{33}$ 2. 2. $a_{31}+2a_{33}-1$ 3. 3. $a_{28}+a_{29}+a_{30}-1$ 4. 4. $a_{26}-a_{27}$ 5. 5. $a_{25}+2a_{27}-1$ 6. 6. $a_{23}-a_{24}$ 7. 7. $a_{22}+2a_{24}-1$ 8. 8. $a_{20}-a_{21}$ 9. 9. $a_{19}+2a_{21}-1$ 10. 10. $a_{8}-a_{9}$ 11. 11. $a_{7}+2a_{9}-1$ 12. 12. $a_{5}-a_{6}$ 13. 13. $a_{4}+2a_{6}-1$ 14. 14. $a_{2}-a_{3}$ 15. 15. $a_{1}+2a_{3}-1$ 16. 16. $3a_{9}*a_{30}+a_{9}-a_{24}-a_{33}$ 17. 17. $a_{24}*a_{29}+2a_{24}*a_{30}+a_{29}*a_{33}-a_{30}*a_{33}-a_{33}$ 18. 18. $3a_{9}*a_{29}-2a_{9}+2a_{24}-a_{33}$ 19. 19. $a_{6}*a_{29}+2a_{6}*a_{30}-a_{6}+a_{21}-a_{27}$ 20. 20. $a_{3}*a_{29}+2a_{3}*a_{30}-3a_{6}*a_{33}$ 21. 21. $3a_{6}*a_{24}-3a_{3}*a_{30}+3a_{6}*a_{33}-a_{3}$ 22. 22. $3a_{9}*a_{21}-3a_{9}*a_{27}+3a_{6}*a_{33}-a_{3}$ 23. 23. $3a_{6}*a_{9}-a_{3}$ * • $N=1211$ 1. 1. $a_{38}-a_{39}$ 2. 2. $a_{37}+2a_{39}-1$ 3. 3. $a_{35}-a_{36}$ 4. 4. $a_{34}+2a_{36}-1$ 5. 5. $a_{28}+a_{29}+a_{30}-1$ 6. 6. $a_{14}-a_{15}$ 7. 7. $a_{13}+2a_{15}-1$ 8. 8. $a_{15}*a_{29}-a_{15}*a_{30}+a_{36}-a_{39}$ * • $N=1212$ 1. 1. $a_{38}-a_{39}$ 2. 2. $a_{37}+2a_{39}-1$ 3. 3. $a_{35}-a_{36}$ 4. 4. $a_{34}+2a_{36}-1$ 5. 5. $a_{32}-a_{33}$ 6. 6. $a_{31}+2a_{33}-1$ 7. 7. $a_{28}+a_{29}+a_{30}-1$ 8. 8. $a_{23}-a_{24}$ 9. 9. $a_{22}+2a_{24}-1$ 10. 10. $a_{17}-a_{18}$ 11. 11. $a_{16}+2a_{18}-1$ 12. 12. $a_{14}-a_{15}$ 13. 13. $a_{13}+2a_{15}-1$ 14. 14. $a_{8}-a_{9}$ 15. 15. $a_{7}+2a_{9}-1$ 16. 16. $a_{18}*a_{30}-a_{15}*a_{33}-a_{9}*a_{36}+a_{9}*a_{39}$ 17. 17. $3a_{9}*a_{30}+a_{9}-a_{24}-a_{33}$ 18. 18. $a_{24}*a_{29}+2a_{24}*a_{30}+a_{29}*a_{33}-a_{30}*a_{33}-a_{33}$ 19. 19. $a_{18}*a_{29}-a_{15}*a_{33}+2a_{9}*a_{36}-2a_{9}*a_{39}$ 20. 20. $a_{15}*a_{29}-a_{15}*a_{30}+a_{36}-a_{39}$ 21. 21. $3a_{9}*a_{29}-2a_{9}+2a_{24}-a_{33}$ 22. 22. $3a_{15}*a_{24}-3a_{9}*a_{36}+3a_{9}*a_{39}-a_{18}$ 23. 23. $3a_{9}*a_{15}-a_{18}$ * • $N=1221$ 1. 1. $a_{38}-a_{39}$ 2. 2. $a_{37}+2a_{39}-1$ 3. 3. $a_{35}-a_{36}$ 4. 4. $a_{34}+2a_{36}-1$ 5. 5. $a_{28}+a_{29}+a_{30}-1$ 6. 6. $a_{26}-a_{27}$ 7. 7. $a_{25}+2a_{27}-1$ 8. 8. $a_{20}-a_{21}$ 9. 9. $a_{19}+2a_{21}-1$ 10. 10. $a_{14}-a_{15}$ 11. 11. $a_{13}+2a_{15}-1$ 12. 12. $a_{11}-a_{12}$ 13. 13. $a_{10}+2a_{12}-1$ 14. 14. $a_{5}-a_{6}$ 15. 15. $a_{4}+2a_{6}-1$ 16. 16. $a_{15}*a_{29}-a_{15}*a_{30}+a_{36}-a_{39}$ 17. 17. $a_{12}*a_{29}-a_{12}*a_{30}+3a_{6}*a_{36}-3a_{6}*a_{39}$ 18. 18. $a_{6}*a_{29}+2a_{6}*a_{30}-a_{6}+a_{21}-a_{27}$ 19. 19. $3a_{15}*a_{21}-3a_{15}*a_{27}+3a_{12}*a_{30}-3a_{6}*a_{36}+3a_{6}*a_{39}-a_{12}$ 20. 20. $3a_{6}*a_{15}-a_{12}$ 21. 21. $a_{21}*a_{29}*a_{36}+2a_{21}*a_{30}*a_{36}-a_{27}*a_{29}*a_{39}+a_{27}*a_{30}*a_{39}-a_{15}*a_{27}+a_{12}*a_{30}-a_{6}*a_{36}-a_{21}*a_{36}-a_{27}*a_{36}+2a_{21}*a_{39}$ * • $N=1222$ 1. 1. $a_{38}-a_{39}$ 2. 2. $a_{37}+2a_{39}-1$ 3. 3. $a_{35}-a_{36}$ 4. 4. $a_{34}+2a_{36}-1$ 5. 5. $a_{32}-a_{33}$ 6. 6. $a_{31}+2a_{33}-1$ 7. 7. $a_{28}+a_{29}+a_{30}-1$ 8. 8. $a_{26}-a_{27}$ 9. 9. $a_{25}+2a_{27}-1$ 10. 10. $a_{23}-a_{24}$ 11. 11. $a_{22}+2a_{24}-1$ 12. 12. $a_{20}-a_{21}$ 13. 13. $a_{19}+2a_{21}-1$ 14. 14. $a_{17}-a_{18}$ 15. 15. $a_{16}+2a_{18}-1$ 16. 16. $a_{14}-a_{15}$ 17. 17. $a_{13}+2a_{15}-1$ 18. 18. $a_{11}-a_{12}$ 19. 19. $a_{10}+2a_{12}-1$ 20. 20. $a_{8}-a_{9}$ 21. 21. $a_{7}+2a_{9}-1$ 22. 22. $a_{5}-a_{6}$ 23. 23. $a_{4}+2a_{6}-1$ 24. 24. $a_{2}-a_{3}$ 25. 25. $a_{1}+2a_{3}-1$ 26. 26. $a_{18}*a_{30}-a_{15}*a_{33}-a_{9}*a_{36}+a_{9}*a_{39}$ 27. 27. $3a_{9}*a_{30}+a_{9}-a_{24}-a_{33}$ 28. 28. $a_{24}*a_{29}+2a_{24}*a_{30}+a_{29}*a_{33}-a_{30}*a_{33}-a_{33}$ 29. 29. $a_{18}*a_{29}-a_{15}*a_{33}+2a_{9}*a_{36}-2a_{9}*a_{39}$ 30. 30. $a_{15}*a_{29}-a_{15}*a_{30}+a_{36}-a_{39}$ 31. 31. $a_{12}*a_{29}-a_{12}*a_{30}+3a_{6}*a_{36}-3a_{6}*a_{39}$ 32. 32. $3a_{9}*a_{29}-2a_{9}+2a_{24}-a_{33}$ 33. 33. $a_{6}*a_{29}+2a_{6}*a_{30}-a_{6}+a_{21}-a_{27}$ 34. 34. $a_{3}*a_{29}+2a_{3}*a_{30}-3a_{6}*a_{33}$ 35. 35. $3a_{15}*a_{24}-3a_{9}*a_{36}+3a_{9}*a_{39}-a_{18}$ 36. 36. $3a_{6}*a_{24}-3a_{3}*a_{30}+3a_{6}*a_{33}-a_{3}$ 37. 37. $a_{18}*a_{21}-a_{12}*a_{24}-a_{18}*a_{27}+a_{12}*a_{33}+a_{3}*a_{36}-a_{3}*a_{39}$ 38. 38. $3a_{15}*a_{21}-3a_{15}*a_{27}+3a_{12}*a_{30}-3a_{6}*a_{36}+3a_{6}*a_{39}-a_{12}$ 39. 39. $3a_{9}*a_{21}-3a_{9}*a_{27}+3a_{6}*a_{33}-a_{3}$ 40. 40. $a_{6}*a_{18}-a_{12}*a_{24}+a_{3}*a_{36}-a_{3}*a_{39}$ 41. 41. $3a_{9}*a_{15}-a_{18}$ 42. 42. $3a_{6}*a_{15}-a_{12}$ 43. 43. $a_{3}*a_{15}-a_{12}*a_{24}+a_{3}*a_{36}-a_{3}*a_{39}$ 44. 44. $a_{9}*a_{12}-a_{12}*a_{24}+a_{3}*a_{36}-a_{3}*a_{39}$ 45. 45. $3a_{6}*a_{9}-a_{3}$ 46. 46. $a_{21}*a_{29}*a_{36}+2a_{21}*a_{30}*a_{36}-a_{27}*a_{29}*a_{39}+a_{27}*a_{30}*a_{39}-a_{15}*a_{27}+a_{12}*a_{30}-a_{6}*a_{36}-a_{21}*a_{36}-a_{27}*a_{36}+2a_{21}*a_{39}$ 47. 47. $6a_{9}*a_{27}*a_{36}-3a_{6}*a_{33}*a_{36}-3a_{21}*a_{33}*a_{36}-3a_{9}*a_{27}*a_{39}-3a_{24}*a_{27}*a_{39}+6a_{6}*a_{33}*a_{39}+a_{18}*a_{27}-a_{12}*a_{33}+a_{3}*a_{36}-a_{3}*a_{39}$ * • $N=2111$ 1. 1. $a_{53}-a_{54}$ 2. 2. $a_{52}+2a_{54}-1$ 3. 3. $a_{50}-a_{51}$ 4. 4. $a_{49}+2a_{51}-1$ 5. 5. $a_{44}-a_{45}$ 6. 6. $a_{43}+2a_{45}-1$ 7. 7. $a_{28}+a_{29}+a_{30}-1$ * • $N=2112$ 1. 1. $a_{53}-a_{54}$ 2. 2. $a_{52}+2a_{54}-1$ 3. 3. $a_{50}-a_{51}$ 4. 4. $a_{49}+2a_{51}-1$ 5. 5. $a_{47}-a_{48}$ 6. 6. $a_{46}+2a_{48}-1$ 7. 7. $a_{44}-a_{45}$ 8. 8. $a_{43}+2a_{45}-1$ 9. 9. $a_{32}-a_{33}$ 10. 10. $a_{31}+2a_{33}-1$ 11. 11. $a_{28}+a_{29}+a_{30}-1$ 12. 12. $a_{23}-a_{24}$ 13. 13. $a_{22}+2a_{24}-1$ 14. 14. $a_{8}-a_{9}$ 15. 15. $a_{7}+2a_{9}-1$ 16. 16. $3a_{9}*a_{30}+a_{9}-a_{24}-a_{33}$ 17. 17. $a_{24}*a_{29}+2a_{24}*a_{30}+a_{29}*a_{33}-a_{30}*a_{33}-a_{33}$ 18. 18. $3a_{9}*a_{29}-2a_{9}+2a_{24}-a_{33}$ * • $N=2121$ 1. 1. $a_{4}+2a_{6}-1$ 2. 2. $a_{5}-a_{6}$ 3. 3. $a_{28}+a_{29}+a_{30}-1$ 4. 4. $a_{19}+2a_{21}-1$ 5. 5. $a_{20}-a_{21}$ 6. 6. $a_{25}+2a_{27}-1$ 7. 7. $a_{26}-a_{27}$ 8. 8. $a_{40}+2a_{42}-1$ 9. 9. $a_{41}-a_{42}$ 10. 10. $a_{43}+2a_{45}-1$ 11. 11. $a_{44}-a_{45}$ 12. 12. $a_{49}+2a_{51}-1$ 13. 13. $a_{50}-a_{51}$ 14. 14. $a_{52}+2a_{54}-1$ 15. 15. $a_{53}-a_{54}$ 16. 16. $a_{5}*a_{29}+2a_{5}*a_{30}-a_{5}+a_{20}-a_{26}$ 17. 17. $2a_{20}*a_{29}^{3}a_{41}+a_{26}*a_{29}^{3}a_{41}+3a_{20}*a_{29}^{2}a_{30}*a_{41}-3a_{20}*a_{29}*a_{30}^{2}a_{41}-3a_{26}*a_{29}*a_{30}^{2}a_{41}-2a_{20}*a_{30}^{3}a_{41}+2a_{26}*a_{30}^{3}a_{41}-3a_{20}*a_{26}*a_{29}^{2}a_{44}-3a_{26}^{2}a_{29}^{2}a_{44}-3a_{20}*a_{26}*a_{29}*a_{30}*a_{44}+6a_{26}^{2}a_{29}*a_{30}*a_{44}+6a_{20}*a_{26}*a_{30}^{2}a_{44}-3a_{26}^{2}a_{30}^{2}a_{44}-6a_{20}^{2}a_{29}^{2}a_{50}+3a_{20}*a_{26}*a_{29}^{2}a_{50}-15a_{20}^{2}a_{29}*a_{30}*a_{50}-6a_{20}*a_{26}*a_{29}*a_{30}*a_{50}-6a_{20}^{2}a_{30}^{2}a_{50}+3a_{20}*a_{26}*a_{30}^{2}a_{50}+3a_{20}*a_{26}*a_{29}^{2}a_{53}+3a_{20}*a_{26}*a_{29}*a_{30}*a_{53}-6a_{20}*a_{26}*a_{30}^{2}a_{53}-4a_{20}*a_{29}^{2}a_{41}-2a_{26}*a_{29}^{2}a_{41}-a_{20}*a_{29}*a_{30}*a_{41}+a_{26}*a_{29}*a_{30}*a_{41}+5a_{20}*a_{30}^{2}a_{41}+a_{26}*a_{30}^{2}a_{41}+3a_{20}^{2}a_{29}*a_{44}+6a_{20}*a_{26}*a_{29}*a_{44}+3a_{26}^{2}a_{29}*a_{44}+6a_{20}^{2}a_{30}*a_{44}+18a_{5}*a_{26}*a_{30}*a_{44}-6a_{20}*a_{26}*a_{30}*a_{44}-3a_{26}^{2}a_{30}*a_{44}+6a_{20}*a_{26}*a_{29}*a_{50}-9a_{5}*a_{20}*a_{30}*a_{50}+9a_{20}^{2}a_{30}*a_{50}-9a_{5}*a_{26}*a_{30}*a_{50}+12a_{20}*a_{26}*a_{30}*a_{50}-3a_{20}^{2}a_{29}*a_{53}-3a_{26}^{2}a_{29}*a_{53}-6a_{20}^{2}a_{30}*a_{53}+3a_{26}^{2}a_{30}*a_{53}+3a_{26}*a_{29}*a_{41}-3a_{26}*a_{30}*a_{41}-3a_{5}^{2}a_{44}+3a_{5}*a_{20}*a_{44}-3a_{5}*a_{26}*a_{44}-6a_{26}^{2}a_{44}+3a_{5}^{2}a_{50}-6a_{5}*a_{20}*a_{50}+3a_{20}^{2}a_{50}+6a_{5}*a_{26}*a_{50}-6a_{20}*a_{26}*a_{50}+3a_{5}*a_{20}*a_{53}-3a_{20}^{2}a_{53}-3a_{5}*a_{26}*a_{53}+6a_{20}*a_{26}*a_{53}$ * • $N=2211$ 1. 1. $a_{13}+2a_{15}-1$ 2. 2. $a_{14}-a_{15}$ 3. 3. $a_{28}+a_{29}+a_{30}-1$ 4. 4. $a_{34}+2a_{36}-1$ 5. 5. $a_{35}-a_{36}$ 6. 6. $a_{37}+2a_{39}-1$ 7. 7. $a_{38}-a_{39}$ 8. 8. $a_{43}+2a_{45}-1$ 9. 9. $a_{44}-a_{45}$ 10. 10. $a_{49}+2a_{51}-1$ 11. 11. $a_{50}-a_{51}$ 12. 12. $a_{52}+2a_{54}-1$ 13. 13. $a_{53}-a_{54}$ 14. 14. $a_{55}+2a_{57}-1$ 15. 15. $a_{56}-a_{57}$ 16. 16. $a_{14}*a_{29}-a_{14}*a_{30}+a_{35}-a_{38}$ 17. 17. $3a_{29}^{2}a_{35}*a_{38}*a_{44}+3a_{29}*a_{30}*a_{35}*a_{38}*a_{44}-6a_{30}^{2}a_{35}*a_{38}*a_{44}+3a_{29}^{2}a_{35}*a_{38}*a_{50}+12a_{29}*a_{30}*a_{35}*a_{38}*a_{50}+12a_{30}^{2}a_{35}*a_{38}*a_{50}-6a_{29}^{2}a_{38}^{2}a_{50}+3a_{29}*a_{30}*a_{38}^{2}a_{50}+3a_{30}^{2}a_{38}^{2}a_{50}-3a_{29}^{2}a_{35}^{2}a_{53}-12a_{29}*a_{30}*a_{35}^{2}a_{53}-12a_{30}^{2}a_{35}^{2}a_{53}-3a_{29}^{2}a_{35}*a_{38}*a_{53}-3a_{29}*a_{30}*a_{35}*a_{38}*a_{53}+6a_{30}^{2}a_{35}*a_{38}*a_{53}-a_{29}^{3}a_{35}*a_{56}-3a_{29}^{2}a_{30}*a_{35}*a_{56}+4a_{30}^{3}a_{35}*a_{56}-2a_{29}^{3}a_{38}*a_{56}-3a_{29}^{2}a_{30}*a_{38}*a_{56}+3a_{29}*a_{30}^{2}a_{38}*a_{56}+2a_{30}^{3}a_{38}*a_{56}+3a_{29}*a_{35}^{2}a_{44}+6a_{30}*a_{35}^{2}a_{44}-6a_{29}*a_{35}*a_{38}*a_{44}-3a_{30}*a_{35}*a_{38}*a_{44}+3a_{29}*a_{38}^{2}a_{44}-3a_{30}*a_{38}^{2}a_{44}-9a_{14}*a_{30}*a_{35}*a_{50}-9a_{14}*a_{30}*a_{38}*a_{50}-12a_{29}*a_{35}*a_{38}*a_{50}-6a_{30}*a_{35}*a_{38}*a_{50}+12a_{29}*a_{38}^{2}a_{50}+6a_{30}*a_{38}^{2}a_{50}+18a_{14}*a_{30}*a_{35}*a_{53}+3a_{29}*a_{35}^{2}a_{53}+6a_{30}*a_{35}^{2}a_{53}-9a_{30}*a_{35}*a_{38}*a_{53}-3a_{29}*a_{38}^{2}a_{53}+3a_{30}*a_{38}^{2}a_{53}+a_{29}^{2}a_{35}*a_{56}+a_{29}*a_{30}*a_{35}*a_{56}-2a_{30}^{2}a_{35}*a_{56}+2a_{29}^{2}a_{38}*a_{56}-a_{29}*a_{30}*a_{38}*a_{56}-a_{30}^{2}a_{38}*a_{56}-3a_{14}*a_{35}*a_{44}-3a_{35}^{2}a_{44}+3a_{14}*a_{38}*a_{44}+9a_{35}*a_{38}*a_{44}-6a_{38}^{2}a_{44}+3a_{14}^{2}a_{50}+6a_{14}*a_{35}*a_{50}-6a_{14}*a_{38}*a_{50}+3a_{35}*a_{38}*a_{50}-3a_{38}^{2}a_{50}-3a_{14}^{2}a_{53}-3a_{14}*a_{35}*a_{53}-6a_{35}^{2}a_{53}+3a_{14}*a_{38}*a_{53}+3a_{35}*a_{38}*a_{53}+3a_{38}^{2}a_{53}-2a_{29}*a_{35}*a_{56}-4a_{30}*a_{35}*a_{56}+2a_{29}*a_{38}*a_{56}+4a_{30}*a_{38}*a_{56}+2a_{35}*a_{56}-2a_{38}*a_{56}$ * • $N=2212$ 1. 1. $a_{16}+2a_{18}-1$ 2. 2. $a_{17}-a_{18}$ 3. 3. $a_{28}+a_{29}+a_{30}-1$ 4. 4. $a_{13}+2a_{15}-1$ 5. 5. $a_{14}-a_{15}$ 6. 6. $a_{31}+2a_{33}-1$ 7. 7. $a_{32}-a_{33}$ 8. 8. $a_{7}+2a_{9}-1$ 9. 9. $a_{8}-a_{9}$ 10. 10. $a_{34}+2a_{36}-1$ 11. 11. $a_{35}-a_{36}$ 12. 12. $a_{37}+2a_{39}-1$ 13. 13. $a_{38}-a_{39}$ 14. 14. $a_{22}+2a_{24}-1$ 15. 15. $a_{23}-a_{24}$ 16. 16. $a_{43}+2a_{45}-1$ 17. 17. $a_{44}-a_{45}$ 18. 18. $a_{46}+2a_{48}-1$ 19. 19. $a_{47}-a_{48}$ 20. 20. $a_{49}+2a_{51}-1$ 21. 21. $a_{50}-a_{51}$ 22. 22. $a_{52}+2a_{54}-1$ 23. 23. $a_{53}-a_{54}$ 24. 24. $a_{55}+2a_{57}-1$ 25. 25. $a_{56}-a_{57}$ 26. 26. $a_{17}*a_{30}-a_{14}*a_{32}-a_{8}*a_{35}+a_{8}*a_{38}$ 27. 27. $3a_{8}*a_{30}+a_{8}-a_{23}-a_{32}$ 28. 28. $a_{23}*a_{29}+2a_{23}*a_{30}+a_{29}*a_{32}-a_{30}*a_{32}-a_{32}$ 29. 29. $a_{17}*a_{29}-a_{14}*a_{32}+2a_{8}*a_{35}-2a_{8}*a_{38}$ 30. 30. $a_{14}*a_{29}-a_{14}*a_{30}+a_{35}-a_{38}$ 31. 31. $3a_{8}*a_{29}-2a_{8}+2a_{23}-a_{32}$ 32. 32. $3a_{14}*a_{23}-3a_{8}*a_{35}+3a_{8}*a_{38}-a_{17}$ 33. 33. $3a_{8}*a_{14}-a_{17}$ 34. 34. $3a_{32}*a_{38}*a_{44}-2a_{29}*a_{38}*a_{47}-a_{30}*a_{38}*a_{47}-3a_{32}*a_{38}*a_{50}+3a_{32}*a_{35}*a_{53}+3a_{8}*a_{38}*a_{53}-3a_{23}*a_{38}*a_{53}+a_{29}*a_{32}*a_{56}-a_{30}*a_{32}*a_{56}-a_{17}*a_{44}+a_{14}*a_{47}-a_{35}*a_{47}+a_{38}*a_{47}+a_{17}*a_{50}-a_{17}*a_{53}-a_{8}*a_{56}+a_{23}*a_{56}$ 35. 35. $3a_{8}*a_{35}*a_{44}-3a_{32}*a_{35}*a_{44}-3a_{8}*a_{38}*a_{44}+a_{29}*a_{35}*a_{47}+2a_{30}*a_{35}*a_{47}+3a_{8}*a_{35}*a_{50}-3a_{23}*a_{38}*a_{50}+a_{29}*a_{32}*a_{56}-a_{30}*a_{32}*a_{56}+a_{17}*a_{44}-a_{14}*a_{47}-a_{35}*a_{47}+a_{38}*a_{47}-a_{8}*a_{56}+a_{23}*a_{56}$ 36. 36. $3a_{17}*a_{32}*a_{44}^{2}-6a_{14}*a_{32}*a_{44}*a_{47}+3a_{32}*a_{35}*a_{44}*a_{47}+3a_{14}*a_{30}*a_{47}^{2}-a_{29}*a_{35}*a_{47}^{2}-2a_{30}*a_{35}*a_{47}^{2}-3a_{8}*a_{17}*a_{44}*a_{50}-3a_{17}*a_{32}*a_{44}*a_{50}+3a_{14}*a_{32}*a_{47}*a_{50}-3a_{8}*a_{35}*a_{47}*a_{50}+3a_{23}*a_{38}*a_{47}*a_{50}+3a_{8}*a_{17}*a_{50}^{2}-3a_{17}*a_{23}*a_{44}*a_{53}+3a_{8}*a_{35}*a_{47}*a_{53}-3a_{8}*a_{38}*a_{47}*a_{53}-3a_{17}*a_{23}*a_{50}*a_{53}+6a_{8}*a_{32}*a_{44}*a_{56}-a_{29}*a_{32}*a_{47}*a_{56}+a_{30}*a_{32}*a_{47}*a_{56}-3a_{8}^{2}*a_{50}*a_{56}+3a_{8}*a_{23}*a_{50}*a_{56}-3a_{8}*a_{32}*a_{50}*a_{56}+3a_{8}^{2}*a_{53}*a_{56}-3a_{8}*a_{23}*a_{53}*a_{56}+3a_{8}*a_{32}*a_{53}*a_{56}-a_{35}*a_{47}^{2}+a_{38}*a_{47}^{2}+a_{17}*a_{47}*a_{50}+a_{17}*a_{47}*a_{53}-2a_{32}*a_{47}*a_{56}$ 37. 37. $9a_{8}*a_{23}*a_{38}*a_{44}*a_{50}-9a_{8}*a_{23}*a_{38}*a_{50}^{2}+9a_{23}*a_{32}*a_{35}*a_{44}*a_{53}+9a_{8}*a_{23}*a_{38}*a_{44}*a_{53}+3a_{29}*a_{32}*a_{35}*a_{47}*a_{53}-3a_{30}*a_{32}*a_{35}*a_{47}*a_{53}+9a_{23}^{2}*a_{38}*a_{50}*a_{53}+9a_{23}*a_{30}*a_{32}*a_{53}*a_{56}+3a_{29}*a_{32}^{2}*a_{53}*a_{56}-3a_{30}*a_{32}^{2}*a_{53}*a_{56}-3a_{17}*a_{23}*a_{44}^{2}-3a_{23}*a_{35}*a_{44}*a_{47}+6a_{32}*a_{35}*a_{44}*a_{47}-3a_{29}*a_{35}*a_{47}^{2}-3a_{30}*a_{35}*a_{47}^{2}+3a_{17}*a_{23}*a_{44}*a_{50}-9a_{8}*a_{35}*a_{47}*a_{50}+3a_{8}*a_{38}*a_{47}*a_{50}+3a_{23}*a_{38}*a_{47}*a_{50}-3a_{17}*a_{23}*a_{44}*a_{53}+3a_{8}*a_{35}*a_{47}*a_{53}-3a_{32}*a_{35}*a_{47}*a_{53}-3a_{8}*a_{38}*a_{47}*a_{53}-3a_{23}*a_{38}*a_{47}*a_{53}-6a_{8}*a_{23}*a_{44}*a_{56}-3a_{23}*a_{30}*a_{47}*a_{56}-3a_{29}*a_{32}*a_{47}*a_{56}+3a_{30}*a_{32}*a_{47}*a_{56}+6a_{8}*a_{23}*a_{50}*a_{56}-3a_{23}^{2}*a_{50}*a_{56}-3a_{8}*a_{23}*a_{53}*a_{56}-3a_{32}^{2}*a_{53}*a_{56}+a_{14}*a_{47}^{2}+2a_{35}*a_{47}^{2}-a_{38}*a_{47}^{2}-a_{17}*a_{47}*a_{50}+a_{17}*a_{47}*a_{53}+2a_{8}*a_{47}*a_{56}+a_{32}*a_{47}*a_{56}$ 38. 38. $9a_{23}*a_{32}*a_{35}*a_{44}^{2}+9a_{23}*a_{30}*a_{35}*a_{44}*a_{47}+a_{29}^{2}*a_{35}*a_{47}^{2}+4a_{29}*a_{30}*a_{35}*a_{47}^{2}-5a_{30}^{2}*a_{35}*a_{47}^{2}-18a_{23}*a_{32}*a_{35}*a_{44}*a_{50}-6a_{29}*a_{32}*a_{35}*a_{47}*a_{50}+6a_{30}*a_{32}*a_{35}*a_{47}*a_{50}+27a_{23}*a_{30}*a_{38}*a_{47}*a_{50}+9a_{29}*a_{32}*a_{38}*a_{47}*a_{50}-9a_{30}*a_{32}*a_{38}*a_{47}*a_{50}+18a_{8}*a_{23}*a_{35}*a_{50}^{2}-9a_{8}*a_{23}*a_{38}*a_{50}^{2}-9a_{23}^{2}*a_{38}*a_{50}^{2}-9a_{23}^{2}*a_{35}*a_{44}*a_{53}+9a_{23}*a_{32}*a_{35}*a_{44}*a_{53}+9a_{8}*a_{23}*a_{38}*a_{44}*a_{53}+9a_{23}*a_{30}*a_{35}*a_{47}*a_{53}+6a_{29}*a_{32}*a_{35}*a_{47}*a_{53}-6a_{30}*a_{32}*a_{35}*a_{47}*a_{53}-9a_{23}^{2}*a_{35}*a_{50}*a_{53}+9a_{23}^{2}*a_{38}*a_{50}*a_{53}+9a_{23}*a_{30}*a_{32}*a_{44}*a_{56}+3a_{29}*a_{32}^{2}*a_{44}*a_{56}-3a_{30}*a_{32}^{2}*a_{44}*a_{56}+9a_{23}*a_{30}^{2}*a_{47}*a_{56}+a_{29}^{2}*a_{32}*a_{47}*a_{56}+a_{29}*a_{30}*a_{32}*a_{47}*a_{56}-2a_{30}^{2}*a_{32}*a_{47}*a_{56}+9a_{23}^{2}*a_{30}*a_{50}*a_{56}-27a_{23}*a_{30}*a_{32}*a_{50}*a_{56}-9a_{29}*a_{32}^{2}*a_{50}*a_{56}+9a_{30}*a_{32}^{2}*a_{50}*a_{56}-9a_{23}^{2}*a_{30}*a_{53}*a_{56}+18a_{23}*a_{30}*a_{32}*a_{53}*a_{56}+6a_{29}*a_{32}^{2}*a_{53}*a_{56}-6a_{30}*a_{32}^{2}*a_{53}*a_{56}-3a_{17}*a_{23}*a_{44}^{2}-9a_{23}*a_{35}*a_{44}*a_{47}-3a_{32}*a_{35}*a_{44}*a_{47}+9a_{23}*a_{38}*a_{44}*a_{47}-3a_{29}*a_{35}*a_{47}^{2}+3a_{29}*a_{38}*a_{47}^{2}-3a_{30}*a_{38}*a_{47}^{2}+6a_{17}*a_{23}*a_{44}*a_{50}-3a_{8}*a_{35}*a_{47}*a_{50}-9a_{23}*a_{35}*a_{47}*a_{50}+6a_{32}*a_{35}*a_{47}*a_{50}+6a_{8}*a_{38}*a_{47}*a_{50}+6a_{23}*a_{38}*a_{47}*a_{50}-9a_{32}*a_{38}*a_{47}*a_{50}-3a_{17}*a_{23}*a_{44}*a_{53}+3a_{8}*a_{35}*a_{47}*a_{53}+3a_{23}*a_{35}*a_{47}*a_{53}-6a_{32}*a_{35}*a_{47}*a_{53}-3a_{8}*a_{38}*a_{47}*a_{53}-3a_{23}*a_{38}*a_{47}*a_{53}-3a_{8}*a_{23}*a_{44}*a_{56}+3a_{23}^{2}*a_{44}*a_{56}-3a_{32}^{2}*a_{44}*a_{56}-12a_{23}*a_{30}*a_{47}*a_{56}-4a_{29}*a_{32}*a_{47}*a_{56}+a_{30}*a_{32}*a_{47}*a_{56}-3a_{23}*a_{32}*a_{50}*a_{56}+9a_{32}^{2}*a_{50}*a_{56}-3a_{8}*a_{23}*a_{53}*a_{56}+3a_{23}^{2}*a_{53}*a_{56}+3a_{23}*a_{32}*a_{53}*a_{56}-6a_{32}^{2}*a_{53}*a_{56}+a_{14}*a_{47}^{2}+4a_{35}*a_{47}^{2}-4a_{38}*a_{47}^{2}-2a_{17}*a_{47}*a_{50}+a_{17}*a_{47}*a_{53}-a_{8}*a_{47}*a_{56}+a_{23}*a_{47}*a_{56}+4a_{32}*a_{47}*a_{56}$ 39. 39. $3a_{14}*a_{17}^{2}*a_{44}^{2}-6a_{14}^{2}*a_{17}*a_{44}*a_{47}+3a_{14}*a_{17}*a_{35}*a_{44}*a_{47}+3a_{14}^{3}*a_{47}^{2}-3a_{14}^{2}*a_{35}*a_{47}^{2}-3a_{14}*a_{17}^{2}*a_{44}*a_{50}-3a_{17}^{2}*a_{38}*a_{44}*a_{50}+3a_{14}^{2}*a_{17}*a_{47}*a_{50}+3a_{14}*a_{17}*a_{38}*a_{47}*a_{50}+3a_{17}^{2}*a_{38}*a_{50}^{2}-3a_{17}^{2}*a_{35}*a_{44}*a_{53}+3a_{14}*a_{17}*a_{35}*a_{47}*a_{53}-3a_{17}^{2}*a_{35}*a_{50}*a_{53}+3a_{8}*a_{17}*a_{35}*a_{50}*a_{56}-3a_{8}*a_{17}*a_{38}*a_{50}*a_{56}-3a_{8}*a_{17}*a_{35}*a_{53}*a_{56}+3a_{8}*a_{17}*a_{38}*a_{53}*a_{56}+2a_{17}^{2}*a_{44}*a_{56}-2a_{14}*a_{17}*a_{47}*a_{56}+a_{17}*a_{35}*a_{47}*a_{56}-a_{17}*a_{38}*a_{47}*a_{56}-a_{17}^{2}*a_{50}*a_{56}+a_{17}^{2}*a_{53}*a_{56}$ 40. 40. $3a_{29}^{2}*a_{35}*a_{38}*a_{44}+3a_{29}*a_{30}*a_{35}*a_{38}*a_{44}-6a_{30}^{2}*a_{35}*a_{38}*a_{44}+3a_{29}^{2}*a_{35}*a_{38}*a_{50}+12a_{29}*a_{30}*a_{35}*a_{38}*a_{50}+12a_{30}^{2}*a_{35}*a_{38}*a_{50}-6a_{29}^{2}*a_{38}^{2}*a_{50}+3a_{29}*a_{30}*a_{38}^{2}*a_{50}+3a_{30}^{2}*a_{38}^{2}*a_{50}-3a_{29}^{2}*a_{35}^{2}*a_{53}-12a_{29}*a_{30}*a_{35}^{2}*a_{53}-12a_{30}^{2}*a_{35}^{2}*a_{53}-3a_{29}^{2}*a_{35}*a_{38}*a_{53}-3a_{29}*a_{30}*a_{35}*a_{38}*a_{53}+6a_{30}^{2}*a_{35}*a_{38}*a_{53}-a_{29}^{3}*a_{35}*a_{56}-3a_{29}^{2}*a_{30}*a_{35}*a_{56}+4a_{30}^{3}*a_{35}*a_{56}-2a_{29}^{3}*a_{38}*a_{56}-3a_{29}^{2}*a_{30}*a_{38}*a_{56}+3a_{29}*a_{30}^{2}*a_{38}*a_{56}+2a_{30}^{3}*a_{38}*a_{56}+3a_{29}*a_{35}^{2}*a_{44}+6a_{30}*a_{35}^{2}*a_{44}-6a_{29}*a_{35}*a_{38}*a_{44}-3a_{30}*a_{35}*a_{38}*a_{44}+3a_{29}*a_{38}^{2}*a_{44}-3a_{30}*a_{38}^{2}*a_{44}-9a_{14}*a_{30}*a_{35}*a_{50}-9a_{14}*a_{30}*a_{38}*a_{50}-12a_{29}*a_{35}*a_{38}*a_{50}-6a_{30}*a_{35}*a_{38}*a_{50}+12a_{29}*a_{38}^{2}*a_{50}+6a_{30}*a_{38}^{2}*a_{50}+18a_{14}*a_{30}*a_{35}*a_{53}+3a_{29}*a_{35}^{2}*a_{53}+6a_{30}*a_{35}^{2}*a_{53}-9a_{30}*a_{35}*a_{38}*a_{53}-3a_{29}*a_{38}^{2}*a_{53}+3a_{30}*a_{38}^{2}*a_{53}+a_{29}^{2}*a_{35}*a_{56}+a_{29}*a_{30}*a_{35}*a_{56}-2a_{30}^{2}*a_{35}*a_{56}+2a_{29}^{2}*a_{38}*a_{56}-a_{29}*a_{30}*a_{38}*a_{56}-a_{30}^{2}*a_{38}*a_{56}-3a_{14}*a_{35}*a_{44}-3a_{35}^{2}*a_{44}+3a_{14}*a_{38}*a_{44}+9a_{35}*a_{38}*a_{44}-6a_{38}^{2}*a_{44}+3a_{14}^{2}*a_{50}+6a_{14}*a_{35}*a_{50}-6a_{14}*a_{38}*a_{50}+3a_{35}*a_{38}*a_{50}-3a_{38}^{2}*a_{50}-3a_{14}^{2}*a_{53}-3a_{14}*a_{35}*a_{53}-6a_{35}^{2}*a_{53}+3a_{14}*a_{38}*a_{53}+3a_{35}*a_{38}*a_{53}+3a_{38}^{2}*a_{53}-2a_{29}*a_{35}*a_{56}-4a_{30}*a_{35}*a_{56}+2a_{29}*a_{38}*a_{56}+4a_{30}*a_{38}*a_{56}+2a_{35}*a_{56}-2a_{38}*a_{56}$ * • $N=2122$ (subset) 1. 1. $a_{7}+2a_{9}-1$ 2. 2. $a_{8}-a_{9}$ 3. 3. $a_{28}+a_{29}+a_{30}-1$ 4. 4. $a_{22}+2a_{24}-1$ 5. 5. $a_{23}-a_{24}$ 6. 6. $a_{31}+2a_{33}-1$ 7. 7. $a_{32}-a_{33}$ 8. 8. $a_{4}+2a_{6}-1$ 9. 9. $a_{5}-a_{6}$ 10. 10. $a_{19}+2a_{21}-1$ 11. 11. $a_{20}-a_{21}$ 12. 12. $a_{25}+2a_{27}-1$ 13. 13. $a_{26}-a_{27}$ 14. 14. $a_{1}+2a_{3}-1$ 15. 15. $a_{2}-a_{3}$ 16. 16. $3a_{8}*a_{30}+a_{8}-a_{23}-a_{32}$ 17. 17. $a_{23}*a_{29}+2a_{23}*a_{30}+a_{29}*a_{32}-a_{30}*a_{32}-a_{32}$ 18. 18. $3a_{8}*a_{29}-2a_{8}+2a_{23}-a_{32}$ 19. 19. $a_{5}*a_{29}+2a_{5}*a_{30}-a_{5}+a_{20}-a_{26}$ 20. 20. $a_{2}*a_{29}+2a_{2}*a_{30}-3a_{5}*a_{32}$ 21. 21. $3a_{5}*a_{23}-3a_{2}*a_{30}+3a_{5}*a_{32}-a_{2}$ 22. 22. $3a_{8}*a_{20}-3a_{8}*a_{26}+3a_{5}*a_{32}-a_{2}$ 23. 23. $3a_{5}*a_{8}-a_{2}$ * • $N=2221$ (subset) 1. 1. $a_{13}+2a_{15}-1$ 2. 2. $a_{14}-a_{15}$ 3. 3. $a_{28}+a_{29}+a_{30}-1$ 4. 4. $a_{34}+2a_{36}-1$ 5. 5. $a_{35}-a_{36}$ 6. 6. $a_{37}+2a_{39}-1$ 7. 7. $a_{38}-a_{39}$ 8. 8. $a_{10}+2a_{12}-1$ 9. 9. $a_{11}-a_{12}$ 10. 10. $a_{4}+2a_{6}-1$ 11. 11. $a_{5}-a_{6}$ 12. 12. $a_{19}+2a_{21}-1$ 13. 13. $a_{20}-a_{21}$ 14. 14. $a_{25}+2a_{27}-1$ 15. 15. $a_{26}-a_{27}$ 16. 16. $a_{14}*a_{29}-a_{14}*a_{30}+a_{35}-a_{38}$ 17. 17. $a_{11}*a_{29}-a_{11}*a_{30}+3a_{5}*a_{35}-3a_{5}*a_{38}$ 18. 18. $a_{5}*a_{29}+2a_{5}*a_{30}-a_{5}+a_{20}-a_{26}$ 19. 19. $3a_{14}*a_{20}-3a_{14}*a_{26}+3a_{11}*a_{30}-3a_{5}*a_{35}+3a_{5}*a_{38}-a_{11}$ 20. 20. $3a_{5}*a_{14}-a_{11}$ 21. 21. $a_{20}*a_{29}*a_{35}+2a_{20}*a_{30}*a_{35}-a_{26}*a_{29}*a_{38}+a_{26}*a_{30}*a_{38}-a_{14}*a_{26}+a_{11}*a_{30}-a_{5}*a_{35}-a_{20}*a_{35}-a_{26}*a_{35}+2a_{20}*a_{38}$ * • $N=2222$ (subset) 1. 1. $a_{16}+2a_{18}-1$ 2. 2. $a_{17}-a_{18}$ 3. 3. $a_{28}+a_{29}+a_{30}-1$ 4. 4. $a_{13}+2a_{15}-1$ 5. 5. $a_{14}-a_{15}$ 6. 6. $a_{31}+2a_{33}-1$ 7. 7. $a_{32}-a_{33}$ 8. 8. $a_{7}+2a_{9}-1$ 9. 9. $a_{8}-a_{9}$ 10. 10. $a_{34}+2a_{36}-1$ 11. 11. $a_{35}-a_{36}$ 12. 12. $a_{37}+2a_{39}-1$ 13. 13. $a_{38}-a_{39}$ 14. 14. $a_{22}+2a_{24}-1$ 15. 15. $a_{23}-a_{24}$ 16. 16. $a_{10}+2a_{12}-1$ 17. 17. $a_{11}-a_{12}$ 18. 18. $a_{4}+2a_{6}-1$ 19. 19. $a_{5}-a_{6}$ 20. 20. $a_{19}+2a_{21}-1$ 21. 21. $a_{20}-a_{21}$ 22. 22. $a_{25}+2a_{27}-1$ 23. 23. $a_{26}-a_{27}$ 24. 24. $a_{1}+2a_{3}-1$ 25. 25. $a_{2}-a_{3}$ 26. 26. $a_{17}*a_{30}-a_{14}*a_{32}-a_{8}*a_{35}+a_{8}*a_{38}$ 27. 27. $3a_{8}*a_{30}+a_{8}-a_{23}-a_{32}$ 28. 28. $a_{23}*a_{29}+2a_{23}*a_{30}+a_{29}*a_{32}-a_{30}*a_{32}-a_{32}$ 29. 29. $a_{17}*a_{29}-a_{14}*a_{32}+2a_{8}*a_{35}-2a_{8}*a_{38}$ 30. 30. $a_{14}*a_{29}-a_{14}*a_{30}+a_{35}-a_{38}$ 31. 31. $a_{11}*a_{29}-a_{11}*a_{30}+3a_{5}*a_{35}-3a_{5}*a_{38}$ 32. 32. $3a_{8}*a_{29}-2a_{8}+2a_{23}-a_{32}$ 33. 33. $a_{5}*a_{29}+2a_{5}*a_{30}-a_{5}+a_{20}-a_{26}$ 34. 34. $a_{2}*a_{29}+2a_{2}*a_{30}-3a_{5}*a_{32}$ 35. 35. $3a_{14}*a_{23}-3a_{8}*a_{35}+3a_{8}*a_{38}-a_{17}$ 36. 36. $3a_{5}*a_{23}-3a_{2}*a_{30}+3a_{5}*a_{32}-a_{2}$ 37. 37. $a_{17}*a_{20}-a_{11}*a_{23}-a_{17}*a_{26}+a_{11}*a_{32}+a_{2}*a_{35}-a_{2}*a_{38}$ 38. 38. $3a_{14}*a_{20}-3a_{14}*a_{26}+3a_{11}*a_{30}-3a_{5}*a_{35}+3a_{5}*a_{38}-a_{11}$ 39. 39. $3a_{8}*a_{20}-3a_{8}*a_{26}+3a_{5}*a_{32}-a_{2}$ 40. 40. $a_{5}*a_{17}-a_{11}*a_{23}+a_{2}*a_{35}-a_{2}*a_{38}$ 41. 41. $3a_{8}*a_{14}-a_{17}$ 42. 42. $3a_{5}*a_{14}-a_{11}$ 43. 43. $a_{2}*a_{14}-a_{11}*a_{23}+a_{2}*a_{35}-a_{2}*a_{38}$ 44. 44. $a_{8}*a_{11}-a_{11}*a_{23}+a_{2}*a_{35}-a_{2}*a_{38}$ 45. 45. $3a_{5}*a_{8}-a_{2}$ 46. 46. $a_{20}*a_{29}*a_{35}+2a_{20}*a_{30}*a_{35}-a_{26}*a_{29}*a_{38}+a_{26}*a_{30}*a_{38}-a_{14}*a_{26}+a_{11}*a_{30}-a_{5}*a_{35}-a_{20}*a_{35}-a_{26}*a_{35}+2a_{20}*a_{38}$ 47. 47. $6a_{8}*a_{26}*a_{35}-3a_{5}*a_{32}*a_{35}-3a_{20}*a_{32}*a_{35}-3a_{8}*a_{26}*a_{38}-3a_{23}*a_{26}*a_{38}+6a_{5}*a_{32}*a_{38}+a_{17}*a_{26}-a_{11}*a_{32}+a_{2}*a_{35}-a_{2}*a_{38}$ ## Appendix C Simulation study Figure 13: Invariant score (y-axis) measured as $L_{2}$-norm of the phylogenetic invariants for each network (x-axis) for the true network (red) and its symmetric network (inverted clades $n_{1}$ and $n_{2}$, in blue) on the cases of true and Gaussian-perturbed CFs. Each panel corresponds to a type of simulation: using true concordance factors (left) and using concordance factors with added Gaussian noise (with increasing standard deviation for noise from left to right). Both the true and symmetric networks have invariant score close to zero, and are thus, easy to distinguish from wrong networks (whose invariant score is far from zero). As the noise increases, the invariant scores moves away from zero, but still within $10^{-4}$. Figure 14: Invariant score (y-axis) measured as $L_{2}$-norm of the phylogenetic invariants for each network (x-axis) for the true network (red) and its symmetric network (inverted clades $n_{1}$ and $n_{2}$, in blue) for the case of true simulated gene trees (“g.t."). Each panel corresponds to a number of simulated gene trees from 100 (left) to 10,000 (right). With true and perturbed concordance factors (top left and top right, respectively). As the number of gene trees increases, the invariant scores of the true and symmetric networks converge to zero, and they are thus, easy to distinguish from wrong networks (whose invariant score is far from zero). Figure 15: Invariant score (y-axis) measured as $L_{2}$-norm of the phylogenetic invariants for each network (x-axis) for the true network (red) and its symmetric network (inverted clades $n_{1}$ and $n_{2}$, in blue) for the case of estimated gene trees. Each panel corresponds to a number of gene trees (g.t. from 1000 to 10,000) and sequence length ($L$ from 500 to 2000). As the number of gene trees increases, the invariant scores of the true and symmetric networks converge to zero, and they are thus, easy to distinguish from wrong networks (whose invariant score is far from zero).
# CLIP-Nav: Using CLIP for Zero-Shot Vision-and-Language Navigation Vishnu Sashank Dorbala University of Maryland, College Park &Gunnar Sigurdsson Amazon Alexa AI &Robinson Piramuthu Amazon Alexa AI &Jesse Thomason Amazon Alexa AI &Gaurav S. Sukhatme Amazon Alexa AI ###### Abstract Household environments are visually diverse. Embodied agents performing Vision-and-Language Navigation (VLN) in the wild must be able to handle this diversity, while also following arbitrary language instructions. Recently, Vision-Language models like CLIP have shown great performance on the task of zero-shot object recognition. In this work, we ask if these models are also capable of zero-shot language grounding. In particular, we utilize CLIP to tackle the novel problem of zero-shot VLN using natural language referring expressions that describe target objects, in contrast to past work that used simple language templates describing object classes. We examine CLIP’s capability in making sequential navigational decisions without any dataset- specific finetuning, and study how it influences the path that an agent takes. Our results on the coarse-grained instruction following task of REVERIE demonstrate the navigational capability of CLIP, surpassing the supervised baseline in terms of both success rate (SR) and success weighted by path length (SPL). More importantly, we quantitatively show that our CLIP-based zero-shot approach generalizes better to show consistent performance across environments when compared to SOTA, fully supervised learning approaches when evaluated via Relative Change in Success (RCS). ## 1 Introduction Vision-and-Language Navigation (VLN) requires agents to follow human instructions in unseen environments. This is a challenging problem since instructions heavily rely on the contents of the scene, possibly unknown to a general agent. A common approach to VLN is to use supervised learning; we argue that this is not practical, given the drastic shift in semantics from scene to scene that impacts the performance of a trained model. We observe this phenomenon in SOTA supervised learning approaches for VLN, which have large performance drops on “unseen” environments that are absent from the training dataset. In this work, we seek to address this issue of generalizing to new environments, and propose to solve VLN in a fully zero-shot manner. The agent is assumed to have no prior knowledge about the instructions or the environment. This setting is does not rely on a particular VLN dataset and is free from any environmental bias that datasets may inherently have. A substantial body of prior work tackles VLN using supervised learning Gu et al. (2022); Guhur et al. (2021); Wang et al. (2019b); Hong et al. (2021). Models are first trained on a set of “seen” environments and instructions, before being evaluated on both “seen” and “unseen” test data. This paradigm often involves encoding instructions and learning topological relations during training Chen et al. (2021); Wang et al. (2021); Chen et al. (2022a), and imitating navigational decisions during inference. These approaches usually see a significant drop in performance on unseen data. Figure 1 compares results of popular VLN approaches on the coarse-grained instruction following task of REVERIE Qi et al. (2020). Observe the significant drop in performance on the “Unseen Val” and “Unseen Test” sets, when compared with “Seen Val” which validates on previously seen environments. This drop holds across all approaches. Performance in comparison with humans is also significantly worse, not just for REVERIE, but also for other popular instruction following tasks such as R2R Anderson et al. (2018) and SOON Zhu et al. (2021). Figure 1: Comparing Model Performance on REVERIE Seen and Unseen Splits: Observe the significant drop in performance of the Unseen Val and Unseen Test sets (blue) when compared with the Val Seen set (green), across all approaches as measured by both Success Rate (SR) (Left) and Success Weighted by Path Length (SPL) (Right). Also observe the poor Unseen Test set performance when compared to the human baseline. Current VLN models are also dataset specific; a trained model from one will not generalize to another. For instance, training on REVERIE and testing on SOON may not give comparable results, despite both involving the coarse- grained instruction following task. A critical factor affecting this generalization is the lack of sufficient training data, and a few articles address this issue by proposing data augmentation techniques Chen et al. (2022b); Li et al. (2022). However, data augmentation has only shown minor improvements in performance. Figure 2: Household images generated from text using a latent diffusion model Rombach et al. (2022). Observe the variance in layout, positioning and lighting. Each home is visually unique, and we hypothesize that this causes people to use unique, environment-specific language in giving out instructions (Orange-striped bathroom for instance). This hypothesis is substantiated by our inspection of REVERIE, and motivates us to treat VLN as a fully zero-shot problem. Household environments are visually diverse, each with unique layouts and objects. As such, instructions given by humans to describe target locations usually contain environment-specific cues. For example, the same target item “bottle” could be described as “next to the mirror” in one instruction and as “above the sink” in another. There could also be unique environment-specific identifiers such as “orange-striped bathroom” or “painting depicting crudely- drawn people”. We hypothesize that the visual distinctiveness of environments causes people to use unique language involving those scenes. In our inspection of the REVERIE dataset we noticed a low pairwise average cosine similarity of $0.32$ between all seen and unseen instructions, substantiating this claim. As all environments are potentially distinct from one another, training datasets fail to capture all the visuo-linguistic diversity needed to generalize (Figure 2). Our work makes the following contributions: * • We present a novel approach to solve coarse-grained instruction following tasks: breaking down guidance language into keyphrases, visually grounding them, and using the grounding scores to drive CLIP-Nav, a novel “zero-shot” navigation scheme. * • Further, we incorporate backtracking into our scheme, and present Seq CLIP-Nav which drastically improves CLIP-Nav’s results, showcasing the importance of backtracking in solving such tasks. * • Our results establish a zero-shot baseline on the task of REVERIE Qi et al. (2020), surpassing the unseen supervised baseline without any form of dataset- specific finetuning in terms of SR and SPL. Our SPL results on this dataset also improves on the SOTA. * • Finally, we establish a new metric to measure generalizability in VLN tasks — Relative Change in Success (RCS). This metric quantitatively showcases the improved performance of our CLIP-based approaches over other supervised methods. ## 2 Related Work Vision-and-Language Navigation (VLN): Coarse-grained Qi et al. (2020); Zhu et al. (2021) and fine-grained Anderson et al. (2018); Ku et al. (2020); Vasudevan et al. (2021) instruction following tasks have recently been of great interest. A majority of approaches attempting to solve these tasks use some form of supervision, employing behavior cloning Huang et al. (2019), reinforcement learning Wang et al. (2018), or even imitation learning Wang et al. (2019a). These approaches are limited to solving VLN either on a single or a specific set of datasets, and do little to analyze cross-dataset performance. Since VLN by definition is intended to be solved in arbitrary environments, we argue that methods should give consistent performance across all scenes, previously seen or not. Zero-Shot VLN: A recent paradigm shift in machine learning has led to the emergence of large deep learning models that are pre-trained on vast amounts of unlabeled data, and finetuned on a variety of downstream tasks. Examples of this include BERT Devlin et al. (2018) and GPT-3 Brown et al. (2020), which have shown SOTA performance on various natural language tasks. For zero-shot VLN, we are particularly interested utilizing CLIP Radford et al. (2021), which has shown improved zero-shot performance on several downstream vision- language tasks Shen et al. (2021). Gadre et al. (2022) in CLIP on Wheels (CoW) attempt to utilize CLIP to perform zero-shot object navigation. Object navigation involves exploring the environment to find a target object without any guidance instructions. CoW modularizes this task into exploration and object localization, and uses CLIP for the latter. While underlying task of navigating to a target location is the same, VLN is very different from object navigation. VLN uses natural human instruction language as opposed to template language. For example, a coarse-grained VLN instruction would be “Go to the kitchen on your right and water the plant there.”, while object navigation uses only the target word “plant”. CoW augments these words with templates (a photo of a $<$object$>$ in a video game) and then grounds them with CLIP. We instead use CLIP not just to ground keyphrases extracted from the VLN instruction, but also to drive our exploration policy. To the best of our knowledge, there are no works that attempt to solve coarse- grained instruction following in a generalized, zero-shot setting, which forms the basis of our research. We look at solving VLN without any form of dataset- specific finetuning, and seek to transfer CLIP’s powerful language grounding capabilities into a sequential decision making pipeline for navigation. ## 3 Approach Performing zero-shot VLN in a household environment requires the agent to have a sense of structural priors to make sequential decisions about how to get to an unseen target location. Consider Figure 3 for instance. The command is for the robot to “Go to the kitchen”, but the panoramic view does not contain visual elements associated with a kitchen. In such a case, we require the agent to pick the best view (and consequently the best direction) that would potentially lead to a kitchen. In this work, we use CLIP to make sequential navigational decisions, asserting its capability in capturing structural priors of indoor hosuehold environments. Figure 3: We look at CLIP’s ability to make sequential navigational decisions. Here, the instruction “Go to the kitchen” suggests that the agent needs to leave the room. However, in order for it to make this decision, it needs to ground this instruction within the panorama, to choose a view with the door leading outwards. Notice that there are no clear visual entities (i.e. spoons or sinks) to suggest the chosen image (in red) is a “kitchen”. The decision is based on pretrained CLIP’s structural prior of the household in picking a view that might lead to the kitchen. Our approach consists of three steps: 1. 1. Instruction Breakdown \- Decompose coarse-grained instructions into keyphrases. 2. 2. Vision-Language Grounding \- Ground keyphrases in the environment using CLIP. 3. 3. Zero-Shot Navigation \- Utilize the CLIP scores to make navigational decisions. Given that our approach is zero-shot and requires no finetuning, we could choose any fixed set of datapoints for evaluation. We opt to use the Seen and Unseen validation splits of the REVERIE Qi et al. (2020), a popular dataset to conduct our experiments. Cross-comparing results between these splits helps us infer the generalizability of our approach, while also enabling similar comparisons on other supervised learning approaches on this dataset. REVERIE contains several human-annotated instructions for paths taken from the Matterport3D Chang et al. (2017) environment. The language used in the instructions uniquely capture various aspects of the environment. Each path is described as a discrete set of adjacent photorealistic panoramic images. They contain around 8 adjacent pano images (or hops) on average, which have been annotated with coarse-grained language guidance to reach target locations. ### 3.1 Instruction Breakdown We first look at breaking down coarse-grained guidance into keyphrases. Our objective is for the keyphrases to contain intermediate goals for the agent to use for sequential decision-making. Consider the instruction: On the second level go the bathroom inside the second bedroom on the right and replace the towels on the towel rack with the clean towels from the linen closet. Observe that the conjunction and separates the Navigational Component (NC) of the instruction from the Activity Component (AC). As such, a coarse-grained instruction can be broken down as, Coarse Instruction (I) = Navigation Component (NC) + Activity Component (AC) We empirically observe that this sentence form holds true with most instructions in REVERIE. The NC tells the agent about how to get to the target location, while the AC tells it what to do once it has reached there. In our case, since our objective is solely navigation, we focus on breaking up the NC into keyphrases and using the AC for target object grounding. One way to obtain keyphrases is to use prepositions in the instruction to break it up. Using this form of retrieval on the example above gives us: While this approach is simplistic, and breaks down instructions into rudimentary keyphrases for grounding, it often fails to capture the temporal nature of the instruction. Notice that “go to the bathroom” comes before “inside the second bedroom”, when the agent needs to enter the second bedroom before going to the bathroom. Alternatively, we look at using a Large Language Model to break it down for us. Using GPT-3 Brown et al. (2020) with the phrase “Break this down into steps.” gives: The instruction breakdown while being far more articulate, and is able to capture the sequential timeline i.e., “Enter the second bedroom” is before “Go to the bathroom”. ### 3.2 Language Grounding using CLIP After breaking down instructions into keyphrases, we use them to navigate inside a Matterport3D environment. Each node in Matterport3D is a panoramic image covering a $360$ degree space around the agent. In order to select a direction for navigation within this panorama, we split it into 4 separate images. Each of these images covers approximately a $90$ degree space around the agent, at a uniform horizontal elevation. We use CLIP to ground keyphrases from both the Navigational (NC) and Activity components (AC) obtained in the instruction breakdown stage, and obtain two scores - a Keyphrase Grounding Score (KGS) and an AC grounding score. KGS helps the agent make sequential navigational decisions, while AC grounding score helps identify if the agent has reached the target location. At each timestep, the image with the highest KGS is selected as the CLIP- chosen image, and is used to drive our navigational scheme. The AC grounding scores are for target object grounding, which gives us threshold or ‘Stop Condition’ for when the agent needs to stop navigating. Figure 4: CLIP Grounding \- We ground the Navigational Component (NC) on all the split images to obtain Keyphrase Grounding Scores (KGS). The “CLIP-chosen image” (highlighted in red) represents the one with the highest KGS, which drives our navigation algorithms. We also simultaneously ground the AC, and use the grounding score to determine if the agent has reached the target location—our “Stop Condition”. The red image in Figure 4 represents the CLIP-chosen image for the given instruction. We observe that changing the instruction by adding more information to it also changes the CLIP-chosen image. For example, if the instruction here were “Go into to the kitchen that is next to the balcony on the second level” instead, the leftmost image is no longer the one with the highest grounding score. CLIP is sensitive to the language given to it for grounding, and this can hinder navigation performance. As such, we limit the words in the extracted keyphrases before computing CLIP scores. ### 3.3 Zero-Shot Navigation The “CLIP-chosen image” obtained drives both our zero-shot navigation strategies. #### 3.3.1 CLIP-Nav Figure 5 shows the overview of CLIP-Nav. At each time step, we split the panorama into 4 images, and obtain the CLIP-chosen image as explained in section 3.2. We obtain adjacent navigable nodes visible from this image using the Matterport Simulator, and choose the closest node. This is done iteratively till the ‘Stop Condition’ described in the previous section is reached. Beyond extracting CLIP-chosen images, the NC grounding score also determines when to select the next keyphrase. For instance, for “Go to the kitchen”, if the grounding score is above a certain threshold, we assume that the agent has successfully navigated to the kitchen, and needs to execute the next keyphrase - “next to the balcony”. In this way, we utilize CLIP not just for choosing navigational directions, but also for determining when the agent has reached intermediate goal locations. Figure 5: CLIP-Nav \- We present a novel approach for zero-shot VLN, that utilizes CLIP to make sequential navigational decisions. At each timestep, a CLIP-Chosen Image is determined by grounding the current NC to each of the panoramic splits. The chosen image represents the direction our model has chosen for zero-shot navigation. In this case, it refers to the bedroom potentially being somewhere in the chosen direction. The AC grounding score gives us a stopping threshold for when to our agent believes it has reached the target. CLIP-Nav runs iteratively until this threshold is reached. #### 3.3.2 Seq CLIP-Nav In order to improve CLIP-Nav, we incorporate a simple backtracking mechanism. Figure 6 presents an overview of this method. We aggregate Keyphrase Grounding Scores across a sequence of $n$ nodes, and average them out to obtain a Sequence Grounding Score (SGS) as follows, $\text{SGS}\>=\frac{\sum_{n}\text{KGS}_{i}}{n},$ where $\text{KGS}_{i}$ is the KGS value of the CLIP-chosen image at at a particular timestep, and $n$ is the number of nodes. This score acts as a backtracking threshold to determine if the agent is heading in the right direction or needs to go back. Figure 6: Seq CLIP-Nav \- In order to improve the performance of CLIP-Nav, we incorporate a backtracking mechanism. CLIP scores across a sequence of timesteps are averaged to obtain a Sequence Grounding Score (SGS). This score is then used to determine if the agent needs to go back a few nodes (backtrack) or not. A low SGS value suggests that the agent might have gone down a path where it is not able to find the intermediate goal defined by the current NC. Alternatively, a high SGS shows that the agent is more confident of its ability in navigating the environment in a zero-shot manner. ## 4 Results Our results establish a baseline for zero-shot VLN, on the task of REVERIE. We use the following metrics to evaluate agent performance. 1. 1. Success Rate (SR) \- This is the fraction of episodes where the agent successfully reaches and stops at the target location. 2. 2. Success Weighted by Inverse Path Length (SPL) \- We use the definition provided by Batra et.al. in Batra et al. (2020). This is a standard metric used in several VLN tasks, and tells us about the optimality of the agent’s success route, when compared with the ground-truth oracle path. 3. 3. Oracle Succss Rate (OSR) \- This is the fraction of episodes where the agent passes through the target location in its path, but does not stop there. This accounts for paths that may have overshot the target, due to inaccurate grounding. 4. 4. Relative Change in Success (RCS) \- We also compute the percentage relative change in performance between Seen and Unseen data, which gives us insight into the generalizability of each approach. This is defined as - $RCS=\frac{\lvert Seen-Unseen\rvert}{\max\\{Seen,Unseen\\}}\times 100$ A lower score indicates that the agent is performing similarly across the splits, while a higher score indicates overfitting on the seen training environments. SR($\%$) $\uparrow$ OSR (%) $\uparrow$ SPL $\uparrow$ Approach Seen Unseen RCS $\downarrow$ Seen Unseen RCS $\downarrow$ Seen Unseen RCS $\downarrow$ ZS Random Walk $3.99$ $5.19$ $23.12$ $8.92$ $11.93$ $25.23$ $0.006$ $0.043$ $86.04$ CLIP-Nav $4.56$ $5.79$ $21.89$ $17.53$ $27.63$ $36.55$ $0.152$ $0.248$ 38.70 Seq CLIP-Nav $12.34$ $14.97$ 17.56 $19.47$ $24.46$ 20.40 $0.212$ 0.450 $52.88$ SV FAST-MATTN $50.53$ $14.40$ $71.50$ $55.17$ $28.20$ $50.69$ $0.455$ $0.072$ $84.17$ AirBERT $47.01$ $27.89$ $40.67$ $48.98$ $34.51$ $29.54$ $0.423$ $0.218$ $46.46$ DUET 71.75 46.98 $34.52$ 73.86 51.07 $30.85$ 0.639 $0.337$ $47.26$ Table 1: Zero-Shot (ZS) and Supervised (SV) results on the Seen and Unseen Val splits of REVERIE. The lowest RCS scores across all metrics indicate that our methods generalize significantly better in new environments over supervised approaches. On the unseen split, Seq. CLIP-Nav gives us the SOTA SPL result, and also improves upon the REVERIE baseline in terms of SR. Table 1 compares our Zero-Shot (ZS) results on the Seen and Unseen Val splits of REVERIE with SOTA fully supervised methods. We also compare with a Random Walk approach that chooses a random neighboring node for $8$ steps. This is the upper limit on the average path length of the dataset which is between $5$-$8$ steps. Observe the improvement in performance of Seq CLIP-Nav over the REVERIE Baseline approach (FAST-MATTN), in terms of SR and SPL on the Unseen split. The Unseen SPL even outperforms the SOTA supervised learning approaches - Airbert and DUET, showing that when our agent does take the right path in an unseen environment, it tends to do so in a more optimal manner. Also notice the significant improvement in the SR performance of Seq CLIP-Nav over CLIP- Nav, even while the OSRs are similar. This indicates the strong influence of backtracking, in preventing the agent from overshooting once it reaches a target location. On the Seen split, the supervised approaches outperform our methods in terms of SR and SPL, and this is quite evident since they have been trained on these environments. The higher RCS scores on these methods however indicate that they perform much worse on unseen environments in comparison to seen ones. This shows poor generalizability, and that they might have overfit on the training dataset. In contrast, the lower RCS scores across SR, OSR and SPL on our CLIP-based approaches indicates a far more consistent performance. We can thus quantitatively infer that our approaches generalize better in new environments over supervised approaches, satisfying one of our primary objectives for zero-shot navigation. This is a promising result, as it shows that our CLIP-based models without any finetuning are able to make consistent embodied navigational decisions irrespective of the type of environment that the agent is placed in. Additionally, we also obtain “Best OSR” results of $47.51\%$ on CLIP-Nav, and $48.41\%$ on Seq CLIP-Nav on the unseen split. These are the single best values across all the house scans. The higher values when compared to the overall OSR shows that our approach is able to perform particularly well in certain types of environments, and analyzing the influencing visuo-lingual factors to transfer knowledge is part of a future study. ## 5 Conclusion and Future Scope We tackle the problem of Vision-and-Language Navigation (VLN) in a fully zero- shot manner to address the issue of generalizability in unseen environments. Our approaches CLIP-Nav, and Seq CLIP-Nav give a far more consistent performance over fully supervised SOTA approaches on the REVERIE dataset when assessed by our generalizability metric — Relative Change in Success (RCS). Seq CLIP-Nav in particular gives improved results over the REVERIE baseline on the unseen validation set in terms of SR, and also improves upon SOTA supervised learning approaches in terms of SPL. This showcases the capability of CLIP without any dataset-specific finetuning in being able to make accurate sequential navigational decisions necessary for zero-shot VLN. In the future, we intend to study the cross-dataset performance of CLIP-Nav on other fine and coarse-grained instruction sets in indoor and outdoor settings as well as when there is a cooperative dialog for instruction following Padmakumar et al. (2022); Gao et al. (2022). Our current backtracking score while simplistic, still drastically influences performance; improving this via meta-learning is a future study. Yet another exciting direction is a Virtual Reality (VR) experiment to learn human patterns in zero-shot instruction following. Analyzing people’s zero-shot paths in VR environments could give us insight into their intuition for sequential decision making, and can be modeled to improve CLIP-Nav. ## References * Anderson et al. (2018) Peter Anderson, Qi Wu, Damien Teney, Jake Bruce, Mark Johnson, Niko Sünderhauf, Ian Reid, Stephen Gould, and Anton Van Den Hengel. Vision-and-language navigation: Interpreting visually-grounded navigation instructions in real environments. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , pp. 3674–3683, 2018. * Batra et al. (2020) Dhruv Batra, Aaron Gokaslan, Aniruddha Kembhavi, Oleksandr Maksymets, Roozbeh Mottaghi, Manolis Savva, Alexander Toshev, and Erik Wijmans. Objectnav revisited: On evaluation of embodied agents navigating to objects. _arXiv preprint arXiv:2006.13171_ , 2020. * Brown et al. (2020) Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. _Advances in neural information processing systems_ , 33:1877–1901, 2020. * Chang et al. (2017) Angel Chang, Angela Dai, Thomas Funkhouser, Maciej Halber, Matthias Niessner, Manolis Savva, Shuran Song, Andy Zeng, and Yinda Zhang. Matterport3d: Learning from rgb-d data in indoor environments. _arXiv preprint arXiv:1709.06158_ , 2017. * Chen et al. (2021) Kevin Chen, Junshen K Chen, Jo Chuang, Marynel Vázquez, and Silvio Savarese. Topological planning with transformers for vision-and-language navigation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 11276–11286, 2021. * Chen et al. (2022a) Shizhe Chen, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, and Ivan Laptev. Think global, act local: Dual-scale graph transformer for vision-and-language navigation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)_ , pp. 16537–16547, June 2022a. * Chen et al. (2022b) Shizhe Chen, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, and Ivan Laptev. Learning from unlabeled 3d environments for vision-and-language navigation. _arXiv preprint arXiv:2208.11781_ , 2022b. * Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. _arXiv preprint arXiv:1810.04805_ , 2018. * Gadre et al. (2022) Samir Yitzhak Gadre, Mitchell Wortsman, Gabriel Ilharco, Ludwig Schmidt, and Shuran Song. Clip on wheels: Zero-shot object navigation as object localization and exploration. _arXiv preprint arXiv:2203.10421_ , 2022. * Gao et al. (2022) Xiaofeng Gao, Qiaozi Gao, Ran Gong, Kaixiang Lin, Govind Thattai, and Gaurav S Sukhatme. Dialfred: Dialogue-enabled agents for embodied instruction following. _arXiv preprint arXiv:2202.13330_ , 2022. * Gu et al. (2022) Jing Gu, Eliana Stefani, Qi Wu, Jesse Thomason, and Xin Eric Wang. Vision-and-language navigation: A survey of tasks, methods, and future directions. _arXiv preprint arXiv:2203.12667_ , 2022. * Guhur et al. (2021) Pierre-Louis Guhur, Makarand Tapaswi, Shizhe Chen, Ivan Laptev, and Cordelia Schmid. Airbert: In-domain pretraining for vision-and-language navigation. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , pp. 1634–1643, 2021. * Hong et al. (2021) Yicong Hong, Qi Wu, Yuankai Qi, Cristian Rodriguez-Opazo, and Stephen Gould. Vln bert: A recurrent vision-and-language bert for navigation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 1643–1653, 2021. * Huang et al. (2019) Haoshuo Huang, Vihan Jain, Harsh Mehta, Alexander Ku, Gabriel Magalhaes, Jason Baldridge, and Eugene Ie. Transferable representation learning in vision-and-language navigation. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , pp. 7404–7413, 2019. * Ku et al. (2020) Alexander Ku, Peter Anderson, Roma Patel, Eugene Ie, and Jason Baldridge. Room-across-room: Multilingual vision-and-language navigation with dense spatiotemporal grounding. _arXiv preprint arXiv:2010.07954_ , 2020. * Li et al. (2022) Jialu Li, Hao Tan, and Mohit Bansal. Envedit: Environment editing for vision-and-language navigation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 15407–15417, 2022. * Padmakumar et al. (2022) Aishwarya Padmakumar, Jesse Thomason, Ayush Shrivastava, Patrick Lange, Anjali Narayan-Chen, Spandana Gella, Robinson Piramuthu, and Dilek Hakkani-Tur Gokhan Tur and. TEACh: Task-driven Embodied Agents that Chat. In _Conference on Artificial Intelligence (AAAI)_ , 2022. URL https://arxiv.org/abs/2110.00534. * Qi et al. (2020) Yuankai Qi, Qi Wu, Peter Anderson, Xin Wang, William Yang Wang, Chunhua Shen, and Anton van den Hengel. Reverie: Remote embodied visual referring expression in real indoor environments. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 9982–9991, 2020. * Radford et al. (2021) Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In _International Conference on Machine Learning_ , pp. 8748–8763. PMLR, 2021. * Rombach et al. (2022) Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 10684–10695, 2022. * Shen et al. (2021) Sheng Shen, Liunian Harold Li, Hao Tan, Mohit Bansal, Anna Rohrbach, Kai-Wei Chang, Zhewei Yao, and Kurt Keutzer. How much can clip benefit vision-and-language tasks? _arXiv preprint arXiv:2107.06383_ , 2021. * Vasudevan et al. (2021) Arun Balajee Vasudevan, Dengxin Dai, and Luc Van Gool. Talk2nav: Long-range vision-and-language navigation with dual attention and spatial memory. _International Journal of Computer Vision_ , 129(1):246–266, 2021. * Wang et al. (2021) Hanqing Wang, Wenguan Wang, Wei Liang, Caiming Xiong, and Jianbing Shen. Structured scene memory for vision-language navigation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 8455–8464, 2021. * Wang et al. (2018) Xin Wang, Wenhan Xiong, Hongmin Wang, and William Yang Wang. Look before you leap: Bridging model-free and model-based reinforcement learning for planned-ahead vision-and-language navigation. In _Proceedings of the European Conference on Computer Vision (ECCV)_ , pp. 37–53, 2018. * Wang et al. (2019a) Xin Wang, Qiuyuan Huang, Asli Celikyilmaz, Jianfeng Gao, Dinghan Shen, Yuan-Fang Wang, William Yang Wang, and Lei Zhang. Reinforced cross-modal matching and self-supervised imitation learning for vision-language navigation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 6629–6638, 2019a. * Wang et al. (2019b) Xin Wang, Qiuyuan Huang, Asli Celikyilmaz, Jianfeng Gao, Dinghan Shen, Yuan-Fang Wang, William Yang Wang, and Lei Zhang. Reinforced cross-modal matching and self-supervised imitation learning for vision-language navigation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 6629–6638, 2019b. * Zhu et al. (2021) Fengda Zhu, Xiwen Liang, Yi Zhu, Qizhi Yu, Xiaojun Chang, and Xiaodan Liang. Soon: Scenario oriented object navigation with graph-based exploration. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 12689–12699, 2021.
# The H$\alpha$ and [O III] $\lambda 5007$ Luminosity Functions of $1.2<z<1.9$ Emission-Line Galaxies from HST Grism Spectroscopy Gautam Nagaraj Department of Astronomy & Astrophysics, The Pennsylvania State University, University Park, PA 16802, USA Institute for Gravitation and the Cosmos, The Pennsylvania State University, University Park, PA 16802, USA Robin Ciardullo Department of Astronomy & Astrophysics, The Pennsylvania State University, University Park, PA 16802, USA Institute for Gravitation and the Cosmos, The Pennsylvania State University, University Park, PA 16802, USA William P. Bowman Astronomy Department, Yale University, New Haven, CT 06511, USA Department of Astronomy & Astrophysics, The Pennsylvania State University, University Park, PA 16802, USA Institute for Gravitation and the Cosmos, The Pennsylvania State University, University Park, PA 16802, USA Alex Lawson Department of Astronomy & Astrophysics, The Pennsylvania State University, University Park, PA 16802, USA Caryl Gronwall Department of Astronomy & Astrophysics, The Pennsylvania State University, University Park, PA 16802, USA Institute for Gravitation and the Cosmos, The Pennsylvania State University, University Park, PA 16802, USA ###### Abstract Euclid and the Roman Space Telescope (Roman) will soon use grism spectroscopy to detect millions of galaxies via their H$\alpha$ and [O III] $\lambda 5007$ emission. To better constrain the expected galaxy counts from these instruments, we use a vetted sample of 4,239 emission-line galaxies from the 3D-HST survey to measure the H$\alpha$ and [O III] $\lambda 5007$ luminosity functions between $1.16<z<1.90$; this sample is $\sim 4$ times larger than previous studies at this redshift. We find very good agreement with previous measurements for H$\alpha$, but for [O III], we predict a higher number of intermediate-luminosity galaxies than previous works. We find that for both lines, the characteristic luminosity, $\mathcal{L}_{*}$, increases monotonically with redshift, and use the H$\alpha$ luminosity function to calculate the epoch’s cosmic star formation rate density. We find that H$\alpha$-visible galaxies account for $\sim 81\%$ of the epoch’s total star formation rate, and this value changes very little over the $1.16<z<1.56$ redshift range. Finally, we derive the surface density of galaxies as a function of limiting flux and find that previous predictions for galaxy counts for the Euclid Wide Survey are unchanged, but there may be more [O III] galaxies in the Roman High Latitude Survey than previously estimated. Galaxy evolution (594), Luminosity function (942), Spectral energy distribution (2129), High-redshift galaxies (734) ††facilities: HST (WFC3), Spitzer (MIPS), Herschel (PACS, SPIRE), GALEX, Swift(UVOT)††software: AstroPy (Astropy Collaboration et al., 2013, 2018), SciPy (Jones et al., 2001–; Virtanen et al., 2020), CLOUDY (Ferland et al., 1998, 2013), FSPS (Conroy et al., 2009; Conroy & Gunn, 2010), MCSED(Bowman et al., 2020) ## 1 Introduction Since its observational discovery by Riess et al. (1998) and Perlmutter et al. (1999), dark energy has been at the forefront of astronomical research. While the $\Lambda$CDM paradigm has been extremely successful in predicting the properties of the cosmological microwave background and reproducing observables such as large scale structure and cosmic abundances, there are many open questions remaining, especially about the nature of dark matter and the evolution of dark energy (e.g., Bull et al., 2016; Amendola et al., 2018, and references therein). In order to better constrain cosmological models, we must continue accruing more accurate and precise observational probes. Such efforts include the use of large-scale spectroscopic surveys to measure baryonic acoustic oscillations and redshift space distortions throughout cosmic time. The most efficient mechanisms for generating these data are multi-fiber spectroscopic surveys, such as WiggleZ (Drinkwater et al., 2010; Blake et al., 2011), the Baryon Oscillation Spectroscopic Survey (BOSS; Dawson et al., 2013), the extended Baryon Oscillation Spectroscopic Survey (eBOSS; Dawson et al., 2016) and the Dark Energy Spectroscopic Instrument (DESI) survey (DESI Collaboration et al., 2016; Abareshi et al., 2022); integrated field unit (IFU) spectroscopic surveys, such as the Hobby-Eberly Telescope Dark Energy Experiment (HETDEX; Gebhardt et al., 2021; Hill et al., 2021); wide-field imaging using large, comprehensive sets of narrow-band filters (e.g., J-PAS; Cepa et al., 2016; Salzano et al., 2021), and slitless (grism) spectroscopic surveys, including Euclid (Laureijs et al., 2011, 2012) and the Nancy Grace Roman Space Telescope (Roman; Green et al., 2012; Dressler et al., 2012; Spergel et al., 2015). Slitless spectroscopy in particular is an extremely efficient method of obtaining spectra over an entire field with no need for the pre-selection of targets. Narrow-band surveys, such as the High-$z$ Emission Line Survey (HiZELS; Geach et al., 2008; Sobral et al., 2009) and the NOAO Extremely Wide Field Infrared Imager (NEWFIRM) H$\alpha$ Survey (Ly et al., 2011), achieve the same efficiency but are usually restricted to a very limited slice of redshift space, and thus require many different filters to survey large volumes. Normal galaxies have no strong emission lines between Ly$\alpha$ at 1216 Å and [O II] $\lambda 3727$, so beyond $z\sim 1$, redshift surveys are most efficiently performed at near-infrared wavelengths, with lines such [O II] $\lambda 3727$, H$\beta$, [O III] $\lambda 5007$, and H$\alpha$. While it is possible to photometrically-select $z\gtrsim 1$ objects and then refine their redshifts with follow-up spectroscopy (e.g., Davis et al., 2003; Steidel et al., 2004; Lilly et al., 2007; DESI Collaboration et al., 2016), emission line galaxy (ELG) surveys to probe large swaths of cosmic time are most easily performed from space. GRAPES (Pirzkal et al., 2004), which used the ACS G800L grism of the Hubble Space Telescope, as well as WISP (Atek et al., 2010) and 3D-HST (Brammer et al., 2012; Momcheva et al., 2016), which used the HST/WFC3 G102 and G141 grisms, represent some of the first efforts to create space- based ELG samples. In this study, we use the 3D-HST sample described in Nagaraj et al. (2021b) and Nagaraj et al. (2021a), hereafter referred to as Paper I and Paper II, to further explore the emission-line properties of $1.2\lesssim z\lesssim 1.9$ galaxies. Two near-future missions, Euclid (Laureijs et al., 2011, 2012) and Roman (Green et al., 2012; Dressler et al., 2012; Spergel et al., 2015), will identify millions of galaxies at redshifts $0.7\lesssim z\lesssim 2.7$ using their H$\alpha$ and [O III] $\lambda 5007$ emission. Given the similarity in the spectroscopic survey designs of 3D-HST, the Euclid Deep Survey, and the Roman High Latitude Survey (see Figure 5 in Paper I for a visual depiction of the survey limits and observed 3D-HST fluxes), we can use 3D-HST as a pathfinder for these missions. In particular, by evaluating the H$\alpha$ and [O III] $\lambda 5007$ luminosity functions and biases with respect to other galaxy samples and dark matter distributions, we can estimate how many galaxies these programs will find and how well they will be able to measure cosmological parameters. Several efforts have been made to calculate the H$\alpha$ and [O III] $\lambda 5007$ luminosity functions in both the local and distant universe, and these have led to deeper understanding of galaxy evolution (e.g., Gallego et al., 1995; Tresse & Maddox, 1998; Sullivan et al., 2000; Jones & Bland-Hawthorn, 2001; Fujita et al., 2003; Hippelein et al., 2003; Glazebrook et al., 2004; Treyer et al., 2005; Wyder et al., 2005; Ly et al., 2007; Geach et al., 2008; Shim et al., 2009; Sobral et al., 2009, 2011; Ly et al., 2011; Tadaki et al., 2011; Sobral et al., 2012; Colbert et al., 2013; Pirzkal et al., 2013; Sobral et al., 2013; Khostovan et al., 2015; Mehta et al., 2015; Sobral et al., 2015; Comparat et al., 2016; Hayashi et al., 2020; Khostovan et al., 2020). At $z\gtrsim 1$, most of these studies are narrow-band surveys targeted at specific redshifts; while these programs involve large numbers of sources and extremely deep exposures (e.g., Khostovan et al., 2020), their ability to investigate cosmic evolution is limited due to the small volumes covered. Other investigations have been limited by sample size (e.g., Shim et al., 2009) or spectral resolution (e.g., Pirzkal et al., 2013, where [O III] and H$\beta$ are a blended feature). Particularly notable is the work of Colbert et al. (2013) and the follow-up study by Mehta et al. (2015), which used data from the HST/WFC3 Infrared Spectroscopic Parallel (WISP) survey (Atek et al., 2010) to create a sample of approximately 1,000 ELGs between $0.3<z<2.3$. Recently, Nagaraj et al. (2021b) carefully vetted a sample of 4350 $1.2<z<1.9$ emission-line galaxies which were originally identified on the 3D-HST grism frames by Momcheva et al. (2016). Here we use a subsample of 3,187 sources to determine the luminosity function and equivalent width distribution of [O III] $\lambda 5007$ and H$\alpha$ in this redshift range. Since the depth and resolution of the 3D-HST data are similar to the grism surveys planned for Euclid and Roman, we can use our measurements to refine the predictions for these studies, and improve our measurement of the amount of intermediate- redshift star formation that is occurring in emission-line galaxies. These data will also allow us to examine the relationship between [O III] $\lambda 5007$ emission and star-formation rate at an epoch intermediate between the local universe, where emission from oxygen is mostly from [O II] $\lambda 3727$, and the redshifts studied by Bowman et al. (2021), where [O III] $\lambda 5007$ dominates. This paper is the third in series that analyses 3D-HST sources at redshifts $1.2<z<1.9$. Paper I showed empirical relations between stellar mass and various observational and physical properties, such as absolute magnitude in a rest-frame optical filter. Paper II focused on the relationships among stars, gas, and dust in galaxies with available mid- and far-IR data. Throughout this paper, we assume a $\Lambda$CDM cosmology with $\Omega_{\Lambda}=0.69$, $\Omega_{M}=0.31$ and $H_{0}=68$ km s-1 Mpc-1 (Bennett et al., 2013). All magnitudes given in the paper are in the AB magnitude system (Oke, 1974). ## 2 Data and Selection Effects In this section we describe the fluxes used for the H$\alpha$ and [O III] $\lambda 5007$ luminosity function calculations, including the critical issue of incompleteness for low-luminosity objects. ### 2.1 Data As the data used to compute the H$\alpha$ and [O III] $\lambda 5007$ luminosity functions have been described in Papers I and II, we give only a brief overview here. From an initial list of 9341 $1.2\lesssim z\lesssim 1.9$ candidates identified on WFC3/G141 grism frames by the 3D-HST survey (GO-11600, 12177, 12328; Brammer et al., 2012; Momcheva et al., 2016), we carefully identified a clean sample of 4350 ELGs brighter than the catalog’s F125W (J) + F140W (JH) + F160W (H) magnitude limit of $m_{J+JH+H}$ $=26$. Since the 625 arcmin2 covered by the 3D-HST survey coincides with regions of the Cosmic Assembly Near-infrared Deep Extragalactic Legacy Survey (CANDELS; Grogin et al., 2011; Koekemoer et al., 2011), the fields have a wealth of photometric observations, and these data helped fuel the results of Paper I and Paper II. Paper I describes the process we used to identify active galactic nuclei (AGN) masquerading as normal galaxies in our sample. First, we used X-ray matching from the deep surveys of the CANDELS fields (especially GOODS-S), to identify 72 AGN in our ELG sample. Specifically, we used a cross-correlation search radius of 1″ to identify possible X-ray counterparts to our ELGs. Any galaxy with an X-ray luminosity greater than $10^{42}$ $\text{erg s}^{-1}$ in the $2-10$ keV band was considered an AGN and eliminated from the analysis. Stacking of the remaining ELGs then found X-rays levels consistent with those expected from simple star formation, suggesting that the vast majority of objects remaining in our sample are normal galaxies with no obvious AGN activity. While X-rays are very effective for AGN identification, they fail when the gas column densities are too high ($N_{H}\gtrsim 5-50\times 10^{23}~{}\textrm{cm}^{-2}$), which tends to correspond to highly dusty, Compton-thick systems (e.g., Brandt & Alexander, 2015, and references therein). To find such objects, we used the IRAC AGN selection criteria described by Donley et al. (2012); the method identified 50 AGN candidates (with 11 being previously excluded via their X-ray luminosity). This whittled down our ELG sample to 4239 objects. Bowman et al. (2019) ran nearly the same procedure on 3D-HST galaxies at $1.9\leq z\leq 2.35$. After removing AGN, they were left with a sample of 1964 ELGs with trustworthy redshifts and clean spectra. These data are used in §4.3, where we combine the datasets to examine the evolution of the [O III] $\lambda 5007$ luminosity function over the full redshift range from $1.16\leq z\leq 2.35$. ### 2.2 Completeness and Selection Effects In this work, we use the H$\alpha$ and [O III] fluxes derived from the 3D-HST grism spectra by Momcheva et al. (2016). Due to the relatively low resolution ($R\sim 130$) of the G141 grism and the morphological broadening that occurs in extended sources, the two components of [O III], [O III] $\lambda 5007$ and [O III] $\lambda 4959$, appear as a single blended feature. However, since the ratio of [O III] $\lambda 5007$ to [O III] $\lambda 4959$ is fixed at 2.98:1 (Storey & Zeippen, 2000), we can simply rescale the Momcheva et al. (2016) measurements to give the de-blended fluxes for primary [O III] $\lambda 5007$ emission line. A larger concern is our inability to separate H$\alpha$ from the bracketing forbidden lines of [N II] $\lambda\lambda 6548,6584$. While corrections for [N II] exist in the literature (e.g., the procedure and references outlined in Price et al., 2014), gas-phase metallicities defined through strong line indicators are subject to degeneracies, inaccuracies, and other issues (e.g., Kewley & Ellison, 2008). Furthermore, the ionization balance between the first and second excited states of oxygen changes dramatically, from predominantly O+ at low redshift to O++ at $z\sim 2$ (e.g., Kewley et al., 2019, and references therein). Since N+ has an ionization potential only slightly less than O+, we can expect a similar trend in our data. This shift in ionization balance means that the [N II]/H$\alpha$ ratio must also evolve with redshift. Another concern associated with our understanding of the data involves the issue of dust attenuation. The wavelength dependence of attenuation has a variety of shapes in different galaxies (see Shivaei et al., 2020; Salim & Narayanan, 2020, and references therein). An accurate modeling of this behavior is difficult, especially given its degeneracy with the other parameters of the complex stellar populations that make up a galaxy’s SED. Even at the wavelength of H$\alpha$ (6563 Å), dust attenuation represents an uncertainty. Given the issues associated with [N II] and dust, we do not attempt to disentangle the flux of H$\alpha$ from that of [N II]; instead, we present a luminosity function for the combined lines. We discuss this further in §4.2 where we calculate the star formation rate density between $1.2\lesssim z\lesssim 1.9$. An extremely important issue for any luminosity function analysis is completeness. Following Bowman et al. (2021), we parameterize the completeness of the [O III] $\lambda 5007$ and H$\alpha$ detections as a function of flux, $f$, using a modified Fleming et al. (1995) function, $\displaystyle F_{F}(f)$ $\displaystyle=\frac{1}{2}\Bigg{[}1+\frac{\alpha_{F}\log(f/f_{{\rm{50}}})}{\sqrt{1+(\alpha_{F}\log(f/f_{{\rm{50}}}))^{2}}}\Bigg{]}$ (1) $\displaystyle\tau(f)$ $\displaystyle=1-e^{-f/f_{\rm 10}}$ (2) $\displaystyle F_{c}(f)$ $\displaystyle=[F_{F}(f)]^{1/\tau(f)}$ (3) In the equations, $F_{c}(f)$ is the completeness, $\alpha_{F}$ describes how quickly the completeness drops off as a function of flux, $f_{\rm 50}$ is the flux where the recovery fraction of objects is 50%, and $f_{\rm 10}$ is the flux at which the sample is 10% complete. As the Fleming function is completely described by $f_{\rm 50}$ and $\alpha_{F}$, $f_{\rm 10}$ is not an independent parameter but a quantity directly derived from Equation 1. An important point to note is that for G141 grism data, the behavior of the completeness curve is virtually independent of wavelength (e.g., Zeimann et al., 2014; Bowman et al., 2021). There are two ways to estimate the parameters $\alpha_{F}$ and $f_{\rm 50}$. The first method, which was used by Bowman et al. (2021) in their analysis of the [O III] $\lambda 5007$ luminosity function of $1.90<z<2.35$ grism-selected galaxies, is to use the data themselves and simultaneously fit the galaxy luminosity function (§3) and each field’s completeness parameters to the observed distribution of emission-line luminosities. Such a procedure is complex, since, even if $\alpha_{F}$ is assumed to be the same across all five CANDELS fields, the calculation still involves at least 9 separate variables (a minimum of three for the luminosity function, one value of $f_{\rm 50}$ for each field, and $\alpha_{F}$). Alternatively, it is possible to fit the completeness curves separately from the luminosity function by assuming the intrinsic flux distribution of faint emission lines is a power law. This approach is reasonable, given the expected nature of the faint galaxy number counts, and was the method employed by Bowman et al. (2019) in their study of the physical properties of $z\sim 2$ grism-selected galaxies. Our experiments show that both methods yield similar results for the galaxy luminosity function. Therefore, to avoid any degeneracies associated with high-dimensional fits and to simplify the analysis, we chose to decouple the question of completeness from the luminosity function calculation and adopt the values of $f_{\rm 50}$ and $\alpha_{F}$ found by Bowman et al. (2021). These parameters, which were derived for $1.90\leq z\leq 2.35$ [O III] $\lambda 5007$ emission on the same 3D-HST frames used here, were found using the simultaneous fitting technique described above. Since completeness should only be a function of line flux, and not depend on the specific line being observed, the Bowman et al. (2021) values should be equally applicable to our survey. There is one potential caveat to this last assertion. As detailed by Momcheva et al. (2016), the $1\sigma$ sensitivity limit of the 3D-HST survey depends not only on the strength of an emission line, but also the angular size of the emitting source. Since the present study focuses on galaxies at lower redshift than those measured by Bowman et al. (2021), this difference has the potential to cause a shift in the survey’s 50% completeness limit. However, based on values from the 3D-HST catalog (Momcheva et al., 2016), the mean size of $1.16\leq z\leq 1.90$ emission-line galaxies is only 1.15 times that of the Bowman et al. (2021) systems. In comparison, the size spread amongst the galaxies in the sample is a factor of $\sim 4$. For this reason, we do not account for this size difference in our analysis. We list the completeness parameters in Table 1 and show the completeness curves in Figure 1. For the purposes of the luminosity functions presented in this paper, we only consider flux measurements above the 50% levels given in Table 1. The inclusion of objects much below this limit induces strong effects on the luminosity function by amplifying the uncertainties in the completeness curves. Conversely, if the flux limit is too strict, the sample of galaxies becomes too small for any reliable measurement of the function’s low- luminosity end. Our 50% cutoff results in 2947 [O III] and 1892 H$\alpha$ measurements being used for our luminosity function calculations. For the analysis of evolution in the [O III] $\lambda 5007$ luminosity function, this number can be incremented using the $1.9\leq z\leq 2.35$ ELGs found by Bowman et al. (2021), which were identified in the same manner as the galaxies used here. This results in a sample of 4519 [O III] $\lambda 5007$ ELGs above the 50% completeness limit. Table 1: Completeness Parameters Field | $f_{\rm 50}~{}(10^{-17}$ $\text{erg cm}^{-2}\text{ s}^{-1}$) ---|--- AEGIS | 2.35 COSMOS | 3.12 GOODS-N | 2.20 GOODS-S | 2.86 UDS | 2.85 Note. — For all fields, $\alpha_{F}=4.56$. Figure 1: Completeness curve for H$\alpha$ and [O III] $\lambda 5007$ in each of the five 3D-HST fields. The parameter values used to generate these curves are listed in Table 1. ## 3 Methodology ### 3.1 Deriving the Luminosity Function The maximum likelihood estimator (MLE) is commonly used in astronomy to measure and fit the parameterized variables of a luminosity function. Simplifications to the MLE integral have also led to computationally efficient procedures for creating discrete luminosity functions derived from techniques such as the $1/V_{\rm max}$ (Schmidt, 1968, 1970; Huchra & Sargent, 1973; Avni & Bahcall, 1980) and the $C^{-}$ (Lynden-Bell, 1971) methods. Here we do both, and derive both a digital representation of the number of galaxies versus emission-line (log) flux, and a convenient analytical representation of the luminosity function. We assume Poisson statistics hold, and define the likelihood of a luminosity function (scripted as $\mathcal{P}$ to avoid confusion with luminosity) following the derivation by Ciardullo et al. (2013). For computational convenience, we use $\mathcal{L}\equiv\log L$ rather than the linear luminosity; in both cases, the math is identical, although the units are different. With the equations in this section, we are able to both measure the galaxies’ discrete luminosity function and fit the data to a parameterization of our choosing (i.e., the Schechter, 1976, function). From Ciardullo et al. (2013), the likelihood of observing any luminosity function, $\phi^{\prime}$ is given by $\ln\mathcal{P}=\sum_{i}^{N}\ln\phi^{\prime}(\mathcal{L}_{i},z_{i})-\int_{z_{1}}^{z_{2}}\int_{\mathcal{L}_{\rm min}(z)}^{\infty}\phi^{\prime}(\mathcal{L},z)\frac{dV}{dz}d\mathcal{L}\,dz$ (4) where $N$ is the number of galaxy luminosities included in the sample. Note that there is a distinction between the true luminosity function $\phi(\mathcal{L},z)$, which we model as a Schechter (1976) function with parameters $\alpha$, $\mathcal{L_{*}}$, and $\log\phi_{*}$, i.e., $\phi(\mathcal{L},z)=\ln(10)10^{\log\phi_{*}(z)}10^{\left[\mathcal{L}-\mathcal{L}_{*}(z)\right]\left[\alpha+1\right]}\exp\left(-10^{\mathcal{L}-\mathcal{L}_{*}(z)}\right)$ (5) and the observed luminosity function, $\phi^{\prime}$, which is affected by incompleteness and measurement error. Following Marshall et al. (1983) and Marshall (1985), we define a function $\Omega(\mathcal{L},z)$ that takes into account both the flux completeness and the limits of the survey area. If $f$ is the flux, which can be thought of as a function of luminosity and redshift, and $p(f,\hat{n})$ is the completeness function in a given direction $\hat{n}$, then we have Equation 6 below, where $d^{2}\hat{n}$ is the integral over the unit sphere. In our case, we consider $p(f,\hat{n})$ constant over any given field used in the analysis. If $\Omega_{i}$ is the effective survey area of field $i$, then $\Omega(\mathcal{L},z)$ can be approximated as $\Omega(\mathcal{L},z)=\int\frac{d^{2}\hat{n}}{4\pi}\,p(f,\hat{n})\approx\sum_{i=0}^{N_{\rm fields}}\frac{\Omega_{i}}{4\pi}p_{i}(f)$ (6) Assuming that the true luminosity is isotropic, we can simply connect the observed and true luminosity functions through $\phi^{\prime}(\mathcal{L},z)=\Omega(\mathcal{L},z)\phi(\mathcal{L},z)$ (7) Going back to Equation 4, $z_{1}$ and $z_{2}$ represent the survey redshift limits. The minimum luminosity, $L_{\rm min}(z)$, is somewhat arbitrary and is simply taken to be a luminosity lower than what is observable. In practice, the upper limit of the integral is also fixed at a value at, or above, that which no galaxies are expected to exist. Finally, $dV/dz$ is the differential volume element. Assuming a spatially flat universe, this is simply $\frac{dV}{dz}=\frac{4\pi d_{A}^{2}(z)}{H(z)}$ (8) where $d_{A}(z)$ represents the comoving angular diameter distance and $H(z)$ is the Hubble parameter. Given this definition, the true luminosity function represents the number of galaxies per dex (log luminosity units) per comoving volume element. Moreover, given the nature of the 3D-HST survey, we assume that the flux completeness function is constant within each CANDELS field, so in a given field, $p(f,\hat{n})\equiv p(f)$. As described in §2, we use the modified Fleming et al. (1995) completeness curve $F_{c}(f)$ for $p(f)$. The likelihood analysis presented here has not included measurement errors in the luminosity (i.e., heteroscedasticity). We follow the convention adopted in Mehta et al. (2015), in which the luminosity errors are assumed to be normal with mean $\mathcal{L}_{i}$ and standard deviation $\sigma_{i}$. In that case, we can replace Equation 4 with $\ln\mathcal{P}=\sum_{i}^{N}\ln\int_{\mathcal{L}_{\rm low}}^{\mathcal{L}_{\rm high}}\phi^{\prime}(\mathcal{L}_{i},z_{i})N(\mathcal{L}|{\mathcal{L}_{i},\sigma_{i}})\frac{dV}{dz}d\mathcal{L}-\int_{z_{1}}^{z_{2}}\int_{\mathcal{L}_{\rm min}(z)}^{\infty}\phi^{\prime}(\mathcal{L},z)\frac{dV}{dz}d\mathcal{L}\,dz$ (9) The drawback of this approach is that it is computationally expensive, as the normal distribution has a possibly different mean and standard deviation for every flux measurement. The results given in §4 are therefore based on Equation 4. We find that including luminosity errors has little to no effect at the bright end of the [O III] luminosity function fit but does slightly lessen the low-luminosity slope $\alpha$. In other words, the observed faint- end slope is affected by Eddington (1913) bias. Bowman et al. (2021) provides a more detailed analysis of the effects of photometric uncertainties on the shape of the luminosity function. If we assume that the luminosity function remains unchanged over the redshift interval of interest, the MLE for discrete points becomes much simpler to compute. Let us parameterize the luminosity function as a sum of discrete Dirac-delta functions at $M$ different luminosities, i.e., $\phi(L)=\sum_{j=1}^{M}\phi_{j}\delta^{D}(L-L_{j})$ (10) Then, using Equations 4 and 10, the likelihood can be expressed as $\ln\mathcal{P}=\sum_{i=1}^{N}\ln\phi_{i}-\sum_{j=1}^{M}\int dz\frac{dV}{dz}\Omega(L_{j},z)\phi_{j}$ (11) Setting the derivative of the likelihood to zero, the MLE solution for the luminosity function $\phi_{i}$ at $L_{i}$ becomes $\phi_{i}^{-1}=\int dz\frac{dV}{dz}\Omega(L_{i},z)\equiv V_{\rm eff}(L_{i})$ (12) This is a slight generalization of the historic $1/V_{\rm max}$ method (Schmidt, 1968, 1970; Huchra & Sargent, 1973; Avni & Bahcall, 1980) for any given completeness curve $p(f,\hat{n})$. We use this formalism, which we will call the $V_{\rm eff}$ method, as a benchmark for our more sophisticated MCMC approaches. ### 3.2 Computing the Luminosity Function Our simplest method of computing the luminosity function is the $V_{\rm eff}$ approach introduced in §3.1. We calculate $\phi_{i}$ (Equation 12) at every source luminosity in our sample and collect the results into luminosity (or log luminosity) bins. Wide bins allow for larger numbers of sources per bin and are thus more reliable, but do not convey as much information given their coarseness. We find that dividing the measurements into $\sim 50$ bins of equal size in log luminosity space works quite well for balancing the number of sources per bin against the complexity of the results. The errors on our fitted parameters are generated via a bootstrap analysis. We take the true $V_{\rm eff}$-method result using the original set of $N$($L_{i},\phi_{i}$) values, and then generate $B$ bootstrap samples, in which the $N$ values of $L_{i}$ and $\phi_{i}$ are generated randomly with replacement. The sample variance is taken to be the error on the luminosity function measurement, $\hat{\sigma}^{2}=\frac{1}{B-1}\sum_{i=1}^{B}\left[\theta_{i}-\left(\frac{1}{B}\sum_{i=1}^{B}\theta_{i}\right)\right]^{2}$ (13) where $\theta$ is the binned luminosity function measured at a particular interval, and $B$ is typically set to $B=100$. We include all fluxes down to the 50% completeness limit (§2) in the $V_{\rm eff}$ method. For the computation of the Schechter (1976) function parameters, our MCMC code uses uniform priors on $\alpha$, $\log\phi_{*}$, and $\mathcal{L}_{*}$ with bounds $[-3,1]$, $[-8,5]$, and $[40,45]$, respectively, while the completeness parameters are fixed at the values given in Table 1. We define the relative likelihood of a solution either through Equation 4 (no observational errors) or Equation 9 (with observational errors) and employ the emcee package (Foreman-Mackey et al., 2013) to explore the parameter space. We compute both a static (non-evolving) luminosity function and one that evolves over time. To explore time evolution in the luminosity function, we use a method based on the technique described in Leja et al. (2020). We let both $\log\phi_{*}$ and $\mathcal{L}_{*}$ be quadratic functions of redshift. However, rather than fitting the coefficients of the quadratic, whose priors are difficult to physically motivate (Leja et al., 2020), we use the values of $\log\phi_{*}$ and $\mathcal{L}_{*}$ at three specific redshifts ($z_{1}=1.20$, $z_{2}=1.76$, and $z_{3}=2.32$ for [O III] $\lambda 5007$ and $z_{1}=1.18$, $z_{2}=1.36$, and $z_{3}=1.54$ for H$\alpha$) to define the quadratic formulation of $\mathcal{L}_{*}(z)$ and $\log\phi_{*}(z)$. As in our static luminosity- function calculation, the prior on $\mathcal{L}_{i*}$ is uniform over the range $[40,45]$ and the prior on $\log\phi_{i*}$ uniform on $[-8,5]$. For the analysis, we fix the completeness parameters and also constrain $\alpha$ to be constant over time. Doing so avoids exacerbating degeneracies between $\alpha$ and the other two Schechter parameters. (In fact, in the case of [O III], we find that leaving $\alpha$ as a free parameter over the large redshift range $1.16\leq z\leq 2.35$ leads to failures in the MCMC fits. For the [O III] $\lambda 5007$ line, we therefore we fix $\alpha=-1.5$.) For the reader’s convenience, in Table 2 we have listed all the parameters used in the non-evolving and redshift-varying luminosity function calculations, as well as the priors applied in the MCMC code. Table 2: H$\alpha$ and [O III] $\lambda 5007$ Luminosity Function Fitting Parameters Parameter(s) | Priors ---|--- $\alpha$ | Uniform on $[-3,1]$; Fixedaa$\alpha$ is fixed at $-1.5$ only for the redshift-varying [O III] $\lambda 5007$ luminosity function. $\mathcal{L}_{*}$, $\mathcal{L}_{1*}$, $\mathcal{L}_{2*}$, $\mathcal{L}_{3*}$ | Uniform on $[40,45]$ $\log\phi_{*}$, $\log\phi_{1*}$, $\log\phi_{2*}$, $\log\phi_{3*}$ | Uniform on $[-8,5]$ $\alpha_{F}$ | Fixed $f_{\rm 50}$ (5 Fields) | Fixed In all cases, we employ 100 walkers and 1000 steps, resulting in 100,000 MCMC realizations. In other words, 100 points in the parameter space are randomly selected as initial states, and the MCMC algorithm takes 1000 steps from each initial state to find regions of higher likelihood. There is no “best-fit” solution, but we find that given these generous numbers for walkers and steps, the solutions generally do converge, suggesting that the true best-fit is closely approached. ### 3.3 Cosmic Variance One source of uncertainty in the normalization of our emission-line luminosity functions is cosmic variance. To estimate the expected amplitude of this effect, we use the cosmic variance calculator111 https://www.ph.unimelb.edu.au/$\sim$mtrenti/cvc/ of Trenti & Stiavelli (2008), which employs both the extended Press-Schechter formalism (Press & Schechter, 1974) and numerical simulations to compute the expected variance in any pencil-beam region of the sky. Following Colbert et al. (2013) and Bowman et al. (2021), we compute the cosmic variance by taking the result for each of the five disconnected CANDELS fields and then dividing the average of these estimates by $\sqrt{5}$. The results of this calculation show that for a non-evolving luminosity function over the redshift range $1.16\leq z\leq 1.56$, the cosmic variance expected for our H$\alpha$-emission luminosity function is $\sim 6.7\%$, while that for [O III] $\lambda 5007$ galaxies between $1.16\leq z\leq 1.90$, this number is $\sim 5.3\%$. For the redshift-varying case, the process of measuring the cosmic variance is not so straightforward, as we have modeled cosmic evolution using a quadratic equation, represented using the values of $\phi_{*}$ and $\mathcal{L}_{*}$ at three redshifts. However, if we divide the surveyed redshift range into three bins, we can estimate the effect of cosmic variance on each bin. We find that for both H$\alpha$ and [O III] $\lambda 5007$, the variance should slightly increase with redshift, with the uncertainties being roughly 12% and 9%, respectively. We include the effects of cosmic variance in our calculations for the expected galaxy counts (Tables 3 and 4) as well as §4.4) and the star formation rate density (§4.2). Given the sizes of our galaxy samples and the volumes of space being surveyed ($\gtrsim 0.7$ and $1.4\times 10^{6}$ Mpc3 for the H$\alpha$ and [O III] $\lambda 5007$ studies, respectively), the effects of cosmic variance should be small in the case of the static luminosity function, but non-negligible for our analysis of cosmic evolution. ## 4 Results In this section, we present the H$\alpha$ \+ [N II] and [O III] $\lambda 5007$ luminosity functions of 3D-HST ELGs, along with the H$\alpha$-based cosmic star formation rate density contained in the emission-line galaxies. Tables 3, 4, and 5 present the overall results for our sample. In Tables 3 and 4, we list our best-fit static luminosity functions along with those of Shim et al. (2009), Colbert et al. (2013), Sobral et al. (2013), Khostovan et al. (2015), and Sobral et al. (2015). As a cautionary statement, the luminosity functions being given in the tables are not directly comparable, as detailed in the columns labeled “Notes”. Moreover, because of the well-known degeneracies between the three Schechter (1976) parameters, our values of $\alpha$, $\mathcal{L}$, and $\phi_{i}$ are not necessarily in agreement with those of the previous studies. Nevertheless, we find that the overall form of our H$\alpha$ luminosity function is compatible with the luminosity functions derived by Colbert et al. (2013) and Sobral et al. (2013), but somewhat distinct from the literature measurements for [O III] $\lambda 5007$. In Table 5, we give the parameters of our best-fit H$\alpha$ and [O III] $\lambda 5007$ redshift-evolving luminosity functions. For the latter, we also extend the redshift range to $z=2.35$ using the measurements of Bowman et al. (2021), since their galaxy sample was defined in exactly the same manner as our dataset. We note that in a grism survey, the true survey area is difficult to calculate as contamination from overlapping spectra and edge effects reduce the number of objects included in analyses. Bowman et al. (2021) studied this censoring by masking out those regions of the 3D-HST survey where emission-line detections are compromised, and fitting a luminosity function using only those galaxies in the unmasked areas. They found that the effective survey area of 3D-HST is $\sim 85\%$ of the total survey area; this is consistent with the correction applied by Colbert et al. (2013) and the estimate made by Ciardullo et al. (2014). In this work, we reduce the quoted area of the 3D-HST survey by 15% to approximate the effects of overlapping spectra and edge-losses. Table 3: H$\alpha$ Luminosity Function Schechter Parameters Reference | $z$ | Sample Size | $\log\phi_{*}$ | $\mathcal{L}_{*}$ | $\alpha$ | $\log\,\int_{0.03L^{*}}^{\infty}\phi(L)\,dL$ | Notes ---|---|---|---|---|---|---|--- Shim et al. (2009) | 0.7 - 1.9 | 80 | $-2.48\pm 0.07$ | $42.54\pm 0.06$ | $-1.39$ (fixed) | $-1.67\pm 0.07$ | HST-NICMOS, H$\alpha$ Colbert et al. (2013) | 0.9 - 1.5 | 517 | $-2.70\pm 0.12$ | $42.18\pm 0.10$ | $-1.43\pm 0.17$ | $-1.8\pm 0.2$ | WISPS, H$\alpha$ Sobral et al. (2013) | 1.47 | 515 | $-2.61^{+0.08}_{-0.09}$ | $42.56^{+0.06}_{-0.05}$ | $-1.62^{+0.25}_{-0.29}$ | $-1.6\pm 0.3$ | HiZELS, H$\alpha$, dust-corrected This work | 1.16 - 1.56 | 1892 | $-2.87^{+0.08}_{-0.09}$ | $42.39^{+0.05}_{-0.05}$ | $-1.60^{+0.07}_{-0.07}$ | $-1.84\pm 0.04$aaIncludes cosmic variance (§3.3) in error budget. | 3D-HST, H$\alpha$ +[N II] This work | 1.16 - 1.56 | 1892 | $-2.86^{+0.03}_{-0.03}$ | $42.39^{+0.02}_{-0.02}$ | $-1.60$ (fixed) | $-1.84\pm 0.04$aaIncludes cosmic variance (§3.3) in error budget. | 3D-HST, H$\alpha$ +[N II] Table 4: [O III] $\lambda 5007$ Luminosity Function Schechter Parameters Reference | $z$ | Sample Size | $\log\phi_{*}$ | $\mathcal{L}_{*}$ | $\alpha$ | $\log\,\int_{0.03L^{*}}^{\infty}\phi(L)\,dL$ | Notes ---|---|---|---|---|---|---|--- Colbert et al. (2013) | 0.7 - 1.5 | 192 | $-3.28\pm 0.09$ | $42.39\pm 0.08$ | $-1.5$ (fixed) | $-2.36\pm 0.09$ | WISPS, [O III] $\lambda\lambda 4959,5007$ Colbert et al. (2013) | 1.5 - 2.3 | 58 | $-3.60\pm 0.14$ | $42.83\pm 0.11$ | $-1.5$ (fixed) | $-2.68\pm 0.16$ | WISPS, [O III] $\lambda\lambda 4959,5007$ Khostovan et al. (2015) | 1.42 | 371 | $-2.61^{+0.10}_{-0.09}$ | $42.06^{+0.06}_{-0.05}$ | $-1.60$ (fixed) | $-1.58\pm 0.09$ | HiZELS, H$\beta$ \+ [O III] Sobral et al. (2015) | 1.37 | 1343 | $-2.71^{+0.08}_{-0.09}$ | $42.10^{+0.05}_{-0.04}$ | $-1.60$ (fixed) | $-1.68\pm 0.09$ | CF-HiZELS, H$\beta$ \+ [O III] This work | 1.16 - 1.9 | 2947 | $-2.67^{+0.06}_{-0.06}$ | $42.23^{+0.04}_{-0.04}$ | $-1.50^{+0.07}_{-0.07}$ | $-1.75\pm 0.03$aaIncludes cosmic variance (§3.3) in error budget. | 3D-HST, [O III] $\lambda 5007$ This work | 1.16 - 1.9 | 2947 | $-2.68^{+0.02}_{-0.02}$ | $42.24^{+0.02}_{-0.02}$ | $-1.50$ (fixed) | $-1.76\pm 0.03$aaIncludes cosmic variance (§3.3) in error budget. | 3D-HST, [O III] $\lambda 5007$ Table 5: Redshift-Evolving Schechter Function Parameters Line | $z$ | Size | $\mathcal{L}_{{z_{1}}*}$ | $\mathcal{L}_{{z_{2}}*}$ | $\mathcal{L}_{{z_{3}}*}$ | $\log\phi_{{z_{1}}*}$ | $\log\phi_{{z_{2}}*}$ | $\log\phi_{{z_{3}}*}$ | $\alpha$ | $z_{1}$ | $z_{2}$ | $z_{3}$ ---|---|---|---|---|---|---|---|---|---|---|---|--- H$\alpha$ | 1.16 - 1.56 | 1892 | $42.31^{+0.09}_{-0.08}$ | $42.37^{+0.06}_{-0.06}$ | $42.52^{+0.09}_{-0.08}$ | $-2.84^{+0.11}_{-0.12}$ | $-2.81^{+0.09}_{-0.10}$ | $-3.02^{+0.10}_{-0.12}$ | $-1.60^{+0.08}_{-0.09}$ | 1.18 | 1.36 | 1.54 [O III] | 1.16 - 1.90 | 2947 | $42.34^{+0.10}_{-0.09}$ | $42.30^{+0.06}_{-0.05}$ | $42.56^{+0.08}_{-0.07}$ | $-3.20^{+0.15}_{-0.18}$ | $-2.87^{+0.10}_{-0.11}$ | $-3.00^{+0.13}_{-0.15}$ | $-1.94^{+0.09}_{-0.09}$ | 1.20 | 1.53 | 1.86 [O III] | 1.16 - 2.35 | 4519 | $42.09^{+0.04}_{-0.03}$ | $42.19^{+0.02}_{-0.02}$ | $42.88^{+0.04}_{-0.03}$ | $-2.70^{+0.04}_{-0.04}$ | $-2.47^{+0.03}_{-0.02}$ | $-3.01^{+0.04}_{-0.04}$ | $-1.50$ (fixed) | 1.20 | 1.76 | 2.32 ### 4.1 H$\alpha$ \+ [N II] Luminosity Function As mentioned in §1, the low resolution of the G141 grism and the morphological broadening associated with grism observations prevent us from separating [N II] from H$\alpha$. While prescriptions for correcting the H$\alpha$ luminosity function for [N II] do exist (e.g., Pettini & Pagel, 2004), their applicability at high-redshift is uncertain. Not only are high-$z$ metallicities generally measured via strong line indicators, which are prone to degeneracies and inconsistencies (e.g., Kewley & Ellison, 2008), but at $z\gtrsim 1$, the ionization parameter of emission-line regions is typically larger than that seen at $z\sim 0$ (see Kewley et al., 2019, and references therein). As the ionization parameter rises, the dominant form of nitrogen should shift to N++, thereby weakening then strength of the [N II] lines and decreasing their contribution to the H$\alpha$ \+ [N II] complex. In this paper, we present H$\alpha$ luminosity functions that are uncorrected for the (presumably minor) contribution of [N II]. In Figure 2, the top plot shows the H$\alpha$ \+ [N II] luminosity function with the three fitted parameters from Equation 5: $\alpha$, $\mathcal{L}_{*}$, and $\log\phi_{*}$. Fits from two hundred MCMC iterations are shown in red, and the median fit in displayed gray. The small spread in the solutions suggests a stable and well-characterized result. The plots in the lower triangle show the marginal posterior distributions for the three parameters, along with 2-D contour plots of the MCMC chains. The strong correlations in the contour plots confirm that the three parameters are not independent. This is expected, as the Schechter parameters are not orthogonal variables. However, the ubiquity of the function in the literature makes continued efforts in such a parameterization worthy. Figure 2: Result of our Bayesian MCMC computation for the H$\alpha$ \+ [N II] luminosity function when assuming invariance over the $1.16\leq z\leq 1.95$ redshift interval. The plots in the left/lower part of the figure show the marginal posterior distributions for each parameter and the 2-D cross-sections of the MCMC realizations. As expected, all three parameters are tightly correlated; this is a common feature of Schechter fits when the data extend less then $\sim 1$ dex below $\mathcal{L}_{*}$. Nevertheless, as illustrated by the plot on the top right, the shape of the luminosity function is very well-defined, with the 200 randomly chosen MCMC runs (shown in red) displaying very little scatter about the most-likely solution (plotted in gray). While a non-evolving luminosity function for H$\alpha$ \+ [N II] $\lambda 6584$ is valuable, the non-negligible redshift range of the data, $1.16<z<1.56$ enables us to study the luminosity function’s evolution. As mentioned in §3.2, we did this by employing an approach based on Leja et al. (2020), in which both $\mathcal{L}_{*}$ and $\log\phi_{*}$ are assigned to be quadratic functions of redshift. Because of the dearth of galaxies at very low luminosities, we forced $\alpha$ to be the same across all redshifts; this avoids the issue of degeneracies between $\alpha$, $\mathcal{L}_{*}$, and $\log\phi_{*}$ seen in Figure 2. We find that the redshift-evolving luminosity function solution is quite similar to the static case, with a statistically indistinguishable low- luminosity slope $\alpha$ and similar $\mathcal{L}_{*}$ and $\log\phi_{*}$ values. All relevant quantities are presented in Table 5. Figure 3 shows the best-fit luminosity functions with redshift indicated by the color. The use of quadratic functions for $\mathcal{L}_{*}$ and $\log\phi_{*}$ allows for smooth, differentiable evolution of the luminosity function. The red points on top of the color curve show the $V_{\rm eff}$ result for the entire redshift range. The points are in greater agreement with the redshift-constant luminosity function but are still within the bounds of the redshift-varying function showed here. We also show how our results compare to the literature. We include comparisons to Shim et al. (2009), who found 80 H$\alpha$ emitters at redshifts $0.7<z<1.9$ through a Hubble-NICMOS grism survey; Colbert et al. (2013), who analyzed a sample of 517 H$\alpha$ emitters at $0.9<z<1.5$ found through the Hubble WISP program; and Sobral et al. (2013), who obtained a dataset of 515 H$\alpha$ emitters at $z=1.47$ through the HiZELS narrow-band imaging. To ensure an apples-to-apples comparison, the literature luminosity functions have been modified to reflect the inclusion of [N II] in our data. Shim et al. (2009) and Colbert et al. (2013) assume $F_{{\rm H}\alpha}=0.71F_{{\rm H}\alpha+{\rm[NII]}}$, while Sobral et al. (2013) uses a formula for correcting [N II] based on equivalent width, which gives an average correction of 25% over all their data. We undo these corrections, along with the 1 mag internal extinction correction applied by Sobral et al. (2013). Figure 3 shows the results. Our sample of H$\alpha$-emitting galaxies is $\sim 4$ times larger than that of any previous study. But from the figure, it is clear that the H$\alpha$ luminosity functions of Colbert et al. (2013) and Sobral et al. (2013) are in good agreement with our work. The only serious discrepancy is with the curve produced by Shim et al. (2009), but since that measurement was based on only 80 objects, this difference is not a concern. From both Table 5 and Figure 3, we see that $\mathcal{L}_{*}$ increases with redshift. Given that we are highly complete at all redshifts above $\mathcal{L}>41.8$, this finding must reflect a physical difference as we go back in cosmic time: there are more high-luminosity ELGs at earlier epochs. From Figure 3, we also find fewer low-luminosity objects at higher redshifts, but this result is more subject to completeness issues, and our result may not be robust. We explore the effects of incompleteness in Figure 4. For our main results, we have fixed the minimum completeness fraction at 50%, i.e., we have excluded from the analysis all objects with monochromatic fluxes fainter than the 50% limit shown in Figure 1. But in Figure 4, we perform an experiment in which we vary the minimum threshold from 1% to 80% and observe the effects on the best- fit luminosity function. As shown in the left panels of the figure, the best- fit value for $\alpha$ decreases (gets steeper), $\mathcal{L}_{*}$ increases, and $\phi_{*}$ decreases when the minimum completeness fraction increases. Nevertheless, from the right panel, we see that the overall luminosity function does not vary significantly in the luminosity range we are able to observe. This is because the three Schechter parameters are correlated: various sets of values can lead to the same overall luminosity function. We see from this experiment that the decision of which flux measurements to include changes the parameterization of the luminosity function, but not the overall shape curve. Figure 3: Redshift evolution of the H$\alpha$ \+ [N II] luminosity function. The curves show the MCMC solutions for the redshift-varying Schechter (1976) function, using each redshift’s median parameter values (closely related to the highest-likelihood solution). Also shown are the best-fit curves from Shim et al. (2009), Colbert et al. (2013), and Sobral et al. (2013). Our results are in general agreement with the literature, especially the results of Colbert et al. (2013). The MCMC results are also compatible with our $V_{\rm eff}$ data point (red triangles). In regard to evolution, we find that the knee of the luminosity function, $\mathcal{L}_{*}$, increases with redshift while the normalization factor $\log\phi_{*}$ decreases, although this may not be due to true redshift evolution (see text for more details). In any case, the evolution is not particularly strong in our $[1.16,1.56]$ redshift range. Figure 4: Effects of the minimum completeness fraction considered for H$\alpha$ fluxes on the fitted parameters. The left panels show that $\alpha$ decreases (gets steeper), $\mathcal{L}_{*}$ increases, and $\phi_{*}$ decreases as the completeness limit for the analysis increases. This experiment demonstrates the importance of completeness in determining the Schechter (1976) function parameters. The right panel shows the luminosity functions for the Schechter parameters in the left panels. The lines are colored according to the minimum completeness fraction. We can see that the effects of completeness on the shape of the luminosity function are much less pronounced than that variables used in the parameterization. ### 4.2 Cosmic Star Formation Rate Density The evolution of the star formation rate density (SFRD) of the universe is an important indicator of galaxy growth and evolution. As reviewed by Madau & Dickinson (2014), we have the general picture that star formation peaked around $z\sim 2$, an era dubbed as “cosmic noon,” and has been declining at $\sim 0.1$ dex per Gyr ever since. However, while this outline is known, further measurements of the SFRD, especially for certain populations of galaxies such as ELGs, are useful for improving our knowledge of the evolution of star formation and quantifying how selection effects propagate into this understanding. In this section, we use the H$\alpha$ luminosity function to calculate the SFRD between $1.16\leq z\leq 1.56$. #### 4.2.1 Conversion between H$\alpha$ luminosity and Star Formation Rate H$\alpha$ luminosity is often used as a direct proxy for very recent star formation (under 10 Myr; e.g., Kennicutt & Evans, 2012). As such, applying a conversion formula from H$\alpha$ luminosity to star formation rate (SFR) is a common process. Typically, the conversion from Hao et al. (2011) and Murphy et al. (2011) as compiled by Kennicutt & Evans (2012) is used. Nevertheless, the aforementioned relation is calibrated in the local universe, therefore raising the concern of its application to higher redshifts. For example, at lower metallicities, main sequence stars tend to be bluer and hotter, changing the amount of ionizing radiation emitted by massive stars, and thus the calibration. Furthermore, the H$\alpha$-SFR calibration summarized by Kennicutt & Evans (2012) applies to dust-corrected H$\alpha$ brightness, and at higher redshifts, the details of dust attenuation constitute a major source of uncertainty (e.g., Bouwens et al., 2012; Nagaraj et al., 2021a). This error propagates directly into the SFR conversion, and is then compounded by the fact that our H$\alpha$ data is contaminated by [N II] $\lambda\lambda 6548,6584$, while the SFR calibration of Kennicutt & Evans (2012) is for H$\alpha$ only. Thus, the applicability of the local conversion of H$\alpha$ luminosity to SFR is unclear. We can address this issue directly by calculating our own relation between SFR and observed H$\alpha$ \+ [N II] luminosity. In Paper I, we discussed the SED fitting procedure of the entire sample of $1.16\leq z\leq 1.90$ 3D-HST emission-line galaxies. We used the Bayesian MCMC SED code MCSED (Bowman et al., 2020) to estimate the physical properties of our ELG sample, assuming a Kroupa (2001) initial mass function (IMF), a single-valued (but free) stellar metallicity, a binned star formation history, with the SFR in each bin a free parameter, and a dust attenuation parameterization from Noll et al. (2009) and Kriek & Conroy (2013). MCSED uses simple stellar population (SSP) SEDs from the Flexible Stellar Population Synthesis (FSPS) library (Conroy et al., 2009; Conroy & Gunn, 2010) with Padova isochrones (Bertelli et al., 1994; Girardi et al., 2000; Marigo et al., 2008). Nebular emission is treated via interpolation in tables of CLOUDY models (Ferland et al., 1998, 2013) computed by Byler et al. (2017). Gas-phase metallicity ($Z_{\rm gas}$) is set equal to the stellar metallicity in MCSED, and we fixed the ionization parameter at $\log U=-2.5$ based on the high [O III]/H$\beta$ ratios observed for our sources. Our SED fits are based primarily on the 3D-HST photometry collected by Skelton et al. (2014), which consists of 147 filter bands distributed over the five CANDELS fields and covering the wavelength range from 3,000 Å to 80,000 Å (observed frame). In Paper I, we merged these data with photometry from Swift and GALEX, which extended the wavelength coverage for over 400 sources to $\sim 2,000$ Å (observed-frame). Also, mid- and far-IR measurements from Spitzer and Herschel were added for over 600 sources in Paper II, but to maintain consistency within the sample, we do not include these dust-sensitive wavelengths here. In addition, MCSED is able to employ emission line fluxes in its SED-solution, making it an ideal analysis tool for grism-based surveys. We included H$\alpha$, H$\beta$, and [O III] $\lambda 5007$ fluxes for galaxies whenever available. Our MCSED-based SFR estimates are fairly robust. We find that the mean uncertainty on our SFRs is $0.24$ dex, with a standard deviation of $0.08$ dex. Moreover, in Paper I, we examined how the basic assumptions underlying our SED fits affected the derived properties of the 3D-HST galaxies. For example, we found that changing the ionization parameter, modifying the weights assigned to the emission line fluxes, or fixing galaxy metallicity all led to statistically indistinguishable distributions for the galaxies’ SFRs. In other words, ionization parameter, metallicity, and the inclusion of emission line fluxes do not noticeably affect the systematics of an SFR measurement. In addition, Bowman et al. (2020) showed that for a large sample of 3D-HST galaxies, the choice of dust attenuation curve does not strongly affect the SFR estimate. On the other hand, we do find that the details of a galaxy’s assumed star formation history (SFH) do affect our SFR estimates. Our fits are based on a “non-parametric” SFH, i.e., one in which the SFR of each epoch in a galaxy’s history is fit independently of the other epochs. Such fits have been proven to reduce a bias in SFR measurements that is introduced by the use of parameterized SFHs (e.g., Conroy, 2013; Leja et al., 2017, 2019; Bowman et al., 2020), though the magnitude of this bias reduction depends on the prior used in each age bin (Leja et al., 2019). Our quoted SFRs represent the star formation rate during the most recent age bin, i.e., over the last 100 Myr of cosmic time. Given that we are trying to calibrate the SFR-H$\alpha$ relation, the lack of influence of the H$\alpha$ flux on the SFR measurement suggests that correlations found between SFR and H$\alpha$ are not artificially induced by the method of measuring SFR. Furthermore, the benefit of using MCSED SFRs is that we have a way of connecting the observed H$\alpha$ \+ [N II] fluxes with SFRs that takes into account dust attenuation and contamination by [N II]. This reduces the bias and uncertainty introduced by applying single values for the dust and [N II] corrections. To calculate the mean SFR-H$\alpha$ relation, we use a procedure very similar to that employed by Bowman et al. (2021) in their analysis of the [O III] $\lambda 5007$ luminosity function. We first divide the H$\alpha$ \+ [N II] luminosities into 25 bins with equal numbers of objects each interval. In each bin, we adopt the mean linear SFR and linear H$\alpha$ luminosity as representative values. We then take the logarithm of those values and fit a line. We chose this process because if we average the measurements in logarithmic space, we would underestimate the total SFR of the population. As suggested by Feigelson & Babu (1992), we use the orthogonal distance regression technique to fit the line, since there is non-negligible heteroscedasticity associated with the measurement errors for both luminosity and SFR. To estimate errors in our solution, we bootstrap the H$\alpha$ \+ [N II] luminosity measurements and repeat the procedure above 1000 times. This error estimation process is the same as that described in §3.2 for the $V_{\rm eff}$ method. We present the results for our sample in Figure 5. The data are plotted as blue dots, while the linear mean binned values are shown as amber diamonds. Finally, the best-fit line is shown in red, with the $1\sigma$ uncertainty displayed via the shaded region (nearly too small to be visible). The correlation between apparent H$\alpha$ \+ [N II] luminosity and SFR is quite clear. The best-fit relation is given by Equation 14. The line has a slope consistent with $1$, suggesting that the relation between SFR and uncorrected H$\alpha$ \+ [N II] luminosity is very close to linear. To be consistent with the SFRD compilation of Madau & Dickinson (2014), we have divided the (linear) SFR by $0.67$ as done in their work; this converts an SFR based on the Kroupa (2001) IMF to one based on the Salpeter (1955) IMF with limits of 0.1 and $100\,M_{\odot}$. In the equation $L_{{\rm H}\alpha+{\rm[NII]}}$ is in units of erg s-1 and SFR is in units of $M_{\odot}$/year. $\begin{split}\log{\rm SFR}&=(1.02\pm 0.02)(\log L_{\rm uncorr}-42)\\\ &+(1.056\pm 0.006)-\log(0.67)\end{split}$ (14) Figure 5: Calibration between SFR and observed (not corrected for dust) H$\alpha$ \+ [N II] luminosity for the sample used in this paper. We show the data as blue dots, the logarithm of the mean linear SFR and luminosity in 25 bins as amber diamonds, and the best-fit line to the binned mean values in red, which has a slope nearly identical to one (a linear relationship in linear space). There is a clear correlation in the data. To be complete, we also calculate the SFRD using the Kennicutt & Evans (2012) relation (shown below in Equation 15), after applying corrections for uniform dust attenuation and [N II] contamination. $\log{\rm SFR}=\log L_{{\rm H}\alpha}-41.27-\log(0.67)$ (15) Both Shim et al. (2009) and Colbert et al. (2013) use a 29% correction for [N II], whereas Sobral et al. (2013) use a formula based on equivalent width and find a median correction of 25%. The 29% correction, which we now adopt, is equivalent to a $-0.15$ dex shift in log luminosity. To test the appropriateness of this 29% correction ($\log$ [N II]/H$\alpha$ $=-0.54$), we determined how [N II]/H$\alpha$ should vary as a function of population age and metallicity using the CLOUDY (Ferland et al., 1998, 2013) nebular emission tables created by Byler et al. (2017), under the assumption of $\log U=-2.5$ (the same value used in our MCSED fits). We show the result in Figure 6. According to the CLOUDY lookup tables, the correction is valid only for most galaxies with $\log(Z/Z_{\odot})\gtrsim-0.2$. For galaxies with $\log(Z/Z_{\odot})\lesssim-0.2$, the true relative strength of [N II] is lower than the correction. Figure 6: CLOUDY prediction for $\log$ [N II]/H$\alpha$ as a function of metallicity and age when $\log U=-2.5$. While $\log$ [N II]/H$\alpha$ is nearly independent of age, it is highly dependent on metallicity. For galaxies with $\log(Z/Z_{\odot})\lesssim-0.7$, [N II] emission is less than 1% of the H$\alpha$ emission. Stellar metallicity is difficult to measure using SED fits to mostly broadband photometric data (e.g., Conroy, 2013; Lower et al., 2020). Nevertheless, from our analysis (see Paper I) we find that 82% of the galaxies in our sample have $\log(Z/Z_{\odot})<-0.2$; this is consistent with the relatively small stellar masses of the galaxies (median $\log M_{*}/M_{\odot}\sim 9.5$). In other words, for the majority of H$\alpha$-emitting galaxies, our 29% correction overestimates the contamination by [N II], and thus underestimates the population’s SFRD. Our MCSED-calibration bypasses this issue by removing the need to correct for [N II]. Topping et al. (2021) do a stacking analysis on $z\sim 1.5$ galaxies found in the Multi-Object Spectrometer for Infra-Red Exploration Deep Evolution Field (MOSDEF) survey. They find that $\log$ [N II]/H$\alpha$ is a strong function of stellar mass, which is correlated with metallicity. From their Figure 2, we notice that for galaxies with $\log(M_{*}/\hbox{$M_{\odot}$})\lesssim 10.5$, the 29% correction for [N II] is an overestimate. We find that 81% of our galaxies have masses $\log(M_{*}/\hbox{$M_{\odot}$})<10.5$, which is in perfect agreement with the aforementioned finding using CLOUDY lookup tables. As for dust, based on results from Paper I, we find that the average differential extinction, $E(B-V)$, for our galaxy sample is 0.16 mag. To calculate $A({\rm H}\alpha)$ for our sample, we use the conversions given by Reddy et al. (2020), $\displaystyle E(B-V)_{\rm nebular}$ $\displaystyle=2.07E(B-V)_{\rm stellar}$ (16) $\displaystyle A({\rm H}\alpha)$ $\displaystyle=2.66E(B-V)_{\rm nebular}$ (17) and apply an average value of $A(H\alpha)=0.88$ to the entire sample. The combination of the [N II] and dust corrections results in a multiplicative factor of $1.60$. #### 4.2.2 Star Formation Rate Density Results Given the relation between SFR and uncorrected H$\alpha$ \+ [N II] luminosity (Equation 14), we can calculate the total SFRD using the luminosity function results in this paper. If $\phi(L)$ is the true luminosity function, $\textrm{SFR}(L)$ is the relation given by Equation 14, and $L$ is the uncorrected H$\alpha$ \+ [N II] luminosity, then the total SFRD contained in 3D-HST ELGs brighter than some luminosity $L_{\rm min}$ is ${\rm SFRD}(z)=\int_{L_{\rm min}}^{\infty}{\rm SFR}(L)\,\phi(L,z)\,dL$ (18) Alternatively, we can use the local calibration between SFR and dust-corrected H$\alpha$ luminosity (Equation 15) and apply the average dust and [N II] corrections as described in §4.2.1. The net effect of these two factors is included in the constant $k=1.6$, and ${\rm SFRD}(z)={\rm SFR}_{\rm corr}\left(k\,\int_{L_{\rm min}}^{\infty}L\,\phi(L,z)\,dL\right)$ (19) which simplifies to ${\rm SFRD}(z)={\rm SFR}_{\rm corr}\left(k\,\Gamma\left(\alpha+2,\frac{L_{\rm min}}{L_{*}}\right)\,\phi_{*}(z)\,L_{*}(z)\right)$ (20) where $\Gamma$ represents the incomplete gamma function. Figure 7 shows the evolution of the cosmic SFRD based on results compiled by Madau & Dickinson (2014) from both UV-based and IR-based measurements, along with the best-fit curve from their review paper. We also include SFRD measurements from Gruppioni et al. (2020) using sub-mm data. Our SFRD results for both the non-evolving and redshift-varying luminosity functions, based on our SED-based SFR calibration, are shown as a green star and red line, respectively. Our SFRD measurement derived from the non-evolving luminosity function with the Kennicutt & Evans (2012) relation is shown as an amber diamond. In order to make a fair comparison to the literature, we use $L_{\rm min}=0.03\mathcal{L}_{*}$, as adopted by Madau & Dickinson (2014). Our measurements derived via the SFR calibration shown in Figure 5 yield an SFRD that is larger by $0.12$ dex than that found using the local H$\alpha$-SFR calibration given by Kennicutt & Evans (2012). This discrepancy is consistent with the conclusion of Section 4.2.1, i.e., that the application of a 29% correction for the contribution of [N II] to the H$\alpha$ measurement leads to an underestimate in star-formation rate and the epoch’s SFRD. We therefore believe that our SED-based SFR calibration is more accurate. However, we do note that the uncertainty on the local-calibration- based SFRD is underestimated since the uncertainties on the dust and [N II] corrections are not propagated into the analysis. Thus, the two numbers may still be consistent. We also observe that the cosmic SFRD calculated for our sample of ELGs is at least $\sim 0.09$ dex below the value expected from best-fit curve of Madau & Dickinson (2014). This implies that H$\alpha$-selected galaxies contain $\lesssim 81\%$ of the star formation in the $z\sim 1.4$ universe. The difference between the cosmic SFRD at $z\sim 1.4$, as determined by Madau & Dickinson (2014), and our H$\alpha$-based measurement suggests that not all of the epoch’s star-formation is detectable via surveys for rest-frame optical emission lines. Such a result is easily explained if some star-forming galaxies are heavily obscured by dust; since emission-line gas is generally attenuated more than star-light, this is a reasonable hypothesis (e.g., Charlot & Fall, 2000; Calzetti et al., 2000; Reddy et al., 2020). These dusty star forming galaxies have been shown to contribute significantly to the star formation rate around cosmic noon, i.e., $1<z<3$ (e.g., Casey et al., 2013). At higher redshifts ($z\gtrsim 3$), far-IR and sub-mm studies have shown that UV-based studies miss a significant portion of the star formation due to dust attenuation in extremely high-SFR galaxies (e.g., Gruppioni et al., 2015; Rowan-Robinson et al., 2016; Wang et al., 2019; Williams et al., 2019; Gruppioni et al., 2020; Loiacono et al., 2021; Khusanova et al., 2021). Nevertheless, the discrepancy between the UV-NIR and FIR-sub-mm methods of calculating star formation (e.g., see the extensive discussion by Katsianis et al., 2021) is much less pronounced at the redshifts of our sample. For example, in Figure 7, we include sub-mm results from Gruppioni et al. (2020) and find that for $z=1$ and $z=2$, the results are still consistent with the Madau & Dickinson (2014) best-fit curve. Coming back to our results, we cannot discount the possibility that our SFRD is consistent with the Madau & Dickinson (2014) curve, as the discrepancy is still within the variance of the literature values. Thus, there is no conclusive evidence that the emission line surveys are missing a significant fraction of sources found in UV-based and/or IR/sub-mm-based surveys between redshifts $1.16<z<1.56$. While the redshift-evolving SFRD is not a flat curve, given the non-negligible uncertainties due to cosmic variance, we cannot dismiss the possibility of no evolution. In other words, our H$\alpha$ emission-line measurements find no strong evidence for SFRD evolution over our $1.16<z<1.56$ redshift range. Figure 7: Our calculated star formation rate density (SFRD) compared to literature values (UV-based results are in blue dots, IR results in red, and sub-mm-based in brown) and the best-fit curve (shown in black) from Madau & Dickinson (2014). The green star and red line show the SFRD when using our MCSED-based H$\alpha$ \+ [N II] SFR calibration with the static and redshift- varying luminosity functions, respectively. By directly connecting the observed H$\alpha$ +[N II] fluxes to SFR, we bypass the issues associated with dust attenuation and [N II] contamination, while propagating in their uncertainties. The amber diamond shows the SFRD calculated using the non- evolving luminosity function with the Kennicutt & Evans (2012) H$\alpha$-SFR calibration; here the uncertainties from dust and [N II] are not included in the error bar. This latter calibration produces an SFRD that is smaller by $0.12$ dex than SED-based value; this is consistent with the overestimate of [N II] in low-mass (low-metallicity) galaxies. At $z\sim 1.4$, the SFRD from ELGs is $\gtrsim 0.09$ dex lower than the Madau & Dickinson (2014) curve, but this difference is still within the variance seen in the literature. We are not able to find conclusive evidence of SFRD evolution between $1.16\leq z\leq 1.56$. ### 4.3 [O III] $\lambda 5007$ Luminosity Function Unlike H$\alpha$, [O III] $\lambda 5007$ does not suffer from uncertainties due to blending. At the redshifts under consideration, the G141 grism has enough resolution and the sources are sufficiently small so that [O III] and H$\beta$ are easily distinguishable. [O III] $\lambda 5007$ is blended with [O III] $\lambda 4959$, but since the ratio of the two lines is fixed by basic physics (2.98:1; Storey & Zeippen, 2000), converting the observed [O III] feature into [O III] $\lambda 5007$ is trivial. We quote only the luminosity of [O III] $\lambda 5007$ in the subsequent analysis. In Figure 8, we show our fitted non-evolving [O III] $\lambda 5007$ luminosity function (over the range $1.16<z<1.90$) and place it in the context of the literature. This includes fits to 192 $0.7<z<1.5$ and 58 $1.5<z<2.3$ [O III] emitters by Colbert et al. (2013), 371 $z=1.42$ [O III] + H$\beta$ emitters from Khostovan et al. (2015), and 1343 $z\sim 1.4$ [O III] + H$\beta$ emitters from Sobral et al. (2015). For our comparison, the [O III] $\lambda\lambda 4959,5007$ values of these works have been converted to [O III] $\lambda 5007$ to match our measurements. The studies by Khostovan et al. (2015) and Sobral et al. (2015) identify H$\beta$ emitters as well as [O III] galaxies, since their narrow-band photometry is unable to distinguish the two object classes without spectroscopic follow-up. Thus, it is unclear how their samples compare to ours. Sobral et al. (2015) find that in their set of sources with spectroscopic confirmation, H$\beta$ emitters constitute around 16% of their $z\sim 1.4$ sample. Interestingly, they notice that these H$\beta$ emitters tend to have lower luminosities than the [O III]-identified emitters, and this may be reflected in the lower $\mathcal{L}_{*}$ values implied by the plot. Still 5/6’s of their (spectroscopically confirmed) sample is made up of [O III] emitters, so the effect of these contaminants should not be large. Khostovan et al. (2015) find a similar phenomenon in their sample. As shown in Figure 8, at $41.5\lesssim\log L\lesssim 42.3$, our luminosity function predicts more [O III] galaxies than any of the other studies, though at higher luminosities our results are well within the bounds of the literature. Moreover, our MCMC result is in good agreement with the $V_{\rm eff}$ result (blue triangles). There are a few factors that may lead to our distinct result for [O III]. The first is that our redshift range is unique. No other study focuses specifically on $1.16<z<1.90$ galaxies. As the luminosity function is an evolving quantity, the particular redshifts involved play a role in determining the measured parameters. Another important consideration is that our sample size is the largest to date, and the larger the sample size, the more accurate the luminosity function. In fact, as shown in Table 4, the value for $\log\,\int_{0.03L^{*}}^{\infty}\phi(L)\,dL$ derived by Sobral et al. (2015) is closest to our study, and it is also the measurement that is most consistent with our value. Alternatively, we note that our analysis encounters problems at both the faint and bright ends of the luminosity function. At the faint end, we are subject to rapid loss of completeness while at the bright end, our measurements of the luminosity function suffer from cosmic variance as we are dealing with small numbers of galaxies. Coupled with degeneracies between Schechter parameters, these effects can lead to divergences between different studies. Extrapolations to low-luminosity galaxies, especially when $\alpha$ is fixed, are often uncertain and should be considered as such. Nevertheless, given our large sample size and the number of galaxy measurements at luminosities significantly less than $L_{*}$, we believe that the differences between our results and those of previous studies are real and not an artifact of our analysis. Figure 8: [O III] $\lambda 5007$ luminosity function from this work (MCMC as the red curves and $V_{\rm eff}$ as blue triangles), as well as fits from Colbert et al. (2013), Khostovan et al. (2015), and Sobral et al. (2015). We predict higher counts of intermediate-luminosity galaxies than the other studies, but at the high-luminosity end, our data are well within the range of values previously derived. There is good agreement between the MCMC and $V_{\rm eff}$ methods. In Figure 9, we combine our [O III] $\lambda 5007$ measurements with those of Bowman et al. (2021) to show the redshift evolution of the [O III] $\lambda 5007$ luminosity function between $1.16\leq z\leq 2.35$. For this analysis, we fix $\alpha=-1.5$, as it very difficult to fit $\alpha$ as a free parameter when the limiting luminosity depends so much on redshift. Like in Figure 2, we show the 1D and 2D cross sections of the MCMC chains in the lower left panels. Once again, we find that $\mathcal{L}_{*}$ and $\log\phi_{*}$ are correlated, but only at the same redshift. In other words, $\mathcal{L}_{1*}$ and $\log\phi_{1*}$ are highly correlated but $\mathcal{L}_{1*}$ and $\log\phi_{2*}$ are not. The data of Figure 9 show that the characteristic luminosity $\mathcal{L}_{*}$ increases with redshift. Moreover, as redshift increases, the overall luminosity function increases at all but the lowest luminosities. In other words, there are many more [O III] $\lambda 5007$-visible galaxies, and especially more [O III] $\lambda 5007$-bright systems, at earlier epochs of cosmic history. This is consistent with the findings of Zeimann et al. (2014), Khostovan et al. (2015), and Bowman et al. (2019), among others, that show the prevalence of [O III] compared to [O II] $\lambda 3727$ at high redshift. Figure 9: MCMC result for the redshift varying [O III] $\lambda 5007$ luminosity function. In the lower left, we show the 1D and 2D cross sections of the parameter chains. Note that at each redshift, $\mathcal{L}_{*}$ and $\log\phi_{*}$ are correlated, but the correlation is not strong across redshifts. The upper-right panel shows the evolution of the luminosity function. Our data show that $\mathcal{L}_{*}$ increases with redshift, and, at most luminosities, the luminosity function is larger at higher redshift. In other words, [O III] $\lambda 5007$-bright galaxies were much more common at $z\sim 2$ than at $z\sim 1$. ### 4.4 Number Counts The precision to which one can measure cosmological parameters through galaxy surveys depends on the number of galaxies and the square of the bias of the observed galaxy population relative to dark matter. Surveys planned for Euclid and Roman will observe millions of H$\alpha$-visible galaxies at $0.9\lesssim z\lesssim 1.8$ and [O III]-visible galaxies at $1.5\lesssim z\lesssim 2.7$, and thus measure quantities such as the angular diameter distance and Hubble parameter at these distant epochs. Specifically, the Euclid Wide Survey (WS) will observe $\sim 15000~{}{\rm deg}^{2}$ of the sky down to a flux limit of $\sim 2\times 10^{-16}$ $\text{erg cm}^{-2}\text{ s}^{-1}$ (Euclid Collaboration et al., 2022), while the Roman High Latitude Survey (HLS) will observe $\sim 2200~{}{\rm deg}^{2}$ down to a flux limit of $\sim 6\times 10^{-17}$ $\text{erg cm}^{-2}\text{ s}^{-1}$ (Spergel et al., 2015). The Euclid Deep Survey will observe $50~{}{\rm deg}^{2}$ in three fields to a similar flux limit (Vavrek et al., 2016). In this section, we use our measurements of the H$\alpha$ and [O III] $\lambda 5007$ luminosity functions to calculate the number of galaxies these surveys are likely to measure, corrected for completeness. One factor to note is that since we have removed AGN from the sample, the number counts we derive will be a slight underestimate. We begin with the total number of galaxies with luminosities greater than $L_{\rm min}(z)$ from redshifts $z_{0}$ to $z_{1}$ over area $\Omega$, $N_{\rm tot}=\int d\Omega\int_{z_{0}}^{z_{1}}dz\frac{dV}{dz\,d\Omega}\int_{L_{\rm min}(z)}^{\infty}dL\,\Omega(L,z)\phi(L,z)$ (21) Given the Schechter function with parameters $\alpha$, $L_{*}$, and $\phi_{*}$ (that are assumed to be invariant over the effective survey area $\Omega_{0}$), we can simplify Equation 21 to $N_{\rm tot}=\Omega_{0}\int_{z_{0}}^{z_{1}}dz\frac{dV}{dzd\Omega}\,\Gamma\left(\alpha+1,\frac{L_{\rm min}(z)}{L_{*}}\right)\phi_{*}(z)L_{*}(z)$ (22) where $\Gamma$ once again represents the incomplete gamma function. In Figure 10, we show the total number of galaxies per square degree that are expected to have H$\alpha$ at $1.2\leq z\leq 1.6$ (left) and [O III] $\lambda 5007$ at $1.5\leq z\leq 1.9$ (right) above a given threshold. We calculate these values using Equation 22 with $\Omega_{0}=1$ deg2, while translating $L_{\rm min}$ to emission line flux using the luminosity distance $D_{L}$ via $F_{\rm min}=\frac{L_{\rm min}(z)}{4\pi D_{L}(z)^{2}}$ (23) For the figure, we use the non-evolving luminosity functions (though the redshift varying luminosity functions yield similar results) and perform the same calculations for the luminosity functions given in the literature. In all cases, the values represent galaxy number counts, assuming 100% completeness. In addition, we include the direct measurements of counts obtained by Bagley et al. (2020) using samples of ELGs from the WFC3 Infrared Parallel Spectroscopic Survey (WISPS), 3D-HST, and A Grism H-Alpha SpecTroscopic survey (AGHAST). This work also made corrections for completeness. In the left panel, we observe that our H$\alpha$ counts are in excellent agreement with those of Colbert et al. (2013) at all values of the limiting flux. The counts also agree with those of Bagley et al. (2020) around the limiting flux of the Euclid WS, but less so at brighter fluxes. Given the consistency of our results with those of Colbert et al. (2013) and Bagley et al. (2020), there is no need to update the prediction of $\sim 3300$ deg-2 for the Euclid WS at 100% completeness (Bagley et al., 2020; Euclid Collaboration et al., 2022) and 16.4 million $7\sigma$ H$\alpha$ galaxy detections at $1.06<z<1.88$ for the Roman HLS assuming 70% completeness (Spergel et al., 2015). On the other hand, our result for [O III] (right side of Figure 10) is distinct from the literature. For limiting fluxes of 1 to $3\times 10^{-16}$ $\text{erg cm}^{-2}\text{ s}^{-1}$, our predicted galaxy counts agree with the results of Colbert et al. (2013) and Bagley et al. (2020). As these limits are similar to those of the Euclid WS ($2\times 10^{-16}$ $\text{erg cm}^{-2}\text{ s}^{-1}$) and the $\geq 7\sigma$ limit of $10^{-16}$ $\text{erg cm}^{-2}\text{ s}^{-1}$ used by Spergel et al. (2015) for predicting Roman HLS counts, our results do not change the predictions for these surveys; at 70% completeness, the Roman HLS should detect 1.4 million [O III] $1.88<z<2.77$ galaxies with $>7\sigma$ confidence (Spergel et al., 2015). However, as we push to lower flux limits, our luminosity function predicts higher counts. For example, for the Roman HLS nominal flux limit of $6.0\times 10^{-17}$ $\text{erg cm}^{-2}\text{ s}^{-1}$, we predict twice as many galaxies as Colbert et al. (2013) in the redshift range $1.5<z<1.9$. Figure 10: Predictions for total number of galaxies per square degree corrected for completeness, as a function of the limiting flux for H$\alpha$ at $1.2\leq z\leq 1.6$ (left) and [O III] $\lambda 5007$ at $1.5\leq z\leq 1.9$ (right). We use our non-evolving luminosity functions for these calculations and include cosmic variance in the error budget. Our H$\alpha$ counts agree remarkably well with those from Colbert et al. (2013) at all limiting fluxes as well as with the direct counts from Bagley et al. (2020) around $F_{\rm lim}\sim 2\times 10^{-16}$ $\text{erg cm}^{-2}\text{ s}^{-1}$. This agreement suggests no changes to the H$\alpha$ predictions for Euclid and Roman are necessary. While our [O III] $\lambda 5007$ count predictions are distinct from the literature, they still do not greatly change the expected counts from Euclid. However, for the HLS flux limit of $6.0\times 10^{-17}$ $\text{erg cm}^{-2}\text{ s}^{-1}$, we predict twice as many galaxies as Colbert et al. (2013). ## 5 Conclusion In Papers I and II, we used the G141 grism data from the Hubble 3D-HST Treasury program (GO-11600, 12177, 12328; Brammer et al., 2012; Momcheva et al., 2016) to identify a clean sample of 4350 normal (i.e., non-AGN) star- forming emission lines galaxies in the redshift range $1.16\leq z\leq 1.90$. In this work, we use the line fluxes of these galaxies, along with a similar set of fluxes measured by Bowman et al. (2021), to measure the galaxies’ emission-line luminosity functions. These data include 1892 H$\alpha$ emitting galaxies between $1.16\leq z\leq 1.56$ and 4519 [O III] emitters with $1.16\leq z\leq 2.35$, all with line fluxes above the 3D-HST 50% completeness limit. While there have been several previous efforts to calculate H$\alpha$ and [O III] luminosity functions in these redshift ranges (see §1 for a detailed list), none used the large sample sizes used here. We employ a generalization of the classical $1/V_{\rm max}$ method to derive the emission-line luminosity functions for our entire sample of galaxies, and samples of galaxies broken down by redshift. We then use Markov Chain Monte Carlo (MCMC) Bayesian techniques to fit these data to the Schechter (1976) luminosity function with $\alpha$ held constant across redshift. We find very good agreement between our H$\alpha$ results and those from the literature, and our [O III] luminosity function is also a good match to prior measurements for line luminosities brighter than $\log L=42.3$ (ergs s-1). However, at fainter [O III] luminosities (where completeness corrections might be an issue), we infer an excess of objects. These results are shown in Figures 2, 3, 8, and 9 and summarized in Tables 3 \- 5. We also compute the star formation rate density (SFRD) of the $1.16\leq z\leq 1.56$ epoch using the H$\alpha$ luminosity function. We find that our SFRD is $\sim 19\%$ smaller than the best-fit $z\sim 1.4$ value found by Madau & Dickinson (2014) (Figure 7), though this discrepancy is within the variance found in the literature. If the difference is real, then one possible explanation is that not all $z\sim 1.4$ star formation takes place in galaxies with observable H$\alpha$ emission lines. In particular, surveys such as 3D-HST will miss heavily obscured galaxies where the emission lines are too extinguished to make it into the sample. We find no evidence for or against cosmic evolution of the SFRD between $1.16<z<1.56$ Finally, we predict total galaxy counts per square degree as a function of the limiting flux (Figure 10). For H$\alpha$ \+ [N II] $\lambda 6584$, our results are consistent with those from Colbert et al. (2013) at all limiting fluxes and Bagley et al. (2020) down to a limiting flux of $\sim 2\times 10^{-16}$ $\text{erg cm}^{-2}\text{ s}^{-1}$, suggesting the previous predictions for the Euclid (Euclid Collaboration et al., 2022) and Roman (Spergel et al., 2015) surveys are accurate. For [O III] $\lambda 5007$, our numbers agree with previous estimates for the Euclid Wide Survey, but depending on where exactly we define the flux limit for the Roman High Latitude Survey, our data may imply a significantly larger number of detectable galaxies. For example, at $10^{-16}$ $\text{erg cm}^{-2}\text{ s}^{-1}$, our results are consistent with the analysis of Colbert et al. (2013), but at $6\times 10^{-17}$ $\text{erg cm}^{-2}\text{ s}^{-1}$, our number counts are higher by a factor of two. Roman may find more faint [O III] emitters than previously anticipated. The $\Lambda$CDM paradigm has garnered many resounding successes in explaining observations of our universe at a variety of scales. However, there are still inconsistencies and unknowns leaving the cosmological model incomplete. To better constrain the model as well as alternate or additional theories, we need to continue honing our observations. Large galaxy surveys represent an important avenue to constrain cosmological parameters through the measurement of baryonic acoustic oscillations and redshift space distortions. Galaxy surveys with precise redshifts will be especially useful for generating the necessary constraints, and IFU and slitless spectroscopy are the most efficient ways of performing these surveys. In the near future, Euclid and Roman will greatly enhance samples of emission line galaxies. The similarities between 3D-HST and these planned surveys make it a perfect pathfinder mission. Our measurements of the H$\alpha$ and [O III] $\lambda 5007$ luminosity functions with galaxy samples that are several times larger than any previous study of the $z\sim 1.5$ redshift range help cement the predictions for the expected yield of the ongoing and future surveys. We thank the anonymous referee for their insightful advice that helped make the paper more thorough. This work has made use of the Rainbow Cosmological Surveys Database, which is operated by the Centro de Astrobiología (CAB/INTA), partnered with the University of California Observatories at Santa Cruz (UCO/Lick,UCSC). This work is based on observations taken by the CANDELS Multi-Cycle Treasury Program with the NASA/ESA HST, which is operated by the Association of Universities for Research in Astronomy, Inc., under NASA contract NAS5-26555. This research has made use of NASA’s Astrophysics Data System. This research has made use of the SVO Filter Profile Service (http://svo2.cab.inta-csic.es/theory/fps/) supported from the Spanish MINECO through grant AYA2017-84089. Computations for this research were performed on the Pennsylvania State University’s Institute for Computational and Data Sciences’ Roar supercomputer. The Institute for Gravitation and the Cosmos is supported by the Eberly College of Science and the Office of the Senior Vice President for Research at the Pennsylvania State University. This material is based upon work supported by the National Science Foundation Graduate Research Fellowship under Grant No. DGE1255832. Any opinion, findings, and conclusions or recommendations expressed in this material are those of the authors(s) and do not necessarily reflect the views of the National Science Foundation. ## References * Abareshi et al. (2022) Abareshi, B., Aguilar, J., Ahlen, S., et al. 2022, AJ, 164, 207 * Amendola et al. (2018) Amendola, L., Appleby, S., Avgoustidis, A., et al. 2018, Living Reviews in Relativity, 21, 2 * Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33 * Astropy Collaboration et al. (2018) Astropy Collaboration, Price-Whelan, A. M., Sipőcz, B. M., et al. 2018, AJ, 156, 123 * Atek et al. (2010) Atek, H., Malkan, M., McCarthy, P., et al. 2010, ApJ, 723, 104 * Avni & Bahcall (1980) Avni, Y., & Bahcall, J. N. 1980, ApJ, 235, 694 * Bagley et al. (2020) Bagley, M. B., Scarlata, C., Mehta, V., et al. 2020, ApJ, 897, 98 * Bennett et al. (2013) Bennett, C. L., Larson, D., Weiland, J. L., et al. 2013, ApJS, 208, 20 * Bertelli et al. (1994) Bertelli, G., Bressan, A., Chiosi, C., Fagotto, F., & Nasi, E. 1994, A&AS, 106, 275 * Blake et al. (2011) Blake, C., Kazin, E. A., Beutler, F., et al. 2011, MNRAS, 418, 1707 * Bouwens et al. (2012) Bouwens, R. J., Illingworth, G. D., Oesch, P. A., et al. 2012, ApJ, 754, 83 * Bowman et al. (2019) Bowman, W. P., Zeimann, G. R., Ciardullo, R., et al. 2019, ApJ, 875, 152 * Bowman et al. (2020) Bowman, W. P., Zeimann, G. R., Nagaraj, G., et al. 2020, ApJ, 899, 7 * Bowman et al. (2021) Bowman, W. P., Ciardullo, R., Zeimann, G. R., et al. 2021, ApJ, 920, 78 * Brammer et al. (2012) Brammer, G. B., van Dokkum, P. G., Franx, M., et al. 2012, ApJS, 200, 13 * Brandt & Alexander (2015) Brandt, W. N., & Alexander, D. M. 2015, A&A Rev., 23, 1 * Bull et al. (2016) Bull, P., Akrami, Y., Adamek, J., et al. 2016, Physics of the Dark Universe, 12, 56 * Byler et al. (2017) Byler, N., Dalcanton, J. J., Conroy, C., & Johnson, B. D. 2017, ApJ, 840, 44 * Calzetti et al. (2000) Calzetti, D., Armus, L., Bohlin, R. C., et al. 2000, ApJ, 533, 682 * Casey et al. (2013) Casey, C. M., Chen, C.-C., Cowie, L. L., et al. 2013, MNRAS, 436, 1919 * Cepa et al. (2016) Cepa, J., Benítez, N., Dupke, R., et al. 2016, in Astronomical Society of the Pacific Conference Series, Vol. 507, Multi-Object Spectroscopy in the Next Decade: Big Questions, Large Surveys, and Wide Fields, ed. I. Skillen, M. Balcells, & S. Trager, 381 * Charlot & Fall (2000) Charlot, S., & Fall, S. M. 2000, ApJ, 539, 718 * Ciardullo et al. (2013) Ciardullo, R., Gronwall, C., Adams, J. J., et al. 2013, ApJ, 769, 83 * Ciardullo et al. (2014) Ciardullo, R., Zeimann, G. R., Gronwall, C., et al. 2014, ApJ, 796, 64 * Colbert et al. (2013) Colbert, J. W., Teplitz, H., Atek, H., et al. 2013, ApJ, 779, 34 * Comparat et al. (2016) Comparat, J., Zhu, G., Gonzalez-Perez, V., et al. 2016, MNRAS, 461, 1076 * Conroy (2013) Conroy, C. 2013, ARA&A, 51, 393 * Conroy & Gunn (2010) Conroy, C., & Gunn, J. E. 2010, ApJ, 712, 833 * Conroy et al. (2009) Conroy, C., Gunn, J. E., & White, M. 2009, ApJ, 699, 486 * Davis et al. (2003) Davis, M., Faber, S. M., Newman, J., et al. 2003, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 4834, Discoveries and Research Prospects from 6- to 10-Meter-Class Telescopes II, ed. P. Guhathakurta, 161 * Dawson et al. (2013) Dawson, K. S., Schlegel, D. J., Ahn, C. P., et al. 2013, AJ, 145, 10 * Dawson et al. (2016) Dawson, K. S., Kneib, J.-P., Percival, W. J., et al. 2016, AJ, 151, 44 * DESI Collaboration et al. (2016) DESI Collaboration, Aghamousa, A., Aguilar, J., et al. 2016, arXiv e-prints, arXiv:1611.00036 * Donley et al. (2012) Donley, J. L., Koekemoer, A. M., Brusa, M., et al. 2012, ApJ, 748, 142 * Dressler et al. (2012) Dressler, A., Spergel, D., Mountain, M., et al. 2012, arXiv e-prints, arXiv:1210.7809 * Drinkwater et al. (2010) Drinkwater, M. J., Jurek, R. J., Blake, C., et al. 2010, MNRAS, 401, 1429 * Eddington (1913) Eddington, A. S. 1913, MNRAS, 73, 359 * Euclid Collaboration et al. (2022) Euclid Collaboration, Scaramella, R., Amiaux, J., et al. 2022, A&A, 662, A112 * Feigelson & Babu (1992) Feigelson, E. D., & Babu, G. J. 1992, ApJ, 397, 55 * Ferland et al. (1998) Ferland, G. J., Korista, K. T., Verner, D. A., et al. 1998, PASP, 110, 761 * Ferland et al. (2013) Ferland, G. J., Porter, R. L., van Hoof, P. A. M., et al. 2013, Rev. Mexicana Astron. Astrofis., 49, 137 * Fleming et al. (1995) Fleming, D. E. B., Harris, W. E., Pritchet, C. J., & Hanes, D. A. 1995, AJ, 109, 1044 * Foreman-Mackey et al. (2013) Foreman-Mackey, D., Hogg, D. W., Lang, D., & Goodman, J. 2013, PASP, 125, 306 * Fujita et al. (2003) Fujita, S. S., Ajiki, M., Shioya, Y., et al. 2003, ApJ, 586, L115 * Gallego et al. (1995) Gallego, J., Zamorano, J., Aragon-Salamanca, A., & Rego, M. 1995, ApJ, 455, L1 * Geach et al. (2008) Geach, J. E., Smail, I., Best, P. N., et al. 2008, MNRAS, 388, 1473 * Gebhardt et al. (2021) Gebhardt, K., Mentuch Cooper, E., Ciardullo, R., et al. 2021, ApJ, 923, 217 * Girardi et al. (2000) Girardi, L., Bressan, A., Bertelli, G., & Chiosi, C. 2000, A&AS, 141, 371 * Glazebrook et al. (2004) Glazebrook, K., Tober, J., Thomson, S., Bland-Hawthorn, J., & Abraham, R. 2004, AJ, 128, 2652 * Green et al. (2012) Green, J., Schechter, P., Baltay, C., et al. 2012, arXiv e-prints, arXiv:1208.4012 * Grogin et al. (2011) Grogin, N. A., Kocevski, D. D., Faber, S. M., et al. 2011, ApJS, 197, 35 * Gruppioni et al. (2015) Gruppioni, C., Calura, F., Pozzi, F., et al. 2015, MNRAS, 451, 3419 * Gruppioni et al. (2020) Gruppioni, C., Béthermin, M., Loiacono, F., et al. 2020, A&A, 643, A8 * Hao et al. (2011) Hao, C.-N., Kennicutt, R. C., Johnson, B. D., et al. 2011, ApJ, 741, 124 * Hayashi et al. (2020) Hayashi, M., Shimakawa, R., Tanaka, M., et al. 2020, PASJ, 72, 86 * Hill et al. (2021) Hill, G. J., Lee, H., MacQueen, P. J., et al. 2021, AJ, 162, 298 * Hippelein et al. (2003) Hippelein, H., Maier, C., Meisenheimer, K., et al. 2003, A&A, 402, 65 * Huchra & Sargent (1973) Huchra, J., & Sargent, W. L. W. 1973, ApJ, 186, 433 * Jones & Bland-Hawthorn (2001) Jones, D. H., & Bland-Hawthorn, J. 2001, ApJ, 550, 593 * Jones et al. (2001–) Jones, E., Oliphant, T., Peterson, P., et al. 2001–, SciPy: Open source scientific tools for Python * Katsianis et al. (2021) Katsianis, A., Yang, X., & Zheng, X. 2021, ApJ, 919, 88 * Kennicutt & Evans (2012) Kennicutt, R. C., & Evans, N. J. 2012, ARA&A, 50, 531 * Kewley & Ellison (2008) Kewley, L. J., & Ellison, S. L. 2008, ApJ, 681, 1183 * Kewley et al. (2019) Kewley, L. J., Nicholls, D. C., & Sutherland, R. S. 2019, ARA&A, 57, 511 * Khostovan et al. (2015) Khostovan, A. A., Sobral, D., Mobasher, B., et al. 2015, MNRAS, 452, 3948 * Khostovan et al. (2020) Khostovan, A. A., Malhotra, S., Rhoads, J. E., et al. 2020, MNRAS, 493, 3966 * Khusanova et al. (2021) Khusanova, Y., Bethermin, M., Le Fèvre, O., et al. 2021, A&A, 649, A152 * Koekemoer et al. (2011) Koekemoer, A. M., Faber, S. M., Ferguson, H. C., et al. 2011, ApJS, 197, 36 * Kriek & Conroy (2013) Kriek, M., & Conroy, C. 2013, ApJ, 775, L16 * Kroupa (2001) Kroupa, P. 2001, MNRAS, 322, 231 * Laureijs et al. (2011) Laureijs, R., Amiaux, J., Arduini, S., et al. 2011, arXiv e-prints, arXiv:1110.3193 * Laureijs et al. (2012) Laureijs, R., Gondoin, P., Duvet, L., et al. 2012, Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 8442, Euclid: ESA’s mission to map the geometry of the dark universe, 84420T * Leja et al. (2019) Leja, J., Carnall, A. C., Johnson, B. D., Conroy, C., & Speagle, J. S. 2019, ApJ, 876, 3 * Leja et al. (2017) Leja, J., Johnson, B. D., Conroy, C., van Dokkum, P. G., & Byler, N. 2017, ApJ, 837, 170 * Leja et al. (2020) Leja, J., Speagle, J. S., Johnson, B. D., et al. 2020, ApJ, 893, 111 * Lilly et al. (2007) Lilly, S. J., Le Fèvre, O., Renzini, A., et al. 2007, ApJS, 172, 70 * Loiacono et al. (2021) Loiacono, F., Decarli, R., Gruppioni, C., et al. 2021, A&A, 646, A76 * Lower et al. (2020) Lower, S., Narayanan, D., Leja, J., et al. 2020, ApJ, 904, 33 * Ly et al. (2011) Ly, C., Lee, J. C., Dale, D. A., et al. 2011, ApJ, 726, 109 * Ly et al. (2007) Ly, C., Malkan, M. A., Kashikawa, N., et al. 2007, ApJ, 657, 738 * Lynden-Bell (1971) Lynden-Bell, D. 1971, MNRAS, 155, 95 * Madau & Dickinson (2014) Madau, P., & Dickinson, M. 2014, ARA&A, 52, 415 * Marigo et al. (2008) Marigo, P., Girardi, L., Bressan, A., et al. 2008, A&A, 482, 883 * Marshall (1985) Marshall, H. L. 1985, ApJ, 299, 109 * Marshall et al. (1983) Marshall, H. L., Tananbaum, H., Avni, Y., & Zamorani, G. 1983, ApJ, 269, 35 * Mehta et al. (2015) Mehta, V., Scarlata, C., Colbert, J. W., et al. 2015, ApJ, 811, 141 * Momcheva et al. (2016) Momcheva, I. G., Brammer, G. B., van Dokkum, P. G., et al. 2016, ApJS, 225, 27 * Murphy et al. (2011) Murphy, E. J., Condon, J. J., Schinnerer, E., et al. 2011, ApJ, 737, 67 * Nagaraj et al. (2021a) Nagaraj, G., Ciardullo, R., Bowman, W. P., & Gronwall, C. 2021a, ApJ, 913, 34 * Nagaraj et al. (2021b) Nagaraj, G., Ciardullo, R., Lawson, A., et al. 2021b, ApJ, 912, 145 * Noll et al. (2009) Noll, S., Pierini, D., Cimatti, A., et al. 2009, A&A, 499, 69 * Oke (1974) Oke, J. B. 1974, ApJS, 27, 21 * Perlmutter et al. (1999) Perlmutter, S., Aldering, G., Goldhaber, G., et al. 1999, ApJ, 517, 565 * Pettini & Pagel (2004) Pettini, M., & Pagel, B. E. J. 2004, MNRAS, 348, L59 * Pirzkal et al. (2004) Pirzkal, N., Xu, C., Malhotra, S., et al. 2004, ApJS, 154, 501 * Pirzkal et al. (2013) Pirzkal, N., Rothberg, B., Ly, C., et al. 2013, ApJ, 772, 48 * Press & Schechter (1974) Press, W. H., & Schechter, P. 1974, ApJ, 187, 425 * Price et al. (2014) Price, S. H., Kriek, M., Brammer, G. B., et al. 2014, ApJ, 788, 86 * Reddy et al. (2020) Reddy, N. A., Shapley, A. E., Kriek, M., et al. 2020, ApJ, 902, 123 * Riess et al. (1998) Riess, A. G., Filippenko, A. V., Challis, P., et al. 1998, AJ, 116, 1009 * Rowan-Robinson et al. (2016) Rowan-Robinson, M., Oliver, S., Wang, L., et al. 2016, MNRAS, 461, 1100 * Salim & Narayanan (2020) Salim, S., & Narayanan, D. 2020, ARA&A, 58, 529 * Salpeter (1955) Salpeter, E. E. 1955, ApJ, 121, 161 * Salzano et al. (2021) Salzano, V., Pigozzo, C., Benetti, M., et al. 2021, J. Cosmology Astropart. Phys, 2021, 033 * Schechter (1976) Schechter, P. 1976, ApJ, 203, 297 * Schmidt (1968) Schmidt, M. 1968, ApJ, 151, 393 * Schmidt (1970) —. 1970, ApJ, 162, 371 * Shim et al. (2009) Shim, H., Colbert, J., Teplitz, H., et al. 2009, ApJ, 696, 785 * Shivaei et al. (2020) Shivaei, I., Reddy, N., Rieke, G., et al. 2020, ApJ, 899, 117 * Skelton et al. (2014) Skelton, R. E., Whitaker, K. E., Momcheva, I. G., et al. 2014, ApJS, 214, 24 * Sobral et al. (2012) Sobral, D., Best, P. N., Matsuda, Y., et al. 2012, MNRAS, 420, 1926 * Sobral et al. (2011) Sobral, D., Best, P. N., Smail, I., et al. 2011, MNRAS, 411, 675 * Sobral et al. (2013) Sobral, D., Smail, I., Best, P. N., et al. 2013, MNRAS, 428, 1128 * Sobral et al. (2009) Sobral, D., Best, P. N., Geach, J. E., et al. 2009, MNRAS, 398, 75 * Sobral et al. (2015) Sobral, D., Matthee, J., Best, P. N., et al. 2015, MNRAS, 451, 2303 * Spergel et al. (2015) Spergel, D., Gehrels, N., Baltay, C., et al. 2015, arXiv e-prints, arXiv:1503.03757 * Steidel et al. (2004) Steidel, C. C., Shapley, A. E., Pettini, M., et al. 2004, ApJ, 604, 534 * Storey & Zeippen (2000) Storey, P. J., & Zeippen, C. J. 2000, MNRAS, 312, 813 * Sullivan et al. (2000) Sullivan, M., Treyer, M. A., Ellis, R. S., et al. 2000, MNRAS, 312, 442 * Tadaki et al. (2011) Tadaki, K.-I., Kodama, T., Koyama, Y., et al. 2011, PASJ, 63, 437 * Topping et al. (2021) Topping, M. W., Shapley, A. E., Sanders, R. L., et al. 2021, MNRAS, 506, 1237 * Trenti & Stiavelli (2008) Trenti, M., & Stiavelli, M. 2008, ApJ, 676, 767 * Tresse & Maddox (1998) Tresse, L., & Maddox, S. J. 1998, ApJ, 495, 691 * Treyer et al. (2005) Treyer, M., Wyder, T. K., Schiminovich, D., et al. 2005, ApJ, 619, L19 * Vavrek et al. (2016) Vavrek, R. D., Laureijs, R. J., Lorenzo Alvarez, J., et al. 2016, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 9911, Modeling, Systems Engineering, and Project Management for Astronomy VI, ed. G. Z. Angeli & P. Dierickx, 991105 * Virtanen et al. (2020) Virtanen, P., Gommers, R., Oliphant, T. E., et al. 2020, Nature Methods, 17, 261 * Wang et al. (2019) Wang, T., Schreiber, C., Elbaz, D., et al. 2019, Nature, 572, 211 * Williams et al. (2019) Williams, C. C., Labbe, I., Spilker, J., et al. 2019, ApJ, 884, 154 * Wyder et al. (2005) Wyder, T. K., Treyer, M. A., Milliard, B., et al. 2005, ApJ, 619, L15 * Zeimann et al. (2014) Zeimann, G. R., Ciardullo, R., Gebhardt, H., et al. 2014, ApJ, 790, 113
# Supercurrent-induced Anomalous Thermal Hall Effect as a New Probe to Superconducting Gap Anisotropy Xiaodong Hu Department of Physics, Boston College, Chestnut Hill, Massachusetts 02467, USA Jung Hoon Han Department of Physics, Sungkyunkwan University, Suwon 16419, Korea Ying Ran Department of Physics, Boston College, Chestnut Hill, Massachusetts 02467, USA ###### Abstract Two-dimensional superconductors have been realized in various atomically thin films such as the twisted bilayer graphene, some of which are anticipated to involve unconventional pairing mechanism. Due to their low dimensionality, experimental probes of the exact nature of superconductivity in these systems have been limited. We propose, by applying a _vertical_ supercurrent to a bilayer superconductor where the mirror symmetry is naturally broken by the twisting, there will be anomalous thermal Hall effect induced by the supercurrent that can serve as a sharp probe for the _in-plane_ anisotropy of the superconducting gap function. This effect occurs in the _absence_ of an external magnetic field and spontaneous breaking of the time-reversal symmetry in the ground state. We derive explicit formulas for the induced thermal Hall conductivity and show them to be significant in the examples of twisted cuprates and twisted FeSe where monolayer superconductivity have already been observed. Though technical challenges still exist, we propose this to be a generic probe of the gap anisotropy in a twisted bilayer superconductor. Introduction.— The discovery of the correlated insulating phases and superconductivity in twisted bilayer graphene (TBG) Cao _et al._ (2018a, b) has spurred intense interest in twisted two-dimensional (2D) heterostructures, leading to the notion of _twistronics_ as a new form of electronic device. To understand the (possibly unconventional) superconductivity realized in these systems, intense experimental investigations have been on-going Oh _et al._ (2021); Kim _et al._ (2022); Lake _et al._ (2022). While the majority of transport studies on twisted bilayers focuses on electrical transport at the moment, thermal transport has long been established as a powerful and complementary tool for investigating the nature of elementary excitations, particularly in superconductors where ordinary electric transport measurement is ineffective Krishana _et al._ (1997); Chiao _et al._ (2000); Sutherland _et al._ (2003); Durst _et al._ (2003); Zhang _et al._ (2001); Cvetkovic and Vafek (2015). More recently, quantized thermal Hall conductivity at low temperature became a signature of the topologically ordered ground states in correlated materials Banerjee _et al._ (2018); Yokoi _et al._ (2021). In this work, we show that the thermal Hall response can be a sensitive probe of the gap anisotropy in twisted bilayer superconductors, or TBS for short. The gap function $\varDelta_{\bm{k}}$ of a superconductor has immense implications for the underlying pairing mechanism and the quasiparticle transport. It may be deduced in the angle-resolved photoemission spectroscopy (ARPES) Shen _et al._ (1993); Ding _et al._ (1996); Zhang _et al._ (2016), or through the quasiparticle interference (QPI) imaging in scanning tunneling spectroscopy Hanaguri _et al._ (2007); Sprau _et al._ (2017). Resolving the gap anisotropy in ARPES becomes challenging though for low temperature superconductors where $\varDelta_{\bm{k}}$ is smaller than the experimental resolution. For twisted heterostructures, resolving the momentum space structure within a small moiré Brillouin zone (BZ) necessitates the QPI imaging over a formidably large area in the real space. As the demand to resolve the gap structure in TBS grows, limitations of existing experimental probes seem to loom larger. A natural question to ask, at this stage, is whether it is possible to invent a new probe of the gap anisotropy for very small $\varDelta_{\bm{k}}$. Here we propose a supercurrent-induced anomalous thermal Hall effect (SATHE) as one possible way to directly probe the gap anisotropy in TBS. Fig 1.(a) shows the schematic setup for SATHE. The TBS may be a vertical Josephson junction (JJ) formed by stacking two atomically thin superconducting films with a certain twist angle, or an intrinsic twisted bilayer superconductor as in TBG. SATHE is a nonlinear response of heat, created by simultaneously applying a _vertical_ supercurrent $\bm{J}_{S}$ and an in-plane temperature gradient, with the resulting transverse in-plane flow of heat. In contrast to the conventional thermal Hall effect (THE) which occurs when the ground state breaks the time-reversal symmetry (TRS), SATHE can occur for ground states that preserve the TRS. No external magnetic field is required to observe SATHE. Figure 1: (a) SATHE for bilayer-SC. The vertical supercurrent $\bm{J}_{S}$, or equivalently, a pairing phase twist $\varphi$ between the two layers of, either a vertical JJ or an intrinsic bilayer superconductor, would induce an in-plane thermal Hall effect in a bilayer superconductor. In the former case, an insulating buffer layer may be present (not shown). (b) an illustration for the dimensionless function $\xi(x)$ in Eq.(2). Our proposal thus differs for most proposals of THE including, for instance, Ref. Can _et al._ (2021) where the JJ spontaneously breaks TRS in the absence of applied supercurrent. It bear resemblance to the Sodemann-Fu proposal for nonlinear electrical Hall effect Sodemann and Fu (2015); Ma _et al._ (2019), in that in both proposals the _unperturbed_ ground state preserves TRS. One can think of the nonlinear electrical Hall effect as arising from the first electric field driving an imbalance of the fermion distribution, and the second electric field is used to probe the Hall response. In our proposal, the supercurrent is employed to drive Berry curvature out of its equilibrium form, then a temperature gradient is applied to induce the thermal Hall response. In this sense, SATHE can be taken as a thermal analogue of the nonlinear Hall effect. When the supercurrent is not too large, one can linearize the current-phase relation $\bm{J}_{S}\propto\varphi$ and SAHTE becomes the perturbative change of the thermal Hall conductivity $\delta\kappa_{xy}$ proportional to the phase twist, $\delta\kappa_{xy}\simeq\varphi\cdot\chi_{\varphi xy}$, with $\chi_{\varphi xy}$ capturing the nonlinear response of the system. Similar to nonlinear Hall effect, the breaking of inversion symmetry is necessary to elicit the desired responses. The observation of SATHE additionally requires the breaking of in-plane and out-of-plane mirror symmetries, which are not conditions normally present in the family of nonlinear Hall effects, but are naturally satisfied in the TBS. The conditions for observing SATHE are therefore not any more stringent than those of other proposed nonlinear Hall effect, at least from the perspective of symmetry requirement. Main results.— For TBS with a multiband electronic structure, the pairing Hamiltonian is generally $\sum_{a,b}\varDelta_{\bm{k}}^{a,b}c^{\dagger}_{a,\bm{k}}c^{\dagger}_{b,-\bm{k}}$, with $a,b$ ranging over the bands. We firstly study the simple situation where only the _intraband pairing_ is present: $\varDelta_{\bm{k}}^{a,b}=0$ for $a\neq b$. Assuming that $|\varDelta_{\bm{k}}|$ is much smaller than the energy difference between bands in the normal state, we find that a small pairing phase difference $\varphi$ between the layers in the TBS induces a change of the Berry curvature and leads to the SATHE formula $\displaystyle\frac{\delta\kappa_{xy}^{\text{intra.}}}{T}=\varphi\cdot\frac{\chi_{\varphi xy}^{\text{intra.}}}{T}=\varphi\cdot\frac{k_{B}^{2}}{16\pi^{2}\hbar}\sum_{\text{FS}}\mathop{\mathrm{sign}}(v_{F})$ $\displaystyle\qquad\times\oint_{\text{FS}}\mathop{}\\!\mathrm{d}k_{\parallel}\,\xi\bigg{(}\frac{\varDelta_{\bm{k}}(T)}{k_{B}T}\bigg{)}\cdot\partial_{k_{\parallel}}[\langle u_{\bm{k}}|\hat{L}|u_{\bm{k}}\rangle].$ (1) Here $\chi_{\varphi xy}^{\text{intra.}}$ is the intraband transport coefficient for SATHE, $\xi(x)$ is the dimensionless function (see Fig.1 (b)) $\displaystyle\xi(x)\equiv\int_{|x|}^{\infty}\mathrm{d}x^{\prime}\sqrt{x^{\prime 2}-x^{2}}\frac{x^{\prime}}{1+\cosh(x^{\prime})},$ (2) and $|u_{\bm{k}}\rangle$ is the Bloch state at the Fermi level in the _normal_ state. The dimensionless Hermitian operator $\hat{L}$ is the generator of Doppler shift due to $\varphi$, and equals $\hat{L}=\mathop{\mathrm{diag}}\\{\mathbf{1},\mathbf{-1}\\}$ in the layer space for a vertical JJ but gets more complicated for an intrinsic bilayer superconductor (see below and Supplemental Material (SM) SM ). The loop integral is performed over each Fermi surface (FS) and $k_{\parallel}$ is the counterclockwise tangential momentum at the Fermi surface. The $\mathop{\mathrm{sign}}(v_{F})=\pm 1$ characterizes whether the FS is electron-like or hole-like (a single band may host multiple FS’s). Clearly if $\varDelta_{\bm{k}}$ is $\bm{k}$-independent, the loop integral over the FS reduces to a total derivative and vanishes identically. The $\hat{L}$ operator and Doppler shift.— The normal state tight-binding Hamiltonian for the bilayer can be written in a block form: $\displaystyle H_{0}(\bm{k})=\begin{pmatrix}H_{0}^{t}(\bm{k})&T_{\perp}(\bm{k})\\\ T^{\dagger}_{\perp}(\bm{k})&H_{0}^{b}(\bm{k})\end{pmatrix},$ (3) where $t/b$ labels the top/bottom layer, and $T_{\perp}$ represents the interlayer hopping. In the superconducting state, pairing terms are introduced. Introducing the pairing phase difference $\varphi$ due to the applied supercurrent is equivalent to performing a gauge transformation $H_{0}\rightarrow U(\varphi)H_{0}U^{\dagger}(\varphi)$ (Doppler shift) while keeping the pairing terms unchanged, and $\hat{L}$ is its generator: $U(\varphi)\equiv e^{-i\frac{\varphi}{4}\hat{L}}$. For a vertical JJ, $\hat{L}=\mathop{\mathrm{diag}}\\{\mathbf{1},\mathbf{-1}\\}$. For an intrinsic superconductor, $\hat{L}$ depends on the atomic orbital positions normal to the bilayer plane (see SM SM ). For a JJ, since $|u_{\bm{k}}\rangle$ is the eigenstate of $H_{0}(\bm{k})$, $\langle u_{\bm{k}}|\hat{L}|u_{\bm{k}}\rangle\in[1,-1]$ serve as the indicator of the layer-component of the state. If $T_{\perp}$ is absent, top/bottom layers decouple and $\langle u_{\bm{k}}|\hat{L}|u_{\bm{k}}\rangle=\pm 1$, leading to vanishing SATHE according to Eq.(1), consistent with physical intuitions. Derivation of Eq. (1).— Here we show how to derive the SATHE formula of Eq. (1) while leaving the more technical parts to SM . The thermal Hall conductance in a 2D superconductor is related to the superconducting Berry curvature Qin _et al._ (2011); Sumiyoshi and Fujimoto (2013) as $\displaystyle\frac{\kappa_{xy}}{T}=-\frac{k_{B}^{2}}{2\hbar}\int_{-\infty}^{\infty}dE\frac{E^{2}}{(k_{B}T)^{2}}\bm{\sigma}(E)f^{\prime}(E),$ (4) where $f(E)$ is the Fermi-Dirac function, and $\displaystyle\bm{\sigma}(E)\equiv-\sum_{\mathbf{a}}\int_{E_{\bm{k}}<E}\frac{d^{2}k}{(2\pi)^{2}}\bm{\Omega}^{\mathbf{a}}_{\bm{k}}.$ (5) (Bold fonts are used for quantities related to the superconducting BdG state, to be distinguished from the normal state ones.) Here $\bm{\Omega}^{\mathbf{a}}_{\bm{k}}=-2\mathop{\mathrm{Im}}\langle\partial_{k_{x}}\mathbf{u}^{\mathbf{a}}_{\bm{k}}|\partial_{k_{y}}\mathbf{u}^{\mathbf{a}}_{\bm{k}}\rangle$ is the superconducting Berry curvature while $\mathbf{a}$ labels bands of the Bogoliubov–de Gennes (BdG) Hamiltonian $\displaystyle\mathbf{H}(\bm{k})=\left(\begin{array}[]{cc}H_{0}(\bm{k})&\bm{\Delta}_{\bm{k}}\\\ \bm{\Delta}_{\bm{k}}&-H_{0}(\bm{k})\end{array}\right).$ (8) The BdG Hamiltonian (8) is written in the Nambu basis $\\{\psi_{\bm{k}},\Lambda^{\dagger}\psi_{-\bm{k}}^{\dagger}\\}$ where $\psi_{\bm{k}}$ is a collection of fermion operators in the band basis, and $\bm{\Delta}_{\bm{k}}^{\dagger}=\bm{\Delta}_{\bm{k}}$. The time reversal transformation works on $\psi_{\bm{k}}$ as $\psi_{\bm{k}}\rightarrow\Lambda K\psi_{\bm{k}}$, where $K$ is complex conjugation and $\Lambda$ is some unitary operation. The energy eigenvalues of the BdG Hamiltonian are ordered into pairs $\pm\mathbf{E}_{1,\bm{k}},\pm\mathbf{E}_{2,\bm{k}},\cdots$ with $\mathbf{E}_{1,\bm{k}}<\mathbf{E}_{2,\bm{k}}<\cdots$ and $\mathbf{a}=\pm 1,\pm 2,\cdots$ are used to label the bands such that $\mathbf{E}_{-\mathbf{a},\bm{k}}=-\mathbf{E}_{\mathbf{a},\bm{k}}$. Assuming only the intraband pairing is present, we have $\mathbf{E}_{|\mathbf{a}|,\bm{k}}=(\epsilon_{|\mathbf{a}|,\bm{k}}^{2}+\varDelta_{|\mathbf{a}|,\bm{k}}^{2})^{1/2}$, where $\epsilon_{a,\bm{k}}$ is the normal state band energy ($a=1,2,...$). We assume that only the $a=1$ band crosses the Fermi level, and all other bands lie strictly above or below it. This allows us to define the interband energy scale as $t\equiv\min_{b}\min_{\bm{k}}|\epsilon_{b,\bm{k}}|$ ($b\neq 1$). The gap function of the first band is defined as $\varDelta_{\bm{k}}\equiv\varDelta_{1,\bm{k}}$. We will consider the weak- pairing limit $|\varDelta_{\bm{k}}|\ll t$, so only the lowest-energy states with $\mathbf{a}=\pm 1$ need to be included to the leading-order $\varDelta_{\bm{k}}/t$ expansion of the Berry curvature $\displaystyle\bm{\Omega}^{\mathbf{a}}_{\bm{k}}\doteq-\mathop{\mathrm{Im}}\left\\{\frac{\mathop{\mathrm{Tr}}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{(\mathbf{E}_{\mathbf{a}}-\mathbf{E}_{-\mathbf{a}})^{2}}-(x\leftrightarrow y)\right\\},$ (9) where $\mathbf{P}_{\mathbf{a}}\equiv|\mathbf{u}^{\mathbf{a}}_{\bm{k}}\rangle\langle{\mathbf{u}}^{\mathbf{a}}_{\bm{k}}|$ is the projector in the Nambu space. After a small phase twist $\varphi$ is turned on, $\mathbf{H},\mathbf{E}_{\pm\mathbf{a}},\mathbf{P}_{\pm\mathbf{a}}$ appearing in Eq. (9) all receive some corrections, which propagate through the THE formulas in Eqs. (4) and (5). However, as shown in SM , only the change in $\mathbf{P}_{\pm\mathbf{a}}$ is important when the unperturbed ground state obeys TRS. In the end, the intraband pairing contribution gives $\displaystyle\delta\bm{\Omega}^{\mathbf{a},\text{intra.}}_{\bm{k}}$ $\displaystyle\doteq\varphi\sum_{\mathbf{b}\neq\pm\mathbf{a}}4\mbox{Im\,Tr}[(\partial_{k_{x}}\mathbf{P}_{\mathbf{a}})(\partial_{k_{y}}\mathbf{P}_{-\mathbf{a}})\mathbf{P}_{\mathbf{b}}(\partial_{\varphi}\mathbf{P}_{\mathbf{a}})\mathbf{P}_{\mathbf{a}}]$ $\displaystyle\qquad-(x\leftrightarrow y),$ (10) To the leading order of $\varDelta_{\bm{k}}/t$ expansion and focusing on the $\mathbf{a}=\mathbf{1}$ band, we find $\delta\bm{\Omega}^{\mathbf{1},\text{intra.}}_{\bm{k}}=-\varphi\cdot\frac{\varDelta_{\bm{k}}^{2}}{4\mathbf{E}_{\mathbf{1}}^{3}}\big{(}v_{x}\partial_{k_{y}}-v_{y}\partial_{k_{x}}\big{)}\langle u^{1}_{\bm{k}}|\hat{L}|u^{1}_{\bm{k}}\rangle,$ (11) with $v_{x,y}\equiv\partial_{k_{x,y}}\epsilon_{1,\bm{k}}$ the normal state Fermi velocity. The change in Berry curvature is now fully described with normal state wave functions, and exhibits high concentration near the Fermi surface due to $\mathbf{E}_{\mathbf{1}}^{3}$ in the denominator. Identifying $\hat{L}|u_{\bm{k}}^{1}\rangle=4i\partial_{\varphi}|u_{\bm{k}}^{1}\rangle$, the derivative $\partial_{k_{y}}\langle u_{\bm{k}}^{1}|\hat{L}|u_{\bm{k}}^{1}\rangle=4\Omega_{y,\varphi}$ becomes the _$\varphi$ -twist Berry curvature_, with one component along the momentum direction and the other along the phase twist $\varphi$. Equation (11) thus demonstrates how the change of Berry curvature in the superconducting state is intricately composed of the gap function, and a mixed Berry curvature in the momentum-phase space. Plugging Eq. (11) into Eq. (4) and after some efforts, the main result of our paper Eq. (1) is established. Interband pairing and nodal superconductivity.— When interband pairing is present, the calculation becomes more sophisticated but the final result turns out to be simple. Introducing the normal state projector $P_{c}\equiv|u^{c}_{\bm{k}}\rangle\langle u^{c}_{\bm{k}}|$, the interband pairing gives $\bm{\Delta}_{\bm{k}}^{\text{inter.}}=\sum_{b\neq c}P_{b}\bm{\Delta}_{\bm{k}}P_{c}$, which can be eliminated from the BdG Hamiltonian in Eq. (8) by a small unitary rotation $e^{i\mathcal{S}\otimes\bm{\tau}_{2}}\mathbf{H}(\bm{k})e^{-i\mathcal{S}\otimes\bm{\tau}_{2}}\doteq H_{0}(\bm{k})\otimes\bm{\tau}_{3}+\bm{\Delta}_{\bm{k}}^{\text{intra.}}\otimes\bm{\tau}_{1}.$ (12) Here $\bm{\tau}$ are Pauli matrices in the Nambu space, and $\mathcal{S}\equiv\sum_{a\neq b}P_{a}\bm{\Delta}_{\bm{k}}P_{b}/(\epsilon_{a,\bm{k}}-\epsilon_{b,\bm{k}})$ can be viewed small simply because $\bm{\Delta}_{\bm{k}}/(\epsilon_{a,\bm{k}}-\epsilon_{b,\bm{k}})\sim\varDelta_{\bm{k}}/t$ is small in the weak-pairing limit. Based on this useful property, all previous perturbative analysis for the intraband pairing can be extended to interband pairing as well, with only one modification replacing the projector $\mathbf{P}_{\mathbf{a}}$ by $\widetilde{\mathbf{P}}_{\mathbf{a}}=e^{-i\mathcal{S}\otimes\bm{\tau}_{2}}\mathbf{P}_{\mathbf{a}}e^{i\mathcal{S}\otimes\bm{\tau}_{2}}$. Collecting all $\varphi$-linear terms contributing to the change of the Berry curvature in Eq. (9) seems to be more complicated, but to the leading order of $\varDelta_{\bm{k}}/t$ the interband contribution can be neatly arranged as SM $\displaystyle\delta\bm{\Omega}^{\mathbf{1},\text{inter.}}_{\bm{k}}=\varphi\cdot\frac{-{\mathbf{d}}_{\bm{k}}\cdot(\partial_{k_{x}}{\mathbf{d}}_{\bm{k}}\times\partial_{k_{y}}{\mathbf{d}}_{\bm{k}})}{2\mathbf{E}^{3}_{\mathbf{1}}},$ (13) $\displaystyle{\mathbf{d}}_{\bm{k}}\equiv(\varDelta_{\bm{k}},G_{\bm{k}},\epsilon_{1,\bm{k}}),\quad G_{\bm{k}}\equiv\frac{-1}{2}\mbox{Re}[\langle u^{1}_{\bm{k}}|\bm{\Delta}^{\text{inter.}}_{\bm{k}}\hat{L}|u^{1}_{\bm{k}}\rangle].$ Equation (13) is clearly reminiscent of the Berry curvature of the effective two-band model $\mathbf{H}_{\text{eff}}=\varphi\cdot\partial_{\varphi}\epsilon_{1,\bm{k}}\bm{\tau}_{0}+\epsilon_{1,\bm{k}}\bm{\tau}_{3}+\varDelta_{\bm{k}}\bm{\tau}_{1}+\varphi\cdot G_{\bm{k}}\bm{\tau}_{2}.$ (14) within the perturbative regime. Indeed, we show that $\mathbf{H}_{\text{eff}}$ is exactly the low-energy effective Hamiltonian of the TBS itself with insertion of phase twist SM $\mathbf{H}_{\text{eff}}^{\mathbf{a}}=\widetilde{\mathbf{P}}_{\mathbf{a}}\mathbf{H}[\varphi]\widetilde{\mathbf{P}}_{\mathbf{a}}\doteq\widetilde{\mathbf{P}}_{\mathbf{a}}\mathbf{H}\widetilde{\mathbf{P}}_{\mathbf{a}}+\varphi\cdot\widetilde{\mathbf{P}}_{\mathbf{a}}(\partial_{\varphi}\mathbf{H})\widetilde{\mathbf{P}}_{\mathbf{a}}$, so could serve as a faithful model to compute $\delta\mathbf{\Omega}_{\bm{k}}^{\mathbf{1},\text{inter.}}$. In calculating $\delta\kappa_{xy}$ in Eq. (4) one should add the two Berry curvature contributions $\delta\bm{\Omega}^{\mathbf{1}}_{\bm{k}}=\delta\bm{\Omega}^{\mathbf{1},\text{intra.}}_{\bm{k}}+\delta\bm{\Omega}^{\mathbf{1},\text{inter.}}_{\bm{k}}$ in Eq. (5). Among the two, the intraband Berry curvature part can be converted to the Fermi surface integral form in Eq. (1) but not the interband part. For calculating the latter contribution we can directly use Eqs. (4)-(5) with $\delta\bm{\Omega}^{\mathbf{1},\text{inter.}}_{\bm{k}}$ given in Eq. (13). Although in principle the interband contribution for SATHE cannot be neglected, there are many physical situations in which the interband pairing and thus the interband contribution to THE does become small. For instance, if each monolayer of TBS itself is superconducting as in a vertical JJ, $\bm{\Delta}_{\bm{k}}$ in Eq. (8) is diagonal in the monolayer band basis. For small twist angle $\theta\ll 1$, a unitary transformation to the band basis also induces interband pairing components in $\bm{\Delta}_{\bm{k}}$, which are proportional to $\theta$ and can be ignored. If the system is an intrinsic TBS like TBG, the intra- and inter-band pairings should be determined self- consistently. For example, for boson-mediated superconductivity, the multiband Eliashberg formulation may be applied, which is characterized by the electron- boson coupling matrix $[\alpha^{2}(\omega)F(\omega)]_{ab}$, where $a,b$ label bands. Either in the regime that intraband coupling is dominant $[\alpha^{2}(\omega)F(\omega)]_{a=b}\gg[\alpha^{2}(\omega)F(\omega)]_{a\neq b}$, or in the regime that the intraband coupling is dominant $[\alpha^{2}(\omega)F(\omega)]_{a=b}\ll[\alpha^{2}(\omega)F(\omega)]_{a\neq b}$, it is easy to show that only the intraband pairing is significant Dolgov _et al._ (2009). However, there is one particular scenario in which the interband contributions may be dominant, and that is when the superconductivity in the TBS is nodal. In this case, the node would develop a mass gap $m_{\bm{k}}=\varphi\cdot G_{\bm{k}}$ right at the nodal point upon the application of a phase twist $\varphi$, and a Chern number transfer $\Delta C=\pm\frac{1}{2}$ between the low-energy BdG bands occurs per node. When the net transferred Chern numbers $n$ from all the nodes is nonzero, the system becomes a chiral topological superconductor with quantized THE $\delta\kappa_{xy}^{\text{inter.}}\equiv\varphi\cdot\chi_{\varphi xy}^{\text{inter.}}=ng_{0}$ in the low-temperature limit $k_{B}T\ll|m|$, where $\chi_{\varphi xy}^{\text{inter.}}$ is the interband transport coefficient for SATHE, and $g_{0}\equiv\pi k_{B}^{2}T/(6\hbar)$ is the quantum of thermal conductance. In our perturbative treatment, $\chi_{\varphi xy}^{\text{inter.}}$ diverges in the low-$T$ regime. This supercurrent-driven topological superconductivity has been discussed in the context of twisted cuprate bilayers Volkov _et al._ (2023a, b); Can _et al._ (2021); Song _et al._ (2022) and twisted NbSe2 heterostructures Hu and Ran (2022). Applications to FeSe and cuprates.— To demonstrate the possibility of observing SATHE in real materials we consider two examples: vertical JJ’s formed by twisted nodeless FeSe, and the nodal cuprate superconducting films. The $\varDelta_{\bm{k}}$ in both of these 2D superconductors has been well characterized by ARPES due to the large energy scale of $\varDelta_{\bm{k}}$ and high transition temperatures. Note that due to the dimensionless nature of the SATHE response, the supercurrent-induced $\kappa_{xy}$ only depends on the ratio $\varDelta_{\bm{k}}/k_{B}T$ as in Eq. (1) and the SATHE response of low-$T_{c}$ superconductors after appropriate rescaling must be similar to the examples we consider now. Monolayer FeSe is reported to host a significant nodeless gap anisotropy of the form $\varDelta_{\bm{k}}=\varDelta_{0}+\varDelta_{2}\cos 2\theta_{\bm{k}}+\varDelta_{4}\cos 4\theta_{\bm{k}}$, where $\varDelta_{0}=9.9$meV, $\varDelta_{2}=-1.4$meV, and $\varDelta_{4}=1.2$meV Zhang _et al._ (2016). The $C_{4}$-rotation-related elliptic $M$-pockets positioned at $(\pm\frac{\pi}{2},\pm\frac{\pi}{2})$ within the two-iron BZ Brouet _et al._ (2012); Borisenko _et al._ (2016); Yi _et al._ (2015) can be described by the $\bm{k}\cdot\bm{p}$ expansion within $\mathrm{Fe}$’s $\\{3d_{xz},3d_{yz}\\}$ orbitals Agterberg _et al._ (2017); Ran _et al._ (2009) as $H_{M}=\left(\frac{1}{2m}(k_{x}^{2}+k_{y}^{2})-\mu\right)+ak_{x}k_{y}\tau_{z}$, where $\tau_{z}$ is a Pauli matrix in the orbital space, $\mu=0.08$eV, $1/2m=1.4$eV$\cdot\text{\AA}^{2}$ and $a=0.6$eV$\cdot\text{\AA}^{2}$. Since all Fermi pockets are near the zone boundary, moiré zone folding effect comes into play, and Bistritzer-MacDonald model Bistritzer and MacDonald (2011) is used to construct the normal state Hamiltonian. For example, to the lowest truncation of the moiré BZ, two hopping proccesses corresponding to $\delta\bm{q}^{t,b}=\pm 2K_{M}\sin\frac{\theta}{2}\hat{k}_{y}$ are included for all four of the $M$-pockets (see Fig.2 in SM SM ). We set $T_{\perp}^{d_{xz}^{t}\text{-}d_{xz}^{b}}(\delta\bm{q}^{t,b})=T_{\perp}^{d_{yz}^{t}\text{-}d_{yz}^{b}}(\delta\bm{q}^{t,b})\simeq T_{\perp}=15$meV and ignore all interorbital hoppings. After turning on such simple $T_{\perp}$ with a twist angle $\theta=11.5^{\circ}$, the Fermi surfaces reconstruct within the moiré BZ (see Fig. 2 (a1)). We secondly consider a twisted bilayer of cuprates 111Note that the currently available cuprate van der Waals materials is Bi2212, which is a bilayer of Cu-O planes. Consequently, twisted _double_ bilayer is more experimental relevant at present Zhu _et al._ (2021); Zhao _et al._ (2021). Our twisted bilayer model can viewed as a minimal illustration on SATHE for twisted nodal superconductors. with a $d$-wave gap anisotropy $\varDelta_{\bm{k}}=\varDelta_{N}\cos 2\theta_{\bm{k}}$ Kanigel _et al._ (2007); Kondo _et al._ (2015) and the reported relation $8.5k_{B}T_{c}=2\varDelta_{N}$ Kendziora _et al._ (1996); Anzai _et al._ (2013). The normal state bilayer Hamiltonian is constructed by first taking the tight-binding model in Ref. Eschrig and Norman (2003) as the monolayer Hamiltonian, and then obtaining the interlayer tunneling $T_{\perp}(\bm{k})=t_{z}\big{(}\frac{1}{4}(\cos k_{x}^{t}-\cos k_{y}^{t})(\cos k_{y}^{b}-\cos k_{y}^{b})+a_{0}\big{)}$ from the detailed orbital analysis in Refs. Song _et al._ (2022); Markiewicz _et al._ (2005). The constant $a_{0}=0.4$ is determined from DFT simulations Markiewicz _et al._ (2005). We set (i) $t_{z}=0.025t_{0}$ ($t_{0}$=leading hopping strength within the $\mathrm{CuO}_{2}$ plane) with $\theta=0.6^{\circ}$ (chiral topological superconductor) and (ii) $t_{z}=0.01t_{0}$ with $\theta=17.2^{\circ}$ (topologically trivial superconductor) and plot the corresponding Fermi surfaces in Fig. 2 (b1) and (c1), respectively Volkov _et al._ (2023a, b). Here we neglect the moiré zone folding effect for several reasons. First, based on the two-center approximation Bistritzer and MacDonald (2011), far away from the zone boundary the moiré zone folding effect may be neglected. Second, the moiré zone folding effect for the Fermi surfaces near the zone boundary does not modify the Fermi surface topology, leaving the results qualitatively unchanged. The interband and intraband contributions to SATHE for twisted bilayers of FeSe and cuprates are plotted in Fig. 2 (a2)-(c2). In our linearized scheme, $\kappa_{xy}$ and $\chi_{\varphi xy}$ are related simply by $\kappa_{xy}=\varphi\cdot\chi_{\varphi xy}$. We plot $\kappa_{xy}$ rather than $\chi_{\varphi xy}$ because the latter apparently diverges as $\varphi\rightarrow 0$ while $\kappa_{xy}^{\text{inter.}}$ remains finite in a supercurrent-driven topological superconductor as shown in Fig. 2 (b2). For illustrative purposes, phase twist $\varphi=1$ rad is chosen for computation 222Josephson phase twist $\varphi=1$ rad is taken, primarily to exhibit the strength of the transport coefficient $\chi_{\varphi xy}^{\text{intra.}}$ from the intraband contribution along with the quantized behavior of $\kappa_{xy}$ arising from the interband one. The former component contributes to the thermal Hall conductivity simply as $\kappa_{xy}^{\text{intra.}}=\varphi\cdot\chi_{\varphi xy}^{\text{intra.}}$, as along as we remain within the _linear regime_ of the sine current-phase relation, which is satisfied simply because $\sin 1.0\doteq 0.84\approx 1.0$. Therefore, the choice $\varphi=1$ rad can still be considered small in our perturbative treatment, and with such choice the plot of $\kappa_{xy}$ also directly reflects the strength of the SATHE transport coefficient $\chi_{\varphi xy}$ that we are concerned with.. The interband contribution is evaluted by using the integral formula of Eq. (1). For the interband contribution we rely on Eqs. (4)-(5) with the Berry curvature obtained in Eq. (13). As shown in Fig. 2, the intraband contribution plays a dominant role for SATHE in both twisted FeSe (a2) and the topological trivial case of twisted cuprates (c2), while interband contribution dominates in topological nontrivial case of twisted cuprates (b2), especially in the low-temperature regime. Note that the $\kappa_{xy}$ in the proposed SATHE can reach $\sim 10^{-1}$ of the thermal conductance quantum when the gap anisotropy is significant, e.g., in the FeSe example. We conclude that SATHE can be a sizable effect, detectable in the foreseeable future. Figure 2: (a): JJ formed by twisted bilayer FeSe. (b) and (c): the topological non-trivial and trivial phases induced by the supercurrent in JJ formed by twisted bilayer cuprates. The twisted Fermi surfaces are shown on the left panels. For FeSe the moiré zone folding effect plays an important role on the topology of reconstructed Fermi surfaces. The color scheme on each Fermi surface represents the strength of $\langle u_{\bm{k}}|\hat{L}|u_{\bm{k}}\rangle\in[-1,1]$, serving as the indicator of the layer-component of states. For all three cases, $\kappa_{xy}$ is computed in unit of the thermal conductance quantum $g_{0}$ as a function of temperature, and a simple mean-field temperature-dependence of $\varDelta_{\bm{k}}(T)=\varDelta_{\bm{k}}(T=0)\sqrt{1-T/T_{c}}$ is implemented. For the topologically nontrivial case (b), gaps $\sim 0.2$meV are generated around superconducting nodes, leading to a Chern number $C=8$ topological superconductivity, while for the topological trivial case (c) the net Chern number is zero. Conclusion and Discussion.— We propose the supercurrent-induced anomalous thermal Hall effect, SATHE, as a new probe to the in-plane gap anisotropy of bilayer superconductors. Different from the conventional thermal Hall effect, the ground state preserve the time reversal symmetry. A pair of probes — a vertically applied supercurrent and a horizontal temperature gradient — is applied to induce the in-plane nonlinear thermal Hall response. Being a thermal response, it works on 2D superconductors where usual electrical probes fail. Since no external magnetic field needs to be applied, SATHE avoids the complications of vortices in the mixed state and probes purely the quasiparticle dynamics in superconductors. Within the BdG framework we showed that SATHE is sensitive to the in-plane gap anisotropy in the twisted bilayer superconductor, and could be large enough to serve as a new experimental probe of gap structure for atomically thin superconducting 2D crystals including twisted bilayer graphene systems, for which experimental probes have been limited due to the low dimensionality. Motivated by the discovery of novel quantum states in van der Waals materials, the experimental community has recently made significant progress towards measuring thermal transports in low dimensional systems. For instance, graphene-based Johnson noise thermometry has been developed Fong and Schwab (2012); Talanov _et al._ (2021) and used to measure thermal transports in graphene, carbon nanotubes and $\alpha$-RuCl3 Fong _et al._ (2013); Waissman _et al._ (2022). We believe SATHE can eventually find its place as an effective probe of 2D twisted materials, in particular for low $T_{c}$ superconductors where SATHE can serve as a sensitive measure of the gap anisotropy. Our results can be straightforwardly generalized to multi-layer superconductors, by replacing the operator $\hat{L}$ with the generator of the multi-layer Doppler shift. ###### Acknowledgements. Y.R. and X.H. acknowledge the support from National Science Foundation under Grant No. DMR-1712128. J.H.H. was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2023R1A2C1002644). ## References * Cao _et al._ (2018a) Y. Cao, V. Fatemi, S. Fang, K. Watanabe, T. Taniguchi, E. Kaxiras, and P. Jarillo-Herrero, Nature 556, 43 (2018a). * Cao _et al._ (2018b) Y. Cao, V. Fatemi, A. Demir, S. Fang, S. L. Tomarken, J. Y. Luo, J. D. Sanchez-Yamagishi, K. Watanabe, T. Taniguchi, E. Kaxiras, _et al._ , Nature 556, 80 (2018b). * Oh _et al._ (2021) M. Oh, K. P. Nuckolls, D. Wong, R. L. Lee, X. Liu, K. Watanabe, T. Taniguchi, and A. Yazdani, Nature 600, 240 (2021). * Kim _et al._ (2022) H. Kim, Y. Choi, C. Lewandowski, A. Thomson, Y. Zhang, R. Polski, K. Watanabe, T. Taniguchi, J. Alicea, and S. Nadj-Perge, Nature 606, 494 (2022). * Lake _et al._ (2022) E. Lake, A. S. Patri, and T. Senthil, arXiv preprint arXiv:2204.12579 (2022). * Krishana _et al._ (1997) K. Krishana, N. Ong, Q. Li, G. Gu, and N. Koshizuka, Science 277, 83 (1997). * Chiao _et al._ (2000) M. Chiao, R. Hill, C. Lupien, L. Taillefer, P. Lambert, R. Gagnon, and P. Fournier, Physical Review B 62, 3554 (2000). * Sutherland _et al._ (2003) M. Sutherland, D. Hawthorn, R. Hill, F. Ronning, S. Wakimoto, H. Zhang, C. Proust, E. Boaknin, C. Lupien, L. Taillefer, _et al._ , Physical Review B 67, 174520 (2003). * Durst _et al._ (2003) A. C. Durst, A. Vishwanath, and P. A. Lee, Physical Review Letters 90, 187002 (2003). * Zhang _et al._ (2001) Y. Zhang, N. Ong, P. Anderson, D. Bonn, R. Liang, and W. Hardy, Physical Review Letters 86, 890 (2001). * Cvetkovic and Vafek (2015) V. Cvetkovic and O. Vafek, Nature Communications 6, 1 (2015). * Banerjee _et al._ (2018) M. Banerjee, M. Heiblum, V. Umansky, D. E. Feldman, Y. Oreg, and A. Stern, Nature 559, 205 (2018). * Yokoi _et al._ (2021) T. Yokoi, S. Ma, Y. Kasahara, S. Kasahara, T. Shibauchi, N. Kurita, H. Tanaka, J. Nasu, Y. Motome, C. Hickey, _et al._ , Science 373, 568 (2021). * Shen _et al._ (1993) Z.-X. Shen, D. Dessau, B. Wells, D. King, W. Spicer, A. Arko, D. Marshall, L. Lombardo, A. Kapitulnik, P. Dickinson, _et al._ , Physical Review Letters 70, 1553 (1993). * Ding _et al._ (1996) H. Ding, M. Norman, J. Campuzano, M. Randeria, A. Bellman, T. Yokoya, T. Takahashi, T. Mochiku, and K. Kadowaki, Physical Review B 54, R9678 (1996). * Zhang _et al._ (2016) Y. Zhang, J. Lee, R. Moore, W. Li, M. Yi, M. Hashimoto, D. Lu, T. Devereaux, D.-H. Lee, and Z.-X. Shen, Physical Review Letters 117, 117001 (2016). * Hanaguri _et al._ (2007) T. Hanaguri, Y. Kohsaka, J. Davis, C. Lupien, I. Yamada, M. Azuma, M. Takano, K. Ohishi, M. Ono, and H. Takagi, Nature Physics 3, 865 (2007). * Sprau _et al._ (2017) P. O. Sprau, A. Kostin, A. Kreisel, A. E. Böhmer, V. Taufour, P. C. Canfield, S. Mukherjee, P. J. Hirschfeld, B. M. Andersen, and J. S. Davis, Science 357, 75 (2017). * Can _et al._ (2021) O. Can, T. Tummuru, R. P. Day, I. Elfimov, A. Damascelli, and M. Franz, Nature Physics 17, 519 (2021). * Sodemann and Fu (2015) I. Sodemann and L. Fu, Physical Review Letters 115, 216806 (2015). * Ma _et al._ (2019) Q. Ma, S.-Y. Xu, H. Shen, D. MacNeill, V. Fatemi, T.-R. Chang, A. M. Mier Valdivia, S. Wu, Z. Du, C.-H. Hsu, _et al._ , Nature 565, 337 (2019). * (22) See Supplemental Material at URL_to_be_inserted_by_publisher for detailed derivations on both intraband and interband contribution of SATHE. Discussion on the nodal origin of the interband contribution and the Fermi surface reconstruction of twisted-bilayer FeSe are also covered. * Qin _et al._ (2011) T. Qin, Q. Niu, and J. Shi, Physical Review Letters 107, 236601 (2011). * Sumiyoshi and Fujimoto (2013) H. Sumiyoshi and S. Fujimoto, Journal of the Physical Society of Japan 82, 023602 (2013). * Dolgov _et al._ (2009) O. V. Dolgov, I. I. Mazin, D. Parker, and A. A. Golubov, Physical Review B 79, 060502 (2009). * Volkov _et al._ (2023a) P. A. Volkov, J. H. Wilson, K. P. Lucht, and J. Pixley, Physical Review Letters 130, 186001 (2023a). * Volkov _et al._ (2023b) P. A. Volkov, J. H. Wilson, K. P. Lucht, and J. Pixley, Physical Review B 107, 174506 (2023b). * Song _et al._ (2022) X.-Y. Song, Y.-H. Zhang, and A. Vishwanath, Physical Review B 105, L201102 (2022). * Hu and Ran (2022) X. Hu and Y. Ran, Physical Review B 106, 125136 (2022). * Brouet _et al._ (2012) V. Brouet, M. F. Jensen, P.-H. Lin, A. Taleb-Ibrahimi, P. Le Fèvre, F. Bertran, C.-H. Lin, W. Ku, A. Forget, and D. Colson, Physical Review B 86, 075123 (2012). * Borisenko _et al._ (2016) S. Borisenko, D. Evtushinsky, Z.-H. Liu, I. Morozov, R. Kappenberger, S. Wurmehl, B. Büchner, A. Yaresko, T. Kim, M. Hoesch, _et al._ , Nature Physics 12, 311 (2016). * Yi _et al._ (2015) M. Yi, Z. Liu, Y. Zhang, R. Yu, J.-X. Zhu, J. Lee, R. Moore, F. Schmitt, W. Li, S. Riggs, _et al._ , Nature Communications 6, 1 (2015). * Agterberg _et al._ (2017) D. Agterberg, T. Shishidou, J. O’Halloran, P. Brydon, and M. Weinert, Physical Review Letters 119, 267001 (2017). * Ran _et al._ (2009) Y. Ran, F. Wang, H. Zhai, A. Vishwanath, and D.-H. Lee, Physical Review B 79, 014505 (2009). * Bistritzer and MacDonald (2011) R. Bistritzer and A. H. MacDonald, Proceedings of the National Academy of Sciences 108, 12233 (2011). * Note (1) Note that the currently available cuprate van der Waals materials is Bi2212, which is a bilayer of Cu-O planes. Consequently, twisted _double_ bilayer is more experimental relevant at present Zhu _et al._ (2021); Zhao _et al._ (2021). Our twisted bilayer model can viewed as a minimal illustration on SATHE for twisted nodal superconductors. * Kanigel _et al._ (2007) A. Kanigel, U. Chatterjee, M. Randeria, M. Norman, S. Souma, M. Shi, Z. Li, H. Raffy, and J. Campuzano, Physical Review Letters 99, 157001 (2007). * Kondo _et al._ (2015) T. Kondo, W. Malaeb, Y. Ishida, T. Sasagawa, H. Sakamoto, T. Takeuchi, T. Tohyama, and S. Shin, Nature Communications 6, 1 (2015). * Kendziora _et al._ (1996) C. Kendziora, R. Kelley, and M. Onellion, Physical Review Letters 77, 727 (1996). * Anzai _et al._ (2013) H. Anzai, A. Ino, M. Arita, H. Namatame, M. Taniguchi, M. Ishikado, K. Fujita, S. Ishida, and S. Uchida, Nature Communications 4, 1 (2013). * Eschrig and Norman (2003) M. Eschrig and M. Norman, Physical Review B 67, 144503 (2003). * Markiewicz _et al._ (2005) R. Markiewicz, S. Sahrakorpi, M. Lindroos, H. Lin, and A. Bansil, Physical Review B 72, 054519 (2005). * Note (2) Josephson phase twist $\varphi=1$ rad is taken, primarily to exhibit the strength of the transport coefficient $\chi_{\varphi xy}^{\text{intra.}}$ from the intraband contribution along with the quantized behavior of $\kappa_{xy}$ arising from the interband one. The former component contributes to the thermal Hall conductivity simply as $\kappa_{xy}^{\text{intra.}}=\varphi\cdot\chi_{\varphi xy}^{\text{intra.}}$, as along as we remain within the _linear regime_ of the sine current-phase relation, which is satisfied simply because $\sin 1.0\mathrel{\mathop{\kern 0.0pt=}\limits^{\textstyle.}}0.84\approx 1.0$. Therefore, the choice $\varphi=1$ rad can still be considered small in our perturbative treatment, and with such choice the plot of $\kappa_{xy}$ also directly reflects the strength of the SATHE transport coefficient $\chi_{\varphi xy}$ that we are concerned with. * Fong and Schwab (2012) K. C. Fong and K. Schwab, Physical Review X 2, 031006 (2012). * Talanov _et al._ (2021) A. V. Talanov, J. Waissman, T. Taniguchi, K. Watanabe, and P. Kim, Review of Scientific Instruments 92, 014904 (2021). * Fong _et al._ (2013) K. C. Fong, E. E. Wollman, H. Ravi, W. Chen, A. A. Clerk, M. Shaw, H. Leduc, and K. Schwab, Physical Review X 3, 041008 (2013). * Waissman _et al._ (2022) J. Waissman, L. E. Anderson, A. V. Talanov, Z. Yan, Y. J. Shin, D. H. Najafabadi, M. Rezaee, X. Feng, D. G. Nocera, T. Taniguchi, _et al._ , Nature Nanotechnology 17, 166 (2022). * Zhu _et al._ (2021) Y. Zhu, M. Liao, Q. Zhang, H.-Y. Xie, F. Meng, Y. Liu, Z. Bai, S. Ji, J. Zhang, K. Jiang, _et al._ , Physical Review X 11, 031011 (2021). * Zhao _et al._ (2021) S. Zhao, N. Poccia, X. Cui, P. A. Volkov, H. Yoo, R. Engelke, Y. Ronen, R. Zhong, G. Gu, S. Plugge, _et al._ , arXiv preprint arXiv:2108.13455 (2021). * Crowley and Fu (2022) P. J. Crowley and L. Fu, Physical Review B 106, 214526 (2022). Supplemental Material for “Supercurrent-induced Anomalous Thermal Hall Effect as a New Probe to Superconducting Gap Anisotropy” Xiaodong Hu, Jung Hun Han, and Ying Ran ## I The Doppler shift and the gauge transformation due to the supercurrent The Doppler shift effect due to supercurrent has been discussed in textbooks and carefully in a recent paper by Crowley and Fu Crowley and Fu (2022). Below we follow the notation in the Appendix B of Ref. Crowley and Fu (2022). To model a superconducting state in the presence of a finite supercurrent with supercurrent velocity $\mathbf{u}$, it is convenient to consider a moving reference frame $S^{\prime}$ which is moving at a velocity $\mathbf{u}$ relative to the lab frame $S$. In the moving frame $S^{\prime}$, there is no supercurrent and the superconducting order parameter is spatially uniform. Microscopically, the Galilean transformation between the frame-$S$ and frame-$S^{\prime}$ is implemented by the unitary: $\displaystyle U_{\mathbf{u}}$ $\displaystyle=e^{\frac{i}{\hbar}\mathbf{u}\cdot\mathbf{g}},$ $\displaystyle\mathbf{g}$ $\displaystyle=M\mathbf{R}-\mathbf{K}t,$ (S1) where $\mathbf{K}$ is the total momentum and $\mathbf{R}$ is the center of mass: $\displaystyle\mathbf{K}$ $\displaystyle=\int d\mathbf{r}\,c_{\mathbf{r},\sigma}^{\dagger}(-i\hbar\nabla)c_{\mathbf{r},\sigma},$ $\displaystyle\mathbf{R}$ $\displaystyle=\frac{1}{M}\int d\mathbf{r}\,c_{\mathbf{r},\sigma}^{\dagger}c_{\mathbf{r},\sigma}m\mathbf{r}.$ (S2) Here $M$ is the total mass, and $\sigma$ labels electron’s spin. Under Galilean transformation, in the real-space basis: $\displaystyle U^{\dagger}_{\mathbf{u}}c^{\dagger}_{\mathbf{r},\sigma}U_{\mathbf{u}}=c^{\dagger}_{\mathbf{r}-\mathbf{u}t,\sigma}e^{\frac{-i}{\hbar}m\mathbf{r}\cdot\mathbf{u}+\frac{i}{\hbar}mu^{2}t/2}.$ (S3) Namely, the electron state $c^{\dagger}_{\mathbf{r},\sigma}$ in frame-$S$ is transformed into the electron state $c^{\dagger}_{\mathbf{r}-\mathbf{u}t,\sigma}$ in frame-$S^{\prime}$. The _spatially uniform_ pairing field in frame-$S^{\prime}$ then can be represented in the real-space basis: $\displaystyle\hat{\Delta}_{S^{\prime}}=\int d\mathbf{r}_{1}d\mathbf{r}_{2}\Delta(\mathbf{r}_{1}-\mathbf{r}_{2})c^{\dagger}_{\mathbf{r}_{1}-\mathbf{u}t,\sigma_{1}}\epsilon_{\sigma_{1},\sigma_{2}}c^{\dagger}_{\mathbf{r}_{2}-\mathbf{u}t,\sigma_{2}},$ (S4) where we assumed spin-singlet pairing for simplicity, and $\epsilon_{\sigma_{1},\sigma_{2}}$ is the Levi-Civita symbol. Performing the inverse Galilean transformation, one finds the pairing field in frame-$S$ is: $\displaystyle\hat{\Delta}_{S}=U_{\mathbf{u}}\hat{\Delta}_{S^{\prime}}U^{\dagger}_{\mathbf{u}}=\int d\mathbf{r}_{1}d\mathbf{r}_{2}e^{\frac{i}{\hbar}m(\mathbf{r}_{1}+\mathbf{r}_{2})\cdot\mathbf{u}}\Delta(\mathbf{r}_{1}-\mathbf{r}_{2})c^{\dagger}_{\mathbf{r}_{1},\sigma_{1}}\epsilon_{\sigma_{1},\sigma_{2}}c^{\dagger}_{\mathbf{r}_{2},\sigma_{2}},$ (S5) One concludes that in the lab frame-$S$, the Cooper pair carries a nonzero center-of-mass momentum $2m\mathbf{u}$ due to the supercurrent. The mean-field Hamiltonian in the frame-$S$ is given by: $\displaystyle H^{MF}_{S}=H_{0}+\hat{\Delta}_{S},$ (S6) where $H_{0}$ is the normal state Hamiltonian. One may now perform a space- dependent and time-independent gauge transformation $\mathbf{U}$ (which is different from the Galilean transformation) to eliminate the spatial dependence in the pairing field $\hat{\Delta}_{S}$: $\displaystyle\mathbf{U}$ $\displaystyle=e^{\frac{-i}{\hbar}\int d\mathbf{r}\,c^{\dagger}_{\mathbf{r},\sigma}c_{\mathbf{r},\sigma}m\mathbf{r}\cdot\mathbf{u}},$ $\displaystyle\mathbf{U}c^{\dagger}_{\mathbf{r},\sigma}\mathbf{U}^{\dagger}$ $\displaystyle=e^{\frac{-i}{\hbar}m\mathbf{r}\cdot\mathbf{u}}c^{\dagger}_{\mathbf{r},\sigma}$ (S7) $H^{MF}_{S}$ after this unitary becomes: $\displaystyle\tilde{H}^{MF}_{S}\equiv\mathbf{U}H^{MF}_{S}\mathbf{U}^{\dagger}=\mathbf{U}H_{0}\mathbf{U}^{\dagger}+\int d\mathbf{r}_{1}d\mathbf{r}_{2}\Delta(\mathbf{r}_{1}-\mathbf{r}_{2})c^{\dagger}_{\mathbf{r}_{1},\sigma_{1}}\epsilon_{\sigma_{1},\sigma_{2}}c^{\dagger}_{\mathbf{r}_{2},\sigma_{2}},$ (S8) In $\tilde{H}^{MF}_{S}$, the pairing field restores the form as if supercurrent is absent, and the normal state Hamiltonian becomes Doppler- shifted due to the gauge transformation. If we denote the phase factor $e^{\frac{-i}{\hbar}m\mathbf{r}\cdot\mathbf{u}}\equiv e^{-i\varphi(\mathbf{r})/2}$, the supercurrent density is: $\displaystyle\mathbf{J}_{s}=-en_{s}\mathbf{u}=-e\frac{\hbar n_{s}}{2m}\nabla\varphi(\mathbf{r}),$ (S9) where we introduced the electron’s superfluid density $n_{s}$, and $\varphi(\mathbf{r})$ can be identified as the phase of the pairing order parameter, which recovers a well-known result. Generally speaking, the superfluid density may be spatially dependent $n_{s}(\mathbf{r})$. For a uniform supercurrent density, due to current conservation, this implies that the superfluid velocity $\mathbf{u}(\mathbf{r})$ would be spatially dependent. Namely, $\mathbf{u}(\mathbf{r})$, or $\nabla\varphi$, would be larger when $n_{s}(\mathbf{r})$ is smaller. In this situation, the unitary $\mathbf{U}$ should be modified as: $\displaystyle\mathbf{U}$ $\displaystyle=e^{\frac{-i}{\hbar}\int d\mathbf{r}\,c^{\dagger}_{\mathbf{r},\sigma}c_{\mathbf{r},\sigma}m\int_{\mathbf{0}}^{\mathbf{r}}d\mathbf{r}^{\prime}\cdot\mathbf{u}(\mathbf{r}^{\prime})},$ $\displaystyle\mathbf{U}c^{\dagger}_{\mathbf{r},\sigma}\mathbf{U}^{\dagger}$ $\displaystyle=e^{\frac{-i}{\hbar}m\int_{\mathbf{0}}^{\mathbf{r}}d\mathbf{r}^{\prime}\cdot\mathbf{u}(\mathbf{r}^{\prime})}c^{\dagger}_{\mathbf{r},\sigma}\equiv e^{-i\varphi(\mathbf{r})/2}c^{\dagger}_{\mathbf{r},\sigma},$ (S10) and one still has $\mathbf{u}(\mathbf{r})=\frac{\hbar}{2m}\nabla\varphi(\mathbf{r})$. The bottom line is that no matter the system is a bilayer Josephson junction or an instrinsic bilayer superconductor, a finite vertical supercurrent state is always modeled by the gauge transformation $c^{\dagger}_{\mathbf{r},\sigma}\rightarrow e^{-i\varphi(\mathbf{r})/2}c^{\dagger}_{\mathbf{r},\sigma}$ that transforms the normal state electronic structure $H_{0}\rightarrow\mathbf{U}H_{0}\mathbf{U}^{\dagger}$ while leaving the pairing field spatially uniform (as if the supercurrent is absent), as we have done in Eq.(4) in the main text. Here $\varphi(\mathbf{r})$ satisfies Eq.(S9), and is only depending on the vertical coordinate $z$: $\varphi(\mathbf{r})=\varphi(z)$. The difference between the two cases lies in the details of the profile of $\varphi(z)$. In the case of a bilayer Josephson junction, the superfluid density is concentrated in each monolayers, while between the two layers $n_{s}$ is small. Consequently, the gradient of the phase $\varphi(z)$ is essentially located between the two layers. In the case of an intrinsic bilayer superconductor, $\varphi(z)$ would be a more smooth function of $z$. We schematically plot the phase $\varphi(z)$ of the gauge transformation for the two cases in Fig.S1. Figure S1: The schematic plot of the phase $\varphi$ involved in the gauge transformation to describe a finite supercurrent state. The two curves represent the case of a Josephson junction (JJ) and an intrinsic bilayer superconductor(Intrinisic) respectively. We now consider the situation that $\varphi(z)$ interpolates the top layer $\delta\varphi/2$ to the bottom layer $-\delta\varphi/2$. The Doppler shift unitary is generated via $\hat{L}$: $\mathbf{U}=e^{-i\frac{\delta\varphi}{4}\hat{L}}$. * • In the case of a vertical JJ, $\hat{L}$ can written in the tight-binding basis, where the $\pm\mathbf{1}$ blocks correspond to the top/bottom layers: $\displaystyle\hat{L}_{\text{JJ}}=\hat{c}^{\dagger}\begin{pmatrix}\mathbf{1}&0\\\ 0&\mathbf{-1}\end{pmatrix}c.$ (S11) * • In the case of an intrinsic bilayer superconductor, assuming the superfluid density is uniform, $\hat{L}$ is a diagonal matrix whose diagonal matrix elements are given by $\frac{z_{\alpha}}{z_{0}}$, where $z_{\alpha}$ is the atomic coordinate of the orbital-$\alpha$ and $\pm z_{0}$ is the position of the top/botom layers. $\displaystyle\hat{L}_{\text{Intrinsic}}=\sum_{\alpha}c_{\alpha}^{\dagger}\frac{z_{\alpha}}{z_{0}}c_{\alpha}$ (S12) ## II Details on the Derivation of the Main Result ### II.1 Supercurrent-indcued Berry Curvature for Intraband Pairing Here we provide details of the derivation of the main result Eq.(1) in the main text. Because only $\mathbf{a}=\pm 1$ bands contribute to the low-energy Berry curvature, we will start with Eq.(8) in the main text. Denoting the $c$-th eigenstate of $H_{0}(\bm{k})$ as $|u^{c}_{\bm{k}}\rangle$: $H_{0}(\bm{k})|u^{c}_{\bm{k}}\rangle=\epsilon_{c,\bm{k}}|u^{c}_{\bm{k}}\rangle$, and introducing the normal state projector $P_{c}\equiv|u^{c}_{\bm{k}}\rangle\langle u^{c}_{\bm{k}}|$, we can define the BdG projector into the low energy Hilbert space: $\displaystyle\mathbf{P}\equiv P_{1}\otimes\bm{\tau}_{0},$ (S13) where $\bm{\tau}_{0}$ is the identity Pauli matrix in the particle-hole space. The low-energy BdG Hamiltonian is simply the two-by-two matrix, which perfectly decouples from the high energy Hilbert space if only intraband pairing is present (which we call _intraband pairing assumption_ below): $\displaystyle\mathbf{P}\mathbf{H}\mathbf{P}=P_{1}\otimes(\epsilon_{1,\bm{k}}\bm{\tau}_{3}+\varDelta_{\bm{k}}\bm{\tau}_{1}).$ (S14) Here $\varDelta_{\bm{k}}$ is chosen to be real due to the time-reversal symmetry. In addition, Projector $\mathbf{P}$ satisfies an important identity with any operator of the following form: $\displaystyle\mathbf{A}^{\vee}\equiv A\otimes\bm{\tau}_{0}\Rightarrow\mathbf{P}\mathbf{A}\mathbf{P}=\langle u^{1}_{\bm{k}}|A|u^{1}_{\bm{k}}\rangle\mathbf{P},$ (S15) which in turn indicates $\mathbf{P}_{-\mathbf{a}}\mathbf{A^{\vee}}\mathbf{P}_{\mathbf{a}}=0$. After the pairing phase twist $\delta\varphi$ are turned on, $\mathbf{H},\mathbf{E}_{\pm\mathbf{a}},\mathbf{P}_{\pm\mathbf{a}}$ all receive perturbations in Eq.(8) in the main text. $\delta\mathbf{H}$ and $\delta\mathbf{E}_{\pm\mathbf{a}}$ do not contribute because $\partial_{\varphi}\mathbf{H}$ and $\partial_{k_{x}}\partial_{\varphi}\mathbf{H}$ both have the form of $\mathbf{A}^{\vee}$ _under the intraband assumption_. For instance: $\displaystyle\partial_{\varphi}\mathbf{H}=\partial_{\varphi}H_{0}(\bm{k})\otimes\bm{\tau}_{0}.$ (S16) Thus we are left with the $\delta\mathbf{P}_{\pm\mathbf{a}}$ contribution only: $\displaystyle\frac{\delta\bm{\Omega}^{\mathbf{a}}}{\delta\varphi}\doteq\frac{-\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{H}\partial_{\varphi}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}^{2}}-(x\leftrightarrow y).$ (S17) Here we used the time-reversal symmetry combined with the particle-hole symmetry, which dictates that the $\delta\mathbf{P}_{-\mathbf{a}}$ contribution and $\delta\mathbf{P}_{\mathbf{a}}$ contribution are identical. After inserting $\mathbf{1}=\sum_{\mathbf{c}}\mathbf{P}_{\mathbf{c}}$ in front of $\partial_{\varphi}\mathbf{P}_{\mathbf{a}}$, and noting that only terms with $\mathbf{c}\neq\pm\mathbf{a}$ contributes _under the intraband pairing assumption_ , the above expression can be rewritten as: $\displaystyle\frac{\delta\bm{\Omega}^{\mathbf{a}}}{\delta\varphi}\doteq$ $\displaystyle\sum_{\mathbf{b}\neq\pm\mathbf{a}}\frac{\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{H}\mathbf{P}_{\mathbf{b}}\partial_{\varphi}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}^{2}(\mathbf{E}_{\mathbf{b}}-\mathbf{E}_{\mathbf{a}})}-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\sum_{\mathbf{b}\neq\pm\mathbf{a}}\frac{\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{P}_{\mathbf{b}}\partial_{\varphi}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}^{2}(\mathbf{E}_{\mathbf{b}}-\mathbf{E}_{\mathbf{a}})}(\mathbf{E}_{\mathbf{b}}+\mathbf{E}_{\mathbf{a}})-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\sum_{\mathbf{b}\neq\pm\mathbf{a}}\frac{2\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{P}_{\mathbf{b}}\partial_{\varphi}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}(\mathbf{E}_{\mathbf{b}}-\mathbf{E}_{\mathbf{a}})}-(x\leftrightarrow y)$ $\displaystyle\qquad+\sum_{\mathbf{b}\neq\pm\mathbf{a}}\frac{\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{P}_{\mathbf{b}}\partial_{\varphi}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}^{2}}-(x\leftrightarrow y).$ (S18) The term in the last line vanishes _under the intraband pairing assumption_ : $\mathbf{P}=\mathbf{P}_{1}+\mathbf{P}_{-1}$, $\sum_{\mathbf{b}\neq\pm\mathbf{a}}\partial_{k_{y}}\mathbf{P}_{\mathbf{b}}=-\partial_{k_{y}}\mathbf{P}$, and the fact that $\partial_{k_{y}}\mathbf{P}\partial_{\varphi}\mathbf{H}$ takes the form of $\mathbf{A}^{\vee}$. We therefore arrive at: $\displaystyle\frac{\delta\bm{\Omega}^{\mathbf{a},\text{intra.}}}{\delta\varphi}\doteq$ $\displaystyle\sum_{\mathbf{b}\neq\pm\mathbf{a}}4\mbox{Im\,Tr}[\partial_{k_{x}}\mathbf{P}_{\mathbf{a}}\partial_{k_{y}}\mathbf{P}_{-\mathbf{a}}\mathbf{P}_{\mathbf{b}}\partial_{\varphi}\mathbf{P}_{\mathbf{a}}\mathbf{P}_{\mathbf{a}}]-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\sum_{\mathbf{b}\neq\pm\mathbf{a}}\frac{2\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{H}\mathbf{P}_{\mathbf{b}}\partial_{\varphi}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}(\mathbf{E}_{\mathbf{b}}-\mathbf{E}_{\mathbf{a}})(\mathbf{E}_{\mathbf{b}}-\mathbf{E}_{-\mathbf{a}})}-(x\leftrightarrow y)$ $\displaystyle\doteq$ $\displaystyle\sum_{b\neq 1}\frac{2\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{H}(\mathbf{P}_{b}+\mathbf{P}_{-b})\partial_{\varphi}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}(\epsilon_{b}-\epsilon_{1})^{2}}-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\sum_{b\neq 1}\frac{2\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{H}P_{b}\otimes\bm{\tau}_{0}\partial_{\varphi}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}(\epsilon_{b}-\epsilon_{1})^{2}}-(x\leftrightarrow y)$ (S19) Now we are ready to connect with the properties of the normal state. Introducing two-dimensional eigenvectors $|\bm{\alpha}_{\mathbf{c}}\rangle$ in the particle-hole subspace: $\displaystyle|\mathbf{u}^{\mathbf{c}}_{\bm{k}}\rangle=|u^{|\mathbf{c}|}_{\bm{k}}\rangle\otimes|\bm{\alpha}_{\mathbf{c}}\rangle,$ (S20) and noting that $\displaystyle\partial_{k}\mathbf{H}=\partial_{\bm{k}}H_{0}(\bm{k})\otimes\bm{\tau}_{3}+O(\partial_{\bm{k}}\bm{\Delta}_{\bm{k}}),$ (S21) together with Eq.(S16), one has: $\displaystyle\frac{\delta\bm{\Omega}^{\mathbf{a},\text{intra.}}}{\delta\varphi}\doteq$ $\displaystyle\sum_{b\neq 1}\frac{2|\langle\bm{\alpha}_{\mathbf{a}}|\tau_{3}|\bm{\alpha}_{-\mathbf{a}}\rangle|^{2}\mbox{Im\,Tr}[P_{1}\partial_{k_{x}}H_{0}P_{1}\partial_{k_{y}}H_{0}P_{b}\partial_{\varphi}H_{0}P_{1}]}{\mathbf{E}_{\mathbf{a}}(\epsilon_{b}-\epsilon_{1})^{2}}-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\sum_{b\neq 1}\frac{2|\langle\bm{\alpha}_{\mathbf{a}}|\tau_{3}|\bm{\alpha}_{-\mathbf{a}}\rangle|^{2}v_{x}\mbox{Im\,Tr}[P_{1}\partial_{k_{y}}H_{0}P_{b}\partial_{\varphi}H_{0}P_{1}]}{\mathbf{E}_{\mathbf{a}}(\epsilon_{b}-\epsilon_{1})^{2}}-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle-\frac{v_{x}|\langle\bm{\alpha}_{\mathbf{a}}|\tau_{3}|\bm{\alpha}_{-\mathbf{a}}\rangle|^{2}}{\mathbf{E}_{\mathbf{a}}}\Omega_{k_{y},\varphi}-(x\leftrightarrow y)=-\frac{\varDelta_{\bm{k}}^{2}}{\mathbf{E}_{\mathbf{a}}^{3}}v_{x}\Omega_{k_{y},\varphi}-(x\leftrightarrow y),$ (S22) where $v_{x}\equiv\partial_{k_{x}}\epsilon_{1,\bm{k}}$, and $\Omega_{k_{y},\varphi}\equiv-2\mbox{Im}[\langle\partial_{k_{y}}u^{1}_{\bm{k}}|\partial_{\varphi}u^{1}_{\bm{k}}\rangle]$ is the _normal-state_ Berry curvature w.r.t. $k_{y}$ and $\varphi$. Using the fact that $|\partial_{\varphi}u^{1}_{\bm{k}}\rangle=\frac{-i}{4}\hat{L}|u^{1}_{\bm{k}}\rangle$, the normal-state Berry curvature $\Omega_{k,\varphi}$ can also be expressed as $\displaystyle\Omega_{k,\varphi}=\frac{1}{4}\partial_{\bm{k}}\langle u^{1}_{\bm{k}}|\hat{L}|u^{1}_{\bm{k}}\rangle,$ (S23) so that Eq.(10) in the main text is obtained. ### II.2 Fermi Surface Integral The SATHE can be computed using Eq.(5) and Eq.(6) in the main text. Define $\displaystyle\bm{\Omega}(E)\equiv\sum_{\mathbf{a}}\int\frac{d^{2}k}{(2\pi)^{2}}\bm{\Omega}^{\mathbf{a}}_{\bm{k}}\delta(E-\mathbf{E}_{\mathbf{a},\bm{k}}),$ (S24) we have $\bm{\sigma}(E)=\bm{\sigma}(0)-\int_{0}^{E}dE^{\prime}\bm{\Omega}(E^{\prime})$. Here $\bm{\sigma}(0)=0$ due to the time-reversal symmetry and the nodeless assumption. Because the change of the Berry curvature is concentrated near the Fermi surface due to the $\mathbf{E}_{\mathbf{a}}$ denominator, and proportional to $1/\varDelta_{\bm{k}}$ right at the Fermi surface, it is convenient to represent $\bm{\Omega}(E)$ as a Fermi surface integral as following: $\displaystyle\frac{\delta\bm{\Omega}(E)}{\delta\varphi}\doteq$ $\displaystyle\sum_{\text{FS}}\frac{1}{(2\pi)^{2}}\oint_{\text{FS},\epsilon_{\bm{k}}\equiv\pm\sqrt{E^{2}-\varDelta_{\bm{k}}^{2}}}dk_{\parallel}\frac{|E|}{|\epsilon_{\bm{k}}|}\frac{1}{\hbar|v_{F}|}\mathbf{\Omega}^{\mathop{\mathrm{sign}}(E)}_{\bm{k}}$ $\displaystyle\doteq$ $\displaystyle\sum_{\text{FS}}\frac{-2}{(2\pi)^{2}}\oint_{\text{FS}}dk_{\parallel}\frac{|E|}{\sqrt{E^{2}-\varDelta_{\bm{k}}^{2}}}\frac{1}{\hbar|v_{F}|}\frac{\varDelta_{\bm{k}}^{2}}{E^{3}}(v_{x}\Omega_{k_{y},\varphi}-v_{y}\Omega_{k_{x},\varphi})$ $\displaystyle=$ $\displaystyle\mathop{\mathrm{sign}}(E)\sum_{\text{FS}}\frac{-1}{2\pi^{2}}\oint_{\text{FS}}dk_{\parallel}\frac{\varDelta_{\bm{k}}^{2}}{E^{2}\sqrt{E^{2}-\varDelta_{\bm{k}}^{2}}}\frac{1}{\hbar|v_{F}|}(v_{x}\Omega_{k_{y},\varphi}-v_{y}\Omega_{k_{x},\varphi})$ $\displaystyle=$ $\displaystyle\mathop{\mathrm{sign}}(E)\sum_{\text{FS}}\frac{-\mathop{\mathrm{sign}}(v_{F})}{2\pi^{2}\hbar}\oint_{\text{FS}}d\vec{k}_{\parallel}\cdot(\Omega_{k_{x},\varphi},\Omega_{k_{y},\varphi})\frac{\varDelta_{\bm{k}}^{2}}{E^{2}\sqrt{E^{2}-\varDelta_{\bm{k}}^{2}}}$ $\displaystyle=$ $\displaystyle\mathop{\mathrm{sign}}(E)\sum_{\text{FS}}\frac{-\mathop{\mathrm{sign}}(v_{F})}{8\pi^{2}\hbar}\oint_{\text{FS}}dk_{\parallel}\partial_{k_{\parallel}}\langle u_{\bm{k}}|\hat{L}|u_{\bm{k}}\rangle\frac{\varDelta_{\bm{k}}^{2}}{E^{2}\sqrt{E^{2}-\varDelta_{\bm{k}}^{2}}}.$ (S25) Integrating out $E$ in advance, we get $\displaystyle\bm{\sigma}(E)=\sum_{\text{FS}}\frac{\mathop{\mathrm{sign}}(v_{F})}{8\pi^{2}\hbar}\oint_{\text{FS}}dk_{\parallel}\partial_{k_{\parallel}}\langle u_{\bm{k}}|\hat{L}|u_{\bm{k}}\rangle\frac{\sqrt{E^{2}-\varDelta_{\bm{k}}^{2}}}{|E|}\Theta(|E|-|\varDelta_{\bm{k}}|).$ (S26) We finally arrive at the result using Eq.(5) in the main text: $\displaystyle\frac{\delta\kappa_{xy}}{\delta\varphi T}=\sum_{\text{FS}}\frac{\mathop{\mathrm{sign}}(v_{F})k_{B}^{2}}{16\pi^{2}\hbar}\oint_{\text{FS}}dk_{\parallel}\partial_{k_{\parallel}}\langle u_{\bm{k}}|\hat{L}|u_{\bm{k}}\rangle\xi\bigg{(}\frac{\varDelta_{\bm{k}}}{T}\bigg{)}.$ (S27) ## III Supercurrent-indcued Berry Curvature for Interband Pairing When interband pairing is present, additional terms need to be considered. However, one can always perform a small unitary rotation $e^{iS\otimes\bm{\tau}_{2}}$ to eliminate the interband pairing: $\displaystyle S\equiv\sum_{a\neq b}\frac{P_{a}\bm{\Delta}^{\text{inter.}}_{\bm{k}}P_{b}}{\epsilon_{a,\bm{k}}+\epsilon_{b,\bm{k}}},\quad S^{\dagger}=S.$ (S28) Since $\mathbf{H}=H_{0}\otimes\bm{\tau}_{3}+(\bm{\Delta}^{\text{intra.}}+\bm{\Delta}^{\text{inter.}})\otimes\bm{\tau}_{1}$, we have: $\displaystyle e^{iS\otimes\bm{\tau}_{2}}\mathbf{H}e^{-iS\otimes\bm{\tau}_{2}}\doteq$ $\displaystyle H_{0}\otimes\bm{\tau}_{3}+(\bm{\Delta}^{\text{intra.}}+\bm{\Delta}^{\text{inter.}})\otimes\bm{\tau}_{1}+(i)[S\otimes\bm{\tau}_{2},H_{0}\otimes\bm{\tau}_{3}]$ $\displaystyle=$ $\displaystyle H_{0}\otimes\bm{\tau}_{3}+(\bm{\Delta}^{\text{intra.}}+\bm{\Delta}^{\text{inter.}})\otimes\bm{\tau}_{1}+(i)(SH_{0}+H_{0}S)\otimes(i)\bm{\tau}_{1}$ $\displaystyle=$ $\displaystyle H_{0}\otimes\bm{\tau}_{3}+(\bm{\Delta}^{\text{intra.}}+\bm{\Delta}^{\text{inter.}})\otimes\bm{\tau}_{1}+(-1)\bm{\Delta}^{\text{inter.}}_{\bm{k}}\otimes\bm{\tau}_{1}=H_{0}\otimes\bm{\tau}_{3}+\bm{\Delta}^{\text{intra.}}\otimes\bm{\tau}_{1}$ (S29) Denoting the projectors in the presence of interband pairing as $\widetilde{\mathbf{P}}_{\mathbf{c}}$, we have $\widetilde{\mathbf{P}}_{\mathbf{c}}\doteq e^{-iS\otimes\bm{\tau}_{2}}\mathbf{P}_{\mathbf{c}}e^{iS\otimes\bm{\tau}_{2}}$. Basically, in all the previous derivations for the intraband case, we only need to replace $\mathbf{P}_{\mathbf{c}}$ by $\widetilde{\mathbf{P}}_{\mathbf{c}}$, which corresponds to performing the small unitary transformation for the operators like $\partial_{\varphi}\mathbf{H}\rightarrow e^{iS\otimes\bm{\tau}_{2}}\partial_{\varphi}\mathbf{H}e^{-iS\otimes\bm{\tau}_{2}}$. In this way $\partial_{\varphi}\mathbf{H}$ has a correction $[iS,\partial_{\varphi}H_{0}]\otimes\bm{\tau}_{2}$ and is no longer $\propto\bm{\tau}_{0}$. We can leave all the energies $\mathbf{E}_{\mathbf{c}}$ unchanged since they receive second order contributions from $\bm{\Delta}^{\text{inter.}}$. After inspection, to the leading order of $\varDelta/t$, one can identify three sources of interband contributions. First, we do need to consider $\delta_{\varphi}\partial_{\bm{k}}\mathbf{H}$ in Eq.(8) in the main text. Second, in the first line of Eq.(S18), $\mathbf{b}=-\mathbf{a}$ needs to be included. Third, the term in the last line of Eq.(S18) is no longer vanishing. We term them as _Part-A,B,C_ and compute them one by one below. _Part-A:_ For the $\delta_{\varphi}\partial_{\bm{k}}\mathbf{H}$ contributions, $\displaystyle\left.\frac{\delta\bm{\Omega}^{\mathbf{a}}_{\bm{k}}}{\delta\varphi}\right|_{A}=$ $\displaystyle\frac{-\mbox{Im\,Tr}\big{[}\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}[iS\otimes\bm{\tau}_{2},\partial_{k_{y}}\partial_{\varphi}\mathbf{H}]\mathbf{P}_{\mathbf{a}}\big{]}}{2\mathbf{E}^{2}_{\mathbf{a}}}-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\frac{-\mbox{Im\,Tr}\big{[}\mathbf{P}_{\mathbf{a}}(\partial_{k_{x}}H_{0}\otimes\bm{\tau}_{3}+\partial_{k_{x}}\bm{\Delta}\otimes\bm{\tau}_{1})\mathbf{P}_{-\mathbf{a}}[iS\otimes\bm{\tau}_{2},\partial_{k_{y}}\partial_{\varphi}H_{0}\otimes\bm{\tau}_{0}]\mathbf{P}_{\mathbf{a}}\big{]}}{2\mathbf{E}^{2}_{\mathbf{a}}}-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle-\langle u^{1}_{\bm{k}}|[iS,\partial_{k_{y}}\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle$ $\displaystyle\qquad\times\frac{v_{x}\mbox{Im}\big{[}\langle\bm{\alpha}_{\mathbf{a}}|\bm{\tau}_{3}|\bm{\alpha}_{-\mathbf{a}}\rangle\langle\bm{\alpha}_{-\mathbf{a}}|\bm{\tau}_{2}|\bm{\alpha}_{\mathbf{a}}\rangle\big{]}+\langle u^{1}_{\bm{k}}|\partial_{k_{x}}\bm{\Delta}|u^{1}_{\bm{k}}\rangle\mbox{Im}\big{[}\langle\bm{\alpha}_{\mathbf{a}}|\bm{\tau}_{1}|\bm{\alpha}_{-\mathbf{a}}\rangle\langle\bm{\alpha}_{-\mathbf{a}}|\bm{\tau}_{2}|\bm{\alpha}_{\mathbf{a}}\rangle\big{]}}{2\mathbf{E}^{2}_{\mathbf{a}}}-(x\leftrightarrow y).$ (S30) Noting a few identities: $\displaystyle\bm{\tau}_{2}|\bm{\alpha}_{-\mathbf{a}}\rangle\langle\bm{\alpha}_{-\mathbf{a}}|\bm{\tau}_{2}=|\bm{\alpha}_{\mathbf{a}}\rangle\langle\bm{\alpha}_{\mathbf{a}}|$ $\displaystyle\partial_{k_{x}}\mathbf{E}_{\mathbf{a}}=\mbox{Tr}[\mathbf{P}_{\mathbf{a}}(\partial_{k_{x}}H_{0}\otimes\bm{\tau}_{3}+\partial_{k_{x}}\bm{\Delta}\otimes\bm{\tau}_{1})\mathbf{P}_{\mathbf{a}}]=v_{x}\frac{\epsilon_{1,\bm{k}}}{\mathbf{E}_{a}}+\langle u^{1}_{\bm{k}}|\partial_{k_{x}}\bm{\Delta}|u^{1}_{\bm{k}}\rangle\frac{\varDelta_{\bm{k}}}{\mathbf{E}_{\mathbf{a}}}$ $\displaystyle\partial_{k_{x}}\mathbf{E}_{\mathbf{a}}=\frac{\varDelta_{\bm{k}}\partial_{k_{x}}\varDelta_{\bm{k}}+\epsilon_{1,\bm{k}}v_{x}}{\mathbf{E}_{\mathbf{a}}}$ $\displaystyle\Rightarrow\langle u^{1}_{\bm{k}}|\partial_{k_{x}}\bm{\Delta}|u^{1}_{\bm{k}}\rangle=\partial_{k_{x}}\varDelta_{\bm{k}}$ $\displaystyle\langle\bm{\alpha}_{\mathbf{a}}|\bm{\tau}_{1}|\bm{\alpha}_{\mathbf{a}}\rangle=\frac{\varDelta_{\bm{k}}}{\mathbf{E}_{\mathbf{a}}},\;\;\langle\bm{\alpha}_{\mathbf{a}}|\bm{\tau}_{3}|\bm{\alpha}_{\mathbf{a}}\rangle=\frac{\epsilon_{1,\bm{k}}}{\mathbf{E}_{\mathbf{a}}}$ (S31) we have: $\displaystyle\left.\frac{\delta\bm{\Omega}^{\mathbf{a}}_{\bm{k}}}{\delta\varphi}\right|_{A}=$ $\displaystyle-\frac{\langle u^{1}_{\bm{k}}|[iS,\partial_{k_{y}}\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle}{2\mathbf{E}^{3}_{\mathbf{a}}}\big{(}-\varDelta_{\bm{k}}v_{x}+\epsilon_{1,\bm{k}}\partial_{k_{x}}\varDelta_{\bm{k}}\big{)}-(x\leftrightarrow y)$ _Part-B:_ This term is: $\frac{\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{\varphi}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}^{2}(\mathbf{E}_{-\mathbf{a}}-\mathbf{E}_{\mathbf{a}})}-(x\leftrightarrow y)$ (S33) Note that $\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}=\partial_{k_{y}}\mathbf{E}_{\mathbf{-}a}\mathbf{P}_{-\mathbf{a}}$, $\displaystyle\left.\frac{\delta\bm{\Omega}^{\mathbf{a}}_{\bm{k}}}{\delta\varphi}\right|_{B}=$ $\displaystyle\frac{\partial_{k_{y}}\mathbf{E}_{\mathbf{a}}}{2\mathbf{E}_{\mathbf{a}}^{3}}\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}(\partial_{k_{x}}H_{0}\otimes\bm{\tau}_{3}+\partial_{k_{x}}\bm{\Delta}\otimes\bm{\tau}_{1})\mathbf{P}_{-\mathbf{a}}[iS,\partial_{\varphi}H_{0}]\otimes\bm{\tau}_{2}\mathbf{P}_{\mathbf{a}}]-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\frac{\partial_{k_{y}}\mathbf{E}_{\mathbf{a}}}{2\mathbf{E}_{\mathbf{a}}^{3}}\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}(\partial_{k_{x}}H_{0}\otimes\bm{\tau}_{3}+\partial_{k_{x}}\bm{\Delta}\otimes\bm{\tau}_{1})\mathbf{P}_{-\mathbf{a}}\bm{\tau}_{2}\mathbf{P}_{\mathbf{a}}]-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\frac{\partial_{k_{y}}\mathbf{E}_{\mathbf{a}}}{2\mathbf{E}_{\mathbf{a}}^{3}}\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}(\partial_{k_{x}}H_{0}\otimes\bm{\tau}_{3}+\partial_{k_{x}}\bm{\Delta}\otimes\bm{\tau}_{1})\bm{\tau}_{2}\mathbf{P}_{\mathbf{a}}]-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\frac{\partial_{k_{y}}\mathbf{E}_{\mathbf{a}}}{2\mathbf{E}_{\mathbf{a}}^{3}}\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle(-v_{x}\langle\bm{\alpha}_{\mathbf{a}}|\bm{\tau}_{1}|\bm{\alpha}_{\mathbf{a}}\rangle+\langle u^{1}_{\bm{k}}|\partial_{k_{x}}\bm{\Delta}|u^{1}_{\bm{k}}\rangle\langle\bm{\alpha}_{\mathbf{a}}|\bm{\tau}_{3}|\bm{\alpha}_{\mathbf{a}}\rangle)-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\frac{\partial_{k_{y}}\mathbf{E}_{\mathbf{a}}}{2\mathbf{E}_{\mathbf{a}}^{4}}\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle\big{(}-\varDelta_{\bm{k}}v_{x}+\epsilon_{1,\bm{k}}\partial_{k_{x}}\varDelta_{\bm{k}}\big{)}-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\frac{(v_{y}\epsilon_{1,\bm{k}}+\varDelta_{\bm{k}}\partial_{k_{y}}\varDelta_{\bm{k}})}{2\mathbf{E}_{\mathbf{a}}^{5}}\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle\big{(}-\varDelta_{\bm{k}}v_{x}+\epsilon_{1,\bm{k}}\partial_{k_{x}}\varDelta_{\bm{k}}\big{)}-(x\leftrightarrow y)$ (S34) After $(x\leftrightarrow y)$ antisymmetrization, we get $\displaystyle\left.\frac{\delta\bm{\Omega}^{\mathbf{a}}_{\bm{k}}}{\delta\varphi}\right|_{B}=$ $\displaystyle\frac{-1}{2\mathbf{E}_{\mathbf{a}}^{3}}\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle v_{x}\partial_{k_{y}}\varDelta_{\bm{k}}-(x\leftrightarrow y)$ (S35) _Part-C:_ Note that $\displaystyle\partial_{k_{y}}\widetilde{\mathbf{P}}_{\mathbf{b}}\doteq\partial_{k_{y}}\big{[}\mathbf{P}_{\mathbf{b}}+[-iS\otimes\bm{\tau}_{2},\mathbf{P}_{\mathbf{b}}]\big{]}\doteq e^{-iS\otimes\bm{\tau}_{2}}\partial_{k_{y}}\mathbf{P}_{\mathbf{b}}e^{iS\otimes\bm{\tau}_{2}}+[-i\partial_{k_{y}}S\otimes\bm{\tau}_{2},\mathbf{P}_{\mathbf{b}}]$ (S36) The contribution from the last line of Eq.(S18) becomes: $\displaystyle\left.\frac{\delta\bm{\Omega}^{\mathbf{a}}_{\bm{k}}}{\delta\varphi}\right|_{C}\doteq$ $\displaystyle\frac{-\mbox{Im\,Tr}[\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}\partial_{k_{y}}\mathbf{P}[iS\otimes\bm{\tau}_{2},\partial_{\varphi}\mathbf{H}]\mathbf{P}_{\mathbf{a}}+\mathbf{P}_{\mathbf{a}}\partial_{k_{x}}\mathbf{H}\mathbf{P}_{-\mathbf{a}}[-i\partial_{k_{y}}S\otimes\bm{\tau}_{2},\mathbf{P}]\partial_{\varphi}\mathbf{H}\mathbf{P}_{\mathbf{a}}]}{\mathbf{E}_{\mathbf{a}}^{2}}-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\frac{-1}{\mathbf{E}_{\mathbf{a}}^{3}}\big{(}-\varDelta_{\bm{k}}v_{x}+\epsilon_{1,\bm{k}}\partial_{k_{x}}\varDelta_{\bm{k}}\big{)}\mbox{Re\,Tr}[\partial_{k_{y}}P_{1}[iS,\partial_{\varphi}H_{0}]P_{1}+P_{1}i\partial_{k_{y}}S\partial_{\varphi}H_{0}P_{1}]-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\frac{-1}{2\mathbf{E}_{\mathbf{a}}^{3}}\big{(}-\varDelta_{\bm{k}}v_{x}+\epsilon_{1,\bm{k}}\partial_{k_{x}}\varDelta_{\bm{k}}\big{)}\mbox{Re\,Tr}[2\partial_{k_{y}}P_{1}[iS,\partial_{\varphi}H_{0}]P_{1}+P_{1}[i\partial_{k_{y}}S,\partial_{\varphi}H_{0}]P_{1}]-(x\leftrightarrow y)$ (S37) Let’s add part-A and part-C together. $\displaystyle\left.\frac{\delta\bm{\Omega}^{\mathbf{a}}_{\bm{k}}}{\delta\varphi}\right|_{A+C}=$ $\displaystyle\frac{-1}{2\mathbf{E}_{\mathbf{a}}^{3}}\big{(}-\varDelta_{\bm{k}}v_{x}+\epsilon_{1,\bm{k}}\partial_{k_{x}}\varDelta_{\bm{k}}\big{)}\partial_{k_{y}}\Big{[}\mbox{Re\,Tr}[P_{1}[iS,\partial_{\varphi}H_{0}]P_{1}]\Big{]}-(x\leftrightarrow y)$ $\displaystyle=$ $\displaystyle\frac{-1}{2\mathbf{E}_{\mathbf{a}}^{3}}\big{(}-\varDelta_{\bm{k}}v_{x}+\epsilon_{1,\bm{k}}\partial_{k_{x}}\varDelta_{\bm{k}}\big{)}\partial_{k_{y}}\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle-(x\leftrightarrow y)$ (S38) The quantity $\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle$ here and in part-B can be easily computed, which we denote as $G_{\bm{k}}$: $\displaystyle G_{\bm{k}}\equiv$ $\displaystyle\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle=-2\mbox{Im}[\langle u^{1}_{\bm{k}}|S\partial_{\varphi}H_{0}|u^{1}_{\bm{k}}\rangle]=-2\sum_{b\neq 1}\mbox{Im}[\langle u^{1}_{\bm{k}}|S|u^{b}_{\bm{k}}\rangle\langle u^{b}_{\bm{k}}|\partial_{\varphi}H_{0}|u^{1}_{\bm{k}}\rangle]$ $\displaystyle=-2\sum_{b\neq 1}\mbox{Im}\Big{[}\frac{\langle u^{1}_{\bm{k}}|\bm{\Delta}^{\text{inter.}}|u^{b}_{\bm{k}}\rangle\langle u^{b}_{\bm{k}}|\partial_{\varphi}H_{0}|u^{1}_{\bm{k}}\rangle}{\epsilon_{1,\bm{k}}+\epsilon_{b,\bm{k}}}\Big{]}\doteq-2\sum_{b\neq 1}\mbox{Im}\Big{[}\frac{\langle u^{1}_{\bm{k}}|\bm{\Delta}^{\text{inter.}}|u^{b}_{\bm{k}}\rangle\langle u^{b}_{\bm{k}}|\partial_{\varphi}H_{0}|u^{1}_{\bm{k}}\rangle}{-\epsilon_{1,\bm{k}}+\epsilon_{b,\bm{k}}}\Big{]}$ $\displaystyle=2\sum_{b\neq 1}\mbox{Im}[\langle u^{1}_{\bm{k}}|\bm{\Delta}^{\text{inter.}}|u^{b}_{\bm{k}}\rangle\langle u^{b}_{\bm{k}}|\partial_{\varphi}u^{1}_{\bm{k}}\rangle]=2\mbox{Im}[\langle u^{1}_{\bm{k}}|\bm{\Delta}^{\text{inter.}}|\partial_{\varphi}u^{1}_{\bm{k}}\rangle]=\frac{-1}{2}\mbox{Re}[\langle u^{1}_{\bm{k}}|\bm{\Delta}^{\text{inter.}}\hat{L}|u^{1}_{\bm{k}}\rangle]$ (S39) Putting together, we finally get $\displaystyle\frac{\delta\bm{\Omega}^{\mathbf{a},\text{inter.}}_{\bm{k}}}{\delta\varphi}=\left.\frac{\delta\bm{\Omega}^{\mathbf{a}}_{\bm{k}}}{\delta\varphi}\right|_{A+B+C}=$ $\displaystyle\frac{-1}{2\mathbf{E}_{\mathbf{a}}^{3}}\big{(}-\varDelta_{\bm{k}}v_{x}+\epsilon_{1,\bm{k}}\partial_{k_{x}}\varDelta_{\bm{k}}\big{)}\partial_{k_{y}}G_{\bm{k}}+\frac{-1}{2\mathbf{E}_{\mathbf{a}}^{3}}G_{\bm{k}}v_{x}\partial_{k_{y}}\varDelta_{\bm{k}}-(x\leftrightarrow y)$ (S40) It is also instructive to study the behavior of $\partial_{\varphi}\mathbf{H}$ in the low energy subspace: $\displaystyle\widetilde{\mathbf{P}}\partial_{\varphi}\mathbf{H}\widetilde{\mathbf{P}}=\mathbf{P}\partial_{\varphi}\mathbf{H}\mathbf{P}+\mathbf{P}[iS\otimes\bm{\tau}_{2},\partial_{\varphi}\mathbf{H}]\mathbf{P}=\langle u^{1}_{\bm{k}}|\partial_{\varphi}H_{0}|u^{1}_{\bm{k}}\rangle\mathbf{P}+\langle u^{1}_{\bm{k}}|[iS,\partial_{\varphi}H_{0}]|u^{1}_{\bm{k}}\rangle\mathbf{P}\bm{\tau}_{2}\mathbf{P}$ (S41) The low energy effective Hamiltonian in the presence of $\delta\varphi$ becomes: $\displaystyle\mathbf{H}_{\text{eff}}=\delta\varphi\partial_{\varphi}\epsilon_{1,\bm{k}}\bm{\tau}_{0}+\epsilon_{1,\bm{k}}\bm{\tau}_{3}+\varDelta_{\bm{k}}\bm{\tau}_{1}+\delta\varphi G_{\bm{k}}\bm{\tau}_{2}=\delta\varphi\partial_{\varphi}\epsilon_{1,\bm{k}}\bm{\tau}_{0}+\mathbf{d}_{\bm{k}}\cdot\vec{\bm{\tau}},$ (S42) where we introduced vector $\mathbf{d}_{\bm{k}}\equiv(\varDelta_{\bm{k}},\delta\varphi G_{\bm{k}},\epsilon_{1,\bm{k}})$. We merely showed that the interband contribution can be faithfully computed using this effective 2-by-2 Hamiltonian: $\displaystyle\delta\bm{\Omega}^{\mathbf{a},\text{inter.}}_{\bm{k}}=-2\mbox{Im\,Tr}[p_{\mathbf{a}}\partial_{k_{x}}p_{\mathbf{a}}\partial_{k_{y}}p_{\mathbf{a}}]=\frac{-d_{\bm{k}}\cdot(\partial_{k_{x}}\mathbf{d}_{\bm{k}}\times\partial_{k_{y}}\mathbf{d}_{\bm{k}})}{2\mathbf{a}|\mathbf{d}_{\bm{k}}|^{3}}\doteq\frac{-\mathbf{d}_{\bm{k}}\cdot(\partial_{k_{x}}\mathbf{d}_{\bm{k}}\times\partial_{k_{y}}\mathbf{d}_{\bm{k}})}{2\mathbf{E}_{\mathbf{a}}^{3}}\propto\delta\varphi$ (S43) $p_{\mathbf{a}}=\frac{1}{2}(1+\mathbf{a}\frac{\mathbf{d}_{\bm{k}}\cdot\vec{\bm{\tau}}}{|\mathbf{d}_{\bm{k}}|})$ is the projector in this effective model. When the superconductivity is nodal, near a node, the effective theory becomes a Dirac equation: $\displaystyle\mathbf{H}^{\text{node}}_{\text{eff}}=\delta\varphi\partial_{\varphi}\epsilon_{1,\bm{k}}\bm{\tau}_{0}+\hbar v_{F}k_{\perp}\bm{\tau}_{3}+\hbar v_{\Delta}k_{\parallel}\bm{\tau}_{1}+m\bm{\tau}_{2},$ (S44) where $\hbar v_{\Delta}=\frac{\partial\varDelta_{\bm{k}}}{\partial k_{\parallel}}$, and $m=\delta\varphi G_{\bm{k}_{\text{node}}}$ is the mass gap generated by the supercurrent. It is easy to show that the nodal contribution to $\bm{\sigma}(E)$ is $\displaystyle\bm{\sigma}^{\text{node}}(E)=\begin{cases}\frac{C}{2\pi},&\text{if }-|m|<E<|m|,\\\ \frac{C|m|}{2\pi|E|},&\text{otherwise}.\end{cases}$ (S45) Here $C=\pm\frac{1}{2}$ is the Chern number transferred due to $m$. Performing the energy integral in Eq.(5) in the main text, in the low temperature limit $k_{B}T\ll|m|$ one recovers the quantized $\kappa^{\text{node}}_{xy}/T=\frac{k_{B}^{2}}{\hbar}\frac{C\pi}{12}$. However, in the high-temperature limit $m\ll k_{B}T$, we have: $\displaystyle\frac{\kappa_{xy}^{\text{node}}}{T}\doteq-\frac{k_{B}^{2}}{\hbar}\frac{C|m|}{2\pi k_{B}T}\int^{\infty}_{0}xf^{\prime}(x)dx=\frac{k_{B}^{2}}{\hbar}\frac{C\ln 2}{2\pi}\frac{|m|}{k_{B}T}.$ (S46) Namely there is a $1/T$ tail in $\frac{\kappa_{xy}}{T}$. As a final remark, in the absence of $\delta\varphi$, it is easy to show that the Berry curvature is nonsingular near the Fermi surface. Consequently, one does not need to consider the contribution to SATHE from $\partial_{\varphi}\mathbf{E}_{\mathbf{a}}$ in the leading order of $\varDelta/t$ expansion. ## IV Fermi Surface Reconstruction for Twisted Bilayer FeSe The effective $\bm{k\cdot p}$ model in the main text is written in the oribtal space of $\\{3d_{xz},3d_{yz}\\}$. Because the dominant hopping processes occur within _the same_ orbitals, we can treat each orbital separately, or equivalently go back to the single-iron Brillouin zone to work with one ellipse on each direction. The right horizontal elliptic $M$-pocket of monolayer FeSe is simply described with $h=-\widetilde{\mu}+a(k_{x}-k_{M})^{2}+bk_{y}^{2}$ (S47) with $\mu=-0.08$eV, $a=1.08\text{eV}\cdot\text{\AA}^{2}$ and $b=1.6\text{eV}\cdot\text{\AA}^{2}$. Denoting the rotated monolayer Hamiltonian as $h_{\pm\theta/2}(\bm{k})$, the rotated bilayer Hamiltonian _without_ interlayer tunnelings is then simply a two-by-two diagonal matrix $H_{\text{w/o }T_{\perp}}=\mathop{\mathrm{diag}}\\{h_{\theta/2}(\bm{k}),h_{-\theta/2}(\bm{k})\\}$. Under small twisting angles, two-center approximation applies Bistritzer and MacDonald (2011) and the general tunneling strength between the top/bottom Bloch states $|\psi^{t}_{\alpha,\bm{k}^{t}}\rangle$ and $|\psi^{b}_{\beta,\bm{k}^{b}}\rangle$ takes the form of $t_{\bm{k}^{t},\bm{k}^{b}}^{\alpha\beta}\equiv\langle\psi_{\alpha}^{t}|H|\psi_{\beta}\rangle=\dfrac{1}{V}\sum_{\bm{G}^{t},\bm{G}^{b}}\delta_{\bm{k}^{t}+\bm{G}^{t},\bm{k}^{b}+\bm{G}^{b}}\cdot e^{-i\bm{G}^{b}\cdot\bm{\tau}^{b}_{\alpha}}\cdot t_{\alpha\beta}(\bm{k}^{t}+\bm{G}^{t})\cdot e^{i\bm{G}^{t}\cdot\bm{\tau}^{t}_{\beta}},$ (S48) where $\bm{G}^{t,b}$ are the reciprocal vector of the top/bottom layer and $\bm{\tau}_{\alpha}^{t}$ and $\bm{\tau}_{\beta}^{b}$ are the sublattice position of the wannier centers of top/bottom states. Now that only one state is left in the single-iron Brillouin zone, we can simply take $t_{\alpha\beta}(\bm{k}^{t}+\bm{G}^{t})=\delta_{\alpha\beta}t(\bm{k}^{t}+\bm{G}^{t})$. The low-energy form of the interlayer tunnelings can be obtained with the expansion $\bm{k}^{t,b}\equiv\bm{q}^{t,b}+\bm{K}_{M}^{t,b}$, $|\bm{q}^{t,b}|\ll 1$. Assuming that the tunneling function $t(\bm{k}^{t}+\bm{G}^{t})$ only depends on the norm of its arguments, we can take the approximation that $t(\bm{k}^{t}+\bm{G}^{t})\equiv t(\bm{q}^{t}+\bm{K}_{M}^{t}+\bm{G}^{t})\doteq t(|\bm{K}_{M}|).$ Since $|\bm{K}_{M}|\gg|\bm{q}^{t,b}|$, only terms with momentum differences $\bm{k}^{t}+\bm{G}^{t}-\bm{k}^{b}-\bm{G}^{b}\propto\theta$ will be left due to the delta function in Eq.(S48). Taking the right $M$-pocket as an example, up to the lowest-order truncation on the grids of the scattering vectors, only two terms of $\delta\bm{q}^{t}$ and $\delta\bm{q}^{b}$ need to be included, as is seen in Fig.S2. Figure S2: (a) Rotated top layer (green) and bottom layer (orange). The interference of all different hopping processes capture the spatial variation of interlayer tunnelings, which determines the moiré Brillouin zone (gray square). Up to the lowest truncation on such moiré k-shell, there are just two vertical hopping vectors $\delta\bm{q}^{t}$ and $\delta\bm{q}^{b}$ involved in the reconstruction of the left/right $M$-pockets (ditto for the top/bottom ones under a $C_{4}$-rotation). (b) Illustration for the moire-pattern- involved Fermi surface reconstruction. Taking the top layer elliptic Fermi surface (green dashed lines) as the reference, both the rotated bottom layer (down orange dashed lines) and the copied $M$-pockets due to $G_{x}$ (up orange dashed lines) involve in the reconstruction of the Fermi surfaces (red solid lines). Denoting $t(|\bm{K}_{M}|)/V\equiv w$, such truncation gives rise to the minimal moiré Hamiltonian $H^{\text{moir\'{e}}}(\bm{k})=\left(\begin{array}[]{ccc}h_{\theta/2}(\bm{k})&w&w\\\ w&h_{-\theta/2}(\bm{k}+\delta\bm{q}^{t})&0\\\ w&0&h_{-\theta/2}(\bm{k}+\delta\bm{q}^{b})\end{array}\right).$ (S49) The moiré pattern-induced Fermi surface reconstruction is then obtained by diagonalization Eq.(S49), which recovers the two vertical _open_ curves in (a1) of Fig.2 in the main text. The other two horizontal curves comes from the interference of the hopping processes between the up/down $M$-pockets.
# Canards in a bottleneck Annalisa Iuorio<EMAIL_ADDRESS>Gaspard Jankowiak <EMAIL_ADDRESS>Peter Szmolyan<EMAIL_ADDRESS>Marie- Therese Wolfram<EMAIL_ADDRESS> ###### Abstract In this paper we investigate the stationary profiles of a nonlinear Fokker- Planck equation with small diffusion and nonlinear in- and outflow boundary conditions. We consider corridors with a bottleneck whose width has a global nondegenerate minimum in the interior. In the small diffusion limit the profiles are obtained constructively by using methods from geometric singular perturbation theory (GSPT). We identify three main types of profiles corresponding to: (i) high density in the domain and a boundary layer at the entrance, (ii) low density in the domain and a boundary layer at the exit, and (iii) transitions from high density to low density inside the bottleneck with boundary layers at the entrance and exit. Interestingly, solutions of the last type involve canard solutions generated at the narrowest point of the bottleneck. We obtain a detailed bifurcation diagram of these solutions in terms of the in- and outflow rates. The analytic results based on GSPT are further corroborated by computational experiments investigating corridors with bottlenecks of variable width. [label1]organization=University of Vienna, Faculty of Mathematics, addressline=Oskar-Morgenstern-Platz 1, city=Vienna, postcode=1090, country=Austria [label2]organization=Universität Konstanz, Fachbereich Mathematik und Statistik, addressline=Fach D 197, city=Konstanz, postcode=78457, country=Germany [label3]organization=Technische Universität Wien, Institute for Analysis and Scientific Computing, addressline=Wiedner Hauptstr. 8-10, city=Vienna, postcode=1040, country=Austria [label4]organization=University of Warwick, Mathematics Institute, city=Coventry, postcode=CV47AL, country=UK lox ## 1 Introduction In this paper we investigate the stationary profiles of a nonlinear Fokker- Planck equation with inflow and outflow boundary conditions, describing the unidirectional cross-sectional average flow of pedestrians in corridors with a single entrance and exit. Changes in the cross section lead to an in- or decrease of the possible flow inside the corridor; different in- and outflow conditions to the formation of boundary layers at the entrance and exit. In [7] the authors derived the investigated 1D area averaged model from a nonlinear convection diffusion equation that was originally proposed by Burger and Pietschmann in [3]. They studied the formation of boundary layers in the case of strictly monotone cross sectional profiles using geometric singular perturbation theory (GSPT). In this paper we extend our analysis to corridors with a unique point of minimal width which we denote as bottlenecks in the following. There has been an increased interest in the analysis of PDE models for pedestrian flows within the applied mathematics community in the last years. These models usually describe the dynamics of a single group of pedestrians having a common goal; for example unidirectional flows in corridor; or several groups with different objectives; as in bidirectional flows see [2, 1]. The resulting PDEs or systems of PDEs are usually highly nonlinear and coupled. In addition to nonlinear boundary conditions, convection dominated terms as well as nonlinear interaction terms require the use of non-standard analytical and computational techniques to show existence of solutions, analyse their long time behavior and perform computational experiments. Stationary profiles of these PDE models provide useful insights into the complex dynamics and allow to predict seggregation dynamics (in the case of multi-species flows) or the formation of boundary layers or high density regions (in the case of low or high inflow and outflow rates or at bottlenecks), see for example [1, 3]. For a general overview on mathematical modeling, analysis and simulation we refer to [4, 13]. We reiterate that the investigated PDE model for area averaged flows comprises a nonlinear convection and linear diffusion term, as well as nonlinear in- and outflow at the entrance and exit. The interplay of small diffusion, the geometry of the domain as well as the in- and outflow rates lead to the formation of boundary layers, which we analyse using Geometric Singular Perturbation Theory (GSPT). GSPT is a dynamical systems approach to singularly perturbed ordinary differential equations started by the pioneering work of Fenichel [6]. The most common form of GSPT considers slow-fast systems of the form $\displaystyle\dot{u}$ $\displaystyle=f(u,v),$ (1) $\displaystyle\varepsilon\dot{v}$ $\displaystyle=g(u,v),$ where $u$ and $v$ are functions of $t$ and $0<\varepsilon\ll 1$. Often $t$ has the interpretation of time but it may represent equally well a spatial variable. For $f=O(1)$ and $g=O(1)$ the variable $u$ varies on the slow time- scale $t$ and the variable $v$ on the fast time-scale $\tau:=\frac{t}{\varepsilon}$, which explains the name slow-fast system. Written on the fast time-scale the equation has the form $\displaystyle u^{\prime}$ $\displaystyle=\varepsilon f(u,v),$ (2) $\displaystyle v^{\prime}$ $\displaystyle=g(u,v).$ Under suitable assumptions, solutions of System (1) for small values of $\varepsilon$ can be constructed as perturbation of concatenations of solutions of the two limiting problems obtained by setting $\varepsilon=0$ in systems (1) and (2), which are referred to as the reduced problem and the layer problem, respectively. In GSPT, these constructions are carried out in the framework of dynamical systems theory; with the theory of invariant manifolds playing a particularly important role. In the specific problem analysed in this paper, well established results and methods from GSPT are used and adapted for the analysis of a boundary value problem. Therefore, we do not give a more detailed summary of GSPT, but refer to [10, 11] for more background on GSPT and its many applications. To name a few recent applications we mention the analysis of multi-scale structures in Micro-Electro-Mechanical Systems [8] and in vegetation patterns [9]. In the context of pedestrian dynamics, GSPT has been successfully applied to study closing channels in [7]. The necessary concepts and results from GSPT are explained in Section 2 as needed in the context of the specific problem at hand. ### 1.1 The mathematical model In the following we briefly discuss the underlying modeling assumption of the area averaged PDE under investigation. A more detailed derivation can be found in [7]. We consider a undirectional flow of a large pedestrian crowd, whose density if given by $\rho=\rho(x,y,t)$, in a 2D domain with a single entrance (at $x=0$) and a single exit (at $x=L$). Furthermore we assume that the pedestrian density is constant across the cross section, that is for fixed $x$. This assumption is satisfied if * • the domain is symmetric with respect to the $x$-axis, and * • the initial pedestrian distribution is symmetric with respect to the $x$-axis. We assume that the dynamics are driven by convective transport and diffusion, in particular the total normalised pedestrian flow is given by $\displaystyle\mathbf{j}=-\varepsilon\nabla\rho+\rho(1-\rho)\mathbf{u},$ (3) where $\mathbf{u}:\mathbb{R}^{2}\rightarrow\mathbb{R}$ is a normalised vector field in the desired direction (in our case pointing in the general direction of the exit) and $\varepsilon>0$ is the diffusion coefficient. We see that the average velocity corresponds to $1-\rho$, hence individuals move at maximum speed $1$ at density $\rho\equiv 0$ and vanishes if the density reaches its maximum value $\rho\equiv 1$. Note that the relation of the average density to the average velocity is commonly referred to as the fundamental diagram, and that similar relations have been investigated in traffic flow; consider for example the well known Lighthill-Whitham-Richard model [12, 14]. In [7] the authors derived a 1D area averaged PDE model, which is based on the above assumptions and a suitable rescaling in space. It reads as $\displaystyle\partial_{t}\rho(x,t)=\partial_{x}\left(k(x)(-\varepsilon\partial_{x}\rho(x,t)+\rho(x,t)\,(1-\rho(x,t)))\right)=0$ (4a) The equation is supplemented with in- and outflow conditions $\displaystyle j(0,t)$ $\displaystyle=\alpha(1-\rho(0,t))$ (4b) $\displaystyle j(L,t)$ $\displaystyle=\beta\rho(0,t)$ (4c) where $j(x,t)=-\varepsilon\partial_{x}\rho(x,t)+\rho(x,t)\,(1-\rho(x,t))$ is the 1D equivalent of (3). In the derivation of (4), the function $k$ is the product of the width with the cross-sectional average of first component of $\mathbf{u}$. For simplicity we refer to $k$ as the width of the bottleneck, which amounts to assuming that the cross-sectional average is 1. The parameters $\alpha>0$ and $\beta>0$ are the inflow and outflow rate, respectively. The boundary condition (4b) describes the inflow at the entrance; the inflow is maximal if the entrance is empty ($\rho\equiv 0$), but decreases to zero when approaching the maximum density $\rho\equiv 1$. At the exit (4c) we do not assume that the outflow is limited by the maximum capacity. Hence, the outflow rate is proportional to the density of individuals at the exit. In [7] the existence of a unique stationary solution of (4) has been established in great generality by PDE methods, thus, it remains to understand its structure and dependence on parameters. ### 1.2 Content and organisation of the paper In this paper we continue and extend the analysis of stationary profiles for system (4) in [7], where a detailed analysis of stationary profiles and their dependence on the in- and outflow rates $\alpha$ and $\beta$ was given for corridors with monotonically decreasing (or increasing) functions $k$. Recall, that smaller values of the function $k$ account for reduced mobility in narrower regions. It was shown in [7] that the nonlinear in- and outflow conditions lead to the formation of boundary layers at the entrance or exit. Building on the approach in [7] we now consider the important case of corridors, whose width has a unique minimum. This setting corresponds to functions $k$ which have a unique global minimum at $x=x_{0}\in(0,L)$. In the following we refer to domains of this type as corridors with bottlenecks (or sometimes only bottlenecks). Throughout this paper we will without loss of generality, assume that $L=1$. Therefore, we investigate the stationary states of system (4) described by $\partial_{x}\mathrm{J}=\partial_{x}(k(x)j(x))=0\,,$ (5a) where $j(x)=-\varepsilon\partial_{x}\rho(x)+\rho(x)\,(1-\rho(x))$ coupled with the following boundary conditions $\displaystyle j$ $\displaystyle=\alpha\left(1-\rho\right)$ $\displaystyle\text{ at }x$ $\displaystyle=0\,,$ (5b) $\displaystyle j$ $\displaystyle=\beta\rho$ $\displaystyle\text{ at }x$ $\displaystyle=1\,.$ We characterise all profiles for different inflow and outflow rates $\alpha$ and $\beta$ in the singular limit $\varepsilon=0$. We identify 8 regions in parameter space corresponding to profiles with different structures. Two of these regions correspond to high-density profiles, two other regions correspond to low density profiles. These profiles are quite similar to profiles considered in [7] and are only weakly affected by the presence of the bottleneck. Due to the bottleneck a new interesting class of profiles exists, which corresponds to solutions starting at high density and making a transition to low density in the region where the function $k$ attains its minimum. Since this type of solutions allows four possible configurations of boundary layers this leads to four types of transitional density profiles. We refer to these four types of profiles as transitional profiles. Our GSPT analysis shows that these transitional profiles are caused by the existence of canard solutions passing through a folded saddle [15]. Canard solutions are solutions of singularly perturbed ODEs which follow repelling slow manifolds for a considerable time. The essence of the canard phenomenon is that these solutions lie exponentially close to the repelling slow manifold and are therefore able to follow it for some time before they are ultimately repelled from it. Clearly, special mechanisms are needed to bring solutions of interest exponentially close to the repelling slow manifold. The occurrence of canard solutions in boundary value problems is conceptually less surprising than their occurrence in initial value problems, nevertheless we are not aware of similar works or results in the context of nonlinear boundary value problems. The three profile types have a similar structure as the low density, high density and maximum current phases observed in Totally Asymmetric Simple Exclusion Process (TASEP). Note that the proposed model (4) was derived from a 2D TASEP, see [3]. In the TASEP, $\alpha$ and $\beta$ are the entry and exit rates, respectively, see for example [5, 17]. The rest of the paper is organised as follows. The GSPT analysis leading to the main result on the structure of solutions is carried out in Section 2. In Section 3 the analytical results are illustrated and confirmed by computational experiments for different channels. We conclude with an interpretation of the main features of the constructed solutions in the various regimes in a manner which could be useful in further studies of pedestrian dynamics. ## 2 GSPT analysis In this section, the stationary states associated to (5b) are investigated in a bottleneck scenario. The problem is rewritten as an equivalent boundary value problem for an autonomous three-dimensional system of first order differential equations in slow-fast form. As explained in the introduction, we will identify 8 regions in the $(\alpha,\beta)$ parameter space, in which the stationary profiles have the same structure in the singular limit $\varepsilon=0$. We construct singular solutions of the boundary value problem as concatenations of solutions of the corresponding layer- and reduced problem. These singular solutions are then shown to persist for $\varepsilon$ small. The profiles which exist in four of these regions involve a canard solution generated at a point corresponding to the minimum of $k$. For the rest of this paper we make the following assumption which is crucial for our approach and results. Main Assumption: The function $k\in C^{2}([0,1])$ is positive and has a unique global nondegenerate minimum at $x=x^{\ast}\in(0,1)$ satisfying $k^{\prime}(x^{\ast})=0,\qquad k^{\prime\prime}(x^{\ast})>0.$ (6) ###### Remark 1. Here we denote the derivative of the coefficient function $k$ as $k^{\prime}$. Below we will also consider the function $g:=k^{\prime}/k$, its derivative will also be denoted as $g^{\prime}$. We would like to point out that starting with Equation (10) the symbol ′ will be mainly used to denote derivatives of the sought solution with respect to a rescaled fast variable. The above slight use of notations should not lead to any confusion. By introducing the function $g:=\frac{k^{\prime}}{k}$ we can rewrite Equation (5b) as the system $\displaystyle\frac{dj}{dx}$ $\displaystyle=-g(x)j,$ (7) $\displaystyle\varepsilon\frac{d\rho}{dx}$ $\displaystyle=\rho(1-\rho)-j.$ Analogously to [7], this system can be transformed into an autonomous system by introducing the variable $\xi=x$ as a new dynamic variable and including the trivial equation $\frac{d\xi}{dx}=1$. From now on, we use the notation $\dot{~{}}=\frac{d}{dx}$. Thus, we obtain the following autonomous reformulation of Equation (7) $\displaystyle\dot{j}$ $\displaystyle=-g(\xi)j,$ (8) $\displaystyle\dot{\xi}$ $\displaystyle=1,$ $\displaystyle\varepsilon\dot{\rho}$ $\displaystyle=\rho(1-\rho)-j,$ where the above assumptions on $k$ identically apply with $\xi^{\ast}=x^{\ast}$, with boundary conditions $\displaystyle j$ $\displaystyle=\alpha\left(1-\rho\right)$ $\displaystyle\text{ at }\xi$ $\displaystyle=0\,,$ (9) $\displaystyle j$ $\displaystyle=\beta\rho$ $\displaystyle\text{ at }\xi$ $\displaystyle=1\,.$ System (8) is a slow-fast system, where the dynamics of $\rho$ occur on the fast scale, while the dynamics of $j$ and $\xi$ take place on the slow scale. By transforming to the fast variable $\chi=\frac{x}{\varepsilon}$, and using the notation ${}^{\prime}=\frac{d}{d\chi}$, we can rewrite System (8) as $\displaystyle j^{\prime}$ $\displaystyle=-\varepsilon g(\xi)j,$ (10) $\displaystyle\xi^{\prime}$ $\displaystyle=\varepsilon,$ $\displaystyle\rho^{\prime}$ $\displaystyle=\rho(1-\rho)-j.$ As explained in the introduction, letting $\varepsilon\to 0$ in Equations (8) and (10) leads to two limiting subproblems – i.e. the _reduced_ problem and the _layer_ problem, respectively – which are simpler to analyse. The layer problem ($\varepsilon=0$ in (10)) is given by $\displaystyle j^{\prime}$ $\displaystyle=0,$ (11) $\displaystyle\xi^{\prime}$ $\displaystyle=0,$ $\displaystyle\rho^{\prime}$ $\displaystyle=\rho(1-\rho)-j,$ and describes the dynamics of the fast variable $\rho$ for fixed $j$ and $\xi$ values. The manifold of its equilibria is known as the _critical manifold_ $\mathcal{C}_{0}:=\left\\{(j,\xi,\rho)~{}:~{}j=\rho(1-\rho)\right\\},$ (12) which is a folded surface in $(j,\xi,\rho)$ space. The critical manifold $\mathcal{C}_{0}$ is the union of two submanifolds $\mathcal{C}_{0}^{a}$ ($\rho>\frac{1}{2}$) and $\mathcal{C}_{0}^{r}$ ($\rho<\frac{1}{2}$) – which are attracting and repelling, respectively – and a line of fold points $F:=\left\\{(j,\xi,\rho)~{}:~{}j=\frac{1}{4},~{}\rho=\frac{1}{2}\right\\},$ (13) as shown in Figure 1. Fenichel Theory [6] implies that away from the fold line $F$ the submanifolds $\mathcal{C}_{0}^{a}$ and $\mathcal{C}_{0}^{r}$ perturb to (non-unique) attracting and repelling slow manifolds $\mathcal{C}_{\varepsilon}^{a}$ and $\mathcal{C}_{\varepsilon}^{r}$ for $\varepsilon$ small. If the reduced flow reaches the fold line $F$ transversally at a point $p\in F$, the point $p$ is a jump point where a transition to fast motion close to solutions of the layer problem occurs, see [16]. At exceptional points $p\in F$ where this transversality condition is violated solutions of the reduced flow may cross through $p$ from $\mathcal{C}_{0}^{a}$ to $\mathcal{C}_{0}^{r}$, or vice versa. Such solutions are called (singular) canards, the corresponding $p\in F$ is a canard point. The least degenerate canard points have been classified and analysed by the blow-up method as folded saddles and folded nodes in [15]. There it is shown that these (singular) canards persist as canard solutions, i.e. solutions corresponding to intersections of the slow manifolds $\mathcal{C}_{\varepsilon}^{a}$ and $\mathcal{C}_{\varepsilon}^{r}$ near $p$ for $\varepsilon$ small. Thus, the existence of canard solutions provides a mechanism that solutions lying in (or exponentially close to) the attracting slow manifold $\mathcal{C}_{\varepsilon}^{a}$ can be continued in (or exponentially close to) the repelling slow manifold $\mathcal{C}_{\varepsilon}^{r}$. The less counter-intuitive situation that solutions lying in the repelling slow manifold can be continued in (or close to) the attracting slow manifold is also possible. Canard solutions of this second type are often referred to as faux canards. \begin{overpic}[scale={.5}]{laypb} \put(40.0,20.0){\footnotesize$\mathcal{C}_{0}^{r}$} \put(40.0,70.0){\footnotesize$\mathcal{C}_{0}^{a}$} \put(58.0,45.0){\footnotesize$F$} \end{overpic} Figure 1: Fast dynamics in $(j,\rho)$-space for a fixed value of $\xi$. The blue curve represents $\mathcal{C}_{0}$ consisting of the two branches $\mathcal{C}_{0}^{a}$ (attracting), $\mathcal{C}_{0}^{r}$ (repelling), and the fold line $F$. The green lines indicate orbits of the layer problem (11), while the blue dot represents the line of fold points $F$. In the following we analyse the reduced flow on $F$. We will show that a canard point of folded saddle type occurs at the point $p^{\ast}=\left(\frac{1}{4},\xi^{\ast},\frac{1}{2}\right)$ (14) where $\xi^{\ast}$ is the location of the global minimum of the function $k$. The reduced problem is very simple $\displaystyle\dot{j}$ $\displaystyle=-g(\xi)j,$ (15a) $\displaystyle\dot{\xi}$ $\displaystyle=1.$ (15b) The phase space for the reduced problem is $[0,1/4]\times[0,1]$, where $j=1/4$ corresponds to the fold line. It follows from Equation (5a) that $k(\xi)\,j$ is a conserved quantity, hence the level lines of this function give the orbits of the reduced problem (15). However, as always for folded critical manifolds, the classification of the reduced flow – in particular at the fold line and at canard points – is more conveniently carried out in the variables $(\xi,\rho)$ by using the constraint $j=\rho(1-\rho)$ which defines $\mathcal{C}_{0}$. Differentiating the constraint with respect to $x$ gives $\dot{j}=(1-2\rho)\dot{\rho}$, which allows to rewrite the reduced problem as $\displaystyle\dot{\xi}$ $\displaystyle=1,$ (16) $\displaystyle(1-2\rho)\,\dot{\rho}$ $\displaystyle=-g(\xi)\,\rho(1-\rho).$ with $\xi\in[0,1]$ and $\rho\in[0,1]$. System (16) is singular at the fold line $F$, i.e. for $\rho=1/2$. This system can be desingularised by multiplying the right hand-side by $1-2\rho$ and dividing out this factor in the $\rho$ equation. This gives the desingularised reduced system $\displaystyle\dot{\xi}$ $\displaystyle=1-2\rho,$ (17) $\displaystyle\dot{\rho}$ $\displaystyle=-g(\xi)\,\rho(1-\rho).$ This multiplication of the right hand side by $(1-2\rho)$ corresponds to a position dependent rescaling of the independent variable $x$, which does not change orbits of the system away from the fold line. However, for $\rho>1/2$ the flow direction is reversed, which needs to be taken into account. We now collect the properties of the reduced problem, which are needed in the analysis of the boundary value problem (8)-(9). These properties depend on properties of the function $g=k^{\prime}/k$. Our main Assumption implies that the global nondegenerate minimum of $k$ at $\xi^{\ast}$ is a simple zero of $g$ corresponding to a saddle point $(\xi^{\ast},1/2)$ of the desingularised reduced problem. Other zeros of $g$ lead to additional equilibria, which are discussed only briefly, since we show later that these play no role in the analysis of the boundary value problem. ###### Lemma 1. The reduced problem (16) has the following properties: 1. 1. The phase portrait is symmetric with respect to the line $\rho=1/2$, which corresponds to the fold line $F$. 2. 2. The variable $\xi$ is increasing along all orbits, i.e. the flow is from left to right. 3. 3. The lines $\rho=0$ and $\rho=1$ are invariant. 4. 4. In regions with $g(\xi)>0$ the variable $\rho$ is decreasing along orbits for $1/2<\rho<1$ and is increasing for $0<\rho<1/2$. In regions with $g(\xi)<0$ this monotonicity is reversed. The variable $\rho$ is constant in regions with $g(\xi)=0$, corresponding to regions where the width of the corridor is constant. 5. 5. The line $\rho=1/2$ is a line of singularities. Points $(\xi,1/2)$ with $g(\xi)>0$ are reached in finite time by the forward flow and the derivative $\dot{\rho}$ blows up there. Similarly, points $(\xi,1/2)$ with $g(\xi)<0$ are reached in finite time by the backward flow. 6. 6. The point $p^{\ast}=\left(\xi^{\ast},\frac{1}{2}\right)$ is a canard point of folded saddle type. 7. 7. There exist two (symmetric with respect to the line $\rho=1/2$) singular canard solutions with orbits $S_{c}$ and $\tilde{S}_{c}$ passing smoothly through the singularity located at $p^{\ast}$. The canard $S_{c}$ crosses from the attracting part of the critical manifold to the repelling one, the (faux) canard $\tilde{S}_{c}$ crosses from the repelling part of the critical manifold to the attracting one. 8. 8. The (faux) canard orbit $\tilde{S}_{c}$ starts at $\xi=0$, $\rho=\rho_{c}^{0}\in(0,1/2)$ and reaches $\xi=1$ at $\rho=\rho_{c}^{1}\in(0,1/2)$. The canard orbit $S_{c}$ starts at $\xi=0$, $\rho=1-\rho_{c}^{0}\in(1/2,1)$ and reaches $\xi=1$ at $\rho=1-\rho_{c}^{1}\in(1/2,1)$. 9. 9. Solutions starting at $\xi=0$ with $\rho\in[0,\rho_{c}^{0})$ reach $\xi=1$ with $\rho\in[0,1-\rho_{c}^{1})$. Solutions starting at $\xi=0$ with $\rho\in(1-\rho_{c}^{0},1]$ reach $\xi=1$ with $\rho\in(\rho_{c}^{1},1])$. 10. 10. Solutions starting at $\xi=0$ with $\rho\in(\rho_{c}^{0},1-\rho_{c}^{0})$ do not cross the line $\xi=\xi^{\ast}$, in particular they do not reach the line $\xi=1$. Solutions reaching $\xi=1$ with $\rho\in(1-\rho_{c}^{1},\rho_{c}^{0})$ do not cross the line $\xi=\xi^{\ast}$ in backwards time, in particular they do not reach the line $\xi=0$. 11. 11. An isolated zero of $g$ at say $\xi_{0}\neq\xi^{\ast}$ corresponds to another folded singularity at $(\xi_{0},1/2)$, which is a folded saddle for $g^{\prime}(\xi_{0})>0$ and a folded center for $g^{\prime}(\xi_{0})<0$. A more degenerate zero of $g$ corresponds to a more degenerate folded singularity. If $g$ is zero on an interval $[\xi_{1},\xi_{2}]$, the density $\rho$ is constant there. In this situation $[\xi_{1},\xi_{2}]\times\\{1/2\\}$ is a line of equilibria, the endpoints of this line are again degenerate folded singularities. The properties of the reduced problem described in the Lemma are illustrated in Figure 2 for a function $k$ which satisfies $k^{\prime}<0$ in $(1,\xi^{\ast})$ and $k^{\prime}>0$ in $(\xi^{\ast},1)$. ###### Remark 2. 1. (a) The notation $S_{c}$ and $\tilde{S}_{c}$ for the canard orbits is chosen to be consistent with the notation we introduce below for other orbits of the reduced problem in the construction of singular solutions of the boundary value problem. 2. (b) The property 11. associated with additional zeros of $g$ (which may occur under our rather general main Assumption on the function $k$) are mainly included for completeness. In Remark 3 below, we show that they play no role in the construction of solutions of the boundary value problem, due to property 10. of the Lemma. ###### Proof. Properties 1.-5. follow directly from the equations. The point $p^{\ast}$ is an equilibrium for the desingularised system (17). The matrix associated with the linearisation of (17) at $p^{\ast}$ is $A:=\left(\begin{array}[]{cc}0&-2\\\\[2.84526pt] -\frac{g^{\prime}(\xi^{\ast})}{4}&0\end{array}\right).$ (18) Since $g^{\prime}(\xi^{\ast})=k^{\prime\prime}(\xi^{\ast})/k(\xi^{\ast})$, the assumption $k^{\prime\prime}(\xi^{\ast})>0$ translates into $g^{\prime}(\xi^{\ast})>0$. This gives $\det A=-\frac{g(\xi^{\ast})}{2}<0$, hence $p^{\ast}$ is a saddle point for (17) with associated smooth stable and unstable manifolds. For the reduced problem (16) – with the flow direction reversed for $\rho>1/2$ – the point $p^{\ast}$ is a folded saddle [15]. Due to a cancellation of a simple zero on both sides of the $\rho$-equation in (16), the stable manifold of the saddle is now the (faux) canard $\tilde{S}_{c}$, corresponding to a smooth solution passing through through the point $p^{\ast}$. Similarly, the unstable manifold of the saddle becomes the canard $S_{c}$. This proves properties 6. and 7. The conserved quantity $k(\xi)j$ of equation (15a) translates into the conserved quantity $H(\xi,\rho)=k(\xi)\rho(1-\rho)>0$ (19) of the desingularised system (17), i.e. the level lines of $H$ give the phase portrait. The canard orbits $S_{c}$ and $\tilde{S}_{c}$ are the level lines $H(\xi,\rho)=\frac{k(\xi^{\ast})}{4}$. Since $k$ has its global minimum at $\xi^{\ast}$, the canard orbits cannot intersect the (fold) line $\rho=1/2$. Since in addition, the canard orbits cannot intersect the lines $\rho=0$, $\rho=1$ where $H=0$, the canard orbits extend to $\xi=0$ and $\xi=1$. Thus assertion 8. follows, with $\rho_{c}^{0}\in(0,1/2)$ and $\rho_{c}^{1}\in(1/2,1)$ defined as the solutions of the equations $H(0,\rho_{c}^{0})=\frac{k(\xi^{\ast})}{4},\qquad H(1,\rho_{c}^{1})=\frac{k(\xi^{\ast})}{4}.$ The solutions described in Assertion 9. lie on level lines with $H(\xi,\rho)>H(\xi^{\ast},1/2)$, the solutions described in Assertion 10. lie on level lines with $H(\xi,\rho)<H(\xi^{\ast},1/2)$. Together with 8. this implies 9. and 10. ∎ The canard $S_{c}$ and the (faux) canard $\tilde{S}_{c}$ on $\mathcal{C}_{0}$ can be described as graphs by means of the following functions $\displaystyle\rho_{c}^{+}(\xi)$ $\displaystyle:=\frac{1}{2}\left(1+\sqrt{1-\frac{k\left(\xi^{\ast}\right)}{k(\xi)}}\right),$ (20a) $\displaystyle\rho_{c}^{-}(\xi)$ $\displaystyle:=\frac{1}{2}\left(1-\sqrt{1-\frac{k(\xi^{\ast})}{k(\xi)}}\right),$ (20b) as follows $\displaystyle S_{c}$ $\displaystyle:=\left\\{(\xi,\rho)\,:\,0\leq\xi\leq\xi^{\ast},\,\rho=\rho_{c}^{+}(\xi)\right\\}\cup\left\\{(\xi,\rho)\,:\,\xi^{\ast}\leq\xi\leq 1,\,\rho=\rho_{c}^{-}(\xi)\right\\},$ (21) $\displaystyle\tilde{S}_{c}$ $\displaystyle:=\left\\{(\xi,\rho)\,:\,0\leq\xi\leq\xi^{\ast},\,\rho=\rho_{c}^{-}(\xi)\right\\}\cup\left\\{(\xi,\rho)\,:\,\xi^{\ast}\leq\xi\leq 1,\,\rho=\rho_{c}^{+}(\xi)\right\\}.$ The values $\rho_{c}^{0}$ and $\rho_{c}^{1}$ introduced in Lemma 1 (corresponding to the $\rho$-values of the (faux) canard) at $\xi=0$ and $\xi=1$, respectively, are then given by $\rho_{c}^{0}=\rho_{c}^{-}(0),\qquad\rho_{c}^{1}=\rho_{c}^{+}(1).$ (22) The points of the canard $S_{c}$ corresponding to $\xi=0$ and $\xi=1$ in $(j,\xi,\rho)$-space which play an important role in the following analysis are $\displaystyle p_{c}^{0}$ $\displaystyle:=\left(\rho_{c}^{0}(1-\rho_{c}^{0}),0,1-\rho_{c}^{0}\right),$ (23) $\displaystyle p_{c}^{1}$ $\displaystyle:=\left(\rho_{c}^{1}(1-\rho_{c}^{1}),1,1-\rho_{c}^{1}\right).$ ###### Remark 3. The function $g$ may have zeros $\xi\neq\xi^{\ast}$. All these points $(\xi,1/2)$ are equilibria of the desingularised system (17) but these equilibria and possible canard solutions associated with them are confined to the open region $\mathcal{N}$ bounded by $\tilde{S}_{c}$ from below and by $S_{c}$ from above for $\xi<\xi^{\ast}$, and by $S_{c}$ from below and by $\tilde{S}_{c}$ from above for $\xi>\xi^{\ast}$ (see Fig. 2). Since no transitions from $\xi=0$ to $\xi=1$ are possible through the region $\mathcal{N}$, it plays no role in the construction of solutions of the boundary value problem. Since other folded singularities associated with local minima or maxima of $k$ and their associated canard solutions are confined to $\mathcal{N}$ these also play no role for boundary value problem. \begin{overpic}[scale={0.7}]{red_flow_new_shade.pdf} \put(1.0,2.0){\small{$0$}} \put(61.0,2.0){$\xi^{\ast}$} \put(-2.0,17.0){$\rho_{c}^{0}$} \put(-8.0,70.0){$1-\rho_{c}^{0}$} \put(83.0,27.0){$1-\rho_{c}^{1}$} \put(83.0,61.0){$\rho_{c}^{1}$} \put(55.0,53.0){$S_{c}$} \put(55.0,33.0){$\tilde{S}_{c}$} \end{overpic} Figure 2: Illustration of the reduced flow associated to Equations (8)-(9) described in Lemma 1 for $k(\xi)=1+a\,\mathrm{cos}\left(\frac{2\pi\xi}{b}\right)$ and $a=0.3$, $b=1.5$. The solid gray line indicates the line of fold points $F$ (see (13)), whereas the blue dot corresponds to the canard point of folded saddle type $p^{\ast}$. The cyan curves correspond to the canard $S_{c}$ (solid line) and the (faux) canard $\tilde{S}_{c}$ (dashed line). The shaded red area represents the region $\mathcal{N}$ defined in Remark 3, which plays no role in the construction of solutions of the boundary value problem (8)-(9). We now begin the construction of solutions of the boundary value problem (8)-(9) by combining solutions of the reduced problem with solutions of the layer problem in such a way that the boundary conditions are satisfied. Here it is important to keep in mind that solutions can jump from points on the repelling branch $\mathcal{C}_{0}^{r}$ of the critical manifold to the attracting branch $\mathcal{C}_{0}^{a}$, but not vice versa. In [7] we have constructed singular solutions in the case of a closing channel using a shooting strategy: we evolved the manifold of boundary conditions at $\xi=0$ forward and checked whether it intersected the manifold of boundary conditions at $\xi=1$. This constructive procedure allowed to identify the initial and final values of $\rho$ (namely $\rho_{0}$ and $\rho_{1}$) for $\varepsilon=0$. In the bottleneck scenario, however, the presence of a canard point lying in the interior of the spatial domain $[0,1]$ implies that singular orbits containing segments of the canards $S_{c}$ or $\tilde{S}_{c}$ can make slow transitions between the branches of the critical manifold. Most importantly, this allows transitions from the attracting branch back to the repelling branch. We will show that this leads to the new type of transitional profiles, described in the introduction. Due to the special role of the canard point $p^{\ast}$ we modify the shooting strategy by evolving also the manifold of boundary conditions at $\xi=1$ (backwards) and checking the intersection with the forward evolution of the manifold of left boundary conditions at $\xi=\xi^{\ast}$, where the canard point $p^{\ast}$ lies. In the dynamical systems framework, boundary conditions (9) correspond to two lines in the $(j,\xi,\rho)$-space, satisfying $j=\alpha(1-\rho)$ at $\xi=0$ and $j=\beta\rho$ at $\xi=1$, respectively. However, due to the fast-slow structure, the set of admissible boundary conditions is restricted to (see Figures 3-4) $\displaystyle\mathcal{L}$ $\displaystyle:=\left\\{\left(\alpha(1-s),\,0,\,s\right)\ :\ \rho_{\alpha}\leq s\leq 1\right\\},$ (24a) $\displaystyle\mathcal{R}$ $\displaystyle:=\left\\{\left(\beta t,\,1,\,t\right)\ :\ 0\leq t\leq\rho_{\beta}\right\\}.$ (24b) Here $\rho_{\alpha}=\begin{cases}\alpha&\quad\text{if }\alpha\leq\frac{1}{2},\\\ 1-\frac{1}{4\alpha}&\quad\text{if }\alpha\geq\frac{1}{2},\end{cases}$ (25) and $\rho_{\beta}=\begin{cases}1-\beta&\quad\text{if }\beta\leq\frac{1}{2},\\\ \frac{1}{4\beta}&\quad\text{if }\beta\geq\frac{1}{2}.\end{cases}$ (26) The lower and upper bounds $\rho_{\alpha}$ and $\rho_{\beta}$ for the density $\rho$ are caused by the fast-slow structure of the flow: if we would consider a starting point $(\alpha(1-\rho),\,0,\,\rho)$ with $0\leq\rho<\rho_{\alpha}$, the orbit would be immediately repelled to infinity from $\mathcal{C}_{0}$, hence connecting to the boundary conditions at $\xi=1$ is impossible. Analogously, points satisfying $(\beta\rho,\,1,\,\rho)$ with $\rho_{\beta}<\rho\leq 1$ cannot be endpoints of the singular orbits, since they are repelling for the layer problem. Thus, the initial and final points of the singular orbits – $p_{0}$ and $p_{1}$, respectively – must satisfy $p_{0}\in\mathcal{L},\textrm{ and }p_{1}\in\mathcal{R}.$ (27) The manifold $\mathcal{L}$ intersects with $\mathcal{C}_{0}$ at $(0,0,1)$ and $l=(\alpha(1-\alpha),0,\alpha),$ (28) while $\mathcal{R}$ intersects with $\mathcal{C}_{0}$ at $(0,1,0)$ and $r=(\beta(1-\beta),1,1-\beta).$ (29) For $\varepsilon=0$, the variable $\xi$ evolves only on $\mathcal{C}_{0}$ according to the reduced flow (16). Therefore, in order for the singular solution to evolve from $\xi=0$ to $\xi=1$, we must connect $\mathcal{L}$ and $\mathcal{R}$ to $\mathcal{C}_{0}$. The points $l$ and $r$ already belong to $\mathcal{C}_{0}$. Other points on $\mathcal{L}$ and $\mathcal{R}$ can reach $\mathcal{C}_{0}$ using the layer problem (11). Tracking the evolution of $\mathcal{L}$ by means of the layer problem at $\xi=0$ until it reaches $\mathcal{C}_{0}$, and analogously the evolution of $\mathcal{R}$ backwards until the layer problem at $\xi=1$ intersects $\mathcal{C}_{0}$, yields two sets (shown in Figures 3-4): $\displaystyle\mathcal{L}^{+}$ $\displaystyle:=\left\\{\left(\alpha(1-s),\,0,\,\rho^{+}(0,s)\right)\ :\ \rho_{\alpha}\leq s\leq 1\right\\},$ (30a) $\displaystyle\mathcal{R}^{-}$ $\displaystyle:=\left\\{\left(\beta t,\,1,\,\rho^{-}(1,t)\right)\ :\ 0\leq t\leq\rho_{\beta}\right\\}.$ (30b) In the following, we use the symbol $\rho^{+}(0,s)$ to indicate the $\rho$-value (greater than or equal to $\frac{1}{2}$) reached by the point $(\alpha(1-s),\,0,\,s)$ after its transition from $\mathcal{L}$ to $\mathcal{C}_{0}$ by means of the layer problem. If the solution $(\xi,\rho)$ of the reduced flow 16 starting at $(0,\rho^{+}(0,s))$ reaches $\xi=\xi^{\ast}$, we denote its value of $\rho$ at $\xi=\xi^{\ast}$ by $\rho^{+}\left(\xi^{\ast},s\right)$. In an analogous manner, we introduce the symbol $\rho^{-}(1,t)$ to indicate the $\rho$-value (less than or equal to $\frac{1}{2}$) reached by the point $(\beta t,\,1,\,t)$ after its transition from $\mathcal{R}$ to $\mathcal{C}_{0}$ by means of the layer problem. If the solution $(\xi,\rho)$ of the reduced flow starting at $(1,\rho^{-}(1,t))$ and flowing backwards reaches $\xi=\xi^{\ast}$, we denote its value of $\rho$ at $\xi=\xi^{\ast}$ by $\rho^{-}\left(\xi^{\ast},t\right)$. When $\alpha<\frac{1}{2}$, the reduced flow can either start on $\mathcal{L}^{+}$ or at $l$, while for $\alpha>\frac{1}{2}$ it must start on $\mathcal{L}^{+}$. Analogously, when $\beta<\frac{1}{2}$, the reduced flow can either end on $\mathcal{R}^{-}$ or at $r$, while for $\beta>\frac{1}{2}$ it must end on $\mathcal{R}^{-}$. (a) (b) Figure 3: Schematic representation of $\mathcal{L}$ (orange line) and $\mathcal{L}^{+}$ (orange curve) for (a) $0<\alpha<\frac{1}{2}$ and (b) $\frac{1}{2}<\alpha<1$. The orange dot corresponds to $l$, the blue curve represents $\mathcal{C}_{0}$, and the green lines correspond to the orbits of the layer problem. (a) (b) Figure 4: Schematic representation of $\mathcal{R}$ (purple line) and $\mathcal{R}^{-}$ (purple curve) for (a) $0<\beta<\frac{1}{2}$ and (b) $\frac{1}{2}<\beta<1$. The purple dot corresponds to $r$, the blue curve represents $\mathcal{C}_{0}$, and the green lines correspond to the orbits of the layer problem. Based on this geometric interpretation of the boundary conditions, we proceed with the construction of the singular orbits by connecting $\mathcal{L}^{+}\cup l$ and $\mathcal{R}^{-}\cup r$ by means of the reduced flow (16) on $\mathcal{C}_{0}$. In doing so, we first let $\mathcal{L}^{+}$ in (30a) flow forward and $\mathcal{R}^{-}$ in (30b) flow backwards by means of the reduced flow until $\xi=\xi^{\ast}$: we call the corresponding sets $\mathcal{L}_{\xi^{\ast}}^{+}$ and $\mathcal{R}_{\xi^{\ast}}^{-}$, respectively. $\displaystyle\mathcal{L}_{\xi^{\ast}}^{+}$ $\displaystyle:=\left\\{\left(\rho^{+}\left(\xi^{\ast},s\right)\left(1-\rho^{+}\left(\xi^{\ast},s\right)\right),\,\xi^{\ast},\,\rho^{+}\left(\xi^{\ast},s\right)\right)\ :\ \rho_{\alpha}\leq s\leq 1\right\\},$ (31a) $\displaystyle\mathcal{R}_{\xi^{\ast}}^{-}$ $\displaystyle:=\left\\{\left(\rho^{-}\left(\xi^{\ast},t\right)\left(1-\rho^{-}\left(\xi^{\ast},t\right)\right),\,\xi^{\ast},\,\rho^{-}\left(\xi^{\ast},t\right)\right)\ :\ 0\leq t\leq\rho_{\beta}\right\\}.$ (31b) If $\alpha\geq\frac{1}{2}$ then $l\in\mathcal{L}^{+}$, and the evolution of $l$ by means of the reduced flow is already included in $\mathcal{L}_{\xi^{\ast}}^{+}$. If $\alpha<\frac{1}{2}$ then $l\notin\mathcal{L}^{+}$, and therefore the corresponding point at $\xi=\xi^{\ast}$ must be defined separately as $l_{\xi^{\ast}}:=\left\\{\left(\rho^{+}\left(\xi^{\ast},\alpha\right)\left(1-\rho^{+}\left(\xi^{\ast},\alpha\right)\right),\,\xi^{\ast},\,1-\rho^{+}\left(\xi^{\ast},\alpha\right)\right)\right\\}.$ (32) Analogously, if $\beta\geq\frac{1}{2}$ then $r\in\mathcal{R}^{-}$, and the backwards evolution of $r$ by means of the reduced flow is already included in $\mathcal{R}_{\xi^{\ast}}^{-}$. If $\beta<\frac{1}{2}$, however, $r\notin\mathcal{R}^{-}$, and therefore the corresponding point at $\xi=\frac{1}{2}$ must be defined separately as $r_{\xi^{\ast}}:=\left\\{\left(\rho^{-}\left(\xi^{\ast},1-\beta\right)\left(1-\rho^{-}\left(\xi^{\ast},1-\beta\right)\right),\,\xi^{\ast},\,1-\rho^{-}\left(\xi^{\ast},1-\beta\right)\right)\right\\}.$ (33) We note that the point $l_{\xi^{\ast}}$ exists if and only if $\alpha\leq\rho_{c}^{0}$; analogously, the point $r_{\xi^{\ast}}$ exists if and only if $\beta\leq 1-\rho_{c}^{1}$ (see Remark 3). A singular orbit is then given by matching the slow and fast pieces obtained by investigating the reduced and layer problems, respectively. More specifically, a singular orbit exists if and only if the intersection between the sets $\mathcal{L}_{\xi^{\ast}}^{+}\cup l_{\xi^{\ast}}$ and $\mathcal{R}_{\xi^{\ast}}^{-}\cup r_{\xi^{\ast}}$ is non-empty, and it is unique if this intersection consists of one point. In addition to the canards $S_{c}$ and $\tilde{S}_{c}$ introduced above, our analysis of the existence and structure of singular orbits is based on four special orbits $S_{\alpha}$, $S_{\beta}$, $\tilde{S}_{\alpha}$, $\tilde{S}_{\beta}$ of the reduced flow (see Figure 5): * • The orbit $S_{\alpha}$, defined for each $\alpha\in(0,1)$, is the one starting at $\rho=\alpha$ at $\xi=0$. For $\alpha<\rho_{c}^{0}$ or $\alpha>1-\rho_{c}^{0}$, the corresponding final value of $\rho$ at $\xi=1$ is denoted by $\rho^{\ast}(\alpha)$. For $\rho_{c}^{0}<\alpha<1-\rho_{c}^{0}$, $S_{\alpha}$ ends on the fold line $F$ and hence $S_{\alpha}\subset\mathcal{N}$. For $\alpha=\rho_{c}^{0}$ or $\alpha=1-\rho_{c}^{0}$, $S_{\alpha}$ ends on the canard point $p^{\ast}$ at $\xi=\xi^{\ast}$, and its continuation for $\xi\in[\xi^{\ast},1]$ is therefore not uniquely defined. * • The orbit $S_{\beta}$, defined for each $\beta\in(0,1)$, is the one ending at $\rho=1-\beta$ at $\xi=1$. For $\beta<1-\rho_{c}^{1}$ or $\beta>\rho_{c}^{1}$, the corresponding initial value of $\rho$ at $\xi=0$ is denoted by $\rho_{\ast}(\beta)$. For $1-\rho_{c}^{1}<\beta<\rho_{c}^{1}$, $S_{\beta}$ ends on the fold line $F$ and hence $S_{\beta}\subset\mathcal{N}$. For $\beta=1-\rho_{c}^{1}$ or $\beta=\rho_{c}^{1}$, $S_{\beta}$ ends on the canard point $p^{\ast}$ at $\xi=\xi^{\ast}$ backward in $\xi$, and its continuation for $\xi\in[0,\xi^{\ast}]$ is therefore not uniquely defined. * • For $i=\alpha,\beta$, we define $\tilde{S}_{i}$ as the reflection of the orbit $S_{i}$ with respect to $\rho=\frac{1}{2}$. Depending on the values of $\alpha$ and $\beta$, one of the orbits $S_{i}$, $\tilde{S}_{i}$, $i=c,\alpha,\beta$, corresponds to the slow part of the singular orbits we will construct. Changing $\alpha$ and $\beta$ influences the orbits $S_{\alpha}$, $\tilde{S}_{\alpha}$ and $S_{\beta}$, $\tilde{S}_{\beta}$. We will show in the following that the $\alpha$, $\beta$ dependent mutual position of these orbits determines the type of singular solution of the boundary value problem. \begin{overpic}[scale={0.7}]{spec_sol_2.pdf} \put(0.0,0.0){$0$} \put(68.0,0.0){$\xi^{\ast}$} \put(90.0,0.0){$1$} \put(100.0,0.0){\Large$\xi$} \put(-11.0,8.0){\footnotesize$1-\rho_{\ast}(\beta)$} \put(0.0,14.0){\footnotesize$\alpha$} \put(0.0,19.0){\footnotesize$\rho_{c}^{0}$} \put(0.0,47.0){$\frac{1}{2}$} \put(-6.0,76.0){\footnotesize$1-\rho_{c}^{0}$} \put(-5.0,80.0){\footnotesize$1-\alpha$} \put(-5.0,85.0){\footnotesize$\rho_{\ast}(\beta)$} \put(0.0,90.0){$1$} \put(0.0,100.0){\Large$\rho$} \put(92.0,10.0){\footnotesize$\beta$} \put(92.0,22.0){\footnotesize$\rho^{\ast}(\alpha)$} \put(92.0,29.0){\footnotesize$1-\rho_{c}^{1}$} \put(92.0,65.0){\footnotesize$\rho_{c}^{1}$} \put(92.0,73.0){\footnotesize$1-\rho^{\ast}(\alpha)$} \put(92.0,83.0){\footnotesize$1-\beta$} \put(60.0,35.0){$\tilde{S}_{c}$} \put(60.0,58.0){$S_{c}$} \put(55.0,20.0){$S_{\alpha}$} \put(55.0,73.0){$\tilde{S}_{\alpha}$} \put(72.0,85.0){$S_{\beta}$} \put(72.0,8.0){$\tilde{S}_{\beta}$} \end{overpic} Figure 5: Schematic illustration of the special orbits $S_{c}$, $S_{\alpha}$, $S_{\beta}$, $\tilde{S}_{c}$, $\tilde{S}_{\alpha}$, $\tilde{S}_{\beta}$ in $(\xi,\rho)$-space in the case $0<\alpha<\rho_{c}^{0}$ and $\rho_{c}^{1}<\beta<1$. Here $k(\xi)=1+a\,\mathrm{cos}\left(\frac{2\pi\xi}{b}\right)$ with $a=0.3$, $b=1.5$. The orbit $S_{c}$ (solid cyan curve) connects $(0,1-\rho_{c}^{0})$ and $(1,1-\rho_{c}^{1})$, the orbit $S_{\alpha}$ (solid magenta curve) connects $(0,\alpha)$ and $(1,\rho^{\ast}(\alpha))$, and the orbit $S_{\beta}$ (solid brown curve) connects $(0,\rho_{\ast}(\beta))$ and $(1,1-\beta)$ with $\alpha,\beta<\frac{1}{2}$ as in (34). The dashed curves represent the orbits $\tilde{S}_{c}$ (cyan), $\tilde{S}_{\alpha}$ (magenta), $\tilde{S}_{\beta}$ (brown), which are symmetric to the corresponding solid ones $S_{c}$, $S_{\alpha}$, $S_{\beta}$ with respect to $\rho=\frac{1}{2}$. We note that the orbits $S_{\alpha}$, $\tilde{S}_{\alpha}$ and $S_{\beta}$, $\tilde{S}_{\beta}$ switch position in the above diagram as $\alpha$, $\beta>\frac{1}{2}$. By using the conserved quantity (19) the respective values of $\rho^{\ast}(\alpha)$ and $\rho_{\ast}(\beta)$ can be computed explicitly: $\displaystyle\rho^{\ast}(\alpha)$ $\displaystyle:=\left\\{\begin{array}[]{ll}\frac{1}{2}\left(1-\sqrt{1-4\alpha(1-\alpha)\frac{k(0)}{k(1)}}\right)\text{ if }\alpha<\rho_{c}^{0},\\\ \frac{1}{2}\left(1+\sqrt{1-4\alpha(1-\alpha)\frac{k(0)}{k(1)}}\right)\text{ if }\alpha>1-\rho_{c}^{0},\end{array}\right.$ (34c) $\displaystyle\rho_{\ast}(\beta)$ $\displaystyle:=\left\\{\begin{array}[]{ll}\frac{1}{2}\left(1+\sqrt{1-4\beta(1-\beta)\frac{k(1)}{k(0)}}\right)\text{ if }\beta<1-\rho_{c}^{1},\\\ \frac{1}{2}\left(1-\sqrt{1-4\beta(1-\beta)\frac{k(1)}{k(0)}}\right)\text{ if }\beta>\rho_{c}^{1}.\\\ \end{array}\right.$ (34f) Note that $\alpha=1-\rho_{\ast}(\beta)$ is equivalent to $\beta=\rho^{\ast}(\alpha)$. Based on this, we divide the $(\alpha,\beta)$-parameter space into eight regions $\mathcal{G}_{i}$, $i=1,\dots,8$ defined via the following curves $\gamma_{ij}$ (here the indices refer to the adjacent regions): $\displaystyle\gamma_{12}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;0<\alpha\leq\rho_{c}^{0},\,\beta=1-\rho^{\ast}(\alpha)\right\\},$ (35a) $\displaystyle\gamma_{13}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\alpha=\rho_{c}^{0},\,1-\rho_{c}^{1}\leq\beta\leq\rho_{c}^{1}\right\\},$ (35b) $\displaystyle\gamma_{17}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\alpha=1-\rho_{\ast}(\beta),\,0<\beta\leq 1-\rho_{c}^{1}\right\\},$ (35c) $\displaystyle\gamma_{24}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\alpha=\rho_{c}^{0},\,\rho_{c}^{1}\leq\beta<1\right\\},$ (35d) $\displaystyle\gamma_{34}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\rho_{c}^{0}\leq\alpha\leq 1-\rho_{c}^{0},\,\beta=\rho_{c}^{1}\right\\},$ (35e) $\displaystyle\gamma_{35}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\alpha=1-\rho_{c}^{0},\,1-\rho_{c}^{1}\leq\beta\leq\rho_{c}^{1}\right\\},$ (35f) $\displaystyle\gamma_{37}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\rho_{c}^{0}\leq\alpha\leq 1-\rho_{c}^{0},\,\beta=1-\rho_{c}^{1}\right\\},$ (35g) $\displaystyle\gamma_{46}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\alpha=1-\rho_{c}^{0},\,\rho_{c}^{1}\leq\beta<1\right\\},$ (35h) $\displaystyle\gamma_{56}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;1-\rho_{c}^{0}\leq\alpha<1,\,\beta=\rho_{c}^{1}\right\\},$ (35i) $\displaystyle\gamma_{58}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;1-\rho_{c}^{0}\leq\alpha<1,\,\beta=1-\rho_{c}^{1}\right\\},$ (35j) $\displaystyle\gamma_{78}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\alpha=\rho_{\ast}(\beta),\,0<\beta\leq 1-\rho_{c}^{1}\right\\}.$ (35k) The above curves correspond to situations where some of the orbits $S_{i}$, $\tilde{S}_{i}$, $i=1,2,3$ defined above coincide. In particular: * • for $(\alpha,\beta)\in\gamma_{12}$, we have $S_{\alpha}=S_{\beta}$ (lying in $\mathcal{C}_{0}^{r}$); * • for $(\alpha,\beta)\in\gamma_{13}\cup\gamma_{24}$, we have $\tilde{S}_{c}=S_{\alpha}$ for $\xi\in[0,\xi^{\ast}]$ (i.e. up to the canard point $p^{\ast}$); * • for $(\alpha,\beta)\in\gamma_{17}$, we have $S_{\alpha}=\tilde{S}_{\beta}$; * • for $(\alpha,\beta)\in\gamma_{34}\cup\gamma_{56}$, we have $S_{c}=S_{\beta}$ for $\xi\in[\xi^{\ast},1]$ (i.e. up to the canard point $p^{\ast}$); * • for $(\alpha,\beta)\in\gamma_{35}\cup\gamma_{46}$, we have $S_{c}=S_{\alpha}$ for $\xi\in[0,\xi^{\ast}]$ (i.e. up to the canard point $p^{\ast}$); * • for $(\alpha,\beta)\in\gamma_{37}\cup\gamma_{58}$, we have $\tilde{S}_{c}=S_{\beta}$ for $\xi\in[\xi^{\ast},1]$ (i.e. up to the canard point $p^{\ast}$); * • for $(\alpha,\beta)\in\gamma_{78}$, we have $S_{\alpha}=S_{\beta}$ (lying in $\mathcal{C}_{0}^{a}$). ###### Remark 4. Whenever two orbits coincide, their symmetric reflections with respect to $\rho=\frac{1}{2}$ coincide as well. The eleven curves in (35) split $(0,1)^{2}$ into $8$ regions $\mathcal{G}_{i}$, $i=1,\dots,8$ (shown in Figure 6): $\displaystyle\mathcal{G}_{1}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;0<\alpha<\rho_{c}^{0},\,\rho^{\ast}(\alpha)<\beta<1-\rho^{\ast}(\alpha)\right\\}$ (36a) $\displaystyle\mathcal{G}_{2}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;0<\alpha<\rho_{c}^{0},\,1-\rho^{\ast}(\alpha)<\beta<1\right\\},$ (36b) $\displaystyle\mathcal{G}_{3}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\rho_{c}^{0}<\alpha<1-\rho_{c}^{0},\,1-\rho_{c}^{1}<\beta<\rho_{c}^{1}\right\\},$ (36c) $\displaystyle\mathcal{G}_{4}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\rho_{c}^{0}<\alpha<1-\rho_{c}^{0},\,\rho_{c}^{1}<\beta<1\right\\},$ (36d) $\displaystyle\mathcal{G}_{5}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;1-\rho_{c}^{0}<\alpha<1,\,1-\rho_{c}^{1}<\beta<\rho_{c}^{1}\right\\},$ (36e) $\displaystyle\mathcal{G}_{6}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;1-\rho_{c}^{0}<\alpha<1,\,\rho_{c}^{1}<\beta<1\right\\},$ (36f) $\displaystyle\mathcal{G}_{7}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;1-\rho_{\ast}(\beta)<\alpha<\rho_{\ast}(\beta),\,0<\beta<1-\rho_{c}^{1}\right\\},$ (36g) $\displaystyle\mathcal{G}_{8}$ $\displaystyle:=\left\\{(\alpha,\beta)\ :\;\rho_{\ast}(\beta)<\alpha<1,\,0<\beta<1-\rho_{c}^{1}\right\\}.$ (36h) In short terms, moving from one region to the other in the $(\alpha,\beta)$ -parameter space leads to a corresponding change in the structure of the singular solutions. Figure 6: Representation of the $(\alpha,\beta)$ bifurcation diagram for $\varepsilon=0$. Here $k(\xi)=1+a\,\mathrm{cos}\left(\frac{2\pi\xi}{b}\right)$ with $a=0.3$, $b=1.5$. Red regions correspond to high density, blue regions to low density, and green regions to transitions from high to low density regimes. In the insets, the density $\rho$ is shown as a function of $\xi$. The blue parts correspond to solutions of the reduced problem (16), whereas the green parts indicate boundary layers. The gray line represents $\rho=\frac{1}{2}$ We will show (in Proposition 1) that within each of those region the structure of the singular solutions is the same. Note that our construction of singular solutions works also on all the boundary curves defined in (35) except for $\gamma_{17}$, $\gamma_{13}\cup\gamma_{24}$, and $\gamma_{37}\cup\gamma_{58}$, where singular solutions are not unique (see Remark 8). To this aim, we introduce the following eight types of singular solutions (see Figure 8-9): Type 1. Singular solutions which start on $\mathcal{C}_{0}^{r}$ at $\xi=0$, follow the reduced flow on $\mathcal{C}_{0}^{r}$ (where $\rho$ increases), and have a layer at $\xi=1$ in which $\rho$ increases. Type 2. Singular solutions which start on $\mathcal{C}_{0}^{r}$ at $\xi=0$, follow the reduced flow on $\mathcal{C}_{0}^{r}$ (where $\rho$ increases), and have a layer at $\xi=1$ in which $\rho$ decreases. Type 3. Singular solutions which have a layer at $\xi=0$ in which $\rho$ increases, follow the reduced flow on $\mathcal{C}_{0}$ (where $\rho$ decreases) passing through the point $p^{\ast}$, and have another layer at $\xi=1$ in which $\rho$ increases. Type 4. Singular solutions which have a layer at $\xi=0$ in which $\rho$ increases, follow the reduced flow on $\mathcal{C}_{0}$ (where $\rho$ decreases) passing through the point $p^{\ast}$, and have another layer at $\xi=1$ in which $\rho$ decreases. Type 5. Singular solutions which have a layer at $\xi=0$ in which $\rho$ decreases, follow the reduced flow on $\mathcal{C}_{0}$ (where $\rho$ decreases) passing through the point $p^{\ast}$, and have another layer at $\xi=1$ in which $\rho$ increases. Type 6. Singular solutions which have a layer at $\xi=0$ in which $\rho$ decreases, follow the reduced flow on $\mathcal{C}_{0}$ (where $\rho$ decreases) passing through the point $p^{\ast}$, and have another layer at $\xi=1$ in which $\rho$ decreases. Type 7. Singular solutions which have a layer at $\xi=0$ in which $\rho$ increases, and follow the reduced flow on $\mathcal{C}_{0}^{a}$ (where $\rho$ decreases). Type 8. Singular solutions which have a layer at $\xi=0$ in which $\rho$ decreases and follow the reduced flow on $\mathcal{C}_{0}^{a}$ (where $\rho$ decreases). More details about the construction and structure of these singular orbits are given in the proof of the following proposition. ###### Proposition 1. Let $k\in C^{2}([0,1])$ be a positive function satisfying Assumption (6). Then for each $(\alpha,\beta)\in\mathcal{G}_{i}$, $i=1,\ldots,8$ there exists a unique singular solution $\Gamma^{i}$ of type $i$ to (5b) composed of segments of orbits of the layer problem (11) and the reduced problem (16) satisfying the boundary conditions. ###### Proof. The proof is based on the shooting technique outlined above. Technically speaking, we show that the intersection of the sets $\mathcal{L}_{\xi^{\ast}}^{+}\cup l_{\xi^{\ast}}$ in (31a)-(32) and $\mathcal{R}_{\xi^{\ast}}^{-}\cup r_{\xi^{\ast}}$ in (31b)-(33) is non-empty, and in particular consists of one point. This gives us the unique values of $\rho_{0},\,\rho_{1}$ for which a singular orbit exists depending on $\alpha$ and $\beta$, which in turn allows us to identify the eight types of singular solutions corresponding to the eight regions defined in (36). While we claim the existence of singular solutions only in the open regions $\Gamma^{i}$, $i=1,\ldots,8$ we also comment on the singular configurations where $(\alpha,\beta)$ lies on the curves $\gamma_{ij}$ from (35). In principle there are four possible ways for the intersection between $\mathcal{L}_{\xi^{\ast}}^{+}\cup l_{\xi^{\ast}}$ and $\mathcal{R}_{\xi^{\ast}}^{-}\cup r_{\xi^{\ast}}$ to occur; one of these defines four possible profiles corresponding to four regions in $(\alpha,\beta)$-parameter space, two of these lead to two possible profiles corresponding to two regions in $(\alpha,\beta)$-parameter space, while the fourth case ($l_{1}\cap r$) leads to an empty intersection, since $l_{\xi^{\ast}}$ and $r_{\xi^{\ast}}$ are separated from $\mathcal{L}_{\xi^{\ast}}^{+}$ and $\mathcal{R}_{\xi^{\ast}}^{-}$, respectively, only for $\alpha\leq\rho_{c}^{0}$ and $\beta\leq 1-\rho_{c}^{1}$, and in this case they can never coincide. Thus, we are left with: Case 1: $l_{\xi^{\ast}}\cap\mathcal{R}_{\xi^{\ast}}^{-}\neq\emptyset$. From the investigation of this case we obtain orbits of type 1, 2. Case 2: $\mathcal{L}_{\xi^{\ast}}^{+}\cap\mathcal{R}_{\xi^{\ast}}^{-}\neq\emptyset$. From the investigation of this case we obtain orbits of type 3, 4, 5, 6. Case 3: $\mathcal{L}_{\xi^{\ast}}^{+}\cap r_{\xi^{\ast}}\neq\emptyset$. From the investigation of this case we obtain orbits of type 7, 8. In the following, we examine Cases 1-3 in more detail. _Case 1: $l_{\xi^{\ast}}\,\cap\,\mathcal{R}_{\xi^{\ast}}^{-}\neq\emptyset$._ By definition of $l_{\xi^{\ast}}$, this occurs only when $\alpha\leq\rho_{c}^{0}$. In this case, we have $l_{\xi^{\ast}}\in\mathcal{R}_{\xi^{\ast}}^{-}$, which implies that $p_{0}=l$ and, consequently, $\rho_{0}=\alpha$. This implies that in this regime no boundary layers exist at $\xi=0$. Moreover, since $\rho(1,s)=\rho^{\ast}(\alpha)$, following the flow of the layer problem until it hits $\mathcal{R}$ we obtain $\rho_{1}=\frac{\alpha(1-\alpha)k(0)}{\beta k(1)}.$ (37) In this case, the singular orbit consists in a slow motion along $\mathcal{C}_{0}^{r}$ followed by a layer at $\xi=1$. The nature of this layer – in particular its orientation – depends on $\alpha$ and $\beta$ as follows: * • When $\alpha<\rho_{c}^{0}$ and $\rho^{\ast}(\alpha)<\beta<1-\rho^{\ast}(\alpha)$, i.e. for $(\alpha,\beta)\in\mathcal{G}_{1}$, $\rho$ increases along the boundary layer at $\xi=1$. The corresponding singular solution is therefore of type $1$ (see Figure 8(a)). * • When $\alpha<\rho_{c}^{0}$ and $\beta>1-\rho^{\ast}(\alpha)$, i.e. for $(\alpha,\beta)\in\mathcal{G}_{2}$, $\rho$ decreases along the boundary layer at $\xi=1$. Therefore, the corresponding singular solution is of type $2$ (see Figure 8(b)). We note that when $\alpha<\rho_{c}^{0}$ and $\beta=1-\rho^{\ast}(\alpha)$ (i.e. on $\gamma_{12}$) there is no layer at $\xi=1$. _Case 2: $\mathcal{L}_{\xi^{\ast}}^{+}\cap\mathcal{R}_{\xi^{\ast}}^{-}\neq\emptyset$._ We observe that by definition $\mathcal{L}_{\xi^{\ast}}^{+}\subset\mathcal{C}_{0}^{a}$ and $\mathcal{R}_{\xi^{\ast}}^{-}\subset\mathcal{C}_{0}^{r}$. Thus, this case corresponds to having $\alpha\geq\rho_{c}^{0}$ and $\beta\geq 1-\rho_{c}^{1}$ and their non-empty intersection is realised at the canard point $p^{\ast}$ (see (14)). This implies that the slow segment of these singular orbits is the canard orbit $S_{c}$. In particular, since $\rho^{+}(\xi^{\ast},s)=\frac{1}{2}=\rho^{-}(\xi^{\ast},t)$, it follows that $\rho^{+}(0,s)=1-\rho_{c}^{0}$ and $\rho^{-}(1,t)=1-\rho_{c}^{1}$. Consequently, the start/end point of the reduced flow are fixed by the canard and correspond to $p_{c}^{0}$ and $p_{c}^{1}$ respectively, whereas boundary layers at $\xi=0$, $1$ may arise depending on $\alpha$ and $\beta$. It is then possible to determine the starting and ending points of the orbit by following the flow of the layer problem (backwards at $\xi=0$ and forward at $\xi=1$); this leads to $\rho_{0}=1-\frac{k(\xi^{\ast})}{4\alpha k(0)},\qquad\rho_{1}=\frac{k(\xi^{\ast})}{4\beta k(1)}.$ (38) In particular, we have: * • When $\rho_{c}^{0}<\alpha<1-\rho_{c}^{0}$, $\rho$ increases along the boundary layer at $\xi=0$. Additionally: * – If $1-\rho_{c}^{1}<\beta<\rho_{c}^{1}$, i.e. for $(\alpha,\beta)\in\mathcal{G}_{3}$, $\rho$ increases along the boundary layer at $\xi=1$. This implies that the singular orbit is of type $3$ (see Figure 8(c)). * – If $\rho_{c}^{1}<\beta<1$, i.e. for $(\alpha,\beta)\in\mathcal{G}_{4}$, $\rho$ decreases along the boundary layer at $\xi=1$. This implies that the singular orbit is of type $4$ (see Figure 8(d)). * • When $1-\rho_{c}^{0}<\alpha<1$, $\rho$ decreases along the boundary layer at $\xi=0$. Additionally: * – If $1-\rho_{c}^{1}<\beta<\rho_{c}^{1}$, i.e. for $(\alpha,\beta)\in\mathcal{G}_{5}$, $\rho$ increases along the boundary layer at $\xi=1$. This implies that the singular orbit is of type $5$ (see Figure 9(a)). * – If $\rho_{c}^{1}<\beta<1$, i.e. for $(\alpha,\beta)\in\mathcal{G}_{6}$, $\rho$ decreases along the boundary layer at $\xi=1$. This implies that the singular orbit is of type $6$ (see Figure 9(b)). We note that when $\alpha=1-\rho_{c}^{0}$ and $\beta\geq 1-\rho_{c}^{1}$ (i.e. on $\gamma_{35}\cup\gamma_{46}$) we have no boundary layer at $\xi=0$. Moreover, when $\beta=\rho_{c}^{1}$ and $\alpha\geq\rho_{c}^{0}$ (i.e. on $\gamma_{34}\cup\gamma_{56}$) we have no boundary layer at $\xi=1$. _Case 3: $\mathcal{L}_{\xi^{\ast}}^{+}\cap r_{\xi^{\ast}}\neq\emptyset$._ By definition of $r_{\xi^{\ast}}$, this occurs only when $\beta<1-\rho_{c}^{1}$. In this case, we have $r_{\xi^{\ast}}\in\mathcal{L}_{\xi^{\ast}}^{+}$, which implies that $p_{1}=r$ and, consequently, $\rho_{1}=1-\beta$ (i.e., no boundary layers emerge at $\xi=1$). Moreover, since $\rho(0,s)=\rho_{\ast}(\beta)$, following the layer problem backwards until it hits $\mathcal{L}$, we obtain $\rho_{0}=1-\frac{\beta(1-\beta)k(1)}{\alpha k(0)}.$ (39) Consequently, the slow motion is here entirely contained in $\mathcal{C}_{0}^{a}$ and there is a boundary layer at $\xi=0$, whose nature depends on $\alpha$ as follows: * • If $1-\rho_{\ast}(\beta)<\alpha<\rho_{\ast}(\beta)$ and $0<\beta<1-\rho_{c}^{1}$, i.e. if $(\alpha,\beta)\in\mathcal{G}_{7}$, $\rho$ is increasing and the singular solution is of type $7$ (see Figure 9(c)). * • If $\rho_{\ast}(\beta)<\alpha<1$ and $0<\beta<1-\rho_{c}^{1}$, i.e. if $(\alpha,\beta)\in\mathcal{G}_{8}$, $\rho$ is decreasing, and we have a singular solution of type $8$ (see Figure 9(d)). We note that when $\alpha=\rho_{\ast}(\beta)$ and $\beta<1-\rho_{c}^{1}$ (i.e. on $\gamma_{78}$), there are no boundary layers. ∎ ###### Remark 5. The construction in Case 3 is essentially the same as the one in Case 1 upon reversal of the flow direction in (8). ###### Remark 6. Singular solutions of type 1, 2, 7, and 8 can be obtained also applying the same strategy used in [7, Proposition 2], as their slow portion is entirely contained in one of the two halves of the critical manifold ($\mathcal{C}_{0}^{r}$ in the case of type 1, 2, $\mathcal{C}_{0}^{r}$ in the case of type 7, 8). Therefore, it would be possible to only focus on the flow of the manifold $\mathcal{L}$ of left boundary conditions up to $\xi=1$ and check its intersection with the projection of the manifold $\mathcal{R}$ of right boundary conditions on $\mathcal{C}_{0}$. ###### Remark 7. Different values of $k(0)$, $k(1)$, and $k(\xi^{\ast})$ influence the structure of the bifurcation diagram sketched in 6 only quantitatively. In particular, the smaller $k(\xi^{\ast})$ is, the larger regions $\mathcal{G}_{i}$, $i=3,4,5,6$ are, consequently reducing the sizes of regions $\mathcal{G}_{i}$, $i=1,2,7,8$. Recall that smaller values of $k(\xi^{\ast})$ correspond to a narrower bottleneck. Figure 7: Schematic representation of a singular solution of (8)-(9) with $a=0.3$, $b=1.5$, $\alpha=0.3$, and $\beta=0.6$ (i.e. $(\alpha,\beta)\in\mathcal{G}_{3}$). The solution has boundary layers at $\xi=0$ and $\xi=1$, while the slow portion of the orbit coincides with $S_{c}$. (a) Region $\mathcal{G}_{1}$: $\alpha=0.1$, $\beta=0.4$ (b) Region $\mathcal{G}_{2}$: $\alpha=0.1$, $\beta=0.9$ (c) Region $\mathcal{G}_{3}$: $\alpha=0.3$, $\beta=0.6$ (d) Region $\mathcal{G}_{4}$: $\alpha=0.3$, $\beta=0.8$ Figure 8: Schematic representation of singular solutions of type 1-4 (rows 1-4, respectively). _First column_ : Boundary conditions at $\xi=0$ in $(j,\rho)$-space: the orange line is $\mathcal{L}$, while the orange curve is $\mathcal{L}^{+}$. The red dot represents $p_{0}$ and the green line illustrates the layer where $\rho$ increases (type 3, 4). _Second column_ : Slow evolution on $\mathcal{C}_{0}$ (blue curve). The orange lines are the projection of $\mathcal{L}$ and $\mathcal{L}^{+}$ on $\mathcal{C}_{0}$, while the purple one represents the projection of $\mathcal{R}^{-}$ on $\mathcal{C}_{0}$. The orange dot corresponds to $l$, while the purple dot corresponds to $r$. For orbits of type 3 and 4 the slow flow involves the passage through the canard point $p^{\ast}$. _Third column_ : Boundary conditions at $\xi=1$ in $(j,\rho)$-space. The red dot corresponds to $p_{1}$, while the purple line and curve represent the manifolds $\mathcal{R}$ and $\mathcal{R}^{-}$, respectively. The green line corresponds to the layer of the singular orbit where $\rho$ increases (type 1-3)/decreases (type 2-4). _Fourth column_ : Singular solution in $(\xi,\rho)$-space. (a) Region $\mathcal{G}_{5}$: $\alpha=0.9$, $\beta=0.6$ (b) Region $\mathcal{G}_{6}$: $\alpha=0.9$, $\beta=0.8$ (c) Region $\mathcal{G}_{7}$: $\alpha=0.7$, $\beta=0.2$ (d) Region $\mathcal{G}_{8}$: $\alpha=0.9$, $\beta=0.2$ Figure 9: Schematic representation of singular solutions of type 5-8 (rows 1-4, respectively). _First column_ : Boundary conditions at $\xi=0$ in $(j,\rho)$-space: the orange line is $\mathcal{L}$, while the orange curve is $\mathcal{L}^{+}$. The red dot represents $p_{0}$ and the green line illustrates the layer where $\rho$ increases (type 7)/decreases (type 5, 6, 8). _Second column_ : Slow evolution on $\mathcal{C}_{0}$ (blue curve). The orange lines are the projection of $\mathcal{L}$ and $\mathcal{L}^{+}$ on $\mathcal{C}_{0}$, while the purple one represents the projection of $\mathcal{R}^{-}$ on $\mathcal{C}_{0}$. The orange dot corresponds to $l$, while the purple dot corresponds to $r$. For orbits of type 5 and 6 the slow flow involves the passage through the canard point $p^{\ast}$. _Third column_ : Boundary conditions at $\xi=1$ in $(j,\rho)$-space. The red dot corresponds to $p_{1}$, while the purple line and curve represent the manifolds $\mathcal{R}$ and $\mathcal{R}^{-}$, respectively. The green line corresponds to the layer of the singular orbit where $\rho$ increases (type 5)/decreases (type 6). _Fourth column_ : Singular solution in $(\xi,\rho)$-space. ###### Remark 8 (Degenerate cases including continua of singular solutions). When $\alpha\leq\rho_{c}^{0}$ and $\beta=\rho^{\ast}(\alpha)$ \- i.e. when $(\alpha,\beta)\in\gamma_{17}$ \- we have that both $\mathcal{L}_{\xi^{\ast}}^{+}\cap r_{\xi^{\ast}}$ and $l_{\xi^{\ast}}\cap\mathcal{R}_{\xi^{\ast}}^{-}$ are non-empty. Consequently, there are two possible reduced solutions, satisfying (see Figure 10(a)) $\text{(a)}\,\begin{cases}\rho(0,s)=\alpha,\\\ \rho(1,s)=\beta,\end{cases}\text{ or }\quad\text{ (b) }\begin{cases}\rho(0,s)=1-\alpha,\\\ \rho(1,s)=\rho_{1}=1-\beta.\end{cases}$ (40) In this case, we have a continuum of singular solutions, since at any $\xi\in[0,1]$ it is possible to jump from the slow trajectory of the reduced flow in (a) to the one in (b) via the flow of the layer problem. Analogously, we obtain a continuum of singular solutions when $\alpha=\rho_{c}^{0}$, $\beta\geq 1-\rho_{c}^{1}$, i.e. when $(\alpha,\beta)\in\gamma_{13}\cup\gamma_{24}$. In this case, in fact, we have that both $\mathcal{L}_{\xi^{\ast}}^{+}\cap\mathcal{R}_{\xi^{\ast}}^{-}$ and $l_{\xi^{\ast}}\cap\mathcal{R}_{\xi^{\ast}}^{-}$ are non-empty, and therefore there are two possible reduced solutions (with jumps possible at any $\xi\in[0,\xi^{\ast}]$ via the flow of the layer problem) satisfying (see Figure 10(b)) $\text{(c)}\,\begin{cases}\rho(0,s)=\rho_{c},\\\ \rho(1,s)=1-\rho_{c}^{1},\end{cases}\text{ or }\quad\text{ (d) }\begin{cases}\rho(0,s)=1-\rho_{c},\\\ \rho(1,s)=1-\rho_{c}^{1}.\end{cases}$ (41) A last example of such a situation is given by $\alpha\geq\rho_{c}^{0}$, $\beta=1-\rho_{c}^{1}$, i.e. when $(\alpha,\beta)\in\gamma_{37}\cup\gamma_{58}$. Here, both $\mathcal{L}_{\xi^{\ast}}^{+}\cap\mathcal{R}_{\xi^{\ast}}^{-}$ and $\mathcal{L}_{\xi^{\ast}}^{+}\cap r_{\xi^{\ast}}$ are non-empty, leading again to two possible reduced solutions (with jumps possible at any $\xi\in[\xi^{\ast},1]$ via the flow of the layer problem) satisfying (see Figure 10(c)) $\text{(c)}\,\begin{cases}\rho(0,s)=1-\frac{k(\xi^{\ast})}{4\alpha k(0)},\\\ \rho(1,s)=\rho_{c}^{1},\end{cases}\text{ or }\quad\text{ (d) }\begin{cases}\rho(0,s)=1-\frac{k(\xi^{\ast})}{4\alpha k(0)},\\\ \rho(1,s)=1-\rho_{c}^{1}.\end{cases}$ (42) Since in these degenerate cases singular solutions are not unique, our method based on transversality arguments to infer persistence of singular solutions to (5a)-(5b) for $0<\varepsilon\ll 1$ do not apply. Moreover, at the point $\alpha=\rho_{c}^{0}$, $\beta=1-\rho_{c}^{1}$ – i.e. at the intersection of $\gamma_{17}$, $\gamma_{13}$, and $\gamma_{37}$ – the situation is even more degenerate as the three previous scenarios collide. We leave the analysis of these more delicate situations for future work. (a) (b) (c) Figure 10: Schematic representation in $(\xi,\rho)$-space of the slow portions (blue curves) of the possible singular orbits for (a) $(\alpha,\beta)\in\gamma_{17}$, (b) $(\alpha,\beta)\in\gamma_{13}\cup\gamma_{24}$, and (c) $(\alpha,\beta)\in\gamma_{37}\cup\gamma_{58}$. The orange and purple curves correspond to the projection of $\mathcal{L}^{+}$ and $\mathcal{R}^{-}$, respectively, on the $(\xi,\rho)$-space. Fast jumps from the slow solution in $\mathcal{C}_{0}^{r}$ to the slow solution in $\mathcal{C}_{0}^{a}$ are possible (a) at each $\xi\in[0,1]$, (b) for $\xi\in[0,\xi^{\ast}]$, and (c) for $\xi\in[\xi^{\ast},1]$. We now prove that the singular solutions from Proposition 1 perturb to solutions of (5a)-(5b) for $\varepsilon$ sufficiently small. ###### Theorem 1. Let $k\in C^{2}([0,1])$ be a positive function satisfying Assumption (6). For each $(\alpha,\beta)\in\mathcal{G}_{i}$, $i=1,\dots,8$, the boundary value problem (5b) has a unique solution $\rho(x,\alpha,\beta,\varepsilon)$ for $\varepsilon$ sufficiently small. In the phase-space formulation (10), this solution corresponds to an orbit $\Gamma^{i}_{\varepsilon}$ which is $\mathcal{O}(\varepsilon^{\mu})$-close to $\Gamma^{i}$ in terms of Hausdorff distance, with $\mu=1$ for $i=1,2,7,8$ and $\mu=1/2$ for $i=3,4,5,6$. ###### Proof. The solutions for $\varepsilon$ small are obtained by perturbing from the singular solutions $\Gamma^{i}$, $i=1,\dots,8$. More precisely, we show that the manifold obtained by flowing the line $\mathcal{L}$ of points corresponding to the boundary conditions at $\xi=0$ to $\xi=\xi^{\ast}$ for $\varepsilon$ small intersects the manifold obtained by flowing the line $\mathcal{R}$ of points corresponding to the boundary conditions at $\xi=1$ to $\xi=\xi^{\ast}$ in a point which is close to the corresponding point of the singular solution. Analogously to Proposition 1, this is done by considering three cases. _Case 1: $(\alpha,\beta)\in\mathcal{G}_{i}$, $i=1,2$_. In this case, the proof is completely analogous to Case 1 in [7, Theorem 2]. In particular, it is possible to show that for $0<\varepsilon\ll 1$ the (forward) flow defined by (8) takes a suitable small segment of $\mathcal{L}$ to a smooth, two- dimensional manifold $\mathcal{M}_{0,\varepsilon}$, which reduces to a curve $\mathcal{L}_{1,\varepsilon}$ when projected in the plane $\xi=1$. Such curve intersects $\mathcal{R}$ in a point $p_{1,\varepsilon}$ which corresponds to the right end-point of the solution of the boundary value problem. The full solution for $\xi\in[0,1]$ is then obtained by following the flow backward from $p_{1,\varepsilon}$ to $\xi=0$. In this case, the perturbed orbits are $\mathcal{O}(\varepsilon)$ close to the corresponding singular ones as all perturbations are $C^{1}$ in $\varepsilon$. _Case 2: $(\alpha,\beta)\in\mathcal{G}_{i}$, $i=3,4,5,6$_. In this case, the singular solution starts with a layer connecting the point $p_{0}\in\mathcal{L}$ to the point $p_{c}^{0}$ on $S_{c}$, then follows the canard through the canard point $p^{\ast}$ up to $\xi=1$, and finally ends with another layer connecting $p_{c}^{1}$ with the point $p_{1}\in\mathcal{R}$. To prove the persistence of this singular orbit, we flow the line $\mathcal{L}$ of boundary conditions at $\xi=0$ forward, the line $\mathcal{R}$ of boundary conditions at $\xi=1$ backward, and show that they intersect transversally at $\xi=\xi^{\ast}$ for $\varepsilon$ small. Since the singular solution involves the point $p^{\ast}$ on the non-hyperbolic fold line $F$ and the emergence of a canard, results on extending GSPT to such problems [15] are needed here. Fenichel theory [6] implies that away from the fold line $F$ (compact subsets of) $\mathcal{C}_{0}^{a}$ and $\mathcal{C}_{0}^{r}$ perturb smoothly to the slow manifolds $\mathcal{C}_{\varepsilon}^{a}$ and $\mathcal{C}_{\varepsilon}^{r}$, respectively. The results in [15, Theorem 4.1] imply that in a neighbourhood of the canard point $p^{\ast}$ the manifolds $\mathcal{C}_{\varepsilon}^{a}$ and $\mathcal{C}_{\varepsilon}^{r}$ intersect transversally in a maximal canard $S_{c}^{\varepsilon}$ (close to $S_{c}$) for $\varepsilon$ sufficiently small. As in case 1, consider a small segment of $\mathcal{L}$ containing $p_{0}$ and denote its extension by the forward flow of (10) by $\mathcal{M}_{0,\varepsilon}$ for $\varepsilon$ small. Analogously, consider a small segment of $\mathcal{R}$ containing $p_{1}$ and denote its extension by the backward flow of (10) by $\mathcal{M}_{1,\varepsilon}$ for $\varepsilon$ small (again a smooth, two- dimensional manifold). By Fenichel theory, the manifolds $\mathcal{M}_{0,\varepsilon}$ and $\mathcal{M}_{1,\varepsilon}$ are exponentially close to $\mathcal{C}_{\varepsilon}^{a}$ and $\mathcal{C}_{\varepsilon}^{r}$, respectively. Therefore, $\mathcal{M}_{0,\varepsilon}$ and $\mathcal{M}_{1,\varepsilon}$ also intersect transversally in a unique orbit, which is the unique solution to the boundary value problem (see Figure 11). Here, the $\mathcal{O}(\varepsilon^{1/2})$ distance between the perturbed and the corresponding singular solutions follows from the blow-up analysis in [15], since the effect of the perturbation in the scaling chart of the blow-up transformation is of the order $\varepsilon^{1/2}$. \begin{overpic}[scale={.7}]{constr_eps_2.pdf} \put(30.0,80.0){$\mathcal{M}_{0,\varepsilon}$} \put(60.0,70.0){$\mathcal{M}_{1,\varepsilon}$} \end{overpic} Figure 11: Schematic representation of a solution (continuous red curve) to the full problem (5a)-(5b) obtained for $(\alpha,\beta)\in\mathcal{G}_{4}$ with the strategy discussed in Case 2 of Theorem 1. The orange and purple manifolds represent $\mathcal{M}_{0,\varepsilon}$ in $[0,\xi^{\ast}+\eta]$ and $\mathcal{M}_{1,\varepsilon}$ in $[\xi^{\ast}-\eta,1]$ with $\eta=1/18$, respectively, whereas the red dots at $\xi=0$ and $\xi=1$ correspond to the initial and final point of the orbit. These manifolds intersect transversally at $\xi=\xi^{\ast}$, providing the uniqueness of the obtained solution. The passage close to the canard point $p^{\ast}$ leads to the $\mathcal{O}(\varepsilon^{1/2})$ distance from the corresponding singular orbit. _Case 3: $(\alpha,\beta)\in\mathcal{G}_{i}$, $i=7,8$_. This case can be proved following the same approach as in [7, Theorem 2], and in particular is completely analogous to Case 1 upon reversal of the flow direction. ∎ ## 3 Numerical experiments In this section we present some numerical results for the steady-state problem (5b) which support the analysis of Section 2. More details about the numerical method employed here can be found in [7]. All results are obtained for $\varepsilon=10^{-3}$. We first set $k(\xi)=1+a\cos\left(\frac{2\pi\xi}{b}\right)$, a choice that was already considered in Figure 6 in the singular case. For $\varepsilon\neq 0$, Figure 12 illustrates some typical profiles, one per region defined by the GSPT analysis. The values chosen for $\alpha$ and $\beta$ are the same as in Figure 6, the solutions are qualitatively very close. Figure 12: $(\alpha,\beta)$ bifurcation diagram for $\varepsilon=10^{-3}$. Here $k(\xi)=1+a\,\mathrm{cos}\left(\frac{2\pi\xi}{b}\right)$ with $a=0.3$, $b=1.5$. The insets show the density $\rho$ as a function of $\xi$. Next investigate more realistic choices for $k$, which should mimic a corridor with a bottleneck. We consider two regions of constant width that are connected by a narrower section in the middle. In particular, we consider a “supergaussian” profile for $k$: $k(\xi)=w_{e}-(w_{e}-w_{m})e^{-\left|\frac{\xi-\xi_{0}}{d}\right|^{6}}\,,$ where $w_{e}$, $w_{m}$, $d$ and $\xi_{0}$ are positive parameters, corresponding to the width of the wider regions at the left and right, the width of the narrow middle section, the neck length and the neck position, respectively. We pick $w_{e}=1$ and consider both $w_{m}=0.9$, corresponding to wider neck and $w_{m}=0.5$ which gives a more pronounced neck. The other parameters are taken as $w_{e}=1$, $d=0.2$, $\xi_{0}=0.6$, which gives a satisfactory, asymmetric width profile as shown in Figure 13. $0$$0.2$$0.4$$0.6$$0.8$$1$$-1/2$$-1/4$$0$$1/4$$1/2$$x$$y$$w_{m}=0.9$$w_{m}=0.5$ Figure 13: Representation of the 2D domain associated with a supergaussian $k$. The thick and bold lines correspond to $w_{m}=0.5$ and $w_{m}=0.9$, respectively. Some characteristic profiles are shown in Figure 14, along with the 8 regions defined by the GSPT analysis above. The selected values of the parameter pair $\alpha$ and $\beta$ are picked with exactly one pair value per region, and the same for both $w_{m}=0.9$ and $w_{m}=0.5$. The parameters $\alpha$ and $\beta$ are also chosen away from the $0$ and $1$, since those values lead to almost constant solutions for regions $\mathcal{G}_{1}$, $\mathcal{G}_{2}$, $\mathcal{G}_{7}$ and $\mathcal{G}_{8}$, which correspond to the blue and red shaded areas. All chosen values are stated in Table 1. | $\mathcal{G}_{1}$ | $\mathcal{G}_{2}$ | $\mathcal{G}_{3}$ | $\mathcal{G}_{4}$ | $\mathcal{G}_{5}$ | $\mathcal{G}_{6}$ | $\mathcal{G}_{7}$ | $\mathcal{G}_{8}$ ---|---|---|---|---|---|---|---|--- $\alpha$ | $0.119$ | $0.128$ | $0.5$ | $0.5$ | $0.881$ | $0.881$ | $0.5$ | $0.941$ $\beta$ | $0.5$ | $0.941$ | $0.5$ | $0.881$ | $0.5$ | $0.881$ | $0.119$ | $0.128$ Table 1: Parameters for Figure 14 Generally speaking, we have three parameters ranges of interest, within which the stationary solutions share the same qualitative behaviour: * • Small $\alpha$ (which corresponds to low inflow as in the blue regions $\mathcal{G}_{1}$ and $\mathcal{G}_{2}$), which leads to _low_ density stationary states with $\rho<\frac{1}{2}$ and a boundary layer on the right boundary. * • Small $\beta$ (which corresponds to low outflow as in the red regions $\mathcal{G}_{7}$ and $\mathcal{G}_{8}$), which leads to _high_ density stationary states with $\rho>\frac{1}{2}$ and a boundary layer on the left boundary. * • Large values of $\alpha$ and $\beta$ (corresponding to high inflow and outflow regimes as in the green regions $\mathcal{G}_{3}$ to $\mathcal{G}_{6}$), leading to density profiles going from high density on the left (before the bottleneck) to low density on the right (after). In this case, boundary layers a present on both boundaries. Inside these three areas, solutions seem to depend only weakly on $\alpha$ and $\beta$, which affect the height of the boundary layers only. It is only across the boundary between these areas (white lines) that pronounced qualitative changes occur. #### Impact of the width of the bottleneck We now turn our attention to the influence of $w_{m}$ on the solutions. The first obvious difference in Figure 14 is the larger square in the center for small $w_{m}$, which corresponds to the region $\mathcal{G}_{3}$ in the singular analysis. This is explained by the very simple dependency of both $\rho_{c}^{0}$ and $\rho_{c}^{1}$ (which bound $\mathcal{G}_{3}$) on $k$, see (34). Figure 14: Phase diagrams in the $(\alpha,\beta)$ parameter space from the GSPT analysis for the supergaussian $k$ along with some typical _non-singular_ solutions for $w_{m}=0.9$ (top) and $w_{m}=0.5$ (bottom). In the regions of low (resp. high) density, in blue (resp. red) in Figure 14, the density $\rho$ is roughly constant on large parts of the domain, with variations at the boundaries as well as at the front and back of the narrow section. Outside of it, $\rho$ takes similar values for both $w_{m}=0.9$ and $w_{m}=0.5$. Inside however, $\rho$ takes values much closer to $1/2$ for $w_{m}=0.5$. Indeed, where $\rho$ is almost constant, the flux can be approximated as $\mathrm{J}=k\rho(1-\rho)$ ; since $\mathrm{J}$ is independent of $x$, lower values of $k$ correspond to $\rho$ closer to $1/2$. This also seems to indicate that in both the low and high density phases, the flux $\mathrm{J}$ for given $(\alpha,\beta)$ only depends weakly on $w_{m}$. This numerical observation confirms the analytical results of Proposition 1 for the singular case ($\varepsilon=0$). We have in fact that in the regions of low (resp. high) density, studied in Case $1$, corresponding to regions $\mathcal{G}_{1}$ and $\mathcal{G}_{2}$ (resp. $3$, corresponding to $\mathcal{G}_{7}$ and $\mathcal{G}_{8}$), the density at the entrance (resp. exit) is given by $\alpha$ (resp. $\beta$) and hence is not affected by the features of the bottleneck. It follows that $\mathrm{J}=\alpha(1-\alpha)$ (resp. $\mathrm{J}=\beta(1-\beta)$). In the green region (which could be argued to correspond to the so-called maximum flux phase for constant $k$), the situation is different. Although the profiles are qualitatively similar with a transition between a high density to a low density plateau, the densities for $w_{m}=0.5$ (wider bottleneck, top) are much closer to $1/2$ for the values of $\alpha$ and $\beta$ which are considered. This relates to a higher flux for the wider bottleneck. The computational results in these regions correspond to the analysis of Case $2$ (regions $\mathcal{G}_{3}$ to $\mathcal{G}_{6}$), a situation in which the density changes significantly (i.e. a boundary layer) in proximity of both the entrance and the exit, immediately preceded by a region where it is approximately constant. The density value in these areas is defined by $1-\rho_{c}^{0}$ and $1-\rho_{c}^{1}$, respectively. With the choice of parameters in Table 1, the approximation $k\simeq 1$ holds, at least in the first and last $10\%$ of the domain; we then get from its definition that $\rho_{c}^{0}$ (resp. $\rho_{c}^{1}$) is increasing (resp. decreasing) w.r.t $w_{m}$. In fact we have $1-\rho_{c}^{0}\simeq\rho_{c}^{1}\simeq\frac{1}{2}\left(1+\sqrt{1-w_{m}}\right)\quad\text{and}\quad\mathrm{J}\simeq\frac{1}{4}w_{m}\,,$ so that $\mathrm{J}$ grows linearly with the width of the neck and eventually reaches $1/4$ for $w_{m}=1$, the maximum value for a straight channel. This is in agreement with the numerical observations described above. To summarize, the influence of the width of the neck in this case is two-fold. First, in terms of $(\alpha,\beta)$, the green region grows larger and eventually completely fills the parameter space as the neck-width goes to zero. Second, it is in this region that $w_{m}$ has a noticeable effect on the flux $\mathcal{J}$, which depends linearly on $w_{m}$, as one would expect intuitively. ###### Remark 9. The observations above are independent of the choice of $k$, provided that $k(0)=k(1)=1$ and that the minimum of $k$ is non degenerate. In the singular case, one obtains an explicit expression for $\mathrm{J}$, which we write as $\mathrm{J}_{k}(\alpha,\beta)$ to emphasize the dependency on $k$, $\alpha$ and $\beta$: $\mathrm{J}_{k}(\alpha,\beta)=\begin{cases}\alpha(1-\alpha)&\alpha\leq\rho_{c}^{0}\wedge\alpha\leq\beta\,,\\\ \beta(1-\beta)&\beta\leq 1-\rho_{c}^{1}\wedge\beta\leq\alpha\,,\\\ \frac{1}{4}\min_{\xi}k(\xi)&\text{otherwise}\,.\end{cases}$ In particular, we have that $\mathrm{J}_{k}(\alpha,\beta)=\min\left\\{\mathrm{J}_{1}(\alpha,\beta),\min_{\xi}k(\xi)\right\\}$. This is illustrated in Figure 15. This means that as $\min_{\xi}k(\xi)$ decreases, the flow $\mathrm{J}_{k}(\alpha,\beta)$ will saturate, _i.e._ reach its maximum, faster as $\alpha$ and $\beta$ increase. The maximum of $\mathrm{J}_{k}$ will also decrease linearly with $\min_{\xi}k(\xi)$. Numerical experiments with two narrow sections of varying width suggest that this applies also for functions $k$ with several (nondegenerate) critical points. Figure 15: Illustration of the flow $\mathrm{J}_{1}$ (solid colors on the left, wireframe on the right) and $\mathrm{J}_{k}$ (solid colors on the right) as a function of $\alpha$ and $\beta$. Recall that $\rho_{c}^{0,1}\rightarrow\frac{1}{2}$ as $k(\xi^{*})\rightarrow 1$, so that from the green rectangles, only the darker one (top right) remains in the limit. ## Conclusion In this work, we investigate the steady-states of a 1D area averaged model describing pedestrian dynamics for unidirectional flows in domains that have a bottleneck. In the proposed model, information about the geometry enters as a nonhomogeneous factor acting both on the diffusive and convective terms. We investigate the case in which this factor admits an isolated minimum, which corresponds to the bottleneck. The stationary profiles exhibit a multi-scale nature, which we analyse using GSPT. This allows us to thoroughly understand the influence of inflow and outflow rates ($\alpha$ and $\beta$, respectively) on the structure of the solutions and, in particular, on the formation of boundary layers. In this framework, the isolated minimum inside the bottleneck corresponds to a canard point where an unusual passage through a repelling branch of the critical manifold occurs. The more complex geometry therefore induces the emergence of two additional regions in the singular bifurcation diagram which have not been observed and investigated before. In general, orbits which include such passage exist for a wide area in the $(\alpha,\beta)$-parameter space, whose size decreases as the neck becomes wider. In order to test the ability of our 1D reduction to capture the essential dynamics of the original two-dimensional model, we plan to suitably calibrate and validate our model as a next step. As observed in [7], the quality of the proposed 1D area averaged approximation depends on the parameter regime considered; we will therefore investigate further averaging assumptions to overcome these issues in the next steps of our research. ## Declaration of competing interests The authors declare no conflict of interest. ## Acknowledgements AI acknowledges support from an FWF Hertha Firnberg Research Fellowship (T 1199-N). ## References * [1] M. Burger, S. Hittmeir, H. Ranetbauer, and M.-T. Wolfram. Lane formation by side-stepping. SIAM Journal on Mathematical Analysis, 48(2):981–1005, 2016. * [2] M. Burger, P. A. Markowich, and J.-F. Pietschmann. Continuous limit of a crowd motion and herding model: analysis and numerical simulations. Kinetic & Related Models, 4(4):1025, 2011. * [3] M. Burger and J.-F. Pietschmann. Flow characteristics in a crowded transport model. Nonlinearity, 29(11):3528–3550, 2016. * [4] E. Cristiani, B. Piccoli, and A. Tosin. Multiscale modeling of pedestrian dynamics, volume 12. Springer, 2014. * [5] B. Derrida, E. Domany, and D. Mukamel. An exact solution of a one-dimensional asymmetric exclusion model with open boundaries. J. Stat. Phys., 69(3-4):667–687, 1992. * [6] N. Fenichel. Geometric singular perturbation theory for ordinary differential equations. Journal of Differential Equations, 31(1):53–98, 1979. * [7] A. Iuorio, G. Jankowiak, P. Szmolyan, and M.-T. Wolfram. A PDE model for unidirectional flows: stationary profiles and asymptotic behaviour. Journal of Mathematical Analysis and Applications, page 126018, 2022\. * [8] A. Iuorio, N. Popović, and P. Szmolyan. Singular perturbation analysis of a regularized MEMS model. SIAM Journal on Applied Dynamical Systems, 18(2):661–708, 2019\. * [9] A. Iuorio and F. Veerman. The influence of autotoxicity on the dynamics of vegetation spots. Physica D: Nonlinear Phenomena, 427:133015, 2021. * [10] C. K. R. T. Jones. Geometric singular perturbation theory. In Dynamical Systems, pages 44–118. Springer Berlin Heidelberg, 1995. * [11] C. Kuehn. Multiple Time Scale Dynamics. Springer International Publishing, 2015. * [12] M. J. Lighthill and G. B. Whitham. On kinematic waves II. A theory of traffic flow on long crowded roads. Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences, 229(1178):317–345, 1955. * [13] B. Maury and S. Faure. Crowds in Equations: An Introduction to the Microscopic Modeling of Crowds. 09 2018. * [14] P. I. Richards. Shock waves on the highway. Operations research, 4(1):42–51, 1956. * [15] P. Szmolyan and M. Wechselberger. Canards in $\mathbb{R}^{3}$. Journal of Differential Equations, 177(2):419–453, 2001. * [16] P. Szmolyan and M. Wechselberger. Relaxation oscillations in $\mathbb{R}^{3}$. Journal of Differential Equations, 200(1):69–104, 2004. * [17] A. J. Wood. A totally asymmetric exclusion process with stochastically mediated entrance and exit. Journal of Physics. A. Mathematical and Theoretical, 42(44):445002, 10, 2009.
# High order asymptotic preserving finite difference WENO schemes with constrained transport for MHD equations in all sonic Mach numbers ###### Abstract. In this paper, a high-order semi-implicit (SI) asymptotic preserving (AP) and divergence-free finite difference weighted essentially nonoscillatory (WENO) scheme is proposed for magnetohydrodynamic (MHD) equations. We consider the sonic Mach number $\varepsilon$ ranging from $0$ to $\mathcal{O}(1)$. High- order accuracy in time is obtained by SI implicit-explicit Runge–Kutta (IMEX- RK) time discretization. High-order accuracy in space is achieved by finite difference WENO schemes with characteristic-wise reconstructions. A constrained transport method is applied to maintain a discrete divergence-free condition. We formally prove that the scheme is AP. Asymptotic accuracy (AA) in the incompressible MHD limit is obtained if the implicit part of the SI IMEX-RK scheme is stiffly accurate. Numerical experiments are provided to validate the AP, AA, and divergence-free properties of our proposed approach. Besides, the scheme can well capture discontinuities such as shocks in an essentially non-oscillatory fashion in the compressible regime, while it is also a good incompressible solver with uniform large-time step conditions in the low sonic Mach limit. ###### Key words and phrases: all sonic Mach number; MHD equations; divergence-free; asymptotic preserving; SI IMEX-RK; finite difference WENO Wei Chen School of Mathematical Sciences, Xiamen University Xiamen, Fujian, 361005, P.R. China Email<EMAIL_ADDRESS> Kailiang Wu***This work is supported in part by NSFC grant 12171227 Department of Mathematics & SUSTech International Center for Mathematics Southern University of Science and Technology National Center for Applied Mathematics Shenzhen (NCAMS) Shenzhen, Guangdong 518055, China Email<EMAIL_ADDRESS> Tao Xiong†††Corresponding author. The work of this author was partially supported by NSFC grant No. 11971025, and the Strategic Priority Research Program of the Chinese Academy of Sciences Grant No. XDA25010401. School of Mathematical Sciences, Xiamen University Fujian Provincial Key Laboratory of Mathematical Modeling and High-Performance Scientific Computing Xiamen, Fujian, 361005, P.R. China Email<EMAIL_ADDRESS> ## 1\. Introduction Ideal magnetohydrodynamic (MHD) equations are widely used in the modeling of weather prediction, astrophysics, as well as laboratory plasma applications such as flows in tokamaks and stellarators. Many shock capturing schemes with explicit time discretizations have been developed for solving compressible ideal MHD equations, including high order discontinuous Galerkin [43, 44, 60, 59], finite difference [49, 15, 18, 16, 17], and finite volume schemes [42, 63, 2, 54, 62], etc. For MHD equations, divergence free of the magnetic field is a very important property. If the initial magnetic field is divergence-free, this condition will be maintained for all later times. Numerically, if this divergence-free condition is violated seriously, an unphysical force will be created to parallel the magnetic field [16], which can produce some numerical instabilities or nonphysical features in the computed solution, cause the loss of pressure positivity [58, 61], and/or lead to the failure of the simulation. The constrained transport (CT) methodology, which was originally introduced by Evans and Hawley[26], is an effective way to maintain the (discrete) divergence-free condition up to machine precision. Many improvements and extensions [4, 1, 20, 21, 15, 16, 17, 18, 29] are followed. Recently, the MHD system in the low sonic Mach limit has attracted a lot of interests [45, 42, 25, 33, 34, 19, 48]. Jiang, Ju, and Li [33] have shown that a weak solution of compressible MHD equations will converge to a strong solution of their corresponding incompressible MHD equations in this low sonic Mach limit. They have also investigated the low Mach limit of full MHD equations with a heat conductivity [34]. Cui, Ou, and Ren [19] have established a uniform convergence from full compressible MHD equations to isentropic incompressible MHD equations with well-prepared initial conditions in the three-dimensional case. Numerically it is also very attractive to develop schemes for MHD flows at any speed. For the above mentioned explicit shock capturing schemes, when applied to the MHD system in the low sonic Mach regime, the time step is subject to a very strict CFL condition which is proportional to the sonic Mach number, making it very undesirable for all speed flows. An implicit time discretization can release the small time step restrictions from an explicit one, but usually it results in a highly nonlinear system, which is not easy to be solved efficiently with an iterative method and it is hard to guarantee a fast convergence. Furthermore, numerical viscosities for such schemes are inversely proportional to the sonic Mach number, which introduce excessive numerical dissipations. Instead, in recent years some researchers are in pursuit of semi-implicit (SI) schemes with low dissipations. In [25], Dumbser et. al. developed a pressure-based SI finite volume all Mach number flow solver for compressible MHD equations. Minoshima and Miyoshi [48] proposed a multistate low-dissipation advection upstream splitting method for MHD with both high and low Mach numbers, by using a Harten-Lax-van Leer discontinuities (HLLD) approximate Riemann solver. Leidi et al. [42] presented an SI finite volume solver with a 5-wave HLLD and a well-balanced method to efficiently simulate MHD flows at low Mach numbers with a gravitational source. For the MHD system with all sonic Mach numbers, it is important to design schemes with asymptotic stability and consistency as the sonic Mach number goes to $0$ in the incompressible limit, namely, asymptotic preserving (AP). AP schemes have been widely used in a variety of areas, such as hydrodynamic or diffusive limits of kinetic models, relaxation limits of hyperbolic models, and low Mach number limits of compressible fluid models. We refer readers to the papers [35, 37, 36, 38, 27, 24, 23] for more information. To achieve high order AP schemes in time, SI implicit-explicit Runge-Kutta (IMEX-RK) methods are widely used, which apply an explicit time discretization for nonstiff terms, while an implicit time discretization for stiff terms [5, 7, 6, 9, 10, 8, 40, 39, 50, 13, 47, 11]. With a suitable choice of explicit and implicit discretizations, both high efficiency and uniform stability independent of the stiff parameter can be obtained. As related, in [57] Tavelli and Dumbser developed an SI space-time discontinuous Galerkin method for all Mach-number flow, which involves staggered meshes and a Picard iteration for solving a large nonlinear system. In [10, 8], Boscarino, Qiu, Russo, and Xiong presented high-order SI IMEX weighted essentially nonoscillatory (WENO) schemes for all- Mach isentropic Euler equations and all-Mach full Euler system, respectively. In [11], Boscheri and Pareschi developed high order pressure-based semi- implicit IMEX schemes for the three dimensional Navier-Stokes equations at all Mach numbers. In [30], Huang, Xing, and Xiong designed a high-order well- balanced SI AP scheme for shallow water equations in all Fraude numbers. We refer to many other related works from the references therein. In this paper, we would like to develop a high-order AP finite difference WENO scheme with CT for MHD equations in all sonic Mach numbers. To our best knowledge, there are very few AP schemes for MHD systems in literature. The new contributions and innovations of this work are outlined as follows: * • In the low sonic Mach number regime, SI IMEX-RK methods are adopted for time discretization to obtain a uniform time step independent of the sonic Mach number. Besides, to avoid a nonlinearity from the equation of state (EOS), an SI approach is used, leading to a linearized elliptic equation for the pressure. With carefully designed explicit and implicit discretizations, only a linear system needs to be solved. The resulting scheme is more efficient than explicit shock capturing schemes, especially in the low sonic Mach regime. * • Numerical viscosities are carefully designed to avoid excessive numerical dissipations, and a CT method is applied to maintain a discrete divergence- free property. The scheme can well capture discontinuities such as shocks in an essentially non-oscillatory fashion in the compressible regime. Meanwhile, the scheme is also a good incompressible solver as $\varepsilon\rightarrow 0$. Due to less dissipation, our scheme can be shown to perform better than explicit schemes for low sonic Mach problems. * • Formal AP and asymptotically accurate (AA) properties in the stiff limit as the sonic Mach number $\varepsilon\rightarrow 0$ are proved, by assuming the implicit part of an SI IMEX-RK scheme is stiffly accurate (SA). The rest of the paper is as follows. In Section 2, we will briefly review the ideal MHD system, corresponding to its low sonic Mach limit based on asymptotic expansions, and a CT methodology for the divergence-free condition. A first-order SI scheme in time will first be introduced in Section 3, and then a high-order SI scheme with IMEX-RK in time and finite difference WENO in space is followed. Formal proofs of AP and AA properties are given in Section 4. Numerical experiments are performed in Section 5. A brief conclusion will be drawn in the last section. ## 2\. Equations of compressible ideal MHD The ideal MHD equations in a conservative form can be written as (2.1) $\frac{\partial}{\partial t}\begin{bmatrix}\rho\\\ \rho\mathbf{u}\\\ \mathbf{B}\\\ E\end{bmatrix}+\nabla\cdot\begin{bmatrix}\rho\mathbf{u}\\\ \rho\mathbf{u}\otimes\mathbf{u}+(p+\frac{1}{2}{\|\mathbf{B}\|}^{2})\mathbf{I}-\mathbf{B}\otimes\mathbf{B}\\\ \mathbf{u}\otimes\mathbf{B}-\mathbf{B}\otimes\mathbf{u}\\\ (E+p+\frac{1}{2}{\|\mathbf{B}\|}^{2})\mathbf{u}-\mathbf{B}(\mathbf{u}\cdot\mathbf{B})\end{bmatrix}=\mathbf{0},$ where $\rho$ is the fluid mass density, $\mathbf{u}=(u,v,w)$ is the velocity, $\mathbf{B}=(B_{x},B_{y},B_{z})$ is the magnetic field, $E$ is the total energy density, $p$ is the gas pressure, and $\|\cdot\|$ is the Euclidean vector norm. The total energy is given by (2.2) $E=\frac{p}{\gamma-1}+\frac{1}{2}(\rho{\|\mathbf{u}\|}^{2}+{\|\mathbf{B}\|}^{2}),$ where $\gamma$ is the specific heat ratio. The system is subject to a divergence-free condition of the magnetic field (2.3) $\nabla\cdot\mathbf{B}=0.$ If initially, the divergence-free condition holds, from (2.1), it also holds for all later times [18]. For this reason, (2.3) is usually not regarded as a constraint (like the $\nabla\cdot{\bf u}=0$ constraint for the incompressible Navier-Stokes equations), but rather an involution [29, 18, 16]. ### 2.1. Eigenstructure of the MHD system For (2.1), the eigenvalues of the Jacobian for the flux function along a normal direction ${\bf n}$ are given as: (2.4) $\lambda_{1,8}={\bf u}\cdot{\bf n}\mp c_{f},\quad\lambda_{2,7}={\bf u}\cdot{\bf n}\mp c_{a},\quad\lambda_{3,6}={\bf u}\cdot{\bf n}\mp c_{s},\quad\lambda_{4,5}={\bf u}\cdot{\bf n},\quad$ where $\lambda_{1,8}$, $\lambda_{2,7}$, $\lambda_{3,6}$, $\lambda_{4}$, and $\lambda_{5}$ are the left/right fast magnetosonic waves, left/right Alfvén waves, left/right slow magnetosonic waves, entropy wave, and divergence wave, correspondingly. The sound speed $a$, Alfvén speed $c_{a}$, fast magnetosonic speed $c_{f}$, and slow magnetosonic speed $c_{s}$ are defined as (2.5) $\left\\{\begin{aligned} a&:=\sqrt{\frac{\gamma p}{\rho}},\\\ c_{a}&:=\sqrt{\frac{{({\bf B}\cdot{\bf n})}^{2}}{\rho}},\\\ c_{f,s}&:={\left\\{\frac{1}{2}\left[a^{2}+\frac{{\|{\bf B}\|}^{2}}{\rho}\pm\sqrt{{\left(a^{2}+\frac{{\|{\bf B}\|}^{2}}{\rho}\right)}^{2}-4a^{2}{c^{2}_{a}}}\right]\right\\}}^{\frac{1}{2}}.\end{aligned}\right.$ A fixed ordering for the MHD eigenvalues is $\lambda_{1}\leq\lambda_{2}\leq\lambda_{3}\leq\lambda_{4}\leq\lambda_{5}\leq\lambda_{6}\leq\lambda_{7}\leq\lambda_{8}.$ Spectral decomposition for this system has already been described in [32, 12, 14, 51, 52, 55], and we will follow the one in [32] for all simulations in the current paper. ### 2.2. Non-dimensionalized MHD To derive a low sonic Mach limit for the ideal MHD system, we start with rewriting (2.1), (2.2), and (2.3) into a dimensionless form. We choose some reference or characteristic values, such as a length $x_{0}$, a time $t_{0}$, a fluid mass density $\rho_{0}$, a velocity $u_{0}$, a magnetic field $B_{0}$, and a gas pressure $p_{0}$, with $u_{0}=x_{0}/t_{0}$. For the sake of simplicity, we consider a regime where the characteristic Alfvén Mach number $u_{0}\sqrt{\rho_{0}}/B_{0}=1$. Then we define the following dimensionless variables: $\hat{{{\bf x}}}=\frac{{{\bf x}}}{x_{0}},\quad\hat{t}=\frac{t}{t_{0}},\quad\hat{\rho}=\frac{\rho}{\rho_{0}},\quad\hat{{\bf u}}=\frac{{\bf u}}{u_{0}},\quad\hat{{\bf B}}=\frac{{\bf B}}{B_{0}},\quad\hat{p}=\frac{p}{p_{0}}.$ Inserting these new variables into (2.1), we obtain the following non- dimensionalised MHD equations[42]: (2.6) $\frac{\partial}{\partial t}\begin{bmatrix}\rho\\\ \rho\mathbf{u}\\\ \mathbf{B}\\\ E\\\ \end{bmatrix}+\nabla\cdot\begin{bmatrix}\rho\mathbf{u}\\\ \rho\mathbf{u}\otimes\mathbf{u}+(\frac{p}{\varepsilon^{2}}+\frac{1}{2}{\|\mathbf{B}\|}^{2})\mathbf{I}-\mathbf{B}\otimes\mathbf{B}\\\ \mathbf{u}\otimes\mathbf{B}-\mathbf{B}\otimes\mathbf{u}\\\ (E+p+\frac{\varepsilon^{2}}{2}{\|\mathbf{B}\|}^{2})\mathbf{u}-\varepsilon^{2}\mathbf{B}(\mathbf{u}\cdot\mathbf{B})\end{bmatrix}=\mathbf{0},$ where we have dropped the hats of the dimensionless variables for ease of presentation. The EOS, divergence-free condition and the characteristic sonic Mach number $\varepsilon$ are now given by (2.7) $E:=\frac{p}{\gamma-1}+\frac{\varepsilon^{2}}{2}(\rho{\|{\bf u}\|}^{2}+{\|{\bf B}\|}^{2}),\quad\nabla\cdot{\bf B}=0,\quad\varepsilon:=u_{0}\sqrt{\frac{\rho_{0}}{p_{0}}}.$ In this dimensionless MHD system (2.6) with (2.7), the expressions of those eigenvalues (2.4) remain unchanged, except the sound speed $a$, Alfvén speed $c_{a}$, fast magnetosonic speed $c_{f}$, and slow magnetosonic speed $c_{s}$ from (2.5) now become $\left\\{\begin{aligned} a&:=\frac{1}{\varepsilon}\sqrt{\frac{\gamma p}{\rho}},\\\ c_{a}&:=\sqrt{\frac{{({\bf B}\cdot{\bf n})}^{2}}{\rho}},\\\ c_{f,s}&:={\left\\{\frac{1}{2}\left[a^{2}+\frac{{\|{\bf B}\|}^{2}}{\rho}\pm\sqrt{{\left(a^{2}+\frac{{\|{\bf B}\|}^{2}}{\rho}\right)}^{2}-4a^{2}{c^{2}_{a}}}\right]\right\\}}^{\frac{1}{2}}.\end{aligned}\right.$ Following the eigenstructure of the dimensionless MHD system discussed above, it is natural to find that if the background flow velocity $u_{0}$ is slow, then the sonic Mach number $\varepsilon=u_{0}\sqrt{\rho_{0}}/\sqrt{p_{0}}\ll 1$, and the fast magnetosonic speed $c_{f}$ becomes very large. We can see that large magnetosonic speed $c_{f}$ will make the fast magnetosonic waves $\lambda_{1,8}$ to be much faster as compared to other waves. In this situation, for explicit shock-capturing schemes, the time step subjects to $\Delta t={\rm CFL}\frac{\Delta x}{\mathop{\max}\limits_{1\leq i\leq 8}|\lambda_{i}|}=\mathcal{O}(\varepsilon\Delta x),$ where $\Delta t$ is the time step size, $\Delta x$ is the mesh size, and ${\rm CFL}$ is the time stability CFL number. On one hand, highly inaccurate solutions will appear due to the excessive numerical viscosity (scales as $\varepsilon^{-1}$) in standard upwind schemes. On the other hand, this restriction results in an increasingly large computational cost for low sonic Mach fluid flows. A straight approach is to apply an implicit time discretization to avoid such severe time step constraints. However, a fully implicit scheme will result in a complicated nonlinear system which is usually very hard to be solved efficiently. Besides, it may not be able to guarantee a correct asymptotic limit. Thus, it is important to design efficient schemes with both asymptotic stability and consistency in the low sonic Mach limit, namely, the AP property. ### 2.3. Low sonic Mach limit Next, we recall a formal derivation of the incompressible MHD system from the non-dimensionalized compressible MHD equations (2.6) with (2.7). We consider an asymptotic expansion ansatz for the following variables [46]: (2.8) $\left\\{\begin{aligned} &p(\mathbf{x},t)=p_{0}(\mathbf{x},t)+\varepsilon p_{1}(\mathbf{x},t)+\varepsilon^{2}p_{2}(\mathbf{x},t)+\cdots,\\\ &\rho(\mathbf{x},t)=\rho_{0}(\mathbf{x},t)+\varepsilon\rho_{1}(\mathbf{x},t)+\varepsilon^{2}\rho_{2}(\mathbf{x},t)+\cdots,\\\ &\mathbf{u}(\mathbf{x},t)=\mathbf{u}_{0}(\mathbf{x},t)+\varepsilon\mathbf{u}_{1}(\mathbf{x},t)+\varepsilon^{2}\mathbf{u}_{2}(\mathbf{x},t)+\cdots,\\\ &\mathbf{B}(\mathbf{x},t)=\mathbf{B}_{0}(\mathbf{x},t)+\varepsilon\mathbf{B}_{1}(\mathbf{x},t)+\varepsilon^{2}\mathbf{B}_{2}(\mathbf{x},t)+\cdots.\end{aligned}\right.$ Inserting (2.8) into (2.6) and (2.7), equating to zero for different orders of $\varepsilon$, corresponding to the first three leading orders, we have * • $\mathcal{O}(\varepsilon^{-2})$ $\nabla p_{0}=0,$ * • $\mathcal{O}(\varepsilon^{-1})$ $\nabla p_{1}=0,$ * • $\mathcal{O}(\varepsilon^{0})$ (2.9) $\frac{\partial}{\partial t}\begin{bmatrix}\rho_{0}\\\ \rho_{0}\mathbf{u}_{0}\\\ \mathbf{B}_{0}\\\ E_{0}\\\ \end{bmatrix}+\nabla\cdot\begin{bmatrix}\rho_{0}\mathbf{u}_{0}\\\ \rho_{0}\mathbf{u}_{0}\otimes\mathbf{u}_{0}+(p_{2}+\frac{1}{2}{\|\mathbf{B}_{0}\|}^{2})\mathbf{I}-\mathbf{B}_{0}\otimes\mathbf{B}_{0}\\\ \mathbf{u}_{0}\otimes\mathbf{B}_{0}-\mathbf{B}_{0}\otimes\mathbf{u}_{0}\\\ (E_{0}+p_{0})\mathbf{u}_{0}\end{bmatrix}=\mathbf{0},$ where we have assumed a similar expansion of $E$ as $p$ in (2.8), and take $E_{0}=p_{0}/(\gamma-1)$. Denoting the material derivative ${\rm d/d}t=\partial/\partial t+{\bf u}_{0}\cdot\nabla$, from the energy equation in (2.9), $E_{0}=p_{0}/(\gamma-1)$, with $\nabla p_{0}=0$, we have (2.10) $\nabla\cdot{\bf u}_{0}=-\frac{1}{p_{0}\gamma}\frac{{\rm d}p_{0}}{{\rm d}t}.$ Taking into account periodic or no-slip ${\bf u}\cdot{\bf n}=0$ boundary conditions on the spatial domain $\Omega$, and integrating (2.10) over $\Omega$, we get $\int_{\Omega}\nabla\cdot{\bf u}_{0}d{\bf x}=\int_{\partial\Omega}{\bf u}_{0}\cdot{\bf n}ds=0$, where ${\bf n}$ is the unit outward normal vector along $\partial\Omega$. This implies $p_{0}$ is constant not only in space but also in time, that is $p_{0}={\rm Const}$. A direct result $\nabla\cdot{\bf u}_{0}=0$ can be obtained through (2.10) with $p_{0}={\rm Const}$. Inserting $\nabla\cdot{\bf u}_{0}=0$ into the mass- continuity equation in (2.9), we get ${\rm d}\rho_{0}/{\rm d}t=0$ which means that the density is constant along the characteristic trajectories. For the order $\mathcal{O}(\varepsilon)$ in (2.6), we have: $\partial E_{1}/\partial t+\nabla\cdot[(E_{0}+p_{0}){\bf u}_{1}+(E_{1}+p_{1}){\bf u}_{0}]=0.$ Due to $E_{1}=p_{1}/(\gamma-1)$ from the EOS in (2.7), similarly as above, we get $p_{1}={\rm Const}$ and $\nabla\cdot{\bf u}_{1}=0$. In this case, we include $\varepsilon\mathbf{u}_{1}$ into $\mathbf{u}_{0}$, and $\varepsilon p_{1}$ into $p_{0}$, the asymptotic expansions of ${\bf u}$ and $p$ become (2.11) $\mathbf{u}=\mathbf{u}_{0}+\varepsilon^{2}\mathbf{u}_{2}+\cdots,$ and (2.12) $p=p_{0}+\varepsilon^{2}p_{2}+\cdots.$ Using such expansions for $\mathbf{u}$ and $p$, as $\varepsilon\rightarrow 0$, the leading order of the system (2.6) yields (2.13) $\left\\{\begin{aligned} &\frac{\partial}{\partial t}\begin{bmatrix}\rho_{0}\mathbf{u}_{0}\\\ \mathbf{B}_{0}\\\ \end{bmatrix}+\nabla\cdot\begin{bmatrix}\rho_{0}\mathbf{u}_{0}\otimes\mathbf{u}_{0}+(p_{2}+\frac{1}{2}{\|\mathbf{B}_{0}\|}^{2})\mathbf{I}-\mathbf{B}_{0}\otimes\mathbf{B}_{0}\\\ \mathbf{u}_{0}\otimes\mathbf{B}_{0}-\mathbf{B}_{0}\otimes\mathbf{u}_{0}\end{bmatrix}=\mathbf{0},\\\ &\frac{{\rm d}\rho_{0}}{{\rm d}t}=0,\quad p_{0}={\rm Const},\quad\nabla\cdot{\bf u}_{0}=0,\quad\nabla\cdot{\bf B}=0.\\\ \end{aligned}\right.$ Numerically, in order to have a correct asymptotic limit from (2.6) to (2.13), it is important to impose well-prepared initial conditions, which are consistent with the expansions given in (2.8), with ${\bf u}$ and $p$ from (2.11) and (2.12). According to [22, 41], the well-prepared initial conditions for (2.6) are given as (2.14) $\left\\{\begin{aligned} &\rho(t=0,{\bf x})=\rho_{0}({\bf x})+\mathcal{O}(\varepsilon)>0,\quad&&p(t=0,{\bf x})=p_{0}+\varepsilon^{2}p_{2}({\bf x})+\cdots,\\\ &{\bf u}(t=0,{\bf x})={\bf u}_{0}({\bf x})+\mathcal{O}(\varepsilon^{2}),\quad&&{\bf B}(t=0,{\bf x})={\bf B}_{0}({\bf x})+\mathcal{O}(\varepsilon),\end{aligned}\right.$ where $\rho_{0}({\bf x})$ is a strictly positive function, $p_{0}={\rm Const}$, $\nabla\cdot{\bf u}_{0}=0$, and $\nabla\cdot{\bf B}_{0}=0$. ### 2.4. CT for $\nabla\cdot{\bf B}=0$ Many CT methods have been proposed to satisfy the discrete divergence-free condition $\nabla\cdot{\bf B}=0$ for the ideal MHD equations [18, 17, 16, 15]. Here we consider the one described in [18], in which the magnetic field is written as the curl of a magnetic vector potential ${\bf A}$: (2.15) ${\bf B}=\nabla\times{\bf A}.$ Due to $\nabla\cdot({\bf u}\otimes{\bf B}-{\bf B}\otimes{\bf u})=\nabla\times({\bf B}\times{\bf u})$, the equation of the magnetic field ${\bf B}$ in the MHD system can be rewritten as (2.16) $\frac{\partial{\bf B}}{\partial t}+\nabla\times({\bf B}\times{\bf u})=0.$ Inserting (2.15) into (2.16), we further get $\nabla\times\left(\frac{\partial{\bf A}}{\partial t}+(\nabla\times{\bf A})\times{\bf u}\right)=0,$ which implies the existence of a scalar function $\psi$ such that $\frac{\partial{\bf A}}{\partial t}+(\nabla\times{\bf A})\times{\bf u}=-\nabla\psi.$ Stable solutions can be obtained by introducing a Weyl gauge, i.e. $\psi\equiv 0$ [29], and the evolution equation for the vector potential becomes $\frac{\partial{\bf A}}{\partial t}+(\nabla\times{\bf A})\times{\bf u}=0.$ It is easy to check that $\nabla\cdot{\bf B}=\nabla\cdot(\nabla\times{\bf A})=0$, i.e. the divergence-free condition will be satisfied for all times. In summary, we get the following MHD system: (2.17) $\left\\{\begin{aligned} &\frac{\partial\rho}{\partial t}+\nabla\cdot\left(\rho{\bf u}\right)=0,\\\ &\frac{\partial\rho{\bf u}}{\partial t}+\nabla\cdot\left(\rho\mathbf{u}\otimes\mathbf{u}+\left(\frac{p}{\varepsilon^{2}}+\frac{1}{2}{\|\mathbf{B}\|}^{2}\right)\mathbf{I}-\mathbf{B}\otimes\mathbf{B}\right)=\mathbf{0},\\\ &\frac{\partial{\bf A}}{\partial t}+(\nabla\times{\bf A})\times{\bf u}=\mathbf{0},\\\ &\frac{\partial E}{\partial t}+\nabla\cdot\left(\left(E+p+\frac{\varepsilon^{2}}{2}{\|\mathbf{B}\|}^{2}\right)\mathbf{u}-\varepsilon^{2}\mathbf{B}\left(\mathbf{u}\cdot\mathbf{B}\right)\right)=0,\\\ &{\bf B}-\nabla\times{\bf A}=\mathbf{0}.\\\ \end{aligned}\right.$ Corresponding to (2.17), with well-prepared initial conditions (2.14), the low sonic Mach limit incompressible MHD system becomes (2.18) $\left\\{\begin{aligned} &\frac{d\rho_{0}}{dt}=0,\quad p_{0}={\rm Const},\quad\nabla\cdot{\bf u}_{0}=0,\quad\nabla\cdot\mathbf{B}_{0}=0,\\\ &\frac{\partial\rho_{0}\mathbf{u}_{0}}{\partial t}+\nabla\cdot\left(\rho_{0}\mathbf{u}_{0}\otimes\mathbf{u}_{0}+\left(p_{2}+\frac{1}{2}{\|\mathbf{B}_{0}\|}^{2}\right)\mathbf{I}-\mathbf{B}_{0}\otimes\mathbf{B}_{0}\right)=\mathbf{0},\\\ &\frac{\partial{\bf A}_{0}}{\partial t}+(\nabla\times{\bf A}_{0})\times{\bf u}_{0}=\mathbf{0},\\\ &{\bf B}_{0}-\nabla\times{\bf A}_{0}=\mathbf{0}.\end{aligned}\right.$ ## 3\. Numerical schemes In this section, we will construct and analyze a high-order finite difference scheme with AP and divergence-free properties for the non-dimensionalized MHD system (2.17). The features of our scheme include the following: we design an SI IMEX-RK time discretization so that the scheme is stable with a time step constraint independent of the sonic Mach number $\varepsilon$; the AP property is preserved in the zero sonic Mach number limit; the scheme can be implemented efficiently without solving a nonlinear system. We will adopt the high order finite difference WENO scheme in [8, 10, 30], and the CT method in [18]. The final scheme can well capture discontinuities such as shocks in the compressible regime, while keeping a discrete divergence-free condition in the low sonic Mach regime, with a uniform time step condition independent of $\varepsilon$. We start with a first-order SI scheme. The high-order IMEX-RK scheme in time and a finite difference WENO reconstruction in space will be followed. After that, an overall scheme will be summarized. ### 3.1. First order SI IMEX scheme In the following, we start with a first-order SI IMEX time discretization for (2.17), while keeping space continuous: (3.1a) $\frac{\rho^{n+1}-\rho^{n}}{\Delta t}+\nabla\cdot\mathbf{q}^{n}=0,$ (3.1b) $\frac{\mathbf{q}^{n+1}-\mathbf{q}^{n}}{\Delta t}+\nabla\cdot\left(\frac{\mathbf{q}^{n}\otimes\mathbf{q}^{n}}{\rho^{n}}-\mathbf{B}^{n}\otimes\mathbf{B}^{n}+\left(\frac{1}{2}{\|\mathbf{B}^{n}\|}^{2}+p^{n}\right)\mathbf{I}\right)+\frac{1-\varepsilon^{2}}{\varepsilon^{2}}\nabla p^{n+1}=\mathbf{0},$ (3.1c) $\frac{\mathbf{A}^{n+1}-\mathbf{A}^{n}}{\Delta t}+\left(\nabla\times{\bf A}^{n}\right)\times\frac{\mathbf{q}^{n}}{\rho^{n}}=\mathbf{0},$ (3.1d) $\frac{E^{n+1}-E^{n}}{\Delta t}+\nabla\cdot\left(\frac{E^{n}+p^{n}}{\rho^{n+1}}\mathbf{q}^{n+1}+\varepsilon^{2}\left(\frac{1}{2}{\|\mathbf{B}^{n}\|}^{2}\frac{\mathbf{q}^{n}}{\rho^{n}}-\left(\frac{\mathbf{q}^{n}}{\rho^{n}}\cdot\mathbf{B}^{n}\right)\mathbf{B}^{n}\right)\right)=0,$ (3.1e) ${\bf B}^{n+1}-\nabla\times{\bf A}^{n+1}=\mathbf{0},$ where $\mathbf{q}=\rho{\bf u}$. The updating flow chart based on the SI IMEX scheme (3.1) is as follows: * • Step 1. Update $\rho^{n+1}$ and ${\bf A}^{n+1}$ from (3.1a) and (3.1c) respectively, then update ${\bf B}^{n+1}$ from (3.1e). * • Step 2. We rewrite (3.1b) as (3.2) $\mathbf{q}^{n+1}=\mathbf{q}^{*}-\Delta t\frac{1-\varepsilon^{2}}{\varepsilon^{2}}\nabla p^{n+1},$ with $\mathbf{q}^{*}=\mathbf{q}^{n}-\Delta t\nabla\cdot\left(\frac{\mathbf{q}^{n}\otimes\mathbf{q}^{n}}{\rho^{n}}-\mathbf{B}^{n}\otimes\mathbf{B}^{n}+\left(\frac{1}{2}{\|\mathbf{B}^{n}\|}^{2}+p^{n}\right)\mathbf{I}\right).$ Substituting $\mathbf{q}^{n+1}$ into (3.1d), we can further get (3.3) $E^{n+1}=E^{*}+\Delta t^{2}\frac{1-\varepsilon^{2}}{\varepsilon^{2}}\nabla\cdot(H^{n}\nabla p^{n+1}),$ where $H^{n}=(E^{n}+p^{n})/\rho^{n+1}$ and $E^{*}=E^{n}-\Delta t\nabla\cdot(H^{n}\mathbf{q}^{*})-\Delta t\varepsilon^{2}\nabla\cdot\left(\frac{1}{2}{\|\mathbf{B}^{n}\|}^{2}\frac{\mathbf{q}^{n}}{\rho^{n}}-\left(\frac{\mathbf{q}^{n}}{\rho^{n}}\cdot\mathbf{B}^{n}\right)\mathbf{B}^{n}\right).$ Corresponding to the pressure in the incompressible limit, we introduce a pressure perturbation $p^{n+1}_{2}$, defined as (3.4) $p^{n+1}_{2}=(p^{n+1}-\bar{p}^{n})/\varepsilon^{2},$ where $\bar{p}^{n}=\int_{\Omega}p^{n}d{\bf x}/|\Omega|$, and we replace $E^{n+1}$ by $p^{n+1}/(\gamma-1)+\varepsilon^{2}({\|\mathbf{q}^{n}\|}^{2}/(2\rho^{n})+{\|\mathbf{B}^{n}\|}^{2}/2)$ from (3.3). We rewrite (3.3) as (3.5) $\frac{\varepsilon^{2}}{\gamma-1}p^{n+1}_{2}-\Delta t^{2}(1-\varepsilon^{2})\nabla\cdot(H^{n}\nabla p^{n+1}_{2})=E^{**},$ where $E^{**}=E^{*}-\bar{p}^{n}/(\gamma-1)-\varepsilon^{2}({\|\mathbf{q}^{n}\|}^{2}/(2\rho^{n})+{\|\mathbf{B}^{n}\|}^{2}/2)$. We can now update $p_{2}^{n+1}$ from the elliptic equation (3.5). * • Step 3. We update $\mathbf{q}^{n+1}$ from (3.2) by using (3.4), and then $E^{n+1}$ from (3.1d). As described in [8], we may set $\nabla\cdot(\alpha p^{n}{\bf I})$ and $(1-\alpha\varepsilon^{2})\nabla p^{n+1}/\varepsilon^{2}$ to replace $\nabla\cdot(p^{n}{\bf I})$ and $(1-\varepsilon^{2})\nabla p^{n+1}/\varepsilon^{2}$ in (3.1b) with $\alpha=1$ for $\varepsilon<1$ and $\alpha=1/\varepsilon^{2}$ for $\varepsilon\geq 1$. Note that if $\varepsilon\geq 1$ the implicit pressure contribution in (3.1b) vanishes, so the momentum $\mathbf{q}^{n+1}$ is evaluated explicitly. With updated $\rho^{n+1}$, ${\bf A}^{n+1}$, ${\bf B}^{n+1}$, and $\mathbf{q}^{n+1}$, $E^{n+1}$ can also be updated in an explicit way from (3.1d). In one dimension, the divergence-free condition is satisfied automatically, so there is no need to apply a CT method. We can replace (3.1c) and (3.1e) with (3.6) $\frac{\mathbf{B}^{n+1}-\mathbf{B}^{n}}{\Delta t}+\nabla\cdot\left(\frac{\mathbf{q}^{n}\otimes\mathbf{B}^{n}}{\rho^{n}}-\frac{\mathbf{B}^{n}\otimes\mathbf{q}^{n}}{\rho^{n}}\right)=0.$ The first step in the flow chart becomes: update $\rho^{n+1}$ and ${\bf B}^{n+1}$ from (3.1a) and (3.6) respectively, and other steps remain unchanged. ### 3.2. High order SI IMEX-RK scheme In this subsection, we will follow a similar procedure as in [8, 10] to extend the first-order SI IMEX scheme (3.1) to high order. For simplicity, we only consider (2.6) with CT methods, namely, (2.17). First, (2.17) can be written as an autonomous system: $\left\\{\begin{aligned} &U_{t}=\mathcal{H}(U,U,{\bf B}),\\\ &{\bf B}=\nabla\times{\bf A},\end{aligned}\right.$ where $U={(\rho,\mathbf{q},{\bf A},E)}^{T}$ and $\mathcal{H}:\mathbb{R}^{n}\times\mathbb{R}^{n}\times\mathbb{R}^{m}\rightarrow\mathbb{R}^{n}$ is a sufficiently regular mapping. We use subscript “E” to express an explicit treatment for the first argument $U$ and the third one ${\bf B}$, and subscript “I” to express an implicit treatment for the second $U$. For the system (2.17), the function $\mathcal{H}(U_{E},U_{I},{\bf B}_{E})$ is defined as $\displaystyle\mathcal{H}(U_{E},U_{I},{\bf B}_{E})=$ $\displaystyle-\nabla\cdot\begin{pmatrix}\mathbf{q}_{E}\\\ \frac{\mathbf{q}_{E}\otimes\mathbf{q}_{E}}{\rho_{E}}-\mathbf{B}_{E}\otimes\mathbf{B}_{E}+(\frac{1}{2}{\|\mathbf{B}_{E}\|}^{2}+p_{E})\mathbf{I}\\\ \mathbf{0}\\\ 0\\\ \end{pmatrix}-\begin{pmatrix}0\\\ \mathbf{0}\\\ (\nabla\times{\bf A}_{E})\times\frac{\mathbf{q}_{E}}{\rho_{E}}\\\ 0\end{pmatrix}$ $\displaystyle-\nabla\cdot\begin{pmatrix}0\\\ \mathbf{0}\\\ \mathbf{0}\\\ \varepsilon^{2}(\frac{1}{2}{\|\mathbf{B}_{E}\|}^{2}\frac{\mathbf{q}_{E}}{\rho_{E}}-(\frac{\mathbf{q}_{E}}{\rho_{E}}\cdot\mathbf{B}_{E})\mathbf{B}_{E})\end{pmatrix}-\nabla\cdot\begin{pmatrix}0\\\ (1-\varepsilon^{2})p_{I,2}\mathbf{I}\\\ \mathbf{0}\\\ H_{I}\mathbf{q}_{I}\end{pmatrix}$ $\displaystyle=$ $\displaystyle-\nabla\cdot\mathcal{F}_{E}^{1}-\mathcal{G}_{E}-\nabla\cdot\mathcal{F}_{E}^{2}-\nabla\cdot\mathcal{F}_{SI},$ with $H_{I}=(E_{E}+p_{E})/\rho_{I}$ and $p_{I,2}=(p_{I}-\bar{p}_{E})/\varepsilon^{2}$. To obtain high-order accuracy in time, we can apply a multi-stage IMEX-RK time discretization with a double Butcher $tableau$, $\begin{array}[]{c|c}\tilde{c}&\tilde{A}\\\ \hline\cr\vspace{-0.25cm}\hfil\\\ &\tilde{b}^{T}\end{array}\ \ \ \ \ \qquad\begin{array}[]{c|c}{c}&{A}\\\ \hline\cr\vspace{-0.25cm}\hfil\\\ &{b^{T}}\end{array},$ where $\tilde{A}=(\tilde{a}_{ij})$ is an $s\times s$ matrix for an explicit scheme, with $\tilde{a}_{ij}=0$ for $j\geq i$, and $A=({a}_{ij})$ is an $s\times s$ matrix for an implicit scheme. For the implicit part of the methods, we use a diagonally implicit scheme, i.e. $a_{ij}=0$, for $j>i$, to guarantee simplicity and efficiency in solving the algebraic equations corresponding to the implicit part of the discretization. The vectors $\tilde{c}=(\tilde{c}_{1},...,\tilde{c}_{s})^{T}$, $\tilde{b}=(\tilde{b}_{1},...,\tilde{b}_{s})^{T}$, $c=(c_{1},...,c_{s})^{T}$, and $b=(b_{1},...,b_{s})^{T}$ complete the characterization of the scheme. The coefficients $\tilde{c}$ and $c$ are given by the usual relation $\displaystyle\tilde{c}_{i}=\sum_{j=1}^{i-1}\tilde{a}_{ij},\ \ \ c_{i}=\sum_{j=1}^{i}a_{ij}.$ Later, for the consideration of AP property, we consider implicit schemes with SA property, i.e., the implicit part of the Butcher table satisfies the condition $b^{T}=e_{s}^{T}A$, with $e_{s}=(0,\cdots 0,1)^{T}$. We will see that the SA property guarantees that the numerical solution is identical to the last internal stage value of the scheme. Now, we may update the solutions as follows. Starting from $U_{E}^{(0)}=U_{I}^{(0)}=U^{n}$ and ${\bf B}_{E}^{(0)}={\bf B}^{n}$, for inner stages $i=1\text{ to }s$: * • First update the solution $U_{E}^{(i)}$ for the explicit part (3.7) $U_{E}^{(i)}=U^{n}+\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\mathcal{H}(U_{E}^{(j)},U_{I}^{(j)},{\bf B}_{E}^{(j)}),\quad{\bf B}_{E}^{(i)}=\nabla\times{\bf A}_{E}^{(i)}.$ * • Update the known values for the implicit part $U_{*}^{(i)}$, where (3.8) $U_{*}^{(i)}=U^{n}+\Delta t\sum^{i-1}_{j=1}a_{ij}\mathcal{H}(U_{E}^{(j)},U_{I}^{(j)},{\bf B}_{E}^{(j)}),$ and then solve (3.9) $U_{I}^{(i)}=U_{*}^{(i)}+\Delta ta_{ii}\mathcal{H}(U_{E}^{(i)},U_{I}^{(i)},{\bf B}_{E}^{(i)}).$ * • Finally, compute the solutions $U^{n+1}$ and ${\bf B}^{n+1}$ at time level $t^{n+1}$ from (3.10) $U^{n+1}=U^{n}+\Delta t\sum_{i=1}^{s}b_{i}\mathcal{H}(U_{E}^{(i)},U_{I}^{(i)},{\bf B}_{E}^{(i)}),\quad{\bf B}^{n+1}=\nabla\times{\bf A}^{n+1}.$ The first order scheme (3.1) is the same as applying the following Butcher table to (2.17): $\begin{array}[]{c|c}0&0\\\ \hline\cr&1\end{array}\qquad\qquad\begin{array}[]{c|c}1&1\\\ \hline\cr&1\end{array},$ namely $U_{E}=U^{n}$ and $U_{I}=U^{n+1}$. ### 3.3. High order finite difference WENO scheme Below we will describe our spatial discretization strategies which incorporate a WENO mechanism in order to capture shocks in the compressible regime and produce a high order incompressible solver for the flow in the zero Mach limit. In particular, we propose to apply a fifth order characteristic-wise WENO procedure to $\mathcal{F}_{E}^{1}$ of the ideal MHD system so that in the compressible regime numerical oscillations could be well controlled, and to apply a component-wise WENO procedure for the flux functions of $\mathcal{F}_{E}^{2}$ and $\mathcal{F}_{SI}$. As for $\mathcal{G}_{E}$, which is a Hamiltonian, a WENO procedure described in [31] will be applied. * • Component-wise WENO $\nabla_{W}$ for $\nabla\cdot\mathcal{F}_{E}^{2}$ and $\nabla\cdot\mathcal{F}_{SI}$: Let ${\bf v}=(\rho,\mathbf{q},{\bf B},E)^{T}$ and $\mathbf{f}=\mathbf{f}({\bf v})$ be the unknown variables and flux of the ideal MHD system (2.1). For simplicity, we consider the 1D case as an example, and we take a uniform mesh with $N+1$ grid points: $a=x_{0}<x_{1}<\cdots<x_{N}=b$. A conservative finite difference WENO scheme for system (2.1) can be written in the following flux- difference form: $(\mathbf{v}_{i})_{t}=\frac{1}{\Delta x}(\hat{\mathbf{f}}_{i+\frac{1}{2}}-\hat{\mathbf{f}}_{i-\frac{1}{2}}),$ where $\hat{\mathbf{f}}_{i\pm\frac{1}{2}}$ are the numerical fluxes, which can be reconstructed by the following procedures: 1. (1) Compute the physical flux at each grid point: $\mathbf{f}_{i}=\mathbf{f}({\bf v}_{i})$. 2. (2) Perform a Lax–Friedrichs flux vector splitting for each component of the physical variables to get (3.11) $\mathbf{f}^{\pm}_{i}=\frac{1}{2}(\mathbf{f}_{i}\pm\alpha\mathbf{v}_{i}),$ with $\alpha=\mathop{\max}\limits_{0\leq k\leq N}(|{\bf u}_{k}|+\hat{c}_{f}^{(k)})$, where (3.12) $\left\\{\begin{aligned} &\hat{a}^{(k)}:={\rm min}\left(\frac{1}{\varepsilon},1\right)\sqrt{\frac{\gamma p_{k}}{\rho_{k}}},\quad\hat{c}_{a}^{(k)}:=\sqrt{\frac{{({\bf B}_{k}\cdot{\bf n})}^{2}}{\rho_{k}}},\\\ &\hat{c}_{f,s}^{(k)}:={\left\\{\frac{1}{2}\left[\left(\hat{a}^{(k)}\right)^{2}+\frac{{\|{\bf B}_{k}\|}^{2}}{\rho_{k}}\pm\sqrt{{\left(\left(\hat{a}^{(k)}\right)^{2}+\frac{{\|{\bf B}_{k}\|}^{2}}{\rho_{k}}\right)}^{2}-\left(2{\hat{a}^{(k)}}{\hat{c}^{(k)}_{a}}\right)^{2}}\right]\right\\}}^{\frac{1}{2}}.\end{aligned}\right.$ 3. (3) Perform a finite difference WENO reconstruction to obtain upwind and downwind numerical fluxes: $\hat{\mathbf{f}}^{+}_{i+\frac{1}{2}}=\Phi_{\rm WENO5}(\mathbf{f}^{+}_{i-2},\mathbf{f}^{+}_{i-1},\mathbf{f}^{+}_{i},\mathbf{f}^{+}_{i+1},\mathbf{f}^{+}_{i+2}),\quad\hat{\mathbf{f}}^{-}_{i+\frac{1}{2}}=\Phi_{\rm WENO5}(\mathbf{f}^{-}_{i+3},\mathbf{f}^{-}_{i+2},\mathbf{\mathbf{f}}^{-}_{i+1},\mathbf{f}^{-}_{i},\mathbf{f}^{-}_{i-1}).$ Here $\Phi_{\rm WENO5}$ denotes a 5th-order WENO reconstruction [31], while other orders can also be used. Then we set $\hat{\mathbf{f}}_{i+\frac{1}{2}}=\hat{\mathbf{f}}^{+}_{i+\frac{1}{2}}+\hat{\mathbf{f}}^{-}_{i+\frac{1}{2}}.$ * • Characteristic-wise WENO $\nabla_{CW}$ for $\nabla\cdot\mathcal{F}_{E}^{1}$: The Jacobian matrix for the flux function $\partial\mathbf{f}/\partial\mathbf{v}$, has a spectral decomposition of the form $\partial\mathbf{f}/\partial\mathbf{v}=RDL$ [18], where $D$ is a diagonal matrix of real eigenvalues, $R$ is the matrix of right eigenvectors, and $L=R^{-1}$ is the matrix of left eigenvectors. Then, for a characteristic-wise WENO reconstruction $\nabla_{CW}$, we have the following procedures: 1. (1) Compute the physical flux at each grid point: $\mathbf{f}_{i}=\mathbf{f}({\bf v}_{i})$. 2. (2) Compute the average state ${\bf v}_{i+\frac{1}{2}}$ by using an arithmetic mean: ${\bf v}_{i+\frac{1}{2}}=\frac{1}{2}({\bf v}_{i}+{\bf v}_{i+1}).$ 3. (3) Compute the right and left eigenvectors of the flux Jacobian matrix $\partial\mathbf{f}/\partial\mathbf{v}$ by taking ${\bf v}={\bf v}_{i+\frac{1}{2}}$ at $x=x_{i+\frac{1}{2}}$: $R_{i+\frac{1}{2}}=R({\bf v}_{i+\frac{1}{2}}),\quad L_{i+\frac{1}{2}}=L({\bf v}_{i+\frac{1}{2}}).$ 4. (4) Project the solution and physical flux into the characteristic space: $\mathbf{w}_{j}=L_{i+\frac{1}{2}}{\bf v}_{j},\quad\mathbf{g}_{j}=L_{i+\frac{1}{2}}\mathbf{f}_{j},$ for all $j$ in the numerical stencil associated with $x=x_{i+\frac{1}{2}}$. In the case of a 5th-order finite difference WENO scheme: $j=i-2,i-1,i,i+1,i+2,i+3$. 5. (5) Perform a Lax–Friedrichs flux vector splitting for each component of the characteristic variables to compute $\mathbf{g}^{\pm}_{j}=\frac{1}{2}(\mathbf{g}_{j}\pm\alpha\mathbf{w}_{j}),$ where $\alpha$ has the same expression in the component-wise WENO (3.11). 6. (6) Perform a WENO reconstruction $\Phi_{\rm WENO5}$ on each component of $\mathbf{g}^{\pm}_{j}$ to obtain their corresponding numerical fluxes, and then set $\hat{\mathbf{g}}_{i+\frac{1}{2}}=\hat{\mathbf{g}}^{+}_{i+\frac{1}{2}}+\hat{\mathbf{g}}^{-}_{i+\frac{1}{2}}.$ 7. (7) Project the numerical flux $\hat{\mathbf{g}}_{i+\frac{1}{2}}$ back to the physical space: $\hat{\mathbf{f}}_{i+\frac{1}{2}}=R_{i+\frac{1}{2}}\hat{\mathbf{g}}_{i+\frac{1}{2}}.$ However, when the sonic Mach number $\varepsilon$ is small, the fast eigenvalue $c_{f}$ will make such a spectral characteristic decomposition to be unstable. Since a characteristic decomposition is mainly required in the compressible regime where $\varepsilon=\mathcal{O}(1)$, while it does not make any significant differences for small Mach numbers; in this case we directly use the eigenvectors in the expression of (2.1) by taking $\varepsilon=1$, which works well from our numerical results. * • A Hamiltonian WENO reconstruction WENO-HJ $(\cdot)_{HJ}$ for $\mathcal{G}_{E}$: We take a 1D Hamilton–Jacobi equation as an example (3.13) $v_{t}+\mathcal{HT}(t,x,v,v_{x})=0,$ where $\mathcal{HT}$ is the Hamiltonian. As described in [31], a semi-discrete form for (3.13) is $v_{t,i}=-\widehat{\mathcal{HT}}(t,x_{i},v_{i},v^{-}_{x,i},v^{+}_{x,i}),$ where $\widehat{\mathcal{HT}}$ is a numerical Hamiltonian which approximates (3.13), and $v^{-}_{x,i},v^{+}_{x,i}$ are upwind and downwind approximations of $v_{x}$ at $x=x_{i}$. A Lax-Friedrichs Hamiltonian from [31] is given as follows (3.14) $\widehat{\mathcal{HT}}\left(t,x,v,u^{-},u^{+}\right)=\mathcal{HT}\left(t,x,v,\frac{u^{-}+u^{+}}{2}\right)-\beta\left(\frac{u^{+}-u^{-}}{2}\right),$ where $\beta=\mathop{\max}\limits_{u\in I(u^{-},u^{+})}|\mathcal{HT}_{u}(t,x,v,u)|,$ and $I(u^{-},u^{+})$ is the interval between $u^{-}$ and $u^{+}$. The values of $v^{+}_{x,i}$ and $v^{-}_{x,i}$ are obtained from $\displaystyle v^{+}_{x,i}$ $\displaystyle=\Phi_{\rm WENO5}\left(\frac{\Delta^{+}v_{i+2}}{\Delta x},\frac{\Delta^{+}v_{i+1}}{\Delta x},\frac{\Delta^{+}v_{i}}{\Delta x},\frac{\Delta^{+}v_{i-1}}{\Delta x},\frac{\Delta^{+}v_{i-2}}{\Delta x}\right),$ $\displaystyle v^{-}_{x,i}$ $\displaystyle=\Phi_{\rm WENO5}\left(\frac{\Delta^{+}v_{i-3}}{\Delta x},\frac{\Delta^{+}v_{i-2}}{\Delta x},\frac{\Delta^{+}v_{i-1}}{\Delta x},\frac{\Delta^{+}v_{i}}{\Delta x},\frac{\Delta^{+}v_{i+1}}{\Delta x}\right),$ with $\Delta^{+}v_{j}=v_{j+1}-v_{j}$. Such a Hamiltonian WENO reconstruction helps us to control unphysical oscillations both in $v_{x}$ and $v$. For $\mathcal{G}_{E}$ in 2D, the Hamiltonian is $\mathcal{HT}\left(t,x,y,\frac{\partial A_{z}}{\partial x},\frac{\partial A_{z}}{\partial y}\right)=u(t,x,y)\frac{\partial A_{z}}{\partial x}+v(t,x,y)\frac{\partial A_{z}}{\partial y},$ and the numerical Hamiltonian (3.14) can be applied along each direction for $u(t,x,y)\frac{\partial A_{z}}{\partial x}$ and $v(t,x,y)\frac{\partial A_{3}}{\partial y}$, respectively. Here $A_{z}$ is the third component of ${\bf A}$. ### 3.4. Flowchart of a high order SI IMEX-RK finite difference WENO scheme With the above space and time discretizations, we now summarize our high order scheme as follows: * • Step 1. Starting from $U^{n}$ at time $t^{n}$, for an intermediate stage $i$ ($1\leq i\leq s$), we first compute $U^{(i)}_{E}$ from (3.7) $\rho^{(i)}_{E}=\rho^{n}-\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\nabla_{CW}\cdot\mathbf{q}^{(j)}_{E},$ $\mathbf{q}^{(i)}_{E}=\mathbf{q}^{n}-\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\left(\nabla_{CW}\cdot\left(\frac{\mathbf{q}^{(j)}_{E}\otimes\mathbf{q}^{(j)}_{E}}{\rho^{(j)}_{E}}-\mathbf{B}^{(j)}_{E}\otimes\mathbf{B}^{(j)}_{E}+\left(\frac{1}{2}{\|\mathbf{B}^{(j)}_{E}\|}^{2}+p^{(j)}_{E}\right)\mathbf{I}\right)+(1-\varepsilon^{2})\nabla_{W}p^{(j)}_{I,2}\right),$ $\mathbf{A}^{(i)}_{E}=\mathbf{A}^{n}-\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\left(\left(\nabla\times{\bf A}^{(j)}_{E}\right)\times\frac{\mathbf{q}^{(j)}_{E}}{\rho^{(j)}_{E}}\right)_{HJ},$ $E^{(i)}_{E}=E^{n}-\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\left(\nabla_{W}\cdot\left(\varepsilon^{2}\left(\frac{1}{2}{\|\mathbf{B}^{(j)}_{E}\|}^{2}\frac{\mathbf{q}^{(j)}_{E}}{\rho^{(j)}_{E}}-\left(\frac{\mathbf{q}^{(j)}_{E}}{\rho^{(j)}_{E}}\cdot\mathbf{B}^{(j)}_{E}\right)\mathbf{B}^{(j)}_{E}\right)\right)+\nabla_{W}\cdot\left(\bar{H}^{(j)}\mathbf{q}^{(j)}_{I}\right)\right),$ with $\bar{H}^{(j)}=(E_{E}^{(j)}+p_{E}^{(j)})/\rho_{I}^{(j)}$. Then we obtain ${\bf B}_{E}^{(i)}=\nabla\times{\bf A}_{E}^{(i)},$ with a 4th-order central difference discretization. * • Step 2. We compute $U^{(i)}_{*}$ in (3.8) $\rho^{(i)}_{*}=\rho^{n}-\Delta t\sum^{i-1}_{j=1}a_{ij}\nabla_{CW}\cdot\mathbf{q}^{(j)}_{E},$ $\mathbf{q}^{(i)}_{*}=\mathbf{q}^{n}-\Delta t\sum^{i-1}_{j=1}a_{ij}\left(\nabla_{CW}\cdot\left(\frac{\mathbf{q}^{(j)}_{E}\otimes\mathbf{q}^{(j)}_{E}}{\rho^{(j)}_{E}}-\mathbf{B}^{(j)}_{E}\otimes\mathbf{B}^{(j)}_{E}+\left(\frac{1}{2}{\|\mathbf{B}^{(j)}_{E}\|}^{2}+p^{(j)}_{E}\right)\mathbf{I}\right)+(1-\varepsilon^{2})\nabla_{W}p^{(j)}_{I,2}\right),$ $E^{(i)}_{*}=E^{n}-\Delta t\sum^{i-1}_{j=1}a_{ij}\left(\nabla_{W}\cdot\left(\varepsilon^{2}\left(\frac{1}{2}{\|\mathbf{B}^{(j)}_{E}\|}^{2}\frac{\mathbf{q}^{(j)}_{E}}{\rho^{(j)}_{E}}-\left(\frac{\mathbf{q}^{(j)}_{E}}{\rho^{(j)}_{E}}\cdot\mathbf{B}^{(j)}_{E}\right)\mathbf{B}^{(j)}_{E}\right)\right)+\nabla_{W}\cdot\left(\bar{H}^{(j)}\mathbf{q}^{(j)}_{I}\right)\right).$ * • Step 3. Solve $U_{I}^{(i)}$ from (3.9). 1. (1) In components, $U_{I}^{(i)}$ satisfies (3.15a) $\rho^{(i)}_{I}=\rho^{(i)}_{*}-\Delta ta_{ii}\nabla_{CW}\cdot\mathbf{q}^{(i)}_{E},$ (3.15b) $\mathbf{q}^{(i)}_{I}=\mathbf{q}^{(i)}_{**}-\Delta ta_{ii}(\frac{1-\varepsilon^{2}}{\varepsilon^{2}}\nabla p^{(i)}_{I}),$ (3.15c) $E^{(i)}_{I}=E^{(i)}_{**}-\Delta ta_{ii}\nabla\cdot(\bar{H}^{(i)}\mathbf{q}^{(i)}_{I}),$ where $\mathbf{q}^{(i)}_{**}=\mathbf{q}^{(i)}_{*}-\Delta ta_{ii}\nabla_{CW}\cdot\left(\frac{\mathbf{q}^{(i)}_{E}\otimes\mathbf{q}^{(i)}_{E}}{\rho^{(i)}_{E}}-\mathbf{B}^{(i)}_{E}\otimes\mathbf{B}^{(i)}_{E}+\left(\frac{1}{2}{\|\mathbf{B}^{(i)}_{E}\|}^{2}+p^{(i)}_{E}\right)\mathbf{I}\right),$ $E^{(i)}_{**}=E^{(i)}_{*}-\Delta ta_{ii}\nabla_{W}\cdot\left(\varepsilon^{2}\left(\frac{1}{2}{\|\mathbf{B}^{(i)}_{E}\|}^{2}\frac{\mathbf{q}^{(i)}_{E}}{\rho^{(i)}_{E}}-\left(\frac{\mathbf{q}^{(i)}_{E}}{\rho^{(i)}_{E}}\cdot\mathbf{B}^{(i)}_{E}\right)\mathbf{B}^{(i)}_{E}\right)\right).$ 2. (2) We can obtain $\rho_{I}^{(i)}$ from (3.15a) directly. To solve (3.15b) and (3.15c), we substitute $\mathbf{q}^{(i)}_{I}$ of (3.15b) into (3.15c), and obtain (3.16) $E^{(i)}_{I}=E^{\circ}_{I}+\left(1-\varepsilon^{2}\right)\Delta t^{2}a^{2}_{ii}\nabla\cdot\left(\bar{H}^{(i)}\left(\frac{\nabla p^{(i)}_{I}}{\varepsilon^{2}}\right)\right),$ with $E_{I}^{(i)}$ following the EOS $E_{I}^{(i)}=p_{I}^{(i)}/(\gamma-1)+\varepsilon^{2}({\|\mathbf{q}_{E}^{(i)}\|}^{2}/(2\rho_{E}^{(i)})+{\|\mathbf{B}_{E}^{(i)}\|}^{2}/2)$ and $E^{\circ}_{I}=E^{(i)}_{**}-\Delta ta_{ii}\nabla_{W}\cdot(\bar{H}^{(i)}\mathbf{q}^{(i)}_{**})$. By substituting (3.17) $p^{(i)}_{I}=\bar{p}^{(i)}_{E}+\varepsilon^{2}p^{(i)}_{I,2}$ into (3.16), where $\bar{p}^{(i)}_{E}=\int_{\Omega}p_{E}^{(i)}d{\bf x}/|\Omega|$, we solve (3.18) $\frac{\varepsilon^{2}}{\gamma-1}p^{(i)}_{I,2}-\left(1-\varepsilon^{2}\right)\Delta t^{2}a^{2}_{ii}\nabla\cdot\left(\bar{H}^{(i)}\left(\nabla p^{(i)}_{I,2}\right)\right)=E^{\circ\circ}_{I},$ with $E^{\circ\circ}_{I}=E^{\circ}_{I}-\bar{p}^{(i)}_{E}/(\gamma-1)-\varepsilon^{2}({\|\mathbf{q}^{(i)}_{E}\|}^{2}/(2\rho^{(i)}_{E})+{\|\mathbf{B}^{(i)}_{E}\|}^{2}/2)$. Notice that in the process of substitution to obtain (3.16) or (3.18), the gradient and the divergence operators are kept to be continuous, obtaining the second order operator $\nabla\cdot(\bar{H}^{(i)}\nabla p^{(i)}_{I,2})$. The second-order spatial derivative is then discretized by a compact discretization as proposed in [10]. 3. (3) With $p^{(i)}_{I,2}$ solved from (3.18), we can update $\mathbf{q}^{(i)}_{I}$ from $\mathbf{q}^{(i)}_{I}=\mathbf{q}^{(i)}_{**}-\Delta ta_{ii}\left(\left(1-\varepsilon^{2}\right)\nabla_{W}p^{(i)}_{I,2}\right),$ and successively update $E^{(i)}_{I}$ from $E^{(i)}_{I}=E^{(i)}_{**}-\Delta ta_{ii}\nabla_{W}\cdot\left(\bar{H}^{(i)}\mathbf{q}^{(i)}_{I}\right).$ * • Step 4. Finally, we update the numerical solution $U^{n+1}=U^{(s)}_{I}$, and ${\bf B}^{n+1}=\nabla\times{\bf A}^{n+1}$. ## 4\. Asymptotic preserving (AP) and Asymptotically Accurate (AA) properties ### 4.1. AP property In this section, we will prove the AP property for the first order SI IMEX scheme (3.1), and the AA property for the high order SI IMEX-RK scheme (3.7)-(3.10). We focus on time discretizations while keeping the space continuous when we discuss the AP or AA property. First, we have the following theorem: ###### Theorem 4.1. The first order SI IMEX scheme (3.1) with space continuous is AP, in the sense that, with well-prepared initial conditions (2.14), at the leading order asymptotic expansions, the scheme (3.1) is a consistent approximation of the incompressible MHD equations (LABEL:S2_E26). ###### Proof. We assume the following expansions of the solutions at time $t^{n}$: (4.1) $\left\\{\begin{aligned} &p^{n}(\mathbf{x})=p^{n}_{0}+\varepsilon^{2}p^{n}_{2}(\mathbf{x}),\quad\rho^{n}(\mathbf{x})=\rho^{n}_{0}(\mathbf{x})+\mathcal{O}(\varepsilon),\\\ &\mathbf{u}^{n}(\mathbf{x})=\mathbf{u}^{n}_{0}(\mathbf{x})+\mathcal{O}(\varepsilon^{2}),\quad\mathbf{A}^{n}(\mathbf{x})=\mathbf{A}^{n}_{0}(\mathbf{x})+\mathcal{O}(\varepsilon),\\\ &\mathbf{B}^{n}(\mathbf{x})=\mathbf{B}^{n}_{0}(\mathbf{x})+\mathcal{O}(\varepsilon),&\end{aligned}\right.$ with periodic or no-slip boundary conditions, and well-prepared conditions $p_{0}^{n}=(\gamma-1)E^{n}_{0}={\rm Const}$, $\nabla\cdot\mathbf{u}^{n}_{0}(\mathbf{x})=0$ and $\nabla\cdot{\bf B}^{n}({\bf x})=0$. No matter what value ${\bf A}^{n+1}$ takes, we can obtain $\nabla\cdot{\bf B}^{n+1}=\nabla\cdot(\nabla\times{\bf A}^{n+1})=0$ from (3.1e). We plug (4.1) into the semi-discrete scheme (3.1) and EOS (3.3). For the leading order $\mathcal{O}(\varepsilon^{-2})$, we obtain $\nabla p_{0}^{n+1}=0,$ i.e. $p_{0}^{n+1}$ is constant in space. Equating to zero the $\mathcal{O}(\varepsilon^{0})$ terms, we have the following equations: (4.2a) $\frac{\rho_{0}^{n+1}-\rho_{0}^{n}}{\Delta t}+\nabla\cdot(\rho_{0}^{n}{\bf u}_{0}^{n})=0,$ (4.2b) $\frac{\rho_{0}^{n+1}\mathbf{u}_{0}^{n+1}-\rho_{0}^{n}\mathbf{u}_{0}^{n}}{\Delta t}+\nabla\cdot\left(\rho_{0}^{n}{\bf u}_{0}^{n}\otimes{\bf u}_{0}^{n}-\mathbf{B}_{0}^{n}\otimes\mathbf{B}_{0}^{n}+\frac{1}{2}{\|\mathbf{B}_{0}^{n}\|}^{2}\mathbf{I}\right)+\nabla p^{n+1}_{2}=\mathbf{0},$ (4.2c) $\frac{\mathbf{A}_{0}^{n+1}-\mathbf{A}_{0}^{n}}{\Delta t}+(\nabla\times{\bf A}_{0}^{n})\times{\bf u}_{0}^{n}=\mathbf{0},$ (4.2d) $\frac{E_{0}^{n+1}-E_{0}^{n}}{\Delta t}+\nabla\cdot\left(\bar{H}_{0}^{n}\left(\rho_{0}^{n+1}{\bf u}_{0}^{n+1}\right)\right)=0,$ (4.2e) ${\bf B}_{0}^{n+1}-\nabla\times{\bf A}_{0}^{n+1}=\mathbf{0},$ (4.2f) $E_{0}^{n+1}=\frac{p_{0}^{n+1}}{\gamma-1},$ with (4.3) $\bar{H}_{0}^{n}=\frac{p_{0}^{n}+E_{0}^{n}}{\rho_{0}^{n+1}}=\frac{\gamma}{\gamma-1}\frac{p_{0}^{n}}{\rho_{0}^{n+1}}.$ Due to (4.2f) and (4.3), it is easy to check that (4.2d) is equivalent to (4.4) $\frac{p_{0}^{n+1}-p_{0}^{n}}{\Delta t}+\gamma p_{0}^{n}\nabla\cdot{\bf u}_{0}^{n+1}=0.$ Taking into account a spatial domain $\Omega$ with periodic or no-slip boundary conditions and integrating (4.4) over $\Omega$, we further get $\int_{\Omega}\nabla\cdot{\bf u}_{0}^{n+1}d{\bf x}=\int_{\partial\Omega}{\bf u}_{0}^{n+1}\cdot{\bf n}ds=0$, where ${\bf n}$ is the unit outward normal vector along $\partial\Omega$. This implies $p_{0}^{n+1}=p^{n}_{0}={\rm Const}$. A direct result $\nabla\cdot{\bf u}_{0}^{n+1}=0$ can be obtained through combing (4.4) with $p_{0}^{n+1}=p^{n}_{0}$. The equations (4.2a)-(4.2c) and (4.2e) correspond to consistent discretizations for other equations in (LABEL:S2_E26), so that the AP property is obtained. ∎ ### 4.2. AA property Similarly to the AP property, here we focus on the AA analysis on time discretizations, while keeping space continuous. Then we have the following theorem: ###### Theorem 4.2. We consider a high order SI IMEX-RK scheme (3.7)-(3.10) of order r applied to system (2.17) on a bounded domain $\Omega\subset\mathbb{R}$ with periodic or compact support boundary conditions, and assume that the implicit part of the SI IMEX-RK scheme is SA and the initial conditions $(\rho^{0}({\bf x}),\rho^{0}({\bf x}){\bf u}^{0}({\bf x}),p^{0},$ ${\bf A}^{0}({\bf x}),{\bf B}^{0}({\bf x}))^{T}$ are well-prepared, namely in the form of (4.1). If we denote $(\rho^{1}({\bf x};\varepsilon),\rho^{1}({\bf x};\varepsilon){\bf u}^{1}({\bf x};\varepsilon),$ $p^{1}({\bf x};\varepsilon),{\bf A}^{1}({\bf x};\varepsilon),{\bf B}^{1}({\bf x};\varepsilon))^{T}$ as the numerical solution after one-time step, then, with $p_{*}$ being a constant, we have: (4.5) $\lim_{\varepsilon\to 0}p^{1}({\bf x};\varepsilon)=p_{*},\quad\lim_{\varepsilon\to 0}\nabla\cdot{\bf u}^{1}({\bf x};\varepsilon)=0,\quad\nabla\cdot{\bf B}^{1}({\bf x};\varepsilon)=0.$ Furthermore, let ${\bf V}_{inc}({\bf x};t)=(\rho_{inc}({\bf x};t),\rho_{inc}({\bf x};t){\bf u}_{inc}({\bf x};t),p_{inc}({\bf x};t),{\bf A}_{inc}({\bf x};t),{\bf B}_{inc}({\bf x};t))^{T}$ be the exact solution of the incompressible MHD equations (LABEL:S2_E26) with the same initial data, one has the following one-step error estimate (4.6) $\lim_{\varepsilon\to 0}{\bf V}^{1}({\bf x};\varepsilon)={\bf V}_{inc}({\bf x};\Delta t)+\mathcal{O}(\Delta t^{r+1}),$ i.e. the scheme is AA. ###### Proof. We consider the first step from $t^{0}=0$ to $t^{1}=\Delta t$ for the SI IMEX- RK scheme (3.7)-(3.10) of order $r$ applied to system (2.17) with well- prepared initial data (4.1): $\left\\{\begin{aligned} &p^{0}(\mathbf{x})=p_{*}+\varepsilon^{2}p^{0}_{2}(\mathbf{x}),\quad\rho^{0}(\mathbf{x})=\rho^{0}_{inc}+\mathcal{O}(\varepsilon),\\\ &\mathbf{u}^{0}(\mathbf{x})=\mathbf{u}^{0}_{inc}+\mathcal{O}(\varepsilon^{2}),\quad\mathbf{A}^{0}(\mathbf{x})=\mathbf{A}^{0}_{inc}+\mathcal{O}(\varepsilon),\quad\mathbf{B}^{0}(\mathbf{x})=\mathbf{B}^{0}_{inc}+\mathcal{O}(\varepsilon),\end{aligned}\right.$ where $\displaystyle\rho^{0}_{inc}:=\rho_{inc}({\bf x},0),\quad\mathbf{u}^{0}_{inc}:=\mathbf{u}_{inc}({\bf x},0),\quad\mathbf{A}^{0}_{inc}:=\mathbf{A}_{inc}({\bf x},0),\quad\mathbf{B}^{0}_{inc}:=\mathbf{B}_{inc}({\bf x},0).$ By well-prepared assumptions we have: $p_{inc}({\bf x},0):=p_{*}$ is a constant independent of time and space, $\nabla\cdot{\bf u}_{inc}^{0}=0$ and $\nabla\cdot{\bf B}^{0}({\bf x})=0$. Now we consider a formal $\varepsilon$-expansion of the quantities: $U_{I}^{(i)}=(\rho_{I}^{(i)},\mathbf{q}_{I}^{(i)},{\bf A}_{I}^{(i)},E_{I}^{(i)}),\quad U_{E}^{(i)}=(\rho_{E}^{(i)},\mathbf{q}_{E}^{(i)},{\bf A}_{E}^{(i)},E_{E}^{(i)}),$ and ${\bf B}_{E}^{(i)}$, with $\mathbf{q}_{I}^{(i)}=\rho_{I}^{(i)}{\bf u}_{I}^{(i)}$ and $\mathbf{q}_{E}^{(i)}=\rho_{E}^{(i)}{\bf u}_{E}^{(i)}$. For example, the expansions of the density and pressure are (4.7) $\left\\{\begin{aligned} &\rho_{I}^{(i)}=\rho_{0,I}^{(i)}+\varepsilon\rho_{1,I}^{(i)}+\cdots,\quad&\rho_{E}^{(i)}&=\rho_{0,E}^{(i)}+\varepsilon\rho_{1,E}^{(i)}+\cdots,\\\ &p_{I}^{(i)}=p_{0,I}^{(i)}+\varepsilon^{2}p_{2,I}^{(i)}+\cdots,\quad&p_{E}^{(i)}&=p_{0,E}^{(i)}+\varepsilon^{2}p_{2,E}^{(i)}+\cdots.\end{aligned}\right.$ To prove the theorem, we use mathematical induction. * • AA property for internal stages $i=1,\cdots,s.$ When $i=1$, it leads to the same AP analysis for the scheme (3.1) with $\Delta t$ replaced by $a_{11}\Delta t$. To prove the result for $i>1$, we make use of the induction hypothesis, assuming the property holds for $j\leq i-1$, and prove that it still holds for $j=i$. For $j=1,\cdots,i-1$, we have (4.8) $p_{0,E}^{(j)}=p_{0,I}^{(j)}=p_{*},\quad E_{0,E}^{(j)}=E_{0,I}^{(j)}=\frac{p_{*}}{\gamma-1},\quad\nabla\cdot{\bf u}_{0,I}^{(j)}=0,\quad\nabla\cdot{\bf B}_{E}^{(j)}=0.$ Now we insert the expansions (4.7) into the explicit step in (3.7), and from the energy equation we get (4.9) $E_{0,E}^{(i)}=E_{inc}^{0}-\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\nabla\cdot\left(\bar{H}_{0}^{(j)}\mathbf{q}_{0,I}^{(j)}\right),$ with $E_{inc}^{0}=p_{*}/(\gamma-1)$ and for $j=1,\cdots,i-1$, (4.10) $\bar{H}_{0}^{(j)}=\frac{E_{0,E}^{(j)}+p_{0,E}^{(j)}}{\rho_{0,I}^{(j)}}=\frac{\gamma}{\gamma-1}\frac{p_{*}}{\rho_{0,I}^{(j)}}.$ Now substituting (4.8), (4.10), and $E_{0,E}^{(i)}=p_{0,E}^{(i)}/(\gamma-1)$ into (4.9), we obtain $p_{0,E}^{(i)}=p_{*}-\Delta t\gamma p_{*}\sum^{i-1}_{j=1}\tilde{a}_{ij}\nabla\cdot{\bf u}_{0,I}^{(j)}=p_{*},$ and $E_{0,E}^{(i)}=p_{*}/(\gamma-1)$ is also a constant for the stage $i$. From (3.7), we have (4.11) $\left\\{\begin{aligned} &\mathbf{A}^{(i)}_{E}=\mathbf{A}^{n}-\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\left(\left(\nabla\times{\bf A}^{(j)}_{E}\right)\times\mathbf{u}^{(j)}_{E}\right),\\\ &{\bf B}_{E}^{(i)}=\nabla\times{\bf A}_{E}^{(i)}.\end{aligned}\right.$ From (LABEL:416), we get $\nabla\cdot{\bf B}_{E}^{(i)}=\nabla\cdot\left(\nabla\times{\bf A}_{E}^{(i)}\right)=0$. From (3.7), in the order of $\mathcal{O}(1)$, the density and momentum equations yield (4.12) $\left\\{\begin{aligned} &\rho_{0,E}^{(i)}=\rho_{inc}^{0}-\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\nabla\cdot\mathbf{q}_{0,E}^{(j)},\\\ &\mathbf{q}^{(i)}_{0,E}=\mathbf{q}^{0}_{inc}-\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\left(\nabla\cdot\left(\rho_{0,E}^{(j)}\mathbf{u}_{0,E}^{(j)}\otimes\mathbf{u}_{0,E}^{(j)}-\mathbf{B}^{(j)}_{0,E}\otimes\mathbf{B}^{(j)}_{0,E}+\left(\frac{1}{2}{\|\mathbf{B}^{(j)}_{0,E}\|}^{2}\right)\mathbf{I}\right)+\nabla p^{(j)}_{2,I}\right),\end{aligned}\right.$ with $\mathbf{q}^{0}_{inc}=(\rho{\bf u})^{0}_{inc}$ and $\nabla p_{0,E}^{(j)}=0$ for $j=1,\cdots,i-1$. Similarly, inserting expansions (4.7) into (3.8), in the order of $\mathcal{O}(1)$, we get (4.13) $\left\\{\begin{aligned} &\rho_{0,*}^{(i)}=\rho_{inc}^{0}-\Delta t\sum^{i-1}_{j=1}{a}_{ij}\nabla\cdot\mathbf{q}_{0,E}^{(j)},\\\ &\mathbf{q}^{(i)}_{0,*}=\mathbf{q}^{0}_{inc}-\Delta t\sum^{i-1}_{j=1}{a}_{ij}\left(\nabla\cdot\left(\rho_{0,E}^{(j)}\mathbf{u}_{0,E}^{(j)}\otimes\mathbf{u}_{0,E}^{(j)}-\mathbf{B}^{(j)}_{0,E}\otimes\mathbf{B}^{(j)}_{0,E}+\left(\frac{1}{2}{\|\mathbf{B}^{(j)}_{0,E}\|}^{2}\right)\mathbf{I}\right)+\nabla p^{(j)}_{2,I}\right),\end{aligned}\right.$ where from (3.17) and (4.8), it follows $\nabla\bar{p}^{(j)}_{E}=0$ for $j=1,\cdots,i-1$. Using (4.10), we get (4.14) $E_{0,*}^{(i)}=E_{inc}^{0}-\Delta t\sum^{i-1}_{j=1}{a}_{ij}\nabla\cdot\left(\bar{H}_{0}^{(j)}\mathbf{q}_{0,I}^{(j)}\right)=\frac{p_{*}}{\gamma-1}-\frac{\gamma p_{*}}{\gamma-1}\Delta t\sum^{i-1}_{j=1}{a}_{ij}\nabla\cdot{\bf u}_{0,I}^{(j)}.$ Therefore, from (4.8) and (4.14), we get $E_{0,*}^{(i)}=\frac{p_{*}}{\gamma-1}.$ Now from (4.14) and (3.9), it follows for the energy equation $E^{(i)}_{0,I}=E^{(i)}_{0,*}-\Delta ta_{ii}\nabla\cdot(\bar{H}^{(i)}_{0}\mathbf{q}^{(i)}_{0,I})=\frac{p_{*}}{\gamma-1}-\frac{\gamma p_{*}}{\gamma-1}\Delta t{a}_{ii}\nabla\cdot{\bf u}_{0,I}^{(i)}$ Considering the EOS $E_{I}^{(i)}=p_{I}^{(i)}/(\gamma-1)+\varepsilon^{2}({\|\mathbf{q}_{E}^{(i)}\|}^{2}/(2\rho_{E}^{(i)})+{\|\mathbf{B}_{E}^{(i)}\|}^{2}/2)$, to zeroth order in $\varepsilon$, we get $E_{0,I}^{(i)}=p_{0,I}^{(i)}/(\gamma-1)$, and we obtain for the pressure $p^{(i)}_{0,I}=p_{*}-\gamma p_{*}\Delta t{a}_{ii}\nabla\cdot{\bf u}_{0,I}^{(i)}.$ Integrating it over the spatial bounded domain $\Omega$, and assuming periodic or compact support boundary conditions, we first obtain $p^{(i)}_{0,I}=p_{*}$, and then we get $\nabla\cdot{\bf u}_{0,I}^{(i)}=0$ at the stage $i$. Finally, from (3.9), considering (4.13), we get for the density and momentum: (4.15) $\left\\{\begin{aligned} &\rho_{0,I}^{(i)}=\rho_{inc}^{0}-\Delta t\sum^{i}_{j=1}{a}_{ij}\nabla\cdot\mathbf{q}_{0,E}^{(j)},\\\ &\mathbf{q}^{(i)}_{0,I}=\mathbf{q}^{0}_{inc}-\Delta t\sum^{i}_{j=1}{a}_{ij}\left(\nabla\cdot\left(\rho_{0,E}^{(j)}\mathbf{u}_{0,E}^{(j)}\otimes\mathbf{u}_{0,E}^{(j)}-\mathbf{B}^{(j)}_{0,E}\otimes\mathbf{B}^{(j)}_{0,E}+\left(\frac{1}{2}{\|\mathbf{B}^{(j)}_{0,E}\|}^{2}\right)\mathbf{I}\right)+\nabla p^{(j)}_{2,I}\right).\end{aligned}\right.$ Then equations (LABEL:416), (4.12), and (4.15), with $\nabla\cdot{\bf B}_{E}^{(i)}=0$, $p_{0,E}^{(i)}=p_{0,I}^{(i)}=p_{*}$, and the divergence-free leading order velocity, i.e. $\nabla\cdot{\bf u}_{0,I}^{(i)}=0$, provide a multi-stage SI IMEX-RK discretization of the limiting system (LABEL:S2_E26) for the internal stage $i$. * • AA property for the updated numerical solution. Assuming that SI IMEX-RK scheme in Section 3.2 is SA, then the numerical solution coincides with the last internal stage $s$, and by setting $i=s$, we get (4.16) $p_{0}^{1}=p_{0,I}^{(s)}=p_{*},\quad\nabla\cdot{\bf u}_{0}^{1}=\nabla\cdot{\bf u}_{0,I}^{(s)}=0,$ and for the magnetic field (4.17) $\nabla\cdot{\bf B}_{E}^{1}=\nabla\cdot(\nabla\times{\bf A}_{E}^{(1)})=0,$ namely we have (4.5). Now if we denote ${\bf V}_{inc}({\bf x};t)=(\rho_{inc}({\bf x};t),\rho_{inc}({\bf x};t){\bf u}_{inc}({\bf x};t),p_{inc}({\bf x};t),{\bf A}_{inc}({\bf x};t),{\bf B}_{inc}({\bf x};t))^{T}$ as the exact solution of the incompressible MHD equations (LABEL:S2_E26), with initial data ${\bf V}_{inc}({\bf x};0)=(\rho^{0}({\bf x}),\rho^{0}({\bf x}){\bf u}^{0}({\bf x}),p^{0},{\bf A}^{0}({\bf x}),{\bf B}^{0}({\bf x}))^{T}$, from equations (LABEL:416), (4.12), (4.15), (4.16), and (4.17), one gets in the low sonic Mach limit where $\varepsilon=0$, a high order SI IMEX-RK scheme of order $r$ for the numerical solutions of (LABEL:S2_E26), that is, the scheme (3.7)-(3.10) is AA, and (4.6) holds. ∎ ## 5\. Numerical tests In this section, we perform some numerical experiments to validate the high order accuracy, AP, AA, divergence-free, and good performances of our proposed scheme for the MHD system with all sonic Mach numbers. The third order SI IMEX-RK with an SA property from [8, 30] is used for time discretizations. The fifth order finite difference WENO reconstruction[8, 30, 53, 31] for the first order spatial derivatives, and a fourth order compact central difference discretization [8, 10] for the second order derivatives will be applied for spatial discretizations. Our scheme is referred as the “IMEX” scheme. For some of the following examples, we will compare our results to reference solutions, which are produced by a fifth-order finite difference WENO scheme with a third-order strong stability preserving RK method. It is referred as an “ERK” scheme if without a CT method [32], while as an “ERKC” scheme if with a CT method [18]. To avoid excessive numerical dissipations for low sonic Mach numbers, for the explicit schemes “ERK” and “ERKC”, here we also take a Lax- Friedrichs flux the same as in (3.11) and (3.12), and numerically we find that it leads to much better results as we can see in the following. Unless otherwise specified, the ${\rm CFL}$ number is taken as $0.25$, and the gas constant is $\gamma=5/3$. For 1D problems, the time step is $\Delta t=\frac{{\rm CFL}\Delta x}{\mathop{\max}\limits_{0\leq k\leq N}\left(|u_{k}|+\hat{c}_{f}^{(k)}\right)},$ with $N+1$ computational grid points, and the definition of $\hat{c}_{f}^{(k)}$ has been given in (3.12). For 2D problems, the time step is $\Delta t={\rm CFL}/\left(\frac{\mathop{\max}\limits_{0\leq k\leq N_{x},0\leq l\leq N_{y}}\left(|u_{k,l}|+\hat{c}_{f,x}^{(k,l)}\right)}{\Delta x}+\frac{\mathop{\max}\limits_{0\leq k\leq N_{x},0\leq l\leq N_{y}}\left(|v_{k,l}|+\hat{c}_{f,y}^{(k,l)}\right)}{\Delta y}\right),$ with $(N_{x}+1)\times(N_{y}+1)$ computational grid points. Here, the fast speeds in the $x$-direciton ($\hat{c}_{f,x}^{(k,l)}$) and in the $y$-direciton ($\hat{c}_{f,y}^{(k,l)}$) are defined as $\left\\{\begin{aligned} \hat{c}_{f,x}^{(k,l)}={\left\\{\frac{1}{2}\left[\left(\hat{a}^{(k,l)}\right)^{2}+\frac{{\|{\bf B}_{k,l}\|}^{2}}{\rho_{k,l}}\pm\sqrt{{\left(\left(\hat{a}^{({k,l})}\right)^{2}+\frac{{\|{\bf B}_{k,l}\|}^{2}}{\rho_{k,l}}\right)}^{2}-\left(2{\hat{a}^{({k,l})}}\sqrt{\frac{B_{x;{k,l}}^{2}}{\rho_{k,l}}}\right)^{2}}\right]\right\\}}^{\frac{1}{2}},\\\ \hat{c}_{f,y}^{(k,l)}={\left\\{\frac{1}{2}\left[\left(\hat{a}^{(k,l)}\right)^{2}+\frac{{\|{\bf B}_{k,l}\|}^{2}}{\rho_{k,l}}\pm\sqrt{{\left(\left(\hat{a}^{({k,l})}\right)^{2}+\frac{{\|{\bf B}_{k,l}\|}^{2}}{\rho_{k,l}}\right)}^{2}-\left(2{\hat{a}^{({k,l})}}\sqrt{\frac{B_{y;{k,l}}^{2}}{\rho_{k,l}}}\right)^{2}}\right]\right\\}}^{\frac{1}{2}},\end{aligned}\right.$ where $\hat{a}^{({k,l})}=\min\left(\frac{1}{\varepsilon},1\right)\sqrt{\frac{\gamma p_{k,l}}{\rho_{k,l}}}.$ ### 5.1. One dimensional case Since the divergence-free condition is satisfied automatically in 1D, therefore, the CT method will not need to be applied in the following 1D numerical simulations. ###### Example 5.1. (Accuracy test) To assess the convergence of our proposed scheme, we simulate the nonlinear circularly polarized Alfvén wave problem in 1D. The smooth initial conditions are given as [16]: (5.1) $\displaystyle(\rho,u,v,w,B_{x},B_{y},B_{z},p)(0,x)$ $\displaystyle\quad=(1,0,0.1\sin(2\pi x),0.1\cos(2\pi x),1,0.1\sin(2\pi x),0.1\cos(2\pi x),0.1)\quad x\in[0,1],$ with a periodic boundary condition. The exact solution to (5.1) propagates with a unit Alfvén speed (that is $U(t,x)=U(0,x+t)$.). We set the time step to be $\Delta t=\frac{{\rm CFL}\Delta x^{\frac{5}{3}}}{\mathop{\max}\limits_{0\leq k\leq N}\left(|u_{k}|+\hat{c}_{f}^{(k)}\right)}.$ The sonic Mach number is $\varepsilon=1$ and we compute up to a final time $T=1$. The numerical $L_{1}$, $L_{2}$, and $L_{\infty}$ errors and orders are shown in Table 5.1. We can observe that a fifth-order spatial accuracy is obtained. Table 5.1. Example 5.1. The $L_{1}$, $L_{2}$, and $L_{\infty}$ errors and orders for $\rho v$ with $\varepsilon=1$ at $T=1$. $N$ | 10 | 20 | 40 | 80 | 160 ---|---|---|---|---|--- $L_{1}$ error | 6.24E-04 | 2.03E-05 | 6.41E-07 | 2.01E-08 | 6.27E-10 order | – | 4.94 | 4.99 | 5.00 | 5.00 $L_{2}$ error | 6.83E-04 | 2.25E-05 | 7.11E-07 | 2.23E-08 | 6.97E-10 order | – | 4.93 | 4.98 | 5.00 | 5.00 $L_{\infty}$ error | 9.65E-04 | 3.15E-05 | 1.00E-06 | 3.15E-08 | 9.85E-10 order | – | 4.94 | 4.97 | 4.99 | 5.00 ###### Example 5.2. (Accuracy test for a range of $\varepsilon$) We take a well-prepared initial data which is $\varepsilon-$dependent: $\left\\{\begin{aligned} \rho(x,0)&=1+\varepsilon^{2}\sin^{2}(2\pi x),\quad&p(x,0)&=(1+\varepsilon^{2}\sin^{2}(2\pi x))^{\gamma},\\\ u(x,0)&=\varepsilon^{2}\sin(2\pi x),\quad&B_{x}(x,0)&=0.5,\\\ v(x,0)&=\sin(2\pi x)+\varepsilon^{2}\cos(2\pi x),\quad&B_{y}(x,0)&=(1+\varepsilon^{2})\sin(2\pi x),\\\ w(x,0)&=0,\quad&B_{z}(x,0)&=(1+\varepsilon^{2})\cos(2\pi x),\\\ \end{aligned}\right.$ on the domain $\Omega=[0,1]$ with $\gamma=1.4$ and periodic boundary conditions. Reference solutions are computed with $N=320$. Four different sonic Mach numbers $\varepsilon=1,10^{-2},10^{-6}$, and $0$ are taken, with a final time $T=0.05$. Numerical errors and orders are shown in Table 5.2. For the intermediate value of $\varepsilon=10^{-2}$, order reduction is observed, which is a typical behavior of high-order IMEX schemes for these multi-scale problems [8]. Around 5th-order accuracy for $\varepsilon=1,10^{-6}$, and $0$ is observed. Table 5.2. Example 5.2. The $L_{1}$ errors and orders for $\rho v$ with $\varepsilon=1,10^{-2},10^{-6}$, and $0$. $N$ $\varepsilon$ | $\varepsilon=1$ | $\varepsilon=10^{-2}$ | $\varepsilon=10^{-6}$ | $\varepsilon=0$ ---|---|---|---|--- $L_{1}$ error | order | $L_{1}$ error | order | $L_{1}$ error | order | $L_{1}$ error | order 10 | 3.08E-02 | – | 4.93E-04 | – | 4.82E-04 | – | 4.82E-04 | – 20 | 6.47E-03 | 2.25 | 7.18E-05 | 2.78 | 1.61E-05 | 4.90 | 1.61E-05 | 4.90 40 | 8.32E-04 | 2.96 | 3.33E-05 | 1.11 | 5.22E-07 | 4.95 | 5.22E-07 | 4.95 80 | 3.74E-05 | 4.48 | 1.01E-05 | 1.72 | 1.76E-08 | 4.89 | 1.76E-08 | 4.89 160 | 1.22E-06 | 4.94 | 3.34E-06 | 1.60 | 6.65E-10 | 4.72 | 6.65E-10 | 4.72 ###### Example 5.3. (MHD shock tube) In this example, we consider a 1D MHD shock tube in the compressible regime where the sonic Mach number is of $\mathcal{O}(1)$. The initial conditions are as follows $\displaystyle(\rho,u,v,w,B_{x},B_{y},B_{z},p)(0,x)$ $\displaystyle\quad=\left\\{\begin{aligned} &(1,0,0,0,0.75,1,0,1)\quad&x&\in[0,0.5),\\\ &(0.125,0,0,0,0.75,-1,0,0.1)\quad&x&\in[0.5,1],\end{aligned}\right.$ with a final time $T=0.1$. The specific heat ratio is set to be $\gamma=2$ with $\varepsilon=1$ on the domain $\Omega=[0,1]$. Reflective boundary conditions are considered, and we take $N=200$. The results are shown in Fig. 5.1, and the solutions obtained by the ERK scheme with $N=2400$ are displayed as references. From these results, it can be seen that our IMEX scheme can capture the MHD waves very well. Figure 5.1. MHD shock tube solutions for Example 5.3. Left: density $\rho$; Right: magnetic field $B_{y}$. Every two points are plotted for the IMEX scheme. ### 5.2. Two dimensional case A CT method described in [18] will be used to maintain the divergence-free condition. In 2D, ${\bf B}=\nabla\times{\bf A}$ can be written as $B_{x}=\partial A_{z}/\partial y$ and $B_{y}=-\partial A_{z}/\partial x$. Therefore, we only need to update the third component $A_{z}$ of ${\bf A}$, and ignore $A_{x}$ and $A_{y}$. All solutions of the following tests maintain $\nabla\cdot{\bf B}=0$ up to machine round off errors[16]. ###### Example 5.4. (Accuracy test) The 2D version of the smooth Alfvén wave problem is obtained by rotating the direction of propagation by an angle of $\theta$, so that the wave now propagates along the direction ${\bf n}=(-\cos{\theta},-\sin{\theta})$ on the domain $\Omega=[0,1/\cos{\theta}]\times[0,1/\sin{\theta}]$ [16]. In this numerical test, periodic boundary conditions are applied on all both directions with $\varepsilon=1,\theta=\pi/4$, and $T=1$. Similarly, if we take $\Delta t={\rm CFL}/\left(\frac{\mathop{\max}\limits_{0\leq k\leq N_{x},0\leq l\leq N_{y}}\left(|u_{k,l}|+\hat{c}_{f,x}^{(k,l)}\right)}{\Delta x^{\frac{5}{3}}}+\frac{\mathop{\max}\limits_{0\leq k\leq N_{x},0\leq l\leq N_{y}}\left(|v_{k,l}|+\hat{c}_{f,y}^{(k,l)}\right)}{\Delta y^{\frac{5}{3}}}\right),$ from Table 5.3 we can find around 5th-order accuracy for the momentum $\rho u$, and 4th-order accuracy for the magnetic field component $B_{x}$ where a fourth-order central difference discretization is used for ${\bf B}=\nabla\times{\bf A}$. Table 5.3. Example 5.4. The $L_{1}$, $L_{2}$, and $L_{\infty}$ errors and orders for $\rho u$ and $B_{x}$ with $\varepsilon=1$ and $T=1$. | $N_{x}\times N_{y}$ | $8\times 8$ | $16\times 16$ | $32\times 32$ | $64\times 64$ | $128\times 128$ ---|---|---|---|---|---|--- $\rho u$ | $L_{1}$ error | 1.54E-03 | 5.73E-05 | 1.86E-06 | 5.88E-08 | 1.84E-09 order | – | 4.75 | 4.94 | 4.99 | 5.00 $L_{2}$ error | 1.80E-03 | 6.47E-05 | 2.08E-06 | 6.55E-08 | 2.07E-09 order | – | 4.80 | 4.96 | 4.99 | 4.99 $L_{\infty}$ error | 2.55E-03 | 9.26E-05 | 2.98E-06 | 9.44E-08 | 3.00E-09 order | – | 4.79 | 4.96 | 4.98 | 4.98 $B_{x}$ | $L_{1}$ error | 9.29E-04 | 6.36E-05 | 3.17E-06 | 1.69E-07 | 9.64E-09 order | – | 3.87 | 4.33 | 4.23 | 4.13 $L_{2}$ error | 1.08E-03 | 7.02E-05 | 3.51E-06 | 1.88E-07 | 1.07E-08 order | – | 3.94 | 4.32 | 4.23 | 4.13 $L_{\infty}$ error | 1.29E-03 | 9.69E-05 | 5.06E-06 | 2.70E-07 | 1.52E-08 order | – | 3.73 | 4.26 | 4.23 | 4.15 ###### Example 5.5. (Accuracy test for a range of $\varepsilon$) In this 2D case, we set a well- prepared initial condition as follows $\left\\{\begin{aligned} \rho(x,y,0)&=1+\varepsilon^{2}\sin^{2}(2\pi(x+y)),\quad&p(x,y,0)&=(1+\varepsilon^{2}\sin^{2}(2\pi(x+y)))^{\gamma},\\\ u(x,y,0)&=\sin(2\pi(x-y))+\varepsilon^{2}\sin(2\pi(x+y)),\quad&B_{x}(x,y,0)&=-\frac{1}{\sqrt{2}}\sin(2\pi(x+y)),\\\ v(x,y,0)&=\sin(2\pi(x-y))+\varepsilon^{2}\cos(2\pi(x+y)),\quad&B_{y}(x,y,0)&=\frac{1}{\sqrt{2}}\sin(2\pi(x+y)),\\\ w(x,y,0)&=0,\quad&B_{z}(x,y,0)&=\cos(2\pi(x+y)),\\\ \end{aligned}\right.$ with the initial magnetic potential $A_{z}(x,y,0)=\cos(2\pi(x+y))/(2\sqrt{2}\pi)$. As in Example 5.2, four different sonic Mach numbers are considered with periodic boundary conditions on the domain $\Omega=[0,1]\times[0,1]$. We compute the solution up to a final time $T=0.01$ on mesh grid points of $N^{2}$. The $L_{1}$ errors and orders of the accuracy are shown in Table 5.4. From this Table, we can see high-order accuracy can be obtained for $\varepsilon=1,10^{-6}$, and $0$. However, under the current mesh sizes, order degeneracy can also be found for $\varepsilon=10^{-2}$, which is similar to the 1D case. Table 5.4. Example 5.5. The $L_{1}$ errors and orders for $\rho u$ with $\varepsilon=1,10^{-2},10^{-6}$, and $0$. $N$ $\varepsilon$ | $\varepsilon=1$ | $\varepsilon=10^{-2}$ | $\varepsilon=10^{-6}$ | $\varepsilon=0$ ---|---|---|---|--- $L_{1}$ error | order | $L_{1}$ error | order | $L_{1}$ error | order | $L_{1}$ error | order 8 | 8.98E-02 | – | 5.18E-03 | – | 7.65E-03 | – | 7.65E-03 | – 16 | 1.39E-02 | 2.69 | 2.09E-03 | 1.31 | 8.74E-04 | 3.13 | 8.74E-04 | 3.13 32 | 6.76E-04 | 4.36 | 2.09E-03 | – | 1.68E-05 | 5.70 | 1.68E-05 | 5.70 64 | 2.28E-05 | 4.89 | 2.41E-03 | – | 7.42E-07 | 4.50 | 7.43E-07 | 4.50 128 | 7.15E-07 | 5.00 | 6.41E-04 | 1.91 | 1.42E-08 | 5.70 | 1.40E-08 | 5.73 ###### Example 5.6. (Orszag-Tang vortex) Next we consider the Orszag–Tang vortex problem, which is widely considered as a standard test for MHD [18, 16, 56, 3]. The problem has smooth initial conditions $\displaystyle(\rho,u,v,w,B_{x},B_{y},B_{z},p)(x,y,0)$ $\displaystyle=$ $\displaystyle(\gamma^{2},-\sin{(y)},\sin{(x)},0,-\sin{(y)},\sin{(2x)},0,\gamma),$ with the initial magnetic potential: $A_{z}=0.5\cos{(2x)}+\cos{(y)}$ on the domain $\Omega=[0,2\pi]\times[0,2\pi]$. Periodic boundary conditions are imposed on all boundaries. As time evolves, the solution forms several shock waves and a vortex structure in the middle of the computational domain. We set $\varepsilon=1$, $N_{x}=N_{y}=192$ and present the density at $T=0.5,T=2,T=3$, and $T=4$ in Fig. 5.2. A slice of the pressure at $y=0.625\pi$ and $T=3$ is shown on the right panel of Fig. 5.3. We find our IMEX scheme can successfully capture the shocks. We do not observe significant oscillations in any of the conserved quantities, and our results are in good agreement with those given in [18, 16]. Figure 5.2. Density for Orszag-Tang problem of Example 5.6 on the mesh grid $192^{2}$. Figure 5.3. Example 5.6. Thermal Pressure for Orszag-Tang problem on the mesh grid $192^{2}$. Every two points are plotted for the IMEX scheme for the cutting plot. ###### Example 5.7. (MHD blast wave) MHD blast wave has been commonly used to test numerical methods for the MHD system [15, 28]. Here, we set periodic boundary conditions on the domain $\Omega=[-0.5,0.5]\times[-0.5,0.5]$. The initial conditions are $\displaystyle(\rho,u,v,w,B_{x},B_{y},B_{z},p)(x,y,0)$ $\displaystyle\quad=\left\\{\begin{aligned} &(1,0,0,0,10\sin{\theta},10\cos{\theta},0,100)\quad&r&\leq 0.125,\\\ &(1,0,0,0,10\sin{\theta},10\cos{\theta},0,10)\quad&r&>0.125,\end{aligned}\right.$ with $\theta=\pi/4$ and $r=\sqrt{x^{2}+y^{2}}$. The initial magnetic potential is given by $A_{z}(x,y,0)=5\sqrt{2}(-x+y)$ with a first order extrapolation if outside the boundary. In Fig. 5.4, we show the density and magnetic pressure obtained by the IMEX and ERKC schemes on a $200\times 200$ mesh with $\varepsilon=0.9$ at $T=0.02$. The slices of the density and the magnetic pressure along $y=-x+0.5$ are also shown in Fig. 5.5. We see that the solution of the IMEX scheme is close to the reference ERKC solutions and no obvious oscillations are observed. In our test, the maximum error of discrete divergence of the magnetic field is $2.87\times 10^{-11}$ for the IMEX scheme and $2.77\times 10^{-11}$ for the ERKC scheme, both are close to machine precision. Figure 5.4. Example 5.7. MHD blast wave problem on the mesh grid $200^{2}$. Top: IMEX; Bottom: ERKC. 35 equally spaced contours are used. Figure 5.5. Example 5.7 MHD blast wave problem. 1D cutting plots along $y=-x+0.5$. Every two points are plotted for the IMEX scheme for the cutting plots. ###### Example 5.8. (Field loop advection) Now, we show the 2D field loop advection test [45, 49] to assess the capability of capturing a tangential discontinuity for a multidimensional flow. The computational domain is $\Omega=[-1,1]\times[-0.5,0.5]$ with periodic boundary conditions. The domain is divided with $256\times 128$ uniform cells. The initial conditions are set as $\displaystyle(\rho,u,v,w,B_{x},B_{y},B_{z},p)(r,0)$ $\displaystyle\quad=(1,v_{0}\cos{(\theta_{0})},v_{0}\sin{(\theta_{0})},0,\partial A_{z}/\partial y,-\partial A_{z}/\partial x,0,1)$ with $v_{0}=1,\cos{(\theta_{0})}=-2/\sqrt{5},\sin{(\theta_{0})}=1/\sqrt{5},r=\sqrt{x^{2}+y^{2}}$, and $\displaystyle A_{z}(r)=\left\\{\begin{aligned} &10^{-3}(0.3-r)\quad&r\leq 0.3,&\\\ &0\quad&otherwise.&\end{aligned}\right.$ In this test, a loop of the magnetic field is advected. In Fig. 5.6, we show both the deviation of the density $\rho$ away from 1 and the magnetic pressure $\|{\bf B}\|^{2}/2$ for the case of $\varepsilon=0.1$ at $T=\sqrt{5}$. For the magnetic pressure, both schemes preserve the initial symmetric loop structure very well, and the IMEX scheme has almost the same results as the ERKC scheme. The one-dimensional cutting plots for the magnetic pressure along $x=0.2$ show that both schemes slightly distort the profile at $y=0.15$. For the deviation of the density, we can find the IMEX scheme is better than the ERKC scheme for preserving a symmetric structure, as compared with the results of the ERKC scheme on a uniform mesh of $384\times 192$ cells. For this problem in the intermediate regime with $\varepsilon=0.1$, we also compare the CPU cost of two schemes in Table 5.5. We can observe that the ERKC scheme costs less when $\varepsilon$ is of $\mathcal{O}(1)$. However, the IMEX scheme is much more efficient in the intermediate and low sonic Mach regimes. The computational cost for the ERKC scheme increases very rapidly, as the sonic Mach number $\varepsilon$ becomes smaller. This example demonstrates the main advantage of an AP scheme for such multiscale problems. Figure 5.6. Example 5.8. Field loop advection on the mesh grid $256\times 128$. Top left: magnetic pressure (IMEX); Middle left: magnetic pressure (ERKC); Bottom left: 1D cutting plot for the magnetic pressure; Top right: deviation of density (IMEX); Middle right: deviation of density (ERKC); Bottom right: 1D cutting plot for the deviation of density. Table 5.5. Example 5.8. The CPU cost (seconds) for the IMEX and ERKC schemes on the mesh grid $256\times 128$ at $T=\sqrt{5}$. $\varepsilon$ | 0.5 | 0.1 | 0.05 ---|---|---|--- IMEX | 7634.880 | 5687.467 | 8015.959 ERKC | 4736.466 | 19719.964 | 38666.192 ###### Example 5.9. (Magnetized Kelvin–Helmholtz instability) Finally, we consider a simulation of the magnetized Kelvin–Helmholtz instability problem[8, 42]. The initial conditions are $\displaystyle(\rho,u,v,w,B_{x},B_{y},B_{z},p)(x,y,0)$ $\displaystyle\quad=(\gamma,1-2\eta(x),0.1\sin{(2\pi x)},0,0.1,0,0,1),$ with a magnetic potential $A_{z}(x,y,0)=0.1y$ on the domain $\Omega=[0,2]\times[-0.5,0.5]$, and $\displaystyle\eta(x)=\left\\{\begin{aligned} &\frac{1}{2}\left(1+\sin\left(16\pi\left(y+\frac{1}{4}\right)\right)\right)\quad&y\in[-\frac{9}{32},-\frac{7}{32}],&\\\ &1\quad&y\in[-\frac{7}{32},\enspace\;\frac{7}{32}],&\\\ &\frac{1}{2}\left(1-\sin\left(16\pi\left(y-\frac{1}{4}\right)\right)\right)\quad&y\in[\enspace\;\frac{7}{32},\enspace\;\frac{9}{32}],&\\\ &0\quad&otherwise.&\end{aligned}\right.$ Periodic boundary conditions are used for the conserved quantities, while a first-order extrapolation is utilized when the magnetic potential is outside of the computational domain. We run the solution up to $T=0.8$ with $\gamma=1.4$ and $\varepsilon=10^{-6}$ on a mesh grid $N_{x}\times N_{y}=256\times 128$. We define the local sonic Mach number as $M_{a}=\sqrt{u^{2}+v^{2}}/\sqrt{\gamma p/\rho}$, and the ratio of the local sonic Mach number $M_{ratio}$ to the maximum value of $M_{a}$ as $M_{ratio}=M_{a}/\max(M_{a})$, where the maximum is taken over all computational grid points. The vorticity $\omega=v_{x}-u_{y}$, where $v_{x}$ and $u_{y}$ are discretized by the 4th order central difference, and $M_{ratio}$ are shown in Fig. 5.7. We observe that they are comparable to the results in [42, 8], namely, our IMEX scheme can capture the incompressible MHD system in the low sonic Mach number regime very well, which verifies the AP property of our scheme. Figure 5.7. Magnetized Kelvin–Helmholtz instability problem for Example 5.9 on the mesh grid $256\times 128$. Left: Vorticity $\omega=v_{x}-u_{y}$; Right: The ratio of the local sonic Mach number $M_{ratio}$. ## 6\. Conclusion In this paper, a high-order SI AP scheme with a discrete divergence-free property for the MHD equations has been developed. High-order accuracy in time has been obtained by SI temporal integrator based on an IMEX-RK framework. High-order accuracy in space has been achieved by finite difference WENO reconstructions. We have formally proved that the scheme is AP. We have also proved the AA property in the stiff limit as the sonic Mach number $\varepsilon\rightarrow 0$, if the SI IMEX-RK scheme is SA. Numerical experiments in 1D and 2D have demonstrated the divergence-free for all ranges of sonic Mach numbers, AP and AA properties in the low sonic Mach limit. Compared to the explicit schemes as reference solutions, such as the ERK [32] and ERKC [18] schemes in 1D and 2D respectively, our IMEX scheme has almost the same performances as the ERK or ERKC scheme, but is much more efficient in the intermediate and low sonic Mach regimes. ## References * [1] D. S. Balsara. Second-order-accurate schemes for magnetohydrodynamics with divergence-free reconstruction. The Astrophysical Journal Supplement Series, 151(1):149, 2004. * [2] D. S. Balsara and M. Dumbser. Divergence-free MHD on unstructured meshes using high order finite volume schemes based on multidimensional Riemann solvers. Journal of Computational Physics, 299:687–715, 2015. * [3] D. S. Balsara, M. Dumbser, and R. Abgrall. Multidimensional HLLC Riemann solver for unstructured meshes–with application to Euler and MHD flows. Journal of Computational Physics, 261:172–208, 2014. * [4] D. S. Balsara and D. S. Spicer. A staggered mesh algorithm using high order Godunov fluxes to ensure solenoidal magnetic fields in magnetohydrodynamic simulations. Journal of Computational Physics, 149(2):270–292, 1999. * [5] S. Boscarino. Error analysis of IMEX Runge–Kutta methods derived from differential-algebraic systems. SIAM Journal on Numerical Analysis, 45(4):1600–1621, 2007. * [6] S. Boscarino, F. Filbet, and G. Russo. High order semi-implicit schemes for time dependent partial differential equations. Journal of Scientific Computing, 68(3):975–1001, 2016. * [7] S. Boscarino, L. Pareschi, and G. Russo. Implicit-explicit Runge–Kutta schemes for hyperbolic systems and kinetic equations in the diffusion limit. SIAM Journal on Scientific Computing, 35(1):A22–A51, 2013. * [8] S. Boscarino, J. Qiu, G. Russo, and T. Xiong. High Order Semi-implicit WENO Schemes for All-Mach Full Euler System of Gas Dynamics. SIAM Journal on Scientific Computing, 44(2):B368–B394, 2022. * [9] S. Boscarino, J.-M. Qiu, and G. Russo. Implicit-explicit integral deferred correction methods for stiff problems. SIAM Journal on Scientific Computing, 40(2):A787–A816, 2018. * [10] S. Boscarino, J.-M. Qiu, G. Russo, and T. Xiong. A high order semi-implicit IMEX WENO scheme for the all-Mach isentropic Euler system. Journal of Computational Physics, 392:594–618, 2019. * [11] W. Boscheri and L. Pareschi. High order pressure-based semi-implicit IMEX schemes for the 3D Navier-Stokes equations at all Mach numbers. Journal of Computational Physics, 434:110206, 2021. * [12] M. Brio and C. C. Wu. An upwind differencing scheme for the equations of ideal magnetohydrodynamics. Journal of Computational Physics, 75(2):400–422, 1988. * [13] W. Cao, F. Zeng, Z. Zhang, and G. E. Karniadakis. Implicit-explicit difference schemes for nonlinear fractional differential equations with nonsmooth solutions. SIAM Journal on Scientific Computing, 38(5):A3070–A3093, 2016. * [14] P. Cargo and G. Gallice. Roe matrices for ideal MHD and systematic construction of Roe matrices for systems of conservation laws. Journal of Computational Physics, 136(2):446–466, 1997. * [15] A. J. Christlieb, X. Feng, Y. Jiang, and Q. Tang. A high-order finite difference WENO scheme for ideal magnetohydrodynamics on curvilinear meshes. SIAM Journal on Scientific Computing, 40(4):A2631–A2666, 2018. * [16] A. J. Christlieb, X. Feng, D. C. Seal, and Q. Tang. A high-order positivity-preserving single-stage single-step method for the ideal magnetohydrodynamic equations. Journal of Computational Physics, 316:218–242, 2016. * [17] A. J. Christlieb, Y. Liu, Q. Tang, and Z. Xu. Positivity-preserving finite difference weighted ENO schemes with constrained transport for ideal magnetohydrodynamic equations. SIAM Journal on Scientific Computing, 37(4):A1825–A1845, 2015. * [18] A. J. Christlieb, J. A. Rossmanith, and Q. Tang. Finite difference weighted essentially non-oscillatory schemes with constrained transport for ideal magnetohydrodynamics. Journal of Computational Physics, 268:302–325, 2014. * [19] W. Cui, Y. Ou, and D. Ren. Incompressible limit of full compressible magnetohydrodynamic equations with well-prepared data in 3-D bounded domains. Journal of Mathematical Analysis and Applications, 427(1):263–288, 2015. * [20] W. Dai and P. R. Woodward. A simple finite difference scheme for multidimensional magnetohydrodynamical equations. Journal of Computational Physics, 142(2):331–369, 1998. * [21] W. Dai and P. R. Woodward. On the divergence-free condition and conservation laws in numerical simulations for supersonic magnetohydrodynamical flows. The Astrophysical Journal, 494(1):317, 1998. * [22] S. Dellacherie. Analysis of Godunov type schemes applied to the compressible Euler system at low Mach number. Journal of Computational Physics, 229(4):978–1016, 2010. * [23] G. Dimarco, R. Loubère, and M.-H. Vignal. Study of a new asymptotic preserving scheme for the Euler system in the low Mach number limit. SIAM journal on Scientific Computing, 39(5):A2099–A2128, 2017. * [24] G. Dimarco and L. Pareschi. High order asymptotic-preserving schemes for the Boltzmann equation. Comptes Rendus Mathematique, 350(9-10):481–486, 2012. * [25] M. Dumbser, D. S. Balsara, M. Tavelli, and F. Fambri. A divergence-free semi-implicit finite volume scheme for ideal, viscous, and resistive magnetohydrodynamics. International Journal for Numerical Methods in Fluids, 89(1-2):16–42, 2019. * [26] C. R. Evans and J. F. Hawley. Simulation of magnetohydrodynamic flows: A constrained transport method. The Astrophysical Journal, 332:659–677, 1988. * [27] F. Filbet and S. Jin. A class of asymptotic-preserving schemes for kinetic equations and related problems with stiff sources. Journal of Computational Physics, 229(20):7625–7648, 2010. * [28] J. Han and H. Tang. An adaptive moving mesh method for two-dimensional ideal magnetohydrodynamics. Journal of Computational Physics, 220(2):791–812, 2007. * [29] C. Helzel, J. A. Rossmanith, and B. Taetz. An unstaggered constrained transport method for the 3D ideal magnetohydrodynamic equations. Journal of Computational Physics, 230(10):3803–3829, 2011. * [30] G. Huang, Y. Xing, and T. Xiong. High order well-balanced asymptotic preserving finite difference WENO schemes for the shallow water equations in all Froude numbers. Journal of Computational Physics, 463:111255, 2022. * [31] G.-S. Jiang and D. Peng. Weighted ENO schemes for Hamilton–Jacobi equations. SIAM Journal on Scientific Computing, 21(6):2126–2143, 2000. * [32] G.-S. Jiang and C.-C. Wu. A high-order WENO finite difference scheme for the equations of ideal magnetohydrodynamics. Journal of Computational Physics, 150(2):561–594, 1999. * [33] S. Jiang, Q. Ju, and F. Li. Incompressible limit of the compressible magnetohydrodynamic equations with periodic boundary conditions. Communications in Mathematical Physics, 297(2):371–400, 2010. * [34] S. Jiang, Q. Ju, and F. Li. Low Mach number limit for the multi-dimensional full magnetohydrodynamic equations. Nonlinearity, 25(5):1351, 2012. * [35] S. Jin. Efficient asymptotic-preserving (AP) schemes for some multiscale kinetic equations. SIAM Journal on Scientific Computing, 21:441–454, 1999. * [36] S. Jin. Asymptotic preserving (AP) schemes for multiscale kinetic and hyperbolic equations: a review. Lecture notes for summer school on methods and models of kinetic theory (M$\&$MKT), Porto Ercole (Grosseto, Italy), pages 177–216, 2010. * [37] S. Jin. Asymptotic-preserving schemes for multiscale physical problems. Acta Numerica, 31:415–489, 2022. * [38] S. Jin, M. Tang, and X. Zhang. A spatial-temporal asymptotic preserving scheme for radiation magnetohydrodynamics in the equilibrium and non-equilibrium diffusion limit. Journal of Computational Physics, 452:110895, 2022. * [39] S. Kadioglu, D. Knoll, M. Sussman, and R. Martineau. A second order JFNK-based IMEX method for single and multi-phase flows. In Computational Fluid Dynamics 2010, pages 549–554. Springer, 2011\. * [40] A. Kanevsky, M. H. Carpenter, D. Gottlieb, and J. S. Hesthaven. Application of implicit–explicit high order Runge–Kutta methods to discontinuous-Galerkin schemes. Journal of Computational Physics, 225(2):1753–1781, 2007. * [41] R. Klein. Semi-implicit extension of a Godunov-type scheme based on low Mach number asymptotics I: One-dimensional flow. Journal of Computational Physics, 121(2):213–237, 1995. * [42] G. Leidi, C. Birke, R. Andrassy, J. Higl, P. Edelmann, G. Wiest, C. Klingenberg, and F. Röpke. A finite-volume scheme for modelling compressible MHD flows at low Mach numbers in stellar interiors. https://arxiv.org/abs/2210.01641, 2022. * [43] F. Li and C.-W. Shu. Locally divergence-free discontinuous Galerkin methods for MHD equations. Journal of Scientific Computing, 22(1):413–442, 2005. * [44] F. Li and L. Xu. Arbitrary order exactly divergence-free central discontinuous Galerkin methods for ideal MHD equations. Journal of Computational Physics, 231(6):2655–2675, 2012. * [45] T. Mamashita, K. Kitamura, and T. Minoshima. SLAU2-HLLD numerical flux with wiggle-sensor for stable low mach Magnetohydrodynamics simulations. Computers $\&$ Fluids, 231:105165, 2021. * [46] W. H. Matthaeus and M. R. Brown. Nearly incompressible magnetohydrodynamics at low Mach number. The Physics of Fluids, 31(12):3634–3644, 1988. * [47] F. Meng, J. Banks, W. Henshaw, and D. Schwendeman. Fourth-order accurate fractional-step IMEX schemes for the incompressible Navier–Stokes equations on moving overlapping grids. Computer Methods in Applied Mechanics and Engineering, 366:113040, 2020. * [48] T. Minoshima and T. Miyoshi. A low-dissipation HLLD approximate Riemann solver for a very wide range of Mach numbers. Journal of Computational Physics, 446:110639, 2021. * [49] T. Minoshima, T. Miyoshi, and Y. Matsumoto. A high-order weighted finite difference scheme with a multistate approximate Riemann solver for divergence-free magnetohydrodynamic simulations. The Astrophysical Journal Supplement Series, 242(2):14, 2019. * [50] O. O’Reilly, E. M. Dunham, and J. Nordström. Simulation of wave propagation along fluid-filled cracks using high-order summation-by-parts operators and implicit-explicit time stepping. SIAM Journal on Scientific Computing, 39(4):B675–B702, 2017. * [51] K. G. Powell. An approximate Riemann solver for magnetohydrodynamics. In Upwind and High-Resolution Schemes, pages 570–583. Springer, 1997. * [52] K. G. Powell, P. L. Roe, T. J. Linde, T. I. Gombosi, and D. L. De Zeeuw. A solution-adaptive upwind scheme for ideal magnetohydrodynamics. Journal of Computational Physics, 154(2):284–309, 1999. * [53] C.-W. Shu. Essentially non-oscillatory and weighted essentially non-oscillatory schemes for hyperbolic conservation laws. Advanced Numerical Approximation of Nonlinear Hyperbolic Equations, pages 325–432, 1998. * [54] A. Susanto, L. Ivan, H. De Sterck, and C. P. Groth. High-order central ENO finite-volume scheme for ideal MHD. Journal of Computational Physics, 250:141–164, 2013. * [55] K. Takahashi and S. Yamada. Regular and non-regular solutions of the Riemann problem in ideal magnetohydrodynamics. Journal of Plasma Physics, 79(3):335–356, 2013. * [56] H. Tang and K. Xu. A high-order gas-kinetic method for multidimensional ideal magnetohydrodynamics. Journal of Computational Physics, 165(1):69–88, 2000. * [57] M. Tavelli and M. Dumbser. A pressure-based semi-implicit space–time discontinuous Galerkin method on staggered unstructured meshes for the solution of the compressible Navier–Stokes equations at all Mach numbers. Journal of Computational Physics, 341:341–376, 2017. * [58] K. Wu. Positivity-preserving analysis of numerical schemes for ideal magnetohydrodynamics. SIAM Journal on Numerical Analysis, 56(4):2124–2147, 2018. * [59] K. Wu, H. Jiang, and C.-W. Shu. Provably positive central DG schemes via geometric quasilinearization for ideal MHD equations. SIAM Journal on Numerical Analysis, in press, 2022. * [60] K. Wu and C.-W. Shu. A provably positive discontinuous Galerkin method for multidimensional ideal magnetohydrodynamics. SIAM Journal on Scientific Computing, 40(5):B1302–B1329, 2018. * [61] K. Wu and C.-W. Shu. Provably positive high-order schemes for ideal magnetohydrodynamics: analysis on general meshes. Numerische Mathematik, 142(4):995–1047, 2019. * [62] K. Wu, D. Xiu, and X. Zhong. A WENO-Based stochastic Galerkin scheme for ideal MHD equations with random inputs. Communications in Computational Physics, 30(2):423–447, 2021. * [63] Z. Xu, D. S. Balsara, and H. Du. Divergence-free WENO reconstruction-based finite volume scheme for solving ideal MHD equations on triangular meshes. Communications in Computational Physics, 19(4):841–880, 2016.
# Some Comments on Unitary Qubit Lattice Algorithms for Classical Problems Paul Anderson 1 Lillian Finegold-Sachs 1 George Vahala 1 Linda Vahala 2 Abhay K. Ram 3 Min Soe 4 Efstratios Koukoutsis 5 Kyriakos Hizandis 51 1 Department of Physics William & Mary Williamsburg VA23185 2 Department of Electrical & Computer Engineering Old Dominion University Norfolk VA 23529 3 Plasma Science and Fusion Center MIT Cambridge MA 02139 4 Department of Mathematics and Physical Sciences Rogers State University Claremore OK 74017 5 School of Electrical and Computer Engineering National Technical University of Athens Zographou 15780 Greece ###### Abstract A qubit lattice algorithm (QLA), which consists of a set of interleaved unitary collision-streaming operators, is developed for electromagnetic wave propagation in tensor dielectric media. External potential operators are required to handle gradients in the refractive indices, and these operators are typically non-unitary. A similar problem arises in the QLA for the Korteweg-de Vries equation, as the potential operator that models the KdV nonlinear term is also non-unitary. Several QLAs are presented here that avoid the need of this non-unitary potential operator by perturbing the collision operator. These QLAs are fully unitary. ## 1 Introduction We have been investigating qubit lattice algorithms (QLA) for some time [1-21]. The aim of QLA is to develop a unitary interleaved sequence of collision-streaming operators which in the continuum limit reduces perturbatively to the desired differential equations describing the system of interest. The first step is to associate a basis set of qubits for the lattice, which on taking appropriate moments will recover the classical fields of interest. Thus the QLA would be immediately encodable onto a quantum computer. However, from our earlier nonlinear studies of 2D and 3D quantum turbulence [8,9, 11], the QLA is ideally parallelized on classical supercomputers with no degradation in parallel performance as the number of cores are ramped up [e.g., to over 750 000 cores on the $IBM$ BlueGene $Mira$ supercomputer at Argonne]. Some care is needed in the choice of the qubit basis. For example, it will be shown in Sec. 2 that the simple basis choice of $(\mathbf{E,H})$ will never lead to a unitary representation of the Maxwell equations of electrodynamics. [$\mathbf{E}$ is the electric field, and $\mathbf{H}$ is the magnetic field]. At the heart of an efficient algorithm on a quantum computer is quantum entanglement of the qubits. For example, consider a 2-qubit representation [e.g., for a basis for a differential equation like the scalar nonlinear Schrodinger (NLS) equation or the Korteweg de Vries (KdV) equation]. A basis is the $2^{2}$ elements $(|00\rangle,|01\rangle,|10\rangle,|11\rangle)$. Now consider a unitary $2\times 2$ collision operator $C=\left[\begin{array}[]{cc}\cos\theta&\sin\theta\\\ -\sin\theta&\cos\theta\end{array}\right]$ (1) acting on the qubit subspace of $(|01\rangle,|10\rangle)$. One of the post- collision qubit elements is $\cos\theta\,|01\rangle+\sin\theta\,|10\rangle.$ (2) However, this post-collision state cannot be represented by a tensor product of the $2^{2}$ \- basis, since the most general tensor product state is $a_{0}b_{0}|00\rangle+a_{0}b_{1}|01\rangle+a_{1}b_{0}|10\rangle+a_{1}b_{1}|11\rangle$ (3) where one qubit state is $a_{0}|0\rangle+a_{1}|1\rangle$ , and the other state is $b_{0}|0\rangle+b_{1}|1\rangle$ , for some coefficients $a_{0}\quad\ldots\quad b_{1}$. To recover the state Eq. (2) from the tenor product state Eq. (3) one would have to eliminate the $|00\rangle$-term. However one must then set either $a_{0}=0$ or $b_{0}=0$. But this would eliminate either the state $|01\rangle$ or the state $|10\rangle$ \- both of which are needed to recover Eq. (2). States which cannot be represented in a tensor product basis of qubits are called entangled states. A maximally entangled state is achieved on taking $\theta=\pi/4$, and is known as a Bell state [22] $B_{1}=\frac{|01\rangle+|10\rangle}{\sqrt{2}}$ Note that the quantum entanglement is achieved here by the unitary collision operator. The streaming operator in QLA will then propagate this entanglement throughtout the lattice. In Sec. 2 we will develop a QLA for the solution of 2D Maxwell equations in a tensor Hermitian dielectric medium. All our previous Maxwell QLA [18, 21] were restricted to scalar dielectrics. We will present a simplified discussion of the Dyson map [23] that will permit us to transform from a non-unitary to unitary basis for the representation of the two curl equations of Maxwell. From these qubit equations we will generate a QLA for tensor dielectric media that is second order accurate. The QLA that we discuss here is not fully unitary. While the collide-stream operator sequence is fully unitary, the external potential operators required to recover the Maxwell equations are not. However these non-unitary matrices are very sparse and could be amenable to some unitary approximate representation. The role of the perturbation parameter $\epsilon$ introduced in the QLA is quite subtle. As to an understanding of the subtlety of $\epsilon$ in QLA we return to the KdV equation in Sec. 3. Our original QLA for KdV [1] consisted of maximally entanglement Bell collision operators together with a required external potential operator to model the KdV nonlinearity. This external potential was not unitary. Here we present a modified collision operator that is fully unitary and which leads to a QLA-KdV that does not require any external potential to be introduced. Finally, in Sec. 4 we make some concluding remarks about future QLA simulations that are needed to elucidate the perturbation parameter $\epsilon$. This parameter is required in order to move the discrete QLA into a continuum representation. ## 2 QLA for Maxwell Equations ### 2.1 Qubit-Electromagnetic field representation Consider a simple dielectric non-magnetic medium with the constitutive equations $\mathbf{D}=\epsilon\mathbf{E},\quad\mathbf{B}=\mu_{0}\mathbf{H}.$ (4) Treating $\mathbf{u}=(\mathbf{E},\mathbf{H})^{\mathbf{T}}$ as the fundamental fields, and $\mathbf{d}=(\mathbf{D},\mathbf{B})^{\mathbf{T}}$ the derived fields , Eq. (2) can be written in matrix form $\mathbf{d}=\mathbf{Wu}$ (5) where $\mathbf{W}$ is a Hermitian $6\times 6$ matrix $\mathbf{W}=\left[\begin{array}[]{cc}\epsilon_{3\times 3}&0\\\ 0&\mu_{0}\mathbf{I}_{3\times 3}\end{array}\right]$ (6) with $\mathbf{I}_{3\times 3}$ the $3\times 3$ identity matrix. and $\mathbf{T}$ is the transpose operator. The curl-curl (source-free) Maxwell equations $\nabla\times\mathbf{E}=-\partial\mathbf{B}/\partial t$, and $\nabla\times\mathbf{H}=\partial\mathbf{D}/\partial t$ in matrix form are just $i\frac{\partial\mathbf{d}}{\partial t}=\mathbf{Mu}$ (7) where, under standard boundary conditions, the curl-matrix operator $\mathbf{M}$ is Hermitian $\mathbf{M}=\left[\begin{array}[]{cc}0_{3\times 3}&i\nabla\times\\\ -i\nabla\times&0_{3\times 3}\end{array}\right].$ (8) Since $\mathbf{W}$ is invertible, Eq. (5) can be written in terms of the basic electromagnetic fields $\mathbf{u}=(\mathbf{E},\mathbf{H})$ $i\frac{\partial\mathbf{u}}{\partial t}=\mathbf{W}^{-\mathbf{1}}\mathbf{M}\mathbf{u}$ (9) In continuum applications, one typically treats the two Maxwell divergence equations $\nabla\cdot\mathbf{B}=0$ and $\nabla\cdot\mathbf{D}=0$ as initial conditions. From the curl-curl equations we see that they will then be satisfied for all time. #### 2.1.1 homogeneous dielectric medium If one is dealing with a homogeneous dielectric medium (e.g., a vacuum), then the constitutive matrix $\mathbf{W}$ is a constant and trivially commutes with the curl-operator $\mathbf{M}$. As a result, the product of the two Hermitian matrices, $\mathbf{W}^{-\mathbf{1}}\mathbf{M}$ is itself Hermitian, and Eq. (7) gives a unitary evolution of the electromagnetic fields $\mathbf{u}=(\mathbf{E},\mathbf{H})^{\mathbf{T}}$. Thus $\mathbf{u}$ is an appropriate basis for the qubit fields and for quantum computation. #### 2.1.2 inhomogeneous dielectric media However, when the matrix $\mathbf{W}$ is spatiaily dependent, then $\mathbf{W}^{-\mathbf{1}}\mathbf{M}\neq\mathbf{M}\mathbf{W}^{-\mathbf{1}}$ and $\mathbf{W}^{-\mathbf{1}}\mathbf{M}$ is not Hermitian. Under these conditions, the qubit representation of the electromagnetic fields $\mathbf{u}=(\mathbf{E},\mathbf{H})^{\mathbf{T}}$ will not yield a unitary evolution of these qubits. However Koukoutsis et. al. [23] have shown how to determine the so-called Dyson map from the fields $\mathbf{u}$ to a new field representation $\mathbf{U}$ such that the resultant representation in terms of the new field $\mathbf{U}$ will result in a unitary evolution of these fields. Indeed, it can be shown [23], that the Dyson map $\mathrm{U}=\mathrm{W}^{1/2}\mathrm{u}$ (10) will yield a unitary evolution equation for $\mathbf{U}$ with $i\frac{\partial\mathbf{U}}{\partial t}=\mathbf{W}^{-1/2}\mathbf{M}\mathbf{W}^{-1/2}\mathbf{U}$ (11) as the matrix operator $\mathbf{W}^{-\mathbf{1}/\mathbf{2}}\mathbf{M}\mathbf{W}^{-\mathbf{1}/\mathbf{2}}$ is Hermitian. Thus one could start to build a QLA based on the electromagnetic fields $\mathbf{U}=\left(\epsilon^{1/2}\mathbf{E},\mu_{0}^{1/2}\mathbf{H}\right)^{T}$ (12) or under the rotation matrix $\mathbf{L}=\frac{1}{\sqrt{2}}\left[\begin{array}[]{cc}I_{3\times 3}&iI_{3\times 3}\\\ I_{3\times 3}&-iI_{3\times 3}\end{array}\right]$ (13) one could base a QLA on the field representation $\mathbf{U}_{\mathbf{RSW}}=\mathbf{LU}$ where $\mathbf{U}_{\mathbf{RSW}}=\frac{1}{\sqrt{2}}\left[\begin{array}[]{c}\epsilon^{1/2}\mathbf{E}+i\mu_{0}^{1/2}\mathbf{H}\\\ \epsilon^{1/2}\mathbf{E}-i\mu_{0}^{1/2}\mathbf{H}\end{array}\right].$ (14) This is nothing but the unitary evolution of the Riemann-Silberstein-Weber (RSW) vector - a representation used to represent Maxwell equations from the early 1920’s [24-26]. Moreover, the theory can be readily extended to diagonal tensor dielectric media, with (assuming non-magnetic materials) the 6-qubit representation $\mathbf{Q}$ of the field $\mathbf{U}=\left(n_{x}E_{x},n_{y}E_{y},n_{z}E_{z},\mu_{0}^{1/2}\mathbf{H}\right)^{T}=\mathbf{Q}$ (15) $(n_{x},n_{y},n_{z})$ is the vector (diagonal) refractive index, with $\epsilon_{x}=n_{x}^{2}$ … . We work in Cartesian coordinates. ### 2.2 $\mathrm{2D}$ QLA for $\mathrm{x-y}$ dependent propagation of Maxwell Equations From Eqs. (9) and (13), Maxwell equations for 2D x-y spatially dependent fields written in terms of the 6-$\mathbf{Q}$ vector components $\displaystyle\frac{\partial q_{0}}{\partial t}=\frac{1}{n_{x}}\frac{\partial q_{5}}{\partial y},\qquad\frac{\partial q_{1}}{\partial t}=-\frac{1}{n_{y}}\frac{\partial q_{5}}{\partial x},\qquad\frac{\partial q_{2}}{\partial t}=\frac{1}{n_{z}}\left[\frac{\partial q_{4}}{\partial x}-\frac{\partial q_{3}}{\partial y}\right]$ (16) $\displaystyle\frac{\partial q_{3}}{\partial t}=-\frac{\partial(q_{2}/n_{z})}{\partial y},\qquad\frac{\partial q_{4}}{\partial t}=\frac{\partial(q_{2}/n_{z})}{\partial x},\qquad\frac{\partial q_{5}}{\partial t}=-\frac{\partial(q_{1}/n_{y})}{\partial x}+\frac{\partial(q_{0}/n_{x})}{\partial y}$ This representation is unitary. Our QLA representation focusses on recovering Eq. (14) perturbatively. One can thus consider developing the representation dimension by dimension. In particular we introduce the following unitary collision operator with collision angles $\theta_{1}$ and $\theta_{2}$ (to be specified later): $C_{X}=\left[\begin{array}[]{cccccc}1&0&0&0&0&0\\\ 0&cos\,\theta_{1}&0&0&0&-sin\,\theta_{1}\\\ 0&0&cos\,\theta_{2}&0&-sin\,\theta_{2}&0\\\ 0&0&0&1&0&0\\\ 0&0&sin\,\theta_{2}&0&cos\,\theta_{2}&0\\\ 0&sin\,\theta_{1}&0&0&0&cos\,\theta_{1}\end{array}\right]$ (17) and the unitary collision operator $C_{Y}=\left[\begin{array}[]{cccccc}cos\,\theta_{0}&0&0&0&0&sin\,\theta_{0}\\\ 0&1&0&0&0&0\\\ 0&0&cos\,\theta_{2}&0&sin\,\theta_{2}&0\\\ 0&0&-sin\,\theta_{2}&cos\,\theta_{2}&0&0\\\ 0&0&0&0&1&0\\\ -sin\,\theta_{0}&0&0&0&0&cos\,\theta_{0}\end{array}\right]$ (18) with collision angles $\theta_{0}$ and $\theta_{2}$. The unitary streaming operator $S^{+x}_{14}$ shifts qubits $q_{1}$ and $q_{4}$ one lattice unit in the $+x$ direction, while leaving the remaining 4 qubits alone. We finally need to introduce the external potential operators $V_{X}=\left[\begin{array}[]{cccccc}1&0&0&0&0&0\\\ 0&1&0&0&0&0\\\ 0&0&1&0&0&0\\\ 0&0&0&1&0&0\\\ 0&0&-sin\,\beta_{2}&0&cos\,\beta_{2}&0\\\ 0&sin\,\beta_{0}&0&0&0&cos\,\beta_{0}\end{array}\right]$ (19) and $V_{Y}=\left[\begin{array}[]{cccccc}1&0&0&0&0&o\\\ 0&1&0&0&0&0\\\ 0&0&1&0&0&0\\\ 0&0&\cos\,\beta_{3}&\sin\,\beta_{3}&0&0\\\ 0&0&0&0&1&0\\\ -sin\,\beta_{1}&0&0&0&0&cos\,\beta_{1}\end{array}\right]$ (20) for particular angles $\beta_{0}$, $\beta_{1}$ and $\beta_{2}$. These potential operators are not unitary, but very sparse. We now consider the following unitary sequence of interleaved collision- streaming operators: $\mathbf{U_{X}}=S^{+x}_{25}.C_{X}^{\dagger}.S^{-x}_{25}.C_{X}.S^{-x}_{14}.C_{X}^{\dagger}.S^{+x}_{14}.C_{X}.S^{-x}_{25}.C_{X}.S^{+x}_{25}.C_{X}^{\dagger}.S^{+x}_{14}.C_{X}.S^{-x}_{14}.C_{X}^{\dagger}$ (21) and $\mathbf{U_{Y}}=S^{+y}_{25}.C_{Y}^{\dagger}.S^{-y}_{25}.C_{Y}.S^{-y}_{03}.C_{Y}^{\dagger}.S^{+y}_{03}.C_{Y}.S^{-y}_{25}.C_{Y}.S^{+y}_{25}.C_{Y}^{\dagger}.S^{+y}_{03}.C_{Y}.S^{-y}_{03}.C_{Y}^{\dagger}$ (22) with the discrete time advancement of the 6-qubit $\mathbf{Q}$ given by $\mathbf{Q}(t+\delta t)=V_{Y}.V_{X}.\mathbf{U_{Y}}.\mathbf{U_{X}}.\mathbf{Q}(t)$ (23) To recover the desired Maxwell equations (14) perturbatively, one employ a small parameter $\epsilon$ as the spatial lattice shift unit (assuming a square $x-y$ lattice), and the unitary collision angles $\theta_{0}=\frac{\epsilon}{4n_{x}}\quad,\qquad\theta_{1}=\frac{\epsilon}{4n_{y}}\quad,\qquad\theta_{2}=\frac{\epsilon}{4n_{z}}.$ (24) so as to recover the coefficients of the $\partial\mathbf{Q}/\partial(x,y)$ terms. Finally, the nonunitary external potential angles need to be defined as $\beta_{0}=\epsilon^{2}\frac{\partial n_{y}/\partial x}{n^{2}_{y}}\quad,\quad\beta_{1}=\epsilon^{2}\frac{\partial n_{x}/\partial y}{n^{2}_{x}}\quad,\quad\beta_{2}=\epsilon^{2}\frac{\partial n_{z}/\partial x}{n^{2}_{z}}\quad,\quad\beta_{3}=\epsilon^{2}\frac{\partial n_{z}/\partial y}{n^{2}_{z}}$ (25) Indeed, using Mathematica to evaluate Eq. (21), one obtains in the continuum spatial limit the desired Maxwell equations to errors of $\epsilon^{4}$ $\displaystyle\frac{\partial q_{0}}{\partial t}=\epsilon^{2}\delta t\frac{1}{n_{x}}\frac{\partial q_{5}}{\partial y},\qquad\frac{\partial q_{1}}{\partial t}=-\epsilon^{2}\delta t\frac{1}{n_{y}}\frac{\partial q_{5}}{\partial x},\qquad\frac{\partial q_{2}}{\partial t}=\epsilon^{2}\delta t\frac{1}{n_{z}}\left[\frac{\partial q_{4}}{\partial x}-\frac{\partial q_{3}}{\partial y}\right]$ (26) $\displaystyle\frac{\partial q_{3}}{\partial t}=-\epsilon^{2}\delta t\frac{\partial(q_{2}/n_{z})}{\partial y},\qquad\frac{\partial q_{4}}{\partial t}=\epsilon^{2}\delta t\frac{\partial(q_{2}/n_{z})}{\partial x},\qquad\frac{\partial q_{5}}{\partial t}=-\epsilon^{2}\delta t\left(\frac{\partial(q_{1}/n_{y})}{\partial x}+\frac{\partial(q_{0}/n_{x})}{\partial y}\right)$ i.e., under diffusion ordering, $\epsilon^{2}\delta t\approx O(1)$, one recovers the continuum Maxwell equations to errors $O(\epsilon^{2})$. We will explore QLA simulations of 2D Maxwell equations in a subsequent paper. ## 3 QLA for KdV without external non-unitary potential operators The KdV equation is an important nonlinear equation and was developed to explore the evolution of shallow water waves. Interestingly, it [27] has also been associated with the Fermi-Pasta-Ulam-Tsingou simulations of the 1950’s. Fermi wanted to examine the equipartition of energy among the modes of a many body problem of weakly coupled nonlinear oscillators. Statistical mechanics indicates that the time- asymptotic state will be one in which there is equipartition of energy among all the oscillator modes. Instead, in the parameter regime they considered, Fermi et. al. found recurrence of initial conditions, but a recurrence that was not a usually extremely long Poincare recurrence time of Hamiltonian systems. Interestingly, this would turn out to be a precursor to soliton theory. The general $\mathrm{KdV}$ equation for arbitrary positive constants $a$ and $b$ $\frac{\partial\psi}{\partial t}+a\psi\frac{\partial\psi}{\partial x}+b\frac{\partial^{3}\psi}{\partial x^{3}}=0$ (27) is exactly integrable. One of its solutions is the right traveling soliton with speed $c$ \- a free parameter $\psi(x,t)=\frac{3c}{a}\operatorname{sech}^{2}\left(\frac{1}{2}\sqrt{\frac{c}{b}}[x-ct]\right)$ (28) Notice that for the KdV soliton, the amplitude and its speed are correlated (unlike the NLS soliton). Since the $\mathrm{KdV}$ equation is a scalar equation for the real function $\psi(x,t)$ one need only to employ 2 qubits / lattice site. First we shall reconsider the QLA for KdV with the use of an external potential to model the nonlinear term in KdV [1]. The collision operator is nothing but Eq. (1). We denote the operator $S_{0}^{+}$ to be the streaming operator that translates the qubit $q_{0}$ one lattice unit in the $+x$-direction. To eliminate the 2nd order spatial derivative one must choose the interleaved sequence of collision-stream unitary operators carefully. In particular the following sequence will generate a second order QLA for the KdV equation $Q(t+\Delta t)=V_{pot}\cdot S_{0}^{+}C\cdot S_{1}^{-}C^{T}\cdot S_{0}^{-}C\cdot S_{1}^{+}C^{T}\cdot S_{0}^{-}C^{T}\cdot S_{1}^{+}C\cdot S_{0}^{+}C^{T}\cdot S_{1}^{-}C\cdot Q(t)$ (29) where the unitary collision operator $C$ is nothing but the maximally entangling operator. Eq, (1), with $\theta=\pi/4$. $Q=\left(q_{0}\,q_{1}\right)^{T}$. The external potential $V_{\text{pot }}$ is the Hermitian matrix $V_{\text{pot }}=\left[\begin{array}[]{cc}\cos\alpha&-\sin\alpha\\\ -\sin\alpha&\cos\alpha\end{array}\right]\quad\text{ with }\quad\alpha=\epsilon^{3}m[x].$ (30) In the continuum limit, one recovers $\frac{\partial\psi}{\partial t}+\epsilon^{3}\left(m[x]\cdot\psi(x,t)+\frac{1}{2}\frac{\partial^{3}\psi}{\partial x^{3}}\right)=0+O\left(\epsilon^{5}\right)$ (31) on defining $\psi=q_{0}+q_{1}$. With the choice of $m[x]=\partial\psi/\partial x$ we have a second order accurate QLA for KdV. Note that the QLA of Eq. (26) is not fully unitary because of the non-unitary property of the external potential operator $V_{\text{pot }}$. ### 3.1 Fully unitary QLAs for KdV There is a large class of unitary QLAs all of which recover the KdV equation to second order accuracy. Here, we will present two QLAs, both having the same unitary collision operator, but with different streaming sequences on the two qubits. Indeed, using Mathematica, it can be shown that the following QLA $Q(t+\Delta t)=S_{0}^{-}C_{1}\cdot S_{0}^{+}C_{1}\cdot S_{1}^{+}C_{1}\cdot S_{1}^{-}C_{1}^{T}\cdot S_{0}^{-}C_{1}^{T}\cdot S_{0}^{+}C_{1}^{T}\cdot S_{1}^{+}C_{1}^{T}\cdot S_{1}^{-}C_{1}\cdot Q(t)$ (32) with unitary collision operator $C_{1}$ $C_{1}=\left[\begin{array}[]{cc}\cos\alpha_{1}&\sin\alpha_{1}\\\ -\sin\alpha_{1}&\cos\alpha_{1}\end{array}\right]\quad\text{ with }\quad\alpha_{1}=\frac{\pi}{4}+\epsilon^{2}m_{1}[x].$ (33) leads in the continuum limit to $\frac{\partial\psi_{1}}{\partial t}+\epsilon^{3}\left(4m_{1}[x]\cdot\frac{\partial\psi_{1}}{\partial x}+\frac{1}{2}\frac{\partial^{3}\psi_{1}}{\partial x^{3}}\right)=0+O\left(\epsilon^{5}\right)$ (34) so that the choice of $m[x]=\psi_{1}$ will recover $\mathrm{KdV}$. Another fully unitary QLA that recovers KdV has the following interleaved sequence of unitary collision-streaming operators: $Q(t+\Delta t)=C_{1}S_{0}^{-}\cdot C_{1}S_{1}^{+}\cdot C_{1}S_{0}^{-}\cdot C_{1}S_{1}^{+}\cdot C_{1}^{T}S_{0}^{+}\cdot C_{1}^{T}S_{1}^{-}\cdot C_{1}^{T}S_{0}^{+}\cdot C_{1}^{T}S_{1}^{-}\cdot Q(t)$ (35) $C_{1}$ is the same collision operator, Eq. (33). In the continuum limit, we find $\frac{\partial\psi_{1}}{\partial t}+\epsilon^{3}\left(-4m_{1}[x]\cdot\frac{\partial\psi_{1}}{\partial x}+\frac{1}{2}\frac{\partial^{3}\psi_{1}}{\partial x^{3}}\right)=0+O\left(\epsilon^{5}\right)$ (36) so that the choice of $m_{1}[x]=-\psi_{1}$ will recover the KdV equation. The implementation of these fully unitary algorithms may not necessarily be straightforward as the perturbation parameter $\epsilon$ introduced into the Mathematica algorithm requires a perturbation in the collision angle of $O\left(\epsilon^{2}\right)$, Eq. (33), while the continuum limit has scaling proceeds as $O\left(\epsilon^{3}\right)$. In previous QLA for nonlinear physics, the order of the function $\psi$ controlled the $\epsilon$-factor. ## 4 SUMMARY The development of a fully unitary QLA for plasma physics [28-31] in particular, is of considerable interest to us as these algorithms are immediately encodable on quantum computers. Of course, since they are time evolution algorithms, they will have to wait till there are error-correcting quantum computers available. In developing QLAs for plasma physics we have taken the tack of first concentrating on the Maxwell equations in a given scalar dielectric media. Then one would eventually generalize to a tensor unitary dielectric description of a cold magnetized plasma. Here, we have shown how to generalize our QLA-scalar dielectric Maxwell equations to handle tensor Hermitian dielectric media. This was facilitated by the use of the Dyson map [23]. Indeed, the explicit determination of the Dyson map proves that there exists a unitary quantum algorithm to describe such a Maxwell system. The problem, of course, is to explicitly construct such an algorithm. We have concentrated on the QLA approach, which employs a non- trivial sequence of interleaved collide-stream unitary operators. To proceed explicitly with QLA, one must resort to perturbation theory and the introduction of a small parameter $\epsilon$. The uinitary collide-stream operator sequence does not in all our Maxwell equation considerations recover the required full set of evolution equations. This has resulted in the need to introduce so-called potential operators in order to recover the equation of interest. At least one of these potential operators turns out to be non- unitary. As these QLAs have parallelized outstandingly on classical supercomputer architectures, outperforming standard computational fluid dynamic codes for the study of quantum turbulence, we have proceeded with the numerical implementation of such QLAs. This seems prudent as an error- correcting quantum computer with long qubit coherence times is still on the somewhat distant horizon. Nevertheless we are also pursuing a fully unitary QLA. In particular, we are revisiting our original non-unitary QLA-KdV [1] to determine a fully unitary QLA. We have found a large class of such unitary QLA-KdV, based on changing the particular collide-stream sequences. The implementation of these fully unitary algorithms may not necessarily be straightforward as the perturbation parameter $\epsilon$ introduced into the Mathematica symbolic manipulations requires a perturbation in the collision angle of $O\left(\epsilon^{2}\right)$, Eq. (33), while the continuum limit has scaling proceeds as $O\left(\epsilon^{3}\right)$, Eq. (34) or (36). In previous QLA for nonlinear physics, the order of the function $\psi$ was used as the $\epsilon$-factor. We believe that understanding the role of $\epsilon$ in the QLA-KdV simulations will be pivotal in handling the role of $\epsilon$ in QLA-Maxwell in both scalar and tensor dielectric media. These simulations will be reported in a future publication. ## 5 Acknowledgments This research was partially supported by Department of Energy grants DE- SC0021647, DE-FG0291ER-54109, DE-SC0021651, DE-SC0021857, and DE-SC0021653. This work has been carried out partially within the framework of the EUROfusion Consortium. E.K has received funding from the Euratom research and training program WPEDU under grant agreement no. 101052200 as well as from the National Program for Controlled Thermonuclear Fusion, Hellenic Republic. K.H is supported by the National Program for Controlled Thermonuclear Fusion, Hellenic Republic. The views and opinions expressed herein do not necessarily reflect those of the European Commission. ## 6 References [1] VAHALA, G, VAHALA, L & YEPEZ, J. 2003 Quantum lattice gas representation of some classical solitons. Phys. Lett A310, 187-196 [2] VAHALA, L, VAHALA, G & YEPEZ, J. 2003 Lattice Boltzmann and quantum lattice gas representations of one-dimensional magnetohydrodynamic turbulence. Phys. Lett A306, 227-234. [3] VAHALA, G, VAHALA, L & YEPEZ, J. 2004. Inelastic vector soliton collisions: a latticebased quantum representation. Phil. Trans: Mathematical, Physical and Engineering Sciences, The Royal Society, 362, 1677-1690 [4] VAHALA, G, VAHALA, L & YEPEZ, J. 2005 Quantum lattice representations for vector solitons in external potentials. Physica A362, 215-221. [5] YEPEZ, J. 2002 An efficient and accurate quantum algorithm for the Dirac equation. arXiv: 0210093. [6] YEPEZ, J. 2005 Relativistic Path Integral as a Lattice-Based Quantum Algorithm. Quant. Info. Proc. 4, 471-509. [7] YEPEZ, J, VAHALA, G & VAHALA, L. 2009a Vortex-antivortex pair in a Bose- Einstein condensate, Quantum lattice gas model of theory in the mean-field approximation. Euro. Phys. J. Special Topics 171, 9-14 [8] YEPEZ, J, VAHALA, G, VAHALA, L & SOE, M. 2009b Superfluid turbulence from quantum Kelvin wave to classical Kolmogorov cascades. Phys. Rev. Lett. 103, 084501. [9] VAHALA, G, YEPEZ, J, VAHALA, L, SOE, M, ZHANG, B, & ZIEGELER, S. 2011 Poincaré recurrence and spectral cascades in three-dimensional quantum turbulence. Phys. Rev. E84, 046713 [10] VAHALA, G, YEPEZ, J, VAHALA, L &SOE, M, 2012 Unitary qubit lattice simulations of complex vortex structures. Comput. Sci. Discovery 5, 014013 [11] VAHALA, G, ZHANG, B, YEPEZ, J, VAHALA. L & SOE, M. 2012 Unitary Qubit Lattice Gas Representation of 2D and 3D Quantum Turbulence. Chpt. 11 (pp. 239 - 272), in Advanced Fluid Dynamics, ed. H. W. Oh, (InTech Publishers, Croatia) [12] YEPEZ, J. 2016 Quantum lattice gas algorithmic representation of gauge field theory. SPIE 9996, paper 9996-2 [13] OGANESOV, A, VAHALA, G, VAHALA, L, YEPEZ, J & SOE, M. 2016a. Benchmarking the Dirac-generated unitary lattice qubit collision-stream algorithm for 1D vector Manakov soliton collisions. Computers Math. with Applic. 72, 386 [14] OGANESOV, A, FLINT, C, VAHALA, G, VAHALA, L, YEPEZ, J & SOE, M 2016b Imaginary time integration method using a quantum lattice gas approach. Rad Effects Defects Solids $171,96-102$ [15] OGANESOV, A, VAHALA, G, VAHALA, L & SOE, M. 2018. Effects of Fourier Transform on the streaming in quantum lattice gas algorithms. Rad. Eff. Def. Solids, 173, 169-174 [16] VAHALA, G., SOE, M., VAHALA, L., & RAM, A. K., 2021 One- and Two- Dimensional quantum lattice algorithms for Maxwell equations in inhomogeneous scalar dielectric media I : theory. Rad. Eff. Def. Solids 176, 49-63. [17] VAHALA, G., SOE, M., VAHALA, L., & RAM, A. K., 2021 One- and Two- Dimensional quantum lattice algorithms for Maxwell equations in inhomogeneous scalar dielectric media II : Simulations. Rad. Eff. Def. Solids 176, 64-72. [18] VAHALA, G, VAHALA, L, SOE, M & RAM, A, K. 2020. Unitary Quantum Lattice Simulations for Maxwell Equations in Vacuum and in Dielectric Media, J. Plasma Phys 86, 905860518 [19] VAHALA, L, SOE, M, VAHALA, G & YEPEZ, J. 2019a. Unitary qubit lattice algorithms for spin-1 Bose-Einstein condensates. Rad Eff. Def. Solids 174, 46-55 [20] VAHALA, L, VAHALA, G, SOE, M, RAM, A & YEPEZ, J. 2019b. Unitary qubit lattice algorithm for three-dimensional vortex solitons in hyperbolic self- defocusing media. Commun Nonlinear Sci Numer Simulat 75, 152-159 [21] RAM, A. K., VAHALA, G., VAHALA, L. & SOE, M 2021 Reflection and transmission of electromagnetic pulses at a planar dielectric interface - theory and quantum lattice simulations AIP Advances 11, 105116 (1-12). [22] MERMIN, N. D., 2007 Quantum computer science, Cambridge University Press, Cambridge [23] KOUKOUTSIS, E., HIZANIDIS, K., RAM, A. K., & VAHALA, G. 2022. Dyson Maps and Unitary Evolution for Maxwell Equations in Tensor Dielectric Media. arXiv:2209.08523 [24] LAPORTE, O. & UHLENBECK, G. E. 1931 Application of spinor analysis to the Maxwell and Dirac equations. Phys. Rev. 37, 1380-1397. [25] OPPENHEIMER, J. R. 1931 Note on light quanta and the electromagnetic field. Phys. Rev. 38, 725-746. [26] MOSES, E. 1959 Solutions of Maxwell’s equations in terms of a spinor notation: the direct and inverse problems, Phys. Rev. 113, 1670-1679 [27] BERMAN, FG. P., & IZRAILEV, F. M., 2005 The Fermi-Pasta-Ulam problem: Fifty years of progress. Chaos 15, 015104. [28] DODIN, I. Y. & STARTEV, E. A. 2021 On appliations of quantum computing to plasma simulations. Phys. Plasmas 28, 092101 [29] ENGEL, A., SMITH, G & PARKER, S. E. 2019 Quantum Algorithm for the Vlasov Equation, Phys. Rev. 100, 062315 [30] ENGEL, A., SMITH, G & PARKER, S. E. 2021. Linear Embedding of nonlinear dynamial systems and prospects for efficient quantum algorithms, Phys. Plasmas 28, 062305 [31] LIU, J-P, KOLDEN, H.O, KROVI, H. K, LOUREIRO, N. F, TRIVISA, K, & CHILDS, A. M. 2021. Proc. Natl. Acad. Sciences 118, e2026805118.
# A Blockchain-based Semantic Exchange Framework for Web 3.0 toward Participatory Economy Yijing Lin, Zhipeng Gao*, Yaofeng Tu, Hongyang Du, Dusit Niyato, Jiawen Kang, Hui Yang ###### Abstract Web 3.0 is the next-generation Internet that enables participants to read, write, and own contents in a decentralized manner. It is mainly driven by blockchain, semantic communication, edge computing, and artificial intelligence, which can construct value networks to achieve participatory economics based on participatory decision making. Web 3.0 can capture the characteristics of blockchain, semantic extraction, and communication to achieve decentralized semantic sharing and transfer information precisely. However, current Web 3.0 solutions focus on the blockchain while overlooking other new technologies’ roles in Web 3.0. To further unleash the advantages of semantic extraction and communication in Web 3.0, in this paper, we propose a blockchain-based semantic exchange framework to realize fair and efficient interactions. In this framework, we first attempt to tokenize semantic information into Non-Fungible Token (NFT) for semantic exchange. Then we utilize a Stackelberg game to maximize buying and pricing strategies for semantic trading. We also leverage Zero-Knowledge Proof to share authentic semantic information without publishing it before receiving payments, which can achieve a fair and privacy-preserving trading compared with current NFT marketplaces. A case study about urban planning is given to show clearly the proposed mechanisms. Finally, several challenges and opportunities are identified. ###### Index Terms: Web 3.0, NFT, Optimization, Zero-Knowledge Proof ## I Introduction Web 3.0 is now attracting the interest of researchers from academics and industries. It is considered to be the next generation of the Internet where participants can control their data, identities, and capabilities to make decentralized decisions and construct participatory economy. There are two mainstream developments of Web 3.0 as follows, including Blockchain Web 3.0 and Semantic Web 3.0. When the concept of Web 3.0 first emerged, it was identified as a semantic web, which can make Internet data machine-readable [1] and reduce energy consumption for processing and transmission. The current definition of Web 3.0 [2] was prevailing with the help of blockchain technologies, which can make participants get rid of tech giants and control user-generated content. Besides, since artificial intelligence and edge computing play important roles to connect Web 3.0 participants in a smart and lightweight way, they are also considered vital components of Web 3.0. The diversified underlying technologies of Web 3.0 are similar to Web 2.0 mainly driven by cloud computing, social networks, and mobile connectivity. Instead of being controlled by tech giants, Web 3.0 allows all participants to make decisions to participate in constructing value networks, which should permit participants without the need of having the expertise of underlying technologies, like DeFi Money Legos [3]. Thus, Web 3.0 can be supported by multiple technologies like blockchain, artificial intelligence, edge computing, and semantic communication, which can easily coordinate to construct value networks to achieve participatory economy based on decentralized decision making among participants. Although several new technologies are emerging as main components to construct value networks in Web 3.0, blockchain is considered the main technology for the participatory economy in many cases. Few researchers focus on how to integrate the components of new technologies into blockchain-based Web 3.0 toward the participatory economy. DeCAST [4] utilizes the integration of blockchain and federated learning to construct parallel intelligent transportation systems, which are effective and practical for smart mobility. A unified blockchain-semantic framework [5] is proposed to capture the benefits of blockchain, semantic extraction, and communication to improve interaction efficiency for Web 3.0. Besides, given the steps of the integration of new technologies in Web 3.0, it is necessary to consider how to make value circulated in the participatory economy to attract participants to join in Web 3.0 decision-making. Figure 1: Difference Between Web 1.0, Web 2.0 and Web 3.0 Figure 2: Applications of Blockchain and Semantic Communication-enabled Web 3.0 Blockchain, semantic extraction and communication can be utilized to reduce information overloaded, reduce energy consumption, and construct efficient interactions toward the participatory economy by extracting semantic information from raw data and implementing sharing in a decentralized manner. However, even capitalizing on the benefits of the above technologies, the design toward the participatory economy still faces many difficulties. On the one hand, it is difficult to quantify and transact semantic information circulated in Web 3.0. The reason is that producers and consumers are difficult to maximize their revenue and utility [6]. On the other hand, Web 3.0 is supported by a public blockchain platform, which makes pending semantic information exposed to both parties. Consumers can copy semantic information without payments to maximize their benefits [7] in current marketplaces. In this paper, we attempt to complement the understanding of blockchain, semantic extraction and communication playing an important role in Web 3.0. We also develop a fair and efficient mechanism based on blockchain and semantic communication to tokenize semantic information and enable semantic exchange by utilizing the Non-Fungible Token (NFT), Stackelberg Game, and Zero Knowledge Proof (ZKP) toward participatory economy. Our contributions are summarized as follows. * • We propose a new semantic exchange framework, that tokenize semantic information to facilitate semantic exchange by using blockchain. * • In this framework, we formulate a Stackelberg game-based semantic trading mechanism to design optimal pricing strategies. * • To facilitate fair and privacy-preserving transactions between producers and consumers, we utilize zero-knowledge proof to transact transformed semantic information without publishing authentic one before payments. * • To show clearly the applications and mechanisms, we utilize urban planning to display how the above contributions work. ## II The Future of Blockchain, Semantic Extraction and Communication-enabled the Internet ### II-A An Accelerating Growth of The Internet The accelerating growth of the iterations of the Internet is witnessed from Web 1.0 to Web 3.0, as shown in Fig. 1. According to Fig. 1, we illustrate the difference between Web 1.0, Web 2.0, and Web 3.0, the connection between Web 3.0, blockchain, edge computing, and semantic communication, and the applications of blockchain and semantic communication-enabled Web 3.0. #### II-A1 What Is The Difference Between Web 1.0, Web 2.0, And Web 3.0 The development of the Internet experiences three iterations of Web 1.0, Web 2.0, and Web 3.0 eras. Web 1.0 is a read-only web, where users can only read information produced by centralized content producers from centralized servers. Web 2.0 is a read-write web driven by mobile connectivity, social networks, and cloud computing, where users can read and write content on platforms provided by giant tech companies. Platforms can utilize content produced by users to make a profit like advertising without paying any to producers. Therefore, the need for Web 3.0 is that users expect to control their data and protect their privacy in a decentralized manner. Web 3.0 is a read-write-own web driven by blockchain, edge computing, artificial intelligence, and semantic communication, where users can read, write, and own content on decentralized value networks. #### II-A2 What Is The Connection Between Web 3.0, Blockchain, Edge Computing, And Semantic Communication Whereas Web 2.0 was mainly driven by mobile connectivity, social networks, and cloud computing to allow users to participate in constructing platform networks, Web 3.0 is enabled by blockchain, edge computing, artificial intelligence, and semantic communication to design an open, trustless, seamless, ubiquitous, and efficient next-generation Internet. Blockchain is often considered the same as Web 3.0, while it is actually an infrastructure to support decentralized data sharing in an open and trustless manner. Edge computing is also vital to construct Web 3.0, which can leverage the computing and storage capabilities of ubiquitous edge devices and servers to make Internet services available and near to users. Artificial intelligence can leverage tons of data produced by edge computing and blockchain to train models, which can improve services and provide seamless experiences. Instead of transmitting information accurately, semantic communication [8] based on edge computing and artificial intelligence pursues how precisely the transmitted information can convey the desired meanings to reduce information overloaded and energy consumed to provide Web 3.0 with efficiency. #### II-A3 What Are Applications of Blockchain And Semantic Communication- enabled Web 3.0 Whereas Web 2.0 mainly focuses on applications on platform networks for controllers to make a profit from user data, blockchain and semantic communication-enabled Web 3.0 concentrate on value networks to facilitate transparent use of data, and decentralized data exchange and sharing. Blockchain and semantic communication-enabled Web 3.0 toward participatory economy can be applied for embracing the digital transformation of healthcare services, the automotive industry, and urban planning, as shown in Fig. 2. ### II-B Research Gaps of Blockchain and Semantic Communication-enabled Web 3.0 Instead of a mansion in the air, Web 3.0 is the next-generation infrastructure consisting of multiple building blocks like blockchain and semantic communication, which is also similar to blockchain made up of cryptography and distributed algorithms. Web 3.0 is expected to construct an entirely new Internet for participants to transact and interact in the participatory economy. Previous works only focus on the blockchain or semantic-enabled Web 3.0, while there are few works focusing on the integration between Web 3.0, blockchain, semantic extraction and communication. DeSci [9] utilized blockchain and decentralized autonomous organizations to construct a reference model for scientific systems. A unified blockchain-semantic framework enabled Web 3.0 was proposed to maintain service security and improve interaction efficiency in Web 3.0 [5]. HyperService [10] was designed to construct an interoperable cross-chain platform for decentralized applications to interact data across blockchains. Besides, blockchain and semantic-enabled Web 3.0 still suffers from the following challenges: 1) how to release the value of semantic information when combining semantic communication and blockchain with Web 3.0, 2) how to maximize the revenue and utility of semantic information producers and consumers in the participatory economy, 3) how to trade semantic information among producers and consumers in a fair manner. To solve the above challenges, we propose a fair and efficient blockchain-based semantic exchange framework for Web 3.0 by NFT-based semantic exchange, Stackelberg game-based semantic trading, and zero-knowledge proof-based semantic sharing mechanisms to form the participatory economy. ## III Fair and Efficient Mechanism Based on Semantic Blockchain In this section, we describe the NFT-based semantic exchange scheme to facilitate semantic circulation. We also propose a Stackelberg game-based dynamic semantic trading mechanism to maximize the utility and revenue of semantic information. Moreover, we construct a zero-knowledge proof-based semantic sharing mechanism to enable fair authentic semantic sharing. ### III-A NFT-based Semantic Exchange Scheme Semantic information extracted from producers and exchanged in semantic communications has been considered how to determine the importance of information [8]. Thus, for semantic information of high importance, it is necessary to decide digital ownerships to facilitate the circulation of valuable semantic information. Therefore, we propose an NFT-based semantic exchange scheme to tokenize semantic information and unalterably prove the digital ownership of producers. Figure 3: Case Study in Urban Planning The NFT-based semantic exchange scheme is composed of three functional parts, including blockchain, off-chain storage, and semantic communication systems. The blockchain is constructed by edge servers to enable consensus and smart contracts, which can support NFT-based semantic exchange transactions between edge devices. Since blockchain has limited storage spaces, off-chain storage is utilized to supplement spaces for blockchain to interact with texts, images, and videos in an efficient way. Edge devices can contribute spare storage spaces to extend the off-chain storage to expand the storage ability of blockchain. Once semantic information is uploaded into off-chain storage, the off-chain storage will return a hash value to uniquely map the semantic information to prevent tampering. The semantic communication systems include semantic information producers and consumers to extract, exchange, and consume semantic information to reduce information overloaded. Moreover, since semantic communication systems need to interact with the blockchain and off- chain storage, edge devices (producers and consumers) should equip with key generation, blockchain client, and off-chain storage client modules to implement semantic information circulating. The workflow of the NFT-based semantic exchange scheme is described as follows. * • Step 1: Write and Deploy NFT Smart Contracts. NFTs are on-chain credentials for semantic information and proofs of the ownership of producers, which are implemented by smart contracts. The function of NFT smart contracts mainly include Mint, Transfer, and Burn. Mint enables producers to utilize smart contracts to release NFT tokens mapping semantic information which is stored in off-chain storage. Transfer can shift the ownership of NFT tokens from producers to consumers, which means the exchange of semantic information. Burn destroys NFT tokens on the blockchain, which excludes semantic information from circulation between producers and consumers. * • Step 2: Publish Semantic information. Due to the limited storage of blockchain, semantic information is first published in off-chain storage by producers via IPFS clients. Producers obtain metadata to access the semantic information via Distributed Hash Table. Metadata can be considered as the hash commitment of semantic information to prevent it from tampering. * • Step 3: Mint Semantic information into NFT. Producers can bind metadata of semantic information, mint them into NFT tokens, and generate token URIs by smart contracts and blockchain clients. Besides metadata, NFT tokens should contain attributes to introduce types of semantic tasks, producers, and time, which can be transmitted in JSON format. * • Step 4: Transfer NFT-mapped Semantic Information. Producers can send a transferFrom transaction to NFT smart contract to assign NFT tokens to consumers, which can exchange semantic information stored in off-chain storage. When consumers own NFT tokens mapped semantic information, they can transfer it to other consumers to make a profit. NFT tokens can be observed to track how the ownership of semantic information changes over time. Besides, producers can send approved transactions to authorize trading marketplaces as operators, and operators can substitute for owners to transfer semantic information from producers to consumers. ### III-B Efficient Stackelberg Game-based Semantic Trading Mechanism Although the proposed NFT-based semantic exchange scheme can circulate semantic information between producers and consumers, it is difficult for both parties to maximize the revenue and utility of semantic information. The fixed pricing strategies may be unfair for producers when prices are lower than cost, while the unlimited pricing strategies may be disadvantaged for consumers when prices are higher than utilities. Thus, we propose a Stackelberg game-based semantic trading mechanism to provide dynamic marketplaces to trade semantic information efficiently. The Stackelberg game approach considers the realistic utility and revenue functions of producers and consumers to obtain the optimal buying and pricing strategies. The revenue function of producers can be defined as income by providing semantic information to consumers minus their costs for computing and extracting semantic information. The income can be obtained from the required amount multiplied by the unit pricing, while the costs should consider the product of the number of CPU cycles, the CPU’s frequency, the energy consumption and the network cost. The utility function of consumers can be given that the precision of received semantic information minuses the computing latency, and minuses the required amount multiplied by the unit pricing. The first-order derivative of the utility function should be greater than zero to incentive consumers to buy semantic information from producers. The above functions are modeled as a Stackelberg game and analyzed the existence and uniqueness of Nash equilibrium. In the first stage, Karush-Kuhn- Tucker (KKT) conditions are utilized to help consumers determine their required amount of semantic information. The optimal solution is reported to producers for consideration. In the second stage, producers maximize their revenue given buying strategies. ### III-C Fair Zero-Knowledge Proof-based Semantic Sharing Mechanism Although tokenizing semantic information can help producers to transact with consumers, it is difficult to preserve the privacy of semantic information. The reason is that blockchain is a public platform where everyone can submit and retrieve semantic information mapped in NFTs. Moreover, since semantic information can be easily duplicated, it is unfair for producers to release it before getting payments, while it is also disadvantageous for consumers to pay before verifying it. Therefore, we propose a fair and privacy-preserving zero- knowledge proof-based semantic sharing mechanism to enable fair authentic semantic sharing. Instead of directly transmitting semantic information once extracted from encoders, producers should transform or process it to protect fairness between both parties. Combining with the above mechanisms, the workflow of the zero-knowledge proof-based semantic sharing mechanism is described as follows. There are three stages corresponding to the three mechanisms. (a) Communication Overhead (b) Revenue of Semantic Trading Figure 4: An Illustration of Efficiency of The Proposed Framework Stage I: Transform semantic information with zero-knowledge proof (ZKP). Producers take a security parameter and a transform ZKP circuit as input, and output a common reference string for setup. The transform ZKP circuit carries on the logic of transformation in a private way to generate proofs of authentic information. Then they take the common reference string, and semantic information, and transformed semantic information as inputs to generate proofs. Stage II: Publishe and mint NFTs for trading given transformed semantic information. Producers bind metadata and proofs of transformed semantic information into NFT. The transformed semantic information is not only mapped by metadata but also represents the source semantic information by proofs. Stage III: Implement transactions between producers and consumers. Producers and consumers configure buying and pricing strategies according to Stackelberg game-based semantic trading. After consumers pay for NFT mapped by transformed semantic information, the NFT is transferred from producers to consumers. Consumers can input proofs, transformed semantic information, and the common string to verify the authenticity of transformed semantic information without revealing the semantic information. ## IV Case Study for Urban Planning in Web 3.0 Urban planning activities are aided significantly by leveraging the proposed framework, which can incorporate the benefits of blockchain and semantic communication to construct Web 3.0 infrastructures. Semantic communication can reduce information overloaded to support real-time semantic interactions and minimize energy consumption for sustainable blockchain implementations. Blockchain can provide traceable semantic information status to enable semantic sharing in untrusted environments. Urban planning requires a tremendous amount of data from other counterparts to support interactions from different locations, as shown in Fig. 3. Producers can extract the contour from images of a certain category of construction, and the corresponding snapshots (semantic information) can be transmitted to respective consumers [11]. Instead of transmitting original images, the extracted contour can reflect key objects and the appearance of certain urban areas to reduce information overloaded. Once urban designers obtain the contour of certain areas from other designers, they can design a panoramic for urban planning. Table I: An Overview of Open Challenges in Blockchain and Semantic Communication-Enabled Web 3.0 Type | Layers | Open Problem ---|---|--- Blockchain-Enabled Web 3.0 | On-Chain | Storage Structure Design, Consensus Protocol Design, Regulation Solutions, Quantum-Resistance Protocol Design Off-Chain | Cross-Chain Solutions, Oracle Protocol Design, Rollup Protocol Design, Zero-Knowledge Proof Layer 2 Solutions Collaboration | Decentralized Identity, Decentralized Finance, Metaverse, Decentralized Autonomous Organization Semantic Communication -Enabled Web 3.0 | Theory | Explainability, Impact of Transmission Rate, Capacity of Semantic Channel [12], Quantity of Semantic Information [13] Metric | Importance of Semantic Information, Tradeoff Between Accuracy and Communication Overhead [8], Inconsistent Background Knowledge Between Source and Destination [13] Application | Implementation of Semantic Communication, Multi-User Collaboration, Video Transmission, Personalization Blockchain and Semantic -Enabled Web 3.0 | Architecture | Unified, Secure, Efficient, Decentralized, Fault-Tolerant, Verifiable, Interoperable, Regulated Collaboration | Verification and Price Bubbles of Semantic Information, Semantic Interoperability, Task-Oriented Layer 2 Solutions Application | Metaverse, Digital Twin, Health Services, Digital Asset Management [5], Automotive Industry However, information circulated in the urban planning activities can be duplicated easily, which makes that other urban planning designers are not willing to exchange it without any benefit and profit. For the same reason, it is unfair for producers to release semantic information first, which makes other urban planning designers can pay nothing to receive the core contour of certain areas. Moreover, since semantic information is valuable, it is necessary to construct a trading mechanism to guide semantic exchange in urban planning activities. The proposed framework is based on the above two infrastructures to build fair and efficient semantic exchange mechanisms, as shown in Fig. 3. The designed zero-knowledge proof-based semantic sharing mechanism can implement fair and efficient transactions among urban designers by downsizing semantic information to exchange the contour of semantic information without releasing core designs first. It utilizes the NFT-based semantic exchange scheme to facilitate downsized semantic information produced from spatial data to circulation in urban planning activities. The proposed Stackelberg game-based semantic trading mechanism can facilitate urban planning activities constructing efficient trading marketplaces to facilitate the circulation of semantic information. To evaluate the proposed framework and mechanisms, we implement the simulation case to illustrate the efficiency, as exhibited in Fig. 4. Fig. 4(a) shows the communication overhead among tradition communication (Tradition), blockchain- based semantic communication (Unified) [5], and our proposed mechanism (Ours). The blockchain module is supported by a practical byzantine fault tolerance algorithm with four nodes. The semantic module is implemented by Pytorch. The size of the transmitted messages among those mechanisms is 1 MB, 100 KB, and 1 KB, referring to Fig. 3. The amount of semantic information represents exchanged number of messages. The communication overhead is the time that semantic information is recorded in the blockchain module. As shown in Fig. 4(a), the proposed mechanism is more efficient than the compared mechanisms. Fig. 4(b) illustrates how the revenue of producers is affected by different parameters of the semantic trading mechanism as the number of iterations increases. With the increase of parameters, the costs take up more weight in the revenue function, which causes a decrease in the income of producers. ## V Open Challenges and Future Directions We overview recent open challenges in terms of blockchain, semantic communication, and Web 3.0, as shown in Table I. Then we also elaborate on several of them in the following. Verification of Semantic Information: In scenarios that require high privacy and security, e.g., automotive industries, verification of semantic information is critical to perform operations correctly. It is impossible to publish semantic information on public Web 3.0 platforms, and leverage semantic information without any verification. Thus, zero-knowledge proof combined with pragmatic functions of semantic communication is expected to share private and authentic semantic information to protect privacy and security in Web 3.0. Besides, since proofs generated by ZKP can be verified given mathematical relationships, reputation mechanisms can be utilized to classify and exclude malicious Web 3.0 participants. Price Bubbles of NFT-mapped Semantic Information: Although semantic information can be considered as NFT-based Web 3.0 assets circulating in the marketplaces, it is difficult to avoid price bubbles in current market mechanisms, like OpenSea [14]. The English Dutch auction mechanism [15] is promising to eliminate price bubbles in the circulation of semantic information. Semantic Interoperability: Since Web 3.0 is integrated with blockchain and semantic communication, semantic interoperability not only includes semantic information circulated among multiple blockchains but also interacted with multiple semantic tasks. The former can be implemented by cross-chain technologies combined with verification, while the latter should consider how to extract the same characteristics between different types of semantic tasks. Task-Oriented Layer 2 Solutions: Participants have to continuously exchange semantic information after perceiving environments to improve communication quality, which is time and resource-consuming. Current Layer 2 solutions may be a practical method to improve the performance of blockchain and semantic communication-based Web 3.0, which can provide off-chain interactions in a decentralized and traceable way. ## VI Conclusion In this paper, we present our understanding and research gaps in Web 3.0, Blockchain, and Semantic Communication. To this end, we propose a fair and efficient blockchain-based semantic exchange framework for the participatory economy. In this framework, we utilize NFT to circulate semantic information in the blockchain to construct value networks. After that, we consider optimal buying and pricing strategies for counterparts transacting with semantic information by a Stackelberg game approach. We also take fairness into account and leverage the zero-knowledge proof to enable authentic semantic sharing. A case study of urban planning and simulation results are implemented to illustrate the efficiency of the proposed framework. Finally, key challenges, opportunities, and future research related to blockchain and semantic communication for Web 3.0 are discussed. ## References * [1] T. Berners-Lee, J. Hendler, and O. Lassila, “The semantic web,” _Scientific american_ , vol. 284, no. 5, pp. 34–43, 2001. * [2] G. Wood, “Đapps: What web 3.0 looks like,” Apr 2014. [Online]. Available: http://gavwood.com/dappsweb3.html * [3] S. Cousaert, J. Xu, and T. Matsui, “Sok: Yield aggregators in defi,” in _2022 IEEE International Conference on Blockchain and Cryptocurrency (ICBC)_. IEEE, 2022, pp. 1–14. * [4] C. Zhao, Y. Lv, J. Jin, Y. Tian, J. Wang, and F.-Y. Wang, “Decast in transverse for parallel intelligent transportation systems and smart cities: Three decades and beyonds,” _IEEE Intelligent Transportation Systems Magazine_ , 2022. * [5] Y. Lin, Z. Gao, H. Du, D. Niyato, J. Kang, R. Deng, and X. S. Shen, “A unified blockchain-semantic framework for wireless edge intelligence enabled web 3.0,” _arXiv preprint arXiv:2210.15130_ , 2022. * [6] Z. Li, W. Wang, Q. Wu, and X. Wang, “Multi-operator dynamic spectrum sharing for wireless communications: A consortium blockchain enabled framework,” _IEEE Transactions on Cognitive Communications and Networking_ , 2022. * [7] R. Song, S. Gao, Y. Song, and B. Xiao, “Zkdet: A traceable and privacy-preserving data exchange scheme based on non-fungible token and zero-knowledge,” in _2022 IEEE 42nd International Conference on Distributed Computing Systems (ICDCS)_. IEEE, 2022, pp. 224–234. * [8] W. Yang, H. Du, Z. Liew, W. Y. B. Lim, Z. Xiong, D. Niyato, X. Chi, X. S. Shen, and C. Miao, “Semantic communications for 6g future internet: Fundamentals, applications, and challenges,” _arXiv preprint arXiv:2207.00427_ , 2022. * [9] W. Ding, J. Hou, J. Li, C. Guo, J. Qin, R. Kozma, and F.-Y. Wang, “Desci based on web3 and dao: A comprehensive overview and reference model,” _IEEE Transactions on Computational Social Systems_ , vol. 9, no. 5, pp. 1563–1573, 2022\. * [10] Z. Liu, Y. Xiang, J. Shi, P. Gao, H. Wang, X. Xiao, B. Wen, Q. Li, and Y.-C. Hu, “Make web3. 0 connected,” _IEEE Transactions on Dependable and Secure Computing_ , 2021. * [11] W. C. Ng, H. Du, W. Y. B. Lim, Z. Xiong, D. Niyato, and C. Miao, “Stochastic resource allocation for semantic communication-aided virtual transportation networks in the metaverse,” _arXiv preprint arXiv:2208.14661_ , 2022. * [12] G. Shi, Y. Xiao, Y. Li, and X. Xie, “From semantic communication to semantic-aware networking: Model, architecture, and open problems,” _IEEE Communications Magazine_ , vol. 59, no. 8, pp. 44–50, 2021. * [13] X. Luo, H.-H. Chen, and Q. Guo, “Semantic communications: Overview, open issues, and future research directions,” _IEEE Wireless Communications_ , 2022. * [14] B. White, A. Mahanti, and K. Passi, “Characterizing the opensea nft marketplace,” in _Companion Proceedings of the Web Conference 2022_ , 2022, pp. 488–496. * [15] C. Deck, M. Servátka, and S. Tucker, “Designing call auction institutions to eliminate price bubbles: Is english dutch the best?” _American Economic Review: Insights_ , vol. 2, no. 2, pp. 225–36, 2020. BIOGRAPHIES Yijing Lin is currently pursuing the Ph.D degree at the State Key Laboratory of Networking and Switching Technology, Beijing University of Posts and Telecommunications, China (e-mail: yjlin@bupt.edu.cn). Zhipeng Gao is the corresponding author of this paper. He is a professor with the State Key Laboratory of Networking and Switching Technology, Beijing University of Posts and Telecommunications, China (e-mail: gaozhipeng@bupt.edu.cn). Yaofeng Tu is the Director of Center Institute of ZTE, China (e-mail: tu.yaofeng@zte.com.cn). Hongyang Du is currently pursuing the Ph.D degree at the School of Computer Science and Engineering, Nanyang Technological University, Singapore (e-mail: hongyang001@e.ntu.edu.sg). Dusit Niyato is a professor with the School of Computer Science and Engineering, Nanyang Technological University, Singapore (e-mail: dniyato@ntu.edu.sg). Jiawen Kang is a professor with the School of Automation, Guangdong University of Technology, China (e-mail: kjwx886@163.com). Hui Yang is a professor with the State Key Laboratory of information Photonics and Optical Communications, Beijing University of Posts and Telecommunications, China (e-mail: yanghui@bupt.edu.cn).
# Geoclidean: Few-Shot Generalization in Euclidean Geometry Joy Hsu Computer Science Stanford University <EMAIL_ADDRESS> &Jiajun Wu Computer Science Stanford University <EMAIL_ADDRESS> &Noah D. Goodman Psychology and Computer Science Stanford University <EMAIL_ADDRESS> ###### Abstract Euclidean geometry is among the earliest forms of mathematical thinking. While the geometric primitives underlying its constructions, such as perfect lines and circles, do not often occur in the natural world, humans rarely struggle to perceive and reason with them. Will computer vision models trained on natural images show the same sensitivity to Euclidean geometry? Here we explore these questions by studying few-shot generalization in the universe of Euclidean geometry constructions. We introduce _Geoclidean_ , a domain- specific language for Euclidean geometry, and use it to generate two datasets of geometric concept learning tasks for benchmarking generalization judgements of humans and machines. We find that humans are indeed sensitive to Euclidean geometry and generalize strongly from a few visual examples of a geometric concept. In contrast, low-level and high-level visual features from standard computer vision models pretrained on natural images do not support correct generalization. Thus Geoclidean represents a novel few-shot generalization benchmark for geometric concept learning, where the performance of humans and of AI models diverge. The Geoclidean framework and dataset are publicly available for download.***The Geoclidean framework can be found at https://github.com/joyhsu0504/geoclidean_framework. †††Datasets can be found at https://downloads.cs.stanford.edu/viscam/Geoclidean/geoclidean.zip. ## 1 Introduction The built world we inhabit is constructed from geometric principles. Yet geometric primitives such as perfect lines and circles, which are the foundations of human-made creations, are uncommon in the natural world. Whether for efficiency or for visual aesthetics, whether innate or learned, humans are sensitive to geometric forms and relations. This natural understanding of geometry enables a plethora of applied skills such as design, construction, and visual reasoning; it also scaffolds the development of rigorous mathematical thinking, historically and in modern education. Thus, understanding the visually-grounded geometric universe is an important desideratum for machine vision systems. Ancient Greek philosophers were amongst the earliest to formalize geometric notions, culminating in Euclid’s geometry in the 4th century BC. With a compass and straight edge, Euclid’s axioms can construct a geometric world that reflects an idealized, or Platonic, physical reality. We hypothesize that Euclidean constructions are intrinsic to human visual reasoning. We thus build a library to define and render such concepts, allowing systematic exploration of geometric generalization. In this paper, we present Geoclidean, a domain- specific language (DSL) for describing Euclidean primitives and construction rules. Sets of construction rules define concepts that can then be realized into infinitely many rendered images capturing the same abstract geometric model. Figure 1: Rendered realizations of Euclidean geometry concepts from the Geoclidean datasets. Based on Geoclidean, we introduce two datasets to study few-shot generalization to novel rendered realizations of Euclidean geometry concepts (See Figure 1). To succeed in solving these tasks, one must understand the underlying geometric concept of a set of rendered images. The first dataset, Geoclidean-Elements, covers mathematical definitions from the first book of Euclid’s Elements (Simson et al., 1838). The second dataset, Geoclidean- Constraints, simplifies and more systematically explores possible relationships between primitives. We publicly release both datasets, as well as the dataset generation library based on Geoclidean. We report findings on the Geoclidean few-shot generalization tasks from human experiments, as well as from evaluation on low-level and high-level visual features from ImageNet-pretrained VGG16 (Deng et al., 2009; Simonyan and Zisserman, 2014), ResNet50 (He et al., 2016), InceptionV3 (Szegedy et al., 2016), and Vision Transformer (Dosovitskiy et al., 2020). We show that humans significantly outperform pretrained vision models, generalizing in a way that is highly consistent with abstract target concepts. ImageNet-pretrained vision models are not as sensitive to intrinsic geometry and do not as effectively encode these geometric abstractions. Our benchmarking process illustrates this gap between humans and models, establishing Geoclidean as an interesting and challenging generalization task for visual representation and geometric concept learning. ## 2 Foundations of Geoclidean In this section, we give an overview of the Geoclidean DSL, which builds on foundations of Euclid’s axioms for constructing with a compass and straightedge. We present a Python library that renders Euclidean geometry concepts described from our DSL into images. We first describe the Euclidean geometry universe in Section 2.1, and then introduce the Geoclidean language and construction rules in Section 2.2. Finally, we show geometric concept realizations into rendered images in Section 2.3. ### 2.1 Euclidean Geometry Universe There exist numerous systems of geometry, each with its own logical system. Each logic can be described in a formal language, each formal language describes construction rules, and each set of construction rules describe concepts in the geometric universe. Geometric concepts can be realized and rendered into images. Euclidean geometry was among the earliest formalized, first described by Euclid in Elements (Simson et al., 1838). The first book of Elements details plane geometry, laying the foundation for basic properties and propositions of geometric objects. Importantly, Euclid’s geometry is constructive—objects only exist if they can be created with a compass and straightedge. Euclidean constructions are abstract models of geometric objects, specified without the use of coordinates and without concrete realizations into images. Hence, changes to the image rendering, such as alterations to size and rotation, do not change an object’s intrinsic Euclidean geometry. Euclid’s axioms build the foundations of this geometric universe via 1) draw a straight line from any point to any point, 2) describe a circle with any centre and distance. The Euclidean geometry universe is determined by these base geometric primitives, and the constraints that parameterize object relationships. There are other universes that contain different logic systems, from Descartes’ analytic geometry, to Euler’s affine geometry, to Einstein’s special relativity. Because of its simplicity and abstraction we are particularly interested in the universe of Euclidean geometry – do humans and machines find these constructions natural? We later explore whether they spontaneously generalize image classes according to Euclidean rules. ### 2.2 Domain-Specific Language: Geoclidean We create a domain-specific language (DSL), Geoclidean, defining Euclidean constructions that arise from the mathematics of Euclidean geometry. Geoclidean allows us to define construction rules for objects and their relations to each other, encompassing concepts in the Euclidean geometry universe. It includes three simple primitives. The first is a point, parameterized by constraints, if any, to an object or a set of objects previously defined. The point is defined without specific coordinates, and only when realized, would be assigned coordinate values $x$ and $y$. The second is a line, which, following the first axiom, is parameterized by two points, representing the beginning and end. The third is a circle which, following the second axiom, is defined by a center point and an edge point. These primitives represent the compass and straightedge constructions that Euclid introduced. Lines and circles are defined by points, while points can be constrained to previously built objects. In this way primitives are sequentially defined to form a concept. Table 1: The Geoclidean DSL for building Euclidean geometry concepts. We assume a pool of variable names for points and objects (point_name, object_name). As a shorthand for point creation followed by reference, we later use e.g. Line(p1(),p2()) to represent p1 = Point(); … Line(p1,p2). The marker * indicates that the object will not be visible in the final rendering. Concept | $\rightarrow$ | Statement; Concept ---|---|--- Statement | $\rightarrow$ | object_name${}^{\text{Visibility}}$ = Object(point_name, point_name) | | | point_name = point(Constraints) Visibility | $\rightarrow$ | [] | * Object | $\rightarrow$ | line | circle Constraints | $\rightarrow$ | [] | [object_name] | [object_name, object_name] Geoclidean’s syntax is defined in Table 1. We can initialize a new point, line, or circle via the constructors Point, Line, Circle, assigning them to a named variable. For points we generally use a shorthand to define and use the variable inline. For example, p1() is a free point that is unconstrained to any objects, and can be realized anywhere in the image, p2(circle1) is a partially constrained point that lies on the object circle1, and p3(line1, line2) is a fully constrained point that lives in the intersection of line1 and line2. A point can be reused by referring to its name. The line Line(p1, p2) is parameterized by two end points, and the circle Circle(p1, p2) is parameterized by points at the center and edge. Not indicated in Table 1 is the semantic constraint that variable names must be defined before being used within constructors. Visibility of rendering is denoted for each object, with * indicating that the object is not visible in the final rendering, as some objects in Euclidean geometry are used solely as helper constructions for other core objects in the concept. Each construction rule is a geometric object, and a sequence of construction rules define a Euclidean geometry concept. ### 2.3 Concept Realization Geoclidean implements this Euclidean language and renders realizations of concepts from the language into images. For each line or circle object, we realize its rendering based on the point parameters required. During rendering, each point parameter is given randomly sampled real-valued coordinates bound by its constraints. If the point exists already, Geoclidean reuses the past component; if it is a free point without constraints, Geoclidean randomly samples values for $x$ and $y$; if constrained, Geoclidean randomly samples a point that lies constrained on the object or the intersection of a set of objects. It is possible for intersection sets to be empty, and we reject sample realizations until finding a satisfying realization for all points. Geoclidean creates the objects sequentially and renders them into an image, if visible. Algorithm 1 Construction rules for the equilateral triangle concept, with colored steps corresponding to the rendered realization in Figure 2. The two circles c1 and c2 are not rendered in the final image as denoted by *. 1: 2:l1 = Line(p1(), p2()) $\triangleright$ Red line 3:c1*= Circle(p1(), p2()) $\triangleright$ Light orange circle 4:c2*= Circle(p2(), p1()) $\triangleright$ Light yellow circle 5:l2 = Line(p1(), p3(c1, c2)) $\triangleright$ Green line 6:l3 = Line(p2(), p3(c1, c2)) $\triangleright$ Blue line Figure 2: Rendered realization of the equilateral triangle concept. We see in Figure 2 an example of the Geoclidean language describing the equilateral triangle concept realized into an image. The construction rules in Algorithm 1 are created step by step; we color each object for clarity. The first construction rule creates a red line between two unconstrained free points, p1() and p2(); when realized with sampled real-valued points, this line can be anywhere in the image with any length and rotation. The second rule creates an invisible orange circle with the ends of the first line as its center and edge point. The third rule creates another invisible yellow circle with the center and edge point flipped, forming intersecting helper circles with the same radius. Then, the fourth rule creates a green line between p1 and a new point p3(c1, c2), which is a point constrained to the intersection of the previously created orange and yellow circles (the realization randomly chooses one of the two intersection points). The last rule creates the final blue line between p2 and p3, completing the constructed triangle and enforcing all sides to be of equal length. This concept is that of the equilateral triangle, and we see that the invisible helper circle objects serve as essential constraints to the final rendering. Importantly, the construction rules do not specify any coordinates, and our Geoclidean framework creates the coordinates upon realization of the concept into an image. Hence, the rendered equilateral triangle can be of any size and orientation, while always respecting the underlying geometric concept. In Geoclidean, every random realization of this concept creates equilateral triangles, as represented in Euclid’s geometric universe. The Geoclidean framework allows us to create rendered image datasets that follow the specified concept language. ## 3 Geoclidean Task and Datasets Do the realizations of Euclidean constructions form natural categories for humans? For computer vision models? To study these questions we introduce a few-shot generalization task based on Geoclidean and two image datasets that realize $37$ Euclidean geometry concepts. #### Task. We explore few-shot generalization from positive examples of a target concept; to the extent that participants generalize to other realizations of the target concept, and not realizations of more general concepts, we conclude the concept is an intuitive kind. For each concept, the task includes five reference examples of the target concept and a test set of 15 images. Among the 15 images, there are five positive examples and two sets of five negative examples. Here, positive examples in the test set are realizations of the target concept, as are the reference examples, and negative examples are realizations of related but different concepts (which are not realizations of the target concept). The goal is to correctly categorize positive examples as positive and negative examples as negative in the test set. The ten negative examples are divided into five Close and five Far examples, where the negative examples in Close are from a closely related concept with a fewer number of constraint differences from the target concept, and negative examples in Far are from a further, less related concept with a larger number of constraint differences. These constraint differences consist of altering a point to have fewer constraints compared to that point in the target concept (yielding a more general and less specified geometric concept). See Figure 3 for examples, with the top representing reference examples, and the bottom representing the test set. Note that, because we are interested in intrinsic sensitivity of visual representations to geometric concepts, we are not introducing a meta-learning task: there are no few-shot generalization sets intended for model training. We now introduce the two datasets we created based on Geoclidean for the generalization task. The first dataset, Geoclidean-Elements, includes the tests of $17$ concepts derived from the first book of Euclid’s Elements; the second dataset, Geoclidean-Constraints, includes the tests of $20$ concepts based on constraints defining relationships between Geoclidean primitives. See Figure 3 for examples from both splits. Figure 3: Examples of Geoclidean-Elements and Geoclidean-Constraints tasks. Each task consists of few-shot reference examples as well as test examples. Few-shot examples and positive test examples derive from the same concept (contained in red boxes), while negative test examples derive from a related but different concept (contained in blue boxes); Close test examples differ by fewer point differences in the construction rules than Far, seen bolded in the Geoclidean language. #### Geoclidean-Elements. The Geoclidean-Elements dataset is derived from definitions in the first book of Euclid’s Elements, which focuses on plane geometry. Geoclidean-Elements includes $17$ target concepts, which, along with Geoclidean primitives, covers definitions in the first book of Elements. These concepts require complex construction rules with helper objects that are not visible in the final renderings. Realizations from Geoclidean-Elements test sensitivity to the exactness of Euclidean constructions without explicit visual constraint differences. The concepts in Geoclidean-Elements include angle (Book I definition IX), perpendicular bisector (def X), angle bisector (def X), sixty degree angle (def XI and XII), radii (def XV), diameter (def XVII and XVIII), segment (def XIX), rectilinear (def XX and XXIII), triangle (def XXI), quadrilateral (def XXII and XXXIV), equilateral triangle (def XXIV, XXV, XXVI in Close and Far), right angled triangle (def XXVII, XXVIII, XXIX in Close and Far), square (def XXX), rhombus (def XXXI), oblong (def XXXII), rhomboid (def XXXIII), and parallel lines (def XXXV). The rest of the definitions are descriptions of Geoclidean primitives (e.g. points, lines). #### Geoclidean-Constraints. The Geoclidean-Constraints dataset consists of $20$ concepts, created from permutations of line and circle construction rules with various constraints describing the relationship between objects. This dataset focuses on explicit constraints between geometric objects. We denote the objects as the following—lines as L, circles as C, and triangles (constructed from three lines) as T. Tasks include three, four, and five object variants, each with specific ordering; the different ordering of objects is significant, as constraints may only depend on previously defined objects. Each concept is defined by object ordering as well as constraints describing the relationships between them; the full set of construction rules for each concept is released with the dataset. The three object concepts are [LLL, CLL, LLC, CCL, LCC, CCC], the four object concepts are [LLLL, LLLC, CLLL, CLCL, LLCC, CCCL, CLCC, CCCC], and the five object concepts are [TLL, LLT, TCL, CLT, TCC, CTT]. These $20$ concepts test the few-shot generalization capability in constrained Euclidean geometry concepts. ## 4 Findings We present our findings on the Geoclidean dataset in benchmarking human performance (Section 4.1) and pretrained vision models’ capabilities (Section 4.2). We show that humans are indeed sensitive to Euclidean geometry concepts, generalizing strongly from five examples across the $37$ concepts. This establishes Geoclidean as an interesting task for evaluating the human-like visual competencies of machine vision. Indeed, we find that state-of-the-art pretrained visual representations perform poorly on this few-shot generalization task. ### 4.1 Human Performance We collected human judgements for the Geoclidean few-shot concept learning task. We recruited $30$ participants for each concept using the Prolific crowd-sourcing platform (Palan and Schitter, 2018). As mentioned above, participants are given five example realizations of the target concept and $15$ test images including five positive examples, five Close negative examples, and five Far negative examples. Each of the test questions states: “Each of the five images shown is a ‘wug’. Is the image below a ‘wug’ or not a ‘wug’?”, where ‘wug’ is a random made-up word for each task. The order of tasks and realizations is randomized for each participant to remove order effects. The experiment interface was implemented on Qualtrics, with details described in the Appendix. Figure 4: Histogram of human accuracies on the Geoclidean-Elements and the Geoclidean-Constraints datasets, as well as accuracies with Close and Far negative examples. The y-axis indicates the percentage of tasks with the specified accuracy on the x-axis. We report task accuracy in Table 2, scored as the percentage of participants correctly categorizing test images, averaged across all examples. We split the $15$ test images into two tasks, with the Close task consisting of $5$ positive examples and $5$ negative examples from Close, and the Far task consisting of the same $5$ positive images with $5$ negative examples from Far. Each task contains $10$ examples in total. We see that human performance is strong across all tasks, with on-average higher scores in Far compared to Close, showing that the number of differences in construction rules affects the semantic distance between rendered realizations. Only tasks LLL, CLCL, and rhomboid yielded slightly better performance in Close than Far. Out of all tasks, LLC, CCC, LLLC with Close negatives are more difficult for humans (with equally poor performance across all test images), which we hypothesize is due to more subtle constraint intersections. In general, humans perform well on this generalization task. We show accuracy histograms in Figure 4, with the left two plots depicting results from concepts in Geoclidean-Elements and Geoclidean-Constraints, and the right two plots depicting results when calculated with Close negative examples and Far negative examples. Humans are more sensitive to concepts in Geoclidean-Elements, which are complex constructions that test the exactness of shapes (e.g., squares and equilateral triangles), and slightly less sensitive to concepts in Geoclidean-Constraints, which test the precise relationships between objects (e.g., constrained contact point between the end of a line and the center of a circle). Participants could generate infinitely many rules consistent with the positive images seen in the few-shot examples (e.g., a “wug” can have its own prototype for each example, as there are no negative examples), and there is potential ambiguity as to which are the correct construction rules of the concept. Despite this wide range of possible generalization patterns, the generalization rule chosen by humans corresponds well to the Euclidean construction universe. Table 2: Human accuracy across all $74$ tasks in Geoclidean. concept | Close | Far | concept | Close | Far ---|---|---|---|---|--- angle | 0.9767 | 0.9833 | lll | 0.9700 | 0.9667 perp bisector | 0.9367 | 0.9833 | cll | 0.9467 | 0.9667 ang bisector | 0.9433 | 0.9533 | llc | 0.6767 | 0.9233 sixty ang | 0.8233 | 0.9533 | ccl | 0.8700 | 0.8833 radii | 0.9233 | 0.9600 | lcc | 0.8867 | 0.9633 diameter | 0.9567 | 1.0000 | ccc | 0.6667 | 0.8767 segment | 0.9300 | 0.9833 | llll | 0.8833 | 0.9767 rectilinear | 0.9000 | 0.9033 | lllc | 0.6667 | 0.8867 triangle | 0.9633 | 0.9767 | clll | 0.8367 | 0.9033 quadrilateral | 0.9167 | 0.9267 | clcl | 0.8700 | 0.8567 eq t | 0.9533 | 0.9800 | llcc | 0.8867 | 0.9333 right ang t | 0.7200 | 0.8133 | cccl | 0.9233 | 0.9333 square | 0.8933 | 0.9867 | clcc | 0.8633 | 0.9000 rhombus | 0.9367 | 0.9667 | cccc | 0.8167 | 0.8800 oblong | 0.9666 | 0.9900 | tll | 0.9467 | 0.9800 rhomboid | 0.9700 | 0.9300 | llt | 0.9267 | 0.9400 parallel l | 0.9500 | 0.9567 | tcl | 0.9533 | 0.9633 | | | clt | 0.9533 | 0.9633 | | | tcc | 0.9533 | 0.9633 | | | cct | 0.9533 | 0.9633 ### 4.2 Model Benchmarks Figure 5: Histogram of human and maximum low-level and high-level feature accuracies of various vision models. We benchmark pretrained vision models’ performance on the Geoclidean task, to evaluate few-shot generalization (with no meta-learning or fine-tuning) in the Euclidean geometry universe. We measure performance of features from ImageNet- pretrained VGG16 (Simonyan and Zisserman, 2014), ResNet50 (He et al., 2016), InceptionV3 (Szegedy et al., 2016), and Vision Transformer (Dosovitskiy et al., 2020), and evaluate both low-level features and high-level features for each of the models. Low-level features are outputs of earlier layers in neural networks, which tend to capture low-level information such as edges and primitive shapes, while high-level features are outputs of later layers that tend to capture more high-level semantic information (Zeiler and Fergus, 2014). We detail how we define the layers for each baseline in the Appendix. To evaluate these features, we extract features $\phi$ from the few-shot $t$ reference images of the target concept, to create a prototype target feature, $T=\frac{1}{n}\sum_{i=1}^{n}\phi(t_{i})$. We classify a test image $r$ as in- concept if it is closer than a threshold to the prototype: $f|T-\phi(r)|<\theta$, where $f$ is the normalizing function between $0\sim 1$. The threshold is fit by selecting the best-performing normalized distance threshold across all 74 tasks, for given features. (By fitting this free threshold, we bias the reported accuracy in favor of models.) Table 3: Human accuracy compared to low and high-level feature accuracy of vision models across $37$ concepts in Geoclidean, each concept containing averaged accuracies between Close and Far tasks. | Human | VGG16 | | RN50 | | InV3 | | ViT | ---|---|---|---|---|---|---|---|---|--- auc | | low | high | low | high | low | high | low | high angle | 0.98 | 0.50 | 0.45 | 0.45 | 0.50 | 0.40 | 0.45 | 0.50 | 0.40 perp bisector | 0.96 | 0.70 | 0.85 | 0.70 | 0.50 | 0.65 | 0.90 | 0.70 | 0.85 ang bisector | 0.95 | 0.50 | 0.60 | 0.50 | 0.65 | 0.60 | 0.60 | 0.50 | 0.75 sixty ang | 0.89 | 0.25 | 0.65 | 0.35 | 0.45 | 0.55 | 0.45 | 0.35 | 0.70 radii | 0.94 | 0.75 | 0.70 | 0.75 | 0.60 | 0.65 | 0.80 | 0.75 | 0.75 diameter | 0.98 | 0.30 | 0.55 | 0.40 | 0.40 | 0.50 | 0.75 | 0.45 | 0.85 segment | 0.96 | 0.45 | 0.60 | 0.60 | 0.40 | 0.30 | 0.65 | 0.55 | 0.65 rectilinear | 0.90 | 0.65 | 0.35 | 0.60 | 0.45 | 0.55 | 0.60 | 0.65 | 0.45 triangle | 0.97 | 0.65 | 0.45 | 0.50 | 0.40 | 0.40 | 0.55 | 0.35 | 0.60 quadrilateral | 0.92 | 0.50 | 0.70 | 0.60 | 0.50 | 0.55 | 0.75 | 0.60 | 0.60 eq t | 0.97 | 0.40 | 0.85 | 0.50 | 0.70 | 0.65 | 0.65 | 0.50 | 0.55 right ang t | 0.77 | 0.65 | 0.75 | 0.60 | 0.70 | 0.60 | 0.55 | 0.70 | 0.55 square | 0.94 | 0.70 | 0.80 | 0.65 | 0.45 | 0.85 | 0.70 | 0.60 | 0.75 rhombus | 0.95 | 0.70 | 0.55 | 0.60 | 0.55 | 0.60 | 0.55 | 0.70 | 0.60 oblong | 0.98 | 0.45 | 0.55 | 0.45 | 0.45 | 0.50 | 0.70 | 0.45 | 0.70 rhomboid | 0.95 | 0.70 | 0.60 | 0.65 | 0.55 | 0.50 | 0.50 | 0.65 | 0.75 parallel l | 0.95 | 0.55 | 0.35 | 0.50 | 0.40 | 0.45 | 0.55 | 0.45 | 0.85 lll | 0.97 | 0.50 | 0.70 | 0.55 | 0.40 | 0.55 | 0.35 | 0.55 | 0.80 cll | 0.96 | 0.25 | 0.60 | 0.40 | 0.30 | 0.30 | 0.60 | 0.30 | 0.60 llc | 0.80 | 0.65 | 0.55 | 0.60 | 0.55 | 0.55 | 0.65 | 0.60 | 0.60 ccl | 0.88 | 0.65 | 0.50 | 0.55 | 0.65 | 0.30 | 0.60 | 0.55 | 0.75 lcc | 0.93 | 0.45 | 0.70 | 0.45 | 0.70 | 0.65 | 0.70 | 0.45 | 0.85 ccc | 0.77 | 0.60 | 0.55 | 0.60 | 0.70 | 0.70 | 0.75 | 0.60 | 0.90 llll | 0.93 | 0.60 | 0.30 | 0.50 | 0.30 | 0.30 | 0.50 | 0.50 | 0.50 lllc | 0.78 | 0.30 | 0.65 | 0.40 | 0.55 | 0.50 | 0.55 | 0.30 | 0.80 clll | 0.87 | 0.55 | 0.65 | 0.60 | 0.65 | 0.65 | 0.50 | 0.70 | 0.65 clcl | 0.86 | 0.60 | 0.65 | 0.65 | 0.45 | 0.40 | 0.35 | 0.60 | 0.85 llcc | 0.91 | 0.75 | 0.65 | 0.85 | 0.55 | 0.80 | 0.70 | 0.80 | 0.70 cccl | 0.93 | 0.35 | 0.60 | 0.50 | 0.65 | 0.35 | 0.65 | 0.45 | 0.75 clcc | 0.88 | 0.60 | 0.55 | 0.55 | 0.60 | 0.60 | 0.75 | 0.55 | 0.80 cccc | 0.85 | 0.80 | 0.65 | 0.70 | 0.60 | 0.75 | 0.50 | 0.75 | 0.95 tll | 0.96 | 0.70 | 0.65 | 0.65 | 0.60 | 0.60 | 0.55 | 0.75 | 0.85 llt | 0.93 | 0.55 | 0.40 | 0.55 | 0.65 | 0.55 | 0.65 | 0.50 | 0.85 tcl | 0.96 | 0.65 | 0.70 | 0.70 | 0.55 | 0.60 | 0.60 | 0.65 | 0.45 clt | 0.88 | 0.50 | 0.60 | 0.45 | 0.50 | 0.55 | 0.65 | 0.45 | 0.70 tcc | 0.84 | 0.30 | 0.45 | 0.35 | 0.35 | 0.55 | 0.45 | 0.30 | 0.45 cct | 0.79 | 0.55 | 0.70 | 0.45 | 0.60 | 0.60 | 0.55 | 0.50 | 0.85 average | 0.91 | 0.55 | 0.60 | 0.55 | 0.53 | 0.54 | 0.60 | 0.55 | 0.70 In Table 3, we present accuracy of ImageNet-pretrained low-level and high- level features across different models. Humans substantially outperform features from vision models, showcasing the gap between human and model capabilities in Euclidean geometry concept learning. We see that, on average across tasks, pretrained vision models perform poorly compared to humans. In general, high-level features perform slightly better than low-level features, though there are cases where this differs. Interestingly, the high-level features from the Vision Transformer (ViT) outperform its convolutional network counterparts, and achieve human-level performance in some tasks. There are a few tasks where the visual feature accuracy outperforms humans, notably tasks CCC, LLLC, CCCC, CCT, where high- level features from ViT perform strongly. We hypothesize that this is because humans are not as sensitive to details of intersections between circles. In Figure 5, we present histograms comparing maximum low and high-level feature accuracy to human accuracy, illustrating the gap in performance. In Table 4, we report the Pearson correlation coefficient between the answers of humans and models. We see that humans are generally more aligned with high- level features than low-level ones. Additionally, though ViT achieves high accuracy, it is not correlated to human performance, indicating failure to generalize in the most human-like way. Table 4: Pearson’s correlation to human accuracy across all $74$ tasks in Geoclidean. VGG16 low | high | RN50 low | high | InV3 low | high | ViT low | high ---|---|---|---|---|---|---|--- -0.08947 | 0.0943 | 0.0341 | -0.1522 | -0.0820 | 0.2575 | -0.0056 | -0.0259 We report additional comparisons of low-level and high-level visual features from ImageNet-pretrained VGG16, ResNet50, InceptionV3, and Vision Transformer in the Appendix, and show that similar trends follow across data splits and performance metrics. We also include low-level and high-level feature visualizations in the Appendix, comparing Geoclidean tasks that require reasoning, to perception tasks involving simple geometric primitives and perturbations. These comparisons highlight Geoclidean as a unique and interesting test for vision models. ## 5 Related Work #### Geometric reasoning datasets. Prior geometric datasets generally fall into two main categories—with geometric objects for computer-aided design (CAD) and for plane geometry. In the first category, the SketchGraphs dataset models relational geometry in CAD design (Seff et al., 2020), and the ABC-Dataset includes parametric representations of 3D CAD models (Koch et al., 2019). In the latter category, CSGNet presented a generated dataset of constructive solid geometry based on 2D and 3D synthetic programs with squares, circles, and triangles (Sharma et al., 2018), while Ellis et al. (2018) connected high–level Latex graphics programs with 2D geometric drawings. Works such as Zhang et al. (2022); Lu et al. (2021) proposed using datasets with annotated geometric primitives and relationships such as containment from geometry diagrams in textbooks. Others introduced reasoning benchmarks with geometric shapes, including Raven’s progressive matrices (Matzen et al., 2010; Wang and Su, 2015; Barrett et al., 2018; Zhang et al., 2019), Bongard problems (Depeweg et al., 2018; Nie et al., 2020), odd-one-out tasks (Mańdziuk and Żychowski, 2019), and a variety of reasoning challenges (Hill et al., 2019; Zhao et al., 2021; El Korchi and Ghanou, 2020; Zhang et al., 2020). Our work is more related to the latter of geometric shapes, and Geoclidean differs by targeting Euclidean geometry concept learning whose construction language 1) does not require specific coordinates, and 2) focuses on the construction steps that form semantically- complex geometric concepts and the constraints between geometric primitives that humans are intrinsically sensitive to. #### Few-shot concept learning. Few-shot learning tasks range in complexity on both the input and task description axis. In the natural language processing domain, tasks such as FewRel (Han et al., 2018) and Few-NERD (Ding et al., 2021) have been proposed for few-shot relation classification and entity recognition. Goodman et al. (2008) introduced concept learning tasks with sequences generated from specified logical rules. In the vision domain, which we are interested in, commonly used tasks include those from Lake et al. (2015), which introduced Omniglot as a collection of simple visual concepts collected from 50 writing systems, and from Vinyals et al. (2016), which proposed miniImageNet (Deng et al., 2009), both for the task of one-shot classification. Works in other vision domains include Massiceti et al. (2021), which explores the few-shot video recognition challenge, and Xiao et al. (2020); Gehler et al. (2008), which examines the few-shot color constancy problem. Triantafillou et al. (2019) created the meta-dataset as a diverse dataset for few-shot learning, with multiple tasks for meta-training such as Maji et al. (2013); Wah et al. (2011); Cimpoi et al. (2014); Nilsback and Zisserman (2008); Houben et al. (2013); Lin et al. (2014). In comparison, we propose Geoclidean as a zero-shot meta-trained, few-shot generalization task that consists of labeled image renderings from a single target concept. ## 6 Discussion An important contribution of our task is that it allows for better testing of vision models that aim to incorporate reasoning and high-level semantics. Additionally, Geoclidean’s zero-shot meta-trained evaluation is especially significant, as many downstream tasks that may leverage pretrained models would greatly benefit from geometric reasoning, such as construction (LegoTron (Walsman et al., 2022), Physical Construction Tasks Bapst et al. (2019)), physical reasoning (CLEVRER (Yi et al., 2019), ThreeDWorld (Gan et al., 2020)), and shape understanding tasks (PartNet (Mo et al., 2019), ShapeNet (Chang et al., 2015)). Furthermore, numerous additional evaluation tasks can be built with the Geoclidean DSL and rendering library, such as those involving natural language or generated large-scale datasets. We include further analyses and discussion in the Appendix. ## 7 Conclusion We have introduced Geoclidean, a domain-specific language for the realization of the Euclidean geometry universe, and presented two datasets of few-shot concept learning to test generalization capability in the geometry domain. Humans considerably outperform vision models on Geoclidean tasks, and we believe that this gap illustrates the potential for improvement in learning visual features that align with human sensitivities to geometry. Geoclidean is thus an important generalization task that vision models are not yet sensitive to, and an effective benchmark for geometric concept learning. Furthermore, such explorations of geometric generalization may help us to understand how human vision made the leap from natural forms to the Platonic forms so prevalent in modern design and engineering. We expect minimal negative societal impact from the release of Geoclidean. We hope future work can build on the foundations of Geoclidean for augmenting vision models in areas such as geometric reasoning and construction, as well as in applications such as education, where geometry is both an essential academic subject and an introduction to proof-based mathematics. #### Acknowledgements We thank Gabriel Poesia and Stephen Tian for providing valuable feedback on the paper. This work is in part supported by the Stanford Institute for Human- Centered Artificial Intelligence (HAI), Center for Integrated Facility Engineering (CIFE), Analog, Autodesk, IBM, JPMC, Salesforce, and Samsung. JH is supported by the Knight Hennessy fellowship and the NSF Graduate Research Fellowship. ## References * Bapst et al. [2019] Victor Bapst, Alvaro Sanchez-Gonzalez, Carl Doersch, Kimberly Stachenfeld, Pushmeet Kohli, Peter Battaglia, and Jessica Hamrick. Structured agents for physical construction. In _International conference on machine learning_ , pages 464–474. PMLR, 2019. * Barrett et al. [2018] David Barrett, Felix Hill, Adam Santoro, Ari Morcos, and Timothy Lillicrap. Measuring abstract reasoning in neural networks. In _International Conference on Machine Learning_ , 2018. * Chang et al. [2015] Angel X Chang, Thomas Funkhouser, Leonidas Guibas, Pat Hanrahan, Qixing Huang, Zimo Li, Silvio Savarese, Manolis Savva, Shuran Song, Hao Su, et al. Shapenet: An information-rich 3d model repository. _arXiv preprint arXiv:1512.03012_ , 2015. * Cimpoi et al. [2014] Mircea Cimpoi, Subhransu Maji, Iasonas Kokkinos, Sammy Mohamed, and Andrea Vedaldi. Describing textures in the wild. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 2014. * Deng et al. [2009] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. ImageNet: A large-scale hierarchical image database. In _IEEE Conference on Computer Vision and Pattern Recognition_ , 2009\. * Depeweg et al. [2018] Stefan Depeweg, Constantin A Rothkopf, and Frank Jäkel. Solving Bongard problems with a visual language and pragmatic reasoning. _arXiv preprint arXiv:1804.04452_ , 2018. * Ding et al. [2021] Ning Ding, Guangwei Xu, Yulin Chen, Xiaobin Wang, Xu Han, Pengjun Xie, Hai-Tao Zheng, and Zhiyuan Liu. Few-nerd: A few-shot named entity recognition dataset. _arXiv preprint arXiv:2105.07464_ , 2021. * Dosovitskiy et al. [2020] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. _arXiv preprint arXiv:2010.11929_ , 2020. * El Korchi and Ghanou [2020] Anas El Korchi and Youssef Ghanou. 2D geometric shapes dataset–for machine learning and pattern recognition. _Data in Brief_ , 32:106090, 2020. * Ellis et al. [2018] Kevin Ellis, Daniel Ritchie, Armando Solar-Lezama, and Josh Tenenbaum. Learning to infer graphics programs from hand-drawn images. In _Advances in Neural Information Processing Systems_ , 2018. * Gan et al. [2020] Chuang Gan, Jeremy Schwartz, Seth Alter, Martin Schrimpf, James Traer, Julian De Freitas, Jonas Kubilius, Abhishek Bhandwaldar, Nick Haber, Megumi Sano, et al. Threedworld: A platform for interactive multi-modal physical simulation. _arXiv preprint arXiv:2007.04954_ , 2020. * Gehler et al. [2008] Peter Vincent Gehler, Carsten Rother, Andrew Blake, Tom Minka, and Toby Sharp. Bayesian color constancy revisited. In _IEEE Conference on Computer Vision and Pattern Recognition_ , 2008\. * Goodman et al. [2008] Noah D Goodman, Joshua B Tenenbaum, Jacob Feldman, and Thomas L Griffiths. A rational analysis of rule-based concept learning. _Cognitive Science_ , 32(1):108–154, 2008. * Han et al. [2018] Xu Han, Hao Zhu, Pengfei Yu, Ziyun Wang, Yuan Yao, Zhiyuan Liu, and Maosong Sun. FewRel: A large-scale supervised few-shot relation classification dataset with state-of-the-art evaluation. _arXiv preprint arXiv:1810.10147_ , 2018. * He et al. [2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 2016. * Hill et al. [2019] Felix Hill, Adam Santoro, David GT Barrett, Ari S Morcos, and Timothy Lillicrap. Learning to make analogies by contrasting abstract relational structure. _arXiv preprint arXiv:1902.00120_ , 2019. * Houben et al. [2013] Sebastian Houben, Johannes Stallkamp, Jan Salmen, Marc Schlipsing, and Christian Igel. Detection of traffic signs in real-world images: The german traffic sign detection benchmark. In _The International Joint Conference on Neural Networks (IJCNN)_ , 2013. * Koch et al. [2019] Sebastian Koch, Albert Matveev, Zhongshi Jiang, Francis Williams, Alexey Artemov, Evgeny Burnaev, Marc Alexa, Denis Zorin, and Daniele Panozzo. ABC: A big cad model dataset for geometric deep learning. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2019. * Lake et al. [2015] Brenden M Lake, Ruslan Salakhutdinov, and Joshua B Tenenbaum. Human-level concept learning through probabilistic program induction. _Science_ , 350(6266):1332–1338, 2015. * Lin et al. [2014] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft COCO: Common objects in context. In _European Conference on Computer Vision_ , 2014. * Lu et al. [2021] Pan Lu, Ran Gong, Shibiao Jiang, Liang Qiu, Siyuan Huang, Xiaodan Liang, and Song-Chun Zhu. Inter-gps: Interpretable geometry problem solving with formal language and symbolic reasoning. _arXiv preprint arXiv:2105.04165_ , 2021. * Maji et al. [2013] Subhransu Maji, Esa Rahtu, Juho Kannala, Matthew Blaschko, and Andrea Vedaldi. Fine-grained visual classification of aircraft. _arXiv preprint arXiv:1306.5151_ , 2013. * Mańdziuk and Żychowski [2019] Jacek Mańdziuk and Adam Żychowski. DeepIQ: A human-inspired AI system for solving IQ test problems. In _International Joint Conference on Neural Networks (IJCNN)_ , 2019\. * Massiceti et al. [2021] Daniela Massiceti, Luisa Zintgraf, John Bronskill, Lida Theodorou, Matthew Tobias Harris, Edward Cutrell, Cecily Morrison, Katja Hofmann, and Simone Stumpf. ORBIT: A real-world few-shot dataset for teachable object recognition. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , 2021. * Matzen et al. [2010] Laura E Matzen, Zachary O Benz, Kevin R Dixon, Jamie Posey, James K Kroger, and Ann E Speed. Recreating Raven’s: Software for systematically generating large numbers of Raven-like matrix problems with normed properties. _Behavior Research Methods_ , 42(2):525–541, 2010\. * Mo et al. [2019] Kaichun Mo, Shilin Zhu, Angel X Chang, Li Yi, Subarna Tripathi, Leonidas J Guibas, and Hao Su. Partnet: A large-scale benchmark for fine-grained and hierarchical part-level 3d object understanding. In _Proceedings of the IEEE/CVF conference on computer vision and pattern recognition_ , pages 909–918, 2019. * Nie et al. [2020] Weili Nie, Zhiding Yu, Lei Mao, Ankit B Patel, Yuke Zhu, and Anima Anandkumar. Bongard-LOGO: A new benchmark for human-level concept learning and reasoning. In _Advances in Neural Information Processing Systems_ , 2020. * Nilsback and Zisserman [2008] Maria-Elena Nilsback and Andrew Zisserman. Automated flower classification over a large number of classes. In _Indian Conference on Computer Vision, Graphics & Image Processing_, 2008. * Palan and Schitter [2018] Stefan Palan and Christian Schitter. Prolific.ac—A subject pool for online experiments. _Journal of Behavioral and Experimental Finance_ , 17:22–27, 2018. * Seff et al. [2020] Ari Seff, Yaniv Ovadia, Wenda Zhou, and Ryan P Adams. Sketchgraphs: A large-scale dataset for modeling relational geometry in computer-aided design. _arXiv preprint arXiv:2007.08506_ , 2020. * Sharma et al. [2018] Gopal Sharma, Rishabh Goyal, Difan Liu, Evangelos Kalogerakis, and Subhransu Maji. CSGNet: Neural shape parser for constructive solid geometry. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 2018. * Simonyan and Zisserman [2014] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. _arXiv preprint arXiv:1409.1556_ , 2014. * Simson et al. [1838] Robert Simson et al. _The elements of Euclid_. Desilver, Thomas, 1838. * Szegedy et al. [2016] Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. Rethinking the inception architecture for computer vision. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 2016. * Triantafillou et al. [2019] Eleni Triantafillou, Tyler Zhu, Vincent Dumoulin, Pascal Lamblin, Utku Evci, Kelvin Xu, Ross Goroshin, Carles Gelada, Kevin Swersky, Pierre-Antoine Manzagol, et al. Meta-dataset: A dataset of datasets for learning to learn from few examples. _arXiv preprint arXiv:1903.03096_ , 2019. * Vinyals et al. [2016] Oriol Vinyals, Charles Blundell, Timothy Lillicrap, and Daan Wierstra. Matching networks for one shot learning. In _Advances in neural information processing systems_ , 2016. * Wah et al. [2011] Catherine Wah, Steve Branson, Peter Welinder, Pietro Perona, and Serge Belongie. The Caltech-UCSD Birds-200-2011 dataset. Technical report, California Institute of Technology, 2011. * Walsman et al. [2022] Aaron Walsman, Muru Zhang, Adam Fishman, Karthik Desingh, Dieter Fox, and Ali Farhadi. Break and make: Interactive structural understanding using lego bricks. In _European Conference on Computer Vision_ , 2022. * Wang and Su [2015] Ke Wang and Zhendong Su. Automatic generation of Raven’s progressive matrices. In _International Joint Conference on Artificial Intelligence_ , 2015\. * Xiao et al. [2020] Jin Xiao, Shuhang Gu, and Lei Zhang. Multi-domain learning for accurate and few-shot color constancy. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2020. * Yi et al. [2019] Kexin Yi, Chuang Gan, Yunzhu Li, Pushmeet Kohli, Jiajun Wu, Antonio Torralba, and Joshua B Tenenbaum. Clevrer: Collision events for video representation and reasoning. _arXiv preprint arXiv:1910.01442_ , 2019. * Zeiler and Fergus [2014] Matthew D Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. In _European conference on computer vision_ , pages 818–833. Springer, 2014. * Zhang et al. [2019] Chi Zhang, Feng Gao, Baoxiong Jia, Yixin Zhu, and Song-Chun Zhu. Raven: A dataset for relational and analogical visual reasoning. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2019. * Zhang et al. [2022] Ming-Liang Zhang, Fei Yin, Yi-Han Hao, and Cheng-Lin Liu. Plane geometry diagram parsing. _arXiv preprint arXiv:2205.09363_ , 2022. * Zhang et al. [2020] Wenhe Zhang, Chi Zhang, Yixin Zhu, and Song-Chun Zhu. Machine number sense: A dataset of visual arithmetic problems for abstract and relational reasoning. In _Proceedings of the AAAI Conference on Artificial Intelligence_ , 2020. * Zhao et al. [2021] Yizhou Zhao, Liang Qiu, Pan Lu, Feng Shi, Tian Han, and Song-Chun Zhu. Learning from the tangram to solve mini visual tasks. _arXiv preprint arXiv:2112.06113_ , 2021. ## Checklist 1. 1. For all authors… 1. (a) Do the main claims made in the abstract and introduction accurately reflect the paper’s contributions and scope? [Yes] 2. (b) Did you describe the limitations of your work? [Yes] , see Section 4.1. 3. (c) Did you discuss any potential negative societal impacts of your work? [Yes] , see Section 7. 4. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. 2. If you are including theoretical results… 1. (a) Did you state the full set of assumptions of all theoretical results? [N/A] 2. (b) Did you include complete proofs of all theoretical results? [N/A] 3. 3. If you ran experiments (e.g. for benchmarks)… 1. (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] , see Appendix. 2. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [N/A] There is no training involved. 3. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [N/A] There is no training involved. 4. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] , see Appendix. 4. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets… 1. (a) If your work uses existing assets, did you cite the creators? [N/A] 2. (b) Did you mention the license of the assets? [Yes] , see Appendix. 3. (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] , see Appendix. 4. (d) Did you discuss whether and how consent was obtained from people whose data you’re using/curating? [Yes] , see Appendix. 5. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [Yes] , see Appendix. 5. 5. If you used crowdsourcing or conducted research with human subjects… 1. (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [Yes] , see Appendix. 2. (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [Yes] , see Appendix. 3. (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [Yes] , see Appendix.
# 3D fictitious wave domain CSEM inversion by adjoint source estimation Pengliang Yang1 1 School of Mathematics, Harbin Institute of Technology, 150001, Harbin, China E-mail<EMAIL_ADDRESS> ###### Abstract Marine controlled-source electromagnetic (CSEM) method has proved its potential in detecting highly resistive hydrocarbon bearing formations. A novel frequency domain CSEM inversion approach using fictitious wave domain time stepping modelling is presented. Using Lagrangian-based adjoint state method, the inversion gradient with respect to resistivity can be computed by the product between the forward and adjoint fields. Simulation of the adjoint field using the same modelling engine is challenging as it requires time domain adjoint source time functions while only a few discrete frequencies of the data residual are available for the inversion. A regularized linear inverse problem is formulated in order to estimate a long time series from very few frequency samples. It can then be solved using linear optimization technique, yielding a matrix-free implementation. Instead of computing adjoint source time function one by one at each receiver location, a basis function implementation has been developed such that the inverse problem can be solved only once and reused every time to construct all time-domain adjoint sources. The method allows computing all frequencies of the EM fields in one go without heavy memory and computational overhead, making efficient 3D CSEM inversion feasible. Numerical examples are employed to demonstrate the application of our method. ## 1 Introduction Controlled-source electromagnetic (CSEM) method is a well established technology to do geophysical exploration (Chave and Cox, 1982; Constable et al., 1986). It can also be configured for air-borne (Chang-Chun et al., 2015) and cross-well (Alumbaugh and Newman, 1997) geometries. Land CSEM has been a commonplace to find mineral deposit (Ward and Hohmann, 1988; Zhdanov and Keller, 1994; Grayver et al., 2014). As a complement to seismic measurement, marine CSEM have been successfully applied to detect hydrocarbon bearing formations in oil and gas industry (Eidesmo et al., 2002; Ellingsrud et al., 2002; Constable and Srnka, 2007; MacGregor and Tomlinson, 2014), using vessel- towed dipole source over an array of receivers deployed on the seabed (hence the name seabed logging). The key to these applications is the good sensitivity of the electromagnetic signals to distinguish high contrast in resistivity between saline-filled rocks and hydrocarbons (Abubakar et al., 2008). This property makes CSEM an ideal tool for de-risking (MacGregor et al., 2007) in reservoir exploration prior to drilling, as well as 4D monitoring during the production (Shantsev et al., 2020). A topical review on marine CSEM inversion has been given by Constable (2010). CSEM inversion, also known as resistivity tomography, has been a standard technique to translate CSEM data into electrical properties of the subsurface. To determine the resistivity model, a nonlinear inverse problem is then formulated to iteratively minimize the difference between the observed EM data and the synthetic data derived from numerical modelling, using different optimization schemes, such as nonlinear conjugate gradient method (Gribenko and Zhdanov, 2007; Commer and Newman, 2008), quasi-Newton l-BFGS (Plessix and Mulder, 2008; Schwarzbach and Haber, 2013), and Gauss-Newton method (Constable et al., 1987; Abubakar et al., 2008; Zaslavsky et al., 2013). The most computation intensive part of CSEM inversion is the numerical simulation of 3D electromagnetic field. One can consider the time-domain finite-difference method (Oristaglio and Hohmann, 1984; Wang and Hohmann, 1993; Taflove and Hagness, 2005), the frequency-domain finite-difference method (Newman and Alumbaugh, 1995; Smith, 1996a; Mulder, 2006; Streich, 2009), and the frequency-domain finite-element method (Li and Key, 2007; da Silva et al., 2012; Key, 2016; Rochlitz et al., 2019). A plethora of CSEM modelling tools directly solve the linear system based on the frequency domain solution of the discretized Maxwell equation. This avoids the high computational cost by direct discretization of time-domain diffusive Maxwell equation involving extremely large number of time steps dictated by the restrictive stability condition. Both the frequency domain finite difference method and finite element method formulate Maxwell equation as a matrix-based linear system, which may be solved using direct (Streich, 2009) or iterative (Smith, 1996b; Mulder, 2006; Puzyrev et al., 2013) solvers. Modelling by direct solver is very attractive for multi-source CSEM problems (Streich, 2009), as the matrix system can be factorized only once and reused for all sources by forward-backward substitution. Since the inversion of the large sparse matrix for 3D problems involves huge amount of memory resources, direct methods may easily go beyond the memory capacity of a desktop computer. Iterative solvers require much less memory storage but may be time-consuming and difficult to converge due to ill-conditioning of the discretized Helmholtz matrix when the conductivity/resistivity model is highly heterogeneous or the modelling grid is severely stretched (Mulder, 2006). While the frequency domain finite element method is flexible to address complex model geometries thanks to the pre-computed meshes, the meshing in 3D geometries using tetrahedral and hexahedral mesh itself is a challenging and time consuming task. In the time-domain, the fields are updated at each time step, allowing the modelling over the same memory units. The frequency domain fields can be integrated on the fly during time-stepping, such that multiple frequencies can be extracted from the same simulation. This avoids repeating several times of the modelling for different frequencies in frequency domain methods. It motivates Maaø (2007) to propose a modified wave domain approach to significantly speed up the computation of diffusive electromagnetic modelling. The method has been adapted in Støren et al. (2008) to efficiently compute the inversion gradient for 3D industrial scale applications. The modified wave domain approach by Maaø (2007) has an attenuation/diffusive term. Inspired by the fictitious wave domain approach initially proposed in Lee et al. (1989), Mittet (2010) transformed the diffusion domain Maxwell equation into a pure wave domain. A straightforward time-domain discretization of the diffusive Maxwell equation leads to the temporal sampling proportional to the square of grid spacing, due to the requirement of the stability condition ($\Delta t\propto\Delta x^{2}$). This means taking half the spatial sampling will lead to a quadratic increase of the number of time steps for a simulation of the same duration. Transforming the diffusive domain into wave domain using fictitious wave domain approach allows a linear increase of time step with the use of finer grid spacing ($\Delta t\propto\Delta x$). This yields a highly efficient scheme to compute the same frequency domain EM field, as it significantly reduces the required number of time steps. These advantages inspire us to adopt the fictitious wave domain method for CSEM inversion. The major contribution of this paper is to develop an efficient frequency domain CSEM inversion scheme based on the fictitious wave domain modelling. The gradient of the misfit functional for frequency domain CSEM inversion requires both a forward and an adjoint field. Efficient computation of the frequency domain adjoint field by time domain modelling manifests itself as a major challenge in this development as it requires the time domain adjoint source time functions at all receiver locations which are unavailable due to the formulation of CSEM inversion in frequency domain. The key novelty of our approach is to estimate the time-domain adjoint source based on only a few discrete frequency samples of the data residual. To do so, we formulate a regularized linear inverse problem, and then present an efficient, matrix-free implementation using very few basis functions. The linear inverse problem can be solved only once and the resulting solution can be reused to derive the adjoint source time functions at all receiver locations. This novel development delivers an efficient 3D CSEM inversion methodology in fictitious wave domain. We finally apply our method to two examples for numerical demonstration. ## 2 CSEM modelling in fictitious wave domain Within a three dimensional space $X$, the diffusive Maxwell equations are governed in the frequency domain $\begin{cases}\nabla\times E(\mathbf{x},\omega;\mathbf{x}_{s})-\mathrm{i}\omega\mu(\mathbf{x})H(\mathbf{x},\omega;\mathbf{x}_{s})=M(\mathbf{x}_{s},\omega),\quad\mathbf{x}\in X,\omega\in\Omega\\\ \nabla\times H(\mathbf{x},\omega;\mathbf{x}_{s})-\sigma(\mathbf{x})E(\mathbf{x},\omega;\mathbf{x}_{s})=J(\mathbf{x}_{s},\omega),\quad\mathbf{x}\in X,\omega\in\Omega\end{cases},$ (1) where the variables $\mathbf{x}$, $t$ and $\omega$ denote space, time and frequency, respectively. The convention of Fourier transform $\partial_{t}\leftrightarrow-\mathrm{i}\omega$ has been adopted above. The electrical and magnetic fields ($E$ and $H$) are vectors consisting of 3 components in x, y and z directions; $J$ and $M$ stand for electrical and magnetic sources at the location $\mathbf{x}_{s}$, respectively. The magnetic permeability is $\mu$. The conductivity $\sigma$ is a symmetric $3\times 3$ tensor, i.e., $\sigma_{ij}=\sigma_{ji}$, $i,j\in\\{x,y,z\\}$. In the isotropic medium, only the diagonal elements of the conductivity tensor are non-zeros and the same in all directions: $\sigma_{xx}=\sigma_{yy}=\sigma_{zz}$; $\sigma_{ij}=0,i\neq j$. In the vertical transverse isotropic (VTI) medium, the diagonal elements are different in horizontal and vertical directions: $\sigma_{h}:=\sigma_{xx}=\sigma_{yy},\quad\sigma_{v}=\sigma_{zz}$, where $\sigma_{h}$ and $\sigma_{v}$ stand for horizontal conductivity and vertical conductivity, respectively. The resistivity is defined as the inverse of the conductivity, i.e., $\rho_{ij}=1/\sigma_{ij}$. The above diffusive Maxwell equation can be converted into wave domain (Mittet, 2010), by defining a fictitious di-electrical permittivity in equation (1) as $\sigma=2\omega_{0}\varepsilon$, while multiplying the 2nd equation in (1) with $\sqrt{-\mathrm{i}\omega/2\omega_{0}}$: $\begin{cases}\nabla\times E^{\prime}(\mathbf{x},\omega;\mathbf{x}_{s})-\mathrm{i}\omega^{\prime}\mu H^{\prime}(\mathbf{x},\omega;\mathbf{x}_{s})=M^{\prime}(\mathbf{x}_{s},\omega),\quad\mathbf{x}\in X,\omega\in\Omega\\\ \nabla\times H^{\prime}(\mathbf{x},\omega;\mathbf{x}_{s})+\mathrm{i}\omega^{\prime}\varepsilon E^{\prime}(\mathbf{x},\omega;\mathbf{x}_{s})=J^{\prime}(\mathbf{x}_{s},\omega),\quad\mathbf{x}\in X,\omega\in\Omega\end{cases}$ (2) based on the following correspondence relation $E^{\prime}=E,\;H^{\prime}=\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}H,\;M^{\prime}=M,\;J^{\prime}=\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}J,\;\omega^{\prime}=(1+\mathrm{i})\sqrt{\omega\omega_{0}}.$ (3) The time domain counterpart of equation (2) reads $\begin{cases}\nabla\times E^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})+\mu(\mathbf{x})\partial_{t^{\prime}}H^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})=M^{\prime}(\mathbf{x}_{s},t^{\prime}),\quad\mathbf{x}\in X,t^{\prime}\in T\\\ \nabla\times H^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})-\varepsilon(\mathbf{x})\partial_{t^{\prime}}E^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})=J^{\prime}(\mathbf{x}_{s},t^{\prime}),\quad\mathbf{x}\in X,t^{\prime}\in T,\end{cases}$ (4) which allows us to do efficient modelling using leap-frog finite-difference time-domain (FDTD) method over the staggered grid. From the electromagnetic fields in the time domain, the frequency domain fields can be integrated during modelling using the transformation $u(\mathbf{x},\omega;\mathbf{x}_{s})=\int_{0}^{T_{\max}}u(\mathbf{x},t^{\prime};\mathbf{x}_{s})e^{\mathrm{i}\omega^{\prime}t^{\prime}}\mathrm{d}t^{\prime}=\int_{0}^{T_{\max}}u(\mathbf{x},t^{\prime};\mathbf{x}_{s})e^{-\sqrt{\omega\omega_{0}}t^{\prime}}e^{\mathrm{i}\sqrt{\omega\omega_{0}}t^{\prime}}\mathrm{d}t^{\prime},$ (5) where $u(\mathbf{x},t^{\prime};\mathbf{x}_{s})\in\\{E^{\prime}_{x},E^{\prime}_{y},E^{\prime}_{z},H^{\prime}_{x},H^{\prime}_{y},H^{\prime}_{z}\\}$ denotes the electric/magnetic field at the spatial location $\mathbf{x}\in X$ and the fictitious time $t^{\prime}\in[0,T_{\max}]$ excited by a source from the spatial location $\mathbf{x}_{s}$; $T_{\max}$ is the final time that the frequency domain field $u(\mathbf{x},\omega;\mathbf{x}_{s})$ reaches its steady state. Equation (5) shows that the complex-valued frequency $\omega^{\prime}=(1+\mathrm{i})\sqrt{\omega\omega_{0}}$ leads to an exponentially decay factor in the time integration, which is the key to realize the attenuation effect of diffusive EM field during time evolution. To model the CSEM response as precise as possible, we resort to our newly developed high-order finite-difference time-domain method over non-uniform grid (Yang and Mittet, 2023; Yang, 2023). To satisfy homogeneous boundary condition approximately, perfectly matched layers (PML) (Komatitsch and Martin, 2007) are padded surrounding the domain of interest to absorb the reflections in the truncated domain. The implementation of (5) can be carried out on the fly during timestepping modelling thanks to the discrete time Fourier transform (DTFT): $u(\mathbf{x},\omega_{k};\mathbf{x}_{s})=\sum_{n=0}^{N_{t}-1}u(\mathbf{x},t^{\prime}_{n};\mathbf{x}_{s})\exp(-\sqrt{\omega_{k}\omega_{0}}t^{\prime}_{n})\exp(\mathrm{i}\sqrt{\omega_{k}\omega_{0}}t^{\prime}_{n}),$ (6) where the discretized fictitious time is $t^{\prime}_{n}=n\Delta t$ with time step $\Delta t$ and time index $n=0,\cdots,N_{t}$, while the interested frequencies are $\omega_{k}$, $k=1,\cdots,N_{\omega}$. Note that the total number of time steps $N_{t}$ is much large than the total number of discrete frequencies $N_{\omega}$, i.e., $N_{t}\gg N_{\omega}$. The modelling terminates when the electric field $E^{\prime}(\mathbf{x},\omega;\mathbf{x}_{s})$ and the magnetic field $H^{\prime}(\mathbf{x},\omega;\mathbf{x}_{s})$ evolves to the stead state. This implies that we may check the convergence of the field regularly during the modelling to avoid additional timesteppings which have negligible contributions to the time integral. It is noteworthy that multiple frequencies can be integrated on the fly during the same timestepping procedure. These advantages make the method attractive for forward modelling. ## 3 Fictitious wave domain CSEM inversion The CSEM inversion is an iterative nonlinear optimization procedure. At each iteration, the gradient of the data misfit with respect to conductivity can be computed via the product between forward and adjoint fields. To use fictitious wave domain modelling engine for adjoint simulation, we present a novel approach to estimating the adjoint source time functions by formulating a regularized linear inverse problem. An in-depth analysis shows that this linear inverse problem needs to be solved only once to result in a limited number of basis functions, which can then be used to construct the time-domain adjoint source at different receiver locations. This key development enables efficient and practical 3D CSEM inversion by fictitous wave domain modelling engine. ### 3.1 CSEM inverse problem The Maxwell equations forms a linear system as $\underbrace{\begin{bmatrix}-\sigma&\nabla\times\\\ \nabla\times&-\mathrm{i}\omega\mu\end{bmatrix}}_{\mathbf{A}(m)}\underbrace{\begin{bmatrix}E\\\ H\end{bmatrix}}_{\mathbf{u}}=\underbrace{\begin{bmatrix}J\\\ M\end{bmatrix}}_{\mathbf{f}},$ (7) where the electromagnetic propagator $\mathbf{A}(m)$ is a linear operator applied to the vector field $\mathbf{u}$, which gathers electrical and magnetic fields (each component has been denoted by $u:=u(\mathbf{x},\omega;\mathbf{x}_{s})$ in the previous section). The source vector with nonzero excitation at the source location $\mathbf{x}_{s}$ is prescribed by $\mathbf{f}:=\mathbf{f}(\mathbf{x},\omega;\mathbf{x}_{s})$. Note that all field variables are functions of both frequency and space ($u(\mathbf{x},\omega;\mathbf{x}_{s}),\mathbf{x}\in X,\omega\in\Omega$), while the medium property is only a function of space ($m(\mathbf{x}),\mathbf{x}\in X$), which can be the conductivity $\sigma_{ij}$ or the permeability $\mu$. To find the resistivity of the subsurface, the data misfit is defined in least-squares sense $\phi_{d}(m)=\frac{1}{2}\Big{\|}\underbrace{\begin{bmatrix}\mathbf{W}_{1}&0\\\ 0&\mathbf{W}_{2}\end{bmatrix}}_{\mathbf{W}}\Big{(}\underbrace{\begin{bmatrix}E^{obs}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})\\\ H^{obs}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})\end{bmatrix}}_{\mathbf{d}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})}-\underbrace{\begin{bmatrix}E(\mathbf{x}_{r},\omega;\mathbf{x}_{s})\\\ H(\mathbf{x}_{r},\omega;\mathbf{x}_{s})\end{bmatrix}}_{\mathbf{d}_{syn}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})[m]}\Big{)}\Big{\|}^{2}=\frac{1}{2}\|\mathbf{W}(\mathbf{d}-\mathbf{R}\mathbf{u})\|^{2},$ (8) where $\mathbf{d}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})$ denotes the observed EM data at the receiver location $\mathbf{x}_{r}$ due to the source at the location $\mathbf{x}_{s}$, while the synthetic data $\mathbf{d}_{syn}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})[m]:=\mathbf{u}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})=\mathbf{R}\mathbf{u}(\mathbf{x},\omega;\mathbf{x}_{s})$ (simulated with the model parameter $m$) are extracted by the restriction operator $\mathbf{R}$ from the modelled wavefield at the receiver location $u(\mathbf{x}_{r},\omega;\mathbf{x}_{s})=\int_{X}u(\mathbf{x},\omega;\mathbf{x}_{s})\delta(\mathbf{x}-\mathbf{x}_{r})\mathrm{d}\mathbf{x},\quad\mathbf{x}\in X,\omega\in\Omega$ (9) where $u(\mathbf{x}_{r},\omega;\mathbf{x}_{s})$ is one component of the vector field $\mathbf{u}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})$. In the remainder of the paper, we shall drop the dependence of $(\mathbf{x}_{r},\omega;\mathbf{x}_{s})$ and the summation over sources and receivers without loss of clarity unless clearly stated if necessary. Two weighting matrices, $\mathbf{W}_{1}$ and $\mathbf{W}_{2}$ together forming the diagonal weighting matrix $\mathbf{W}$, are employed to weight the electric and magnetic fields respectively. These weighting matrices may be specified according to the uncertainty model in terms of the real acquisition and the equipment (Mittet and Morten, 2012). If the magnetic data are not considered in the inversion, we simply set $\mathbf{W}_{2}=0$. Taking the first derivative of equation (7) with respect to model parameter $m$ gives $\frac{\partial\mathbf{u}}{\partial m}=-\mathbf{A}^{-1}(m)\frac{\partial\mathbf{A}(m)}{\partial m}\mathbf{u}.$ (10) The gradient of the data misfit with respect to the model parameter $m$ is $\begin{split}\frac{\partial\phi_{d}(m)}{\partial m}=&\Re\langle\mathbf{W}\mathbf{R}\frac{\partial\mathbf{u}}{\partial m},\mathbf{W}(\mathbf{R}\mathbf{u}-\mathbf{d})\rangle\\\ =&-\Re\langle\mathbf{W}\mathbf{R}\mathbf{A}^{-1}(m)\frac{\partial\mathbf{A}(m)}{\partial m}\mathbf{u},\mathbf{W}(\mathbf{R}\mathbf{u}-\mathbf{d})\rangle\\\ =&\Re\langle\frac{\partial\mathbf{A}(m)}{\partial m}\mathbf{u},\underbrace{(\mathbf{A}^{\dagger})^{-1}(m)\mathbf{R}^{\dagger}\mathbf{W}^{\dagger}\mathbf{W}(\mathbf{d}-\mathbf{R}\mathbf{u}}_{\mathbf{v}})\rangle\\\ =&\Re\langle\frac{\partial\mathbf{A}(m)}{\partial m}\mathbf{u},\mathbf{v}\rangle\\\ =&\Re\sum_{s}\sum_{\omega}\bar{\mathbf{v}}^{\mathrm{T}}\frac{\partial\mathbf{A}(m)}{\partial m}\mathbf{u},\end{split}$ (11) where $\Re$ takes the real part of a complex number, $\dagger$ is the complex conjugate transpose. It is noteworthy that the newly introduced variable $\mathbf{v}$, coined adjoint variable, or co-state variable in optimal control theory, must satisfy $\mathbf{A}^{\dagger}(m)\mathbf{v}=\mathbf{R}^{\mathrm{T}}\mathbf{W}^{\mathrm{T}}\mathbf{W}(\mathbf{d}-\mathbf{R}\mathbf{u}).$ (12) This shows that $\mathbf{v}$ satisfies another Maxwell equation based on adjoint Maxwell operator $\mathbf{A}^{\dagger}(m)$. The right hand side of the adjoint equation acts as the virtual source to emanate the adjoint field. Equation (11) is a generic expression applicable in fully anisotropic medium for the model parameters $m\in\\{\sigma_{ij},\mu\\}$. Let us point out that the above gradient expression may be scaled with the local cell volume $\Delta V(\mathbf{x})$ due to the discretization of the spatial integral $\int_{X}\mathrm{d}\mathbf{x}$. In the CSEM settings, the magnetic permeability $\mu$ is considered as the same constant as in the vacuum. We are interested in retrieving the conductivity $\sigma$ (or the resistivity $\rho=1/\sigma$) which may exhibit anisotropy. Equation (7) implies $\frac{\partial\mathbf{A}(m)}{\partial\sigma}=\frac{\partial}{\partial\sigma}\begin{bmatrix}-\sigma&\nabla\times\\\ \nabla\times&-\mathrm{i}\omega\mu\end{bmatrix}=\begin{bmatrix}-1&0\\\ 0&0\end{bmatrix}.$ (13) Denote the adjoint field $\mathbf{v}=(\underline{E},\underline{H})^{\mathrm{T}}$, where the underline is utilized to distinguish it from the forward field. The gradient in equation (11) becomes $\frac{\partial\phi_{d}}{\partial\sigma_{ij}}=-\Re\sum_{s}\sum_{\omega}\underline{\bar{E}}_{j}\cdot E_{i},\qquad i,j\in\\{x,y,z\\}.$ (14) For commonly used VTI medium, the gradients of the misfit with respect to horizontal conductivity $\sigma_{h}$ and vertical conductivity $\sigma_{v}$ are $\frac{\partial\phi_{d}}{\partial\sigma_{h}}=-\Re\sum_{s}\sum_{\omega}(\underline{\bar{E}}_{x}\cdot E_{x}+\underline{\bar{E}}_{y}\cdot E_{y}),\quad\frac{\partial\phi_{d}}{\partial\sigma_{v}}=-\Re\sum_{s}\sum_{\omega}\underline{\bar{E}}_{z}\cdot E_{z}.$ (15) In isotropic medium ($\sigma_{ii}=\sigma$, $i=x,y,z$), we have the simple gradient expression $\frac{\partial\phi_{d}}{\partial\sigma}=-\Re\sum_{s}\sum_{\omega}\underline{E}^{\dagger}\cdot E=-\Re\sum_{s}\sum_{\omega}(\underline{\bar{E}}_{x}\cdot E_{x}+\underline{\bar{E}}_{y}\cdot E_{y}+\underline{\bar{E}}_{z}\cdot E_{z}).$ (16) ### 3.2 Adjoint modelling in fictitious wave domain To facilitate the computation of the adjoint field and the inversion gradient, one may take the conjugate of equation (12): $\mathbf{A}^{\mathrm{T}}(m)\bar{\mathbf{v}}=\overline{\mathbf{R}^{\mathrm{T}}\mathbf{W}^{\mathrm{T}}\mathbf{W}(\mathbf{d}-\mathbf{R}\mathbf{u})}.$ (17) Since $\mathbf{A}^{\mathrm{T}}(m)=\mathbf{A}(m)$, the above operation allows us to use the same modelling engine to compute the conjugate of the adjoint field, as long as proper adjoint source can be provided for adjoint simulation. Unfortunately, for CSEM inversion in the frequency domain, only few discrete frequencies are available, leading to the adjoint source (i.e. the data residual) specified only at few discrete frequencies. A time domain adjoint source is therefore not directly accessible. To still use fictitious wave domain timestepping modelling for adjoint simulation, let us now split the weighted data residual into electrical and magnetic components $\overline{\mathbf{R}^{\mathrm{T}}\mathbf{W}^{\mathrm{T}}\mathbf{W}(\mathbf{d}-\mathbf{R}\mathbf{u})}=(\overline{\delta d_{E}},\overline{\delta d_{H}})^{\mathrm{T}}$ and repeat the substitution and multiplication operations as in Section 2: $\begin{cases}\nabla\times\underbrace{\overline{\underline{E}}}_{\underline{\overline{E}}^{\prime}}+\mu\underbrace{\sqrt{-\mathrm{i}2\omega\omega_{0}}}_{-\mathrm{i}\omega^{\prime}}\underbrace{\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\underline{\overline{H}}}_{\underline{\overline{H}}^{\prime}}=\overline{\delta d_{E}},\\\ \nabla\times\underbrace{\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\underline{\overline{H}}}_{\underline{\overline{H}}^{\prime}}-\underbrace{\sqrt{-\mathrm{i}2\omega\omega_{0}}}_{-\mathrm{i}\omega^{\prime}}\varepsilon\underline{\overline{E}}=\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\overline{\delta d_{H}}.\end{cases}$ (18) We need to consider the time domain counterpart of the above system for efficient adjoint modelling by fictitious timestepping modelling. Switching from diffusive frequency domain back to fictitious time domain with wave should be understood as a linear transformation rather than the true inverse Fourier transform. This is because the mapping of fictitious transform is not bijective, as the number of time steps $N_{t}$ for numerical simulation is significantly larger than the number of discrete frequencies $N_{\omega}$ used for CSEM investigation ($N_{\omega}\ll N_{t}$). There exists an infinite number of time series which may match the few discrete frequencies. Searching for the inverse is therefore an under-determined problem. Let us denote the right hand side of equation (18) ($\overline{\delta d_{E}}$ or $\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\overline{\delta d_{H}}$) at a specific receiver location as $s(\omega)$. To retrieve a long time series for adjoint modelling, a linear inverse problem is formulated to convert the adjoint source $s(\omega)$ from frequency to time domain based on DTFT in equation (6) by minimizing the following misfit functional $\begin{split}\Psi&=\sum_{k=1}^{N_{\omega}}|s(\omega_{k})-\sum_{n=0}^{N_{t}-1}e^{-\sqrt{\omega_{k}\omega_{0}}t^{\prime}_{n}}e^{\mathrm{i}\sqrt{\omega_{k}\omega_{0}}t^{\prime}_{n}}s(t^{\prime}_{n})|^{2}+\gamma\sum_{n=0}^{N_{t}-1}|s(t^{\prime}_{n})|^{2}\\\ &=\sum_{k=1}^{N_{\omega}}\Big{(}|\Re\\{s(\omega_{k})\\}-\sum_{n=0}^{N_{t}-1}e^{-\sqrt{\omega_{k}\omega_{0}}t^{\prime}_{n}}\cos(\sqrt{\omega_{k}\omega_{0}}t^{\prime}_{n})s(t^{\prime}_{n})|^{2}\\\ &\qquad+|\Im\\{s(\omega_{k})\\}-\sum_{n=0}^{N_{t}-1}e^{-\sqrt{\omega_{k}\omega_{0}}t^{\prime}_{n}}\sin(\sqrt{\omega_{k}\omega_{0}}t^{\prime}_{n})s(t^{\prime}_{n})|^{2}\Big{)}+\gamma\sum_{n=0}^{N_{t}-1}|s(t^{\prime}_{n})|^{2}\end{split}$ where $\Im$ takes the imaginary part of the complex variable. If we discretize the time as $t_{n}=n\Delta t$ and define $\mathbf{s}(\omega):=\begin{bmatrix}\Re\\{s(\omega_{1})\\}\\\ \vdots\\\ \Re\\{s(\omega_{N_{\omega}})\\}\\\ \Im\\{s(\omega_{1})\\}\\\ \vdots\\\ \Im\\{s(\omega_{N_{\omega}})\\}\end{bmatrix},\mathbf{s}(t^{\prime}):=\begin{bmatrix}s(t^{\prime}_{0})\\\ \vdots\\\ s(t^{\prime}_{N_{t}-1})\end{bmatrix}$ (19) and $\mathbf{B}:=\begin{bmatrix}e^{-\sqrt{\omega_{1}\omega_{0}}t^{\prime}_{0}}\cos(\sqrt{\omega_{1}\omega_{0}}t^{\prime}_{0})&\cdots&e^{-\sqrt{\omega_{1}\omega_{0}}t^{\prime}_{N_{t}-1}}\cos(\sqrt{\omega_{1}\omega_{0}}t^{\prime}_{N_{t}-1})\\\ \vdots&\ddots&\vdots\\\ e^{-\sqrt{\omega_{N_{\omega}}\omega_{0}}t^{\prime}_{0}}\cos(\sqrt{\omega_{N_{\omega}}\omega_{0}}t^{\prime}_{0})&\cdots&e^{-\sqrt{\omega_{N_{\omega}}\omega_{0}}t^{\prime}_{N_{t}-1}}\cos(\sqrt{\omega_{N_{\omega}}\omega_{0}}t^{\prime}_{N_{t}-1})\\\ e^{-\sqrt{\omega_{1}\omega_{0}}t^{\prime}_{0}}\sin(\sqrt{\omega_{1}\omega_{0}}t^{\prime}_{0})&\cdots&e^{-\sqrt{\omega_{1}\omega_{0}}t^{\prime}_{N_{t}-1}}\sin(\sqrt{\omega_{1}\omega_{0}}t^{\prime}_{N_{t}-1})\\\ \vdots&\ddots&\vdots\\\ e^{-\sqrt{\omega_{N_{\omega}}\omega_{0}}t^{\prime}_{0}}\sin(\sqrt{\omega_{N_{\omega}}\omega_{0}}t^{\prime}_{0})&\cdots&e^{-\sqrt{\omega_{N_{\omega}}\omega_{0}}t^{\prime}_{N_{t}-1}}\sin(\sqrt{\omega_{N_{\omega}}\omega_{0}}t^{\prime}_{N_{t}-1})\\\ \end{bmatrix},$ (20) the objective of the above linear inverse problem can be compactly written in matrix form $\Psi=\|\mathbf{s}(\omega)-\mathbf{B}\mathbf{s}(t^{\prime})\|^{2}+\gamma\|\mathbf{s}(t^{\prime})\|^{2},$ (21) where a regularization term taking into account the minimum energy of the solution has been penalized by a parameter $\gamma$ to stabilize the linear inversion and to determine a unique solution. Note that the complex factors have been split into real and imaginary part. To solve this ill-posed problem, regularization has been introduced to recover the well-posedness during the inversion of the matrix $\mathbf{B}\in\mathbb{R}^{2N_{\omega}\times N_{t}}$. The solution $\mathbf{s}(t)$ can easily be found using linear optimization algorithms, e.g. LSQR (Paige and Saunders, 1982) and the conjugate gradient method for the resulting normal equation (CGNR) (Saad, 2003). Based on the frequency and time index, the each element of $\mathbf{B}$ can be formed on the fly when computing matrix vector product, implying a matrix free implementation. It should be noted that fictitious wave domain modelling is simply a mathematical tool to efficiently compute frequency domain EM fields. Disguising a diffusive phenomenon as a wave event will not change the diffusive nature of the underlying physics. Indeed, Mittet (2010) proposed another method to calculate the adjoint source time function by superposition of delayed causal wavelets. However, it yields a nonlinear inverse problem which is very difficult to solve. Clearly, embedding a linear inverse problem in a nonlinear inverse problem is a much better option than embedding double nonlinear inverse problems. The truncated singular value decomposition (SVD) have been applied (Støren et al., 2008) to invert for an adjoint source, following the modified wave domain formulation of Maaø (2007). Mathematically, our approach using damped least-squares should give the same solution as the one from truncated SVD, since SVD is equivalent to finding the Moore-Penrose pseudo-inverse for $\mathbf{B}$ (Björck, 1996, theorem 1.2.10 in p. 15, and section 2.7.2, p. 101). However, our empirical experience shows that the approximate inverse given by SVD suffers from numerical instability while consuming too much memory. The above linear inversion method gets rid of these issues, incorporating all frequencies in one run with better memory and computational efficiency. ### 3.3 A matrix-free basis function implementation Let us denote $\mathbf{B}^{+}=(\mathbf{B}^{\mathrm{T}}\mathbf{B}+\gamma\mathbf{I})^{-1}\mathbf{B}^{\mathrm{T}}\in\mathbb{R}^{N_{t}\times 2N_{\omega}}$ the Moore-Penrose pseudo-inverse of $\mathbf{B}\in\mathbb{R}^{2N_{\omega}\times N_{t}}$ such that $\mathbf{s}(t)=\mathbf{B}^{+}\mathbf{s}(\omega).$ (22) The matrix $\mathbf{B}^{+}$ may be written down using the $2N_{\omega}$ columns: $\mathbf{B}^{+}=\begin{bmatrix}\mathbf{b}_{1},\cdots,\mathbf{b}_{N_{\omega}},\mathbf{b}_{N_{\omega}+1},\cdots,\mathbf{b}_{2N_{\omega}}\end{bmatrix}$ (23) in which $\mathbf{b}_{k}\in\mathbb{R}^{N_{t}}$ is the $i$-th column of $\mathbf{B}^{+}$. As a result, equation (22) translates into $\mathbf{s}(t)=\sum_{k=1}^{N_{\omega}}\Re\\{s(\omega_{k})\\}\mathbf{b}_{k}+\Im\\{s(\omega_{k})\\}\mathbf{b}_{k+N_{\omega}}.$ (24) It then becomes clear that the columns of $\mathbf{B}^{+}$ are the very few number of basis functions to construct the adjoint source time function. Given $\Delta t$ and $N_{t}$, these basis functions are uniquely determined, hence independent of the frequency spectrum $s(\omega)$ (which are simply the coefficients for the linear combination of these basis functions). This means the iterative solution procedure needs to be performed only once to find all $\mathbf{b}_{k}$, which can then be used for each receiver locations to infer the corresponding adjoint source time function. Equation (24) also gives a recipe to efficiently compute these functions: * • If we set $\Re\\{s(\omega_{k})\\}=1$, $\Im\\{s(\omega_{k})\\}=0$ and $\Re\\{s(\omega_{j})\\}=\Im\\{s(\omega_{j})\\}=0$ ($j\neq k$), we obtain $\mathbf{s}(t)=\mathbf{b}_{k}$. * • If we set $\Im\\{s(\omega_{k})\\}=1$, $\Re\\{s(\omega_{k})\\}=0$ and $\Re\\{s(\omega_{j})\\}=\Im\\{s(\omega_{j})\\}=0$ ($j\neq k$), we obtain $\mathbf{s}(t)=\mathbf{b}_{k+N_{\omega}}$. By repeatedly feeding the iterative optimization algorithm $2N_{\omega}$ times using different Dirac delta input $\mathbf{s}(\omega)=(0,\cdots,0,1,0,\cdots,0)^{\mathrm{T}}$, we obtain the $2N_{\omega}$ columns of $\mathbf{B}^{+}$. Figure 1 gives an example of these basis functions using the parameters $\Delta t=0.002$, $N_{t}=1000$ and $f_{0}=1$ and $N_{\omega}=3$. To solve this linear optimization problem, Claerbout’s conjugate gradient algorithm (Claerbout and Fomel, 2008, chapter 2.3.6) (which is essentially a CGNR algorithm according to Saad (2003, chapter 8.3, pp. 266-268)) has been applied. As can be seen from Figure 2, within 15 CG iterations, the proposed method leads to highly accurate estimation for these basis functions, with the error less than $10^{-8}$. In this test, the regularization parameter $\gamma$ was chosen to be $10^{-3}$. It is found the shape of the basis functions do not change much using different choices of $\gamma$, indicating that the method is robust. For different receiver locations, the data residuals in the frequency domain may be dramatically different, but these basis functions are the same. We therefore only solve the above regularized linear inverse problem once to construct all adjoint source time functions through their linear combination. The frequency domain data residuals are the weighting coefficients to modulate these basis. Figure 1: The basis functions using the parameters $\Delta t=0.004$ s, $N_{t}=1000$ and $N_{\omega}=3$. The amplitude of these basis functions decays to zero with the increase of the time, which is quite suitable for frequency domain EM fields converging to steady state after simulation using sufficient number of time steps. Figure 2: The convergence of CG iterations for the solution of basis functions One thing to remark is that the adjoint modelling in the time domain should have reverse time order compared with the forward modelling. Taking the conjugate reverse the time again. This is the reason why a decaying trend of the estimated basis function is observed in Figure 1. ### 3.4 Connections and distinctions with existing methods Note the incident field $E_{i}$ is the product between the Green’s function with the source current $E_{i}(\mathbf{x},\omega;\mathbf{x}_{s})=\sum_{k}G_{ik}^{EE}(\mathbf{x},\omega;\mathbf{x}_{s})J_{k}(\mathbf{x}_{s},\omega),$ (25) while the conjugate of adjoint field can be specified as $\overline{\underline{E}}_{j}(\mathbf{x},\omega;\mathbf{x}_{s})=\int_{X}(\sum_{k}G_{jk}^{EE}(\mathbf{x},\omega;\mathbf{x}_{r})\overline{\delta d^{E}_{k}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})}+\sum_{i}G_{ji}^{EH}(\mathbf{x},\omega;\mathbf{x}_{r})\overline{\delta d^{H}_{i}(\mathbf{x}_{r},\omega;\mathbf{x}_{s})})\mathrm{d}\mathbf{x}_{r},$ (26) where $i,j,k\in\\{x,y,z\\}$ indicate different components of the EM field, while $G_{ij}^{EE}(\mathbf{x},\omega;\mathbf{x}_{s})$ and $G_{ik}^{EH}(\mathbf{x},\omega;\mathbf{x}_{s})$ stand for the $i$th component of the electrical Green’s function for angular frequency $\omega$ at the spatial location $\mathbf{x}$ due to the electrical and magnetic sources directed in $j$\- and $k$\- th directions. This implies that we can still use fictitious wave modelling for adjoint simulation: one can, for each frequency and each receiver, first compute the Green’s function (which is independent of the source time function) in frequency domain, and then form the adjoint field by linear combination of them using the data residual. This method, ensured by the superposition principle thanks to the linearity of the Maxwell equation, enables the approach still working out. Such a computing scheme repeats many times of simulation depending on the number of receivers, which is very inefficient when the number of receivers becomes extremely large. Equation (17) shows that the adjoint field can be computed using only one modelling, provided that a time-domain adjoint source time function is available and then injected at once to do time-domain simulation. The spirit of our method is to perform frequency domain inversion using time domain modelling engine, similar to Sirgue et al. (2008, 2010) for seismic full waveform inversion (FWI). However, there are a number of differences. Switching between time and frequency domain is straightforward in seismic FWI using the definition of discrete inverse Fourier transform. One can convert frequency domain data residual using the definition of inverse discrete Fourier transform $s(t)=\frac{1}{2\pi}\int_{-\infty}^{\infty}s(\omega)e^{-i\omega t}\mathrm{d}\omega\approx\sum_{k=1}^{N_{\omega}}s(\omega_{k})e^{-i\omega_{k}t}.$ (27) Of course, the same issue presented in Sirgue et al. (2010) approach: only limited number of frequencies are available to reconstruct a long time series. The last equality in (27) assumes all the absent frequencies are zeros. Since Fourier basis is orthonormal, the computation of adjoint source time function under this assumption leads to a minimum energy solution equivalent to solving the following least-squares minimization problem $\min_{s(t)}\sum_{k=1}^{N_{\omega}}\|s(\omega_{k})-\sum_{j=1}^{N_{t}}s(t_{j})e^{\mathrm{i}\omega_{k}t_{j}}\|^{2}.$ (28) For fictitious wave domain approach, plugging the complex-valued frequency $\omega^{\prime}=(1+\mathrm{i})\sqrt{\omega\omega_{0}}$ into exponential factor yields an exponentially decay factor $e^{\mathrm{i}\omega^{\prime}t^{\prime}}=e^{\mathrm{i}(1+\mathrm{i})\sqrt{\omega\omega_{0}}t^{\prime}}=e^{-\sqrt{\omega\omega_{0}}t^{\prime}}e^{\mathrm{i}\sqrt{\omega\omega_{0}}t^{\prime}},$ (29) leaving the converted EM field possessing strong attenuation behavior, while making the basis no more orthogonal. Precisely speaking, the transformation in (5) is a Laplace transform rather than normal Fourier transform. As a result, the inverse fictitious transformation is not well defined, that is, $u(\mathbf{x},t^{\prime};\mathbf{x}_{s})\neq\frac{1}{2\pi}\int_{\Omega}\mathrm{d}\omega u(\mathbf{x},\omega;\mathbf{x}_{s})e^{-\mathrm{i}\omega^{\prime}t^{\prime}}.$ (30) Since direct use of inverse Fourier transform as done by Sirgue et al. (2010) does not apply, a matrix-free iterative solution of adjoint source time function in this work is evidently desirable. ### 3.5 The final algorithm The total misfit functional for CSEM inversion consists of both data fitting and model regularization: $\phi(m)=\phi_{d}(m)+\beta\phi_{m}(m),$ (31) where $\phi_{m}(m)$, penalized by the parameter $\beta$, is the model misfit term to enforce smoothness on the inverted model. A popular choice is to use Tikhonov regularization to minimize the roughness of the model compared with a reference model $m_{ref}$, i.e., $\phi_{m}(m)=\frac{1}{2}\|\nabla^{\alpha}(m-m_{ref})\|^{2}$, where the anisotropic first order difference operator $\nabla^{\alpha}=(\alpha_{x}\partial_{x},\alpha_{y}\partial_{y},\alpha_{z}\partial_{z})^{\mathrm{T}}$ with coefficients $\alpha_{x}$, $\alpha_{y}$ and $\alpha_{z}$ can be tuned by the user. This leads to the total gradient of the misfit functional prescribed by $\frac{\partial\phi}{\partial m}=\frac{\partial\phi_{d}}{\partial m}+\frac{\partial\phi_{m}}{\partial m}$. Based on the gradient information computed above, a descent direction $\delta m^{k}$ can be constructed to update the model parameters iteratively $m^{k+1}=m^{k}+\alpha\delta m^{k},$ (32) where $\alpha$ is the step length estimated by line search method. In this paper, the descent direction $\delta m^{k}$ at the $k$-th iteration is computed using l-BFGS algorithm (Nocedal and Wright, 2006) by storing the gradients in the previous iterations. It is worth noting that the inversion gradient computed using the proposed approach is approximate rather than exact. Consequently, the descent direction estimated using l-BFGS after several iterations may have difficulty to succeed in the line search procedure. We therefore propose to restart l-BFGS using the steepest descent direction after the first failure of the line search. The algorithm will eventually be terminated if the maximum number of iterations is reached. Using conductivity as model parameter creates some kinds of ill-conditioning due to its value varying in a large range (the seawater is around 0.3 $\Omega\cdot$m while a thin layer of hydrocarbon bearing sediment is more than 100 $\Omega\cdot$m). To capture the high contrast of the parameter variations, the inversion is re-parameterized using the logarithmic transformation $m:=\ln\rho$, which is a dimensionless parameter with comparable magnitude varying in a much smaller dynamic range. Indeed, there exists other types of parameter scaling, see another example in Abubakar et al. (2008). This re- parametrization gives a relation $\rho=e^{m}$. Switching the parametrization is trivial thanks to the chain rule: $\frac{\partial\phi}{\partial m}=\frac{\partial\rho}{\partial m}\cdot\frac{\partial\sigma}{\partial\rho}\cdot\frac{\partial\phi}{\partial\sigma}=-\frac{1}{\rho}\frac{\partial\phi}{\partial\sigma}.$ (33) During the iterative inversion, the model parameter may be bounded within an upper bound $m_{\max}$ and a lower bound $m_{\min}$ according to a priori knowledge. These bounds help to stabilize the inversion in a more physically sensible manner. ## 4 Application examples In this section we present two numerical examples to demonstrate the application of our method. The two examples are designed to validate the gradient expressions for the isotropic (the first example) and the VTI anisotropic (the second example) cases. In both examples, I use in-line and azimuth data with $E_{x}$ and $E_{y}$ components. This motivates us to utilize the uncertainty model proposed in Morten et al. (2009) to compute the diagonal weight matrix $\mathbf{W}$ in order to capture the varying sensitivity of broadside EM data. The observed data are poluted by 3% of Gaussian white noise before inversion. The regularization parameters are configured with $\beta=0.01$ using a cooling factor (between 0.7 and 1) through iterations. We set $\alpha_{x}=\alpha_{y}=1$ and $\alpha_{z}=0.1$ in both tests. We choose the initial model as the reference model for both inverse exercises. To mitigate the weak sensitivity of the EM field to the deeper part of the model, a depth preconditioning has been applied following the work of Plessix and Mulder (2008). The code is parallelized over the sources to achieve the best scaling performance since the sources are independent of each other. ### 4.1 A two-block model Our first example is a two-block land model similar to the one presented in Grayver et al. (2013). The model has a background resistivity of 5 $\Omega\cdot$m, including two anomalies: a low resistivity inclusion of 1 $\Omega\cdot$m and a high resistivity inclusion of 100 $\Omega\cdot$m, as shown in Figure 3a. The physical dimension of the model spans over a 3D domain $X=X_{1}\times X_{2}\times X_{3}$, where $X_{1}=[-2000,2000]$ m, $X_{2}=[-2000,2000]$ m and $X_{3}=[0,1500]$ m. There are 16 receivers and 256 transmitters with equal distance on the surface of this model. Figure 4 gives the survey layout sheet of the acquisition geometry. The model has been discretized with the regular grid using $\Delta x=\Delta y=60$ m and $\Delta z=25$ m. The observed data are generated at two frequencies: 0.25 Hz and 1 Hz. Figure 3: (a) The true land model (5 $\Omega\cdot$m background with two anomalous inclusions of 1 $\Omega\cdot$m (left) and 100 $\Omega\cdot$m (right) ); (b) The homogeneous initial model of 5 $\Omega\cdot$m. The models are clipped at 25 $\Omega\cdot$m for display purpose. Figure 4: Survey layout sheet for the two-block model. In total, 16 transmitters (marked by dots) and 256 receivers (labeled with triangles) are deployed. This 3D CSEM inversion runs for 30 iterations. In order to check how the synthetic data fits the observations, we plot both the amplitude and the phase of the synthetic data at all receivers together with the observed data using the initial model and the retrieved resistivity model after inversion. In Figure 5, the synthetic data created from the homogeneous initial model cannot match the observed data very well. In Figure 6, the synthetic data generated from inverted resistivity volume indeed are better aligned with the observed data. Note that the horizontal axis in Figures 5 and 6 are receiver indices rather than offset. This is due to the broadside survey configuration. To better visualize the data matching connected to acquisition geometry, we design a scatter plot with hot colors indicating the data error at each receiver location. Figures 7 shows that at the beginning of the inversion, the significant misfit (defined as $\|\mathbf{W}(\mathbf{d}_{obs}-\mathbf{d}_{syn})\|$ at each receiver location) is very large. Since we mute the data within 600 m offset, the significant misfit surrounding transmitter Tx-10 are zeros. The plot in Figure 7 also highlights the importance of azimuthal data has much larger significant misfit based than inline directions. After 30 inversion iterations, the significant misfit becomes much lower, see Figure 8. Figure 5: Comparison between observed data and synthetic data from initial model according to (a) amplitude and (b) phase for transmitter Tx-10. The horizontal axis is the index of the receivers rather than offset, since we consider broadside configuration. It is clear that both the amplitude and phase does not match very well. Note also that the level of the data matching has a strong correlation with the receiver indices. Figure 6: Comparison between observed data and synthetic data from inverted model according to (a) amplitude and (b) phase for transmitter Tx-10. The synthetic data are now better aligned with the observed data after overlapping display. Figure 7: Scatter plot of the significant misfit for transmitter Tx-10 at iteration 1. Figure 8: Scatter plot of the significant misfit for transmitter Tx-10 at iteration 30. The significant misfit becomes much lower after inversion. The above data comparison gives us a confidence of our inversion scheme working properly, according to a specific source (Tx-10). In Figure 9, we plot the normalized data misfit which is a global measure of the inversion. We see that after 30 iterations, it arrives at a relatively low normalized misfit. According to Figure 10a, the frequency of large significant misfit at receiver locations are distributed in a large range from 0 to 15. After the inversion, it has been compressed a lot (most of them are clustered within 3), as shown in Figure 10b. The final inverted model is capable to retrieve both the low resistivity inclusion (Figure 11a) and the high resistivity inclusion (Figure 11b). Figure 9: The convergence history in terms of the normalized misfit for the inversion of the two-block model. Figure 10: The histogram of the significant misfit at (a) iteration 1 and (b) iteration 30. Figure 11: Inversion result for (a) low resistivity anomaly and (b) high resistivity anomaly. ### 4.2 A marine CSEM example In practical CSEM survey for reservoir exploration, a number of towlines using very low-frequency (0.1-10 Hz), high-energy electrical source will be deployed with hundreds of receivers. The source-receiver offset extends up to 10 km is very standard. Here, a 3D CSEM inversion is carried out for 30 iterations based on a model of size $X=X_{1}\times X_{2}\times X_{3}$, where $X_{1}=X_{2}=[-9000,9000]$ m and $X_{3}=[0,3500]$ m. Figure 12 illustrates the survey layout sheet for the test: 10 towlines (5 in x direction and 5 in y direction) are deployed in order to cover the region of interest. Each towline has 81 source locations with a separation of 200 m. The receivers (25 in total) are deployed at the crossings of these towlines. We consider the reciprocity to switch the source and receiver, to achieve efficiency in inversion. As shown in Figure 13a, this synthetic model includes the seawater of 0.3 $\Omega\cdot$m. The resistivity of the background sediment is varying from 1.5 to 2.5 $\Omega\cdot$m along the depth. The most striking feature is a resistor of 10 $\Omega\cdot$m at shallow part and a disk-shaped resistor of 100 $\Omega\cdot$m sitting at the depth between 2200 m and 2350 m, which mimics the canonical reservoir. The starting model for the inversion in Figure 13b takes a homogeneous value 1.5 $\Omega\cdot$m to mimic the situation that no accurate a priori information is known for inversion. A seafloor bathymetry with varying depth has been embedded in the model. The model is densely gridded around bathymetry and above, with the grid spacing $\Delta z=25$ m, see Figure 14. The grid has been stretched with a constant growing factor at a certain distance below the seabed. Three frequencies (0.25 Hz, 0.75 Hz and 2.25 Hz) have been used to perform this inversion. Figure 15 shows that this CSEM inversion was converging well. The comparison between Figure 16 and Figure 17 shows that the significant misfit has been largely reduced for the receiver at the location (0, 0) m. Note that in Figure 17c, the data misfit for the third frequency at the near offset is still large, indicating the presence of shallow resistor, which has not been well recovered in the inverted model (cf. Figure 19). To recover also the shallow resistor, we should use more near offset data rather than drop off them since 1000 m. A frequency dependent weighting strategy can be applied to boost the importance of high frequencies, as they have shallow penetrating depth according to skin depth. Since the scatter plot of the signficant misfit does not reflect the global misfit, the significant misfit at each receiver location has been shown in Figure 18: at the beginning of the inversion, they are far from 1; after the inversion, the significant misfit at all receiver locations are close to 1, which is the target value after adding noise into the observed data. After 30 l-BFGS iterations, the inversion was successful to recover the resevoir in both vertical resistivity ($\rho_{v}=1/\sigma_{h}$) and the horizontal resistivity ($\rho_{h}=1/\sigma_{v}$), but the shallow resistor is difficult to obtain. The imprint of the shallow resistor can be found just above the resevoir in our reconstructed model: the horizontal location is correct but the depth is obviously incorrect. It can be seen that the recovered anomaly in $\rho_{v}$ is of 100-150 m shallower than the true location. This highlights the low resolution of CSEM inversion and the ambuiguity of the depth. The inverted vertical resistivity in Figure 19a is much better resolved than the horizontal resistivity in Figure 19b. The value of the retrieved $\rho_{v}$ is higher than the retrieved $\rho_{h}$, while both are smaller than 100 $\Omega\cdot$m and thicker than the truth. This is due to the effect of anomalous transverse resistance (ATR) (Mittet and Morten, 2012, equation 11): because the CSEM response is proportional to the product of resistivity difference times the thickness, lower resistivity with larger thickness can create similar amplitude response compared with higher resistivity with thin depth distribution. There is also weak increase of the resistivity in $\rho_{h}$ in Figure 19b, suggesting that the initial model of homogeneous background lower than the true model, as can be seen by comparing with Figure 19c. Figure 12: Survey layout sheet for marine CSEM inversion. Figure 13: (a) The true resistivity model (note that the resistivity of the sediment below the seabed is varying from 1.5 to 2.5 $\Omega\cdot$m); (b) The initial resistivity model (the resistivity of the sediment takes a homogeneous value 1.5, to mimic the situation that no accurate a priori information is known). The models are clipped at 5 $\Omega\cdot$m for display purpose. Figure 14: The vertical section of the nonuniform grid to mesh the 3D marine CSEM resistivity Figure 15: The convergence history of the normalized misfit for marine CSEM inversion. Figure 16: Scatter plot of the significant misfit for receiver at (0, 0) m in iteration 1. Figure 17: Scatter plot of the significant misfit for receiver at (0, 0) m in iteration 30. The significant misfit becomes much lower after inversion. Figure 18: The significant misfit at each receiver location before and after inversion Figure 19: The result of 3D marine CSEM inversion: (a) vertical resistivity $\rho_{v}$, (b) horizontal resistivity $\rho_{h}$ and (c) true resistivity model. The display is in logorithmic scale. Note that in the inverted vertical resistivity, there is a small uphill variation just above the deep resistor which has a good correspondence to the shallow resistor in horizontal direction, where the highest resistivity has been retrieved at the center. ## 5 Discussion In our method, the modelling jobs for each source can be parallelized independently with perfect scalability. Since it is based on time stepping, the EM fields at the next time step will overwrite the memory unit at the previous step. This makes the memory usage of the method very economically. Our computation was performed on a Intel(R) Xeon(R) Gold 6258R CPU @ 2.70GHz, possessing 56 CPU cores (each core has 2 threads) and 128 GB memory. For the two-block model of 3D grid size 139*139*99 (after padding with PML boundaries on each side), combination of forward and adjoint modelling for building the gradient associated with one source only takes the memory of approximately 402 MB to compute all frequencies. (The memory requirement using frequency domain finite difference direct solver for such a large model size could not even fit 128 GB memory.) This implies that we can simulateneously launch more than 100 modelling jobs (corresponding to 100 independent sources) to fully explore the small memory footprint of fictitous wave domain modelling in performing large scale 3D inversion. For 16 sources, we can use 4 threads to parallelize the computation by OpenMP using shared memory. The modelling takes less time at the beginning of the inversion, as the model is conductive and has less inhomogeneties. At the later iterations, each gradient building takes more computing time than before, because the resistive anomalies gradually come into the model, the time step $\Delta t$ becomes smaller and the number of timesteps becomes larger due to stability condition. The inversion with 30 iterations can be completed within 10 hours. On average, building one gradient (equivalent to one forward plus one adjoint solve) takes roughly 20 minutes. It should be noted that the time series generated by our method only allows the matching of the field at given frequencies. The above showcases simply demonstrate that the proposed method works well, making fictitious wave domain an attractive modelling engine for 3D practical inversion. However, one has to be aware that the choice of regularization parameters $\beta$, $\alpha_{x}$, $\alpha_{y}$, $\alpha_{z}$ are of paramount importance to the final inversion result. Lelièvre and Farquharson (2013) have made a thorough investigation and concluded that regularization of noisy data usually requires a lot of apriori knowledge about the geologic structure, such that the related parameters can be specified properly. Our choices may be far from perfect, and better choices should be possible to significantly improve the inversion result, while avoiding the false positives in imaging. These are, however, irrelevant to the core contribution of this work. The derivation in the appendix A develops an alternative adjoint formulation to avoid solving the regularized linear inverse problem. Equation (40) gives an explicit expression to build CSEM inversion gradient directly by cross- correlation of fictitious time domain forward and adjoint fields. However, 3D implementation of such a strategy is not trivial as it is mandatory to simultaneously access fictitious time domain forward and adjoint field at each time step. This is challenging due to the opposite directions in time between forward and adjoint fields. In doing so, one may think of storing all 3D wavefield snapshots during thunsands of times and accessing them one by one in reverse order, which is extremely inefficient due to slow memory acess speed and heavy IO traffic. A feasible approach is to use the wavefield reconstruction method by storing the boundary values (Yang et al., 2016b), or resort to optimal checkpointing strategy (Symes, 2007; Yang et al., 2016a). Wavefield reconstruction by boundary storage implies each gradient computation involves three times of simulation, two for forward fields and one for adjoint fields. Optimal checkpointing requires much higher recomputation ratio and thus even more computationally expensive. In summary, these strategies demand either significant memory consumption for storing boundaries of forward fields at every time step, or sophisticated operations to achieve optimal combinatorial recomputation of the forward fields, thus left as the future work. The scheme used in this paper is a hybrid combination of time domain modelling and frequency domain inversion. To build the gradient, only two times of modelling are needed, one for forward fields and one for adjoint fields. Consequently, the gradient computed by product between forward and adjoint fields in frequency domain becomes the most efficient option among all equivalent formulations. ## 6 Conclusion A nonlinear CSEM inversion based on fictitious wave domain modelling using timestepping simulation has been presented. To construct the gradient using forward and adjoint fields, the adjoint source time functions are estimated by solving a regularized linear inverse problem. A matrix-free implementation using the basis functions is proposed such that the linear inverse problem is iteratively solved only once and then the resulting basis functions can be reused to build up all adjoint source time functions at every receiver location. The method enables efficient 3D CSEM inversion which are demonstrated by numerical examples. The method will become more attractive for land CSEM which normally requires much more number of frequencies than marine CSEM. The proposed method can equally be applied to magnetotelluric (MT) inversion to decipher the subsurface structure in even larger scale using much lower frequencies of the EM recordings. ## Computer Code Availability Name of the code: `libEMMI` Functionality: Library for 3D controlled-source electromagnetic modelling and inversion Contact: Pengliang Yang, Harbin Institute of Technology, China E-mail<EMAIL_ADDRESS> System requirements: Linux OS Programming language: C, Fortran Result visualization: Python3, Madagascar (Scons) and Gnuplot Compilation requirement: gcc compiler, mpicc compiler and make Software required: fftw3 (http://fftw.org/) The code has been released publically via github repository https://github.com/yangpl/libEMMI. ## Acknowledgements The author acknowledges the support from Chinese Fundamental Research Funds for the Central Universities (AUGA5710010121) and National Natural Science Fundation of China (42274156). The author is indebted to Rune Mittet for the inspiring discussions and the encouragement of this work. The alternative adjoint formulation in the appendix A was due to René-Édourt Plessix. ## Appendix A An alternative adjoint formulation The above approach uses diffusive Maxwell equation in frequency domain as the state equation, which holds no matter what kind of modelling engine is used as the solver. The misfit in (8) can also be expanded as $\phi_{d}(m)=\frac{1}{2}\sum_{s}\sum_{k}|\mathbf{W}_{1}(E^{obs}(\mathbf{x}_{r},\omega_{k};\mathbf{x}_{s})-RE(\mathbf{x},\omega_{k};\mathbf{x}_{s}))|^{2}+|\mathbf{W}_{2}(H^{obs}(\mathbf{x}_{r},\omega_{k};\mathbf{x}_{s})-RH(\mathbf{x},\omega_{k};\mathbf{x}_{s}))|^{2}.$ Based on equation (4) and the correspondence between wave and diffusive EM fields, the Lagrangian functional translates a constrained minimization into an unconstrained optimization $\begin{split}\mathcal{L}=&\frac{1}{2}\sum_{s}\sum_{k}|\mathbf{W}_{1}(E^{obs}(\mathbf{x}_{r},\omega_{k};\mathbf{x}_{s})-RE(\mathbf{x},\omega_{k};\mathbf{x}_{s}))|^{2}+|\mathbf{W}_{2}(H^{obs}(\mathbf{x}_{r},\omega_{k};\mathbf{x}_{s})-RH(\mathbf{x},\omega_{k};\mathbf{x}_{s}))|^{2}\\\ +&\Re\sum_{s}\sum_{k}\int_{X}\mathrm{d}\mathbf{x}\overline{\lambda_{1}(\mathbf{x},\omega_{k};\mathbf{x}_{s})}\bigg{(}E(\mathbf{x},\omega_{k};\mathbf{x}_{s})-\int_{0}^{T_{\max}}dt^{\prime}E^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})e^{-\sqrt{\omega_{k}\omega_{0}}t^{\prime}}e^{\mathrm{i}\sqrt{\omega_{k}\omega_{0}}t^{\prime}}\bigg{)}\\\ +&\Re\sum_{s}\sum_{k}\int_{X}\mathrm{d}\mathbf{x}\overline{\lambda_{2}(\mathbf{x},\omega_{k};\mathbf{x}_{s})}\bigg{(}H(\mathbf{x},\omega_{k};\mathbf{x}_{s})-\sqrt{-\frac{2\omega_{0}}{\mathrm{i}\omega_{k}}}\int_{0}^{T_{\max}}dt^{\prime}H^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})e^{-\sqrt{\omega_{k}\omega_{0}}t^{\prime}}e^{\mathrm{i}\sqrt{\omega_{k}\omega_{0}}t^{\prime}}\bigg{)}\\\ +&\sum_{s}\int_{0}^{T_{\max}}dt^{\prime}\int_{X}\mathrm{d}\mathbf{x}\underline{H}^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})\cdot\bigg{(}\nabla\times E^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})+\mu^{\prime}\partial_{t}H^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})-M^{\prime}(\mathbf{x}_{s},t^{\prime})\bigg{)}\\\ +&\sum_{s}\int_{0}^{T_{\max}}dt^{\prime}\int_{X}\mathrm{d}\mathbf{x}\underline{E}^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})\cdot\bigg{(}\nabla\times H^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})-\epsilon\partial_{t}E^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})-J^{\prime}(\mathbf{x}_{s},t^{\prime})\bigg{)},\end{split}$ (34) where $\lambda_{1}(\mathbf{x},\omega;\mathbf{x}_{s})$ and $\lambda_{2}(\mathbf{x},\omega;\mathbf{x}_{s})$ are Lagrangian multipliers in the frequency domain, while $\underline{E}^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})$ and $\underline{H}^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})$ are Lagrangian multipliers in the fictitious time domain. Setting the Lagrangian with respect to state variables to zero ($\partial\mathcal{L}/\partial\lambda_{1}=\partial\mathcal{L}/\partial\lambda_{2}=\partial\mathcal{L}/\partial\underline{E}^{\prime}=\partial\mathcal{L}/\underline{H}^{\prime}=0$) gives exactly the state equations. The adjoint equations are obtained by setting the Lagrangian with respect to state variables, i.e., $\displaystyle\frac{\partial\mathcal{L}}{\partial E(\mathbf{x},\omega_{k};\mathbf{x}_{s})}=0$ $\displaystyle\Leftrightarrow\lambda_{1}(\mathbf{x},\omega_{k};\mathbf{x}_{s})=R^{\dagger}\mathbf{W}_{1}^{\dagger}\mathbf{W}_{1}(E^{obs}(\mathbf{x}_{r},\omega_{k})-RE(\mathbf{x},\omega_{k};\mathbf{x}_{s})),$ (35a) $\displaystyle\frac{\partial\mathcal{L}}{\partial H(\mathbf{x},\omega_{k};\mathbf{x}_{s})}=0$ $\displaystyle\Leftrightarrow\lambda_{2}(\mathbf{x},\omega_{k};\mathbf{x}_{s})=R^{\dagger}\mathbf{W}_{2}^{\dagger}\mathbf{W}_{2}(H^{obs}(\mathbf{x}_{r},\omega_{k};\mathbf{x}_{s})-RH(\mathbf{x},\omega_{k};\mathbf{x}_{s})),$ (35b) $\displaystyle\frac{\partial\mathcal{L}}{\partial E^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})}=0$ $\displaystyle\Leftrightarrow\nabla\times\underline{H}^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})+\epsilon\partial_{t}\underline{E}^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})=\Re\sum_{k}\overline{\lambda_{1}(\mathbf{x},\omega_{k};\mathbf{x}_{s})}e^{-\sqrt{\omega_{k}\omega_{0}}t^{\prime}}e^{\mathrm{i}\sqrt{\omega_{k}\omega_{0}}t^{\prime}},$ (35c) $\displaystyle\frac{\partial\mathcal{L}}{\partial H^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})}=0$ $\displaystyle\Leftrightarrow-\mu\partial_{t}\underline{H}^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})+\nabla\times\underline{E}^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})=\Re\sum_{k}\sqrt{-\frac{2\omega_{0}}{\mathrm{i}\omega_{k}}}\overline{\lambda_{2}(\mathbf{x},\omega_{k};\mathbf{x}_{s})}e^{-\sqrt{\omega_{k}\omega_{0}}t^{\prime}}e^{\mathrm{i}\sqrt{\omega_{k}\omega_{0}}t^{\prime}}.$ (35d) Equations (35a) and (35b) reveals that $\lambda_{1}$ and $\lambda_{2}$ are nothing more than the electric and magnetic components of the weighted data residual, while equations (35c) and (35d) form a new set of Maxwell system. It is important to remark that in deriving the above equations, we have tacitly applied integration by parts in time and space, assuming zero initial condition of the forward EM fields $E^{\prime}(\mathbf{x},t^{\prime}=0)=H^{\prime}(\mathbf{x},t^{\prime}=0)=0,\quad\mathbf{x}\in X,$ (36) zero final conditions of the adjoint fields $\underline{E}^{\prime}(\mathbf{x},t^{\prime}=T_{\max})=\underline{H}^{\prime}(\mathbf{x},t^{\prime}=T_{\max})=0,\quad\mathbf{x}\in X,$ (37) and homogeneous boundary condition in space for both forward and adjoint fields $E^{\prime}(\mathbf{x},\cdot)=H^{\prime}(\mathbf{x},\cdot)=\underline{E}^{\prime}(\mathbf{x},\cdot)=\underline{H}^{\prime}(\mathbf{x},\cdot)=0,\quad\mathbf{x}\in\partial X.$ (38) At the saddle point, we again obtain $\frac{\partial\phi_{d}}{\partial\epsilon}=\frac{\partial\mathcal{L}}{\partial\epsilon}=-\sum_{s}\int_{0}^{T_{\max}}\mathrm{d}t^{\prime}\underline{E}^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s})\cdot\partial_{t}E^{\prime}(\mathbf{x},t^{\prime};\mathbf{x}_{s}).$ (39) Since $\sigma_{ij}=2\omega_{0}\epsilon_{ij}$, the application of the chain rule yields $\frac{\partial\phi_{d}}{\partial\sigma_{ij}}=\frac{\partial\phi_{d}}{\partial\epsilon_{ij}}\frac{\partial\epsilon_{ij}}{\partial\sigma_{ij}}=-\sum_{s}\frac{1}{2\omega_{0}}\int_{0}^{T_{\max}}\mathrm{d}t^{\prime}\underline{E}^{\prime}_{j}(\mathbf{x},t^{\prime};\mathbf{x}_{s})\cdot\partial_{t}E^{\prime}_{i}(\mathbf{x},t^{\prime};\mathbf{x}_{s}).$ (40) With the gradient at hand, one can then construct the descent direction based on nonlinear minimization methods to solve the inverse problem. ## References * Abubakar et al. (2008) Abubakar, A., Habashy, T., Druskin, V., Knizhnerman, L., Alumbaugh, D., 2008. 2.5 D forward and inverse modeling for interpreting low-frequency electromagnetic measurements. Geophysics 73, F165–F177. * Alumbaugh and Newman (1997) Alumbaugh, D., Newman, G., 1997\. Three-dimensional massively parallel electromagnetic inversion-II. Analysis of a crosswell electromagnetic experiment. Geophysical Journal International 128, 355–363. * Björck (1996) Björck, Å., 1996. Numerical methods for least squares problems. SIAM, Society for Industrial and Applied Mathematics, Philadelphia. * Chang-Chun et al. (2015) Chang-Chun, Y., Xiu-Yan, R., Yun-He, L., Yan-Fu, Q., Chang-Kai, Q., Jing, C., 2015\. Review on airborne electromagnetic inverse theory and applications. Geophysics 80, W17–W31. * Chave and Cox (1982) Chave, A.D., Cox, C.S., 1982\. Controlled Electromagnetic Sources for Measuring Electrical Conductivity Beneath the Oceans 1. Forward Problem and Model Study. Journal of Geophysical Research 87, 5327–5338. * Claerbout and Fomel (2008) Claerbout, J.F., Fomel, S., 2008\. Image estimation by example: geophysical soundings image construction: multidimensional autoregression. Stanford University. * Commer and Newman (2008) Commer, M., Newman, G.A., 2008\. New advances in three-dimensional controlled-source electromagnetic inversion. Geophysical Journal International 172, 513–535. * Constable (2010) Constable, S., 2010. Ten years of marine CSEM for hydrocarbon exploration. Geophysics 75, 75A67–75A81. * Constable and Srnka (2007) Constable, S., Srnka, L.J., 2007\. An introduction to marine controlled-source electromagnetic methods for hydrocarbon exploration. Geophysics 72, WA3–WA12. * Constable et al. (1986) Constable, S.C., Cox, C.S., Chave, A.D., 1986. Offshore electro-magnetic surveying techniques, in: 56th Annual International Meeting, SEG, Expanded Abstracts, Society of Exploration Geophysicists. pp. 81–82. * Constable et al. (1987) Constable, S.C., Parker, R.L., Constable, C.G., 1987. Occam’s inversion: A practical algorithm for generating smooth models from electromagnetic sounding data. Geophysics 52, 289–300. * Eidesmo et al. (2002) Eidesmo, T., Ellingsrud, S., MacGregor, L., Constable, S., Sinha, M., Johansen, S., Kong, F., Westerdahl, H., 2002\. Sea bed logging (SBL), a new method for remote and direct identification of hydrocarbon filled layers in deepwater areas. First break 20. * Ellingsrud et al. (2002) Ellingsrud, S., Eidesmo, T., Johansen, S., Sinha, M., MacGregor, L., Constable, S., 2002\. Remote sensing of hydrocarbon layers by seabed logging (SBL): Results from a cruise offshore Angola. The Leading Edge 21, 972–982. * Grayver et al. (2013) Grayver, A.V., Streich, R., Ritter, O., 2013. Three-dimensional parallel distributed inversion of CSEM data using a direct forward solver. Geophysical Journal International 193, 1432–1446. * Grayver et al. (2014) Grayver, A.V., Streich, R., Ritter, O., 2014. 3D inversion and resolution analysis of land-based CSEM data from the Ketzin CO2 storage formation. Geophysics 79, E101–E114. * Gribenko and Zhdanov (2007) Gribenko, A., Zhdanov, M., 2007\. Rigorous 3D inversion of marine CSEM data based on the integral equation method. Geophysics 72, WA73–WA84. * Key (2016) Key, K., 2016. MARE2DEM: a 2-D inversion code for controlled-source electromagnetic and magnetotelluric data. Geophysical Journal International 207, 571–588. * Komatitsch and Martin (2007) Komatitsch, D., Martin, R., 2007\. An unsplit convolutional perfectly matched layer improved at grazing incidence for the seismic wave equation. Geophysics 72, SM155–SM167. * Lee et al. (1989) Lee, K.H., Liu, G., Morrison, H., 1989. A new approach to modeling the electromagnetic response of conductive media. Geophysics 54, 1180–1192. * Lelièvre and Farquharson (2013) Lelièvre, P.G., Farquharson, C.G., 2013\. Gradient and smoothness regularization operators for geophysical inversion on unstructured meshes. Geophysical Journal International 195, 330–341. URL: https://doi.org/10.1093/gji/ggt255, doi:10.1093/gji/ggt255. * Li and Key (2007) Li, Y., Key, K., 2007. 2D marine controlled-source electromagnetic modeling: Part 1—an adaptive finite-element algorithm. Geophysics 72, WA51–WA62. * Maaø (2007) Maaø, F., 2007. Fast finite-difference time-domain modeling for marine subsurface electromagnetic problems. Geophysics 72, A19–A23. * MacGregor et al. (2007) MacGregor, L., Barker, N., Overton, A., Moody, S., Bodecott, D., 2007. Derisking exploration prospects using integrated seismic and electromagnetic data-A Falkland Islands case study. The Leading Edge 26, 356–359. * MacGregor and Tomlinson (2014) MacGregor, L., Tomlinson, J., 2014\. Marine controlled-source electromagnetic methods in the hydrocarbon industry: A tutorial on method and practice. Interpretation 2, SH13–SH32. * Mittet (2010) Mittet, R., 2010. High-order finite-difference simulations of marine CSEM surveys using a correspondence principle for wave and diffusion fields. Geophysics 75, F33–F50. * Mittet and Morten (2012) Mittet, R., Morten, J.P., 2012\. Detection and imaging sensitivity of the marine CSEM method. Geophysics 77, E411–E425. * Morten et al. (2009) Morten, J.P., Bjørke, A.K., Støren, T., 2009. CSEM data uncertainty analysis for 3D inversion, in: SEG Technical Program Expanded Abstracts 2009, Society of Exploration Geophysicists. pp. 724–728. * Mulder (2006) Mulder, W., 2006. A multigrid solver for 3D electromagnetic diffusion. Geophysical prospecting 54, 633–649. * Newman and Alumbaugh (1995) Newman, G.A., Alumbaugh, D.L., 1995\. Frequency-domain modelling of airborne electromagnetic responses using staggered finite differences. Geophysical Prospecting 43, 1021–1042. * Nocedal and Wright (2006) Nocedal, J., Wright, S.J., 2006\. Numerical Optimization. 2nd ed., Springer. * Oristaglio and Hohmann (1984) Oristaglio, M.L., Hohmann, G.W., 1984\. Diffusion of electromagnetic fields into a two-dimensional earth: A finite-difference approach. Geophysics 49, 870–894. * Paige and Saunders (1982) Paige, C.C., Saunders, M.A., 1982\. ALGORITHM 583 LSQR : Sparse linear equations and least squares problems. ACM Transactions on Mathematical Software 8, 195–209. * Plessix and Mulder (2008) Plessix, R.E., Mulder, W.A., 2008\. Resistivity imaging with controlled-source electromagnetic data: depth and data weighting. Inverse Problems 24, 034012\. * Puzyrev et al. (2013) Puzyrev, V., Koldan, J., de la Puente, J., Houzeaux, G., Vázquez, M., Cela, J.M., 2013\. A parallel finite-element method for three-dimensional controlled-source electromagnetic forward modelling. Geophysical Journal International 193, 678–693. * Rochlitz et al. (2019) Rochlitz, R., Skibbe, N., Günther, T., 2019. custEM: Customizable finite-element simulation of complex controlled-source electromagnetic data. Geophysics 84, F17–F33. * Saad (2003) Saad, Y., 2003. Iterative Methods for Sparse Linear Systems. SIAM, Philadelphia. * Schwarzbach and Haber (2013) Schwarzbach, C., Haber, E., 2013\. Finite element based inversion for time-harmonic electromagnetic problems. Geophysical Journal International 193, 615–634. * Shantsev et al. (2020) Shantsev, D.V., Nerland, E.A., Gelius, L.J., 2020. Time-lapse CSEM: how important is survey repeatability? Geophysical Journal International 223, 2133–2147. * da Silva et al. (2012) da Silva, N.V., Morgan, J.V., MacGregor, L., Warner, M., 2012\. A finite element multifrontal method for 3D CSEM modeling in the frequency domain. Geophysics 77, E101–E115. * Sirgue et al. (2008) Sirgue, L., Etgen, J.T., Albertin, U., 2008. 3D Frequency Domain Waveform Inversion using Time Domain Finite Difference Methods, in: Proceedings 70th EAGE, Conference and Exhibition, Roma, Italy, p. F022. * Sirgue et al. (2010) Sirgue, L., Etgen, J.T., Albertin, U., Brandsberg-Dahl, S., 2010\. System and method for 3D frequency domain waveform inversion based on 3D time-domain forward modeling. US Patent 7,725,266. * Smith (1996a) Smith, J.T., 1996a. Conservative modeling of 3-D electromagnetic fields, Part i: Properties and error analysis. Geophysics 61, 1308–1318. * Smith (1996b) Smith, J.T., 1996b. Conservative modeling of 3-D electromagnetic fields, Part II: Biconjugate gradient solution and an accelerator. Geophysics 61, 1319–1324. * Støren et al. (2008) Støren, T., Zach, J.J., Maaø, F.A., 2008. Gradient calculations for 3D inversion of CSEM data using a fast finite-difference time-domain modelling code, in: 70th EAGE Conference and Exhibition incorporating SPE EUROPEC 2008. * Streich (2009) Streich, R., 2009. 3D finite-difference frequency-domain modeling of controlled-source electromagnetic data: Direct solution and optimization for high accuracy. Geophysics 74, F95–F105. * Symes (2007) Symes, W.W., 2007. Reverse time migration with optimal checkpointing. Geophysics 72, SM213–SM221. URL: http://link.aip.org/link/?GPY/72/SM213/1, doi:10.1190/1.2742686. * Taflove and Hagness (2005) Taflove, A., Hagness, S.C., 2005\. Computational Electrodynamics: The Finite-Difference Time-Domain Method. 3rd ed., Artech House. * Wang and Hohmann (1993) Wang, T., Hohmann, G.W., 1993\. A finite-difference, time-domain solution for three-dimensional electromagnetic modeling. Geophysics 58, 797–809. * Ward and Hohmann (1988) Ward, S.H., Hohmann, G.W., 1988\. Electromagnetic theory for geophysical applications, in: Electromagnetic Methods in Applied Geophysics: Volume 1, Theory. Society of Exploration Geophysicists, pp. 130–311. * Yang (2023) Yang, P., 2023. libEMM: A fictious wave domain 3D CSEM modelling library bridging sequential and parallel GPU implementation. Computer Physics Communications 288, 108745. doi:https://doi.org/10.1016/j.cpc.2023.108745. * Yang et al. (2016a) Yang, P., Brossier, R., Métivier, L., Virieux, J., 2016a. Wavefield reconstruction in attenuating media: A checkpointing-assisted reverse-forward simulation method. Geophysics 81, R349–R362. doi:10.1190/geo2016-0082.1. * Yang et al. (2016b) Yang, P., Brossier, R., Virieux, J., 2016b. Wavefield reconstruction from significantly decimated boundaries. Geophysics 80, T197–T209. doi:10.1190/GEO2015-0711.1. * Yang and Mittet (2023) Yang, P., Mittet, R., 2023. Controlled-source electromagnetics modelling using high order finite-difference time-domain method on a nonuniform grid. Geophysics 88, E53–E67. doi:10.1190/geo2022-0134.1. * Zaslavsky et al. (2013) Zaslavsky, M., Druskin, V., Abubakar, A., Habashy, T., Simoncini, V., 2013. Large-scale gauss-newton inversion of transient csem data using the model order reduction framework. Geophysics 78, E161–E171. * Zhdanov and Keller (1994) Zhdanov, M.S., Keller, G.V., 1994\. The geoelectrical methods in geophysical exploration. volume 31.
# FedGPO: Heterogeneity-Aware Global Parameter Optimization for Efficient Federated Learning Young Geun Kim Carole-Jean Wu Korea University ASU / Meta <EMAIL_ADDRESS><EMAIL_ADDRESS> ###### Abstract Federated learning (FL) has emerged as a solution to deal with the risk of privacy leaks in machine learning training. This approach allows a variety of mobile devices to collaboratively train a machine learning model without sharing the raw on-device training data with the cloud. However, efficient edge deployment of FL is challenging because of the system/data heterogeneity and runtime variance. This paper optimizes the energy-efficiency of FL use cases while guaranteeing model convergence, by accounting for the aforementioned challenges. We propose FedGPO based on a reinforcement learning, which learns how to identify optimal global parameters (B, E, K) for each FL aggregation round adapting to the system/data heterogeneity and stochastic runtime variance. In our experiments, FedGPO improves the model convergence time by 2.4 times, and achieves 3.6 times higher energy efficiency over the baseline settings, respectively. ## 1 Introduction Federated learning (FL) has recently emerged as a practical framework for training machine learning (ML) models on a large variety of mobile devices privately [4, 36, 51, 24]. A shared ML model is trained over E epochs with a minibatch size of B on K selected devices, where K is a subset of N client devices participating in the FL task. The K devices then upload the respective model gradients (or trained model parameters) to the cloud, in order to update the global model while keeping all the raw data on the device. Thus, FL deals with the risk of privacy leaks for deep neural network (DNN) training. While many existing approaches have been proposed to deploy FL efficiently [51, 11, 6, 42, 35, 36, 26, 69], a fundamental challenge remains — setting global parameters (B, E, K) round-by-round to ensure efficient edge execution. These global parameters significantly affect the model quality and convergence time, as they directly determine the amount of data reflected on the model gradients [29]. Moreover, they also affect the energy efficiency of participant devices, because the amount of training computation on each client device depends on the parameter settings [35]. Therefore, to achieve efficient FL execution at the edge, finding optimal global parameter settings is crucial. Hyperparameter optimization (HPO) has been extensively studied for the centralized training. Common approaches include grid- and genetic-based searches [1]. ML-based methods, such as Bayesian Optimization [66, 25], are also applicable. Typically, ML-based HPO methods tune the hyperparameters using the accuracy results obtained from iterative DNN training on the entire dataset. Since it is infeasible to train a model on the entire dataset for each set of global parameters int the resource-constrained edge execution environment, tuning the global parameters round-by-round has been considered as a common practice for FL [49, 29]. However, round-by-round global parameter tuning is still challenging due to the following unique aspects of FL: * • System Heterogeneity and Runtime Variance. There exist a variety types of system-on-chips (SoCs) with distinct computing performance at the edge [70], which results in large performance gaps across participating devices. Furthermore, stochastic runtime variance, including on-device interference [63] and network stability [30], can even exacerbate the performance variability [15, 16] across the devices. This results in the straggler problem, where the training time per aggregation round is determined by the slowest device, making it difficult to find the optimal global parameters for each round. * • Data Heterogeneity. For model convergence, ensuring training data are independently and identically distributed (IID) for each and every participating device is crucial. However, in edge execution environment, client training data are not guaranteed to be non-IID, as training samples of an individual user are often not representative of the entire population [5, 50]. The inclusion of non-IID data in training can defer model convergence [44, 51]. Since global parameters influence the degree of non-IID data reflected in the model gradients; it is also crucial to adjust the parameters considering data heterogeneity. The high degree of heterogeneity and runtime variance makes it challenging to optimize global parameters using offline, server-based simulations. Static simulation-based optimization studies cannot adapt to dynamic system/data heterogeneity or runtime variance. To identify efficient global parameters for each FL training round under system/data heterogeneity, several approaches have been proposed recently [49, 29]. However, these prior approaches do not consider the stochastic nature of edge computing, including performance interference and network variability. In addition, prior work did not consider optimizing the energy efficiency of FL, which can lead to increasing energy footprint at scale [71]. To the best of our knowledge, this is the first work to tackle energy-efficient global parameter optimization for FL. This paper proposes an FL global parameter optimization framework based on reinforcement learning — FedGPO — that dynamically adjusts the global parameters (B, E, K) to maximize the FL energy efficiency guaranteeing model quality. The optimization is performed round-by-round over the entire training process, considering system and data heterogeneity, as well as runtime variance. Since the optimal (B, E, K) varies with the computation characteristics of neural networks (NN), performance profiles of participant device systems, local training sample distributions, and runtime variance, the enormous design space makes it difficult to enumerate exhaustively. Hence, we propose a technique based on a reinforcement learning to addres this optimization formulation. FedGPO identifies the characteristics of NNs and profiles of devices such as the intensity on-device interference, network instability, and the distributions of data samples every aggregation round. It then determines the global parameters for the round, which maximizes energy efficiency while not deteriorating the training accuracy. Based on the result of the decision, FedGPO continuously learns and predicts the efficient global parameters. The key contributions of this work are as follows: * • We present performance and energy efficiency characterization for the FL global parameter design space. The characterization results show that optimal settings vary across the FL training rounds due to the varying level of the system/data heterogeneity and the stochastic runtime variance (Section 2). * • We propose a global parameter optimization framework, FedGPO, that identifies the near-optimal global parameter setting for each round, enabling energy- efficient federated learning (Section 3). * • We implement and evaluate FedGPO for FL use cases with 200 mobile devices encompassing three performance categories: high, medium, and low (Section 5). Real system-based experiments demonstrate that FedGPO improves the energy efficiency of the participant devices by 3.6x, while satisfying the accuracy requirements. ## 2 Background and Motivation ### 2.1 Global Impact of FL Parameters Algorithm 1 FedAvg Variable: B, E, K B is the local minibatch size E is the number of local epochs K is the number of participant devices Constants: N, $\eta$ N is the number of entire devices $\eta$ is the learning rate Server executes: initialize $(B,E,K)$ initialize $w_{0}$ for each round t = 1,2,… do $S_{t}$ $\leftarrow$ (random set of K clients among N clients) for each client k $\in$ $S_{t}$ in parallel do $w^{k}_{t+1}$ $\leftarrow$ ClientUpdate($k$, $w_{t}$) $w_{t+1}$ $\leftarrow$ $\sum_{k=1}^{K}\frac{n_{k}}{n}w_{t+1}^{k}$ ClientUpdate(k, w): // Run on client k $B$ $\leftarrow$ (split $P_{k}$ into batches of size B for each local epoch i from 1 to E do for batch $b\in B$ do $w$ $\leftarrow$ $w-\eta\nabla\ell(w;b)$ return w to server To prevent privacy leaks in ML training, FL is proposed, where edge devices train a shared global model collaboratively without sharing the on-device data samples with the cloud [4, 36, 51]. FedAvg is the de-facto FL algorithm [36, 51] (Algorithm 1). For N devices, the server initializes a global model along with the number of local training epochs E, the local training minibatch size B, and the number of participant devices K. It also initializes the model parameters $w_{0}$. In every round t, the server randomly selects K devices among the N devices and transmits the global model to them. Each selected device trains the model locally using the on-device data with a batch size of B over E epochs. After the local execution of the training is finished, each device transmits the model parameters back to the server. The server then updates the global model with the average of local parameters. Global parameters (B, E, K) significantly impact the FL model convergence and energy efficiency. It is therefore crucial to carefully select the parameters for better model quality and the energy efficiency of participating devices. Figure 1 shows the (a) convergence time and (b) global performance per watt (PPW) of a CNN model with the MNIST dataset (CNN-MNIST) [38, 67] for varying FL settings of (B, E, K). B determines the number of local data samples used for one iteration of training on each device. Typically, B is associated with the generalization problem — using larger batch sizes usually yield poor generalizability [64, 21]. For this reason, B largely affects the model convergence and global energy efficiency, as shown in Figure 1. E represents the number of training iterations for each device with the same data samples. Since E is related to the over- versus under-fitting to specific data samples [51], it also has a global impact on the model convergence, as shown in Figure 1. The number of participant devices K can be considered as the global batch size in FL, as it is related to the amount of global data used per round. Although smaller values of K enable efficient FL deployment by reducing the impact of communication overhead [51], a careful selection is still required — K is also associated with the generalization problem [51], affecting the model convergence and global energy efficiency as shown in Figure 1. Figure 1: Depending on the global parameters, the FL convergence performance and global energy efficiency vary significantly. The convergence round and global PPW are normalized to (1, 10, 20). Figure 2: The most energy-efficient global parameter combination shifts in accordance with the NN characteristics. The convergence round and global PPW are normalized to (1, 10, 20). It is also important to consider the NN characteristics when selecting the global parameters, as the two are interrelated. Figure 2 shows the global PPW of two NNs under the different FL settings of (B, E, K). In case of CNN-MNIST [38, 67], (B, E, K) of (8, 10, 20) shows the best energy efficiency among the selected global parameter combinations. In contrast, when we use the LSTM model with the Shakespeare dataset (LSTM-Shakespeare) [51, 36], the best energy efficient global parameter combination shifts to (4, 20, 20) as the learning characteristics of LSTM-Shakespeare differ from those of CNN-MNIST. Furthermore, LSTM-Shakespeare comprises more memory-intensive RC layers, whereas CNN-MNIST mainly consists of computation-intensive convolutional and fully-connected layers. Owing to the memory pressure, LSTM-Shakespeare exhibits higher energy efficiency with smaller input batch sizes and more iterations. ### 2.2 FL Global Parameter Optimization Figure 3: Training time per round of devices significantly varies with (a) B and (b) E introducing large performance gaps across the devices. The training time per round is normalized to that of H with B of 1 and that of H with E of 10 for (a) and (b), respectively. In traditional centralized training, hyperparameter optimization (HPO) is the process of finding a set of hyperparameters that minimizes loss or maximizes accuracy for a DNN [66]. Generally, HPO first randomly selects a set of hyperparameters, and trains the DNN using the entire dataset. The training results (i.e., loss and accuracy) obtained with the selected hyperparameters are measured, and used to select the next set of hyperparameters. Since the hyperparameters can be any integer or floating point values, their search space is usually large. Hence, to efficiently explore the search space, machine learning-based optimization techniques, such as Bayesian Optimization (BO) or Tree Parzan Estimator (TPE), are widely employed] for the HPO process. Despite a variety of previous works on centralized training, global parameter optimization in FL has unique challenges: Straggler Problem: For each device, the training time of each device substantially varies with the global parameters, introducing the straggler problem — the overall training time per round is determined by the slowest device. Figure 3 illustrates the training time per round for different device categories (i.e., H, M, and L for high-end, mid-end, and low-end devices, respectively111A detailed specification of each device category is presented in Section 4.1.), depending on different B and E values. As B determines the amount of on-device data to be processed in each iteration, training time on each device category significantly depends on its computation- and memory- capabilities. In addition, as E determines the number of iterations, it has a linear impact on the training time per round. Figure 4: Runtime variance significantly affects computation and communication time exacerbating the straggler problem. The training time per round is normalized to that of H in the absence of runtime variance. Stochastic runtime variance exacerbates the straggler problem. In a real use case, there can be several applications co-running with the FL execution, since modern mobile device support multi-tasking features [62]. This causes on-device shared resource interference [31, 34, 40] degrading computation performance of FL. In addition, signal strength variations in wireless network can affect the performance and energy efficiency of global aggregations in FL — the data transmission latency and energy increase exponentially at weak signal strength [33, 12]. Figure 4 shows the training time per round on different device categories, (a) when there is no runtime variance, (b) when there is on-device interference, and (c) when the network is not stable. As shown in Figure 4(a) and (b), the on-device interference deteriorates the computation time for each device. Since the impact of interference depends on the capabilities of each device in terms of computation and memory, it exacerbates the inter-device performance gaps. Further, as shown in Figure 4(a) and (c), the network instability deteriorates the communication time of each device, thus affecting the percentage of performance gaps across the devices. Figure 5: The adaptive adjustment of global parameters resolves the straggler problem saving energy consumption of each device category. The energy consumption is normalized to H with fixed parameters. Figure 6: Adaptive parameters can improve global PPW by resolving the straggler problem while guaranteeing model convergence. The adaptive round-by-round adjustment of the global parameters for different devices can resolve the straggler problem, improving the FL energy efficiency. Figure 5 shows the energy consumption of each device (a) when using the same fixed parameters and (b) when adaptively adjusting the parameters for different devices round-by-round. In the former case, faster devices (e.g., H and M) need to wait for the slower devices (e.g., L) consuming energy, as shown in Figure 5(a). Using smaller B or E for the slower devices can reduce the performance gaps across the devices, saving the energy as shown in Figure 5(b). This significantly improves the average training time per round (2.3x) and global PPW (3.6x), as shown in Figure 6(b) and (c) respectively — a careful adjustment of the parameters can still guarantee the model convergence as shown in Figure 6(a). Figure 7: The optimal global parameters shift depending on the presence of data heterogeneity. Data Heterogeneity: It is also crucial for parameter optimization strategies to consider the impact of data heterogeneity. Figure 7(a) shows the global PPW over different global parameter (B, E, K) setting, in the absence of data heterogeneity. In this case, the most energy-efficient (B, E, K) is (8, 10, 20). In the presence of data heterogeneity, however the global energy efficiency of all (B, E, K) is degraded as shown in Figure 7(b), since the data heterogeneity significantly affects model convergence [35]. In this case, the most energy-efficient (B, E, K) shifts to (8, 5, 10), as decreasing E or K reduces the amount of non-IID data reflected to the model parameters — E affects the number of iterations for parameter updates with the given data and K affects the number of non-IID devices participating for gradient updates. Since the degree of data heterogeneity can vary round-by-round depending on the participant compositions, it is also important to carefully tune the global parameters round-by-round taking into account the data heterogeneity. ## 3 FedGPO Figure 8: FedGPO Design Overview. Under the heterogeneity and runtime variance, it is infeasible to enumerate the large search space associated with FL global parameter optimization. To efficiently explore the optimization space and accurately predict the optimal global parameters, we propose an approach called FedGPO, based on reinforcement learning (RL). Due to its low complexity yet high sample efficiency [28, 73], RL has been widely used for the system optimization in the edge domain [58]. In the following sections, we first provide an overview of the FedGPO design, and then elaborate on its RL design and algorithm. ### 3.1 Overview Figure 8 presents an overview of the FedGPO design. Based on RL, FedGPO attempts to learn an optimal action decision (i.e., global parameter selection) from prior information based on the current state (i.e., heterogeneity and runtime variance) and the given reward (i.e., the amount of global/local energy efficiency and model accuracy improvement of the selected action). In each aggregation round, FedGPO identifies the global execution states of FL (\small{1}⃝), such as characteristics of neural network model architectures and the composition of randomly-selected K’ participant devices. Note K’ is K determined in the previous aggregation round. FedGPO then identifies the local execution states of the selected devices (\small{1}⃝), such as the usage of resources, network instability, and the number of data classes each device has. With the identified state information, FedGPO selects the action (\small{2}⃝), i.e., sets per-device global parameters expected to improve the FL energy efficiency without degrading the model convergence and accuracy. It selects the actions using lookup tables (i.e., Q-tables shared across the devices in the same performance category) [14, 37, 58], which store the accumulated rewards of previously selected parameter combinations. Using the selected global parameters, FedGPO executes the training on each selected device (\small{3}⃝). After the aggregation round ends, FedGPO measures its result (i.e., training time, energy consumption, and test accuracy) for calculating the reward (\small{4}⃝). Finally, FedGPO updates the Q-tables with the calculated reward (\small{5}⃝). By repeating the aforementioned process (\small{1}⃝-\small{5}⃝), FedGPO learns how to select the optimal global parameters. ### 3.2 FedGPO RL Design Table 1: Discrete values for FedGPO states. State | Discrete Values ---|--- $S_{CONV}$ | Small ($<$10), medium ($<$20), large ($<$30), larger ($>=$40) $S_{FC}$ | Small ($<$10), large ($>=$10) $S_{RC}$ | Small ($<$5), medium ($<$10), large ($>=$10) $S_{Co\\_CPU}$ | None (0%), small ($<$25%), medium ($<$75%), large ($<=$100%) $S_{Co\\_MEM}$ | None (0%), small ($<$25%), medium ($<$75%), large ($<=$100%) $S_{Network}$ | Regular ($>$40Mbps), bad ($<=$40Mbps) $S_{Data}$ | Small ($<$25%), medium ($<$100%), large (=100%) To produce accurate predictions of RL, it is crucial to model the core components in a realistic execution environment: 1) state, 2) action, and 3) reward. This section defines the core components for system energy efficiency optimization of FL. State: We define the FL execution state based on our observations in Section 2. As demonstrated in Section 2.1, the optimal global parameters depend on the NN characteristics. To model the impact of these characteristics, we identify $S_{CONV}$, $S_{FC}$, and $S_{RC}$ which represent the numbers of convolutional, fully-connected, and recurrent layers, respectively — the three layers are typically highly interrelated with the training/inference efficiency [35]. In addition, as shown in Section 2.2, the optimal global parameters are also affected by on-device interference. To model this impact, we identify per- device states of $S_{Co\\_CPU}$ and $S_{Co\\_MEM}$ which represent the CPU utilization and memory usage of co-running applications, respectively. Because the optimal global parameters also depend on network stability, we identify the per-device network stability of $S_{Network}$ with the wireless network (e.g., Wi-Fi and 5G) bandwidth. We also model the data heterogeneity impact by identifying $S_{Data}$ which represents the number of data classes of each device for the aggregation round. It is difficult to encode continuous values into the RL lookup table. Thus, we convert continuous values of each state into discrete values, by applying a clustering algorithm [9, 35]. Table 1 summarizes the discretized values. Note FedGPO can support larger search space by further reducing the search space size with different clustering algorithms. Action: Actions in RL model the customizable control knobs of the system. In the context of FL global parameter optimization, we define the actions as the selection of global parameters for each aggregation round. For the local mini- batch size of B, we define the discrete batch size numbers as a feasible range for resource-constrained edge devices [51, 35]. We also define the discrete numbers of local epochs E and participant devices K as ranges based on the de- facto FL algorithm [51] to ensure a balanced computation-communication ratio. For example, larger E with smaller K typically increases the computation- communication ratio. Table 2 summarizes the discrete values of all global parameters as actions. Table 2: Discrete global parameter values for FedGPO actions. Parameter | Discrete Values ---|--- B | {1, 2, 4, 8, 16, 32} E | {1, 5, 10, 15, 20} K | {1, 5, 10, 15, 20} Reward: The reward in RL models the optimization objective. To ensure that FedGPO selects global parameters that maximize energy efficiency without degrading model convergence and accuracy, we define the reward R as in (1). Here, $R_{energy\\_local}$ represents the energy consumption of each participant device, whereas $R_{energy\\_global}$ denotes that of all devices. $R_{accuracy}$ is the test accuracy of the NN model, while $R_{accuracy\\_prev}$ is that of the prior round. Note, since time-to- convergence is not measurable before the convergence, we substitute it with the improvement in accuracy — a similar practice has been used for the hyperparameter optimization of ML training [65]. $\displaystyle if\;\;\;R_{accuracy}\;-\;R_{accuracy\\_prev}\;\;<=\;\;0,$ (1) $\displaystyle\qquad R=R_{accuracy}-100$ $\displaystyle else$ $\displaystyle\qquad R=-R_{energy\\_global}-R_{energy\\_local}$ $\displaystyle\qquad\;\;\;\;\;\;\;\;+\alpha R_{accuracy}+\beta(R_{accuracy}-R_{accuracy\\_prev})$ Among the encoded reward values, we estimate $R_{energy\\_local}$ using a commonly-used energy formulation, such as [27, 32, 30]. Algorithm 2 Training the Q-learning model Variable: S, A S is the variable for the state A is the variable for the action Constants: $\gamma$, $\mu$, $\epsilon$ $\gamma$ is the learning rate $\mu$ is the discount factor $\epsilon$ is the exploration probability Initialize $Q(S,A)$ as random values Repeat (whenever an aggregation round begins): Observe state and store in S if rand() $<\epsilon$ then Choose action A randomly else Choose action A which maximizes Q(S,A) Run training with global parameters defined by A (when local training and aggregation terminate) Obtain $R_{energy\\_global}$, $R_{energy\\_local}$, and $R_{accuracy}$ Calculate reward R Observe new state S’ Choose action A’ which maximizes Q(S’,A’) Q(S,A) $\leftarrow$ Q(S,A) \+ $\gamma$[R \+ $\mu$Q(S’,A’) \- Q(S,A)] S $\leftarrow$ S’ For each device, we calculate computation energy, $E_{comp}$, using utilization-based CPU and GPU power models [27, 32], as in (2). Here, $E^{i}_{CPU\\_core}$ and $E_{GPU}$ represent the energy consumed by the ith CPU core and GPU, respectively, $E_{PU\\_core}$ is the energy consumed by either of the CPU cores or GPU, $t^{f}_{busy}$ is the time spend in the busy state at frequency f and $t_{idle}$ is that in the idle state, and $P^{f}_{busy}$ is the power consumed during $t^{f}_{busy}$ at f and $P_{idle}$ is that during $t_{idle}$. Note $P^{f}_{busy}$ and $P_{idle}$ for CPU/GPU are obtained by measuring the corresponding processing unit’s power at each frequency. $\displaystyle E_{comp}$ $\displaystyle=\sum_{i}{E_{CPU\\_core}^{i}}+E_{GPU},$ (2) $\displaystyle E_{PU\\_core}$ $\displaystyle=\sum_{f}(P_{busy}^{f}\times t_{busy}^{f})+P_{idle}\times t_{idle}$ We also calculate communication energy [30], $E_{comm}$, as in (3), where $t_{TX}$ is the measured transmission latency of the gradient (or parameter) updates, and $P^{S}_{TX}$ is the power consumed by the device during $t_{TX}$ at signal strength S. Note $P^{S}_{TX}$ is obtained based on the measured transmission power consumption of devices at each signal strength. $\displaystyle E_{comm}=$ $\displaystyle P_{TX}^{S}\times t_{TX}$ (3) We then calculate the idle energy, $E_{idle}$, for the rest of devices (i.e., devices not participating in the round), as in (4), where $t_{round}$ denotes the training time of the round. $\displaystyle E_{idle}=$ $\displaystyle P_{idle}\times t_{round}$ (4) Based on $E_{comp}$, $E_{comm}$, and $E_{idle}$, $R_{energy\\_local}$ is calculated for each device, as in (5). $\displaystyle if\;\;device\;\;\subset\;\;S_{t}$ (5) $\displaystyle\qquad R_{energy\\_local}=E_{comp}+E_{comm}$ $\displaystyle else$ $\displaystyle\qquad R_{energy\\_local}=E_{idle}$ Based on $R_{energy\\_local}$, $R_{energy\\_global}$ is also calculated for entire N devices (6). $\displaystyle R_{energy\\_global}=$ $\displaystyle\sum_{i}^{N}R_{energy\\_local}$ (6) ### 3.3 RL Algorithm To enable real-time decision making for each FL round, we employ Q-learning [9], as it provides the advantage of low latency overhead by employing lookup tables to determine the best action. For RL, it is also crucial to consider the balance between exploitation and exploration to avoid the local optima [14, 37, 35]. To overcome this issue, we use the epsilon-greedy algorithm [57, 58] along with the Q-learning, which chooses an action with the highest reward, or a uniformly-random action based on a pre-specified exploration probability ($\epsilon$). In Q-learning, the value function $Q(S,A)$ accepts the state S and the action A as parameters in the form of a lookup table (Q-table). To permit a large number of participants and address the scalability requirement of FL, FedGPO exploits shared Q-tables222Shared Q-tables can be a potential source of system usage leakage. To overcome this, FedGPO can exploit per-device Q-tables instead, which imporves the prediction accuracy by 2.7% degrading 12.2% of convergence overhead. for devices within the same performance category. Sharing the learned results across the devices can also expedite the design space exploration process [35], as each client device experiences different level of heterogeneity and runtime variance. Algorithm 2 presents the algorithm for training the shared Q-tables. FedGPO first randomly initializes the Q-tables. During each aggregation round, it observes the execution state S as identified in Section 3.1. It then generates a random value and compares it with $\epsilon$ 333We use 0.1 for $\epsilon$. To determine $\epsilon$, we tested the accuracy and convergence overhead of FedGPO with 0.1, 0.5, and 0.9 of $\epsilon$.. If the value is lower than $\epsilon$, FedGPO selects the per-device global parameters randomly for exploration. Otherwise, it selects A with the largest $Q(S,A)$ using the Q-tables. After the local training execution in an FL aggregation round is finished using the selected A, FedGPO calculates the reward R as described in Section 3.1. FedGPO then identifies the new execution state S’, and selects the corresponding A’ using $Q(S^{\prime},A^{\prime})$ for each device. It then updates the $Q(S,A)$ based on the equation in Algorithm 2, where $\gamma$ and $\mu$ are hyperparameters of the learning rate and discount factor, respectively — $\gamma$ and $\mu$ determined based on the sensitivity analysis. When the learning phase is completed, i.e., the largest $Q(S,A)$ value is converged for each S, FedGPO uses the shared Q-tables to select A (i.e., global parameters) for each device to maximize $Q(S,A)$ for S. ## 4 Experimental Methodology ### 4.1 System Infrastructure Table 3: System profiles of Amazon EC2 instances. Category | Instance | Performance | RAM ---|---|---|--- (GFLOPS) | (GB) H | m4.large | 153.6 | 8 M | t3a.medium | 80.0 | 4 L | t2.small | 52.8 | 2 We emulate FL with 200 mobile devices referring to prior studies pertaining to FL [42, 51, 35, 36]. Since it is difficult to run experiments with 200 real smartphones, we emulate the FL performance by using Amazon EC2 instances [2] (Table 3), which feature equivalent theoretical GFLOP performance and memory capacity to those of the three smartphone performance categories: high-end (H), mid-end (M), and low-end (L) devices. By referring to in-the-field system performance distribution [70], we composed 200 instances with 30 H, 70 M, and 100 L devices. For the model aggregation server, we use a c5d.24xlarge Amazon EC2 instance, whose theoretical performance and RAM are 448 GFLOPS and 32GB, respectively. For measuring the power, we use three representative smartphones for each performance category [34]: Mi8Pro [23], Galaxy S10e [61], and Moto X Force [55] (Table 4). We use an external Monsoon Power Meter [54] to measure three smartphones’ power consumption while running on-device training (implemented with DL4j [13]) — a similar approach was used in prior studies [59, 63, 35]. Based on the measured performance and power consumption, we evaluate the energy efficiency of participant devices in FL. To evaluate the effectiveness of FedGPO, we implement it on top of the FedAvg algorithm [36, 51] with PyTorch [60]. We compare FedGPO with three baselines: * • Fixed (Best), which uses the most energy-efficient parameter combination identified by grid search, yet fixed during the entire FL rounds. * • Adaptive (BO) which adjusts the global parameters each round using a Bayesian Optimization algorithm where many state-of-the-art approaches are based [66]. * • Adaptive (GA) which adjusts the global parameters every round using a genetic algorithm [1]. We also compare FedGPO with two previous approaches: FedEX [29] and ABS [49]. FedEX adjusts the parameters with an exponentiated gradient updates, whereas ABS only adjusts the local minibatch size with a Deep RL. Table 4: Specifications of mobile devices. Device | CPU | GPU ---|---|--- Mi8Pro (H) | Cortex-A75 (2.8GHz) | Adreno 630 (0.7GHz) 23 V/F steps | 7 V/F steps 5.5 W | 2.8 W Galaxy S10e (M) | Mongoose (2.7GHz) | Mali-G76 (0.7GHz) 21 V/F steps | 9 V/F steps 5.6 W | 2.4 W Moto X Force (L) | Cortex-A57 (1.9GHz) | Adreno 430 (0.6GHz) 15 V/F steps | 6 V/F steps 3.6 W | 2.0 W We determine two hyperparameters (i.e., learning rate and discount factor) of FedGPO by evaluating the three values of 0.1, 0.5, and 0.9 for each one [9, 35]. We find that a higher learning rate improves prediction accuracy, as FedGPO works better when a higher amount of the reward is reflected in the Q-tables — FedGPO needs to adapt to the heterogeneity and stochastic variance within the limited rounds. In contrast, we observe that a lower discount factor improves prediction accuracy, as FedGPO exhibits higher performance when a lower amount of the reward or the following state is reflected in that of the current state — sequential states have a weak mutual relationship because of their stochastic nature. Accordingly, in our evaluation, we use 0.9 and 0.1 for the respective hyperparameter. ### 4.2 Workloads and Execution Scenarios Workloads: We evaluate FedGPO with two workloads [42, 51, 36]: (1) training a CNN model with the MNIST dataset (CNN-MNIST) for image classification [38, 67] and (2) training an LSTM model with the Shakespeare dataset (LSTM-Shakespeare) for the next character prediction [51, 36]. We also use a state-of-the-art NN workload: (3) training the MobileNet with the ImageNet dataset (MobileNet- ImageNet) for image classification [10, 22]. Note our study focuses on mobile- centric neural networks by referring to recent FL deployment and use cases [18, 17, 3, 52, 19, 75, 24], as larger networks have been infeasible for resource-constrained mobile devices [20]. Runtime variance: To emulate realistic on-device interference, we run a synthetic co-running application on a random subset of devices. The synthetic application exhibits the same CPU and memory usage as the real-world mobile application of web browsing [63, 59]. In addition, as the variability of real- world network follows a Gaussian distribution [12, 30], we generate a random network bandwidth following such a distribution using a Wi-Fi AP. Data distribution: We evaluate different degrees of data heterogeneity with two different data sample distributions [5, 44]: Ideal IID and Non-IID. For Ideal IID, all the data classes are evenly distributed to the devices. In contrast, for non-IID, each data class is randomly distributed following a Dirichlet distribution with a 0.1 concentration parameter [35, 5, 41, 44, 47, 39, 56, 72]. ## 5 Evaluation Results and Analysis ### 5.1 Result Overview Figure 9 compares the PPW energy efficient, the convergence performance, and the training accuracy among the three FL applications. The PPW and the convergence time speedup are normalized to the Fixed (Best) case. Compared to the Fixed (Best), Adaptive (BO), and Adaptive (GA), FedGPO achieves 3.6x, 3.1x, and 1.7x of the average FL energy efficiency improvement, respectively. It also maintains the training accuracy with the improved convergence time. For all FL use cases, FedGPO alleviates the performance gaps across the participant devices, by identifying more efficient per-device global parameters compared to the baseline settings. This improves the average training time per round by 2.3x over the Fixed (Best). Additionally, by reducing the performance gaps across the devices, the redundant energy consumption of the devices is saved by 57.5% over the Fixed (Best). As a result, significantly better energy efficiency is achieved compared to the baseline settings. The global parameters selected by FedGPO also guarantees the convergence (i.e., the training loss settles to a certain value [53] while the training accuracy gets to an error range of the value achieved by the baseline in an ideal environment). FedGPO maintains the similar number of convergence rounds with Fixed (Best), improving the convergence time while maintaining the model quality; the convergence round difference of Fixed (Best) and FedGPO is only 0.2%. Compared to Adaptive (BO), FedGPO shows 2.4x better convergence time. This is mainly because the Adaptive (BO) has lower sample efficiency than FedGPO, and thus fails to adapt to the heterogeneity and runtime variance round-by-round. GA has a relatively higher sample efficiency than BO [28], but it still requires a number of mutation/crossover generations for the convergence. Due to the quick adaptability to the hetergeneity and runtime variance every round, FedGPO exhibits 1.6x better convergence time, compared to Adaptive (GA). Figure 9: FedGPO improves the PPW by 4.1x, 3.2x, and 3.5x compared to the baseline Fixed (Best) for CNN-MNIST, LSTM-Shakespeare, and MobileNet-ImageNet, respectively. It also maintains the training accuracy with the improved convergence time. ### 5.2 Adaptability and Accuracy Analysis Adaptability to stochastic variance: Figure 10 compares the PPW, convergence performance, and model accuracy of CNN-MNIST, (a) in the absence of runtime variance, (b) in the presence of on-device interference from co-running applications, and (c) in the presence of network variance. When there exists runtime variance, FedGPO achieves 5.0x, 4.2x, and 3.0x of the average energy efficiency improvement, compared to Fixed (Best), Adaptive (BO), and Adaptive (GA), respectively. Furthermore, it also improves the convergence time while maintaining the training accuracy. Note other NNs show similar result trends. Under the runtime variance, the performance gap across participating devices significantly varies round-by-round due to the varying on-device computation and communication time. Nevertheless, FedGPO selects better per-device global parameters every round compared to the baseline settings, by quickly adapting varying on-device computation or communication time with high sample efficiency. As a result, the convergence time is improved by 3.2x, 2.9x, and 2.5x, compared with that of Fixed (Best), Adaptive (BO), and Adaptive (GA), respectively. Additionally, by eliminating the redundant energy consumption of the participating devices, FedGPO also significantly improves the energy efficiency compared to the baseline settings. Note, in this case, the training accuracy of the baseline setting is significantly degraded due to the exacerbated straggler problems — previous works just drop the gradient updates from the stragglers [41, 42]. Adaptability to data heterogeneity: Figure 11 illustrates the energy efficiency, convergence time, and training accuracy of CNN-MNIST, (a) in the absence of data heterogeneity (i.e., Ideal IID) and (b) in the presence of data heterogeneity (Non-IID). Even under the latter scenario, FedGPO still improves the PPW by 6.2x, 1.9x, and 1.3x, compared with Fixed (Best), Adaptive (BO), and Adaptive (GA), respectively. It also improves convergence time and training accuracy against the baseline settings. Figure 10: Under the runtime variance, FedGPO significantly improves the FL energy efficiency by 5.0x, 4.2x, and 3.0x, on average, compared to Fixed (Best), Adaptive (BO), and Adaptive (GA), respectively. It also improves the convergence time maintaining the training accuracy. In the presence of non-IID participants, neither E nor K is adjusted by the baseline settings depending on the degree of data heterogeneity, maintaining the amount of non-IID data reflected on the model gradients. On the other hand, FedGPO learns how data heterogeneity affects the energy efficiency and convergence performance, and adjusts gradient updates with E and K along with B. Therefore, it significantly improves the PPW, convergence performance, and model accuracy even under the data heterogeneity. The prediction accuracy of the baseline settings is also significantly degraded in this case, as they accept the gradient (or parameter) updates from non-IID participants as equally as those from IID-participants [68]. Prediction accuracy: FedGPO accurately selects the near-optimal global parameters round-by-round. Table 5 lists the mean absolute percentage accuracy of FedGPO over the optimal global parameters for each round — these parameters are identified in terms of minimizing the performance gap across the devices, rather than global convergence. FedGPO achieves an average prediction accuracy of 94.7%. FedGPO also adapts to the stochastic features of edge execution environments. As shown in Table 5, in the presence of runtime variance (i.e., on-device interference and network variability), FedGPO successfully selects the near- optimal global parameters, achieving a 94.4% average prediction accuracy. In the presence of data heterogeneity, FedGPO exhibits relatively lower prediction accuracy (88.9% on average), as the minimization of the performance gaps across devices does not guarantee the model convergence in this case, whereas FedGPO selects the parameters that guarantee the model convergence while improving the energy efficiency, as shown in Figure 11. ### 5.3 Comparison with Prior Work Figure 11: Even in the presence of data heterogeneity, FedGPO achieves 6.2x, 1.9x, and 1.3x higher energy efficiency than Fixed (Best), Adaptive (BO), and Adaptive (GA), respectively, by adaptively adjusting E and K along with B round-by-round. We evaluate FedGPO compared to two prior works for CNN-MNIST: FedEX [29] and ABS [49]. On average, FedGPO achieves 1.5x and 2.1x improvements in energy efficiency compared to FedEX and ABS, respectively (Figure 12). Note other NNs show similar result trends. Under the runtime variance, FedEX and ABS improves the convergence performance and PPW over the baseline, as they reduce the device performance gap, by explicitly considering the straggler problem. However, FedEX does not adapt to the runtime variance as quickly as FedGPO does, due to the lower sample efficiency of exponentiated gradient updates. Furthermore, ABS does not adjust E and K, which helps to deal with the straggler problem and data heterogeneity. As a result, FedGPO further improves PPW by 1.5x and 1.7x over FedEX and ABS, respectively (Figure 12). Compared to the baseline, FedEX is robust to data heterogeneity by adjusting the amount of non-IID data reflected on the model gradients with E and K. However, it still fails to quickly adapt to the data heterogeneity due to its low sample efficiency. In contrast, ABS is not robust to data heterogeneity, as B is not closely related to the data heterogeneity as explained in Section 2. Consequently, FedGPO achieves 1.4x and 3.6x higher PPW over FedEX and ABS, respectively (Figure 12). ### 5.4 Convergence and Overhead Analysis Table 5: Accuracy for Global Parameter Selection. Runtime | Data | Prediction ---|---|--- Variance | Heterogeneity | Accuracy No | No | 94.7% Yes (On-device Interference) | No | 94.2% Yes (Unstable Network) | No | 94.5% No | Yes | 87.7% Yes | Yes | 90.1% When training the shared Q-tables of FedGPO, the reward converges after 30-40 aggregation rounds. Prior to convergence, FedGPO shows 24.2% lower energy efficiency than Fixed (Best), on average. After the convergence, however, FedGPO selects more efficient global parameters than the baselines, as we observed in Section 5.2. Consequently, the global energy efficiency is eventually improved compared with that of the baselines. The average runtime cost associated with training the shared Q-tables is 499.6 $\mu$s except for the FL execution time, corresponding to 0.7% of the training time per round. The overhead includes 1) identifying the per-device states (496.8 $\mu$s), 2) choosing global parameters (0.2 $\mu$s), 3) calculating the reward (2.1 $\mu$s), and 4) updating the Q-tables (0.5 $\mu$s). The total memory requirement of FedGPO is also efficient — in our experiments with three device categories, 0.4MB, 0.0125% of the 32GB DRAM capacity, was only required. ## 6 Related Work Optimization for FL: FL enables collaborative training of a shared ML model in a private manner [4, 36, 51, 46]. To deploy FL on the edge efficiently, FedAvg has been employed as the standard algorithm [36, 51]. This algorithm alleviates the communication overheads by employing fewer participants with higher number of per-device training iterations [51, 11, 6]. Subsequently, various methods have been proposed for model accuracy improvement [43] and security robustness [45, 48]. Although FedAvg demonstrates the potential of FL deployment, it is still faceing the important challenges for optimization — namely the straggler problem and data heterogeneity. As the straggler problem leads to the excess energy consumption of participant devices, prior studies attempted to mitigate it via asynchronous gradient updates [8] or intelligent participant selection [26, 51, 35]. In addition, to deal with the adverse impact of data heterogeneity, other approaches attempted to reduce the amount of non-IID data reflected on model gradients by allowing asynchronous aggregations [8, 7], the use of globally shared data [74], or partial updates [42]. As the main target of the techniques does not encompass the global parameter adjustment, they can be applied with FedGPO. Global parameter optimization for FL: Along with the general optimizations, global parameter optimization is also crucial in FL, as the global parameters can significantly affect the FL execution efficiency. In traditional centralized training, many hyperparameter optimization (HPO) techniques have been proposed to expedite the space exploration based on ML. Such techniques include Bayesian Optimization (BO) [66, 25] and the Genetic Algorithm (GA) [1]. However, as these HPO techniques require training with the entire dataset for each set of global parameters, they are not feasible for the resource-constrained edge execution environment. Furthermore, they also do not directly address the unique challenges of FL including the straggler problem and data heterogeneity. Considering the system and data heterogeneity in FL, several approaches have been proposed to adjust the global parameters round-by-round based on exponentiated gradient updates [29] or deep RL [49]. However, these techniques do not consider the stochastic nature of the edge-cloud execution environment including performance interference and network variability. In addition, they do not account for the energy efficiency of the participant devices. Different from the prior approaches, FedGPO explores the energy efficient global parameter optimization for FL, when there exists system/data heterogeneity and runtime variance. Based on a customized reinforcement learning, FedGPO can identify a near-optimal global parameters for each round by adapting to the heterogeneity and runtime variance. Figure 12: FedGPO outperforms both FedEX [29] and ABS [49], in terms of convergence time and energy efficiency, regardless of the presence of runtime variance or data heterogeneity. ## 7 Conclusion To enable energy-efficient FL on the edge, we propose a global parameter optimization framework called FedGPO. The FL performance and energy efficiency characterization in the edge execution environment demonstrates that optimal global parameters depend on various features: NN characteristics, system/data heterogeneity, and stochastic runtime variance. FedGPO successfully determines near-optimal global parameters in consecutive FL aggregation rounds, by considering these features. We implement representative FL use cases on an emulated edge-cloud execution environment using off-the-shelf systems. FedGPO improves the average FL energy efficiency by 3.6x, compared with the baseline settings. Compared to FedEX and ABS, FedGPO improves the energy efficiency by 1.5x and 2.1x, on average, respectively, by considering system/data heterogeneity and runtime variance. We demonstrate the viability of FedGPO as a solution to global parameter optimization for energy-efficient FL in realistic edge-cloud execution environments. ## References * [1] H. Alibrahim and S. A. Ludwig, “Hyperparameter optimization: Comparing genetic algorithm against grid search and bayesian optimization,” in _IEEE Congress on Evolutionary Computation_ , 2021. * [2] Amazon, “Ec2.” [Online]. Available: https://aws.amazon.com/ec2 * [3] Apple, “Protection against reconstruction and its applications in private federated learning.” [Online]. Available: https://machinelearning.apple.com/research * [4] K. Bonawitz, H. Eichner, W. Grieskamp, D. Huba, A. Ingerman, V. Ivanov, C. Kiddon, J. Konecny, S. Mazzocchi, H. B. McMahan, T. V. Overveldt, D. Petrou, D. Ramage, and J. Roselander, “Towards federated learning at scale: System design,” _arXiv:1902.01046_ , 2019. * [5] Z. Chai, H. Fayyaz, Z. Fayyaz, A. Anwar, Y. Zhou, H. Ludwig, and Y. Cheng, “Towards taming the resource and data heterogeneity in federated learning,” in _Proceedings of the USENIX Conference on Operational Machine Learning (OpML)_ , 2019. * [6] C. Chen, H. Xu, W. Wang, B. Li, B. Li, L. Chen, and G. Zhang, “Communication-efficient federated learning with adaptive parameter freezing,” in _Proc. of the Int’l Conf. on Distributed Computing Systems_ , 2021. * [7] Y. Chen, S. Biookaghazadeh, and M. Zhao, “Exploring the capabilities of mobile devices in supporting deep learning,” in _Proceedings of the ACM/IEEE Symposium on Edge Computing (SEC)_ , 2019, pp. 127–138. * [8] Y. Chen, Y. Ning, M. Slawski, and H. Rangwala, “Asynchronous online federated learning for edge devices with non-iid data,” _arXiv:1911.02134_ , 2019. * [9] Y. Choi, S. Park, and H. Cha, “Optimizing energy efficiency of browsers in energy-aware scheduling-enabled mobile devices,” in _Proceedings of the Int’l Conf. on Mobile Computing and Networking (MobiCom)_ , 2019. * [10] J. Deng, W. Dong, R. Socher, L. J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in _Proceedings of the IEEE Conf. on Computer Vision and Pattern Recognition (CVPR)_ , 2009. * [11] Y. Deng, F. Lyu, J. Ren, Y. Zhang, Y. Zhou, Y. Zhang, and Y. Yang, “Share: Shaping data distribution at edge for communication-efficient hierarchical federated learning,” in _Proc. of the Int’l Conf. on Distributed Computing Systems_ , 2021. * [12] N. Ding, D. Wagner, X. Chen, A. Pathak, Y. C. Hu, and A. Rice, “Characterizing and modeling the impact of wireless signal strength on smartphone battery drain,” in _Proceedings of the International Conference on Measurement and Modeling of Computer Systems (SIGMETRICS)_ , 2013, pp. 29–40. * [13] DL4j, “Deeplearning4j.” [Online]. Available: https://deeplearning4j.org * [14] E. Even-Dar, S. Mannor, and Y. Mansour, “Action elimination and stopping conditions for the multi-armed bandit and reinforcement learning problems,” _Journal of Machine Learning Research_ , vol. 7, pp. 1079–1105, 2006. * [15] B. Gaudette, C.-J. Wu, and S. Vrudhula, “Improving smartphone user experience by balancing performance and energy with probabilistic qos guarantee,” in _Proc. of the Int’l Symp. on High Performance Computer Architecture_ , 2016\. * [16] B. Gaudette, C.-J. Wu, and S. Vrudhula, “Optimizing user satisfaction of mobile workloads subject to various sources of uncertainties,” _IEEE Transactions on Mobile Computing_ , vol. 18, no. 12, pp. 2941–2953, 2019. * [17] Google, “Federated learning.” [Online]. Available: https://federated.withgoogle.com * [18] Google, “Federated learning: Collaborative machine learning without centralized training data.” [Online]. Available: https://ai.googleblog.com/2017/04/federated-learning-collaborative.html * [19] A. Hard, K. Rao, R. Mathews, S. Ramaswamy, F. Beaufays, S. Augenstein, H. Eichner, C. Kiddon, and D. Ramage, “Federated learning for mobile keyboard prediction,” _arXiv:1811.03604_ , 2018. * [20] M. Hejazinia, D. Huba, I. Leontiadis, K. Maeng, M. Malek, L. Melis, I. Mironov, M. Nasr, K. Wang, and C.-J. Wu, “Fel: High capacity learning for recommendation and ranking via federated ensemble learning,” 2022. * [21] E. Hoffer, I. Hubara, and D. Soudry, “Train longer, generalize better: Closing the generalization gap in large batch training of neural networks,” _arXiv:1705.08741v2_ , 2017. * [22] A. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, “Mobilenets: Efficient convolutional neural networks for mobile vision applications,” _arXiv:1704.04861_ , 2017. * [23] Huawei, “Kirin 980, the world’s first 7nm process mobile ai chipset.” [Online]. Available: https://consumer.huawei.com/en/campaign/kirin980/ * [24] D. Huba, J. Nguyen, K. Malik, R. Zhu, M. Rabbat, A. Yousefpour, C.-J. Wu, H. Zhan, P. Ustinov, H. Srinivas, K. Wang, A. Shoumikhin, J. Min, and M. Malek, “Papaya: Practical, private, and scalable federated learning,” in _Proceedings of Machine Learning and Systems_ , 2022. * [25] F. Hutter, L. Kotthoff, and J. Vanschoren, “Automl: Methods, systems, challenges,” 2019. * [26] Y. Jin, L. Jiao, Z. Qian, S. Zhang, S. Lu, and X. Wang, “Resource-efficient and convergence-preserving online participant selection in federated learning,” in _Proc. of the Int’l Conf. on Distributed Computing Systems_ , 2020. * [27] R. Joseph and M. Martonosi, “Run-time power estimation in high performance microprocessors,” in _Proceedings of International Symposium on Low Power Electronics and Design (ISLPED)_ , 2001. * [28] S. C. Kao, G. Jeong, and T. Krishna, “Confuciux: Autonomous hardware resource assignment for dnn accelerators using reinforcement learning,” in _Proceedings of the Int’l Symp. on Microarchitecture (MICRO)_ , 2020. * [29] M. Khodak, R. Tu, T. Li, L. Li, M.-F. Balcan, V. Smith, and A. Talwalkar, “Federated hyperparameter tuning: Challenges, baselines, and connections to weight-sharing,” _arXiv:2106.04502v2_ , 2021. * [30] Y. G. Kim and S. W. Chung, “Signal strength-aware adaptive offloading for energy efficient mobile devices,” in _Proceedings of the International Symposium on Low Power Electronics and Design_ , 2017, pp. 1–6. * [31] Y. G. Kim, M. Kim, and S. W. Chung, “Enhancing energy efficiency of multimedia applications in heterogeneous mobile multi-core processors,” _IEEE Transactions on Computers_ , vol. 66, no. 11, pp. 1878–1889, 2017. * [32] Y. G. Kim, M. Kim, J. M. Kim, M. Sung, and S. W. Chung, “A novel gpu power model for accurate smartphone power breakdown,” _ETRI Journal_ , vol. 37, no. 1, pp. 157–164, 2015. * [33] Y. G. Kim, Y. S. Lee, and S. W. Chung, “Signal strength-aware adaptive offloading with local image preprocessing for energy efficient mobile devices,” _IEEE Transactions on Computers_ , vol. 69, no. 1, pp. 99–101, 2020. * [34] Y. G. Kim and C.-J. Wu, “Autoscale: Energy efficiency optimization for stochastic edge inference using reinforcement learning,” in _Proceedings of the IEEE/ACM International Symposium on Microarchitecture (MICRO)_ , 2020, pp. 1082–1096. * [35] Y. G. Kim and C.-J. Wu, “Autofl: Enabling heterogeneity-aware energy efficient federated learning,” in _Proceedings of the IEEE/ACM International Symposium on Microarchitecture (MICRO)_ , 2021. * [36] J. Konecny, H. B. McMahan, F. X. Yu, P. Richtarik, A. T. Suresh, and D. Bacon, “Federated learning: Strategies for improving communication efficiency,” _arXiv:1610.05492_ , 2016. * [37] D. E. Koulouriotis and A. Xanthopoluos, “Reinforcement learning and evolutionary algorithms for non-stationary multi-armed bandit problems,” _Applied Mathematics and Computation_ , vol. 196, 2008. * [38] Y. LeCun, C. Cortes, and C. J. C. Burges, “The mnist database of handwritten digits.” [Online]. Available: http://yann.lecun.com/exdb/mnist/ * [39] G. Lee, M. Jeong, Y. Shin, S. Bae, and S. Y. Yun, “Perservations of the global knowledge by not-true self knowledge distillation in federated learning,” _arXiv:2106.03097v3_ , 2021. * [40] S.-Y. Lee and C.-J. Wu, “Performance characterization, prediction, and optimization for heterogeneous systems with multi-level memory interference,” in _Proceedings of the IEEE International Symposium on Workload Characterization (IISWC)_ , 2017, pp. 43–53. * [41] Q. Li, Y. Diao, Q. Chen, and B. He, “Federated learning on non-iid data silos: An experimental study,” _arXiv:2102:02079v2_ , 2021. * [42] T. Li, A. K. Sahu, M. sanjabi, M. Zaheer, A. Talwalkar, and V. Smith, “Federated optimization in heterogeneous networks,” in _Proceedings of Int’l Conf. on Machine Learning and Systems (MLSys)_ , 2020. * [43] T. Li, M. Sanjabi, A. Beirami, and V. Smith, “Fair resource allocation in federated learning,” in _Proceedings of the International Conference on Learning Representations (ICLR)_ , 2020. * [44] X. Li, K. Huang, W. Yang, S. Wang, and Z. Zhang, “On the convergence of fedavg on non-iid data,” in _Proceedings of the International Conference on Learing Representation (ICLR)_ , 2020. * [45] Y. Li, M. Alian, Y. Yuan, Z. Qu, P. Pan, R. Wang, A. Schwing, H. Esmaeilzadeh, and N. S. Kim, “A network-centric hardware/algorithm co-design to accelerate distributed training of deep neural networks,” in _Proc. of the Int’l Symp. on Microarchitecture (MICRO)_ , 2018. * [46] S. Lin, G. Yang, and J. Zhang, “A collaborative learning framework via federated meta-learning,” in _Proc. of the Int’l Conf. on Distributed Computing Systems_ , 2020. * [47] T. Lin, L. Kong, S. U. Stich, and M. Jaggi, “Ensemble distillation for robust model fusion in federated learning,” in _Proceedings of the Int’l Conf. on Neural Information Processing Systems (NIPS)_ , 2020. * [48] Y. Lu, X. Huang, Y. Dai, S. Maharjan, and Y. Zhang, “Differentially private asynchronous federated learning for mobile edge computing in urban informatics,” _IEEE Transactions on Industrial Informatics_ , vol. 16, pp. 2134–2143, 2020. * [49] Z. Ma, Y. Xu, H. Xu, Z. Meng, L. Huang, and Y. Xue, “Adaptive batch size for federated learning in resource-constrained edge computing,” _IEEE Transactions on Mobile Computing_ , 2021. * [50] K. Maeng, H. Lu, L. Melis, J. Nguyen, M. Rabbat, and C.-J. Wu, “Towards fair federated recommendation learning: Characterizing the inter-dependence of system and data heterogeneity,” in _Proceedings of the 16th ACM Conference on Recommender Systems_ , 2022, p. 156–167. * [51] H. B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. Arcas, “Communication-efficient learning of deep networks from decentralized data,” _arXiv:1602.05629_ , 2017. * [52] Meta, “Applying federated learning to protect data on mobile devices.” [Online]. Available: https://engineering.fb.com/2022/06/14/production-engineering/federated-learning-differential-privacy * [53] T. Mitchell, “Machine learning,” _McGraw Hill_ , 1997. * [54] Monsoon, “High voltage power monitor.” [Online]. Available: https://www.msoon.com/high-voltage-power-monitor * [55] Motorola, “Moto x force - technical specs.” [Online]. Available: https://support.motorola.com/uk/en/solution/MS112171 * [56] X. Mu, Y. Shen, K. Cheng, X. Geng, J. Fu, T. Zhang, and Z. Zhang, “Fedproc: Prototypical contrastive federated learning on non-iid data,” _arXiv:2019.12273v1_ , 2021. * [57] R. Nishtala, P. Carpenter, V. Petrucci, and X. Martorell, “Hipster: Hybrid task manager for latency-critical cloud workloads,” in _Proceedings of the International Symposium on High Performance Computer Architecture (HPCA)_ , 2017, pp. 409–420. * [58] S. Pagani, S. Manoj, A. Jantsch, and J. Henkel, “Machine learning for power, energy, and thermal management on multicore processors: A survey,” _IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems_ , vol. 39, no. 1, pp. 101–116, 2020. * [59] D. Pandiyan, S.-Y. Lee, and C.-J. Wu, “Performance, energy characterizations and architectural implications of an emerging mobile platform benchmark suit,” in _Proceedings of IEEE International Symposium on Workload Characterization (IISWC)_ , 2013. * [60] PyTorch, “Pytorch.” [Online]. Available: https://pytorch.org * [61] Samsung, “Samsung galaxy s10e, s10, & s10+.” [Online]. Available: https://www.samsung.com/global/galaxy/galaxy-s10 * [62] D. Shingari, A. Arunkumar, B. Gaudette, S. Vrudhula, and C.-J. Wu, “Dora: Optimizing smartphone energy efficiency and web browser performance under interference,” in _Proceedings of the IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS)_ , 2018, pp. 64–75. * [63] D. Shingari, A. Arunkumar, and C.-J. Wu, “Characterization and throttling-based mitigation of memory interference for heterogeneous smartphones,” in _Proceedings of the IEEE International Symposium on Workload Characterization (IISWC)_ , 2015. * [64] S. L. Smith, P. J. Kindermans, C. Ying, and Q. V. Le, “Don’t decay the learning rate, increase the batch size,” in _Proceedings of International Conference on Learning Representations (ICLR)_ , 2018. * [65] J. Snoek, H. Larochelle, and R. P. Adams, “Practical bayesian optimization of machine learning algorithms,” in _Proceedings of the International Conference on Neural Information Processing Systems (NIPS)_ , 2012. * [66] A. Souza, L. Nardi, L. Oliveira, K. Olukotun, M. Lindauer, and F. Hutter, “Prior-guided bayesian optimization,” in _Proceedings of Workshop on Meta-Learning (NeurlPS)_ , 2020. * [67] J. T. Springenberg, A. Dosovitskiy, T. Brox, and M. Riedmiller, “Striving for simplicity: The all convolutional net,” in _Proceedings of the International Conference on Language Representations (ICLR)_ , 2015. * [68] C. Wang, X. Wei, and P. Zhou, in _Proceedings of IEEE International Parallel and Distributed Processing Symposium (IPDPS)_ , 2020, pp. 212–221. * [69] C. Wang, Y. Yang, and P. Zhou, “Towards efficient scheduling of federated mobile devices under computational and statistical heterogeneity,” _IEEE Transactions on Parallel and Distributed Systems_ , vol. 32, pp. 394–410, 2021. * [70] C.-J. Wu, D. Brooks, K. Chen, D. Chen, S. Choudhury, M. Dukhan, K. Hazelwood, E. Isaac, Y. Jia, B. Jia, T. Leyvand, H. Lu, Y. Lu, L. Qiao, B. Reagen, J. Spisak, F. Sun, A. Tulloch, P. Vajda, X. Wang, Y. Wang, B. Wasti, Y. Wu, R. Xian, S. Yoo, and P. Zhang, “Machine learning at facebook: Understanding inference at the edge,” in _Proceedings of the IEEE International Symposium on High Performance Computer Architecture (HPCA)_ , 2019, pp. 331–344. * [71] C.-J. Wu, R. Raghavendra, U. Gupta, B. Acun, N. Ardalani, K. Maeng, G. Chang, F. A. Behram, J. Huang, C. Bai, M. Gschwind, A. Gupta, M. Ott, A. Melnikov, S. Candido, D. Brooks, G. Chauhan, B. Lee, H. H. S. Lee, B. Akyildiz, M. Balandat, J. Spisak, R. Jain, M. Rabbat, and K. Hazelwook, “Sustainable ai: Environmental implications, challenges and opportunities,” _arXiv:2111.00364v2_ , 2022. * [72] X. Wu, J. Niu, X. Liu, T. Ren, Z. Huang, and Z. Li, “pfedgf: Enabling personalized federated learning via gradient fusion,” in _Proceedings of IEEE Internatiaonl Parallel and Distributed Processing Symposium (IPDPS)_ , 2022, pp. 639–649. * [73] D. Yarats, A. Zhang, I. Kostrikov, B. Amos, J. Pineau, and R. Fergus, “Improving sample efficiency in model-free reinforcement learning from images,” _arXiv:1910.01741v3_ , 2019. * [74] Y. Zhao, M. Li, L. Lai, N. Suda, D. Civin, and V. Chandra, “Federated learning with non-iid data,” _arXiv:1806:00582_ , 2018. * [75] J. Zhou, S. Zhang, Q. Lu, W. Dai, M. Chen, X. Liu, S. Pirttikangas, Y. Shi, and W. Zhang, “A survey on federated learning and its applications for accelerating industrial internet of things,” _arXiv:2104.10501_.
# Regret Pruning for Learning Equilibria in Simulation-Based Games Bhaskar Mishra, Cyrus Cousins, Amy Greenwald ###### Abstract In recent years, empirical game-theoretic analysis (EGTA) has emerged as a powerful tool for analyzing games in which an exact specification of the utilities is unavailable. Instead, EGTA assumes access to an oracle, i.e., a simulator, which can generate unbiased noisy samples of players’ unknown utilities, given a strategy profile. Utilities can thus be empirically estimated by repeatedly querying the simulator. Recently, various progressive sampling (PS) algorithms have been proposed, which aim to produce PAC-style learning guarantees (e.g., approximate Nash equilibria with high probability) using as few simulator queries as possible. A recent work by Areyan Viqueira, Cousins, and Greenwald introduces a pruning technique called _regret-pruning_ which further minimizes the number of simulator queries placed in PS algorithms which aim to learn _pure_ Nash equilibria. In this paper, we address a serious limitation of this original regret pruning approach – it is only able to guarantee that _true_ pure Nash equilibria of the empirical game are approximate equilibria of the true game, and is unable to provide any strong guarantees regarding the efficacy of approximate pure Nash equilibria. This is a significant limitation since in many games, pure Nash equilibria are computationally intractable to find, or even non-existent. We introduce three novel regret pruning variations. The first two variations generalize the original regret pruning approach to yield guarantees for approximate pure Nash equilibria of the empirical game. The third variation goes further to even yield strong guarantees for all approximate mixed Nash equilibria of the empirical game. We use these regret pruning variations to design two novel progressive sampling algorithms, PS-REG+ and PS-REG-M, which experimentally outperform the previous state-of-the-art algorithms for learning pure and mixed equilibria, respectively, of simulation-based games. ## Introduction Game theory is the standard conceptual framework used to analyze strategic interactions among rational agents in multi-agent systems. A game comprises a collection of players, each with a set of strategies and a utility function, mapping strategy profiles (i.e., combinations of strategies) to values. Traditionally, game-theoretic analysis presumes complete access to a game’s structure, including the utility functions. In recent years, empirical game-theoretic analysis (EGTA) has emerged as a powerful tool for analyzing games in which such an exact specification of the utilities is unavailable. Instead, EGTA assumes access to an oracle, i.e., a (stochastic) simulator, which produces unbiased noisy samples of players’ unknown utilities given a strategy profile (Wellman 2006; Tuyls et al. 2020; Areyan Viqueira, Cousins, and Greenwald 2020). Such games are called _simulation-based games_ (Vorobeychik and Wellman 2008), or _black-box_ games (Picheny, Binois, and Habbal 2016), and their empirical counterparts, which are derived from simulation data, are called _empirical games_. Simulation- based games have been studied in many practical settings including trading agent analyses in supply chains (Vorobeychik, Kiekintveld, and Wellman 2006; Jordan, Kiekintveld, and Wellman 2007), ad auctions (Jordan and Wellman 2010; Areyan Viqueira et al. 2019), and energy markets (Ketter, Peters, and Collins 2013); designing network routing protocols (Wellman, Kim, and Duong 2013); strategy selection in real-time games (Tavares et al. 2016); and the dynamics of RL algorithms, like AlphaGo (Tuyls et al. 2018). A typical EGTA goal is to produce PAC-style learning guarantees (e.g., approximate Nash equilibria with high probability) with minimal _query complexity_ , i.e., the number of simulation queries placed (Tuyls et al. 2020; Areyan Viqueira, Cousins, and Greenwald 2020; Cousins et al. 2022). This goal has led to the development of _progressive sampling_ algorithms (Areyan Viqueira, Cousins, and Greenwald 2020; Cousins et al. 2022), which place simulation queries in progressive batches, until the desired guarantee is reached. On top of progressive sampling, two papers introduce various pruning techniques, to further minimize query complexity. One of these techniques, _well-estimated pruning_ , prunes strategy profiles whose utilities are very likely to already be sufficiently close to the true utilities (Cousins et al. 2022). This technique is useful for learning a variety of game properties, including regret, pure or mixed Nash equilibria, welfare-maximizing outcomes, and more. A second technique, called _regret pruning_ , is intended for use only when learning pure Nash equilibria, as it prunes strategy profiles that are highly unlikely to be best responses, and hence unlikely to be necessary for finding pure Nash equilibria (Areyan Viqueira, Cousins, and Greenwald 2020). In this paper, we focus predominantly on regret pruning. Areyan Viqueira, Cousins, and Greenwald (2020) claim that their regret pruning criterion can be used to learn an empirical game which, with high probability, satisfies a certain dual pure Nash containment guarantee – all pure Nash equilibria of the simulation-based game are approximate equilibria of the empirical game, and all approximate pure Nash equilibria of the empirical game are approximate equilibria of the simulation-based game. We show via a direct counterexample, however, that their progressive sampling algorithm using regret pruning fails to satisfy this second inclusion. Rather, their algorithm only yields the guarantee that all _true_ pure Nash equilibria of the empirical game are approximate equilibria of the simulation-based game, and is unable to directly yield any non-trivial guarantee regarding _approximate_ pure Nash equilibria of the empirical game. This difference is crucial, since in many games, computing a pure Nash equilibrium is computationally intractable (e.g., it is NP-complete in graphical games (Gottlob, Greco, and Scarcello 2005)), and sometimes such an equilibrium does not even exist. In such cases, an approximate pure Nash equilibrium is the best that can be hoped for, but Areyan Viqueira, Cousins, and Greenwald’s progressive sampling algorithm using regret pruning yields no guarantees for such equilibria. In response to this limitation of the original regret pruning technique, we design three new variations of regret pruning (and new corresponding progressive sampling algorithms). At the cost of a slightly tighter regret pruning criterion, our first regret pruning variation yields the guarantee regarding approximate equilibria of the empirical game which Areyan Viqueira, Cousins, and Greenwald’s regret pruning technique was originally intended to yield. Our second variation incorporates the non-uniform utility deviation bounds used by Cousins et al. (2022) for well-estimated pruning in order to design a looser regret pruning criterion which nonetheless yields the same guarantees as the first variation. Finally, our third variation also takes advantage of non-uniform bounds, but uses different proof techniques than those used in the first two variations to yield both pure _and_ mixed Nash containment guarantees. The third regret pruning variation is a particularly significant contribution, as it is one of the first pruning techniques beyond simple well-estimated pruning for learning mixed equilibria. The only available alternative is rationalizability pruning introduced by Areyan Viqueira, Cousins, and Greenwald (2020), which requires the use of a computationally expensive iterative dominance algorithm, has a very tight pruning criterion which often prunes few to no strategy profiles in practice, and most importantly, like the original regret pruning technique, only yields guarantees regarding _true_ mixed Nash equilibria of the empirical game. In contrast, our novel third regret pruning variation utilizes a pruning criterion which is very cheap to compute, can prune a very significant number of simulation queries (which we confirm experimentally), and yields Nash containment guarantees for approximate mixed equilibria of the empirical game. In addition to presenting the guarantees progressive sampling algorithms using these novel variations of regret pruning can satisfy, we also derive sample complexity bounds for these new algorithms. In particular, we present PAC- style upper bounds on the number of samples our progressive sampling algorithms will take to prune each respective strategy profile. Finally, we conclude by demonstrating experimentally that our novel progressive sampling algorithms which incorporate both well-estimated pruning and novel regret pruning variations significantly outperform Cousins et al.’s progressive sampling algorithm which used well-estimated pruning alone, in some cases requiring up to 50% fewer simulation queries to learn equilibria of similar quality. ### Related Works The EGTA literature, while relatively young, is growing rapidly, with researchers actively contributing methods for myriad game models. Some of these methods are designed for normal-form games (Cousins et al. 2022; Areyan Viqueira, Cousins, and Greenwald 2020; Areyan Viqueira et al. 2019; Tavares et al. 2016; Fearnley et al. 2015; Vorobeychik and Wellman 2008), and others, for extensive-form games (Marchesi, Trovò, and Gatti 2020; Gatti and Restelli 2011; Zhang and Sandholm 2021). Most methods apply to games with finite strategy spaces, but some apply to games with infinite strategy spaces (Marchesi, Trovò, and Gatti 2020; Vorobeychik, Wellman, and Singh 2007; Wiedenbeck, Yang, and Wellman 2018). A related line of work aims to empirically design mechanisms via EGTA methodologies (Vorobeychik, Kiekintveld, and Wellman 2006; Areyan Viqueira et al. 2019). The progressive sampling algorithms and pruning techniques which we design in this paper extend work done by Cousins et al. (2022); Areyan Viqueira, Cousins, and Greenwald (2020); Areyan Viqueira et al. (2019) in designing algorithms for learning equilibria in normal-form simulation-based games with finite strategy spaces. ## Learning Framework We begin with the standard definition of standard normal-form games, and some related properties . We then introduce our formal model of simulation-based games and empirical games. Finally, we state the concentration inequalities we use to guide pruning in our progressive sampling algorithms. ### Basic Game Theory ###### Definition 1 (Normal-Form Game). A _normal-form game_ $\Gamma\doteq\langle P,\\{S_{p}\\}_{p\in P},\bm{u}\rangle$ consists of a set of players $P$, each with a corresponding _pure strategy set_ $S_{p}$. We define $\bm{S}\doteq S_{1}\times\dots\times S_{|P|}$ to be the _pure strategy profile space_ , and then $\bm{u}:\bm{S}\to\mathbb{R}^{|P|}$ is a vector-valued _utility function_ (equivalently, a vector of $|P|$ scalar utility functions $\bm{u}_{p}$). Given an NFG $\Gamma$, we denote by $S_{p}^{\diamond}$ the set of distributions over $S_{p}$; this set is called player $p$’s _mixed strategy set_. We define $\bm{{S}^{\diamond}}={S}^{\diamond}_{1}\times\dots\times{S}^{\diamond}_{|P|}$ to be the _mixed strategy profile space_ , and, overloading notation, we write $\bm{u}(\bm{s})$ to denote the expected utility of a mixed strategy profile $\bm{s}\in\bm{{S}^{\diamond}}$. Each pure strategy profile $\bm{s}\in\bm{S}$ is contained in the mixed strategy profile space $\bm{{S}^{\diamond}}$, represented by the profile with each mixed strategy concentrated entirely at the respective pure strategy. Given player $p$ and strategy profile $\bm{s}\in\bm{{S}^{\diamond}}$, the set $\operatorname{Adj}_{p,\bm{s}}\doteq\\{(\bm{s}_{1},\dots,\bm{s}_{p-1},t,\bm{s}_{p+1},\dots,\bm{s}_{\left\lvert{}P\right\rvert{}})\mid t\in S_{p}\\}$ contains all adjacent strategy profiles, meaning those in which the strategies of all players $q\neq p$ are fixed at $\bm{s}_{q}$, while player $p$’s strategy may vary across their pure strategy set. ###### Definition 2 (Regret). A player $p$’s regret at strategy profile $\bm{s}\in\bm{{S}^{\diamond}}$ is defined as $\displaystyle\operatorname{Reg}_{p}(\bm{s};\bm{u})\doteq\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\bm{u}_{p}(\bm{s}^{\prime})-\bm{u}_{p}(\bm{s})$. We further define $\operatorname{Reg}(\bm{s};\bm{u})\doteq\max_{p\in P}\operatorname{Reg}_{p}(\bm{s};\bm{u})$. A strategy profile $\bm{s}\in\bm{S}$ is player $p$’s _best response_ if $\operatorname{Reg}_{p}(\bm{s};\bm{u})=0$ (i.e., the player does not regret choosing this strategy profile as opposed to an adjacent one). We say it is an $\varepsilon$-_best response_ if $\operatorname{Reg}_{p}(\bm{s};\bm{u})\leq\varepsilon$. A strategy profile $\bm{s}\in\bm{{S}^{\diamond}}$ is an $\varepsilon$-_Nash equilibrium_ if it is an $\varepsilon$-best response for each player $p\in P$ (i.e., if $\operatorname{Reg}(\bm{s};\bm{u})\leq\varepsilon$). If $\bm{s}$ corresponds to a pure strategy profile, then we call it an $\varepsilon$-_pure Nash equilibrium_ ($\varepsilon$-PNE); otherwise, we call it an $\varepsilon$-_mixed Nash equilibrium_ ($\varepsilon$-MNE). A $0$-PNE is simply called a PNE, and a $0$-MNE is called an MNE. The set of $\varepsilon$-pure (resp. mixed) Nash equilibria is denoted $\textup{E}_{\varepsilon}(\bm{u})$ (resp. $\textup{E}^{\diamond}_{\varepsilon}(\bm{u})$), and the set of pure (resp. mixed) Nash equilibria is denoted $\textup{E}(\bm{u})$ (resp. $\textup{E}^{\diamond}(\bm{u})$). ### Formal Model of Simulation Based Games In simulation-based games, we assume access to a _simulator_ $\mathscr{S}(\cdot)$, which can be queried to produce unbiased noisy samples of the players’ utilities when $\bm{s}\in\bm{S}$ is played. We denote such a sample by $\dot{\bm{u}}(\bm{s})\sim\mathscr{S}(\bm{s})$, where $\dot{\bm{u}}(\bm{s})$ is a $\left\lvert{}P\right\rvert{}$-vector comprising utilities for each player. ###### Definition 3 (Simulation-Based Game). A simulation-based game $\Gamma_{\mathscr{S}}\doteq\langle P,\bm{S},\mathscr{S}\rangle$ consists of a set of players $P$, a pure strategy profile space $\bm{S}\doteq S_{1}\times\dots\times S_{\left\lvert{}P\right\rvert{}}$, and a simulator $\mathscr{S}(\cdot)$ that produces noisy samples $\dot{\bm{u}}(\bm{s})\sim\mathscr{S}(\bm{s})$ upon simulation of a strategy profile $\bm{s}\in\bm{S}$. Corresponding to each simulation-based game $\Gamma_{\mathscr{S}}$ is an expected normal-form game. ###### Definition 4 (Expected Normal-Form Game). Given a simulation-based game $\Gamma_{\mathscr{S}}\doteq\langle P,\bm{S},\mathscr{S}\rangle$, we define the “underlying” utility function $\bm{u}:\bm{S}\to\mathbb{R}^{\left\lvert{}P\right\rvert{}}$ by $\bm{u}(s)\doteq\operatorname*{\mathbb{E}}_{\dot{\bm{u}}(\bm{s})\sim\mathscr{S}(\bm{s})}\left[\dot{\bm{u}}(\bm{s})\right]$. The expected game corresponding to $\Gamma_{\mathscr{S}}$ is then the normal- form game $\langle P,\bm{S},\bm{u}\rangle$. Overloading notation, we also let $\Gamma_{\mathscr{S}}$ denote this (unknown) expected normal-form game. Since we do not have direct access to the expected normal-form game, its utilities must be learned by repeatedly querying the simulator. The resulting empirical estimate of the expected game is called an empirical game. ###### Definition 5 (Empirical Normal-Form Game). Given a simulation-based game $\Gamma_{\mathscr{S}}\doteq\langle P,\bm{S},\mathscr{S}\rangle$, let $\dot{\bm{u}}^{(1)}(\bm{s})$, $\dots$, $\dot{\bm{u}}^{(m_{\bm{s}})}(\bm{s})\sim\mathscr{S}(\bm{s})$ denote the sample utilities produced by $m_{\bm{s}}>0$ queries to the simulator at strategy profile $\bm{s}\in\bm{S}$. We define the empirical utility function $\hat{\bm{u}}:\bm{S}\to\mathbb{R}^{|P|}$ by $\hat{\bm{u}}(\bm{s})\doteq\frac{1}{m_{\bm{s}}}\sum_{i=1}^{m_{\bm{s}}}\dot{\bm{u}}^{(i)}(\bm{s})$ for all $\bm{s}\in\bm{S}$, and the ensuing empirical normal-form game by $\hat{\Gamma}_{\mathscr{S}}\doteq\langle P,\bm{S},\hat{\bm{u}}\rangle$. From here onwards, let $\Gamma_{\mathscr{S}}\doteq\langle P,\bm{S},\mathscr{S}\rangle$ be an arbitrary simulation-based game with underlying utility function $\bm{u}$, and let $\hat{\Gamma}_{\mathscr{S}}\doteq\langle P,\bm{S},\hat{\bm{u}}\rangle$ be a corresponding empirical game. Using our formalization, we can now present one of the foundational results of EGTA (Tuyls et al. 2020). ###### Lemma 1. If $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$, then $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})\subseteq\textup{E}_{4\varepsilon}(\bm{u})\textrm{ and }\textup{E}^{\diamond}(\bm{u})\subseteq\textup{E}^{\diamond}_{2\varepsilon}(\hat{\bm{u}})\subseteq\textup{E}^{\diamond}_{4\varepsilon}(\bm{u})\enspace,$ or more generally, $\textup{E}_{\gamma}(\bm{u})\subseteq\textup{E}_{2\varepsilon+\gamma}(\hat{\bm{u}})$ and $\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u})$ for all $\gamma\geq 0$ (resp. for mixed equilibria). This result can be understood as stating that given a sufficiently strong approximation $\hat{\Gamma}_{\mathscr{S}}$ of a simulation-based game $\Gamma_{\mathscr{S}}$, we can approximate pure (resp. mixed) Nash equilibria in $\Gamma_{\mathscr{S}}$ with perfect recall – all pure (resp. mixed) Nash equilibria in $\Gamma_{\mathscr{S}}$ are approximate Nash equilibria in $\hat{\Gamma}_{\mathscr{S}}$ – and with _approximately_ perfect precision – all approximate pure (resp. mixed) Nash equilibria in $\hat{\Gamma}_{\mathscr{S}}$ are approximate pure (resp. mixed) Nash equilibria in $\Gamma_{\mathscr{S}}$. Lemma 1 is one of the primary motiviations for EGTA’s pursuit of designing efficient algorithms for learning strong approximations of simulation-based games, as it guarantees that the better an approximation an empirical game is of a simulation-based game, the more strategically representative the empirical game will be of the underlying simulation-based game. ### Tail Bounds Next, we state the tail bounds upon which our novel regret pruning techniques and progressive sampling algorithms depend. These are the same bounds derived and used by Cousins et al.; for a more thorough discussion of them see Cousins et al. (2022). For all subsequent results, we make the following “bounded utilities” assumption. ###### Assumption 1 (Bounded Utilities). For each strategy profile $\bm{s}\in\bm{S}$, the sample utilities produced via $\mathscr{S}(\bm{s})$ lie on the bounded interval $[a_{\bm{s}},b_{\bm{s}}]$ for some fixed $a_{\bm{s}},b_{\bm{s}}\in\mathbb{R}$. We define $c:=\sup_{\bm{s}\in\bm{S}}(b_{\bm{s}}-a_{\bm{s}})$. The most straight-forward tail bound for mean-estimation is Hoeffding’s Inequality, which was used by Tuyls et al. (2020). We use Hoeffding’s inequality to bound each individual utility, combined with a union bound to yield a guarantee for all utilities. ###### Theorem 1 (Hoeffding’s Inequality). Let $\hat{\Gamma}_{\mathscr{S}}\doteq\langle P,\bm{S},\hat{\bm{u}}\rangle$ be an empirical game. Then, with probability at least $1-\updelta$, for all $(p,\bm{s})\in\bm{\mathcal{I}}$, it holds that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq c\sqrt{\frac{\ln\left(\nicefrac{{2\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}}{{\updelta}}\right)}{2m_{\bm{s}}}}\doteq\varepsilon^{\textup{H}}_{p}(\bm{s})\enspace.$ Let $\bm{v}_{p}(\bm{s})$ denote $\operatorname*{\mathbb{V}}_{\dot{\bm{u}}(\bm{s})\sim\mathscr{S}(\bm{s})}\left[\dot{\bm{u}}_{p}(\bm{s})\right]$ for all $(p,\bm{s})\in P\times\bm{S}$. Since Hoeffding’s Inequality assumes a worst-case variance on the utilities (i.e., $\bm{v}_{p}(\bm{s})=\nicefrac{{c^{2}}}{{4}}$), when variances are small, it yields a very loose bound. When the variances of utilities are known, Bennett’s inequality provides a non-uniform, variance-sensitive guarantee. ###### Theorem 2 (Bennett’s Inequality). Let $\hat{\Gamma}_{\mathscr{S}}\doteq\langle P,\bm{S},\hat{\bm{u}}\rangle$ be an empirical game. Then, with probability at least $1-\updelta$, for all $(p,\bm{s})\in\bm{\mathcal{I}}$, it holds that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\frac{c\ln\left(\nicefrac{{2\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}}{{\updelta}}\right)}{3m_{\bm{s}}}+\sqrt{\frac{2\bm{v}_{p}(\bm{s})\ln\left(\nicefrac{{2\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}}{{\updelta}}\right)}{m_{\bm{s}}}}$ Of course, since our only access to the utilities of the simulation-based game are via the simulator, we do not know their variances. Cousins et al. (2022) circumvent this limitation by deriving an “empirical Bennett’s inequality” depending on empirical estimates of the true utility variances. ###### Theorem 3 (Empirical Bennett’s Inequality). Let $\hat{\Gamma}_{\mathscr{S}}\doteq\langle P,\bm{S},\hat{\bm{u}}\rangle$ be an empirical game. Let $\kappa_{\updelta}\doteq\left(\frac{1}{3}+\frac{1}{2\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}\right)$. For all $(p,\bm{s})\in\bm{\mathcal{I}}$, define $\displaystyle{\hat{\bm{v}}_{p}(\bm{s})}$ $\displaystyle\doteq{\frac{1}{m-1}}{\sum}\limits_{j=1}^{m}\left(\bm{u}_{p}(\bm{s};y_{j})-\hat{\bm{u}}_{p}(\bm{s};\bm{Y})\right)^{2};$ $\displaystyle\varepsilon^{\hat{\bm{v}}}_{p}(\bm{s})$ $\displaystyle\doteq{\frac{2c^{2}\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{3m}+\sqrt{\kappa_{\updelta}\Bigl{(}\frac{c^{2}\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m-1}\Bigr{)}^{\smash{2}}+\frac{2c^{2}\hat{\bm{v}}_{p}(\bm{s})\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m}}};$ $\displaystyle\varepsilon^{\hat{\textup{B}}}_{p}(\bm{s})$ $\displaystyle\doteq{\frac{c\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{3m}+\sqrt{\frac{2\left(\hat{\bm{v}}_{p}(\bm{s})+\varepsilon^{\hat{\bm{v}}}_{p}(\bm{s})\right)\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m}}}.$ Then, with probability at least $1-\updelta$, for all $(p,\bm{s})\in\bm{\mathcal{I}}$, it holds that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon^{\hat{\textup{B}}}_{p}(\bm{s})$. This empirical Bennett guarantee forms the basis for our progressive sampling algorithms. ### Progressive Sampling Algorithms 1:procedure PSP($\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},c,\updelta,\varepsilon$) 2: input: Conditional game $\Gamma_{\mathcal{Y}}$, condition distribution $\mathscr{D}$, index set $\bm{\mathcal{I}}$, failure probability $\updelta\in(0,1)$, target error $\varepsilon>0$ 3: Initialize empirical utilities $\hat{\bm{u}}^{(0)}_{p}(\bm{s})=0$ for $(p,\bm{s})\in\bm{\mathcal{I}}$ 4: Initialize utility deviation bounds $\hat{\varepsilon}^{(0)}_{p}(\bm{s})=\infty$ for $(p,\bm{s})\in\bm{\mathcal{I}}$ 5: Initialize active utility index set $\bm{\mathcal{I}}^{(0)}\leftarrow\bm{\mathcal{I}}$ 6: Initialize a sampling schedule $m_{1},\dots,m_{T}$ and cumulative sample size $M_{0}\leftarrow 0$ 7: for $t\in 1,\dots,T$ do 8: for $\bm{s}\in\bm{S}$ do 9: Determine unpruned player indices $P(\bm{s};\bm{\mathcal{I}}^{(t)})\leftarrow\\{p\in P\mid(p,\bm{s})\in\bm{\mathcal{I}}^{(t)}\\}$ at strategy profile $\bm{s}$ 10: Query simulator for utilities of unpruned players: $\\{\dot{\bm{u}}_{p}(\bm{s})\mid p\in P(\bm{s};\bm{\mathcal{I}}^{(t)})\\}\leftarrow\mathscr{S}(\bm{s};P(\bm{s};\bm{\mathcal{I}}^{(t)}))$ 11: Update empirical utilities $\hat{\bm{u}}^{(t)}_{p}(\bm{s})\leftarrow\frac{M_{t-1}}{M_{t-1}+m_{t}}\cdot\hat{\bm{u}}^{(t-1)}_{p}(\bm{s})+\frac{m_{t}}{M_{t-1}+m_{t}}\cdot\dot{\bm{u}}_{p}(\bm{s})$ for $p\in P(\bm{s};\bm{\mathcal{I}}^{(t)})$ 12: Compute new utility deviation bounds $\hat{\varepsilon}^{(t)}_{p}(\bm{s})$ for $p\in P(\bm{s};\bm{\mathcal{I}}^{(t)})$, each with failure probability $\nicefrac{{\updelta}}{{\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}}$ 13: end for 14: Update cumulative sample size $M_{t}\leftarrow M_{t-1}+m_{t}$ 15: Prune any indices in $\bm{\mathcal{I}}^{(t)}$ which do not require further estimation (e.g., well-estimated pruning: $\bm{\mathcal{I}}^{(t)}\leftarrow\smash{\\{(p,\bm{s})\in\bm{\mathcal{I}}^{(t-1)}\mid\hat{\varepsilon}^{(t)}_{p}(\bm{s})>\varepsilon\\}}$, i.e., prune indices that have met the target $\varepsilon$ error guarantee) 16: if all indices in $\bm{\mathcal{I}}^{(t)}$ are pruned (i.e., $\bm{\mathcal{I}}^{(t)}=\emptyset$) then 17: return empirical utilities $\hat{\bm{u}}^{(t)}$ 18: end if 19: end for 20:end procedure Algorithm 1 General Progressive Sampling Algorithm Finally, we present the general class of progressive sampling (PS) algorithms (see Algorithm 1) for learning simualation-based games. As the name suggests, progressive sampling algorithms work by progressively sampling utilities, pruning those which are sufficiently estimated for the relevant learning goal at hand. One core component of progressive sampling algorithms is the sampling schedule. On each iteration, a progressive sampling algorithm will collect the number of samples dictated by the sampling schedule for each active (i.e., unpruned) utility index. It will then use the samples to update the empirical game and to compute new utility deviation bounds (which in our case will be dependent on the tail bounds introduced earlier). Notice that the utility deviation bounds must each have individual failure probability $\frac{\updelta}{\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}$, as opposed to just $\frac{\updelta}{\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}$ as in Theorem 3. This is to ensure that, via an additional union bound, all pruned indices will, with high probability (w.h.p.), have been pruned justifiably with respect to the true game. Finally, the empirical game and utility deviation bounds will be used to inform which utility indices can be pruned on the current iteration. Progressive sampling algorithms terminate once either all utility indices are pruned, or the sampling schedule is exhausted. In this paper, we focus on PS algorithms for learning equilibria. On the basis of Lemma 1, one sufficient condition for pruning an index $(p,\bm{s})\in\bm{\mathcal{I}}$ is that it is estimated (w.h.p.) to within some target error $\varepsilon$. If this is the only pruning criteria used, then upon termination of the algorithm, if all indices have been pruned, the resulting empirical game will (w.h.p.) satisfy $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$, and will thus (w.h.p.) satisfy the pure and mixed dual Nash containment results in Lemma 1. Cousins et al. (2022) design precisely this algorithm, using Hoeffding (Theorem 1) and empirical Bennett (Theorem 3) bounds to inform their pruning (pruning once the guarantee corresponding to an index is tighter than the target error $\varepsilon$), and carefully crafting a sampling schedule which guarantees that all indices will be pruned prior to its exhaustion. They refer to this pruning approach as _well-estimated pruning_ (Cousins et al. 2022). We show an example of how the pruning criteria may be implemented in Algorithm 1 15. ## Regret Pruning Having covered the requisite background, we can now begin our discussion of regret pruning and present our novel regret-pruning variations. As discussed earlier, Lemma 1 immediately suggests well-estimated pruning as a pruning approach, and this approach was used to design the PS algorithm presented in Cousins et al. (2022) (which we henceforth refer to as PS-WE for Progressive Sampling with Well-Estimated Pruning). ###### Theorem 4. If PS- WE$(\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},\updelta,\varepsilon)$ returns an empirical utility function $\hat{\bm{u}}$, then with probability at least $1-\updelta$, for all $\gamma\geq 0$, it holds that 1. 1. $\textup{E}_{\gamma}(\bm{u})\subseteq\textup{E}_{2\varepsilon+\gamma}(\hat{\bm{u}})$ and $\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u})$ 2. 2. $\textup{E}^{\diamond}_{\gamma}(\bm{u})\subseteq\textup{E}^{\diamond}_{2\varepsilon+\gamma}(\hat{\bm{u}})$ and $\textup{E}^{\diamond}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}^{\diamond}_{2\varepsilon+\gamma}(\bm{u})$ . The condition presented in Lemma 1 (i.e., $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$), however, is not the only condition to yield these kinds of Nash containment results. Consider the PS algorithm designed in Areyan Viqueira, Cousins, and Greenwald (2020) (which we refer to as PS-REG-0 for Progressive Sampling with Regret Pruning; the 0 will distinguish this algorithm from our new variations). Unlike PS-WE, PS-REG-0 uses uniform utility deviation bounds – they simply bound all utility deviations by $\sup_{(p,\bm{s})\in\bm{\mathcal{I}}}\varepsilon^{\hat{\textup{B}}}_{p}(\bm{s})$ ($\varepsilon^{\hat{\textup{B}}}$ defined as in Theorem 3)111Areyan Viqueira, Cousins, and Greenwald (2020) use a version of the empirical Bennett tail bounds that is a constant factor looser than the one presented here, as the tighter bound had not been derived until Cousins et al. (2022). PS-REG-0 uses well-estimated pruning (though the authors do not explicitly call it that), but it additionally uses what the authors call “regret pruning” (Areyan Viqueira, Cousins, and Greenwald 2020). The authors claim that though using this pruning approach does _not_ guarantee that the condition in Lemma 1 is met upon termination of the algorithm, it nonetheless guarantees that (w.h.p.) the pure Nash containment result $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})\subseteq\textup{E}_{4\varepsilon}(\bm{u})$ is satisfied. In this section, we present a counterexample which shows that their pruning approach does not in fact guarantee this pure Nash containment result. We show that, instead, their pruning approach is only able to guarantee a weaker Nash containment result. Furthermore, we also present 3 novel variations of the regret pruning criterion presented in Areyan Viqueira, Cousins, and Greenwald (2020) which each have varying benefits and costs in comparison to the original. The first variation is a generalization of the original regret pruning criterion with respect to a hyper-parameter $\gamma^{*}\geq 0$. We show that when $\gamma^{*}=0$, this variation is identical to the regret pruning criterion from PS-REG-0. When $\gamma^{*}=2\varepsilon$, however, we show that, at the cost of taking slightly longer to prune indices, this new variation yields the stronger pure Nash containment guarantee which Areyan Viqueira, Cousins, and Greenwald (2020) originally intended their pruning criterion to meet (i.e., $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})\subseteq\textup{E}_{4\varepsilon}(\bm{u})$). The second variation takes advantage of non-uniform utility deviation bounds to yield the same guarantee as the first variation, while pruning indices significantly sooner in practice than otherwise. Finally, the third variation modifies the second variation, yielding a mixed Nash containment guarantee in addition to the same pure Nash guarantee, at the cost of a slightly tighter pruning criterion than variation 2. ### Old Regret Pruning We begin by presenting the regret pruning criterion used in PS-REG-0 from Areyan Viqueira, Cousins, and Greenwald (2020). Since PS-REG-0 uses uniform utility deviation bounds we simply use $\hat{\varepsilon}^{(t)}$ to denote the utility deviation bound on $\hat{\bm{u}}^{(t)}$ at iteration $t$. PS-REG-0 prunes an index $(p,\bm{s})\in\bm{\mathcal{I}}$ on an iteration $t$ if any of the following holds: 1. 1. $\hat{\varepsilon}^{(t)}\leq\varepsilon$ (well-estimated pruning) 2. 2. $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}}^{(t)})\geq 2\hat{\varepsilon}^{(t)}$ (regret pruning). Areyan Viqueira, Cousins, and Greenwald (2020) claim that PS-REG-0 satisfies the following guarantee. ###### Claim 1. If PS- REG-0$(\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},\updelta,\varepsilon)$ returns an empirical utility function $\hat{\bm{u}}$, then with probability at least $1-\updelta$, it holds that $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})\subseteq\textup{E}_{4\varepsilon}(\bm{u})$. We present a counter-example that shows that the second inclusion $\textup{E}_{2\varepsilon}(\hat{\bm{u}})\subseteq\textup{E}_{4\varepsilon}(\bm{u})$ does not necessarily hold. ###### Counterexample 1. Consider a two-player game $\Gamma$ in which player $A$ has two pure strategies, $a_{1}$ and $a_{2}$, while player $B$ has just one pure strategy $b$. Define player $A$’s utility function by $\bm{u}_{A}(a_{1},b)=2$ and $\bm{u}_{A}(a_{2},b)=1$. Suppose we run PS-REG-0 with target error $\varepsilon\doteq 0.2$ and get the following: $\displaystyle\textup{Iteration 1:}\quad\boxed{\begin{aligned} &\hat{\bm{u}}^{(1)}_{A}(a_{1},b)=2.5\\\ &\hat{\bm{u}}^{(1)}_{A}(a_{2},b)=1.45\\\ &\hat{\varepsilon}^{(1)}=0.5\\\ \end{aligned}}$ $\displaystyle\quad\Longrightarrow\quad\begin{aligned} &\textup{Index $(A,(a_{2},b))$ regret pruned}\\\ &\textup{(since $\hat{\varepsilon}^{(1)}=0.5<\frac{\operatorname{Reg}_{A}((a_{2},b);\hat{\bm{u}}^{(1)})}{2}=0.525$)}\end{aligned}$ $\displaystyle\textup{Iteration 2:}\quad\boxed{\begin{aligned} &\hat{\bm{u}}^{(2)}_{A}(a_{1},b)=1.8\\\ &\hat{\bm{u}}^{(2)}_{A}(a_{2},b)=1.45\\\ &\hat{\varepsilon}^{(2)}=0.2\\\ \end{aligned}}$ $\displaystyle\quad\Longrightarrow\quad\begin{aligned} &\textup{Index $(A,(a_{1},b))$ well-estimated pruned;}\\\ &\textup{PS-REG-0{} terminates with $\hat{\bm{u}}_{A}=\hat{\bm{u}}^{(2)}_{A}$}\end{aligned}$ We have that $(a_{2},b)\in\textup{E}_{2\varepsilon}(\hat{\bm{u}})$, since $\operatorname{Reg}_{A}((a_{2},b);\hat{\bm{u}})=1.8-1.45=0.35<2\varepsilon=0.4$ and $\operatorname{Reg}_{B}((a_{2},b);\hat{\bm{u}})=0$. But we have $(a_{2},b)\not\in\textup{E}_{4\varepsilon}(\bm{u})$, since $\operatorname{Reg}_{A}((a_{2},b);\bm{u})=2-1=1>4\varepsilon=0.8$. Hence, we have $\textup{E}_{2\varepsilon}(\hat{\bm{u}})\not\subseteq\textup{E}_{4\varepsilon}(\bm{u})$. Since all utility deviation guarantees have been held, this is not a failure case. Thus, 1 cannot be true. Though the second inclusion in 1 cannot be guaranteed, we show that an alternative inclusion does hold (w.h.p.). ###### Theorem 5. If PS- REG-0$(\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},\updelta,\varepsilon)$ returns an empirical utility function $\hat{\bm{u}}$, then with probability at least $1-\updelta$, we have that $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})$ and $\textup{E}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon}(\bm{u})$. From this guarantee, we see that if a game analyst is using PS-REG-0 to learn an approximate pure Nash equilibrium of the simulation-based game, they will need to compute a (true) pure Nash equilibrium of the resulting empirical game. Of course, in many games, computing a pure Nash equilibrium is computationally intractable (e.g., it is NP-complete in graphical games (Gottlob, Greco, and Scarcello 2005)); the best that can be hoped for is an approximate pure Nash equilibrium. Furthermore, even if the simulation-based game has a pure Nash equilibrium, it is not then guaranteed that the empirical game will also have a pure Nash equilibrium, but rather only that it will have a $2\varepsilon$-pure Nash equilibrium. A limitation of PS-REG-0 is that it is not able to provide any guarantees regarding the efficacy of approximate pure Nash equilibria from the empirical game in the true game. ### New Regret Pruning Variations We now introduce our novel regret pruning criteria. We begin by presenting a variation which resolves the limitation observed in PS-REG-0 of lacking guarantees regarding empirical approximate Nash equilibria. This variation is derived on the basis of a stronger version of Lemma 1. ###### Lemma 2. Let $\gamma^{*}\geq 0$. If $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$ satisfying $\operatorname{Reg}_{p}(\bm{s};\bm{u})=0$ or $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})\leq\gamma^{*}$, then $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})$ and $\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u})$ for all $0\leq\gamma\leq\gamma^{*}$. Whereas Lemma 1 required all indices to be well-estimated, Lemma 2 requires only indices with sufficiently low regret in both the true game and empirical game to be well-estimated. This gives room for indices with provably high regret to be regret-pruned. Of course, as Areyan Viqueira, Cousins, and Greenwald (2020) also observed, this potential for regret pruning seems to come at the cost of any guarantees regarding mixed Nash equilibria. On the basis of Lemma 2, we design our first regret pruning variation. ###### Theorem 6. Consider a PS algorithm, PS-REG, using uniform utility deviation bounds, which conducts well-estimated pruning and on each iteration $t$, also _regret- prunes_ any index $(p,\bm{s})\in\bm{\mathcal{I}}$ which satisfies $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}}^{(t)})>\max\\{2\hat{\varepsilon}^{(t)},\gamma^{*}+\varepsilon+\hat{\varepsilon}^{(t)}\\}\enspace.$ If PS- REG$(\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},\updelta,\varepsilon,\gamma^{*})$ returns an empirical utility function $\hat{\bm{u}}$, then with probability at least $1-\updelta$, it holds that $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})$ and $\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u})$ for all $0\leq\gamma\leq\gamma^{*}$. Notice that when $\gamma^{*}=0$, PS-REG is identical to PS-REG-0, even yielding the same exact guarantees. PS-REG is thus a generalization of PS- REG-0 to cases where $\gamma^{*}>0$. When $\gamma^{*}>0$, PS-REG yields, at the cost of potentially reduced pruning, a stronger guarantee upon termination than PS-REG-0, ensuring that even an approximate empirical pure Nash equilibrium (so long as it is at worst a $\gamma^{*}$-pure Nash equilibrium) will be an approximate pure Nash equilibrium in the true game. In practice, the parameter $\gamma^{*}$ can be set to the smallest value for which the game analyst is still certain they will be able to compute a $\gamma^{*}$-pure Nash equilibria of the empirical game. If we set $\gamma^{*}=2\varepsilon$, we get the dual pure Nash containment guarantee, $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})\subseteq\textup{E}_{4\varepsilon}(\bm{u})$, which PS-REG-0 was originally designed to meet. One limitation of both PS-REG and PS-REG-0 is that they both depend only on uniform utility deviation bounds. The next algorithm and regret pruning variation takes advantage of non-uniform utility deviation bounds (Theorem 3) to prune potentially more (and in practice significantly more; see Figure 1) indices via both well-estimated pruning and regret-pruning, while yielding the same guarantees as PS-REG. In the following result, we use $\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}})$ to denote the high- probability lower-bound on $\operatorname{Reg}_{p}(\bm{s};\bm{u})$ defined by $\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}})\doteq\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}(\bm{u}_{p}(\bm{s}^{\prime})-\varepsilon_{p}(\bm{s}^{\prime}))-(\bm{u}_{p}(\bm{s})+\varepsilon_{p}(\bm{s})).$ ###### Theorem 7. Consider a PS algorithm, PS-REG+, using non-uniform utility deviation bounds, which conducts well-estimated pruning and on each iteration $t$, also _regret- prunes_ any index $(p,\bm{s})\in\bm{\mathcal{I}}$ which satisfies $\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}}^{(t)})>\max\\{0,\gamma^{*}+\varepsilon-\hat{\varepsilon}^{(t)}_{p}(\bm{s})\\}\enspace.$ If PS- REG+$(\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},\updelta,\varepsilon,\gamma^{*})$ returns an empirical utility function $\hat{\bm{u}}$, then with probability at least $1-\updelta$, it holds that $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})$ and $\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u})$ for all $0\leq\gamma\leq\gamma^{*}$. Notice again that when $\gamma^{*}=0$, the pruning criterion becomes simply $\operatorname{Reg}_{p}^{\downarrow}(\bm{s};\hat{\bm{u}}^{(t)})>0$ and yields the same guarantees as PS-REG-0. Thus, when $\gamma^{*}=0$, PS-REG+ can also reasonably be called PS-REG-0+. Since all the aforementioned regret pruning variations derive from the result presented in Lemma 2, they are only able to provide guarantees regarding pure equilibria. This is the most glaring limitation of the new PS algorithms presented so far, since many games do not even have strong approximate pure Nash equilibria which the algorithms could potentially be used to learn. In contrast, PS-WE derives from Lemma 1, and thus yields mixed Nash containment guarantees, but of course does not allow for regret-pruning. We now present a lemma which serves as a middle ground between Lemma 2 and Lemma 1. ###### Lemma 3. If $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\right\\}$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$, then for all $0\leq\gamma\leq 2\varepsilon$, it holds that $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})\textup{ and }\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u}),$ and for all $\gamma\geq 0$, it holds that $\textup{E}^{\diamond}(\bm{u})\subseteq\textup{E}^{\diamond}_{4\varepsilon}(\hat{\bm{u}})\textup{ and }\textup{E}^{\diamond}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}^{\diamond}_{2\varepsilon+\frac{3\gamma}{2}}(\bm{u}).$ The condition in Lemma 3 is looser than that in Lemma 1, opening up the potential for regret-pruning, but is (strictly) tighter than the condition in Lemma 2 when $\gamma^{*}=2\varepsilon$ (i.e., empirical games which satisfy the condition in Lemma 3 necessarily satisfy the condition in Lemma 2 when $\gamma^{*}=2\varepsilon$, but the converse does not hold; see Appendix for proof), which allows for the additional mixed Nash containment guarantee. Notice further that unlike Lemma 2, Lemma 3 does not depend on any additional parameter $\gamma^{*}$ and yields guarantees for all $\gamma\geq 0$, rather than just $\gamma\in[0,\gamma^{*}]$. We use Lemma 3 to derive yet another regret pruning variation. ###### Theorem 8. Consider a PS algorithm, PS-REG-M, using non-uniform utility deviation bounds, which conducts well-estimated pruning and on each iteration $t$, also _regret- prunes_ any index $(p,\bm{s})\in\bm{\mathcal{I}}$ which satisfies $\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}}^{(t)})>\varepsilon+\hat{\varepsilon}^{(t)}_{p}(\bm{s})\enspace.$ If PS- REG-M$(\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},\updelta,\varepsilon)$ returns an empirical utility function $\hat{\bm{u}}$, then with probability at least $1-\updelta$, for all $0\leq\gamma\leq 2\varepsilon$, it holds that $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})\textup{ and }\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u}),$ and for all $\gamma\geq 0$, it holds that $\textup{E}^{\diamond}(\bm{u})\subseteq\textup{E}^{\diamond}_{4\varepsilon}(\hat{\bm{u}})\textup{ and }\textup{E}^{\diamond}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}^{\diamond}_{2\varepsilon+\frac{3\gamma}{2}}(\bm{u}).$ The mixed Nash containment guarantee achieved by PS-REG-M is a $\frac{\gamma}{2}$ factor looser than that achieved by PS-WE (Lemma 1). As a result, slightly stronger approximate empirical mixed Nash equilibria will need to be computed when using PS-REG-M in order to guarantee an equally strong approximate true mixed Nash equilibrium as in PS-WE. ### Efficiency Bounds and Correctness Using similar proof techniques to those used in Cousins et al. (2022) to derive efficiency bounds for PS-WE, we derive upper bounds on the number of samples each utility index requires prior to being pruned by PS-REG+ and PS- REG-M, respectively. In the following results, suppose that the utility deviation bounds being used by the mentioned PS algorithms are the minimum of Hoeffding bounds (Theorem 1) and empirical Bennett bounds (Theorem 3). ###### Theorem 9. When running PS- REG+$(\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},\updelta,\varepsilon,\gamma^{*})$, with probability at least $1-\frac{\updelta}{3}$, the index $(p,\bm{s})\in\bm{\mathcal{I}}$ will be pruned prior to the first iteration $t$ with cumulative sample size $M_{t}\geq$ $2+2\ln\smash{\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}{\updelta}}\min\begin{cases}\frac{10c}{\operatorname{Reg}_{p}(\bm{s};\bm{u})-\gamma^{*}}+\frac{25\left\lVert{}\bm{v}_{p}(\operatorname{Adj}_{p,\bm{s}})\right\rVert{}_{\infty}}{(\operatorname{Reg}_{p}(\bm{s};\bm{u})-\gamma^{*})^{2}}\\\ \frac{5c}{2\varepsilon}+\frac{\bm{v}_{p}(\bm{s})}{\varepsilon^{2}}\end{cases}$ (defaulting to the second option when $\operatorname{Reg}_{p}(\bm{s};\bm{u})\leq\gamma^{*}$). When running PS- REG-M$(\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},\updelta,\varepsilon)$, with probability at least $1-\frac{\updelta}{3}$, the index $(p,\bm{s})\in\bm{\mathcal{I}}$ will be pruned prior to the first iteration $t$ with cumulative sample size $M_{t}\geq$ $2+2\ln\smash{\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}{\updelta}}\min\begin{cases}\frac{12.5c}{\operatorname{Reg}_{p}(\bm{s};\bm{u})-\varepsilon}+\frac{25\left\lVert{}\bm{v}_{p}(\operatorname{Adj}_{p,\bm{s}})\right\rVert{}_{\infty}}{(\operatorname{Reg}_{p}(\bm{s};\bm{u})-\varepsilon)^{2}}\\\ \frac{5c}{2\varepsilon}+\frac{\bm{v}_{p}(\bm{s})}{\varepsilon^{2}}\end{cases}$ queries at profile $\bm{s}$ (defaulting to the second option when $\operatorname{Reg}_{p}(\bm{s};\bm{u})\leq\varepsilon$). The above result reinforces the idea that care is required when choosing a sampling schedule for these algorithms. If the marginal sample size $m_{t}$ is small at each iteration $t$, then very few queries will be wasted from when an index is ready to be pruned to when it is actually pruned by the algorithm. On the other hand, if very small marginal sample sizes are used, then a very large schedule length $T$ will be required to reach a sufficiently large cumulative sample size to prune all utility indices. But a larger schedule length $T$ yields looser utility deviation bounds, thus resulting in _all_ indices requiring more queries to be pruned than otherwise. Hence, there is a trade-off in designing a sampling schedule between keeping marginal sample sizes small and keeping the schedule length small. We discuss our particular choice of sampling schedule in the next section, and in greater detail in the Appendix. In the following result, we use Hoeffding’s Inequality to derive an upper bound on the requisite total cumulative sample size a sampling schedule needs to ensure that all utility indices will be pruned prior to its exhaustion. ###### Theorem 10. Suppose that the total samples $M_{T}$ allocated in the sampling schedule is greater than or equal to the maximum number of samples needed to prune an arbitrary index, i.e., $M_{T}\doteq\sum_{t=1}^{T}m_{t}\geq\frac{c^{2}\ln\frac{2\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}{\updelta}}{2\varepsilon^{2}}\enspace.$ Then for each algorithm $\textup{PS}\in$ $\\{\textup{PS-WE{}, PS-REG-0{}, PS-REG{}, PS-REG+{}, PS-REG-M{}}\\},$ it is guaranteed that PS$(\Gamma_{\mathcal{Y}},\mathscr{D},\bm{\mathcal{I}},\updelta,\varepsilon)$ will terminate and return an empirical game with the guarantees corresponding to the respective algorithm. ## Experiments Figure 1: Average query complexity vs $\nicefrac{{1}}{{\varepsilon}}$ for 10 runs of each algorithm for each target error $\varepsilon\in\\{\frac{c}{2},\frac{c}{3},\dots,\frac{c}{100}\\}$. All algorithm runs are conducted on a single randomly-generated two-player zero- sum game with non-uniform additive noise. In this section, we experimentally explore the behavior of all the aforementioned PS algorithms. When choosing a sampling schedule for each algorithm, we follow Cousins et al. (2022) in designing sampling schedules which begin at sample size $\alpha$, a lower bound on the minimum number of samples needed to prune an arbitrary index, and end with a cumulative sample size $\omega$, an upper bound on the maximum number of samples needed to prune an arbitrary index. For all algorithms, we use the upper bound from Theorem 10 for $\omega$, i.e., $\omega\doteq\frac{c^{2}\ln\frac{2\left\lvert{}\Gamma\right\rvert{}T}{\updelta}}{2\varepsilon^{2}}$, thus guaranteeing that each algorithm will return an empirical game satisfying the respective guarantees of the algorithm upon termination. For algorithms using regret pruning, we set $\alpha$ to be a lower bound on the number of samples to estimate a zero-variance utility to (w.h.p.) within $\frac{c}{2}$ error (since no regret pruning can occur prior to at least one index achieving such an error guarantee; see Appendix for proof). For PS-WE, we follow Cousins et al. (2022) in setting $\alpha$ to be a lower bound on the number of samples needed to estimate a zero-variance utility to (w.h.p.) within a target error $\varepsilon$, though we improve their lower bound by a small constant factor. Finally, while our PS-WE sampling schedule has a geometric sampling schedule (i.e., geometrically increasing cumulative sample size) as in Cousins et al. (2022), using such a geometric schedule for our regret pruning algorithms results in too many iterations spent on very small sample sizes, yielding looser bounds with very few additional queries saved in return. To correct for this, our regret pruning algorithms use a sampling schedule which is linear until it reaches the corresponding $\alpha$ derived for PS-WE, and then follows the same geometric sampling schedule used in PS-WE. For more details regarding the sampling schedules, see the Appendix. ### Query Complexity vs. Target Error In the following experiments, we test our algorithms on two-player random zero-sum games (generated via the game-generator GAMUT (Nudelman et al. 2004)) with 40 actions for each player and utility values in the range $[-2,2]$. In order to emulate a noisy simulator, we add noise to each sample utility value. For each utility index $(p,\bm{s})\in\bm{\mathcal{I}}$, we sample a variance modifier $\nu_{p,\bm{s}}\sim\textup{Beta(1.5, 3)}$, and then each time the simulator is queried for $\dot{\bm{u}}(\bm{s})\sim\mathscr{S}(\bm{s})$, we set $\dot{\bm{u}}_{p}(\bm{s})\doteq\bm{u}_{p}(\bm{s})+\mathcal{N}(\nu_{p,\bm{s}})$ where $\mathcal{N}(\nu_{p,\bm{s}})$ is a scaled and shifted Bernoulli random variable, generating either $10\nu_{p,\bm{s}}$ or $-10\nu_{p,\bm{s}}$ with equal probability. Notice then that our final utility range for these random zero-sum simulation-based games is $c=24$. Sampling variance modifiers from $\textup{Beta}(1.5,3)$ ensures that our utility indices have a wide range of noise variables with mostly moderate variance, but with some noise variables having particularly high variance and some particularly low variance. For our first experiment (Figure 1), we compare the query complexities (i.e., the number of simulation queries placed prior to termination) of our algorithms for varying target errors $\varepsilon$. We begin by generating a two-player random zero-sum game $\Gamma_{\mathscr{S}}$ (and variance modifiers $\nu_{p,\bm{s}}$ for each utility index $(p,\bm{s})\in\bm{\mathcal{I}}$). For each target error $\varepsilon\in\\{\frac{c}{2},\frac{c}{3},\dots,\frac{c}{100}\\}$, we then run each of our aforementioned PS algorithms (with failure probability $\updelta\doteq 0.05$) on $\Gamma_{\mathscr{S}}$ a total of 10 times and plot the average query complexity across those runs. For each algorithm, we then connect these average query complexities for each target error by a line plot in Figure 1. In Figure 1, we see that, as expected, PS-REG obtains its guarantees for approximate pure equilibria of the empirical game at the cost of a slightly greater query complexity than PS-REG-0. In a similar vein, PS-REG+ with $\gamma^{*}=2\varepsilon$ also requires a greater number of queries than PS- REG+ with $\gamma^{*}=0$ (i.e, PS-REG-0+) in order to yield its guarantees for $\gamma^{*}$-pure equilibria of the empirical game. We also observe that PS- REG-0 and PS-REG with $\gamma^{*}=2\varepsilon$, which both use uniform utility deviation bounds, consume significantly more queries than PS-WE, which takes advantage of non-uniform bounds but conducts no regret pruning. This suggests that utilizing non-uniform utility deviation bounds is crucial for designing query efficient progressive sampling algorithms. This idea is further reinforced when looking at PS-REG+ and PS-REG-M, both of which use regret pruning criteria that take advantage of non-uniform utility deviation bounds, and outperform PS-WE by a very significant margin, especially when the target error $\varepsilon$ is small. Another particularly surprising result is that PS-REG-M only consumes marginally more queries than PS-REG+ with $\gamma^{*}=2\varepsilon$, despite yielding strong mixed Nash containment guarantees in return. Figure 2: Average ratio of query complexity to query complexity of PS-WE vs $\nicefrac{{1}}{{\varepsilon}}$ for a single run of each algorithm on each of 10 randomly generated two-player zero-sum games with non-uniform additive noise for each target error $\varepsilon\in\\{\frac{c}{2},\frac{c}{3},\dots,\frac{c}{100}\\}$. Standard deviation bands are plotted and are just barely visible due to low variation. Since our first experiment only tests on a single randomly generated simulation-based game, it is possible that the generated game was just particularly amenable to regret pruning. For our second experiment (Figure 2), we observe the proportion of additional queries our new PS algorithms are able to save on average (across 10 random zero-sum simulation-based games) in comparison to PS-WE. This time, we run each PS algorithm (again with failure probability $\updelta\doteq 0.05$) only once for each generated game. In Figure 2, we observe that the comparative performance of the algorithms in Figure 1 remain consistent across many different random two-player zero-sum games. We further observe that past a certain turning point, progressive sampling algorithms which use regret pruning techniques save a greater proportion of queries with respect to PS-WE as smaller target errors $\varepsilon$ are used. When the target error $\varepsilon$ is very small, PS- REG+ (both $\gamma^{*}=0$ and $\gamma^{*}=2\varepsilon$) and PS-REG-M are able to obtain their respective guarantees while saving more than 50% and 40%, respectively, of the queries used by PS-WE. ## Conclusion In this paper, we address a serious limitation of Areyan Viqueira, Cousins, and Greenwald’s progressive sampling algorithm with regret pruning – it is only able to yield guarantees regarding _true_ pure Nash equilibria of the empirical game. We design two primary novel progressive sampling algorithms for practitioners to use to learn equilibria in simulation-based games: PS- REG+ and PS-REG-M. PS-REG+ combines well-estimated pruning with a novel regret pruning variation which is modified to ensure the algorithm yields pure Nash containment guarantees for approximate $\gamma^{*}$-pure Nash equilibria of the empirical game and to take advantage of non-uniform utility deviation bounds to prune utility indices as soon as possible. When using PS-REG+, a game analyst will set $\gamma^{*}$ according to the weakest approximate pure Nash equilibria for which they desire guarantees. PS-REG-M also incorporates well-estimated pruning and a novel regret pruning variation, except unlike PS- REG+, it yields strong Nash containment guarantees for all approximate pure or mixed Nash equilibria of the empirical game, at the cost of a slightly greater query complexity. Both PS-REG+ and PS-REG-M significantly outperform PS-WE, the prior state-of-the-art algorithm for learning equilibria in simulation- based games. In light of this, game analysts seeking such an algorithm should use PS-REG+ if they only aim to learn pure Nash equilibria and should otherwise use PS-REG-M. In this work, we have only applied our progressive sampling algorithms and pruning techniques to normal-form games. In future work, we aim to extend this methodology to other game models such as extensive-form games. Additionally, EGTA algorithms for learning equilibria of simulation-based game have thus far been completely detached from algorithms for computing equilibria. A game analyst must first use EGTA algorithms to learn a sufficiently strong approximation of the simulation-based game, and then must compute equilibria of this empirical game. Future work can incorporate variance-sensitive and regret-sensitive progressive sampling techniques into an existing game-solving algorithm to make an EGTA-aware game-solving algorithm. ## References * Areyan Viqueira, Cousins, and Greenwald (2020) Areyan Viqueira, E.; Cousins, C.; and Greenwald, A. 2020. Improved Algorithms for Learning Equilibria in Simulation-Based Games. In _Proceedings of the 19th International Conference on Autonomous Agents and MultiAgent Systems_ , 79–87. * Areyan Viqueira et al. (2019) Areyan Viqueira, E.; Cousins, C.; Mohammad, Y.; and Greenwald, A. 2019. Empirical Mechanism Design: Designing Mechanisms from Data. In _UAI_ , 406. AUAI Press. * Areyan Viqueira et al. (2019) Areyan Viqueira, E.; Greenwald, A.; Cousins, C.; and Upfal, E. 2019. Learning Simulation-Based Games from Data. In _Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems_ , 1778–1780. International Foundation for Autonomous Agents and Multiagent Systems. * Cousins et al. (2022) Cousins, C.; Mishra, B.; Viqueira, E. A.; and Greenwald, A. 2022. Computational and Data Requirements for Learning Generic Properties of Simulation-Based Games. _arXiv preprint arXiv:2208.06400_. * Fearnley et al. (2015) Fearnley, J.; Gairing, M.; Goldberg, P. W.; and Savani, R. 2015. Learning equilibria of games via payoff queries. _The Journal of Machine Learning Research_ , 16(1): 1305–1344. * Gatti and Restelli (2011) Gatti, N.; and Restelli, M. 2011. Equilibrium Approximation in Simulation-Based Extensive-Form Games. In _The 10th International Conference on Autonomous Agents and Multiagent Systems - Volume 1_ , AAMAS ’11, 199–206. Richland, SC: International Foundation for Autonomous Agents and Multiagent Systems. ISBN 0982657153. * Gottlob, Greco, and Scarcello (2005) Gottlob, G.; Greco, G.; and Scarcello, F. 2005. Pure Nash Equilibria: Hard and Easy Games. In _AAAI_. * Jordan, Kiekintveld, and Wellman (2007) Jordan, P. R.; Kiekintveld, C.; and Wellman, M. P. 2007. Empirical game-theoretic analysis of the TAC supply chain game. In _Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems_ , 193. ACM. * Jordan and Wellman (2010) Jordan, P. R.; and Wellman, M. P. 2010. Designing an ad auctions game for the trading agent competition. In _Agent-Mediated Electronic Commerce. Designing Trading Strategies and Mechanisms for Electronic Markets_. Springer. * Ketter, Peters, and Collins (2013) Ketter, W.; Peters, M.; and Collins, J. 2013. Autonomous Agents in Future Energy Markets: The 2012 Power Trading Agent Competition. In _AAAI_. * Marchesi, Trovò, and Gatti (2020) Marchesi, A.; Trovò, F.; and Gatti, N. 2020. _Learning Probably Approximately Correct Maximin Strategies in Simulation-Based Games with Infinite Strategy Spaces_ , 834–842. Richland, SC: International Foundation for Autonomous Agents and Multiagent Systems. ISBN 9781450375184. * Nudelman et al. (2004) Nudelman, E.; Wortman, J.; Shoham, Y.; and Leyton-Brown, K. 2004. Run the GAMUT: A comprehensive approach to evaluating game-theoretic algorithms. In _Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems-Volume 2_ , 880–887. IEEE Computer Society. * Picheny, Binois, and Habbal (2016) Picheny, V.; Binois, M.; and Habbal, A. 2016. A Bayesian optimization approach to find Nash equilibria. _arXiv preprint arXiv:1611.02440_. * Tavares et al. (2016) Tavares, A.; Azpurua, H.; Santos, A.; and Chaimowicz, L. 2016. Rock, paper, starcraft: Strategy selection in real-time strategy games. In _The Twelfth AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment (AIIDE-16)_. * Tuyls et al. (2020) Tuyls, K.; Perolat, J.; Lanctot, M.; Hughes, E.; Everett, R.; Leibo, J. Z.; Szepesvári, C.; and Graepel, T. 2020. Bounds and dynamics for empirical game theoretic analysis. _Autonomous Agents and Multi-Agent Systems_ , 34(1): 7. * Tuyls et al. (2018) Tuyls, K.; Perolat, J.; Lanctot, M.; Leibo, J. Z.; and Graepel, T. 2018. A Generalised Method for Empirical Game Theoretic Analysis. _arXiv preprint arXiv:1803.06376_. * Vorobeychik, Kiekintveld, and Wellman (2006) Vorobeychik, Y.; Kiekintveld, C.; and Wellman, M. P. 2006. Empirical mechanism design: methods, with application to a supply-chain scenario. In _Proceedings of the 7th ACM conference on Electronic commerce_. ACM. * Vorobeychik and Wellman (2008) Vorobeychik, Y.; and Wellman, M. P. 2008. Stochastic search methods for Nash equilibrium approximation in simulation-based games. In _Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems-Volume 2_ , 1055–1062. International Foundation for Autonomous Agents and Multiagent Systems. * Vorobeychik, Wellman, and Singh (2007) Vorobeychik, Y.; Wellman, M. P.; and Singh, S. 2007. Learning payoff functions in infinite games. _Machine Learning_ , 67(1-2): 145–168. * Wellman (2006) Wellman, M. P. 2006. Methods for Empirical Game-Theoretic Analysis (extended abstract). In _aaai06_ , 1552–1556. * Wellman, Kim, and Duong (2013) Wellman, M. P.; Kim, T. H.; and Duong, Q. 2013. Analyzing incentives for protocol compliance in complex domains: A case study of introduction-based routing. _arXiv preprint arXiv:1306.0388_. * Wiedenbeck, Yang, and Wellman (2018) Wiedenbeck, B.; Yang, F.; and Wellman, M. P. 2018. A Regression Approach for Modeling Games with Many Symmetric Players. In _32nd AAAI Conference on Artificial Intelligence_. * Zhang and Sandholm (2021) Zhang, B. H.; and Sandholm, T. 2021. Finding and Certifying (Near-)Optimal Strategies in Black-Box Extensive-Form Games. _Proceedings of the AAAI Conference on Artificial Intelligence_ , 35(6): 5779–5788. ## Appendix A Appendix ### Notation To improve readability of the following proofs, we introduce a few notational short-hands. 1. 1. Given a utility index $(p,\bm{s})\in\bm{\mathcal{I}}$ and a utility function $\bm{u}$, let $\bm{s}^{*}_{\bm{u}}$ denote the best response of player $p$ to the opponents’ strategies in $\bm{s}$, i.e., $\bm{s}^{*}_{\bm{u}}\doteq\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\bm{u}_{p}(\bm{s}^{\prime})$. 2. 2. Given a mixed strategy profile $\bm{s}\in\bm{{S}^{\diamond}}$, player $p\in P$, and strategy $t\in S_{p}$, we let $\bm{s}|_{t}$ denote the strategy profile $\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}$ satisfying $\bm{s}^{\prime}_{p}=t$, and let $\operatorname*{\mathbb{P}}[t|\bm{s}]$ denote the probability that mixed strategy profile $\bm{s}$ assigns to strategy $t$. ### Nash Containment Lemmas We begin by proving the three different Nash containment lemmas. See 1 ###### Proof. Suppose that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$. Let $\gamma\geq 0$ be arbitrary, and suppose that $\bm{s}\in\textup{E}_{\gamma}(\bm{u})$. Then we have that $\displaystyle\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})$ $\displaystyle=\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\hat{\bm{u}}_{p}(\bm{s})$ $\displaystyle\leq(\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})+\varepsilon)-(\bm{u}_{p}(\bm{s})-\varepsilon)$ By assumption $\displaystyle\leq(\bm{u}_{p}(\bm{s})+\gamma+\varepsilon)-(\bm{u}_{p}(\bm{s})-\varepsilon)$ Since $\bm{s}\in\textup{E}_{\gamma}(\bm{u})$, it holds that $\bm{u}_{p}(\bm{s})+\gamma\geq\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})$ $\displaystyle=2\varepsilon+\gamma,$ and hence $\bm{s}\in\textup{E}_{2\varepsilon+\gamma}(\hat{\bm{u}})$. Since $\bm{s}$ was arbitrary, we have that $\textup{E}_{\gamma}(\bm{u})\subseteq\textup{E}_{2\varepsilon+\gamma}(\hat{\bm{u}})$. By completely analogous reasoning, we see that $\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u})$. ∎ See 2 ###### Proof. Suppose that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$ satisfying $\operatorname{Reg}_{p}(\bm{s};\bm{u})=0$ or $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})\leq\gamma^{*}$. Suppose that $\bm{s}\in\textup{E}(\bm{u})$. Then we have that $\displaystyle\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})$ $\displaystyle=\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\hat{\bm{u}}_{p}(\bm{s})$ $\displaystyle\leq(\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})+\varepsilon)-(\bm{u}_{p}(\bm{s})-\varepsilon)$ $\displaystyle\begin{aligned} &\textsf{$\operatorname{Reg}_{p}(\bm{s}^{*}_{\hat{\bm{u}}};\hat{\bm{u}})=0\leq\gamma^{*}$, so $(p,\bm{s}^{*}_{\hat{\bm{u}}})$ is well-estimated (WE)}\\\ &\textsf{$\operatorname{Reg}_{p}(\bm{s};\bm{u})=0$ (since $\bm{s}\in\textup{E}(\bm{u})$), so $(p,\bm{s})$ is WE}\end{aligned}$ $\displaystyle\leq(\bm{u}_{p}(\bm{s})+\varepsilon)-(\bm{u}_{p}(\bm{s})-\varepsilon)$ Since $\bm{s}\in\textup{E}(\bm{u})$, it holds that $\bm{u}_{p}(\bm{s})\geq\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})$ $\displaystyle=2\varepsilon,$ and hence $\bm{s}\in\textup{E}_{2\varepsilon}(\hat{\bm{u}})$. Since $\bm{s}$ was arbitrary, we have that $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})$. Now let $\gamma\in[0,\gamma^{*}]$ and suppose instead that $\bm{s}\in\textup{E}_{\gamma}(\hat{\bm{u}})$. Then we have that $\displaystyle\operatorname{Reg}_{p}(\bm{s};\bm{u})$ $\displaystyle=\bm{u}_{p}(\bm{s}^{*}_{\bm{u}})-\bm{u}_{p}(\bm{s})$ $\displaystyle\leq(\hat{\bm{u}}_{p}(\bm{s}^{*}_{\bm{u}})+\varepsilon)-(\hat{\bm{u}}_{p}(\bm{s})-\varepsilon)$ $\displaystyle\begin{aligned} &\textsf{$\operatorname{Reg}_{p}(\bm{s}^{*}_{\bm{u}};\bm{u})=0$ (by definition), so $(p,\bm{s}^{*}_{\bm{u}})$ is WE}\\\ &\textsf{$\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})\leq\gamma^{*}$ (since $\bm{s}\in\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{\gamma^{*}}(\hat{\bm{u}})$), so $(p,\bm{s})$ is WE}\end{aligned}$ $\displaystyle\leq(\hat{\bm{u}}_{p}(\bm{s})+\gamma+\varepsilon)-(\hat{\bm{u}}_{p}(\bm{s})-\varepsilon)$ Since $\bm{s}\in\textup{E}_{\gamma}(\hat{\bm{u}})$, it holds that $\hat{\bm{u}}_{p}(\bm{s})+\gamma\geq\hat{\bm{u}}_{p}(\bm{s}^{*}_{\bm{u}})$ $\displaystyle=2\varepsilon+\gamma,$ and hence $\bm{s}\in\textup{E}_{2\varepsilon+\gamma}(\bm{u})$. Since $\bm{s}$ was arbitrary, we have that $\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u})$ and we are done. ∎ Before proving the third Nash containment lemma, we prove an accessory lemma. ###### Lemma 4. Let $\varepsilon>0$ be arbitrary. Suppose that for all $(p,\bm{s})\in\bm{\mathcal{I}}$, it holds that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\right\\}\enspace.$ Then for all $(p,\bm{s})\in\bm{\mathcal{I}}$ satisfying $\operatorname{Reg}_{p}(\bm{s};\bm{u})=0$, it must hold that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$. ###### Proof. Suppose that $(p,\bm{s})\in\bm{\mathcal{I}}$ satisfies $\operatorname{Reg}_{p}(\bm{s};\bm{u})=0$. Since $\operatorname{Reg}_{p}(\bm{s}^{*}_{\hat{\bm{u}}};\hat{\bm{u}})=0$, we have that $\left\lvert{}\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})\right\rvert{}\leq\varepsilon$, which implies that $\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})\geq\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\varepsilon$. But by definition, we have that $\bm{u}_{p}(\bm{s})\geq\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})$, and hence, it must hold that $\bm{u}_{p}(\bm{s})\geq\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\varepsilon$. By hypothesis, we also have that $\bm{u}_{p}(\bm{s})\leq\hat{\bm{u}}_{p}(\bm{s})+\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\right\\}$. Chaining these two inequalities, we have that $\displaystyle\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\varepsilon$ $\displaystyle\leq\hat{\bm{u}}_{p}(\bm{s})+\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\right\\}$ $\displaystyle\Longleftrightarrow\quad$ $\displaystyle\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\hat{\bm{u}}_{p}(\bm{s})$ $\displaystyle\leq\varepsilon+\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\right\\}$ $\displaystyle\Longleftrightarrow\quad$ $\displaystyle\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})$ $\displaystyle\leq\varepsilon+\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\right\\}\enspace.$ If $\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}>\varepsilon$, then the second term in the max wins out, but solving for $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})$ immediately yields that $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})\leq 2\varepsilon$, a contradiction. Hence, it must hold that $\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\leq\varepsilon$, and hence that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$. ∎ See 3 ###### Proof. Suppose that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\right\\}$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$. We will first show the pure Nash containment result, and then show the mixed Nash containment result. Suppose that $\bm{s}\in\textup{E}({\bm{u}})$. Then we have that $\displaystyle\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})$ $\displaystyle=\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\hat{\bm{u}}_{p}(\bm{s})$ By definition $\displaystyle\leq(\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})+\varepsilon)-(\bm{u}_{p}(\bm{s})-\varepsilon)$ $\displaystyle\begin{aligned} &\textsf{Since $\operatorname{Reg}_{p}(\bm{s}^{*}_{\hat{\bm{u}}};\hat{\bm{u}})=0$, by hypothesis $(p,\bm{s}^{*}_{\hat{\bm{u}}})$ is WE}\\\ &\textsf{Since $\operatorname{Reg}_{p}(\bm{s};\bm{u})=0$, by \lx@cref{creftypecap~refnum}{lem:zero-regret} $(p,\bm{s})$ is WE.}\end{aligned}$ $\displaystyle\leq(\bm{u}_{p}(\bm{s})+\varepsilon)-(\bm{u}_{p}(\bm{s})-\varepsilon)$ Since $\bm{s}\in\textup{E}(\bm{u})$, it holds that $\bm{u}_{p}(\bm{s})\geq\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})$ $\displaystyle=2\varepsilon,$ and hence $\bm{s}\in\textup{E}_{2\varepsilon}(\hat{\bm{u}})$. Since $\bm{s}$ was arbitrary, we have that $\textup{E}(\bm{u})\subseteq\textup{E}_{2\varepsilon}(\hat{\bm{u}})$. Now let $\gamma\in[0,2\varepsilon]$ suppose instead that $\bm{s}\in\textup{E}_{\gamma}(\hat{\bm{u}})$. Then we have that $\displaystyle\operatorname{Reg}_{p}(\bm{s};\bm{u})$ $\displaystyle=\bm{u}_{p}(\bm{s}^{*}_{\bm{u}})-\bm{u}_{p}(\bm{s})$ By definition $\displaystyle\leq(\hat{\bm{u}}_{p}(\bm{s}^{*}_{\bm{u}})+\varepsilon)-(\hat{\bm{u}}_{p}(\bm{s})-\varepsilon)$ $\displaystyle\begin{aligned} &\textsf{Since $\operatorname{Reg}_{p}(\bm{s}^{*}_{\bm{u}};\bm{u})=0$, by \lx@cref{creftypecap~refnum}{lem:zero-regret} $(p,\bm{s}^{*}_{\bm{u}})$ is WE}\\\ &\textsf{Since $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})\leq\gamma\leq 2\varepsilon$ (by def. of $\textup{E}_{\gamma}(\hat{\bm{u}})$), by hypothesis $(p,\bm{s})$ is WE}\end{aligned}$ $\displaystyle\leq(\hat{\bm{u}}_{p}(\bm{s})+\gamma+\varepsilon)-(\hat{\bm{u}}_{p}(\bm{s})-\varepsilon)$ Since $\bm{s}\in\textup{E}_{\gamma}(\hat{\bm{u}})$, it holds that $\hat{\bm{u}}_{p}(\bm{s})+\gamma\geq\hat{\bm{u}}_{p}(\bm{s}^{*}_{\bm{u}})$ $\displaystyle=2\varepsilon+\gamma,$ and hence $\bm{s}\in\textup{E}_{2\varepsilon+\gamma}(\bm{u})$. Since $\bm{s}$ was arbitrary, we have that $\textup{E}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}_{2\varepsilon+\gamma}(\bm{u})$, and we have shown the pure Nash containment result. Now we will prove the mixed Nash containment result. Suppose that $\bm{s}\in\textup{E}^{\diamond}(\bm{u})$. Then we have that $\displaystyle\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})$ $\displaystyle=\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\hat{\bm{u}}_{p}(\bm{s})$ By definition $\displaystyle=\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\sum_{t\in S_{p}}\hat{\bm{u}}_{p}(\bm{s}|_{t})\operatorname*{\mathbb{P}}[t|\bm{s}]$ By definition (of utility at mixed strategy profile) $\displaystyle\leq(\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})+\varepsilon)-\sum_{t\in S_{p}}\left(\bm{u}_{p}(\bm{s}|_{t})-\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s}|_{t};\hat{\bm{u}})}{2}\right\\}\right)\operatorname*{\mathbb{P}}[t|\bm{s}]$ By hypothesis $\displaystyle\leq(\bm{u}_{p}(\bm{s})+\varepsilon)-\sum_{t\in S_{p}}\left(\bm{u}_{p}(\bm{s}|_{t})-\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s}|_{t};\hat{\bm{u}})}{2}\right\\}\right)\operatorname*{\mathbb{P}}[t|\bm{s}]$ Since $\bm{s}\in\textup{E}^{\diamond}(\bm{u})$, we have $\bm{u}_{p}(\bm{s})\geq\bm{u}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})$ $\displaystyle=\varepsilon+\sum_{t\in S_{p}}\operatorname*{\mathbb{P}}[t|\bm{s}]\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s}|_{t};\hat{\bm{u}})}{2}\right\\}$ By definition (of utility at mixed strategy profile) $\displaystyle\leq\varepsilon+\sum_{t\in S_{p}}\operatorname*{\mathbb{P}}[t|\bm{s}]\left(\varepsilon+\frac{\operatorname{Reg}_{p}(\bm{s}|_{t};\hat{\bm{u}})}{2}\right)$ $\max\\{A,B\\}\leq A+B$ $\displaystyle=2\varepsilon+\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}.$ By definition (of utility at mixed strategy profile) Solving the above for $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})$, we get that $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})\leq 4\varepsilon$, and hence $\bm{s}\in\textup{E}^{\diamond}_{4\varepsilon}(\hat{\bm{u}})$. Since $\bm{s}$ was arbitrary, we have that $\textup{E}^{\diamond}(\bm{u})\subseteq\textup{E}^{\diamond}_{4\varepsilon}(\hat{\bm{u}})$. Now instead let $\gamma\geq 0$, and suppose that $\bm{s}\in\textup{E}^{\diamond}_{\gamma}(\hat{\bm{u}})$. Then we have that $\displaystyle\operatorname{Reg}_{p}(\bm{s};\bm{u})$ $\displaystyle=\bm{u}_{p}(\bm{s}^{*}_{\bm{u}})-\bm{u}_{p}(\bm{s})$ By definition $\displaystyle=\bm{u}_{p}(\bm{s}^{*}_{\bm{u}})-\sum_{t\in S_{p}}\bm{u}_{p}(\bm{s}|_{t})\operatorname*{\mathbb{P}}[t|\bm{s}]$ By definition (of utility at mixed strategy profile) $\displaystyle\leq(\hat{\bm{u}}_{p}(\bm{s}^{*}_{\bm{u}})+\varepsilon)-\sum_{t\in S_{p}}\left(\hat{\bm{u}}_{p}(\bm{s}|_{t})-\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s}|_{t};\hat{\bm{u}})}{2}\right\\}\right)\operatorname*{\mathbb{P}}[t|\bm{s}]$ Hypothesis + Lemma 4 $\displaystyle\leq(\hat{\bm{u}}_{p}(\bm{s})+\gamma+\varepsilon)-\sum_{t\in S_{p}}\left(\hat{\bm{u}}_{p}(\bm{s}|_{t})-\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s}|_{t};\hat{\bm{u}})}{2}\right\\}\right)\operatorname*{\mathbb{P}}[t|\bm{s}]$ Since $\bm{s}\in\textup{E}^{\diamond}_{\gamma}(\hat{\bm{u}})$, we have $\hat{\bm{u}}_{p}(\bm{s})+\gamma\geq\hat{\bm{u}}_{p}(\bm{s}^{*}_{\bm{u}})$ $\displaystyle=\gamma+\varepsilon+\sum_{t\in S_{p}}\operatorname*{\mathbb{P}}[t|\bm{s}]\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s}|_{t};\hat{\bm{u}})}{2}\right\\}$ By definition (of utility at mixed strategy profile) $\displaystyle\leq\gamma+\varepsilon+\sum_{t\in S_{p}}\operatorname*{\mathbb{P}}[t|\bm{s}]\left(\varepsilon+\frac{\operatorname{Reg}_{p}(\bm{s}|_{t};\hat{\bm{u}})}{2}\right)$ $\max\\{A,B\\}\leq A+B$ $\displaystyle=\gamma+2\varepsilon+\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}$ By definition (of utility at mixed strategy profile) $\displaystyle\leq\frac{3\gamma}{2}+2\varepsilon\enspace,$ Since $\bm{s}\in\textup{E}_{\gamma}(\hat{\bm{u}})$, we have $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})\leq\gamma$ and hence $\bm{s}\in\textup{E}^{\diamond}_{2\varepsilon+\frac{3\gamma}{2}}(\bm{u})$. Since $\bm{s}$ was arbitrary, we have that $\textup{E}^{\diamond}_{\gamma}(\hat{\bm{u}})\subseteq\textup{E}^{\diamond}_{2\varepsilon+\frac{3\gamma}{2}}(\bm{u})$. ∎ In the text, we claim that the condition in Lemma 3 is strictly tighter than the condition in Lemma 2 when $\gamma^{*}=2\varepsilon$. We show this by proving the forward implication from Lemma 3’s condition to Lemma 2’s condition, and then showing the reverse implication to be false. Suppose the condition from Lemma 3, i.e., that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\right\\}$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$. Then by Lemma 4, we have that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$ with $\operatorname{Reg}_{p}(\bm{s};\bm{u})=0$. But by the hypothesis, we directly have that for all $(p,\bm{s})\in\bm{\mathcal{I}}$ with $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})\leq 2\varepsilon$, it holds that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})}{2}\right\\}\leq\max\\{\varepsilon,\varepsilon\\}=\varepsilon$. Hence, the condition from Lemma 2 is satisfied. But obviously the converse does not hold, since indices with $\operatorname{Reg}_{p}(\bm{s};\bm{u})>0$ and $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}})>2\varepsilon$ can have arbitrarily bad approximations (e.g., the empirical utilities can be arbitrarily low), even if the condition in Lemma 2 holds. ### Algorithm Correctness Proofs A crucial component to all of our progressive sampling algorithm correctness proofs is the use of a union bound to ensure that (w.h.p.) all pruning that occurs is justified with respect to the true game. We see this line of reasoning in the following correctness proof for PS-WE. See 4 ###### Proof. Recall that for each index $(p,\bm{s})\in\bm{\mathcal{I}}$ and iteration $t\in\\{1,\dots,T\\}$, the scalar $\hat{\varepsilon}^{(t)}_{p}(\bm{s})$ is an upper bound on $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}^{(t)}_{p}(\bm{s})\right\rvert{}$ with probability at least $1-\frac{\updelta}{\left\lvert{}\Gamma\right\rvert{}T}$. Thus, via a union bound, with probability $1-\updelta$, it holds _for all_ indices $(p,\bm{s})\in\bm{\mathcal{I}}$ and iterations $t\in\\{1,\dots,T\\}$ that if $\hat{\bm{u}}_{p}^{(t)}(\bm{s})$ has been computed (i.e., the index hasn’t been pruned on a prior iteration), it satisfies $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}^{(t)}_{p}(\bm{s})\right\rvert{}\leq\hat{\varepsilon}^{(t)}_{p}(\bm{s})$. Now suppose that PS-WE returns an empirical utility function $\hat{\bm{u}}$. This implies that PS-WE managed to prune all utility indices prior to the exhaustion of its sampling schedule. Combining this with the result from the previous paragraph, we get that with probability at least $1-\updelta$, for all $(p,\bm{s})\in\bm{\mathcal{I}}$, if we let $t$ denote the iteration on which $(p,\bm{s})$ was well-estimated pruned, it holds that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}=\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}^{(t)}_{p}(\bm{s})\right\rvert{}\leq\hat{\varepsilon}^{(t)}_{p}(\bm{s})\leq\varepsilon.$ The correctness guarantee then follows from Lemma 1 ∎ As the same union bound argument is applied in all of the following correctness proofs, we do not repeat it again. We begin by proving the correctness of our PS-REG+ algorithm, and show PS-REG and PS-REG-0 to simply be special cases of this algorithm. Before proving the correctness of PS-REG+, we prove another accessory lemma. ###### Lemma 5. Suppose that for all $(p,\bm{s})\in\bm{\mathcal{I}}$, it holds that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon_{p}(\bm{s})$. Then for all $(p,\bm{s})\in\bm{\mathcal{I}}$, we have that $\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\left(\hat{\bm{u}}_{p}(\bm{s}^{\prime})-\varepsilon_{p}(\bm{s}^{\prime})\right)\leq\bm{u}_{p}(\bm{s}^{*}_{\bm{u}})\leq\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\left(\hat{\bm{u}}_{p}(\bm{s}^{\prime})+\varepsilon_{p}(\bm{s}^{\prime})\right)$ ###### Proof. Let $(p,\bm{s})\in\bm{\mathcal{I}}$. We have that $\displaystyle\bm{u}_{p}(\bm{s}^{*}_{\bm{u}})=\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\bm{u}_{p}(\bm{s}^{\prime})\geq\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}(\hat{\bm{u}}_{p}(\bm{s}^{\prime})-\varepsilon_{p}(\bm{s}^{\prime})).$ The second inequality holds by analogous reasoning. ∎ See 7 ###### Proof. Suppose that PS-REG+ returns an empirical utility function $\hat{\bm{u}}$, thus guaranteeing that all indices have been pruned. Suppose that an index $(p,\bm{s})\in\bm{\mathcal{I}}$ is regret-pruned on iteration $t$. Then, we have (w.h.p.) that $\displaystyle\operatorname{Reg}_{p}(\bm{s},\bm{u})$ $\displaystyle=\bm{u}_{p}(\bm{s}^{*}_{\bm{u}})-\bm{u}_{p}(\bm{s})$ $\displaystyle\geq\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s}^{\prime})-\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})\right)-\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s})+\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right)$ Lemma 5 \+ Definition of $\varepsilon^{(t)}_{p}(\bm{s})$ $\displaystyle=\operatorname{Reg}^{\downarrow}_{p}(\bm{s},\hat{\bm{u}}^{(t)})>0$ $\displaystyle\textsf{Definition of $\operatorname{Reg}^{\downarrow}$ + Pruning Criterion}.$ We also have (w.h.p.) that $\displaystyle\operatorname{Reg}_{p}(\bm{s},\hat{\Gamma}_{\mathscr{S}})$ $\displaystyle=\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\hat{\bm{u}}_{p}(\bm{s})$ $\displaystyle=\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})-\hat{\bm{u}}^{(t)}_{p}(\bm{s})$ Since $(p,\bm{s})$ was pruned on iteration $i$ $\displaystyle\geq\hat{\bm{u}}_{p}(\bm{s}^{*}_{\bm{u}})-\hat{\bm{u}}^{(t)}_{p}(\bm{s})$ $\hat{\bm{u}}_{p}(\bm{s}^{*}_{\hat{\bm{u}}})\geq\hat{\bm{u}}_{p}(\bm{s}^{*}_{\bm{u}})$ by definition $\displaystyle\geq(\bm{u}_{p}(\bm{s}^{*}_{\bm{u}})-\varepsilon)-\hat{\bm{u}}^{(t)}_{p}(\bm{s})$ $\displaystyle\begin{aligned} &\textsf{$(p,\bm{s}^{*}_{\bm{u}})$ cannot have been regret pruned since $\operatorname{Reg}_{p}(\bm{s}^{*}_{\bm{u}},\Gamma_{\mathscr{S}})=0$.}\\\ &\textsf{Hence, it must have been WE pruned.}\end{aligned}$ $\displaystyle\geq\left(\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s}^{\prime})-\hat{\varepsilon}_{p}^{(t)}(\bm{s}^{\prime})\right)-\varepsilon\right)-\hat{\bm{u}}^{(t)}_{p}(\bm{s})$ $\displaystyle=\operatorname{Reg}^{\downarrow}_{p}(\bm{s},\hat{\Gamma}_{\mathscr{S}}^{(t)})+\hat{\varepsilon}^{(t)}_{p}(\bm{s})-\varepsilon$ By definition $\displaystyle>(\gamma^{*}+\varepsilon-\hat{\varepsilon}^{(t)}_{p}(\bm{s}))+\hat{\varepsilon}^{(t)}_{p}(\bm{s})-\varepsilon$ Pruning criterion $\displaystyle=\gamma^{*}.$ Since $(p,\bm{s})$ was arbitrary, we have (w.h.p.) that all regret-pruned indices have positive corresponding regret in the true game and greater than $\gamma^{*}$ corresponding regret in the empirical game. But since all indices have been pruned, this implies that, with probability at least $1-\updelta$, all indices $(p,\bm{s})\in\bm{\mathcal{I}}$ with $\operatorname{Reg}_{p}(\bm{s},\Gamma_{\mathscr{S}})=0$ or $\operatorname{Reg}_{p}(\bm{s},\hat{\Gamma}_{\mathscr{S}})\leq\gamma^{*}$ will be well-estimated pruned by PS-REG+ prior to termination, and will hence satisfy $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\varepsilon$. The correctness result then follows from Lemma 2. ∎ See 6 ###### Proof. Since PS-REG uses uniform utility deviation bounds, we have that for each index $(p,\bm{s})\in\bm{\mathcal{I}}$ and iteration $t$, it holds that $\hat{\varepsilon}^{(t)}_{p}(\bm{s})=\hat{\varepsilon}^{(t)}$. If we applied the regret pruning technique from PS-REG+ to such an algorithm, the pruning criterion would simplify as follows: $\displaystyle\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}}^{(t)})$ $\displaystyle>\max\\{0,\gamma^{*}+\varepsilon-\hat{\varepsilon}^{(t)}_{p}(\bm{s})\\}$ $\displaystyle\Longleftrightarrow$ $\displaystyle\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s}^{\prime})-\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})\right)-\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s})+\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right)$ $\displaystyle>\max\\{0,\gamma^{*}+\varepsilon-\hat{\varepsilon}^{(t)}_{p}(\bm{s})\\}$ $\displaystyle\Longleftrightarrow$ $\displaystyle\left[\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\bm{u}}^{(t)}_{p}(\bm{s}^{\prime})-\hat{\bm{u}}^{(t)}_{p}(\bm{s})\right]-2\hat{\varepsilon}^{(t)}$ $\displaystyle>\max\\{0,\gamma^{*}+\varepsilon-\hat{\varepsilon}^{(t)}\\}$ $\displaystyle\Longleftrightarrow$ $\displaystyle\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\bm{u}}^{(t)}_{p}(\bm{s}^{\prime})-\hat{\bm{u}}^{(t)}_{p}(\bm{s})$ $\displaystyle>2\hat{\varepsilon}^{(t)}+\max\\{0,\gamma^{*}+\varepsilon-\hat{\varepsilon}^{(t)}\\}$ $\displaystyle\Longleftrightarrow$ $\displaystyle\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}}^{(t)})$ $\displaystyle>\max\\{2\hat{\varepsilon}^{(t)},\gamma^{*}+\varepsilon+\hat{\varepsilon}^{(t)}\\}\enspace.$ But this is precisely the pruning criterion of PS-REG. Thus, PS-REG+ is simply a generalization of PS-REG to cases with non-uniform utility deviation bounds, and PS-REG must then yield the same guarantees as PS-REG+. ∎ See 5 ###### Proof. Consider the regret pruning criterion in PS-REG when $\gamma^{*}=0$. We know that the first iteration $t$ on which $\hat{\varepsilon}^{(t)}\leq\varepsilon$, all indices will be well-estimated pruned and the algorithm will terminate. Thus, regret pruning will only occur on iterations on which $\hat{\varepsilon}^{(t)}>\varepsilon$. But then the regret pruning criterion simplifies to $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}}^{(t)})>\max\\{2\hat{\varepsilon}^{(t)},\varepsilon+\hat{\varepsilon}\\}=2\hat{\varepsilon}^{(t)}$, which is precisely the pruning condition of PS-REG-0. Thus, PS-REG is simply a generalization of PS-REG-0 to $\gamma^{*}>0$, and hence PS-REG-0 must yield the same guarantees as PS-REG when $\gamma^{*}=0$. ∎ See 8 ###### Proof. Suppose an index $(p,\bm{s})\in\bm{\mathcal{I}}$ is regret-pruned on iteration $i$. In the proof of Theorem 7, we see that $\operatorname{Reg}^{\downarrow}_{p}(\bm{s},\hat{\Gamma}_{\mathscr{S}}^{(i)})>0$ implies that $\operatorname{Reg}_{p}(\bm{s},\Gamma_{\mathscr{S}})$. Hence, this pruning criteria is also guaranteed not to regret prune any index $(p,\bm{s}^{\prime})\in\bm{\mathcal{I}}$ satisfying $\operatorname{Reg}_{p}(\bm{s},\Gamma_{\mathscr{S}})=0$. Further following the proof of Theorem 7, we have that $\operatorname{Reg}_{p}(\bm{s},\hat{\Gamma}_{\mathscr{S}})\geq\operatorname{Reg}^{\downarrow}_{p}(\bm{s},\hat{\Gamma}_{\mathscr{S}}^{(i)})+\hat{\varepsilon}^{(i)}_{p}(\bm{s})-\varepsilon$, which when combined with our pruning criteria, yields (w.h.p.) that $\operatorname{Reg}_{p}(\bm{s},\hat{\Gamma}_{\mathscr{S}})\geq 2\hat{\varepsilon}^{(i)}_{p}(\bm{s})\geq 2\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}.$ Since $(p,\bm{s})$ was arbitrary, we have that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\frac{\operatorname{Reg}_{p}(\bm{s},\hat{\Gamma}_{\mathscr{S}})}{2}$ for each index $(p,\bm{s})\in\bm{\mathcal{I}}$ that is regret-pruned. Since the remaining indices must all be well-estimated, we have (w.h.p.) that for all $(p,\bm{s})\in\bm{\mathcal{I}}$, it holds that $\left\lvert{}\bm{u}_{p}(\bm{s})-\hat{\bm{u}}_{p}(\bm{s})\right\rvert{}\leq\max\left\\{\varepsilon,\frac{\operatorname{Reg}_{p}(\bm{s},\hat{\Gamma}_{\mathscr{S}})}{2}\right\\}.$ The conclusion then follows from Lemma 3. ∎ ### Efficiency Bounds (Cousins et al. 2022) derive high-probability sample complexity bounds for their empirical Bennett tail bounds. We state these sample complexity results below, and use them to derive our efficiency bounds for PS-REG+ and PS-REG-M. ###### Lemma 6. Consider an index $(p,\bm{s})\in\bm{\mathcal{I}}$. If the sample size $m_{\bm{s}}\geq 2+2\ln\frac{3}{\updelta}\left(\frac{5c}{2\varepsilon}+\frac{\bm{v}_{p}(\bm{s})}{\varepsilon^{2}}\right)$, then with probability at least $1-\frac{\updelta}{3}$, it will hold that $\varepsilon^{\hat{\textup{B}}}_{p}(\bm{s})\leq\varepsilon$. See 9 ###### Proof. Each of our efficiency bounds is presented as a minimum over two bounds, the first corresponding to regret pruning and the second to well-estimated pruning. It is clear that the second is a direct consequence of Lemma 6. We show the regret pruning bounds, beginning with PS-REG+. Recall that PS-REG+ prunes an index $(p,\bm{s})\in\bm{\mathcal{I}}$ on an iteration $t$ if $\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}}^{(t)})>\max\\{0,\gamma^{*}+\varepsilon-\hat{\varepsilon}^{(t)}_{p}(\bm{s})\\}$. We have (w.h.p.) that $\displaystyle\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}}^{(t)})$ $\displaystyle=\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s}^{\prime})-\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})\right)-\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s})+\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right)$ $\displaystyle>\left[\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\bm{u}}^{(t)}_{p}(\bm{s}^{\prime})\right]-\left[\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}_{p}^{(t)}(\bm{s}^{\prime})\right]-\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s})+\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right)$ $\displaystyle>\left[\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\left(\bm{u}_{p}(\bm{s})-\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right)\right]-\left[\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}_{p}^{(t)}(\bm{s}^{\prime})\right]-\left(\bm{u}_{p}(\bm{s})+2\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right)$ $\displaystyle>\left[\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\bm{u}_{p}(\bm{s}^{\prime})\right]-2\left[\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}_{p}^{(t)}(\bm{s}^{\prime})\right]-\left(\bm{u}_{p}(\bm{s})+2\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right)$ $\displaystyle=\operatorname{Reg}_{p}(\bm{s};\bm{u})-2\hat{\varepsilon}^{(t)}_{p}(\bm{s})-2\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})\enspace.$ Hence, a strictly tighter pruning criterion for PS-REG+ would be pruning an index $(p,\bm{s})\in\bm{\mathcal{I}}$ when $\displaystyle\operatorname{Reg}_{p}(\bm{s};\bm{u})-\gamma^{*}$ $\displaystyle>2\hat{\varepsilon}^{(t)}_{p}(\bm{s})+2\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})-\gamma^{*}+\max\\{0,\gamma^{*}+\varepsilon-\hat{\varepsilon}^{(t)}_{p}(\bm{s})\\}$ $\displaystyle=2\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})+\max\left\\{2\hat{\varepsilon}^{(t)}_{p}(\bm{s})-\gamma^{*},\varepsilon+\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right\\}\enspace.$ We can make the pruning criterion even tighter by increasing the right-hand side: $\displaystyle 2\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})+\max\left\\{2\hat{\varepsilon}^{(t)}_{p}(\bm{s})-\gamma^{*},\varepsilon+\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right\\}$ $\displaystyle\leq 4\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime}).$ Hence, the latest (w.h.p.) an index $(p,\bm{s})\in\bm{\mathcal{I}}$ will be regret-pruned by PS-REG+ is when $\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})<\frac{\operatorname{Reg}_{p}(\bm{s};\bm{u})-\gamma^{*}}{4}.$ Our result then follows via Lemma 6. By analogous reasoning, we have that a strictly tighter regret pruning criterion than the one in PS-REG-M would be pruning an index $(p,\bm{s})\in\bm{\mathcal{I}}$ when $\displaystyle\operatorname{Reg}_{p}(\bm{s};\bm{u})-\varepsilon>3\hat{\varepsilon}^{(t)}_{p}(\bm{s})+2\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime}).$ Similar to before, we can make the pruning criterion even tighter by increasing the right hand side: $3\hat{\varepsilon}^{(t)}_{p}(\bm{s})+2\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})\leq 5\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})$ Hence, the latest (w.h.p.) an index $(p,\bm{s})\in\bm{\mathcal{I}}$ will be regret pruned by PS-REG-M is when $\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})<\frac{\operatorname{Reg}_{p}(\bm{s};\bm{u})-\varepsilon}{5}\enspace.$ Once again, our result follows from Lemma 6. ∎ See 10 ###### Proof. Recall that all the aforementioned progressive sampling algorithms use well- estimated pruning. By Hoeffding’s Inequality (Theorem 1), we have that on iteration $T$ of algorithm PS, for all $(p,\bm{s})\in\bm{\mathcal{I}}$, it holds that $\displaystyle\hat{\varepsilon}^{(T)}_{p}(\bm{s})\leq c\sqrt{\frac{\ln\left(\frac{2\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}{\updelta}\right)}{M_{T}}}\leq c\sqrt{\ln\left(\frac{2\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}{\updelta}\right)\cdot\frac{2\varepsilon^{2}}{c^{2}\ln\left(\frac{2\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}{\updelta}\right)}}=\varepsilon,$ and thus all indices which remain active until iteration $T$ will be pruned on iteration $T$. Hence, the aforementioned algorithms are guaranteed to prune all indices prior to the exhaustion of the sampling schedule, and thus will return an empirical game satisfying the respective guarantees of the algorithm. ∎ ### Sampling Schedule Our sampling schedule is derived via a sample complexity lower bound for the empirical Bennett tail bounds presented in Theorem 3. (Cousins et al. 2022) lower bound the empirical Bennett bounds via the zero-variance case of Bennett’s inequality (Theorem 2). We, however, use a tighter lower bound which we derive below. ###### Lemma 7. Consider an index $(p,\bm{s})\in\bm{\mathcal{I}}$. If $\varepsilon^{\hat{\textup{B}}}_{p}(\bm{s})\leq\varepsilon$, then it must hold that the sample size $m_{\bm{s}}>\left(\frac{1}{3}+\sqrt{\frac{4+2\sqrt{3}}{3}}\right)\cdot\frac{c\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{\varepsilon}$. ###### Proof. We have that $\displaystyle\varepsilon^{\hat{\bm{v}}}_{p}(\bm{s})$ $\displaystyle=\frac{2c^{2}\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{3m}+\sqrt{\left(\frac{1}{3}+\frac{1}{2\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}\right)\left(\frac{c^{2}\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m-1}\right)^{2}+\frac{2c^{2}\hat{\bm{v}}_{p}(\bm{s})\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m}}$ $\displaystyle>\frac{2c^{2}\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{3m}+\sqrt{\frac{1}{3}\left(\frac{c^{2}\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m-1}\right)^{2}}$ $\displaystyle>\frac{2+\sqrt{3}}{3}\cdot\frac{c^{2}\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m}\enspace.$ We further have that $\displaystyle\varepsilon^{\hat{\textup{B}}}_{p}(\bm{s})$ $\displaystyle=\frac{c\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{3m}+\sqrt{\frac{2\left(\hat{\bm{v}}_{p}(\bm{s})+\varepsilon^{\hat{\bm{v}}}_{p}(\bm{s})\right)\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m}}$ $\displaystyle>\frac{c\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{3m}+\sqrt{\frac{2\cdot\frac{2+\sqrt{3}}{3}\cdot\frac{c^{2}\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m}\cdot\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m}}$ $\displaystyle=\left(\frac{1}{3}+\sqrt{\frac{4+2\sqrt{3}}{3}}\right)\cdot\frac{c\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}}{\updelta}\right)}{m}$ The conclusion follows directly. ∎ Thus, our sampling schedule for PS-WE begins at $\alpha=\left(\frac{1}{3}+\sqrt{\frac{4+2\sqrt{3}}{3}}\right)\cdot\frac{c\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}{\updelta}\right)}{\varepsilon}$ and ends at a cumulative sample size that is at least $\omega\doteq\frac{c^{2}\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}{\updelta}\right)}{2\varepsilon^{2}}$ to satisfy Theorem 10. Following (Cousins et al. 2022), we then use a schedule with a geometrically increasing cumulative sample size with a geometric factor $\beta$ (for our experiments, we use $\beta=1.1$). Our schedule length is then $T\doteq\lceil\log_{\beta}\left(\frac{\omega}{\alpha}\right)\rceil$. The first sample size is defined by $m_{1}\doteq\alpha\beta$, and each following sample size is defined by $m_{t}\doteq\alpha\beta^{t}-m_{t-1}$. Of course, for all of our regret pruning algorithms, it may be possible for regret pruning to occur prior to at least one index $(p,\bm{s})\in\bm{\mathcal{I}}$ achieving $\hat{\varepsilon}^{(t)}_{p}(\bm{s})\leq\varepsilon$. Regret pruning cannot, however, happen prior to at least one index $(p,\bm{s})\in\bm{\mathcal{I}}$ achieving $\hat{\varepsilon}^{(t)}_{p}(\bm{s})\leq\frac{c}{2}$. This can be seen by looking at the loosest regret pruning criterion we discuss, that used in PS-REG+ with $\gamma^{*}=0$. An index $(p,\bm{s})\in\bm{\mathcal{I}}$ is regret-pruned on iteration $t$ if $\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}}^{(t)})>0$. Notice that if $\hat{\varepsilon}^{(t)}_{p}(\bm{s})>\frac{c}{2}$ for all $(p,\bm{s})\in\bm{\mathcal{I}}$, then for any given index $(p,\bm{s})\in\bm{\mathcal{I}}$, we have that $\displaystyle\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}}^{(t)})$ $\displaystyle=\sup_{\bm{s}^{\prime}\in\operatorname{Adj}_{p,\bm{s}}}\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s}^{\prime})-\hat{\varepsilon}^{(t)}_{p}(\bm{s}^{\prime})\right)-\left(\hat{\bm{u}}^{(t)}_{p}(\bm{s})+\hat{\varepsilon}^{(t)}_{p}(\bm{s})\right)$ $\displaystyle<\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}}^{(t)})-c,$ and hence $(p,\bm{s})$ will be regret-pruned only if it holds that $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}}^{(t)})-c>\operatorname{Reg}^{\downarrow}_{p}(\bm{s};\hat{\bm{u}}^{(t)})>0$. But the latter is impossible since $\operatorname{Reg}_{p}(\bm{s};\hat{\bm{u}}^{(t)})\leq c$ by definition. Hence, no index can be regret-pruned (by any of our regret pruning criteria) prior to at least one index being achieving an estimation guarantee of at least $\frac{c}{2}$. Based on the above, we start our sampling schedule for all our regret pruning algorithms on $\alpha^{\prime}\doteq\left(\frac{1}{3}+\sqrt{\frac{4+2\sqrt{3}}{3}}\right)\cdot 2\ln\left(\frac{3\left\lvert{}\bm{\mathcal{I}}\right\rvert{}T}{\updelta}\right)$. But, as argued in the text, using a schedule with a strictly geometrically increasing schedule, we waste too many iterations on small sample sizes and yield a schedule length that is too large. Hence, we instead fix the schedule length to be $1.5$ times the schedule length used for PS-WE. We then occupy the final two-thirds of our schedule with the same sample sizes used in the sampling schedule for PS-WE, and occupy the first third of our schedule with one that has linearly increasing cumulative sample size beginning at $\alpha^{\prime}$ and ending at $\alpha$ (from above).
# Double layers in the Earth’s bow shock Jiepu Sun Space Sciences Laboratory University of California at Berkeley CA 94720, USA <EMAIL_ADDRESS> &Ivan Y. Vasko Space Sciences Laboratory University of California at Berkeley CA 94720, USA <EMAIL_ADDRESS> &Stuart D. Bale Space Sciences Laboratory University of California at Berkeley CA 94720, USA &Rachel Wang Space Sciences Laboratory University of California at Berkeley CA 94720, USA &Forrest S. Mozer Space Sciences Laboratory University of California at Berkeley CA 94720, USA ###### Abstract We present Magnetospheric Multiscale observations of electrostatic double layers in quasi-perpendicular Earth’s bow shock. These double layers have predominantly parallel electric field with amplitudes up to 100 mV/m, spatial widths of 50–700 m, and plasma frame speeds within 100 km/s. The potential drop across a single double layer is 2–7% of the cross-shock potential in the de Hoffmann-Teller frame and occurs over the spatial scale of ten Debye lengths or one tenth of electron inertial length. Some double layers can have spatial width of 70 Debye lengths and potential drop up to 30% of the cross- shock potential. The electron temperature variation observed across double layers is roughly consistent with their potential drop. While electron heating in the Earth’s bow shock occurs predominantly due to the quasi-static electric field in the de Hoffmann-Teller frame, these observations show that electron temperature can also increase across Debye-scale electrostatic structures. Key points: 1. 1. The first identification of ion-acoustic double layers and electron temperature variation across them in the Earth’s bow shock. 2. 2. Double layers have typical spatial width around ten Debye lengths or one tenth of electron inertial length. 3. 3. Typical potential drop across a double layer is 2–7% of the cross-shock potential in the de Hoffmann-Teller frame. ## 1 Introduction Electron heating remains one of the not entirely resolved problems in the physics of collisionless shock waves [1, 2, 3]. Previous spacecraft measurements showed that the electron heating across quasi-perpendicular Earth’s bow shock is mainly determined by the quasi-static cross-shock electrostatic field in the de Hoffmann-Teller frame [1, 4, 5]. This electrostatic field is rarely possible to be measured directly, but based on the observed electron heating we expect it to have typical amplitude of a few mV/m and spatial scale intermediate between electron and ion inertial lengths [6, 7, 2, 8]. The major effect of high-frequency fluctuations produced by plasma instabilities in the Earth’s bow shock is assumed to be a smoothing of the electron distribution function shaped by quasi-static magnetic and electric fields [1, 4, 5]. There are currently experimental [9, 8, 10] and theoretical [3, 11, 12] indications that high-frequency fluctuations not only smooth the electron distribution function, but can also contribute to the electron heating. In this study we present experimental evidence that a fraction of the electron heating in the Earth’s bow shock can occur across Debye-scale electrostatic double layers. Note that the term heating used in this letter is equivalent of temperature increase, because we observe spatial profiles of the electron temperature, rather than its temporal evolution. It is noteworthy that double layers in the form of more or less unipolar spikes in the parallel (magnetic field-aligned) electric field are universally observed during the energy release processes in space plasma [13]. Double layers were originally reported in the auroral region [14, 15], while later studies showed they had plasma frame speeds around local ion-acoustic speed and spatial widths of about ten Debye lengths [16, 17]. Similar ion-acoustic double layers were observed around plasma injection fronts in the inner magnetosphere [18, 19], fast plasma flows in the plasma sheet [20, 21], and reconnection current sheets in the magnetosphere [22, 23]. Parker Solar Probe measurements have recently revealed double layers, presumably of ion-acoustic type, in the Venusian bow shock [24]. The spacecraft measurements showed that ion-acoustic double layers can indeed accelerate or heat electrons [20, 21]. The presence of electrostatic double layers in the Earth’s bow shock was pointed out in the past [25, 26], but no in-depth analysis of their properties and the associated electron heating was carried out. In this letter we present Magnetospheric Multiscale measurements of electrostatic double layers in the Earth’s bow shock and demonstrate that they are of ion-acoustic type. We show that electron temperature variation across the double layers is roughly consistent with the potential drop across them. The origin of these double layers is discussed. ## 2 Observations We first consider a quasi-perpendicular Earth’s bow shock crossing by Magnetospheric Multiscale (MMS) spacecraft around 07:56:30 UT on 4th November 2015 that was previously used to study electrostatic solitary and ion-acoustic waves [27, 10]. We use burst mode measurements of the DC-coupled magnetic field at 128 S/s (samples per second) resolution provided by Digital and Analogue Fluxgate Magnetometers [28], electron and ion moments and velocity distribution functions provided at 0.03 and 0.15 s cadence respectively by the Fast Plasma Investigation instrument [29], electric field fluctuations along with voltage signals of individual voltage-sensitive probes at 8,192 S/s resolution provided by Axial Double Probe [30] and Spin-Plane Double Probe [31], and magnetic field fluctuations at 8,192 S/s resolution provided by the Search Coil Magnetometer [32]. The electric field was measured by two pairs of probes in the spin plane, with opposing probes mounted on tips of 60 m antennas, and a pair of axial probes mounted on tips of 14.6 m antennas along the spin axis. The electric field components in the orthogonal coordinate system associated with the probes, whose schematic can be found in Supporting Materials (SM), were computed as $E_{12}=(V_{2}-V_{1})/120\;{\rm m}$, $E_{34}=(V_{4}-V_{3})/120\;{\rm m}$, $E_{56}=(V_{6}-V_{5})/29.2\;{\rm m}$, and then multiplied by frequency response factors to be specified below. The electric and magnetic fields will also be presented in the Geocentric Solar Ecliptic (GSE) coordinate system whose $z-$axis essentially coincides with the spacecraft spin axis. Figure 1 presents MMS4 measurements in the aforementioned Earth’s bow shock crossing. Panel (a) presents the magnetic field magnitude and highlights a region around the ramp, where double layers were observed. The expanded view of that region is shown in following panels. Panel (b) demonstrates that the magnetic field is predominantly along the $y-$axis and, thus, almost within the spacecraft spin plane. Panel (c) shows that the plasma flows at about 100 km/s parallel to the local magnetic field and 200 km/s perpendicular to it. Three electric field components in panels (d)–(f) and their omnidirectional wavelet spectrum in panel (g) show a sporadic occurrence of electric field fluctuations at frequencies above a few hundred Hz. These are electrostatic ion-acoustic waves considered in the previous study [10]. The intriguing, though not universal, feature is that each burst of ion-acoustic waves is preceded by a solitary structure, whose typical frequency is below 200 Hz. These solitary structures are electrostatic, since no magnetic counterparts were observed, with the electric field predominantly along the $y-$axis which is parallel to the local magnetic field. We only present in-depth analysis of the solitary structure with the largest electric field amplitude, because similar analysis of other structures revealed they were of identical nature. Figure 1: MMS4 observation of double layers in a quasi-perpendicular Earth’s bow shock crossing on 4th November, 2015. Panel (a) presents the magnetic field magnitude observed across the shock along with a highlighted region, whose expanded view is shown in the following panels: (b), (c) magnetic field and plasma flow velocity in the GSE coordinate system; (d)–(f) three electric field components measured at 8,192 S/s in the GSE coordinate system and (g) their omnidirectional continuous Morse wavelet spectrum in arbitrary units. There are sporadic bursts of high-frequency electrostatic fluctuations (above about 200 Hz) preceded by lower-frequency electrostatic structures (below 200 Hz). These electrostatic structures are highlighted in panels (d)–(f) with the largest-amplitude one highlighted in red. Figure 2 presents an interferometry analysis of the largest-amplitude solitary structure highlighted in Figure 1. The electric field in the magnetic field- aligned coordinate in panel (a) demonstrates that the solitary structure has a full temporal width of 20 ms and predominantly parallel electric field with amplitude of 100 mV/m. Note electrostatic fluctuations with the temporal scale of about 1 ms, whose electric field is rather oblique to the electric field of the solitary structure. These high-frequency fluctuations can be excluded by a low-pass filter with the cutoff frequency of 200 Hz. Panel (b) presents electric field components $E_{12}$, $E_{34}$, and $E_{56}$ computed using low- pass filtered voltage signals and frequency response factors of 1.35 and 1.2 for spin plane and axial antennas respectively. The axial antenna’s factor may actually vary between 0.8 and 1.6 [27], but our results are not sensitive to its specific value, because the electric field of the solitary structure is essentially in the spin plane; moreover, it is directed from probe 2 to probe 1. Applying the Maximum Variance Analysis [33] to the electric field in panel (b), we determined a unit vector $\hat{\bf k}$ along the electric field polarization direction; in the coordinate system associated with the probes we have $\hat{\bf k}\approx(-0.91,0.12,0.39)$ that is within 20∘ of the local magnetic field. Low-pass filtered voltage signals are presented in panels (c)–(e) along with correlation coefficients and time delays between signals of opposing probes. The highest correlation and the largest time delay were observed between $V_{1}$ and $-V_{2}$. The speed of a locally planar electrostatic structure can be estimated as follows [34, 27]: $V_{s}=\hat{k}_{ij}l_{ij}/\Delta t_{ij}$, where $\Delta t_{ij}$ is the time delay between signals $V_{i}$ and $-V_{j}$ of a pair of opposing probes, while $l_{ij}$ is the corresponding antenna length. Taking into account that $\hat{k}_{12}\approx-0.91$, $\Delta t_{12}\approx-2.93$ ms, and $l_{12}=60$ m, we obtained the speed of the solitary structure in the spacecraft frame, $V_{s}\approx 19$ km/s. Note that the electric field polarization direction $\hat{\bf k}$, intrinsically ambiguous by 180∘, was chosen to be consistent with propagation from probe 1 to probe 2 as observed in panel (c). Voltage signals of other pairs of opposing probes are not perfectly correlated and corresponding time delays are less reliable. The estimated speed allowed translating temporal profiles into spatial ones and computing the electrostatic potential of the solitary structure, $\Phi=\int E_{k}V_{s}dt$, where $E_{k}$ is the electric field along the polarization direction. Panel (f) shows that the solitary structure has a net potential drop across itself of $\Delta\Phi\approx 25$ V; thus, it is actually a double layer. The spatial width of the double layer defined as $w=\Delta\Phi/{\rm max}(E_{k})$ is about 240 m, which is 15$\lambda_{D}$ or 0.2$\lambda_{e}$ in units of local Debye length or electron inertial length. With the spatial width defined that way, the electrostatic potential of the double layer can be approximated as $\Phi(X)\approx 0.5\Delta\Phi\left(1+{\rm tanh}(2X/w)\right)$, where $X$ is the spatial coordinate along the electric field polarization direction. Note that the spatial extent of the double layer in the plane perpendicular to the polarization direction is less than 10 km, since the double layer was not observed aboard the other MMS spacecraft spatially separated from MMS4 by at least 10 km in that plane. We estimated the speed of the double layer in the plasma frame as $V_{s}^{*}=|V_{s}-\hat{\bf k}\cdot{\bf V}_{i}|$, where ${\bf V}_{i}$ is the plasma flow velocity. We found $V_{s}^{*}\approx 90$ km/s or about 0.9$c_{\rm IA}$, where $c_{\rm IA}=(T_{e\parallel}/m_{p})^{1/2}\approx 100$ km/s is a rough estimation of the local ion-acoustic speed ($m_{p}$ is the proton mass). Thus, this double layer is of ion-acoustic type. The potential drop of 25 V is about 24% of the local electron parallel temperature and about 6% of the entire cross-shock potential $\Delta\Phi_{\rm HT}$ corresponding to the electric field in the de Hoffmann- Teller frame; the cross-shock potential of $\Delta\Phi_{\rm HT}\approx 400$ V was estimated using the electron momentum balance (SM). Figure 2: The interferometry analysis of the double layer highlighted in red in Figure 1: (a) the electric field measured at 8,192 S/s in the magnetic field-aligned coordinate system (components parallel and perpendicular to the local magnetic field are shown); (b) the electric field in the coordinate system associated with voltage-sensitive probes and computed using voltage signals $V_{1}$–$V_{6}$ of probes in panels (c)–(e): $E_{12}=1.35\cdot(V_{2}-V_{1})/120\;{\rm m}$, $E_{34}=1.35\cdot(V_{4}-V_{3})/120\;{\rm m}$, and $E_{56}=1.2\cdot(V_{6}-V_{5})/29.2\;{\rm m}$; (c)–(e) low-pass filtered (cutoff frequency of 200 Hz) and offset eliminated voltage signals of four probes in the spacecraft spin plane ($V_{1}$–$V_{4}$) and of two probes mounted on the axial antenna along the spin axis ($V_{5}$ and $V_{6}$); cross- correlation coefficients and time delays between voltage signals of opposing probes are indicated in panels; (f) the electric field $E_{k}$ along the polarization direction $\hat{\bf k}$ computed by applying the Maximum Variance Analysis to the electric field in panel (b) and the electrostatic potential computed as $\Phi=\int E_{k}V_{s}dt$, where $V_{s}\approx 19$ km/s is the double layer speed in the spacecraft frame estimated using the time delay between $V_{1}$ and $-V_{2}$. The estimated speed $V_{s}$ allowed translating the temporal axis into the spatial axis shown at the bottom. Figure 3: The analysis of the electron temperature variation across the double layer shown in Figure 2: (a) the electrostatic potential $\Phi=\int E_{k}V_{s}dt$ obtained by integrating the low-pass filtered (with cutoff frequency of 200 Hz) electric field $E_{k}$ over an interval about twice as long as the interval shown in Figure 2; (b) and (c) the electron density and the parallel temperature measured at 30 ms cadence; (d) and (e) the phase space density (PSD) of electrons collected upstream and downstream of the double layer over 30 ms intervals highlighted in panels (a)–(c). PSDs correspond to electrons with pitch angles within $(0^{\circ},30^{\circ})$ and $(150^{\circ},180^{\circ})$ in panel (d) and within $(60^{\circ},120^{\circ})$ in panel (e). The dashed line in panel (d) corresponds to the expected PSD downstream of the double layer that was computed using Liouville mapping of the upstream PSD (the energy of upstream electrons was increased by 25 eV, while their phase space density was conserved). The fluxes of electrons below about 30 eV were contaminated by secondary and photoelectrons. Figure 3 presents an analysis of the electron temperature variation and momentum balance across the double layer. Panel (a) shows the electrostatic potential obtained by integrating the low-pass filtered electric field over about 120 ms time interval, which is twice longer than that in Figure 2. Panels (b) and (c) present electron density $n_{e}$ and parallel temperature $T_{e||}$ over the same interval. Because the double layer can only be considered isolated over this interval (Figure 1e), there is merely one point of plasma measurements available upstream and downstream of it. The 30 ms time intervals where upstream and downstream electron distribution functions were collected are highlighted in panels (a)–(c). The projection of the Ohm’s law onto the local magnetic field indicates that the electrostatic field of the double layer must be balanced by the electron pressure gradient (e.g., [35]) $\displaystyle e\nabla_{{\color[rgb]{0,0,0}||}}\Phi\approx\nabla_{{\color[rgb]{0,0,0}||}}(n_{e}T_{e\parallel})/n_{e},$ (1) where we omitted negligible inertia terms and neglected anomalous resistivity; the latter is a reasonable assumption in the Earth’s bow shock [36, 37, 5]. The electron density increases across the double layer from 18 to 25 cm-3, while parallel electron temperature increases from 87 to 107 eV. Therefore, the expected potential drop is $\Delta\Phi\approx\Delta(n_{e}T_{e\parallel})/e\langle n_{e}\rangle\approx 50$ V, where $\langle n_{e}\rangle\approx 21.5$ cm-3 is the averaged electron density. The Ohm’s law prediction is consistent with the observed potential drop of 25 V within a factor of two. The analysis of electron distribution functions upstream and downstream of the double layer reveals the reason for the quantitative disagreement. Panel (d) presents the distribution function of upstream and downstream electrons streaming quasi-parallel to the local magnetic field line. In accordance with the presence of a net potential drop, the downstream distribution function is wider than the upstream one. It is wider however than the expected distribution computed via Liouville mapping of the upstream distribution. The expected distribution in panel (d) was obtained by increasing the energy of upstream electrons by 25 eV, while keeping their phase space density conserved. Note that the distribution of downstream electrons could not be computed below about 50 eV, because upstream electron fluxes below about 30 eV were contaminated by secondary and photoelectrons (e.g., Gershman et al. [38]). In addition, the distribution of downstream electrons below 25 eV would not be possible to compute using the Liouville mapping, because these electrons are trapped downstream of the double layer; see a schematics of the electron phase space in the SM. We do not expect any widening of the distribution function of electrons streaming quasi- perpendicular to the local magnetic field, but observations in panel (e) demonstrate the opposite. This is a strong indication of temporal variations that occurred within 90 ms between observations upstream and downstream of the double layer. The temporal variations are not necessarily local, because each distribution function was collected over 30 ms, which translates into spatial resolution of a few hundred kilometers ($>50$ eV electrons cover at least 100 km over 30 ms). We believe that the quantitative inconsistency within a factor of two between the Ohm’s law prediction and the observed potential drop is caused by temporal variations, which resulted in parallel temperature increase and widening of the distribution function larger than expected for the local potential drop of 25 V. Table 1: The table presents various properties of double layers shown in Figures 2 and 4: spacecraft frame speed $V_{s}$ and plasma frame speed $V_{s}^{*}$ along with its value in units of local ion-acoustic speed estimation, $c_{\rm IA}=(T_{e\parallel}/m_{p})^{1/2}$; spatial width $w$ in physical units as well as in units of local Debye length $\lambda_{D}$ and electron inertial length $\lambda_{e}$; potential drop $\Delta\Phi$ in physical units as well as in units of local parallel electron temperature $T_{e\parallel}$ and in units of cross-shock potential $\Delta\Phi_{\rm HT}$ in the de Hoffmann-Teller frame. | $V_{s}$ & $V_{s}^{*}$ [km/s] | $w$ [m] | $\Delta\Phi$ [V] | $w/\lambda_{D}$ | $w/\lambda_{e}$ | $e\Delta\Phi/T_{e\parallel}$ | $V_{s}^{*}/c_{\rm IA}$ | $\Delta\Phi/\Delta\Phi_{\rm HT}$ ---|---|---|---|---|---|---|---|--- Figure 2 | 19 & 90 | 240 | 25 | 15 | 0.23 | 24% | 0.9 | 6% Figure 4a | 52 & 104 | 90 | 7.5 | 8 | 0.1 | 8% | 1.1 | 2% Figure 4b | 51 & 62 | 670 | 32 | 67 | 0.9 | 33% | 0.6 | 32% Figure 4c | 27 & 50 | 55 | 3 | 9 | 0.07 | 9% | 0.9 | 5% Figure 4d | 12 & 95 | 65 | 2.2 | 13 | 0.1 | 7% | 1.8 | 7% Figure 4 demonstrates double layers observed in several other Earth’s bow shock crossings considered previously by Wang et al. [27]. The upper panel in each panel set (a)–(d) presents the magnetic field magnitude along with a highlighted region where the double layer was observed, while the middle panel presents the parallel electric field over that interval. These panels demonstrate that double layers can occur within or close to the shock ramp as well as in the downstream region. The electric fields of all double layers are parallel to the local magnetic field within 20∘ (SM). We carried out interferometry analysis to estimate speeds and electrostatic potentials of the double layers (SM). Bottom panels in (a)–(d) demonstrate electrostatic potentials, while various parameters of the double layers are presented in Table 1. These double layers have plasma and spacecraft frame speeds within 100 km/s, spatial widths within 50–700 m, and potential drops up to 30 V. They are of ion-acoustic type, because their plasma frame speeds are comparable with local ion-acoustic speed, $V_{s}^{*}/c_{\rm IA}\approx 0.5$–$2$. The typical spatial width of the double layers is 8–15$\lambda_{D}$ or 0.07–0.23$\lambda_{e}$, though some double layers can have spatial width up to 70$\lambda_{D}$ or 0.9$\lambda_{e}$. The potential drop across a single double layer is typically 2–7%, though can be up to 30%, of the cross-shock potential in the de Hoffmann-Teller frame. In units of local parallel electron temperature, these potential drops vary between 7% and 30%. For all the double layers the prediction of the Ohm’s law (1) is consistent within 30% with the observed potential drop; the only exception is the double layer in panel (d), whose potential drop is relatively small and the Ohm’s law estimation is most likely dominated by temporal variations (SM). Figure 4: Double layers observed in other quasi-perpendicular Earth’s bow shock crossings. In each set of panels (a)–(d), the upper panel presents the magnetic field magnitude observed across a shock, the middle panel shows the parallel electric field over the interval highlighted in the upper panel, and the bottom panel presents the electrostatic potential computed using the double layer speed estimated using interferometry (SM). Electrostatic potentials are shown over time intervals highlighted in the middle panels; the spatial axis corresponding to the temporal axis is at the top of the bottom panels. Table 1 summarizes various parameters of the double layers. . ## 3 Discussion and Conclusion Previous spacecraft measurements showed that electron heating in the Earth’s bow shock was determined by the quasi-static electric field in the de Hoffmann-Teller frame [1, 4, 5], which arises from different electron and ion dynamics [39]. The typical spatial scale of the quasi-static electric field is between electron and ion inertial length [6, 7, 2, 8]. While we believe that electron heating in the Earth’s bow shock is indeed dominated by the quasi- static electric field, in this letter we demonstrated that the electron temperature can also increase across electrostatic double layers. The electron temperature variation across a single double layer can be 2–7% of the cross- shock potential in the de Hoffmann-Teller frame and occurs over spatial scale of only about ten Debye lengths or one tenth of electron inertial length. Some double layers can have spatial width of 70 Debye lengths and potential drops up to 30% of the cross-shock potential. Importantly, double layers can have their high-potential side facing magnetosheath (Figures 2 and 4b) as well as solar wind (Figures 4a,c,d). In the former the electron temperature across double layers increases toward magnetosheath, while in the latter it increases toward solar wind. Note that the quasi-static electric field always results in electron temperature increase toward the magnetosheath. Double layers presented in this letter are electrostatic and they fundamentally differ from electric field spikes reported in the foot of a quasi-perpendicular Earth’s bow shock [9]; those spikes were nonlinear whistler mode structures [40, 41]. Similar electrostatic double layers were previously reported in the auroral region [16, 17], plasma sheet [20, 21], inner magnetosphere [18, 19], reconnection current sheet [22, 23], and Venusian bow shock [24]. In rare cases the double layer propagation velocity was revealed by interferometry [16, 17, 21], while typically it was assumed to coincide with the local ion-acoustic speed. Our interferometry analysis showed that double layers in the Earth’s bow shock have plasma frame speed in the range of 0.5–2$c_{\rm IA}$, where we used a rough estimation of the local ion- acoustic speed, $c_{\rm IA}=(T_{e\parallel}/m_{p})^{1/2}$, because its exact value depends on non-Maxwellian features of electron and ion distribution functions (e.g., Vasko et al. [42]). These double layers are of ion-acoustic type and, thus, identical with double layers in the auroral region [16, 17] and plasma sheet [21]. Similarly to previous observations, the double layers are associated with high-frequency electric field fluctuations (Figures 1 and 4), though that is not a universal feature (Figure 4a). The origin of these high-frequency fluctuations is not apparent, because the double layers are rather weak to produce electron beams at the high-potential side (Figure 3). The causal relation between the double layers and the temperature variation across them is not obvious. Double layers formed by a current-driven ion- acoustic instability can indeed result in electron heating [43, 44, 45], but double layers formed at a contact boundary between cold and hot electron populations result in no actual electron heating [46, 47]. In the case of a current-driven ion-acoustic instability in a uniform plasma [43], a double layer is associated with a negative potential spike (ion hole). Although ion holes are abundant in the Earth’s bow shock [48, 49, 27], the observed double layers are not associated with them (Figures 2 and 4) and, hence, fundamentally different from ones observed in simulations by Sato and Okuda [43]. The double layers are similar to those produced by a current-driven instability developing around a localized plasma density perturbation [44], but they are also not different from double layers formed at a contact boundary between cold and hot electron populations, the latter produced by some other heating mechanism [46, 47]. The origin of the double layers need to be established in the future to reveal their ability in actual electron heating. In conclusion, we showed that the electron temperature in the Earth’s bow shock can increase across Debye-scale electrostatic structures. The estimation of net contribution of these structures to the electron heating remains a challenge and the mechanism producing these double layers remains to be revealed. ## 4 Open Research All the additional information on double layers considered in this letter can be found in the Supporting Materials. We thank the MMS team for excellent data available at https://lasp.colorado.edu/mms/sdc/public/about/browse-wrapper/. The work of J.S. and R.W. was supported by National Science Foundation grants No. 2026680. The work of I.V. and F.M. was supported by NASA Heliophysics Guest Investigator grant No. 80NSSC21K0730. I.V. also thanks the International Space Science Institute, Bern, Switzerland for supporting the working group "Resolving the Microphysics of Collisionless Shock Waves". I.V. thanks Anton Artemyev and Sergey Kamaletdinov for valuable contributions. ## References * [1] J. D. Scudder. A review of the physics of electron heating at collisionless shocks. Advances in Space Research, 15:181–223, 1995. * [2] V. Krasnoselskikh, M. Balikhin, S. N. Walker, S. Schwartz, D. Sundkvist, V. Lobzin, M. Gedalin, S. D. Bale, F. Mozer, J. Soucek, Y. Hobara, and H. Comisel. The Dynamic Quasiperpendicular Shock: Cluster Discoveries. Space Sci. Rev., 178:535–598, October 2013. * [3] Michael Gedalin. Large-scale versus Small-scale Fields in the Shock Front: Effect on the Particle Motion. Astrophys. J. , 895(1):59, May 2020. * [4] A. J. Hull, J. D. Scudder, D. E. Larson, and R. Lin. Electron heating and phase space signatures at supercritical, fast mode shocks. J. Geophys. Res., 106:15711–15734, August 2001. * [5] B. Lefebvre, S. J. Schwartz, A. F. Fazakerley, and P. Décréau. Electron dynamics and cross-shock potential at the quasi-perpendicular Earth’s bow shock. Journal of Geophysical Research (Space Physics), 112:A09212, September 2007. * [6] S. D. Bale, M. A. Balikhin, T. S. Horbury, V. V. Krasnoselskikh, H. Kucharek, E. Möbius, S. N. Walker, A. Balogh, D. Burgess, B. Lembège, E. A. Lucek, M. Scholer, S. J. Schwartz7 10, and M. F. Thomsen. Quasi-perpendicular Shock Structure and Processes. Space Sci. Rev., 118(1-4):161–203, June 2005. * [7] S. J. Schwartz, E. Henley, J. Mitchell, and V. Krasnoselskikh. Electron Temperature Gradient Scale at Collisionless Shocks. Physical Review Letters, 107(21):215002, November 2011. * [8] Lynn B. Wilson, L.J Chen, and V. Roytershteyn. The discrepancy between simulation and observation of electric fields in collisionless shock. Frontiers in Astronomy and Space Sciences, 2021. * [9] L. J. Chen, S. Wang, L. B. Wilson, S. Schwartz, N. Bessho, T. Moore, D. Gershman, B. Giles, D. Malaspina, F. D. Wilder, R. E. Ergun, M. Hesse, H. Lai, C. Russell, R. Strangeway, R. B. Torbert, A. F. -Vinas, J. Burch, S. Lee, C. Pollock, J. Dorelli, W. Paterson, N. Ahmadi, K. Goodrich, B. Lavraud, O. Le Contel, Yu. V. Khotyaintsev, P. A. Lindqvist, S. Boardsen, H. Wei, A. Le, and L. Avanov. Electron Bulk Acceleration and Thermalization at Earth’s Quasiperpendicular Bow Shock. Phys. Rev. Lett., 120(22):225101, June 2018. * [10] I. Y. Vasko, F. S. Mozer, S. D. Bale, and A. V. Artemyev. Ion-Acoustic Waves in a Quasi-Perpendicular Earth’s Bow Shock. Geophys. Res. Lett., 49(11):e98640, June 2022. * [11] Aaron Tran and Lorenzo Sironi. Electron Heating in Perpendicular Low-beta Shocks. Astrophys. J. Lett., 900(2):L36, September 2020. * [12] S. R. Kamaletdinov, I. Y. Vasko, A. V. Artemyev, R. Wang, and F. S. Mozer. Quantifying electron scattering by electrostatic solitary waves in the earth’s bow shock. Physics of Plasmas, 29(8):082301, 2022. * [13] L. Andersson and R. E. Ergun. The Search for Double Layers in Space Plasmas. Washington DC American Geophysical Union Geophysical Monograph Series, 197:241–249, January 2012. * [14] M. Temerin, K. Cerny, W. Lotko, and F. S. Mozer. Observations of double layers and solitary waves in the auroral plasma. Phys. Rev. Lett., 48(17):1175–1179, Apr 1982. * [15] F. S. Mozer and C. A. Kletzing. Direct observation of large, quasi-static, parallel electric fields in the auroral acceleration region. Geophys. Res. Lett., 25(10):1629–1632, January 1998. * [16] R. E. Ergun, Y. J. Su, L. Andersson, C. W. Carlson, J. P. McFadden, F. S. Mozer, D. L. Newman, M. V. Goldman, and R. J. Strangeway. Direct Observation of Localized Parallel Electric Fields in a Space Plasma. Phys. Rev. Lett., 87(4):045003, July 2001. * [17] L. Andersson, R. E. Ergun, D. L. Newman, J. P. McFadden, C. W. Carlson, and Y. J. Su. Characteristics of parallel electric fields in the downward current region of the aurora. Physics of Plasmas, 9(8):3600–3609, August 2002. * [18] Xiaohua Deng, Maha Ashour-Abdalla, Meng Zhou, Raymond Walker, Mostafa El-Alaoui, Vassilis Angelopoulos, R. E. Ergun, and David Schriver. Wave and particle characteristics of earthward electron injections associated with dipolarization fronts. Journal of Geophysical Research (Space Physics), 115(A9):A09225, September 2010. * [19] D. M. Malaspina, L. Andersson, R. E. Ergun, J. R. Wygant, J. W. Bonnell, C. Kletzing, G. D. Reeves, R. M. Skoug, and B. A. Larsen. Nonlinear electric field structures in the inner magnetosphere. Geophys. Res. Lett., 41(16):5693–5701, August 2014. * [20] R. E. Ergun, L. Andersson, J. Tao, V. Angelopoulos, J. Bonnell, J. P. McFadden, D. E. Larson, S. Eriksson, T. Johansson, C. M. Cully, D. N. Newman, M. V. Goldman, A. Roux, O. Lecontel, K. H. Glassmeier, and W. Baumjohann. Observations of Double Layers in Earth’s Plasma Sheet. Phys. Rev. Lett., 102(15):155002, April 2009. * [21] Zhigang Yuan, Yue Dong, Shiyong Huang, Zuxiang Xue, and Xiongdong Yu. Direct Observation of Acceleration and Thermalization of Beam Electrons Caused by Double Layers in the Earth’s Plasma Sheet. Geophys. Res. Lett., 49(13):e99483, July 2022. * [22] Rongsheng Wang, Quanming Lu, Yuri V. Khotyaintsev, Martin Volwerk, Aimin Du, Rumi Nakamura, Walter D. Gonzalez, Xuan Sun, Wolfgang Baumjohann, Xing Li, Tielong Zhang, Andrew N. Fazakerley, Can Huang, and Mingyu Wu. Observation of double layer in the separatrix region during magnetic reconnection. Geophys. Res. Lett., 41(14):4851–4858, July 2014. * [23] M. Øieroset, T. D. Phan, R. Ergun, N. Ahmadi, K. Genestreti, J. F. Drake, Y. H. Liu, C. Haggerty, J. P. Eastwood, M. A. Shay, P. S. Pyakurel, S. Haaland, M. Oka, M. Goodbred, S. Eriksson, J. L. Burch, R. B. Torbert, Y. Khotyaintsev, C. T. Russell, R. J. Strangeway, D. J. Gershman, and B. L. Giles. Spatial evolution of magnetic reconnection diffusion region structures with distance from the X-line. Physics of Plasmas, 28(12):122901, December 2021. * [24] D. M. Malaspina, K. Goodrich, R. Livi, J. Halekas, M. McManus, S. Curry, S. D. Bale, J. W. Bonnell, T. Dudok de Wit, K. Goetz, P. R. Harvey, R. J. MacDowall, M. Pulupa, A. W. Case, J. C. Kasper, K. E. Korreck, D. Larson, M. L. Stevens, and P. Whittlesey. Plasma Double Layers at the Boundary Between Venus and the Solar Wind. Geophys. Res. Lett., 47(20):e90115, October 2020. * [25] Y. Hobara, S. N. Walker, M. Balikhin, O. A. Pokhotelov, M. Gedalin, V. Krasnoselskikh, M. Hayakawa, M. André, M. Dunlop, H. RèMe, and A. Fazakerley. Cluster observations of electrostatic solitary waves near the Earth’s bow shock. Journal of Geophysical Research (Space Physics), 113:A05211, May 2008. * [26] Katherine A. Goodrich et al. MMS Observations of Electrostatic Waves in an Oblique Shock Crossing. Journal of Geophysical Research (Space Physics), 123(11):9430–9442, Nov 2018. * [27] R. Wang, I. Y. Vasko, F. S. Mozer, S. D. Bale, I. V. Kuzichev, A. V. Artemyev, K. Steinvall, R. Ergun, B. Giles, Y. Khotyaintsev, P. A. Lindqvist, C. T. Russell, and R. Strangeway. Electrostatic Solitary Waves in the Earth’s Bow Shock: Nature, Properties, Lifetimes, and Origin. Journal of Geophysical Research (Space Physics), 126(7):e29357, July 2021. * [28] C. T. Russell, B. J. Anderson, W. Baumjohann, K. R. Bromund, D. Dearborn, D. Fischer, G. Le, H. K. Leinweber, D. Leneman, W. Magnes, J. D. Means, M. B. Moldwin, R. Nakamura, D. Pierce, F. Plaschke, K. M. Rowe, J. A. Slavin, R. J. Strangeway, R. Torbert, C. Hagen, I. Jernej, A. Valavanoglou, and I. Richter. The Magnetospheric Multiscale Magnetometers. Space Sci. Rev., 199:189–256, March 2016. * [29] C. Pollock, T. Moore, A. Jacques, J. Burch, U. Gliese, Y. Saito, T. Omoto, L. Avanov, A. Barrie, V. Coffey, J. Dorelli, D. Gershman, B. Giles, T. Rosnack, C. Salo, S. Yokota, M. Adrian, C. Aoustin, C. Auletti, S. Aung, V. Bigio, N. Cao, M. Chandler, D. Chornay, K. Christian, G. Clark, G. Collinson, T. Corris, A. De Los Santos, R. Devlin, T. Diaz, T. Dickerson, C. Dickson, A. Diekmann, F. Diggs, C. Duncan, A. Figueroa-Vinas, C. Firman, M. Freeman, N. Galassi, K. Garcia, G. Goodhart, D. Guererro, J. Hageman, J. Hanley, E. Hemminger, M. Holland, M. Hutchins, T. James, W. Jones, S. Kreisler, J. Kujawski, V. Lavu, J. Lobell, E. LeCompte, A. Lukemire, E. MacDonald, A. Mariano, T. Mukai, K. Narayanan, Q. Nguyan, M. Onizuka, W. Paterson, S. Persyn, B. Piepgrass, F. Cheney, A. Rager, T. Raghuram, A. Ramil, L. Reichenthal, H. Rodriguez, J. Rouzaud, A. Rucker, Y. Saito, M. Samara, J.-A. Sauvaud, D. Schuster, M. Shappirio, K. Shelton, D. Sher, D. Smith, K. Smith, S. Smith, D. Steinfeld, R. Szymkiewicz, K. Tanimoto, J. Taylor, C. Tucker, K. Tull, A. Uhl, J. Vloet, P. Walpole, S. Weidner, D. White, G. Winkert, P.-S. Yeh, and M. Zeuch. Fast Plasma Investigation for Magnetospheric Multiscale. Space Sci. Rev., 199:331–406, March 2016. * [30] R. E. Ergun, S. Tucker, J. Westfall, K. A. Goodrich, D. M. Malaspina, D. Summers, J. Wallace, M. Karlsson, J. Mack, N. Brennan, B. Pyke, P. Withnell, R. Torbert, J. Macri, D. Rau, I. Dors, J. Needell, P.-A. Lindqvist, G. Olsson, and C. M. Cully. The Axial Double Probe and Fields Signal Processing for the MMS Mission. Space Sci. Rev., 199:167–188, March 2016. * [31] P.-A. Lindqvist, G. Olsson, R. B. Torbert, B. King, M. Granoff, D. Rau, G. Needell, S. Turco, I. Dors, P. Beckman, J. Macri, C. Frost, J. Salwen, A. Eriksson, L. Åhlén, Y. V. Khotyaintsev, J. Porter, K. Lappalainen, R. E. Ergun, W. Wermeer, and S. Tucker. The Spin-Plane Double Probe Electric Field Instrument for MMS. Space Sci. Rev., 199:137–165, March 2016. * [32] O. Le Contel et al. The Search-Coil Magnetometer for MMS. Space Sci. Rev., 199(1-4):257–282, Mar 2016. * [33] B. U. Ö. Sonnerup and M. Scheible. Minimum and Maximum Variance Analysis. ISSI Scientific Reports Series, 1:185–220, 1998. * [34] Ivan Y. Vasko, Rachel Wang, Forrest S. Mozer, Stuart D. Bale, and Anton V. Artemyev. On the nature and origin of bipolar electrostatic structures in the Earth’s bow shock. Frontiers in Physics, 8:156, June 2020. * [35] M. C. Kelley and F. S. Mozer. A technique for making dispersion relation measurements of electrostatic waves. J. Geophys. Res., 77(34):6900–6903, December 1972. * [36] Steven J. Schwartz, Michelle F. Thomsen, S. J. Bame, and John Stansberry. Electron heating and the potential jump across fast mode shocks. J. Geophys. Res., 93:12923–12931, November 1988. * [37] Steven J. Schwartz, Robert Ergun, Harald Kucharek, Lynn Wilson, Li-Jen Chen, Katherine Goodrich, Drew Turner, Imogen Gingell, Hadi Madanian, Daniel Gershman, and Robert Strangeway. Evaluating the deHoffmann-Teller Cross-Shock Potential at Real Collisionless Shocks. Journal of Geophysical Research (Space Physics), 126(8):e29295, August 2021. * [38] Daniel J. Gershman, Levon A. Avanov, Scott A. Boardsen, John C. Dorelli, Ulrik Gliese, Alexander C. Barrie, Conrad Schiff, William R. Paterson, Roy B. Torbert, Barbara L. Giles, and Craig J. Pollock. Spacecraft and Instrument Photoelectrons Measured by the Dual Electron Spectrometers on MMS. Journal of Geophysical Research (Space Physics), 122(11):11,548–11,558, November 2017. * [39] C. C. Goodrich and J. D. Scudder. The adiabatic energy change of plasma electrons and the frame dependence of the cross-shock potential at collisionless magnetosonic shock waves. J. Geophys. Res., 89(A8):6654–6662, August 1984. * [40] M. A. Balikhin, T. Dudok de Wit, H. St. C. K. Alleyne, L. J. C. Woolliscroft, S. N. Walker, V. Krasnosel’skikh, W. A. C. Mier-Jedrzejeowicz, and W. Baumjohann. Experimental determination of the dispersion of waves observed upstream of a quasi-perpendicular shock. Geophys. Res. Lett., 24(7):787–790, April 1997. * [41] S. N. Walker, M. A. Balikhin, and M. N. Nozdrachev. Ramp nonstationarity and the generation of whistler waves upstream of a strong quasiperpendicular shock. Geophys. Res. Lett., 26(10):1357–1360, May 1999. * [42] I. Y. Vasko, O. V. Agapitov, F. S. Mozer, J. W. Bonnell, A. V. Artemyev, V. V. Krasnoselskikh, G. Reeves, and G. Hospodarsky. Electron-acoustic solitons and double layers in the inner magnetosphere. Geophys. Res. Lett., 44:4575–4583, May 2017. * [43] T. Sato and H. Okuda. Numerical simulations on ion acoustic double layers. J. Geophys. Res., 86(A5):3357–3368, May 1981. * [44] D. L. Newman, M. V. Goldman, R. E. Ergun, and A. Mangeney. Formation of Double Layers and Electron Holes in a Current-Driven Space Plasma. Phys. Rev. Lett., 87(25):255001, December 2001. * [45] Alexander R. Vazsonyi, Kentaro Hara, and Iain D. Boyd. Non-monotonic double layers and electron two-stream instabilities resulting from intermittent ion acoustic wave growth. Physics of Plasmas, 27(11):112303, November 2020. * [46] T. C. Li, J. F. Drake, and M. Swisdak. Suppression of Energetic Electron Transport in Flares by Double Layers. Astrophys. J. , 757(1):20, September 2012. * [47] T. C. Li, J. F. Drake, and M. Swisdak. Dynamics of Double Layers, Ion Acceleration, and Heat Flux Suppression during Solar Flares. Astrophys. J. , 793(1):7, September 2014. * [48] I. Y. Vasko, F. S. Mozer, V. V. Krasnoselskikh, A. V. Artemyev, O. V. Agapitov, S. D. Bale, L. Avanov, R. Ergun, B. Giles, P. A. Lindqvist, C. T. Russell, R. Strangeway, and R. Torbert. Solitary Waves Across Supercritical Quasi-Perpendicular Shocks. Geophys. Res. Lett., 45(12):5809–5817, June 2018. * [49] R. Wang, I. Y. Vasko, F. S. Mozer, S. D. Bale, A. V. Artemyev, J. W. Bonnell, R. Ergun, B. Giles, P. A. Lindqvist, C. T. Russell, and R. Strangeway. Electrostatic Turbulence and Debye-scale Structures in Collisionless Shocks. Astrophys. J. Lett., 889(1):L9, January 2020.
# High order asymptotic preserving well-balanced finite difference WENO schemes for all Mach full Euler equations with gravity ###### Abstract. In this paper, we propose a high order semi-implicit well-balanced finite difference scheme for all Mach Euler equations with a gravitational source term. To obtain the asymptotic preserving property, we start from the conservative form of full compressible Euler equations and add the evolution equation of the perturbation of potential temperature. The resulting system is then split into a (non-stiff) nonlinear low dynamic material wave to be treated explicitly, and (stiff) fast acoustic and gravity waves to be treated implicitly. With the aid of explicit time evolution for the perturbation of potential temperature, we design a novel well-balanced finite difference WENO scheme for the conservative variables, which can be proven to be both asymptotic preserving and asymptotically accurate in the incompressible limit. Extensive numerical experiments were provided to validate these properties. ###### Key words and phrases: compressible Euler equations; all Mach numbers; gravity; finite difference WENO; high order method; asymptotic preserving; well-balanced. Guanlan Huang School of Mathematical Sciences, Xiamen University, Xiamen, Fujian, 361005, PR China <EMAIL_ADDRESS> Yulong Xing Department of Mathematics, The Ohio State University, Columbus, OH 43210, USA <EMAIL_ADDRESS> Tao Xiong***Corresponding author. School of Mathematical Sciences, Fujian Provincial Key Laboratory of Mathematical Modeling and High-Performance Scientific Computing, Xiamen University Xiamen, Fujian 361005, PR China <EMAIL_ADDRESS> ## 1\. Introduction In this paper, we are interested in the compressible Euler equations with a gravitational source (1.1) $\left\\{\begin{array}[]{ll}\rho_{t}+\nabla\cdot(\rho{\bf u})=0,\\\\[2.84526pt] (\rho{\bf u})_{t}+\nabla\cdot(\rho{\bf u}\otimes{\bf u})+\nabla p=-\rho\nabla\Phi,\\\\[2.84526pt] E_{t}+\nabla\cdot\left((E+p){\bf u}\right)=-\rho{\bf u}\cdot\nabla\Phi,\end{array}\right.$ where $\rho$ is the density, ${\bf u}$ is the velocity, $p$ is the pressure, and $\rho{\bf u}$ is the momentum. $\Phi=\Phi({\bf x})$ is the gravitational potential which is assumed to be time-independent. $E=\frac{1}{2}\rho|{\bf u}|^{2}+\rho e$ is the total non-gravitational energy, with $e$ being the specific internal energy. The system (1.1) needs to be closed by providing an equation of state (EOS), which is usually given in the form $e=\mathcal{E}(\rho,p)$. For an ideal gas, the EOS can be written as $e=p/(\gamma-1)/\rho$, and the energy $E$ becomes (1.2) $E=\frac{1}{2}\rho|{\bf u}|^{2}+\frac{p}{\gamma-1},$ with $\gamma>1$ being the ratio of specific heat. On one hand, for hyperbolic systems with a source term, one important feature is that they admit equilibrium state solutions, and well-balanced numerical methods are desirable to exactly preserve such equilibrium states, so that small perturbations around an equilibrium state can be well captured on relatively coarse mesh sizes. In recent years, well-balanced schemes are very attractive for shallow water equations with source terms, see [8, 29, 1, 43, 45, 44, 33], the review papers [42, 28] and the references therein. The compressible Euler equations with a gravitational source (1.1) have a zero- velocity hydrostatic equilibrium state [46] of the form (1.3) ${\bf u}={\bf 0},\qquad\nabla p=-\rho\nabla\Phi.$ Numerically, similar to the shallow water equations, it is essential to develop well-balanced schemes for the Euler equations (1.1) to exactly preserve such an equilibrium state, especially for long-time simulations. Many well-balanced schemes have been studied, including finite difference schemes [46, 32], finite volume schemes [24, 15, 5, 17, 30, 40, 7, 6, 20, 26], and discontinuous Galerkin finite element methods [41, 16, 31]. These schemes are usually based on explicit time discretizations for compressible flows. On the other hand, the Euler equations (1.1) have many applications with a wide range of Mach numbers. We can rewrite (1.1) into a dimensionless form, by introducing a set of dimensionless variables with some suitable reference values (1.4) $\hat{{{\bf x}}}=\frac{{{\bf x}}}{\ell_{ref}},\,\hat{t}=\frac{t}{t_{ref}},\,\hat{\rho}=\frac{\rho}{\rho_{ref}},\,\hat{{\bf u}}=\frac{{\bf u}}{U_{ref}},\,\hat{p}=\frac{p}{p_{ref}},\,\hat{E}=\frac{E}{p_{ref}},\,\hat{\Phi}=\frac{\Phi}{\Phi_{ref}},$ where $U_{ref}=\ell_{ref}/t_{ref}$. Under these dimensionless variables, the system (1.1) becomes [38, 9, 4] (1.5a) $\displaystyle\rho_{t}+\nabla\cdot(\rho{\bf u})=0,$ (1.5b) $\displaystyle(\rho{\bf u})_{t}+\nabla\cdot(\rho{\bf u}\otimes{\bf u})+\frac{1}{\varepsilon^{2}}\nabla p=-\frac{1}{\varepsilon^{2}}\rho\nabla\Phi,$ (1.5c) $\displaystyle E_{t}+\nabla\cdot\left((E+p){\bf u}\right)=-\rho{\bf u}\cdot\nabla\Phi.$ Here we drop the hats for those dimensionless variables for clarity. The parameter $\varepsilon={U_{ref}}/{c_{ref}}$ is a referenced global Mach number, with $c_{ref}=\sqrt{{p_{ref}}/{\rho_{ref}}}$ being the acoustic velocity depending on the background flow. The EOS (1.2) becomes (1.6) $E=\frac{1}{2}\varepsilon^{2}\rho|{\bf u}|^{2}+\frac{p}{\gamma-1}.$ The dimensionless system (1.5) with the EOS (1.6) is still hyperbolic, and its eigenvalues along the normal direction ${\bf{n}}$ are (1.7) $\lambda_{1}={\bf u}\cdot{\bf{n}}-c/\varepsilon,\quad\lambda_{2}={\bf u}\cdot{\bf{n}},\quad\lambda_{3}={\bf u}\cdot{\bf{n}}+c/\varepsilon,$ where $c=\sqrt{\gamma p/\rho}$ is the scaled speed of sound. The model (1.5)-(1.6) with the global Mach number $\varepsilon$ ranging from $0$ to $\mathcal{O}(1)$ is widely used for all-speed flows [10, 38]. For the system (1.5), due to the stiffness of eigenvalues given in (1.7), shock capturing schemes with explicit time discretizations are subject to a very strict time step restriction [3] (1.8) $\Delta t=\text{CFL}\frac{\Delta x}{\max(|{\bf u}|+c/\varepsilon)}\sim\varepsilon\Delta x,$ when $\varepsilon$ is small. Here $\Delta t$ is the time step, $\Delta x$ is the mesh size and CFL denotes the CFL number. Besides, for schemes with numerical viscosities depending on the eigenvalues (1.7), e.g. schemes with a Lax-Friedrichs numerical flux, their numerical viscosities would be inversely proportional to $\varepsilon$. As $\varepsilon$ goes to $0$, either the time step is too small, or the numerical viscosities are too large, leading to very inefficient numerical schemes [46, 32]. To avoid that, several semi-implicit schemes have been developed for all-Mach Euler systems [3, 27, 38, 10, 4], which devote to efficient schemes of easy implementation and uniform time stabilities with respect to $\varepsilon$, and avoid nonlinear iterative solvers as used in the fully implicit schemes. In this work, we aim to provide an efficient high order solver for all-Mach full Euler equations with a gravitational source (1.5), which enjoy the benefit of being high order, well-balanced and asymptotic preserving (AP) simultaneously. The design of high order AP method for Euler equations with gravity is a nontrivial task. The target model (1.5) is different from all- Mach homogeneous Euler equations considered in [13], where the stiff acoustic wave only connects to the energy equation, so that only the gradient of pressure in the momentum equation and the flux in the energy equation deserve implicit treatments. All Froude number shallow water equations with a non-flat bottom topography were recently studied in [22], where the source term (the gradient of the bottom function) can be merged with the pressure gradient, due to the polytropic EOS $p=p(h)$ with $h$ being the water height. For our model (1.5), the stiff acoustic wave is balanced by the gravity wave, therefore both of them require implicit treatments, which in turn leads to the requirement of implicit treatments for both the density equation (1.5a) and the energy equation (1.5c). As a result, one new challenge is that, due to the implicit discretizations for these conservative variables, it is very hard to numerically derive a consistent discretization for the perturbation of the potential temperature $\theta_{2}$, which appears in the incompressible limit, see (2.8). This is mainly due to the highly nonlinear relationships between conservative and primitive variables, and the nonlinear EOS (1.6). Therefore, the existing approaches in [13, 22] cannot be directly applied. To address such an issue, one novel contribution of this paper is to add the evolution of the perturbation $\theta_{2}$ of the potential temperature to (1.5) in the numerical discretization. The original energy equation corresponds to the updating of hydrostatic pressure in the compressible regime for conservational purpose, and the added equation of $\theta_{2}$ extracts the contribution of the energy equation in the scale of $\mathcal{O}(\varepsilon^{2})$, and is corresponding to the update of the hydrodynamic pressure $p_{2}$ in the incompressible regime. The added equation of $\theta_{2}$ has simplified the treatment of the nonlinear EOS (1.6), which helps us to ensure that the correct asymptotic limit is achieved. Then, we split the system (1.5) into a (non-stiff) nonlinear low dynamic material wave to be treated explicitly, and (stiff) fast acoustic and gravity waves to be treated implicitly. A high order implicit-explicit (IMEX) method is utilized as the temporal discretization. With the aid of explicit time evolution for the perturbation of potential temperature $\theta_{2}$, we design a novel well-balanced finite difference WENO scheme for the spatial discretization. The temporal update of the conservative variables $(\rho,\rho{\bf u},E)$ is through the conservative scheme, and $\theta_{2}$ is only an auxiliary variable to aid with the design of AP method. The proposed method can be formally proved to be AP and asymptotically accurate (AA). In addition, in the AP and AA analyses, we can show that, as $\varepsilon\rightarrow 0$, (1.5c) is a consistent updating of hydrostatic pressure, while $\theta_{2}$ contributes to the hydrodynamic pressure. Extensive one- and two-dimensional numerical experiments demonstrate the high order accuracy, well-balanced, AP and AA properties of our proposed approach, as well as good performances in both high and low Mach regimes. The rest of the paper is organized as follows. In Section 2, a low Mach limit of all-Mach Euler equations with gravity is reviewed. AP well-balanced numerical methods are described in Section 3, with AP and AA analyses given in Section 4. Numerical experiments are performed in Section 5, followed by concluding remarks in Section 6. The detailed step-by-step flowchart of the fully discrete high order scheme with a multi-stage IMEX time discretization and a corresponding detailed proof of its AA property are presented as supplementary materials. ## 2\. Low Mach limit for full Euler equations with gravity In this section, let us briefly review the low Mach limit of Euler equations with gravity (1.5). To derive such a limit, it would be more convenient to start with an equation for the pressure $p$, instead of (1.5c) for the total energy. By utilizing the EOS (1.6), (1.5c) can be replaced by (2.1) $p_{t}+{\bf u}\cdot\nabla p+\gamma p\nabla\cdot{\bf u}=0.$ If we further define a potential temperature $\theta$ from $p=(\rho\theta)^{\gamma}$, it yields (2.2) $(\rho\theta)_{t}+\nabla\cdot(\rho\theta{\bf u})=0,$ which reduces to (after combined with (1.5a)) (2.3) $\theta_{t}+{\bf u}\cdot\nabla\theta=0,$ namely, the potential temperature $\theta$ satisfies a simple transport equation with the velocity ${\bf u}$. Now the model (1.5) can be rewritten as (2.4a) $\displaystyle\rho_{t}+\nabla\cdot(\rho{\bf u})=0,$ (2.4b) $\displaystyle(\rho{\bf u})_{t}+\nabla\cdot(\rho{\bf u}\otimes{\bf u})+\frac{1}{\varepsilon^{2}}\nabla p=-\frac{1}{\varepsilon^{2}}\rho\nabla\Phi,$ (2.4c) $\displaystyle\theta_{t}+{\bf u}\cdot\nabla\theta=0.$ Let us perform a Chapman-Enskog expansion in the following form [10] (2.5) $\left\\{\begin{array}[]{ll}\rho({\bf x},t)=\rho_{0}({\bf x},t)+\varepsilon^{2}\rho_{2}({\bf x},t)+\cdots,\\\\[2.84526pt] {\bf u}({\bf x},t)={\bf u}_{0}({\bf x},t)+\varepsilon{\bf u}_{1}({\bf x},t)+\cdots,\\\\[2.84526pt] p({\bf x},t)=p_{0}({\bf x},t)+\varepsilon^{2}p_{2}({\bf x},t)+\cdots,\\\\[2.84526pt] \theta({\bf x},t)=\theta_{0}({\bf x},t)+\varepsilon^{2}\theta_{2}({\bf x},t)+\cdots.\end{array}\right.$ By substituting them into (2.4b), we can collect the leading order $\mathcal{O}(1/\varepsilon^{2})$ terms (2.6) $\nabla p_{0}=-\rho_{0}\nabla\Phi,$ where $p_{0}$ is the hydrostatic pressure. The $\mathcal{O}(1)$ terms of (2.4b) yield (2.7) $(\rho_{0}{\bf u}_{0})_{t}+\nabla\cdot(\rho_{0}{\bf u}_{0}\otimes{\bf u}_{0})+\nabla p_{2}=-\rho_{2}\nabla\Phi,$ where $p_{2}$ corresponds to the hydrodynamic pressure. To obtain a closed system for the low Mach limit, a further assumption is needed due to the extra source term $-\rho_{2}\nabla\Phi$ appearing in (2.7), which is different from the homogeneous all Mach full Euler system without gravity [13]. One commonly used assumption [25, 10] is a constant background potential temperature $\theta_{0}$ with hydrostatic background states $\rho_{0}({\bf x})$ and $p_{0}({\bf x})$ satisfying (2.6). With such assumption, from (2.4a), (2.7), (2.4c) and the expansion (2.5), we obtain the following limiting incompressible equations (2.8) $\left\\{\begin{array}[]{ll}\nabla\cdot\left(\rho_{0}{\bf u}_{0}\right)=0,\\\\[5.69054pt] \left(\rho_{0}{\bf u}_{0}\right)_{t}+\nabla\cdot(\rho_{0}{\bf u}_{0}\otimes{\bf u}_{0})+\nabla p_{2}=-\rho_{2}\nabla\Phi,\\\\[5.69054pt] (\theta_{2})_{t}+{\bf u}_{0}\cdot\nabla\theta_{2}=0,\end{array}\right.$ which is known as the Ogura and Phillips model [34]. The model (2.8) contains four unknowns ${\bf u}_{0}$, $\rho_{2}$, $p_{2}$, $\theta_{2}$, and is closed by the relation $p=(\rho\,\theta)^{\gamma}$, which yields (2.9) $p_{2}=\lim_{\varepsilon\rightarrow 0}\frac{p-p_{0}}{\varepsilon^{2}}=\lim_{\varepsilon\rightarrow 0}\frac{(\rho\,\theta)^{\gamma}-(\rho_{0}\,\theta_{0})^{\gamma}}{\varepsilon^{2}}=\frac{\gamma p_{0}}{\rho_{0}\theta_{0}}(\rho_{0}\,\theta_{2}+\theta_{0}\,\rho_{2}),$ where the Chapman-Enskog expansion (2.5) is used. As compared to the homogeneous all Mach full Euler system without gravity [13], an extra equation of $\theta_{2}$ exists in the incompressible system (2.8). Due to the existence of the source term in the momentum equation of (2.8), the limit model (2.8) requires (2.9) to close the system, which will lead to additional numerical challenges as described in the next section. ###### Remark 2.1. Another model assumes a stratified background potential temperature $\theta_{0}({\bf x})$ with $\nabla\theta_{0}({\bf x})=\mathcal{O}(\varepsilon^{2})$, which is typically used in a nearly incompressible regime. Similarly if we have hydrostatic background states $\rho_{0}({\bf x})$ and $p_{0}({\bf x})$ satisfying (2.6), the nearly incompressible system is given as (2.10) $\left\\{\begin{array}[]{ll}\nabla\cdot\left(\rho_{0}{\bf u}_{0}\right)=0,\\\\[5.69054pt] \left(\rho_{0}{\bf u}_{0}\right)_{t}+\nabla\cdot(\rho_{0}{\bf u}_{0}\otimes{\bf u}_{0})+\nabla p_{2}=-\rho_{2}\nabla\Phi,\\\\[5.69054pt] (\theta_{2})_{t}+{\bf u}_{0}\cdot\nabla\theta_{2}+\frac{1}{\varepsilon^{2}}{\bf u}_{0}\cdot\nabla\theta_{0}=0.\end{array}\right.$ This is known as the Bannon’s anelastic model [2]. As $\varepsilon\rightarrow 0$, $\theta_{0}({\bf x})$ converges to a constant and (2.10) becomes (2.8). ## 3\. Numerical scheme In this section, we propose a class of semi-implicit high order finite difference weighted essentially non-oscillatory (WENO) schemes, with both AP and well-balanced properties, for the all-Mach Euler system with gravity (1.5). A novel new framework is presented to design a well-balanced high order semi-implicit scheme, that is AP and AA in the incompressible limit, and is conservative and robust in the compressible regime. ### 3.1. Model problem and semi-implicit splitting For the model (1.5)-(1.6), the gravitational source appears both in the second momentum and third energy equations. It couples all the conservative unknowns together, which is different from the source term due to a non-flat bottom topography in the shallow water equations [22]. Compared to all Mach homogeneous Euler equations without gravity in [13], this introduces two extra variables $\rho_{2}$ and $\theta_{2}$ in the limiting equations (2.8). As a result, a fully nonlinear coupling of all conservative variables exists, and it is highly nontrivial to extract an elliptic equation for the extra variables $\rho_{2}$ or $p_{2}$ from (1.5) through the nonlinear EOS (1.6). Note that such elliptic equation is very important to ensure a correct low Mach limit (2.8). Therefore, the frameworks developed in [22, 13] cannot be directly applied to (1.5). In order to address the difficulty to ensure a correct asymptotic limit from (1.5), one novel idea of this paper is to include the evolution of perturbation of the potential temperature $\theta_{2}$ to the system (3.1) $\left\\{\begin{array}[]{ll}\rho_{t}+\nabla\cdot(\rho{\bf u})=0,\\\\[5.69054pt] (\rho{\bf u})_{t}+\nabla\cdot(\rho{\bf u}\otimes{\bf u})+\frac{1}{\varepsilon^{2}}\nabla p=-\frac{1}{\varepsilon^{2}}\rho\nabla\Phi,\\\\[5.69054pt] E_{t}+\nabla\cdot\left(\left(E+p\right){\bf u}\right)=-\rho{\bf u}\cdot\nabla\Phi,\\\\[5.69054pt] (\theta_{2})_{t}+{\bf u}\cdot\nabla\theta_{2}+\frac{1}{\varepsilon^{2}}{\bf u}\cdot\nabla\theta_{0}=0.\end{array}\right.$ Note that the equation of $\theta_{2}$ is taken from (2.4c) by assuming an expansion $\theta({\bf x},t)=\theta_{0}({\bf x})+\varepsilon^{2}\theta_{2}({\bf x},t)$, mimicking the one in (2.5). Here $\theta_{0}({\bf x})$ is assumed to be a pre-determined time independent background potential temperature and $\nabla\theta_{0}({\bf x})=\mathcal{O}(\varepsilon^{2})$. This equation can ensure the right incompressible limit (2.8), and more generally is consistent with (2.10) in the nearly incompressible regime. The advantage of (3.1), compared with (1.5), is that the third energy equation corresponds to the updating of hydrostatic pressure $p_{0}$ (in the compressible regime) for conservational purpose, while the fourth equation of $\theta_{2}$ extracts the contribution of the energy equation in the scale of $\mathcal{O}(\varepsilon^{2})$, and is corresponding to the hydrodynamic pressure $p_{2}$ which plays the role as a Lagrangian multiplier in the incompressible regime. The fourth equation of $\theta_{2}$, which is derived from (2.4c), has simplified the treatment of the nonlinear EOS (1.6) as compared to the third energy equation, and this helps us to easily form an elliptic equation for $\rho_{2}$ as described in the design of our scheme below. We first follow the idea in [21, 37] to split the system (3.1) into a stiff part and a non-stiff component (3.2) $\frac{dU}{dt}=-\nabla\cdot\mathcal{F}_{E}-\nabla\cdot\mathcal{F}_{I}-\mathcal{S}_{E}-\mathcal{S}_{I},$ where $U=(\rho,\rho{\bf u},E,\theta_{2})^{T}$ and (3.3a) $\displaystyle\nabla\cdot\mathcal{F}_{E}=\alpha\Big{(}\nabla\cdot(\rho{\bf u})_{E},\nabla\cdot(\rho{\bf u}\otimes{\bf u}+p\,{\mathbb{I}})_{E},\nabla\cdot((E+p){\bf u})_{E},0\Big{)}^{T}$ $\displaystyle+\Big{(}0,(1-\alpha)\,\nabla\cdot(\rho{\bf u}\otimes{\bf u})_{E}+\alpha\frac{1-\varepsilon^{2}}{\varepsilon^{2}}\nabla p_{E},0,{\bf u}_{E}\cdot\nabla\theta_{2,E}+\frac{1}{\varepsilon^{2}}{\bf u}_{E}\cdot\nabla\theta_{0}\Big{)}^{T},$ (3.3b) $\displaystyle\nabla\cdot\mathcal{F}_{I}=(1-\alpha)\left(\nabla\cdot(\rho{\bf u})_{I},\frac{1}{\varepsilon^{2}}\nabla p_{I},\nabla\cdot((E+p){\bf u})_{I},0\right)^{T},$ (3.3c) $\displaystyle\mathcal{S}_{E}=\alpha\left(0,\frac{1}{\varepsilon^{2}}\rho_{E}\nabla\Phi,(\rho{\bf u})_{E}\cdot\nabla\Phi,0\right)^{T},$ (3.3d) $\displaystyle\mathcal{S}_{I}=(1-\alpha)\left(0,\frac{1}{\varepsilon^{2}}\rho_{I}\nabla\Phi,(\rho{\bf u})_{I}\cdot\nabla\Phi,0\right)^{T}.$ The subscripts $E$ and $I$ represent non-stiff and stiff parts of the flux function $\mathcal{F}$ and the source term $S$, which will be discretized explicitly and implicitly, respectively. In the splitting, we take the splitting parameter $\alpha=\min(\varepsilon^{2},1)$, that is, $\alpha=\varepsilon^{2}$ when $\varepsilon\leq 1$ and (3.2) turns to be fully explicit if $\varepsilon\geq 1$. In the following, we only consider $\alpha=\varepsilon^{2}$ with $\varepsilon\leq 1$ for ease of presentation. ### 3.2. First order IMEX scheme We start with presenting the first order IMEX AP time discretization for (3.1). We focus on time discretization only and keep space to be continuous. The spatial discretization will be discussed afterward. Following the splitting in (3.2), the semi-discrete (in time) scheme takes the form (3.4a) $\displaystyle\frac{\rho^{n+1}-\rho^{n}}{\Delta t}+\varepsilon^{2}\,\nabla\cdot(\rho{\bf u})^{n}+(1-\varepsilon^{2})\,\nabla\cdot(\rho{\bf u})^{n+1}=0,$ (3.4b) $\displaystyle\frac{(\rho{\bf u})^{n+1}-(\rho{\bf u})^{n}}{\Delta t}+\nabla\cdot(\rho{\bf u}\otimes{\bf u}+p\,{\mathbb{I}})^{n}+\frac{1-\varepsilon^{2}}{\varepsilon^{2}}\nabla p^{n+1}$ $\displaystyle\hskip 99.58464pt=-\rho^{n}\nabla\Phi-\frac{1-\varepsilon^{2}}{\varepsilon^{2}}\rho^{n+1}\nabla\Phi,$ (3.4c) $\displaystyle\frac{E^{n+1}-E^{n}}{\Delta t}+\varepsilon^{2}\,\nabla\cdot\left((E+p){\bf u}\right)^{n}+(1-\varepsilon^{2})\,\nabla\cdot\left((E+p){\bf u}\right)^{n+1}$ $\displaystyle\hskip 99.58464pt=-\varepsilon^{2}\,\left(\rho{\bf u}\right)^{n}\cdot\nabla\Phi-(1-\varepsilon^{2})\,\left(\rho{\bf u}\right)^{n+1}\cdot\nabla\Phi,$ (3.4d) $\displaystyle\frac{\theta_{2}^{n+1}-\theta_{2}^{n}}{\Delta t}+{\bf u}^{n}\cdot\nabla\theta_{2}^{n}+\frac{1}{\varepsilon^{2}}{\bf u}^{n}\cdot\nabla\theta_{0}=0.$ Introducing the variables (3.5a) $\displaystyle\widetilde{\rho}=\rho^{n}-\Delta t\,\varepsilon^{2}\,\nabla\cdot(\rho{\bf u})^{n},$ (3.5b) $\displaystyle\widetilde{\rho{\bf u}}=(\rho{\bf u})^{n}-\Delta t\Big{(}\nabla\cdot(\rho{\bf u}\otimes{\bf u}+p{\mathbb{I}})^{n}+\rho^{n}\nabla\Phi\Big{)},$ (3.5c) $\displaystyle\widetilde{E}=E^{n}-\Delta t\,\varepsilon^{2}\,\Big{(}\nabla\cdot\left((E+p){\bf u}\right)^{n}+\,\left(\rho{\bf u}\right)^{n}\cdot\nabla\Phi\Big{)},$ we can rewrite (3.4) as (3.6a) $\displaystyle\rho^{n+1}=\widetilde{\rho}-\Delta t\,(1-\varepsilon^{2})\,\nabla\cdot(\rho{\bf u})^{n+1},$ (3.6b) $\displaystyle(\rho{\bf u})^{n+1}=\widetilde{\rho{\bf u}}-\Delta t\,\frac{1-\varepsilon^{2}}{\varepsilon^{2}}\left(\nabla p^{n+1}+\rho^{n+1}\nabla\Phi\right),$ (3.6c) $\displaystyle E^{n+1}=\widetilde{E}-\Delta t\,(1-\varepsilon^{2})\,\left(\nabla\cdot\left((E+p){\bf u}\right)^{n+1}+\,\left(\rho{\bf u}\right)^{n+1}\cdot\nabla\Phi\right),$ (3.6d) $\displaystyle\theta_{2}^{n+1}=\theta_{2}^{n}-\Delta t\,\left({\bf u}^{n}\cdot\nabla\theta_{2}^{n}\right)-\Delta t\,\left(\frac{1}{\varepsilon^{2}}{\bf u}^{n}\cdot\nabla\theta_{0}\right),$ which will be solved with the following three steps: * • Step 1: Combining (3.6a) and (3.6b) leads to (3.7) $\rho^{n+1}=\widetilde{\rho}-(1-\varepsilon^{2})\,\Delta t\,\nabla\cdot(\widetilde{\rho{\bf u}})+\Delta t^{2}\,\frac{(1-\varepsilon^{2})^{2}}{\varepsilon^{2}}\nabla\cdot\left(\nabla p^{n+1}+\rho^{n+1}\nabla\Phi\right),$ We now introduce $p_{2}$ and $\rho_{2}$ from (3.8) $p^{n+1}=p_{0}({\bf x})+\varepsilon^{2}\,p_{2}^{n+1},\quad\rho^{n+1}=\rho_{0}({\bf x})+\varepsilon^{2}\,\rho_{2}^{n+1},$ according to (2.5), where $p_{0}({\bf x})$ and $\rho_{0}({\bf x})$ satisfy $\nabla p_{0}=-\rho_{0}\nabla\Phi$ as given in (2.6). Therefore (3.7) becomes (3.9) $\varepsilon^{2}\rho_{2}^{n+1}=\widetilde{\rho}-(1-\varepsilon^{2})\,\Delta t\,\nabla\cdot(\widetilde{\rho{\bf u}})+\Delta t^{2}(1-\varepsilon^{2})^{2}\nabla\cdot\left(\nabla p_{2}^{n+1}+\rho_{2}^{n+1}\nabla\Phi\right)-\rho_{0},$ which cannot be solved due to two unknowns $p^{n+1}_{2}$ and $\rho_{2}^{n+1}$. From the relation $p=(\rho\theta)^{\gamma}$, we have (3.10) $p_{2}^{n+1}=(p-p_{0})/\varepsilon^{2}=\left[\Big{(}(\rho_{0}+\varepsilon^{2}\,\rho_{2}^{n+1})(\theta_{0}+\varepsilon^{2}\,\theta_{2}^{n+1})\Big{)}^{\gamma}-p_{0}\right]/\varepsilon^{2}.$ Substituting it into (3.9) leads to a nonlinear system for $\rho_{2}^{n+1}$, which can be solved iteratively. To simplify the computation, we approximate (3.10) with a linearization up to an approximation error $\mathcal{O}(\varepsilon^{2})$, (3.11) $p_{2}^{n+1}=\frac{\gamma p_{0}}{\rho_{0}\theta_{0}}(\rho_{0}\theta_{2}^{n+1}+\rho^{n+1}_{2}\,\theta^{n+1}),$ where we denote $\theta^{n+1}=\theta_{0}+\varepsilon^{2}\,\theta_{2}^{n+1}$. Substituting (3.11) into (3.9) yields (3.12) $\varepsilon^{2}\rho_{2}^{n+1}=\widetilde{\widetilde{\rho}}+\Delta t^{2}(1-\varepsilon^{2})^{2}\nabla\cdot\left(\nabla\left(\frac{\gamma p_{0}}{\rho_{0}\theta_{0}}\theta^{n+1}\,\rho_{2}^{n+1}\right)+\rho_{2}^{n+1}\nabla\Phi\right),$ where (3.13) $\widetilde{\widetilde{\rho}}=\widetilde{\rho}-(1-\varepsilon^{2})\,\Delta t\,\nabla\cdot(\widetilde{\rho{\bf u}})+\Delta t^{2}(1-\varepsilon^{2})^{2}\Delta\left(\frac{\gamma p_{0}}{\theta_{0}}\theta^{n+1}_{2}\right)-\rho_{0}.$ The purpose of adding the fourth equation on $\theta_{2}$ in (3.1) is to evaluate $\theta^{n+1}_{2}$ from (3.6d) explicitly. With such information, $\theta^{n+1}$ can be computed and (3.12) is only a linear system for $\rho_{2}^{n+1}$. Otherwise, without (3.6d), (3.12) will be coupled to (3.6c), which is very complicated to solve for $\rho_{2}$ and $\theta_{2}$. After obtaining $\rho^{n+1}_{2}$, $p_{2}^{n+1}$ is available from (3.11). * • Step 2: $(\rho{\bf u})^{n+1}$ directly follows from (3.6b) and (3.8). In order to ensure exact mass and energy conservation, especially in the compressible regime, we further update $\rho^{n+1}$ from (3.6a), and finally $E^{n+1}$ from (3.6c). We note that in (3.6c), the implicit term $((E+p){\bf u})^{n+1}$ is computed based on (3.14) $((E+p){\bf u})^{n+1}=H^{n+1}\,(\rho{\bf u})^{n+1},$ where the enthalpy $H^{n+1}$ is given by (3.15) $H^{n+1}=\left(\frac{E+p}{\rho}\right)^{n+1}=\frac{\gamma}{\gamma-1}\frac{p^{n+1}}{\rho^{n+1}}+\frac{1}{2}\varepsilon^{2}\frac{|(\rho{\bf u})^{n+1}|^{2}}{(\rho^{n+1})^{2}},$ following the EOS (1.6). To compute $H^{n+1}$, we use $\rho^{n+1}$ and $p^{n+1}$ from (3.8) available after step 1. In this way, $(\rho{\bf u})^{n+1}$, $\rho^{n+1}$ and $E^{n+1}$ are updated in a conservative form and explicitly via a sequential way. ###### Remark 3.1. In the momentum equation (3.4b), the stiff gravitational source $\rho\nabla\Phi$ which depends on the unknown variable $\rho$ needs to be discretized implicitly. Correspondingly, we need an implicit discretization for the density equation (3.4a) in the low Mach regime. On the other hand, the implicit discretization of $\nabla p$ needs to be coupled with an implicit discretization of the energy equation (3.4c). In this way, the mass, momentum and energy equations in (3.4) all have components needing implicit discretizations, and uniform time stability independent of $\varepsilon$ can be obtained. This can be viewed as, in some sense, a combination of two approaches for the isentropic and full Euler equations [12, 13]. We refer to [12] for the Fourier analysis with respect to the rationale behind these implicit discretizations. ###### Remark 3.2. In [10], the authors used (2.2) instead of (1.5c), and designed an AP scheme in the low Mach regime. $\rho\theta$ in (2.2) is not a physically conservative variable. In this work, we keep the energy equation (1.5c), which is important for shock capturing purpose in the high Mach regime. Besides, a numerical discretization of (2.2) does not straightforwardly result in a consistent discretization of $\theta_{2}$ in the limiting equation (2.8) as $\varepsilon\rightarrow 0$. It requires $\nabla\cdot(\rho{\bf u})=0$ to be satisfied at the discrete level as $\varepsilon\rightarrow 0$, which is not an easy task. ### 3.3. High order IMEX Runge-Kutta (IMEX-RK) scheme We now generalize the first-order semi-implicit scheme (3.4) to high-order, by utilizing a high order IMEX-RK scheme proposed in [11], which has already been used in [12, 13]. We first denote the split system (3.2) as (3.16) $U_{t}=\mathcal{H}(U_{E},U_{I}),\qquad U(t_{0})=U_{0}.$ where $\mathcal{H}(U_{E},U_{I})=-\nabla\cdot\mathcal{F}_{E}-\nabla\cdot\mathcal{F}_{I}-\mathcal{S}_{E}-\mathcal{S}_{I}$, with $U_{E}=(\rho_{E},(\rho{\bf u})_{E},E_{E},\theta_{2,E})^{T}$ and $U_{I}=(\rho_{I},(\rho{\bf u})_{I},E_{I},\theta_{2,I})^{T}$ representing the terms that will be discretized explicitly and implicitly. The solutions $U_{E}$ and $U_{I}$ will be updated separately, that is (3.17) $\left\\{\begin{array}[]{l}(U_{E})_{t}=\mathcal{H}(U_{E},U_{I}),\\\\[5.69054pt] (U_{I})_{t}=\mathcal{H}(U_{E},U_{I}),\end{array}\right.$ with the same initial condition (3.18) $U_{E}(t_{0})=U_{I}(t_{0})=U_{0}.$ Following [12, 13], we apply an IMEX-RK scheme to (3.17) with a double Butcher tableau [14] (3.19) $\begin{array}[]{c|c}\tilde{c}&\tilde{A}\\\ \hline\cr\vspace{-0.25cm}\hfil\\\ &\tilde{b}^{T}\end{array},\ \ \ \ \ \begin{array}[]{c|c}{c}&{A}\\\ \hline\cr\vspace{-0.25cm}\hfil\\\ &{b^{T}}\end{array},$ where $\tilde{A}=(\tilde{a}_{ij})$ is an $s\times s$ matrix for an explicit scheme with $\tilde{a}_{ij}=0$ for $j\geq i$, and $A=({a}_{ij})$ is an $s\times s$ matrix for an implicit scheme. A diagonally implicit RK (DIRK) scheme with $a_{ij}=0$ for $j>i$ is used, which is simple and efficient for implicit discretization. Other vectors are $\tilde{c}=(\tilde{c}_{1},...,\tilde{c}_{s})^{T}$, $\tilde{b}=(\tilde{b}_{1},...,\tilde{b}_{s})^{T}$, and $c=(c_{1},...,c_{s})^{T}$, $b=(b_{1},...,b_{s})^{T}$, where $\tilde{c}$ and $c$ satisfy the relation $\tilde{c}_{i}=\sum_{j=1}^{i-1}\tilde{a}_{ij}$, $c_{i}=\sum_{j=1}^{i}a_{ij}$. For the high order IMEX-RK scheme, as discussed in [12, 13], a stiffly accurate property with $b^{T}={\bf e}_{s}^{T}A$ where ${\bf e}_{s}=(0,\cdots,0,1)^{T}$ for the implicit part is preferred, which can yield the AA property of the scheme. A sketched procedure of the IMEX-(DI)RK scheme is given as follows: 1. (1) Set $U_{E}^{(0)}=U_{I}^{(0)}=U^{n}$. For each inner stage $i=1\text{ to }s$, * • Update the explicit solution $U_{E}^{(i)}$ (3.20) $U_{E}^{(i)}=U^{n}+\Delta t\sum^{i-1}_{j=1}\tilde{a}_{ij}\mathcal{H}(U_{E}^{(j)},U_{I}^{(j)});$ * • Compute $U_{\star}^{(i)}$ based on known solutions at previous stages (3.21) $U_{\star}^{(i)}=U^{n}+\Delta t\sum^{i-1}_{j=1}a_{ij}\mathcal{H}(U_{E}^{(j)},U_{I}^{(j)}),$ then update the implicit solution $U_{I}^{(i)}$ via (3.22) $U_{I}^{(i)}=U_{\star}^{(i)}+\Delta ta_{ii}\mathcal{H}(U_{E}^{(i)},U_{I}^{(i)});$ 2. (2) Update the solution $U^{n+1}$ at the next time level $t^{n+1}$ by (3.23) $U^{n+1}=U^{n}+\Delta t\sum_{i=1}^{s}b_{i}\mathcal{H}(U_{E}^{(i)},U_{I}^{(i)}).$ ### 3.4. Fully discrete scheme with high order spatial discretizations In this part, we introduce high order WENO spatial discretizations for (3.2). We expect the following desirable properties from the spatial discretization. Firstly, it should be suitable for all Mach numbers, that is, for $\varepsilon$ ranging from $0$ to $\mathcal{O}(1)$. Secondly, it can preserve the well-balanced property, namely, the zero-velocity steady state solution (1.3) can be maintained exactly on the discrete level. To achieve both purposes, we adopt the spatial discretizations proposed in [13], and well- balanced reconstruction strategies from [46, 32]. Without loss of generality, we present (3.2) in the two-dimensional spatial setting. Taking $\alpha=\varepsilon^{2}$, we denote $U=(\rho,\rho u,\rho v,E,\theta_{2})^{T}$ and (3.24a) $\displaystyle\nabla\cdot\mathcal{F}_{E}=\partial_{x}\mathcal{F}_{E}^{x}+\partial_{y}\mathcal{F}_{E}^{y},\quad\mathcal{S}_{E}=\mathcal{S}_{E}^{x}+\mathcal{S}_{E}^{y},$ (3.24b) $\displaystyle\nabla\cdot\mathcal{F}_{I}=(1-\varepsilon^{2})(\partial_{x}\mathcal{F}_{I}^{x}+\partial_{y}\mathcal{F}_{I}^{y}),\quad\mathcal{S}_{I}=(1-\varepsilon^{2})(\mathcal{S}_{I}^{x}+\mathcal{S}_{I}^{y}),$ with (3.25a) $\displaystyle\partial_{x}\mathcal{F}_{E}^{x}=\varepsilon^{2}\Big{(}\partial_{x}(\rho u),\,\partial_{x}(\rho u^{2}+p),\,\partial_{x}(\rho uv),\,\partial_{x}((E+p)u),\,0\Big{)}_{E}^{T}$ $\displaystyle\hskip 36.98866pt+(1-\varepsilon^{2})\Big{(}0,\,\partial_{x}(\rho u^{2}+p),\,\partial_{x}(\rho uv),\,0,\,0\Big{)}_{E}^{T}$ $\displaystyle\hskip 36.98866pt+\Big{(}0,\,0,\,0,\,0,\,u\partial_{x}\theta_{2}+\frac{1}{\varepsilon^{2}}u\partial_{x}\theta_{0}\Big{)}_{E}^{T}$ (3.25b) $\displaystyle\partial_{y}\mathcal{F}_{E}^{y}=\varepsilon^{2}\Big{(}\partial_{y}(\rho v),\,\partial_{y}(\rho uv),\,\partial_{y}(\rho v^{2}+p),\,\partial_{y}((E+p)v),\,0\Big{)}_{E}^{T}$ $\displaystyle\hskip 36.98866pt+(1-\varepsilon^{2})\Big{(}0,\,\partial_{y}(\rho uv),\,\partial_{y}(\rho v^{2}+p),\,0,\,0\Big{)}_{E}^{T}$ $\displaystyle\hskip 36.98866pt+\Big{(}0,\,0,\,0,\,0,\,v\partial_{y}\theta_{2}+\frac{1}{\varepsilon^{2}}v\partial_{y}\theta_{0}\Big{)}_{E}^{T},$ (3.25c) $\displaystyle\partial_{x}\mathcal{F}_{I}^{x}=\Big{(}\partial_{x}(\rho u),\,\partial_{x}p_{2},\,0,\,\partial_{x}((E+p)u),\,0\Big{)}_{I}^{T},$ (3.25d) $\displaystyle\partial_{y}\mathcal{F}_{I}^{y}=\Big{(}\partial_{y}(\rho v),\,0,\,\partial_{y}p_{2},\,\partial_{y}((E+p)v),\,0\Big{)}_{I}^{T},$ and (3.26a) $\displaystyle\mathcal{S}_{E}^{x}=\Big{(}0,\rho\,\partial_{x}\Phi,0,\varepsilon^{2}\rho u\partial_{x}\Phi,0\Big{)}_{E}^{T},\quad\mathcal{S}_{E}^{y}=\Big{(}0,0,\rho\,\partial_{y}\Phi,\varepsilon^{2}\rho v\partial_{y}\Phi,0\Big{)}_{E}^{T},$ (3.26b) $\displaystyle\mathcal{S}_{I}^{x}=\Big{(}0,\rho_{2}\,\partial_{x}\Phi,0,\rho u\partial_{x}\Phi,0\Big{)}_{I}^{T},\quad\mathcal{S}_{I}^{y}=\Big{(}0,0,\rho_{2}\,\partial_{y}\Phi,\rho v\partial_{y}\Phi,0\Big{)}_{I}^{T}.$ Here (3.8) with $\nabla p_{0}=-\rho_{0}\nabla\Phi$ is used in (3.25c), (3.25d) and (3.26b), by replacing $p$ and $\rho$ with $p_{2}$ and $\rho_{2}$ respectively. For a high order finite difference discretization, we take a uniform cartesian mesh with mesh sizes $\Delta x$ and $\Delta y$ along $x$ and $y$ directions respectively. The grid points are $(x_{i},y_{j})$ for $i=1,2,\dots,N_{x}$, $j=1,2,\dots,N_{y}$. We denote $w_{ij}=w(x_{i},y_{j})$ for short with any variable $w$, and the interface values are denoted as $w_{i\pm\frac{1}{2},j}$ and $w_{i,j\pm\frac{1}{2}}$ respectively. We now briefly describe our spatial discretizations. Denoting ${\bf q}=\rho{\bf u}$, the high order WENO approximation of the spatial operator $\mathcal{H}(U_{E},U_{I})$ defined in (3.16) is given by (3.27) $\mathcal{H}_{WENO}(U_{E},U_{I})=-\left(\begin{aligned} &\varepsilon^{2}\nabla_{CW}\cdot{\bf q}_{E}+(1-\varepsilon^{2})\nabla_{W}\cdot{\bf q}_{I}\\\ &\nabla^{WB}_{CW}\cdot\left(\frac{{\bf q}\otimes{\bf q}}{\rho}+p\,{\mathbb{I}}\right)_{E}-\frac{\rho_{E}}{\rho_{0}}\nabla^{WB}_{CW}p_{0}\\\ &\hskip 34.14322pt+(1-\varepsilon^{2})\left(\nabla_{W}p_{2,I}+\rho_{2,I}\nabla\Phi\right)\\\ &\varepsilon^{2}\left(\nabla_{CW}\cdot(H{\bf q})_{E}+{\bf q}_{E}\cdot\nabla\Phi\right)\\\ &\hskip 34.14322pt+(1-\varepsilon^{2})\left(\nabla_{W}\cdot(H{\bf q})_{I}+{\bf q}_{I}\cdot\nabla\Phi\right)\\\ &{\bf u}_{E}^{(j)}\cdot\nabla_{UW}\theta_{2,E}+\frac{1}{\varepsilon^{2}}{\bf u}_{E}\nabla\theta_{0}\end{aligned}\right),$ where the WENO operators $\nabla_{CW}$, $\nabla_{W}$, $\nabla_{UW}$ and $\nabla^{WB}_{CW}$ will be defined below. The fully discrete scheme can be obtained by combining the spatial discretization (3.27) with the high order IMEX temporal discretization outlined in (3.20)-(3.23). Note that this involves an implicit step in (3.22). Following the approach to solve the first order IMEX method (3.6) in Section 3.2, we can convert it into an elliptic equation and then update the solution in a sequential way. For the high order scheme in (3.22), we form the following elliptic equation for $\rho_{2}$ (3.28) $\varepsilon^{2}\rho_{2,I}^{(i)}=\rho_{\star\star\star}^{(i)}+(\Delta ta_{ii})^{2}(1-\varepsilon^{2})^{2}\left(\Delta\Big{(}\frac{\gamma p_{0}}{\theta_{0}}\,\theta_{2,E}^{(i)}+\frac{\gamma p_{0}}{\rho_{0}\theta_{0}}\theta^{(i)}\,\rho_{2,I}^{(i)}\Big{)}+\nabla\cdot\Big{(}\rho_{2,I}^{(i)}\nabla\Phi\Big{)}\right),$ with (3.29) $\rho_{\star\star\star}^{(i)}=\rho_{\star\star}^{(i)}-\Delta ta_{ii}(1-\varepsilon^{2})\nabla_{W}\cdot{\bf q}_{\star\star}^{(i)}-\rho_{0},$ where (3.30a) $\displaystyle\rho_{\star\star}^{(i)}=\rho_{\star}^{(i)}-\Delta ta_{ii}\varepsilon^{2}\nabla_{CW}\cdot{\bf q}_{E}^{(i)},$ (3.30b) $\displaystyle{\bf q}_{\star\star}^{(i)}={\bf q}_{\star}^{(i)}-\Delta ta_{ii}\left(\nabla^{WB}_{CW}\cdot\left(\frac{{\bf q}\otimes{\bf q}}{\rho}+p{\mathbb{I}}\right)_{E}^{(i)}-\frac{\rho_{E}^{(i)}}{\rho_{0}}\nabla^{WB}_{CW}p_{0}\right).$ Here $\rho_{\star}^{(i)}$ and ${\bf q}_{\star}^{(i)}$ are updated from (3.21). For this elliptic equation, all the derivative operators in (3.28) are discretized with high order central differences, e.g., the compact central difference for the Laplacian operator as in [12]. After obtaining $\rho_{2,I}^{(i)}$, we can update $U_{I}^{(i)}$ from (3.22) explicitly in a sequential way. The detailed step-by-step flowchart of the fully discrete high order scheme is provided in the supplementary material. Taking the two dimensional model (3.24)-(3.26) described above as an example, these WENO operators $\nabla_{CW}$, $\nabla_{W}$, $\nabla_{UW}$ and $\nabla^{WB}_{CW}$ are defined as follows. * • Characteristic-wise WENO reconstruction $\nabla_{CW}$. The traditional characteristic-wise WENO reconstruction is applied to those conservative variables of the explicit part. Let us denote $U_{EC}=(\rho,\rho u,\rho v,E)^{T}_{E}$ and (3.31) $\mathcal{F}_{EC}^{x}=(\rho u,\rho u^{2}+p,\rho uv,(E+P)u)^{T}_{E},\quad\mathcal{F}_{EC}^{y}=(\rho v,\rho uv,\rho v^{2}+p,(E+P)v)^{T}_{E},$ with (3.32) $\mathcal{S}_{EC}^{x}=(0,-\rho\partial_{x}\Phi,-\rho u\partial_{x}\Phi,0)^{T}_{E},\quad\mathcal{S}_{EC}^{y}=(0,-\rho\partial_{y}\Phi,-\rho v\partial_{y}\Phi,0)^{T}_{E}.$ We note that $\mathcal{F}_{EC}^{x}$ and $\mathcal{F}_{EC}^{y}$ are the fluxes of 2D compressible Euler equations in a conservative form. They are exactly the first terms on the right-hand side of (3.25a) and (3.25b) (with a factor $\varepsilon^{2}$), corresponding to the conserved variable $U_{EC}$. We take $\mathcal{F}_{EC}^{x}$ and $\mathcal{F}_{EC}^{y}$ to form our Jacobian matrices and perform a 2D characteristic-wise finite difference WENO reconstruction, as detailed in [23, 36], to obtain their numerical approximation. The same reconstructed values are also used to evaluate the second terms, e.g., $\partial_{x}(\rho u^{2}+p)$ and $\partial_{x}(\rho uv)$ in the second term on the right side of (3.25a) (with a factor $(1-\varepsilon^{2})$). Such a characteristic-wise WENO reconstruction is denoted by $\nabla_{CW}$. $\nabla_{CW}$ is very important for essentially non- oscillatory shock capturing, compared with component-wise WENO reconstructions, especially for multi-dimensional cases [13]. * • Component-wise WENO reconstructions $\nabla_{W}$ and $\nabla_{UW}$. When the characteristic direction is unclear, the component-wise finite difference WENO reconstruction is applied to some first-order derivative terms, e.g., the third terms in (3.25a) and (3.25b) and those in (3.25c) and (3.25d). We refer to [23, 36] for detailed reconstruction procedures. Such reconstructions with a Lax-Friedrichs flux is frequently used. Denoting (3.33) $\mathcal{F}_{IC}^{x}=(\rho u,p_{2},0,(E+p)u)^{T}_{I},\quad\mathcal{F}_{IC}^{y}=(\rho v,0,p_{2},(E+p)v)^{T}_{I},$ and taking $\mathcal{F}_{IC}^{x}$ as an example, a Lax-Friedrichs flux splitting is given by (3.34) $(\mathcal{F}_{IC}^{x})^{\pm}_{ij}=\frac{1}{2}\Big{(}(\mathcal{F}_{IC}^{x})_{ij}\pm\Lambda(U_{EC})_{ij}\Big{)},$ here we take $\Lambda=\max\limits_{\rho,{\bf u},p}\\{|{\bf u}|+\min(1,1/\varepsilon)\sqrt{\frac{\gamma p}{\rho}}\\}$ and use $U_{EC}$ to obtain numerical viscosities. For component-wise WENO reconstructions with a Lax-Friedrichs numerical flux, we denote them as $\nabla_{W}$. On the other hand, the component-wise WENO reconstruction with an upwind numerical flux is used in the equation of $\theta_{2}$, and we denote it as $\nabla_{UW}$, e.g. the $\nabla_{UW}\theta_{2,E}$ term in (3.27). * • Well-balanced WENO reconstruction. For all Mach Euler equations with static gravitational fields (1.5), it admits an equilibrium steady-state solution (1.3). Namely, if the equilibrium (1.3) holds initially, it will hold for all later times. Numerically, well-balanced schemes are designed in order to preserve such a steady-state solution on the discrete level. Here we follow the approach proposed in [46, 32] to construct well-balanced WENO approximation. The main idea is to use exactly the same reconstructions to the flux and source terms, and revise the numerical viscosity to a prebalanced form so that the numerical viscosities vanish for equilibrium solutions. For the system (1.5), an equilibrium static solution would have $\rho=\rho_{0}({\bf x})$ and $p=p_{0}({\bf x})$, which satisfies (3.35) $\nabla p_{0}=-\rho_{0}\nabla\Phi.$ This is consistent with the asymptotic limit (2.6), namely, the equilibrium solution is also a solution to (2.8) [10]. Here we still take $\partial_{x}\mathcal{F}^{x}_{EC}$ as an example to demonstrate how such well- balanced WENO scheme is designed. First, a well-balanced Lax-Friedrichs flux splitting is based on (3.36) $(\mathcal{F}_{EC}^{x})^{\pm}_{ij}=\frac{1}{2}\Big{(}(\mathcal{F}_{EC}^{x})_{ij}\pm\Lambda(\tilde{U}_{EC})_{ij}\Big{)},$ where $\tilde{U}_{EC}=(\rho-\rho_{0},\rho{\bf u},E-\frac{p_{0}}{\gamma-1})^{T}_{E}$. For the source term (3.26), we have $\nabla\Phi=-\nabla p_{0}/\rho_{0}$ from (3.35), and can rewrite (3.26) as (3.37) $\mathcal{S}_{EC}^{x}=(0,-\frac{\rho}{\rho_{0}}\partial_{x}p_{0},\rho u\partial_{x}\Phi,0)^{T},\quad\mathcal{S}_{EC}^{y}=(0,-\frac{\rho}{\rho_{0}}\partial_{y}p_{0},0,\rho v\partial_{y}\Phi)^{T}.$ To approximate $\mathcal{S}_{EC}^{x}$ numerically, in analogous to (3.36), we split $p_{0}$ as (3.38) $(p_{0})^{\pm}_{ij}=\frac{1}{2}(p_{0})_{ij}.$ The finite difference characteristic-wise WENO reconstructions used to approximate $\\{(\mathcal{F}_{EC}^{x})^{\pm}_{ij}\\}$ will then be applied to $\\{(p_{0})^{\pm}_{ij}\\}$ with exactly the same weights (evaluated based on the smoothness indicators of $\\{(\mathcal{F}_{EC}^{x})^{\pm}_{ij}\\}$). In this way, we can obtain the exact balance of these two terms on the discrete level, when the equilibrium state is reached. We refer to [46, 32] for more detailed discussions. Similar approximations of $\partial_{y}\mathcal{F}^{y}_{EC}$ and $\mathcal{S}_{EC}^{y}$ can be applied. For those terms with above described well-balanced WENO reconstructions, we denote them as $\nabla_{CW}^{WB}$. ## 4\. Asymptotic preserving and asymptotically accurate properties In this section, we will prove that our first order IMEX scheme (3.4) is AP and the high order IMEX-RK scheme in Section 3.3 is AA. For any variable $w$, let us denote $w^{n}({\bf x})=w({\bf x},t^{n})$. During the discussion, we focus on time discretization and keep the space continuous. The initial solutions are assumed to be well-prepared, namely, the following expansions (4.1a) $\displaystyle\rho^{n}({\bf x})=\rho^{n}_{0}({\bf x})+\varepsilon^{2}\rho^{n}_{2}({\bf x}),$ (4.1b) $\displaystyle(\rho{\bf u})^{n}({\bf x})=(\rho{\bf u})^{n}_{0}({\bf x})+\varepsilon(\rho{\bf u})^{n}_{1}({\bf x}),$ (4.1c) $\displaystyle p^{n}({\bf x})=p^{n}_{0}({\bf x})+\varepsilon^{2}p^{n}_{2}({\bf x}),$ hold at $n=0$. The relation $p=(\rho\theta)^{\gamma}$ yields (4.2) $\theta^{n}({\bf x})=\theta^{n}_{0}({\bf x})+\varepsilon^{2}\theta^{n}_{2}({\bf x}).$ In addition, the initial conditions should satisfy (4.3) $\nabla\cdot(\rho{\bf u})_{0}^{0}=0,\quad\nabla p^{0}_{0}=-\rho_{0}^{0}\nabla\Phi.$ ### 4.1. Asymptotic preserving We start with proving the AP property of the first order IMEX scheme (3.4). ###### Theorem 4.1. With well-prepared initial conditions (4.1)-(4.2) at $n=0$ which satisfy (4.3), and assuming the expansion (4.1)-(4.2) at all later times, the first order IMEX scheme (3.4) with (3.9), under the assumptions of (3.8), $\nabla p_{0}=-\rho_{0}\nabla\Phi$ and $\theta_{0}$ is a constant, is asymptotic preserving, namely, as $\varepsilon\rightarrow 0$, the limiting scheme of (3.4) is a consistent discretization of (2.8). ###### Proof. As $\varepsilon\rightarrow 0$, under the assumptions that $\nabla p_{0}=-\rho_{0}\nabla\Phi$, $\theta_{0}$ is a constant, the scheme (3.4) becomes (4.4a) $\displaystyle\frac{\rho_{0}^{n+1}-\rho_{0}^{n}}{\Delta t}+\nabla\cdot(\rho{\bf u})_{0}^{n+1}=0,$ (4.4b) $\displaystyle\frac{(\rho{\bf u})_{0}^{n+1}-(\rho{\bf u})_{0}^{n}}{\Delta t}+\nabla\cdot(\rho{\bf u}\otimes{\bf u}+p{\mathbb{I}})_{0}^{n}+\nabla p^{n+1}_{2}=-\rho^{n}_{0}\nabla\Phi-\rho^{n+1}_{2}\nabla\Phi,$ (4.4c) $\displaystyle\frac{E_{0}^{n+1}-E_{0}^{n}}{\Delta t}+\nabla\cdot\left((E_{0}+p_{0}){\bf u}_{0}\right)^{n+1}=-\left(\rho{\bf u}\right)_{0}^{n+1}\cdot\nabla\Phi,$ (4.4d) $\displaystyle\frac{\theta_{2}^{n+1}-\theta_{2}^{n}}{\Delta t}+{\bf u}_{0}^{n}\cdot\nabla\theta_{2}^{n}=0,$ and the EOS (1.6) reduces to (at the time levels $t^{n}$ and $t^{n+1}$) (4.5) $E_{0}^{n}=\frac{p_{0}^{n}}{\gamma-1},\qquad E_{0}^{n+1}=\frac{p_{0}^{n+1}}{\gamma-1}.$ We now prove that starting from the well-prepared initial conditions (4.1)-(4.3) at $n=0$, the limiting scheme (4.4) is a consistent discretization to (2.8) after one time step. First of all, due to $\nabla p_{0}^{0}=-\rho_{0}^{0}\nabla\Phi$, (4.4b) becomes (4.6) $\frac{(\rho{\bf u})_{0}^{1}-(\rho{\bf u})_{0}^{0}}{\Delta t}+\nabla\cdot(\rho{\bf u}\otimes{\bf u})_{0}^{0}+\nabla p^{1}_{2}=-\rho^{1}_{2}\nabla\Phi.$ Taking divergence on both sides of (4.6) and using $\nabla\cdot(\rho{\bf u})^{0}_{0}=0$, we have (4.7) $\frac{\nabla\cdot(\rho{\bf u})_{0}^{1}}{\Delta t}+\nabla\cdot(\nabla\cdot(\rho{\bf u}\otimes{\bf u})_{0}^{0})+\nabla\cdot(\nabla p^{1}_{2}+\rho^{1}_{2}\nabla\Phi)=0.$ As $\varepsilon\rightarrow 0$, the elliptic equation (3.9) becomes (4.8) $0=\Delta t^{2}\,\nabla\cdot(\nabla\cdot(\rho{\bf u}\otimes{\bf u})^{0}_{0})+\Delta t^{2}\nabla\cdot\left(\nabla p_{2}^{1}+\rho_{2}^{1}\nabla\Phi\right),$ since $\widetilde{\rho}-(1-\varepsilon^{2})\,\Delta t\,\nabla\cdot(\widetilde{\rho{\bf u}})-\rho^{0}_{0}$ in (3.9) approaches $\Delta t^{2}\,\nabla\cdot(\nabla\cdot(\rho{\bf u}\otimes{\bf u})^{0}_{0})$, due to (3.5a) and (3.5b). Substituting (4.8) into (4.7) leads to (4.9) $\nabla\cdot(\rho{\bf u})_{0}^{1}=0,$ and from (4.4a), we obtain $\rho^{1}_{0}=\rho^{0}_{0}$. Utilizing (4.5), the energy equation (4.4c) becomes (4.10) $\frac{1}{\gamma-1}\frac{p_{0}^{1}-p_{0}^{0}}{\Delta t}+\nabla\cdot\left(\frac{\gamma\,p^{1}_{0}}{(\gamma-1)\rho^{1}_{0}}(\rho{\bf u})^{1}_{0}\right)=-\left(\rho{\bf u}\right)_{0}^{1}\cdot\nabla\Phi.$ If we assume a constant background potential temperature $\theta_{0}$ as in Section 2, from $p=(\rho\theta)^{\gamma}$, we have $p_{0}=C\,\rho^{\gamma}_{0}$ with $C=\theta_{0}^{\gamma}$, which leads to (4.11) $\nabla\left(\frac{p^{1}_{0}}{\rho_{0}^{1}}\right)=C\nabla(\rho^{1}_{0})^{(\gamma-1)}=C(\gamma-1)(\rho^{1}_{0})^{\gamma-2}\nabla\rho_{0}^{1},$ and using (4.9), we have (4.12) $\displaystyle\nabla\cdot\left(\frac{\gamma\,p^{1}_{0}}{(\gamma-1)\rho^{1}_{0}}(\rho{\bf u})^{1}_{0}\right)$ $\displaystyle=\frac{\gamma}{\gamma-1}\left((\rho{\bf u})^{1}_{0}\cdot\nabla\left(\frac{p^{1}_{0}}{\rho^{1}_{0}}\right)+\frac{p^{1}_{0}}{\rho^{1}_{0}}\nabla\cdot(\rho{\bf u})^{1}_{0}\right)$ $\displaystyle=C\gamma(\rho^{1}_{0})^{\gamma-2}(\rho{\bf u})^{1}_{0}\cdot\nabla\rho_{0}^{1}.$ And $\nabla p^{1}_{0}=-\rho^{1}_{0}\nabla\Phi$ yields (4.13) $-\left(\rho{\bf u}\right)_{0}^{1}\cdot\nabla\Phi=(\rho{\bf u})^{1}_{0}\cdot\frac{\nabla p^{1}_{0}}{\rho^{1}_{0}}=C\gamma(\rho^{1}_{0})^{\gamma-2}(\rho{\bf u})^{1}_{0}\cdot\nabla\rho_{0}^{1},$ therefore, the equation (4.10) leads to $p^{1}_{0}=p^{0}_{0}$. Note that the update of density equation in (4.4a) yields $\rho^{1}_{0}=\rho^{0}_{0}$, and the update of energy equation in (4.4c) yields $p^{1}_{0}=p^{0}_{0}$, which is consistent with the assumption $\nabla p_{0}^{1}=-\rho^{1}_{0}\nabla\Phi$ at the next time level. From (4.9), (4.6) and (4.4d), we observe that the limiting scheme (4.4) is a consistent discretization to (2.8) at the next time level, as well as $p^{1}_{0}=p^{0}_{0}$ and $\rho^{1}_{0}=\rho^{0}_{0}$. By the mathematical induction, we can prove that the same conclusion holds for all later time levels in a similar way, under the assumptions of the expansion (4.1)-(4.2) at any later time level $t^{n}$, which confirms the AP property of our scheme. ∎ ### 4.2. Asymptotically accurate The AP property can ensure the first-order scheme (3.4) to be a consistent discretization of the limiting equation (2.8) when $\varepsilon\rightarrow 0$. The high order scheme (3.20)-(3.23) is named AA [35, 12, 13], if it is AP and preserves the high order temporal accuracy as $\varepsilon$ goes to zero. ###### Theorem 4.2. Consider the high order semi-implicit IMEX-RK scheme (3.20)-(3.23) with space continuous, and use a $k$-th order stiffly accurate IMEX-RK scheme (3.19). Under the same assumptions for the AP property, for the solutions after one time step, we have $\lim_{\varepsilon\to 0}\theta^{1}({\bf x};\varepsilon)=\theta_{0},\,\lim_{\varepsilon\to 0}\nabla\cdot\Big{(}\rho^{1}({\bf x};\varepsilon){\bf u}^{1}({{\bf x}};\varepsilon)\Big{)}=0,\,\lim_{\varepsilon\to 0}(\nabla p^{1}({\bf x};\varepsilon)+\rho^{1}({\bf x};\varepsilon)\nabla\Phi)=0.$ If we denote ${\bf V}^{1}({\bf x};\varepsilon)=\big{(}\rho^{1}({\bf x};\varepsilon),\rho^{1}({\bf x};\varepsilon){\bf u}^{1}({\bf x};\varepsilon),p^{1}({\bf x};\varepsilon),\theta_{2}^{1}({\bf x};\varepsilon)\big{)}$ and let ${\bf V}^{e}({\bf x},t)=\big{(}\rho^{e}({\bf x},t),\rho^{e}({\bf x},t){\bf u}^{e}({\bf x},t),p^{e}({\bf x},t),\theta_{2}^{e}({\bf x},t)\big{)}$ be the exact solutions of (2.8), one has the one-step error estimate (4.14) $\lim_{\varepsilon\to 0}{\bf V}^{1}({\bf x};\varepsilon)={\bf V}^{e}({\bf x},\Delta t)+\mathcal{O}(\Delta t^{k+1}),$ namely, the high order semi-implicit scheme is AA. The proof follows from the mathematical induction and similar derivations of the AP analysis. A detailed proof is provided in the supplementary material. ## 5\. Numerical tests In this section, we will perform some numerical tests for the all-Mach full Euler equations with gravity, where the Mach number ranges from $0$ to ${\mathcal{O}}(1)$. For one-dimensional (1D) problems, we consider the vertical direction with coordinate denoted by $x$. For two-dimensional (2D) problems, we use the coordinate $(x,y)$ with $y$ being the vertical direction, unless specified otherwise. The proposed method was outlined in Section 3.4. The third order SA IMEX-RK scheme in [13, Section 3.2.2] will be used as temporal discretization, and the fifth order finite difference WENO scheme [36] is adopted in space. The time step is taken to be $\Delta t=\text{CFL}\,\Delta x/\Lambda$ with $\Lambda=\max\\{|{\bf u}|+\min(1,1/\varepsilon)\sqrt{{\gamma p}/{\rho}}\\}$, and we set $\text{CFL}=0.2$. Our proposed scheme is denoted as the “IMEX” scheme in the following. We will compare our scheme to an explicit fifth order well-balanced finite difference WENO scheme, as was developed in [32], which is denoted as “WB-Xing”. We will demonstrate that our approach is high order accurate, AP and AA, as well as that it is more computationally efficient as compared to WB-Xing in the low Mach regime. ### 5.1. 1D accuracy test We first consider the following 1D example with a linear gravitational field $\Phi_{x}=1$. The hydrostatic steady state satisfying (2.6) is given by (5.1) $\rho_{0}(x)=\left(1-\frac{\gamma-1}{\gamma}x\right)^{\frac{1}{\gamma-1}},\quad p_{0}(x)=\left(1-\frac{\gamma-1}{\gamma}x\right)^{\frac{\gamma}{\gamma-1}},\quad\theta_{0}(x)=1,$ with the adiabatic index $\gamma=1.4$. The corresponding perturbations are also at a steady state, where $\rho_{2}(x)=1+0.2\sin(\pi x),\qquad p_{2}(x)=4.5-x+0.2\cos(\pi x)/\pi,$ so that we have steady-state well-prepared solutions for (3.1) $\rho(x)=\rho_{0}(x)+\varepsilon^{2}\rho_{2}(x),u(x)=0,p(x)=p_{0}(x)+\varepsilon^{2}p_{2}(x),\theta_{2}(x)=\frac{1}{\rho}\left(\frac{\rho_{0}\theta_{0}}{\gamma p_{0}}p_{2}-\rho_{2}\theta_{0}\right).$ Here the computational domain is $\Omega=[0,2]$. We divide $\Omega$ into $N$ uniform cells, where $N=8\times 2^{i}\,(i=1,2,\ldots,5)$. The errors are computed by comparing the numerical solution to the exact solution at a final time $T=0.1$. We show the $L_{1}$ errors and orders for the density $\rho$ in Table 5.1, by taking four different global Mach numbers $\varepsilon=1,10^{-2},10^{-4}$ and $0$. We observe that our scheme has at least fifth order accuracy for all $\varepsilon$’s, as the spatial error is dominant and the solutions reach the steady state. The convergence orders of $\rho u$ and $E$ are similar, and we omit them to save space. Table 5.1. Example 5.1. $L_{1}$ errors and orders of $\rho$ with $\varepsilon=1,10^{-2},10^{-4},0$. $T=0.1$. N | $\varepsilon=1$ | $\varepsilon=10^{-2}$ | $\varepsilon=10^{-4}$ | $\varepsilon=0$ ---|---|---|---|--- error | order | error | order | error | order | error | order 16 | 9.19E-05 | – | 1.20E-08 | – | 1.33E-08 | – | 1.33E-08 | – 32 | 2.94E-06 | 4.96 | 2.94E-10 | 5.35 | 9.63E-11 | 7.11 | 9.65E-11 | 7.11 64 | 8.90E-08 | 5.05 | 5.80E-12 | 5.66 | 4.25E-12 | 4.50 | 4.25E-12 | 4.51 128 | 2.74E-09 | 5.02 | 8.99E-14 | 6.01 | 6.03E-14 | 6.14 | 6.04E-14 | 6.14 256 | 8.52E-11 | 5.01 | 1.64E-15 | 5.77 | 7.30E-16 | 6.37 | 7.33E-16 | 6.36 ### 5.2. 1D shock tube problem In this example, we consider a 1D shock tube problem in a high Mach regime where we take $\varepsilon=0.9$, and the initial conditions are given by $(\rho,u,p)=(1,0,1),\quad\text{ if }x<0.5;\qquad(0.125,0,0.1),\quad\text{ otherwise; }$ which has been studied in [13] without the gravitational source. Here we consider a linear gravity with $\Phi_{x}=1$. The background steady state is still (5.1). The computational domain is $[0,1]$, with inflow and outflow boundary conditions on the left and right boundaries respectively. We take $N=200$ and show the results at $T=0.1$ in Fig. 5.1. We can observe that the IMEX numerical solutions match those of WB-Xing very well, and both schemes work well in the high Mach regime when the shock exists. Figure 5.1. Example 5.2. Numerical solutions of $\rho$, $\rho u$ and $p$ with $N=200$ at $T=0.1$. ### 5.3. 2D accuracy test In this example, we consider a smooth 2D problem with a linear gravitational field $(\Phi_{x},\Phi_{y})=(1,1)$. The steady state is given by $\rho_{0}(x,y)=\left(1-\frac{\gamma-1}{\gamma}(x+y-2)\right)^{\frac{1}{\gamma-1}},\,\,p_{0}(x,y)=\left(1-\frac{\gamma-1}{\gamma}(x+y-2)\right)^{\frac{\gamma}{\gamma-1}},$ with $\theta_{0}(x,y)=1$ and the adiabatic index $\gamma=1.4$. Similar to Example 5.1, the perturbations are also at a steady state $\rho_{2}(x,y)=1+0.2\sin(\pi(x+y)),\quad p_{2}(x,y)=4.5-x+0.2\cos(\pi(x+y))/\pi,$ so that the steady-state exact solutions are $\rho(x,y)=\rho_{0}(x,y)+\varepsilon^{2}\rho_{2}(x,y),p(x,y)=p_{0}(x,y)+\varepsilon^{2}p_{2}(x,y),\theta_{2}=\frac{1}{\rho}\left(\frac{\rho_{0}\theta_{0}}{\gamma p_{0}}p_{2}-\rho_{2}\theta_{0}\right).$ Here ${\bf u}=(u_{0},-u_{0})$ which is perpendicular to the gravitational field, and we take $u_{0}=1$. The computational domain is $\Omega=[0,2]^{2}$, which is divided into $N^{2}$ uniform cells, with $N=8\times 2^{i}\,(i=1,2,\ldots,5)$. The errors are computed by comparing the numerical solution to the exact solution at the final time $T=0.5$. Similarly we show the $L_{1}$ errors and orders of $\rho$ in Tables 5.2, with four different global Mach numbers $\varepsilon=1,10^{-2},10^{-4}$ and $0$. We can observe the expected convergence rates for all $\varepsilon$’s, similar to the 1D case. Table 5.2. Example 5.3. $L_{1}$ errors and orders of $\rho$ with $\varepsilon=1,10^{-2},10^{-4},0$. $T=0.05$. $N^{2}$ uniform cells. $N$ | $\varepsilon=1$ | $\varepsilon=10^{-2}$ | $\varepsilon=10^{-4}$ | $\varepsilon=0$ ---|---|---|---|--- error | order | error | order | error | order | error | order $16$ | 4.86E-04 | – | 1.83E-06 | – | 1.83E-06 | – | 1.83E-06 | – $32$ | 3.70E-05 | 3.72 | 8.17E-08 | 4.49 | 8.14E-08 | 4.49 | 8.14E-08 | 4.49 $64$ | 1.48E-06 | 4.65 | 2.32E-09 | 5.14 | 2.31E-09 | 5.14 | 2.31E-09 | 5.14 $128$ | 5.51E-08 | 4.74 | 5.48E-11 | 5.41 | 5.41E-11 | 5.42 | 5.41E-11 | 5.42 $256$ | 1.38E-09 | 5.32 | 1.09E-12 | 5.66 | 1.06E-12 | 5.67 | 1.06E-12 | 5.67 ### 5.4. Traveling vortex This test has been considered for the 2D shallow water equations [22]. In (3.1), if we assume $\theta=1$ and take $c_{p}=2,c_{v}=1$, then the specific gas constant is $R=c_{p}-c_{v}=1$, and $\gamma=c_{p}/c_{v}=2$, it is equivalent to the shallow water equations and in this case the pressure is $p=\rho^{2}/2$ with $\varepsilon=0.05$. Following the settings in [22], for the source term we take $\Phi(x,y)=e^{-5(x-1)^{2}},$ namely, the gravity is along the $x$-direction for this example. Other initial conditions are given by $\displaystyle\rho(x,y,0)=110-\Phi+\left\\{\begin{aligned} &\left(\frac{\varepsilon\Gamma}{\omega}\right)^{2}(k(\omega\Gamma_{c})-k(\pi)),\quad&\text{if }\omega\Gamma_{c}\leq\pi;\\\ &0,\quad&\text{otherwise},\end{aligned}\right.$ $\displaystyle u(x,y,0)=2+\left\\{\begin{aligned} &\Gamma(1+\cos(\omega\Gamma_{c}))(0.5-y),\quad&\text{if }\omega\Gamma_{c}\leq\pi;\\\ &0,\quad&\text{otherwise},\end{aligned}\right.$ $\displaystyle v(x,y,0)=\left\\{\begin{aligned} &\Gamma(1+\cos(\omega\Gamma_{c}))(x-0.5),\quad&\text{if }\omega\Gamma_{c}\leq\pi;\\\ &0,\quad&\text{otherwise},\end{aligned}\right.$ with $\Gamma_{c}=\sqrt{(x-0.5)^{2}+(y-0.5)^{2}},\,\Gamma=8,\,\omega=4\pi,$ and $k(\xi)=2\cos(\xi)+2\xi\sin(\xi)+\frac{1}{8}\cos(2\xi)+\frac{\xi}{4}\sin(2\xi)+\frac{3}{4}\xi^{2}.$ The hydrostatic steady state takes the form $\rho_{0}(x,y)=110-\Phi,\qquad\theta_{0}=1,\qquad p_{0}(x,y)=\frac{1}{2}\left(110-\Phi\right)^{2}.$ Here the computational domain is $[0,2]\times[0,1]$, with periodic boundary conditions. We use this example to demonstrate that the new modeling (3.1) and our proposed scheme can have similar good performances as in [22] for this special case. We strictly follow our scheme in Section 3.4 with a mesh grid $200\times 100$ and show the perturbation of density $\rho-\rho_{0}$ in Fig. 5.2. We can observe that the numerical results are very similar to those obtained in [22], which demonstrates the nice performance of the proposed AP method in the low Mach regime. Figure 5.2. Example 5.4. Numerical solutions for the perturbation of density $\rho-\rho_{0}$. Mesh $200\times 100$. From left to right: $t=0.3,0.6,1.0$, respectively. ### 5.5. 2D isothermal equilibrium This example is designed to test the well-balanced property of a given scheme, as well as its ability to capture small disturbances around an equilibrium state [41]. We consider an isothermal equilibrium state (5.2) $\rho(x,y)=\bar{\rho}\exp\left(-\frac{\bar{\rho}g}{\bar{p}}(x+y)\right),\,{\bf u}=0,\,p(x,y)=\bar{p}\exp\left(-\frac{\bar{\rho}g}{\bar{p}}(x+y)\right),$ with $\bar{\rho}=1.21$, $\bar{p}=1$ and $g=1$. The linear gravitational field with $(\Phi_{x},\Phi_{y})=(1,1)$ is taken. The adiabatic index is $\gamma=1.4$ and the global Mach number is set to be $\varepsilon=0.9$. The problem is defined on a unit square $[0,1]^{2}$. We run the simulation up to the final time $T=1$, on three different meshes $50\times 50$, $100\times 100$ and $200\times 200$. We compare the numerical solutions to the steady-state exact solutions. It was observed that both IMEX and WB-Xing can preserve the solutions up to machine errors, and a well-balanced property is achieved for both schemes. We omit the results here to save space. Next, we add a small perturbation to the pressure $p_{2}(x,y)=\frac{1}{810}\exp\left(-\frac{100\bar{\rho}g}{\bar{p}}((x-0.3)^{2}+(y-0.3)^{2}\right)$ which is a small Gaussian hump centered at $(0.3,0.3)$. The other terms are kept to be the same. We take a mesh grid $100\times 100$ and compute the numerical solution up to $T=0.15$. Here a transmissive boundary condition [39] is used. The numerical results are shown in Fig. 5.3, and again we can observe that both schemes can capture the small perturbations very well in this high Mach regime. Figure 5.3. Example 5.5. Numerical solutions of the pressure perturbation and the density perturbation at $T=0.15$ with $N\times N=100\times 100$. Left: WB- Xing; middle: IMEX; right: cuts along the line $y=0.305\,m$. Top: pressure perturbation; bottom: density perturbation. ###### Remark 5.1. For this 2D isothermal equilibrium example, numerical boundary conditions are treated as follows. We split our main variable $U=(\rho,\rho{\bf u},E,\theta)^{T}$ into two parts: a hydrostatic one $U^{0}$ and a perturbation $U^{\prime}$, that is, $U=U^{0}+U^{\prime}$ with $U^{0}=\left(\rho_{0},{\bf 0},\frac{p_{0}}{\gamma-1},\theta_{0}\right)^{T},\qquad U^{\prime}=\left(\varepsilon^{2}\rho_{2},\rho{\bf u},E-\frac{p_{0}}{\gamma-1},\varepsilon^{2}\theta_{2}\right)^{T}.$ We apply a transmissive boundary condition [39] to the perturbation $U^{\prime}$, i.e., ghost values of $U^{\prime}$ are assigned by mirror symmetry. Ghost values of the hydrostatic component $U^{0}$ are obtained via extrapolation. The same boundary treatments are also applied to the following two examples. ### 5.6. Rising thermal bubble This is a benchmark test problem for atmospheric flows, simulating the dynamics of a warm bubble, which has been studied in [18, 19, 41]. Here we consider a linear gravitational field in the vertical direction with $\Phi_{x}=0$ and $\Phi_{y}=g$, and the gravitational constant is $g=9.8m/s^{2}$. The Exner pressure takes the form $\Pi=1-(\gamma-1)gy/\gamma/R/\bar{T}$, with $\bar{T}=300\,K$, $\gamma=1.4,\,$ $R=287.058\,\text{J/kg K}$ being the gas constant, and the potential temperature is $\theta(x,y,0)=\theta_{0}+\Delta\theta(x,y,0)=300\,K+\left\\{\begin{aligned} &0,&r>r_{c};\\\ &\frac{\theta_{c}}{2}(1+cos(\pi\,r/r_{c}))&r\leq r_{c}.\end{aligned}\right.$ Here $r=\sqrt{(x-x_{c})^{2}+(y-y_{c})^{2}}$, $\theta_{c}=0.5\,K$, $(x_{c},y_{c})=(500,350)\,m$ and $r_{c}=250\,m$, so that the initial condition can be denoted as $\rho(x,y,0)=\frac{\bar{p}}{R\theta}\,\Pi^{\frac{1}{\gamma-1}},\quad{\bf u}(x,y,0)=(0,0),\quad p(x,y,0)=\bar{p}\,\Pi^{\frac{\gamma}{\gamma-1}},$ with the hydrostatic steady state $\rho_{0}(x,y)=\frac{\bar{p}}{R\,\theta_{0}}\,\Pi^{\frac{1}{\gamma-1}},\qquad\theta_{0}(x,y)=300\,K,\qquad p_{0}(x,y)=\bar{p}\,\Pi^{\frac{\gamma}{\gamma-1}}.$ where $\bar{p}=10^{5}N/m^{2}$ is a reference pressure at $y=0\,m$, and the computational domain is $[0,1000]\times[0,1000]\,m^{2}$. This problem can be rewritten into a dimensionless form by choosing the reference values $\displaystyle p_{ref}=10^{5}\,N/m^{2},\quad\rho_{ref}=10\,kg/m^{3},\quad t_{ref}=10^{3}\,s,\quad l_{ref}=10^{3}\,m,$ $\displaystyle U_{ref}=\frac{l_{ref}}{t_{ref}}=1\,m/s,\qquad\theta_{ref}=\frac{p_{ref}}{R\rho_{ref}}=\frac{10^{4}}{R}\,K,$ so that the global Mach number is $\varepsilon=\frac{U_{ref}}{\sqrt{p_{ref}/\rho_{ref}}}=10^{-2}.$ We run the simulation for a very long time until the final stopping time $t=700s$. We show the perturbation of potential temperature $\Delta\theta$ on a mesh gird of $200\times 200$ in Fig. 5.4. Inviscid wall boundary conditions [41] are used. We can see that both schemes can correctly capture the shear movement of the rising bubble, which is along the opposite direction of gravity, and eventually, it forms a shape of a mushroom cloud. For this example with smooth solutions, we only use component-wise WENO reconstruction with linear weights for saving computational cost. However, comparing the WB- Xing scheme with our IMEX scheme, our scheme is much less diffusive on the same mesh grid, as the explicit WB-Xing scheme has a numerical viscosity inversely proportional to the Mach number $\varepsilon$, while our IMEX scheme does not. We also compare the CPU cost of two schemes in Table 5.3, from which we can observe that our IMEX scheme is much more efficient for this low Mach problem. Figure 5.4. Example 5.6. Numerical results for the perturbation of potential temperature $\Delta\theta$. From left to right, $t=400s,600s,700s$, respecively. Top: WB-Xing; Bottom: IMEX. Mesh: $200\times 200$. ### 5.7. Inertia-gravity wave This test arises from atmospheric flows, and has also been studied in [18, 19, 41]. The computational domain is set as $[0,300000]\times[0,10000]\,m^{2}$, with a periodic boundary condition in the $x$ direction and an inviscid wall boundary condition in the $y$ direction. The linear gravitational field with $(\Phi_{x},\Phi_{y})=(0,g)$ and $g=9.8\,m/s^{2}$ is considered. The Exner pressure is $\Pi=1+\frac{(\gamma-1)g^{2}}{\gamma RT_{0}\mathscr{N}^{2}}\biggl{[}\exp\biggl{(}-\frac{\mathscr{N}^{2}}{g}y\biggr{)}-1\biggr{]},$ and the initial potential temperature is denoted as $\theta(x,y,0)=\theta_{0}(y)+\Delta\theta(x,y,0),$ where $\theta_{0}(y)=\bar{T}\exp\biggl{(}\frac{\mathscr{N}}{g}y\biggr{)}\,K,\quad\Delta\theta(x,y,0)=\theta_{c}\sin\left(\frac{\pi y}{h_{c}}\right)\biggl{[}1+(x-x_{c})^{2}/a_{c}^{2}\biggr{]}^{-1},$ with the Brunt–Väisälä frequency $\mathscr{N}=0.01/s$, $\bar{T}=300\,K$, $\theta_{c}=0.01\,K$, $h_{c}=10000\,m$, $x_{c}=100000\,m$ and $a_{c}=5000\,m$. The initial conditions are given by $\rho(x,y,0)=\frac{\bar{p}}{R\theta}\,\Pi^{\frac{1}{\gamma-1}},\quad{\bf u}(x,y,0)=(20,0)\,m/s,\quad p(x,y,0)=\bar{p}\,\Pi^{\frac{\gamma}{\gamma-1}},$ and the equilibrium state is $p_{0}=\bar{p}\,\Pi^{\frac{\gamma}{\gamma-1}},\qquad\theta_{0}=\bar{T}\exp\biggl{(}\frac{\mathscr{N}}{g}y\biggr{)}\,K,\qquad\rho_{0}=\frac{\bar{p}}{R\theta_{0}}\Pi^{\frac{1}{\gamma-1}},$ with $\bar{p}=10^{5}\,N/m^{2}$. As in Example 5.6, we consider the problem in a dimensionless form, by choosing the following reference values $\displaystyle p_{ref}=10^{5}N/m^{2},\quad\rho_{ref}=10^{-1}kg/m^{3},\quad t_{ref}=10^{5}m/s^{2},\quad l_{ref}=10^{5}m,$ $\displaystyle U_{ref}=\frac{l_{ref}}{t_{ref}}=1\,m/s,\qquad\theta_{ref}=\frac{p_{ref}}{R\rho_{ref}}=\frac{10^{6}}{R}\,K,$ and the corresponding global Mach number is $\varepsilon=\frac{U_{ref}}{\sqrt{p_{ref}/\rho_{ref}}}=10^{-3}.$ In this example, for the WB-Xing scheme, an HLLC flux which is less diffusive than the Lax-Friedrichs flux, is used in the simulation. We run the solution up to $t=3000s$ for both schemes. For this example, we take CFL=$0.01$ for our IMEX scheme, and CFL=$0.005$ for the WB-Xing scheme. We show the perturbation of potential temperature $\Delta\theta$ on the mesh grid $400\times 50$ and $800\times 50$ in Fig. 5.5 and Fig. 5.6. From these numerical solutions, especially the cuts in Fig. 5.6, we can see the IMEX scheme has better resolution than the WB-Xing scheme on the coarse grid. We also compare the CPU cost of two schemes in Table 5.3, and observe that the IMEX scheme is much more efficient in this low Mach regime. Figure 5.5. Example 5.7. Numerical results for the perturbation of potential temperature $\Delta\theta$ at $t=3000s$. Mesh grid: $400\times 50$ (left); $800\times 50$ (right). Top: WB-Xing; bottom: IMEX. Figure 5.6. Example 5.7. Numerical results for the perturbation of potential temperature $\Delta\theta$ along the line $y=4900\,m$ at $t=3000s$. Mesh grid: $400\times 50$ (left); $800\times 50$ (right). Table 5.3. CPU cost (seconds) for the WB-Xing and IMEX schemes. | $N_{x}\times N_{y}$ | WB-Xing | IMEX ---|---|---|--- Example 5.6 | $100\times 100$ | 7,328.77 | 1,354.02 $\varepsilon=10^{-2}$ | $200\times 200$ | 55,719.46 | 11,712.07 Example 5.7 | $400\times 50$ | 300,512.70 | 12,578.01 $\varepsilon=10^{-3}$ | $800\times 50$ | 585,356.01 | 32,064.40 ## 6\. Conclusion In this work, we designed a high order semi-implicit AP well-balanced finite difference WENO scheme for the all Mach full Euler system with a gravitational field. It is much more challenging than the shallow water equations with a non-flat bottom [22], due to the existence of gravity which couples all equations together in the limit. We proposed to add the evolution for the perturbation of potential temperature in the design of our scheme, which provides a general and easy framework for the development of AP schemes to ensure the correct incompressible limit. The AP and AA properties are formally analyzed. Numerical tests on both 1D and 2D problems have demonstrated the high order accuracy, well-balanced, AP and AA properties of our proposed scheme. Our IMEX scheme has also been shown to yield better resolution and higher efficiency in the low Mach regime when compared with the explicit scheme. ## References * [1] E. Audusse, F. Bouchut, M.-O. Bristeau, R. Klein, and B. Perthame. A fast and stable well-balanced scheme with hydrostatic reconstruction for shallow water flows. SIAM Journal on Scientific Computing, 25(6):2050–2065, 2004. * [2] P. R. Bannon. On the anelastic approximation for a compressible atmosphere. Journal of the Atmospheric Sciences, 53(23):3618–3628, 1996. * [3] W. Barsukow, P. V. Edelmann, C. Klingenberg, and F. K. Röpke. A low-Mach Roe-type solver for the euler equations allowing for gravity source terms. ESAIM: Proceedings and Surveys, 58:27–39, 2017. * [4] T. Benacchio. A blended semi-implicit numerical model for weakly compressible atmospheric dynamics. PhD thesis, 2014. * [5] J. P. Berberich, P. Chandrashekar, and C. Klingenberg. A general well-balanced finite volume scheme for Euler equations with gravity. In XVI International Conference on Hyperbolic Problems: Theory, Numerics, Applications, pages 151–163. Springer, 2016. * [6] J. P. Berberich, P. Chandrashekar, C. Klingenberg, and F. K. Röpke. Second order finite volume scheme for Euler equations with gravity which is well-balanced for general equations of state and grid systems. Communications in Computational Physics, 26(2):599–630, 2019. * [7] A. Bermúdez, X. López, and M. E. Vázquez-Cendón. Finite volume methods for multi-component Euler equations with source terms. Computers & Fluids, 156:113–134, 2017. * [8] A. Bermudez and M. E. Vazquez. Upwind methods for hyperbolic conservation laws with source terms. Computers & Fluids, 23(8):1049–1071, 1994. * [9] C. Birke, C. Chalons, and C. Klingenberg. A low Mach two-speed relaxation scheme for the compressible Euler equations with gravity. arXiv preprint arXiv:2112.02986, 2021. * [10] G. Bispen, M. Lukáčová-Medvid’ová, and L. Yelash. Asymptotic preserving IMEX finite volume schemes for low Mach number Euler equations with gravitation. Journal of Computational Physics, 335:222–248, 2017. * [11] S. Boscarino, F. Filbet, and G. Russo. High order semi-implicit schemes for time dependent partial differential equations. Journal of Scientific Computing, 68(3):975–1001, 2016. * [12] S. Boscarino, J.-M. Qiu, G. Russo, and T. Xiong. A high order semi-implicit IMEX WENO scheme for the all-Mach isentropic Euler system. Journal of Computational Physics, 392:594–618, 2019. * [13] S. Boscarino, J.-M. Qiu, G. Russo, and T. Xiong. High order semi-implicit WENO schemes for all Mach full Euler system of gas dynamics. SIAM Journal on Scientific Computing, 4:B368–B394, 2022. * [14] J. C. Butcher. Numerical Methods for Ordinary Differential Equations. Third Edition, John Wiley & Sons Ltd, 2016. * [15] P. Chandrashekar and C. Klingenberg. A second order well-balanced finite volume scheme for Euler equations with gravity. SIAM Journal on Scientific Computing, 37(3):B382–B402, 2015. * [16] P. Chandrashekar and M. Zenk. Well-balanced nodal discontinuous Galerkin method for Euler equations with gravity. Journal of Scientific Computing, 71(3):1062–1093, 2017. * [17] V. Desveaux, M. Zenk, C. Berthon, and C. Klingenberg. A well-balanced scheme to capture non-explicit steady states in the Euler equations with gravity. International Journal for Numerical Methods in Fluids, 81(2):104–127, 2016. * [18] D. Ghosh and E. M. Constantinescu. Well-balanced, conservative finite difference algorithm for atmospheric flows. AIAA Journal, 54(4):1370–1385, 2016. * [19] F. X. Giraldo and M. Restelli. A study of spectral element and discontinuous Galerkin methods for the Navier–Stokes equations in nonhydrostatic mesoscale atmospheric modeling: Equation sets and test cases. Journal of Computational Physics, 227(8):3849–3877, 2008. * [20] L. Grosheintz-Laval and R. Käppeli. High-order well-balanced finite volume schemes for the Euler equations with gravitation. Journal of Computational Physics, 378:324–343, 2019. * [21] J. Haack, S. Jin, and J.-G. Liu. An all-speed asymptotic-preserving method for the isentropic Euler and Navier-Stokes equations. Communications in Computational Physics, 12(4):955–980, 2012. * [22] G. Huang, Y. Xing, and T. Xiong. High order well-balanced asymptotic preserving finite difference WENO schemes for the shallow water equations in all Froude numbers. Journal of Computational Physics, 463:111255, 2022. * [23] G.-S. Jiang and C.-W. Shu. Efficient implementation of weighted ENO schemes. Journal of Computational Physics, 126(1):202–228, 1996. * [24] R. Käppeli and S. Mishra. Well-balanced schemes for the Euler equations with gravitation. Journal of Computational Physics, 259:199–219, 2014. * [25] R. Klein. Asymptotics, structure, and integration of sound-proof atmospheric flow equations. Theoretical and Computational Fluid Dynamics, 23(3):161–195, 2009\. * [26] C. Klingenberg, G. Puppo, and M. Semplice. Arbitrary order finite volume well-balanced schemes for the Euler equations with gravity. SIAM Journal on Scientific Computing, 41(2):A695–A721, 2019. * [27] M. A. Kopera and F. X. Giraldo. Analysis of adaptive mesh refinement for IMEX discontinuous galerkin solutions of the compressible Euler equations with application to atmospheric simulations. Journal of Computational Physics, 275:92–117, 2014. * [28] A. Kurganov. Finite-volume schemes for shallow-water equations. Acta Numerica, 27:289–351, 2018. * [29] A. Kurganov and E. Tadmor. Solution of two-dimensional Riemann problems for gas dynamics without Riemann problem solvers. Numerical Methods for Partical Differential Equations, 18(5):584–608, 2002. * [30] G. Li and Y. Xing. High order finite volume WENO schemes for the Euler equations under gravitational fields. Journal of Computational Physics, 316:145–163, 2016. * [31] G. Li and Y. Xing. Well-balanced discontinuous Galerkin methods with hydrostatic reconstruction for the Euler equations with gravitation. Journal of Computational Physics, 352:445–462, 2018. * [32] G. Li and Y. Xing. Well-balanced finite difference weighted essentially non-oscillatory schemes for the Euler equations with static gravitational fields. Computers & Mathematics with Applications, 75(6):2071–2085, 2018\. * [33] S. Noelle, N. Pankratz, G. Puppo, and J. R. Natvig. Well-balanced finite volume schemes of arbitrary order of accuracy for shallow water flows. Journal of Computational Physics, 213(2):474–499, 2006. * [34] Y. Ogura and N. A. Phillips. Scale analysis of deep and shallow convection in the atmosphere. Journal of the Atmospheric Sciences, 19(2):173–179, 1962. * [35] L. Pareschi and G. Russo. Implicit-explicit Runge-Kutta schemes and applications to hyperbolic systems with relaxation. Journal of Scientific computing, 25(1-2):129–155, 2005. * [36] C.-W. Shu. Essentially non-oscillatory and weighted essentially non-oscillatory schemes for hyperbolic conservation laws. Advanced Numerical Approximation of Nonlinear Hyperbolic Equations, pages 325–432, 1998. * [37] M. Tang. Second order method for isentropic Euler equation in the low Mach number regime. Kinetic and Related Models, 5(1):155–184, 2012. * [38] A. Thomann, G. Puppo, and C. Klingenberg. An all speed second order well-balanced IMEX relaxation scheme for the Euler equations with gravity. Journal of Computational Physics, 420:109723, 2020. * [39] E. F. Toro. Riemann solvers and numerical methods for fluid dynamics: a practical introduction. Springer, 2009. * [40] R. Touma, U. Koley, and C. Klingenberg. Well-balanced unstaggered central schemes for the Euler equations with gravitation. SIAM Journal on Scientific Computing, 38(5):B773–B807, 2016. * [41] K. Wu and Y. Xing. Uniformly high-order structure-preserving discontinuous Galerkin methods for Euler equations with gravitation: Positivity and well-balancedness. SIAM Journal on Scientific Computing, 43(1):A472–A510, 2021. * [42] Y. Xing. Numerical methods for the nonlinear shallow water equations. In R. Abgrall and C.-W. Shu, editors, Handbook of Numerical Analysis, Volume 18, Handbook of Numerical Methods for Hyperbolic Problems: Applied and Modern Issues, volume 18, pages 361–384. North-Holland, Elsevier, Amsterdam, 2017. * [43] Y. Xing and C.-W. Shu. High order finite difference WENO schemes with the exact conservation property for the shallow water equations. Journal of Computational Physics, 208(1):206–227, 2005. * [44] Y. Xing and C.-W. Shu. A new approach of high order well-balanced finite volume WENO schemes and discontinuous Galerkin methods for a class of hyperbolic systems with source terms. Comput. Phys, 1(1):100–134, 2006. * [45] Y. Xing and C.-W. Shu. High order well-balanced finite volume WENO schemes and discontinuous Galerkin methods for a class of hyperbolic systems with source terms. Journal of Computational Physics, 214(2):567–598, 2006. * [46] Y. Xing and C.-W. Shu. High order well-balanced WENO scheme for the gas dynamics equations under gravitational fields. Journal of Scientific Computing, 54(2):645–662, 2013.
# FREDSR: Fourier Residual Efficient Diffusive GAN for Fast Single Image Super Resolution Kyoungwan Woo1 * Achyuta Rajaram 2 * ###### Abstract _FREDSR is a GAN variant that aims to outperform traditional GAN models in specific tasks such as Single Image Super Resolution with extreme parameter efficiency at the cost of per-dataset generalizeability. FREDSR integrates fast Fourier transformation, residual prediction, diffusive discriminators, etc to achieve strong performance in comparisons to other models on the UHDSR4K dataset for Single Image 3x Super Resolution from 360p and 720p with only 37000 parameters. The model follows the characteristics of the given dataset, resulting in lower generalizeability but higher performance on tasks such as real time up-scaling._ ## 1 Introduction Image to Image Translation(I2I) is a highly general meta-task that can summarize most application-specific subtasks in image generation or processing. Broadly, we use the definition from [1], where we seek to convert an an input image from domain A to domain B, while preserving ”intrinsic source content”, i.e. the meaningful semantic features of the image, while converting only the ”extrinsic style”, pertaining to the specific domain. This process can be defined as a mapping that synthesizes images that are indistinguishable from a sample from the target distribution B, using the semantic information present from a given sample of distribution A. This definition can generalize to many problems in image processing, with past work being used for image synthesis [2], image segmentation [3], style transfer [4], image inpainting [5], image super-resolution [6], among many other applications. Given the formulation of the I2I problem stated above, it seems natural that Generative Adversarial Networks(GAN) would be applicable for these tasks. GANs seek to generate images indistinguishable from those in a given distribution by playing a mini-max game between a discriminator and generator, where the generator seeks to decrease the discriminator’s accuracy in determining if the outputs of the generator belong to a specified distribution of images, while the discriminator seeks to maximise said accuracy [7]. This methodology circumvents the problems of using a pixel-level loss, or other naive supervised losses, as when they are used, blurry images are creates due to averaging of the multiple plausible solutions for a given translation. As a result of this natural applicability to GANs, past research has sought to apply GAN architectures to I2I translation problems, such as the pix2pix network, which pioneered the development of GANs for I2I translation [8]. Pix2pix allowed for high-quality low-resolution I2I translation to be computed without the need for problem-specific hand crafted loss functions. However, the initial formulation has not been crafted for high-resolutions, with [9] stating that conditional GANs such as pix2pix can struggle to generate high definition images, due to training instability. Furthermore, they indicate that including Perceptual Loss, an error function based on the outputs of classification networks, that seek to perform large-scale feature extraction, has significant benefits. Using this as a supervised objective allowed for stable images and was an improvement over raw pixel-level objectives. More recently, an improvement to pix2pix has been suggested, referred to as pix2pixHD, combining classifier-based perceptual loss with discriminator based loss in order to allow for high resolution image generation [10]. However, this model requires 3 discriminators to allow for the processing of information at every scale, which creates large increases to the computational resources required during training. Furthermore, using several discriminators increases learning instability, as matching gradient updates between the Generator and Discriminator is one of the largest challenges in GAN training, as not doing so properly leads to common GAN failure modes such as Catastrophic Forgetting and Mode Collapse [11]. An alternative to this architecture, optimized for large mask high resolution image inpainting, uses Fast Fourier Convolutions (FFCs) to allow for efficient processing of global-scale information along with the local-scale details [5],[12]. With this prior, we contribute: i. Using the previously introduced FFC block, we create a novel model architecture for I2I translation, adapting the original encoder-decoder stack from pix2pix. The use of the FFC allows the model to quickly gain access to global information. This allows for the model to operate on extrinsic global information throughout its entirety, improving its parameter efficiency for translation tasks. Additionally, this architecture is easily adaptable for Single Image Super Resolution(SISR) problems, and the maintaining of global information proves crucial to the performance of this architecture. ii. We extend the diffusive discriminator from [13], which seeks to combat discriminator overfitting by using a tool from the highly successful de- noising Diffusion Probabilistic Model architecture [14], adding diffusion from a Gaussian mixture distribution to the discriminator inputs, by using a residual discriminator design. Overall, we use all of these techniques to create a novel GAN variant, which experimentation has shown to be performant for specialized SISR tasks, while having much higher parameter efficiency in orders of magnitude. We motivate the existence of such a technique in the application of real-time 3D rendering, such as those found in modern video games. Modern gaming requires high resolution rendering of complicated 3D meshes with potentially billions of polygons, as well as expensive processes such as ray-traced per pixel lighting, ambient occlusion, and motion blur. Performance in this application has traditionally scaled harshly with output resolution, due to a quadratic increase in total calculations. Thus, a real-time application of SISR can provide efficiency benefits. One of the examples of utilizing unsupervised learning techniques for this task was introduced by NVIDIA with their Deep Learning Super Sampling application. FREDSR follows a similar notion of extremely performant, yet specialized SISR problem task, and was optimized for high performance on a specific dataset with extreme parameter efficiency. ## 2 Method We seek to develop a novel Single Image Super Resolution(SISR) architecture where we take a three channel input image high resolution and low resolution pair $q_{hr}$ and $q_{lr}$. We propose a model architecture that takes in $q_{lr}$ and outputs a three channel image $q_{h}$, which has the same resolution as $q_{hr}$. Figure 2: The model architecture of the FREDSR generator. This network uses the recently proposed FFCs [12], along with a complex loss function that combines adversarial, perceptual, and pixel-level losses. Graphic style was inspired by [5]. ### 2.1 Maintaining global information through all layers Challenging problems in I2I translation, including that of large-scale modification to high-resolution data, requires processing of global information. We argue that for successful high-resolution I2I translation, a high receptive field that encapsulates global information, and maintains it throughout the network is necessary in order to efficiently process global extrinsic parameters on an image, such as the time of day, lighting conditions, the artistic style associated with a painter, etc. Traditional fully convolutional models suffer from slow growth of effective receptive field, due to the small kernels used [15]. To improve on this, [16] proposed the dilated convolution, which introduces ”gaps” between the values of a convolution, ”spreading out” a 3x3 convolution, for example over an 8x8 area, while only using 9 parameters. However, this loses some local information present, meaning that a complex architecture that combines these with traditional convolutions is necessary in order to achieve usable results, as was seen in the I2I density map generation network proposed in [17]. Additionally, even the introduction of such dilated convolutions only slightly increases the receptive field. Thus, as a result, improvement is possible, which comes by way of the Fast Fourier Convolution (FFC)[12], based on the Fast Fourier Transform [18]. FFC splits information into two main channels, global and local, where the local information is processed using traditional convolutions, and the global information is processed using a real FFT2d. During each FCC block, the information of both branches are combined to allow for more efficient use of parameters. An illustration of the exact construction of the FFC block can be seen in Figure 2. This allows for the efficient use of parameters. This unit are fully differentiable, and we show them in used as a drop-in replacement for traditional convolutions. Additionally, FFCs have been showed to be well suited to capture periodic structures, which are extremely common both in natural and artificial environments, such as brick buildings or waves on the surface of water [5]. ### 2.2 Model Architecture Here, we introduce our novel I2I architecture, and the modifications necessary for high resolution SISR tasks with an extremely efficient network. We have already described the foundational units of this architecture, the FFC block. This architecture is illustrated in Figure 2. #### SISR Generator A traditional I2I network cannot perform SISR, due to the differing resolutions of the input and output. However, our architecture is easily adaptable to perform this task. We perform this adaption based off of the foundation of the VDSR model proposed in [19]. to begin, we upscale the low- resolution input by using a bicubic upscaler. Then we use our model to predict the difference in value between the input bicubic upscaled image, and the true image. This process is visualized in figure 2. This model has several advantages over other super-resolution networks, as the FFC module allow for the more efficient use of parameters, preservation of low-frequency features maintained through the upscaling. Additionally the residual learning allows for much more efficient use of parameters and faster training, as no training time or parameters are ”wasted” learning to copy the original image. This allows our model to achieve competitive performance to SOTA methods with with only 37115 total parameters on the specialized tasks. Figure 3: The model architecture of the FREDSR discriminator. This was based on a traditional discriminator, with special modifications for the super- resolution task. This discriminator uses diffusion applied to its inputs, among other tactics to improve discriminator performance and prevent overfitting. Graphic style was inspired by [5]. #### SISR Discriminator We allow the discriminator to train to distinguish between the residuals, also known as the differences between the upscaled images and the true images, and our model’s predicted residuals. This architecture is visualized in full in figure 3. ### 2.3 Generative Loss Functions In FREDSR, we combine various L1 and L2 losses. The benefits of combining various level losses have been previously discussed in papers such as [8]. #### SSIM loss In the case of L2 loss, we utilize the Structural Similarity Index Metric for the SISR model. In conjunction with the residual discriminator, SSIM works to deblurr the image. Compared to common L1/L2 losses such as MAE and MSE, SSIM has been proved to outperform in super resolution tasks, such as in [20]. The loss function is as follows: $\mathcal{L}_{SSIM}=-SSIM(x,y)=-\frac{(2\mu_{x}\mu_{y}+C_{1})+(2\sigma_{xy}+C_{2})}{(\mu_{x}^{2}+\mu_{y}^{2}+C_{1})(\sigma_{x}^{2}+\sigma_{y}^{2}+C_{2})}$ (1) With x,y defined as the generated and real images respectively. #### Charbonnier Loss As a method of enforcing color similarity, as well as mixing L1 and L2 losses, we use the Charbonnier Loss function. This is a variation of MAE which is differentiable and has shown to mix beneficial properties of L1 and L2 losses, and has shown significant success. $\mathcal{L}_{CHARB}=E(\sqrt{(x-y)^{2}+\epsilon})$ (2) With x,y defined as the generated and real images respectively, and $\epsilon$ a constant chosen. #### MGE loss To improve the performance of the model for SISR tasks, we add an additional loss on top of the global feature reconstruction loss. This loss is focused on the maintaining of edges, and was introduced for SISR in [21]. The maintaining of sharp edges allows for a more accurate image to that captured at high resolution. MGE uses a classical sobel operator, first introduced in [22]. After this, we can compute the gradients for each pixel, then define MGE as follows. $\mathcal{L}_{MGE}=E((G(x)-G(y))^{2})$ (3) With x,y defined as the generated and real images respectively. #### Adversarial Loss In the case of adversarial losses, as the model implementation follows a GAN structure, we utilize the modified MiniMax loss from [7].The loss function is as follows. $\mathcal{L}_{adv}=E_{(y_{gen})}[log(D(y_{gen}))]$ (4) where $y_{gen}$ are corresponding images taken from the original and high- resolution generated images. #### Final Generative loss Additionally to the previous losses, we add a Perceptual Loss, as they have shown incredible benefits within the training and evaluation of image generation networks [23],[24],[25],[26]. Combining this with the above loss functions, we end up with the final generative loss function to be minimized by gradient descent as: $\mathcal{L}_{final}=\lambda_{1}*\mathcal{L}_{adv}+\lambda_{2}*\mathcal{L}_{PL}+\lambda_{3}*\mathcal{L}_{MGE}+\lambda_{4}*\mathcal{L}_{SSIM}+\\\ {\lambda_{5}*\mathcal{L}_{CHARB}}$ (5) with $\lambda_{1}$,$\lambda_{2}$,$\lambda_{3}$,$\lambda_{4}$, $\lambda_{5}$ as scaling parameters. ### 2.4 Discriminative Loss Functions #### Modified Minimax Loss In the case of discriminative losses, as the model implementation follows a GAN structure, we train the model to recognize real images as greater than 0.5, and fake images as less than 0.5. Thus, our loss is given by: $\mathcal{L}_{disc}=E_{(x,y)}[log(D(y)]+E_{(y)}[log(1-D(G(x))]$ (6) where x,y are corresponding images taken from the original and high-resolution images respectively. ## 3 Experiments FREDSR was developed on Tensorflow [27] and Keras [28], then trained on UHDSR4K, a SIRS benchmark dataset from [29]. The images were downscaled using bicubic algorithms, consistent with the original dataset. Along with the techniques outlined in Sections 1 and 2, we utilize various traditional and novel GAN training methods, such as discriminator restart learning from [30] and variable Gaussian noise, an improved version of adaptive blur and control from [31]. It is important to note that FREDSR utilized minimal to no hyperparameter tuning: all $\lambda$ values were set to equalize the importance of each loss function. ### 3.1 Diffusive Discriminator A common problem with GAN stability is combating discriminator overfitting, and several methods have been used to address this, including the use of random data augmentations which adapt with the discriminator’s performance. [32] For stable training of GANs, we implement the solution proposed in [13], using adaptive noise samples from the forward chain of a Gaussian Mixture distribution. This method has been shown both theoretically and experimentally to provide stable and data-efficient GAN training, and improve over baselines for high-quality image generation [13]. In this work, we implement the diffusive inputs along with our residual discriminator. ### 3.2 Discriminator Restart Learning Even with diffusion applied to the discriminator inputs, the discriminator tends to converge faster than the generator, or simply learn the wrong weights after an extended period of training. By resetting the discriminator learning rate mid training, or a harsh full weight reinitialization, we ensure that the discriminator continues learning based on a more converged generator output. In this reset step, the discriminator receives a learning rate increase. The generator’s focus on the adversarial loss is lowered as well, to allow the discriminator to learn more freely. Once the discriminator is near convergence again, the changes to the coefficients are reversed. ### 3.3 Gaussian Random Noise Adding randomized noise to each convolution of the generator is a commonly used tactic in deep learning to increase the generalizability of a model while decreasing the chance of overfitting. [33] Although our datasets are near the larger scale, to improve the higher-resolution generalizability of our model as well as speed up training, we implement Gaussian random noise generations in between each of our fast fourier convolutions. The gaussian noise levels should decrease similarly to the generator loss, as in the later phases of training, randomized data can lead to less favorable outputs. ### 3.4 AdamW Optimizer We utilize the AdamW optimizer proposed by [34] to combat the weight regularization introduced by Adam. Adam tends to regularize larger weights less than smaller weights, and thus does not converge as well as traditional methods such as stochastic gradient descent. AdamW has been proposed to improve this issue, and we notice a significant improvement over the Adam optimizer. ### 3.5 Cosine Annealing Decay Learning Rate We utilize a cosine annealing decay learning rate scheduler to combine both decaying learning rate schedulers and warm restart learning rate schedulers proposed by [35]. Restart learning has been used commonly with stochastic gradient descent to speed up training processes, and exponential decay has been used to help models converge. Figure 4: A graph plot of the learning rate for the SISR generator: Each peak is five percent lower than the previous, and the learning rate will decay until halved prior to the next reset. ### 3.6 UHDSR4K Image Super Resolution For our main analysis, we focus on the super-resolution objective of the UHDSR4K dataset introduced in [29]. This dataset is the largest-scale UHD dataset in the field of 4K image super-resolution. For training, we downsample the original 4K images to 1920x1080. From here, we construct our 3x upsampling dataset, derived from the original UHDSR4K, by further downsampling to 640x360 resolution, and training our model to perform 3x super-resolution. This dataset is composed of diverse environments including city scenes, people, animals, buildings, cars, natural landscapes, and sculptures scraped from the internet in order to diversify the camera processing applied, along with allowing for high resolution images from many surroundings. In order to compute our downsampling, we follow [29] by using bicubic interpolation. For our metrics, we use standard super-resolution metrics, specifically SSIM and PSNR [36],[37]. Drawing comparisons using all of these methods allows for verifying that our model can both perform the super-resolution task and successfully match pixels, as well as is globally closer to the original image. It is important to note that our model’s hyperparameters were not fully tuned, yielding suboptimal performance. We visualize generator loss with respect to training steps in figure 5. Figure 5: A partial graph plot of the generator loss over train steps. Note than our model converges much slower than other models such as VDSR[19]. We perform training on 12x Nvidia Tesla V100s for over 300 epochs. #### Comparisons to Baselines For our comparison, we use several baseline models, which allow for us to directly benchmark our model’s performance on its original training dataset. We note that although it would be preferred to train our baseline models on the UHDSR4K dataset, as we have done with our model, due to time and resource constraints we were unable to do so. However, as this dataset has high diversity, we believe that it has coverage of the training distributions of the baseline models. We test all models at 3x super-resolution on the UHDSR4K dataset, with our training-time downscaling applied. Thus, we test 3x super resolution from 640 pixels by 360 pixels to 1920 pixels by 1080 pixels on the test dataset of UHDSR4K. For our baseline comparison models, we use only publically available pretrained models. We select 3 strong deep learning baselines with diverse techniques used. These baselines are EDSR[38], ESPCN[39], and FSRCNN [40]. For completeness, we also report bilinear and bicubic upscaling as a control. The results, calculated using the SSIM, and PSNR metrics as described previously are tabulated in Table 1. UHDSR4K 640x320 3x Upscaling --- Method | Params | SSIM $\uparrow$ | PSNR$\uparrow$ Bicubic | 0 | 0.856 | 26.75 Bilinear | 0 | 0.848 | 26.49 EDSR | 43000k | 0.879 | 27.61 ESPCN | 20k | 0.858 | 27.02 FSRCNN | 12k | 0.857 | 27.03 FREDSR | 37k | 0.883 | 27.776 Table 1: Quantitative evaluation of 3x SISR performance on downscaled UHDSR4K dataset. We report Structural Similarity Index Measure(SSIM) and Peak Signal to Noise Ratio(PSNR) metrics. FREDSR compares favorably against these baselines across all metrics. As shown by this, our method can outperforms the baselines in the task of SISR on its original training resolution and dataset, strictly outperforming significantly larger models, indicating high overall performance and parameter efficiency. Further ablation analysis is necessary in order to determine the exact root of this high performance, but we believe it can be attributed to the poor generalization of the other models, high receptive field of FREDSR, and ability to take advantage of the information present in repetitive structures in order to perform more accurate prediction. #### High resolution Performance Across Resolutions Figure 6: Positive samples for high resolution performance. These are examples of our model’s performance on the UHDSR4K dataset, in order to display our model’s high-resolution super-resolution capabilities. The top image is our model’s output, while the bottom is the true high resolution image. As shown here, along with our results from table 2, show an interesting improvement in our models as the resolution improves. This can be attributed to the task of super-resolution becoming easier at higher resolutions. Along with our model’s performance on low resolutions, we seek to find out how it performs on high resolution upscaling. To this end, we study performance on high resolution upscaling. We point out the high amount of repetitive patches present in urban scenes, and study the ability of our network to handle repeated signals deformed by changes in perspective. This signal deformation has been shown to cause challenges to past networks implementing FFCs [5], thus we study our model’s performance in this domain. We compare our model against a variety of strong baseline solutions for upscaling, and train a model, using the pretrained 360p model as a base, to do 3x super resolution from 720p to 4k. #### UHDSR4K For our full resolution UHDSR4K analysis, we perform 3x upscaling from the standard resolutions 720p to 4k resolutions. For our comparison, we use the PNSR and SSIM results from [29], testing our models against SRCNN[Dong2015], FSRCNN [40],VSDR [19], EDSR[38], RCAN[41], RDN[42], HAN[43], DRLN[44], and MANet [29]. We report PSNR and SSIM results from [29] along with our results and compare our models generalization performance at higher resolution against models that have been natively trained on upscaling the higher resolution. We also include Bilinear and Bicubic upscaling for control. UHDSR4K 1280x720 3x Upscaling --- Method | Params | SSIM $\uparrow$ | PSNR$\uparrow$ Bilinear | 0 | 0.9240 | 30.781 Bicubic | 0 | 0.9375 | 31.984 SRCNN | 57l | 0.9503 | 34.082 FSRCNN | 12k | 0.9462 | 33.614 VDSR | 665k | 0.9575 | 35.115 EDSR | 43000k | 0.9608 | 35.674 RCAN | 16000k | 0.9608 | 35.576 RDN | 21900k | 0.9614 | 35.769 HAN | 20000k | 0.9601 | 35.547 DRLN | 34000k | 0.9617 | 35.808 MANet | 27000k | 0.9618 | 35.842 FREDSR | 37k | 0.9645 | 35.3344 Table 2: Quantitative evaluation of 3x SISR generalization performance on original UHDSR4K dataset. We report Structural Similarity Index Measure(SSIM), and Peak Signal to Noise Ratio(PSNR) metrics. We report metrics for other deep learning based methods from [29]. FREDSR compares favorably against these baselines across all metrics, even though it is comprised of far fewer parameters. This can be attributed to the scale invariance caused by the inductive biases in the FFC modules. As demonstrated here, we outperform several strong baselines, despite having orders of magnitudes of fewer parameters. This indicates an ability to efficiently capture the requisite information, which we can attribute to the use of FFC and the process of residual learning. Additionally, we see an increase in performance as resolution increases, where a model trained on multiple resolutions outperforms a model trained on only one. This is potentially due to the nature of the task of super-resolution itself, however further analysis is required to determine the root cause of this phenomenon. #### Other datasets However, we note that as expected, FREDSR does not perform as highly across other datasets, such as Manga109 or Urban100, from [45] and [46] respectively. This is visualized in Figure 7, where we show our poor generalization performance. The reasoning behind this is clear: bicubic upscaling does not perform as well on such datasets at all, and the learned residual images are generally lacking of color and of a much higher resolution, compared to the low resolution. This shows that FREDSR is highly specialized, losing generalizability at the benefit of extreme parameter efficiency. However, for the main real-world application of video game upscaling, this should not be a major issue, due to the consistent resolution and environment present. Figure 7: Poor performance, with low clarity in edges, can be seen on the Manga109 or Urban100 datasets, from [45] and [46] respectively. This is potentially caused by the large distributional shift present, namely the change in resolution and color in residual images. ## 4 Related Work Within the domain of super-resolution, several traditional algorithms and data-driven models have been designed to accurately upscale images. Such classical algorithms are aptly covered in [47], and will not be discussed further here, as most state of the art solutions to SISR are based off of learning based techniques [48]. The use of convolutional networks for super resolution was first introduced in [Dong2015], which also initiated the use of deep learning for this task. Skip connections, integral to stabilizing training, improving parameter efficiency, and increasing performance [49] were first introduced with the development of DRCN [19]. The use of GANs for this task, which have the major benefit of generating high-quality images when compared to more traditional methods, was introduced in [50]. The main challenge faced with GANs, their low mode coverage, or output diversity, is not applicable to the super-resolution objective, meaning that the main advantages of GANs, namely their comparatively faster sampling and high quality outputs than diffusion networks and VAEs respectively, can be utilized [51]. In a similar fashion to [50], [10] combines adversarial and perceptual loss functions in order to train a GAN for super-resolution. Other works have noted the lack of high receptive field present in most super-resolution networks, pointing out that redundant patches within an image, common in both artificial and natural landscapes, allow for global context to be more informative to the computation of local upscaling. As a result, several works have taken advantage of the increased receptive field of the dilated, or atrous, convolution [52], [21]. This further motivates our use of Fourier convolutions, where we take advantage of their theoretically unbounded receptive fields. Other works have also used recurrent and attention mechanisms to allow for global cross-patch information usage [41], [43], [53], [54]. These attention-based mechanisms are an extremely interesting novel field for further works to explore. ## 5 Discussion In this study, we investigate the use of the Fast Fourier convolution for solving the specialized SISR objective. Using these convolutions, we have constructed a novel GAN architecture that has shown to outperform several baselines, while being significantly smaller. Our model arguably shows good performance in urban settings due to the highly repeated sectors present, which seems challenging to other methods, as seen in figure 1 and 5. Our model seems to have extremely high parameter efficiency, outperforming models which are comprised of more than a thousand times as many parameters. In exchange for this, we lose generalizability outside of a given image distribution. We believe that this tradeoff can be harmful, but mitigated under a circumstance where we believe our method is strongest-video game upscaling. This opinion is corroborated by the fact that video games have similar highly repeating sectors to urban scenes, due to their general usage of built-in textures for three-dimensional models. This model is highly efficient, which is necessary for real-time application on consumer graphics hardware, and can be trained for a given game, as within this narrow domain significant generalization is not required. This would allow for rendering at a low resolution and upscaling. Further work would be required in this domain to allow for enforced temporal coherence. However, convolutions are not the only approach to increasing the receptive field of image processing networks, and models like MLP Mixer models [55] as well as vision transformers [56] are exciting topics for future study in this field. Especially in the case of super resolution, higher receptive fields will unlock possibilities for the development of deep learning for high- resolution image processing, allowing for an increase in fidelity, diversity, and processing speed. Improved super resolution will allow for advancements in medical imaging [57] and compression of image and video formats [58], among many other applications. ## Acknowledgments The authors acknowledge the MIT SuperCloud and Lincoln Laboratory Supercomputing Center for providing (HPC, database, consultation) resources that have contributed to the research results reported within this paper. We additionally thank Professor Leslie Kaelbling and Ge Yang from the MIT Computer Science and Artificial Intelligence Laboratory for their valuable discussions and providing us with access to resources to complete this research. Lastly, we acknowledge Tensorflow and Horovod as the development language and platform of choice for this project. ## References * [1] Yingxue Pang, Jianxin Lin, Tao Qin, and Zhibo Chen. Image-to-image translation: Methods and applications, 2021. * [2] Peihao Zhu, Rameen Abdal, Yipeng Qin, and Peter Wonka. Sean: Image synthesis with semantic region-adaptive normalization. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020. * [3] Xi Guo, Zhicheng Wang, Qin Yang, Weifeng Lv, Xianglong Liu, Qiong Wu, and Jian Huang. Gan-based virtual-to-real image translation for urban scene semantic segmentation. Neurocomputing, 394:127–135, 2020. * [4] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A. Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In 2017 IEEE International Conference on Computer Vision (ICCV), pages 2242–2251, 2017. * [5] Roman Suvorov, Elizaveta Logacheva, Anton Mashikhin, Anastasia Remizova, Arsenii Ashukha, Aleksei Silvestrov, Naejin Kong, Harshith Goka, Kiwoong Park, and Victor Lempitsky. Resolution-robust large mask inpainting with fourier convolutions. CoRR, abs/2109.07161, 2021. * [6] Yuan Yuan, Siyuan Liu, Jiawei Zhang, Yongbing Zhang, Chao Dong, and Liang Lin. Unsupervised image super-resolution using cycle-in-cycle generative adversarial networks. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pages 814–81409, 2018. * [7] Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial networks, 2014. * [8] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A. Efros. Image-to-image translation with conditional adversarial networks, 2016\. * [9] Qifeng Chen and Vladlen Koltun. Photographic image synthesis with cascaded refinement networks, 2017. * [10] Ting-Chun Wang, Ming-Yu Liu, Jun-Yan Zhu, Andrew Tao, Jan Kautz, and Bryan Catanzaro. High-resolution image synthesis and semantic manipulation with conditional gans, 2017. * [11] Hoang Thanh-Tung and Truyen Tran. On catastrophic forgetting and mode collapse in generative adversarial networks, 2018. * [12] Lu Chi, Borui Jiang, and Yadong Mu. Fast fourier convolution. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 4479–4488. Curran Associates, Inc., 2020. * [13] Zhendong Wang, Huangjie Zheng, Pengcheng He, Weizhu Chen, and Mingyuan Zhou. Diffusion-gan: Training gans with diffusion, 2022. * [14] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models, 2020. * [15] Wenjie Luo, Yujia Li, Raquel Urtasun, and Richard Zemel. Understanding the effective receptive field in deep convolutional neural networks. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. * [16] Fisher Yu and Vladlen Koltun. Multi-scale context aggregation by dilated convolutions, 2015. * [17] Yuhong Li, Xiaofan Zhang, and Deming Chen. Csrnet: Dilated convolutional neural networks for understanding the highly congested scenes, 2018. * [18] E. O. Brigham and R. E. Morrow. The fast fourier transform. IEEE Spectrum, 4(12):63–70, 1967. * [19] Jiwon Kim, Jung Kwon Lee, and Kyoung Mu Lee. Accurate image super-resolution using very deep convolutional networks. CoRR, abs/1511.04587, 2015. * [20] Hang Zhao, Orazio Gallo, Iuri Frosio, and Jan Kautz. Loss functions for image restoration with neural networks. IEEE Transactions on Computational Imaging, 3(1):47–57, 2017. * [21] Zhengyang Lu and Ying Chen. Single image super resolution based on a modified u-net with mixed gradient loss, 2019. * [22] N. Kanopoulos, N. Vasanthavada, and R.L. Baker. Design of an image edge detection filter using the sobel operator. IEEE Journal of Solid-State Circuits, 23(2):358–367, 1988. * [23] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and super-resolution, 2016\. * [24] Cheng Shi and Chi-Man Pun. Adaptive multi-scale deep neural networks with perceptual loss for panchromatic and multispectral images classification. Information Sciences, 490:1–17, 2019. * [25] Yifan Liu, Hao Chen, Yu Chen, Wei Yin, and Chunhua Shen. Generic perceptual loss for modeling structured output dependencies, 2021\. * [26] Richard Zhang, Phillip Isola, Alexei A. Efros, Eli Shechtman, and Oliver Wang. The unreasonable effectiveness of deep features as a perceptual metric, 2018. * [27] Martín Abadi, Ashish Agarwal, Paul Barham, Eugene Brevdo, Zhifeng Chen, Craig Citro, Greg S. Corrado, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Ian Goodfellow, Andrew Harp, Geoffrey Irving, Michael Isard, Yangqing Jia, Rafal Jozefowicz, Lukasz Kaiser, Manjunath Kudlur, Josh Levenberg, Dandelion Mané, Rajat Monga, Sherry Moore, Derek Murray, Chris Olah, Mike Schuster, Jonathon Shlens, Benoit Steiner, Ilya Sutskever, Kunal Talwar, Paul Tucker, Vincent Vanhoucke, Vijay Vasudevan, Fernanda Viégas, Oriol Vinyals, Pete Warden, Martin Wattenberg, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng. TensorFlow: Large-scale machine learning on heterogeneous systems, 2015\. Software available from tensorflow.org. * [28] Francois Chollet et al. Keras, 2015. * [29] Kaihao Zhang, Dongxu Li, Wenhan Luo, Wenqi Ren, Björn Stenger, Wei Liu, Hongdong Li, and Ming-Hsuan Yang. Benchmarking ultra-high-definition image super-resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 14769–14778, October 2021. * [30] Kun Li and Dae-Ki Kang. Enhanced generative adversarial networks with restart learning rate in discriminator. Applied Sciences, 12:1191, 01 2022. * [31] Igor Susmelj, Eirikur Agustsson, and Radu Timofte. Abc-gan : Adaptive blur and control for improved training stability of generative adversarial networks. 2017\. * [32] Tero Karras, Miika Aittala, Janne Hellsten, Samuli Laine, Jaakko Lehtinen, and Timo Aila. Training generative adversarial networks with limited data, 2020. * [33] Guozhong An. The effects of adding noise during backpropagation training on a generalization performance. Neural Computation, 8(3):643–674, 1996. * [34] Ilya Loshchilov and Frank Hutter. Fixing weight decay regularization in adam. CoRR, abs/1711.05101, 2017. * [35] Ilya Loshchilov and Frank Hutter. SGDR: stochastic gradient descent with restarts. CoRR, abs/1608.03983, 2016. * [36] Zhou Wang, Alan Bovik, Hamid Sheikh, and Eero Simoncelli. Image quality assessment: From error visibility to structural similarity. Image Processing, IEEE Transactions on, 13:600 – 612, 05 2004. * [37] Alain Horé and Djemel Ziou. Image quality metrics: Psnr vs. ssim. In 2010 20th International Conference on Pattern Recognition, pages 2366–2369, 2010. * [38] Bee Lim, Sanghyun Son, Heewon Kim, Seungjun Nah, and Kyoung Mu Lee. Enhanced deep residual networks for single image super-resolution. In 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pages 1132–1140, 2017. * [39] Wenzhe Shi, Jose Caballero, Ferenc Huszár, Johannes Totz, Andrew P. Aitken, Rob Bishop, Daniel Rueckert, and Zehan Wang. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1874–1883, 2016. * [40] Chao Dong, Chen Change Loy, and Xiaoou Tang. Accelerating the super-resolution convolutional neural network, 2016. * [41] Yulun Zhang, Kunpeng Li, Kai Li, Lichen Wang, Bineng Zhong, and Yun Fu. Image super-resolution using very deep residual channel attention networks. CoRR, abs/1807.02758, 2018. * [42] Yulun Zhang, Yapeng Tian, Yu Kong, Bineng Zhong, and Yun Fu. Residual dense network for image super-resolution. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2472–2481, 2018. * [43] Ben Niu, Weilei Wen, Wenqi Ren, Xiangde Zhang, Lianping Yang, Shuzhen Wang, Kaihao Zhang, Xiaochun Cao, and Haifeng Shen. Single image super-resolution via a holistic attention network, 2020. * [44] Saeed Anwar and Nick Barnes. Densely residual laplacian super-resolution, 2019. * [45] Yusuke Matsui, Kota Ito, Yuji Aramaki, Azuma Fujimoto, Toru Ogawa, Toshihiko Yamasaki, and Kiyoharu Aizawa. Sketch-based manga retrieval using manga109 dataset. Multimedia Tools and Applications, 76(20):21811–21838, nov 2016\. * [46] Jia-Bin Huang, Abhishek Singh, and Narendra Ahuja. Single image super-resolution from transformed self-exemplars. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5197–5206, 2015. * [47] Jos Ouwerkerk. Image super-resolution survey. Image Vision Comput., 24:1039–1052, 10 2006. * [48] Zhihao Wang, Jian Chen, and Steven C. H. Hoi. Deep learning for image super-resolution: A survey. CoRR, abs/1902.06068, 2019. * [49] A. Emin Orhan. Skip connections as effective symmetry-breaking. CoRR, abs/1701.09175, 2017. * [50] Christian Ledig, Lucas Theis, Ferenc Huszar, Jose Caballero, Andrew P. Aitken, Alykhan Tejani, Johannes Totz, Zehan Wang, and Wenzhe Shi. Photo-realistic single image super-resolution using a generative adversarial network. CoRR, abs/1609.04802, 2016. * [51] Zhisheng Xiao, Karsten Kreis, and Arash Vahdat. Tackling the generative learning trilemma with denoising diffusion GANs. In International Conference on Learning Representations, 2022. * [52] George Seif and Dimitrios Androutsos. Large receptive field networks for high-scale image super-resolution. CoRR, abs/1804.08181, 2018. * [53] Tao Dai, Jianrui Cai, Yongbing Zhang, Shu-Tao Xia, and Lei Zhang. Second-order attention network for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. * [54] Zhisheng Lu, Juncheng Li, Hong Liu, Chaoyan Huang, Linlin Zhang, and Tieyong Zeng. Transformer for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, pages 457–466, June 2022. * [55] Ilya Tolstikhin, Neil Houlsby, Alexander Kolesnikov, Lucas Beyer, Xiaohua Zhai, Thomas Unterthiner, Jessica Yung, Andreas Steiner, Daniel Keysers, Jakob Uszkoreit, Mario Lucic, and Alexey Dosovitskiy. Mlp-mixer: An all-mlp architecture for vision, 2021. * [56] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale, 2020. * [57] Wenzhe Shi, Jose Caballero, Christian Ledig, Xiahai Zhuang, Wenjia Bai, Kanwal Bhatia, Antonio M. Simoes Monteiro de Marvao, Tim Dawes, Declan O’Regan, and Daniel Rueckert. Cardiac image super-resolution with global correspondence using multi-atlas patchmatch. In Kensaku Mori, Ichiro Sakuma, Yoshinobu Sato, Christian Barillot, and Nassir Navab, editors, Medical Image Computing and Computer-Assisted Intervention – MICCAI 2013, pages 9–16, Berlin, Heidelberg, 2013. Springer Berlin Heidelberg. * [58] Sheng Cao, Chao-Yuan Wu, and Philipp Krähenbühl. Lossless image compression through super-resolution, 2020.
# Trace the Accretion Geometry of H 1743–322 with Type C Quasi-periodic Oscillations in Multiple Outbursts Qing C. Shui Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China University of Chinese Academy of Sciences, Chinese Academy of Sciences, 100049, Beijing, China S. Zhang Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China Yu P. Chen Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China Shuang N. Zhang Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China University of Chinese Academy of Sciences, Chinese Academy of Sciences, 100049, Beijing, China Ling D. Kong Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China Institut für Astronomie und Astrophysik, Kepler Center for Astro and Particle Physics, Eberhard Karls, Universität, Sand 1, D-72076 Tübingen, Germany Peng J. Wang Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China University of Chinese Academy of Sciences, Chinese Academy of Sciences, 100049, Beijing, China L. Ji School of Physics and Astronomy, Sun Yat-Sen University, Zhuhai, 519082, China Hong X. Yin Shandong Key Laboratory of Optical Astronomy and Solar- Terrestrial Environment, School of Space Science and Physics, Institute of Space Sciences, Shandong University, Weihai, Shandong 264209, China Jin L. Qu Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China L. Tao Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China Ming Y. Ge Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China Jing Q. Peng Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China University of Chinese Academy of Sciences, Chinese Academy of Sciences, 100049, Beijing, China Z. Chang Key Laboratory of Particle Astrophysics, Institute of High Energy Physics, Chinese Academy of Sciences, 100049, Beijing, China J. Li CAS Key Laboratory for Research in Galaxies and Cosmology, Department of Astronomy, University of Science and Technology of China, Hefei 230026, China School of Astronomy and Space Science, University of Science and Technology of China, Hefei 230026, China P. Zhang College of Science, China Three Gorges University, Yichang 443002, China ###### Abstract We present a systematic analysis of type C quasi-periodic oscillation (QPO) observations of H 1743–322 throughout the _Rossi X-ray Timing Explorer_ (_RXTE_) era. We find that, while different outbursts have significant flux differences, they show consistent positive correlations between the QPO fractional root-mean-square (rms) amplitude and non-thermal fraction of the emission, which indicate an independence of the intrinsic QPO rms on individual outburst brightness in H 1743–322. However, the dependence of the QPO rms on frequency is different between the outburst rise and decay phases, where QPO fractional rms of the decay phase is significantly lower than that of the rise phase at low frequencies. The spectral analysis also reveals different ranges of coronal temperature between the two outburst stages. A semi-quantitative analysis shows that the Lense-Thirring precession model could be responsible for the QPO rms differences, requiring a variable coronal geometric shape. However, the variable-Comptonization model could also account for the findings. The fact that the rms differences and the hysteresis traces in the hardness-intensity diagram (HID) accompany each other indicates a connection between the two phenomena. By correlating the findings with QPO phase lags and the quasi-simultaneous radio flux previously published, we propose there could be corona-jet transitions in H 1743–322 similar to those that have been recently reported in GRS 1915+105. black hole physics – accretion, accretion disc – binaries, close – X-rays: binaries – X-rays: individual (H 1743–322) ## 1 Introduction Undergoing outbursts occasionally after staying with faint luminosity for a long time in quiescence is the primary feature of low mass black hole X-ray binaries (BHXRBs) (Remillard & McClintock, 2006; Done et al., 2007). Most complete outbursts are observed to have four typical states: the low/hard state (LHS), hard intermediate state (HIMS), soft intermediate state (SIMS) and high soft state (HSS), characterized by different X-ray spectral and variability properties (Belloni et al., 2005; Fender et al., 2009). In the outburst rise phase, the system starts from quiescence, increases luminosity with strong variability and non-thermal dominated spectra in the LHS, experiences the hard-to-soft transition in the HIMS and SIMS, and then stays in the HSS with the thermal dominated spectra and weakest variability for weeks. With the decreasing accretion rate, the system goes through the soft- to-hard transition back to the LHS, and fades in quiescence. The different state transition luminosity between the rise and decay phases directly leads to the trace of canonical ‘q’ shape in the hardness-intensity diagram (HID, Homan et al., 2001; Homan & Belloni, 2005), and the so-called hysteresis effect which is still not well understood in BHXRBs (Maccarone, 2003; Zdziarski et al., 2004; Weng et al., 2021). In addition to these features in the X-ray energy bands, BHXRBs are also characterized by radio/infrared emission, which is generally believed to be associated with relativistic jets (Fender, 2001; Fender & Belloni, 2004; Méndez et al., 2022, and references therein). In the hard state, radio emission with a flat spectrum ($S_{\nu}\propto\nu^{\alpha}$ with $\alpha\sim 0$) is interpreted as self- absorbed synchrotron emission from an optically thick, steady and compact jet (Blandford & Königl, 1979; Fender, 2001).During the hard-to-soft transition, the steady and compact jet is gradually quenched, where the radio emission, if present, is thought to be attributed to optically thin synchrotron emission from transient ejected plasma clouds with relativistic speeds ($v\sim c$) (Fender et al., 2004). The X-ray spectrum of a BHXRB usually consists of a thermal and a non-thermal component, the former is believed to come from a geometrically thin and optically thick disc (Shakura & Sunyaev, 1973; Lynden-Bell & Pringle, 1974), while the latter is produced by the Comptonization of soft photons. However, the geometry of the Comptonizing medium is relatively less clear, which could be an extended cloud consisting of hot electrons ($\sim 100$ keV) called “corona” (Bisnovatyi-Kogan & Blinnikov, 1976; Sunyaev & Titarchuk, 1980; Titarchuk, 1994; Zdziarski et al., 1996, 1999; Życki et al., 1999) or/and the jet-base (Markoff et al., 2005; You et al., 2021). A portion of Comptonized photons can irradiate the disc and then end up as the reflection component which has abundant features, like broad emission lines and Compton hump, etc. (Dauser et al., 2010; García et al., 2014, 2015b, and references therein). Low frequency quasi-periodic oscillations (LFQPOs, roughly 0.1–30 Hz, van der Klis, 1989) are the most prominent features observed in the power density spectrum (PDS) of BHXRBs, with the classification of type A, B and C based on the centroid frequency, quality factor and root-mean-square (rms) amplitude (Wijnands et al., 1999; Casella et al., 2005; Remillard & McClintock, 2006). The appearance of type C QPOs is frequent in the LHS and HIMS with strong amplitudes (fractional rms $\sim 10\%$) and flat-top noise components in the PDS. In the past few decades, several models has been proposed to explain the dynamical origin of QPOs based either on the geometric or the intrinsic properties of the accretion flow. Some examples of the _intrinsic_ models are trapped corrugation modes (Kato, 1990; Wagoner, 1999), the Accretion-ejection instability model (AEI; Tagger & Pellat, 1999) and the Two-Component Advection Flow model (TCAF; Molteni et al., 1996), etc.. For the _geometric_ models, most of them are related to the relativistic Lense-Thirring (L-T) precession, which was originally invoked to be the dynamic mechanism of QPOs by Stella & Vietri (1998). As an extension of the relativistic precession model (RPM; Stella et al., 1999), the L-T precession model proposed by Ingram et al. (2009) assumes the entire hot flow precesses within the inner radius of the truncated disc (Esin et al., 1997). We refer readers to Ingram & Motta (2019) for recent reviews of observations and theories of LFQPOs. In addition to the frequency, the radiative properties of QPOs, e.g. rms amplitudes and time lags, etc., also provide extra useful information. Since several observational studies have found that, for most type C QPOs, the variability increases with the photon energy (Zhang et al., 2017; Huang et al., 2018; Kong et al., 2020; Zhang et al., 2020) and no prominent disc-like component exists in the rms spectra (Sobolewska & Życki, 2006; Axelsson et al., 2013; Axelsson & Done, 2016), the radiative mechanism of the type C QPO should be strongly related to the Comptonized emission. The inclination dependence of amplitudes and time lags (see Motta et al., 2015; van den Eijnden et al., 2017) and reflection variability extracted from the phase- resolved spectroscopy (see Ingram & van der Klis, 2015; Ingram et al., 2016, 2017) add support to a geometrical origin, especially the L-T precession model (Ingram et al., 2009). However, these observational findings from different sources show very diverse lacking quantitative explanations, and most recently, Nathan et al. (2022) applied the L-T precession model to fit the phase-resolved spectroscopy of GRS 1915+105 and found an unexpectedly long thermalization time-scale of $\sim 70$ ms, which is incompatible with the soft lags ($\sim 1$ ms) found in other BHXRBs, e.g. MAXI J1820+070 (Kara et al., 2019) and GX 339–4 (Uttley et al., 2011). These inconsistent findings indicate L-T precession model, especially its radiative part, could be incomplete. Karpouzas et al. (2020) have recently proposed a time-dependent Comptonization model following the work of Lee & Miller (1998) and Kumar & Misra (2014) to quantitatively explain the energy-dependence of phase lags and rms amplitudes of kilohertz QPOs in neutron-star systems. This model has been successfully applied to BHXRBs, e.g. GRS 1915+105 and MAXI J1348–630 (see Karpouzas et al., 2021; García et al., 2021; Méndez et al., 2022; García et al., 2022). The updated version of the model (Bellavita et al., 2022), incorporating a disc- blackbody as the seed-photon source, has been applied to fit the rms amplitude and phase lag spectra of type C QPOs in MAXI J1535–571, using _Insight_ -HXMT data in the 1–100 keV energy range (Zhang et al., 2022). The transient BHXRB H 1743–322 was first observed by _Ariel_ -V satellite in 1977 August (Kaluzienski & Holt, 1977), located at $\rm RA=17^{h}46^{m}15^{s}.596$ and $\rm Dec=-32^{\circ}14^{\prime}00^{\prime\prime}.860$. Based on the X-ray/radio observations of the two-sided jet in the 2003 outburst, Steiner et al. (2012) determined a distance of $8.5\pm 0.8\ \rm kpc$, and an inclination angle of $75^{\circ}\pm 3^{\circ}$, respectively. By applying the relativistic accretion disc model in the spectral fitting, they also estimated a black hole spin of $a_{*}=0.2\pm 0.3$. H 1743–322 is an active black hole transient source and was monitored by _RXTE_ to undergo outbursts frequently between 2003 and 2011. The 2003 outburst is the brightest one with observations in multiple wavelengths (see Parmar et al., 2003; Homan et al., 2005; McClintock et al., 2009), then two much fainter outbursts in 2004 and 2005, respectively, followed behind (Bhattacharjee et al., 2017; Coriat et al., 2011). In 2008, two outbursts (2008a and 2008b) were observed in both of X-ray and radio bands, but the 2008a outburst is classified as failed-transition outburst for the short outburst cycle without experiencing soft states (Capitanio et al., 2009; Jonker et al., 2010; Coriat et al., 2011). Then this source entered into a new outburst during 2009 (Chen et al., 2010; Motta et al., 2010), and exhibited the last three outbursts in 2010 (2010a and 2010b) and 2011 (Zhou et al., 2013). Recently, Aneesha & Mandal (2020) systematically analysed the spectral evolution of the outbursts of H 1743–322 in the _RXTE_ era. In the post _RXTE_ era, there were 2012, 2014, 2016 and 2018 outburst monitored by other instruments (Shidatsu et al., 2014; Stiele & Yu, 2016; Chand et al., 2020; Williams et al., 2020; Wang et al., 2022). In this work, we systematically investigate type C QPO data born out of _RXTE_ observations in both of the rise and decay phases of seven outbursts from H 1743–322. We introduce observations and data reductions in Section 2, present the data analysis and results in Section 3 , discuss these in Section 4 and finally summarize in Section 5. ## 2 Observations and Data reductions Fig. 1 shows the _RXTE_ /All Sky Monitor (ASM) light-curve of H 1743–322 in the 1.5–12 keV energy band. Between 2003 and 2011, this source experienced nine outbursts, and we focus on the 2003, 2008a, 2008b, 2009, 2010a, 2010b and 2011 outburst which contain abundant type C QPO observations carried out by _RXTE_. The `HEASOFT` software package version 6.28 is used for the data analysis. We generate good time internals (GTIs) with the constraints that the elevation angle (ELV) is larger than $10^{\circ}$ and angular distance between the pointing position and source (OFFSET) is less than $0.02^{\circ}$. The standard data products of e.g. light curve, spectrum and background are produced from observational data of Proportional Counter Array (PCA). Since combining large spectral data sets with different PCA configurations can produce large systematic errors (see Smith et al., 2009), we only use data from Proportional Counter Unit (PCU) 2 in our subsequent spectral analysis because it is the only unit which was 100% on during the observations. Data of standard 2 mode are adopted for the spectral analysis in the energy range of 3–30 keV without any groupings and binnings. To account for calibration uncertainties, a systematic error of 0.5 per cent is added to spectral fittings. We generate the power density spectrum (PDS) using the light curve born out of ‘Event’ and ‘Binned’ modes in the energy range of 3–30 keV (PCA channels 7–71 of the calibration epoch 5111https://heasarc.gsfc.nasa.gov/docs/xte/e-c_table.html), which is the same as that of the spetral fittings. Then both of the spectral and timing analyses are performed using XSPEC version 12.12.0. Figure 1: The _RXTE_ /ASM light curve of H 1743–322 in the 1.5–12 keV energy band with a resolution of one point per day. In the _RXTE_ era, there are nine outbursts between 2003 and 2011, where we focus on the 2003, 2008a, 2008b, 2009, 2010a, 2010b and 2011 outburst which contain abundant type C QPO observations. ## 3 Analysis and results ### 3.1 Power Density Spectra For each observation in the present study, we produce the PDS in the 1/64–-64 Hz frequency range with the 8-ms time resolution by taking Miyamoto normalization (Belloni & Hasinger, 1990; Miyamoto et al., 1991). All the PDS are fitted with several Lorentzian functions (see Fig. 2). In the PDS fittings, we use at least two Lorentzian functions to fit the broad band noise and QPO signal, respectively. The centriod frequency of the Lorentzian fitted the low-frequency broad band noise is fixed at zero. However, if there are any other significant residual structures (e.g. the second QPO harmonic), more Lorentzian functions would be added to the fitting model. Since we primarily focus on the QPO signal, whether to add more Lorentzians depends on whether the Lorentzian functions used to fit the QPO harmonics describe the QPO components well in the case of the current total model. To do this, after fitting with the current model, we retain the best-fit parameters and remove the Lorentzian components fitted the QPO harmonics to check whether the residual structures from the data/model ratio plots are Lorentzian-like shapes and broad band noises are well fitted. Following Motta et al. (2015), if a QPO consists of multiple harmonic peaks, the QPO fractional rms is computed by adding in quadrature the rms of the harmonic peaks (i.e. QPO fractional rms is computed by $\sqrt{\sum{P_{i}}}$, where $P_{i}$ is the power calculated with integration of the $i$-th QPO harmonic Lorentzian function). Additionally, considering the background contribution, the QPO fractional rms reported in the present study is finally computed by ${\rm rms}=\sqrt{\sum_{i}{P_{i}}}\times[(S+B)/S],$ (1) where $S$ and $B$ are the source and background average rate, respectively (see Bu et al., 2015). The best-fit QPO parameters of the present study are presented in Appendix D. Figure 2: Representative power density spectra from the outburst rise (plotted in black) and decay phase (plotted in blue). Power density spectra are produced in the 3–30 keV energy range and 1/64–64 Hz frequency range, then fitted with the model consisting of multiple Lorentzian functions (dotted lines). For panel (a), strong QPO components appear at high frequencies ($f_{\rm QPO}>2$ Hz), where QPO amplitudes of the two outburst phases are comparable. For panel (b), the QPO components appear at low frequencies ($f_{\rm QPO}<2$ Hz), where QPO amplitude of the rise phase is significantly larger than that of the decay phase. ### 3.2 Energy Spectra In our spectral fittings, the Galactic absorption effect is always included in our models by implementing $tbabs$ model (Wilms et al., 2000). The hydrogen column density ($N_{\rm H}$) is fixed to $2.3\times 10^{22}\ {\rm cm^{-2}}$ following Miller et al. (2006). Firstly, we start with a simple model consists of a power-law and a disc blackbody component (Model 1: $tbabs\times(diskbb+powerlaw)$) and obtain large reduced-$\rm\chi^{2}$ ($\gg 2$) in the most observations. Then we replace $powerlaw$ with $nthcomp$, a physically motivated thermal Comptonization model which describes the spectrum of Compton upscattering photons (Zdziarski et al., 1996; Życki et al., 1999). The spectral fitting with the second model (Model 2: $tbabs\times(diskbb+nthcomp)$) gives a smaller reduced-$\rm\chi^{2}$ but, as shown in Fig. 3a and b, the residuals still show obvious structures at energies around the iron line ($\sim 6.4$ keV) and Compton hump ($\sim 10$–30 keV), providing evidence of a relativistic reflection component (García et al., 2013, and references therein). Hence we add a reflection model $relxillcp$ (Dauser et al., 2014; García et al., 2014) in our spectral fittings (Model 3: $tbabs\times(diskbb+nthcomp+relxillcp)$). For the spectral fitting with Model 3, we link the seed photon temperature ($kT_{\rm bb}$) of $nthcomp$ with the inner disk temperature ($kT_{\rm in}$) of $diskbb$ and choose the flavor of disk-blackbody seed photons for $nthcomp$. Since $relxillcp$ calculates the reflection component using the $nthcomp$ continuum, it is self-consistent to link the relevant parameters like the photon index ($\Gamma$) and electron temperature ($kT_{\rm e}$) between two models. In the reflection model $relxillcp$, we assume the canonical power-law emissivity profile, $\epsilon\propto r^{-3}$, for the disc (Fabian et al., 1989). The parameter $R_{\rm out}$ is the outer radius of the accretion disc which turns out to be not sensitive to the overall fitting and hence is frozen at the maximum value ($1000R_{\rm g}$, where $R_{\rm g}=GM/c^{2}$ is the gravitational radius). In order to make $relxillcp$ only calculate the reflection component, we fix the reflection fraction ($R_{\rm f}$) to $-1$. Referring to previous studies, the disc inclination ($i$), spin parameter of the black hole ($a_{*}$) and iron abundance ($A_{\rm Fe}$, in solar units) are set to $75^{\circ}$, 0.2 and 3.0, respectively (Steiner et al., 2012; Chand et al., 2020). The shape of $nthcomp$ continuum is set by the combination of the electron temperature ($kT_{\rm e}$) and scattering optical depth ($\tau_{\rm s}$), where the higher cut-off energy is parameterized by $kT_{\rm e}$ (Życki et al., 1999). However, in many observations, the cut-off energy is beyond the energy band of our spectral analysis (3–30 keV). Although $kT_{\rm e}$ can influence the reflection hump at energies around 20–40 keV which gives a chance to estimate it beyond the spectral coverage (see García et al., 2015a), we note that $kT_{\rm e}$ is not completely reliable in our spectral fitting. For a less constrained $kT_{\rm e}$, we fix it to 300 keV. Additionally, we notice the $diskbb$ contribution is marginal in where its parameters are not well constrained by using only PCA data and the spectral fitting does not even need this component in some observations (see also Plant et al., 2014; Aneesha et al., 2019). For these observations, we fix the parameter $kT_{\rm in}$ and normalization of $diskbb$ at 0, then let $kT_{\rm bb}$ of $nthcomp$ as a free parameter in the fitting. We calculate the unabsorbed flux for each component in the energy range 3–30 keV by convolving $cflux$ model with these required models. After fitting energy spectra with Model 3, we obtain a reasonable reduced-$\chi^{2}$ ($\sim 1$) for most observations (see Appendix A) and compute the 90 per cent confident-level uncertainties using the Markov Chain Monte Carlo (MCMC) technique, with length 40 000. The best-fit parameters of the two representative spectra presented in Fig. 3 with Model 3 are summarized in Table 1. Figure 3: Representative _RXTE_ /PCA energy spectra (the top panels) and fitting residuals (the three bottom narrow panels are plotted for Model 1, 2 and 3, respectively) for both of the rise phase (the left panel, obs.ID: 80146-01-30-00) and the decay phase (the right panel, obs.ID: 93427-01-04-00). The $diskbb$, $nthcomp$ and $relxillcp$ model are plotted in red, orange and blue, respectively. Table 1: Best-fit Parameters and the Corresponding 90% Confidence Intervals Obtained from Fitting the Representative _RXTE_ /PCA Spectra Presented in Fig. 3 Using the Model 3: $tbabs\times(diskbb+nthcomp+relxillcp)$. Model | Parameters | Rise Phase | Decay Phase ---|---|---|--- $tbabs$ | $N_{\rm H}$ ($10^{22}\rm cm^{-2}$) | 2.3 (fixed) $relxillcp$ | $i$ (Deg) | 75 (fixed) | $a_{*}$ | 0.2 (fixed) | $R_{\rm out}\ (R_{\rm g})$ | 1000 (fixed) | $q^{\rm a}$ | 3 (fixed) | $A_{\rm Fe}$ (Solar Units) | 3 (fixed) | $R_{\rm f}$ | $-1$ (fixed) $diskbb$ | $T_{\rm in}$ (keV) | $0.76^{+0.05}_{-0.05}$ | $0.62^{+0.08}_{-0.18}$ | $N_{\rm disk}^{\rm b}$ | $866^{+280}_{-201}$ | $99^{+842}_{-53}$ $nthcomp$ | $\Gamma$ | $2.29^{+0.04}_{-0.03}$ | $1.83^{+0.03}_{-0.05}$ | $kT_{\rm e}$ (keV) | $8.6^{+1.0}_{-0.5}$ | 300(fixed) | $N_{\rm nth}^{\rm c}$ | $2.52^{+0.29}_{-0.21}$ | $0.15^{+0.05}_{-0.08}$ $relxillcp$ | $R_{\rm in}\ (R_{\rm ISCO})$ | $17.9^{+76.8}_{-8.1}$ | 100 (fixed) | $\log_{10}(\xi)^{\rm d}$ | $2.81^{+0.13}_{-0.17}$ | $3.53^{+0.67}_{-0.31}$ | $(\rm erg\ cm\ s^{-1})$ | $N_{\rm rel}^{\rm e}\ (10^{-3})$ | $42.0^{+17.1}_{-9.7}$ | $2.3^{+4.50}_{-0.10}$ | $\chi^{2}/{\rm d.o.f}$ | 50.18/46 | 38.29/48 ### 3.3 Timing-spectral Joint Analysis The timing and spectral analyses presented in Section 3.1 and 3.2, respectively, provide the essential inputs for a spectral-timing joint diagnostic of the outburst what is shown in follows. #### 3.3.1 Correlations between the QPO rms and non-thermal component Fig. 4 presents the relations between the QPO fractional rms and non-thermal component during the rise and decay phases, separating the different outbursts. For clarity’s sake, data points of the rise phase from different outbursts are displayed in different colors and shapes in Fig. 4a and b, and the data points of the decay phase are plotted in gray without separating different outbursts, while in the Fig. 4c and d, we display data sets in the opposite way. As shown in Fig. 4a, in the rise phase, there are no remarkable correlations between the fractional rms and non-thermal fluxes ($F_{\rm nthcomp}$), and data points from different outbursts distribute widely in $F_{\rm nthcomp}$ which is consistent with the large differences in the outburst peak fluxes shown in Fig. 1. However, if we display the QPO fractional rms as a function of the non-thermal fraction ($F_{\rm nthcomp}/F_{\rm total}$), the positive correlations between the fractional rms and non-thermal component become significant and consistent among different outbursts (see Fig. 4b). In the decay phase, different outbursts show the similar non-thermal fluxes ($\rm\sim 10^{-9}\ ergs\ cm^{-2}\ s^{-1}$). However, compared with the rise phase, the positive correlation between the fractional rms and non-thermal fraction is not clear within the decay phase because of the relatively larger error bar of the non-thermal fraction (see Fig. 4c and d). Figure 4: The relations between QPO fractional rms and the non-thermal component. For clarity’s sake, data of the different outburst rise phases are displayed in different colors in the panels (a) and (b), and data points of the decay phases are plot in gray without separating different outbursts, while the panels (c) and (d) display the data points of the rise and decay phases in the opposite way. #### 3.3.2 QPO rms dependence on frequency Type C QPOs usually appear in the HIMS, a stage shows the significant transition in both of the timing and spectral domains (Homan & Belloni, 2005; Remillard & McClintock, 2006), where properties of the type C QPO, like central frequency and fractional rms, evolve in a large value range. Previous studies have presented that the fractional rms of type C QPOs varies with the frequency (see Motta et al., 2015; van Doesburgh & van der Klis, 2020; Zhang et al., 2020; Wang et al., 2022). Here, we analyse a number of type C QPO samples across different outbursts of H 1743–322 and present the QPO fractional rms dependence on fundamental frequency in both the rise and decay phases in Fig. 5a for comparisons. Data points displayed are not be distinguished among different outbursts, but only between the rise and decay phases. As one can see, during the rise phase, the QPO fractional rms increases slightly with frequency below 1 Hz and then remains roughly flat around 1–2 Hz, while shows a significant drop at higher frequencies. However, the presented relation of the decay phase deviates from that of the rise phase at low frequencies, where the deviation becomes larger for the lower frequency. Also, these relations between the QPO fractional rms and frequency in both of the rise and decay phases are consistent among different outbursts. The fractional rms presented in the study is computed by adding in quadrature the rms of the harmonic peaks. However, we have checked the case that taking the rms of QPO fundamental only, and find the dependence of fractional rms on frequency deviates only slightly with respect to the results presented in Fig. 5, where the two outburst stages remain different branches in the relations between QPO fractional rms and frequency clearly. Since the fractional rms is defined as the fractional variability of the flux, the fractional rms we compute in the Section 3.1 and presented in Fig. 5a can be described by ${\rm rms}=\frac{F_{\rm var}}{F_{\rm total}}=\sigma\times\frac{F_{\rm c}}{F_{\rm total}},$ (2) where $F_{\rm var}$ is the variable flux (absolute rms), $F_{\rm total}$ is the total time-averaged flux, $\sigma$ is a function for the intrinsic rms, and $F_{\rm c}$ represents the flux contributing to the QPO variability (see also Kong et al., 2020; Shui et al., 2021). If we assume that only the non- thermal component contributes to the type C QPO, while the fractional rms is diluted by the other components, i.e. the thermal and reflected components, then $F_{\rm c}$ is equal to $F_{\rm nthcomp}$. So the intrinsic rms is the variability amplitude of non-thermal emission, which can be computed by $\sigma={\rm rms}\times\frac{F_{\rm total}}{F_{\rm nthcomp}}.$ (3) The dependence of the intrinsic rms on frequency is presented in Fig. 5b, and we find it is similar to that of the fractional rms displayed in Fig. 5a. It is clear to see the two separate branches of the rise and decay phases. Figure 5: QPO fractional rms (the left panel) and intrinsic rms (the right panel) plotted as a function of frequency. The rise and decay phase are distinguished in different colors and shapes: data points from the rise phase are plotted as blue circles, while those from the decay phase are plotted as black squares. ### 3.4 Coronal Parameters and Radio Emission We have shown that the dependence of the QPO intrinsic rms (hereafter QPO rms) on frequency shows two branches: the QPO rms in the outburst rise phase is significantly larger than that in the decay phase at low frequencies. To investigate this bi-modality in more detail, we present the dependence of the rms-$f_{\rm QPO}$ relation on radio fluxes and spectral parameters in Fig. 6. The quasi-simultaneous measurements of the radio flux density at $\sim 8.5$ GHz ($S_{\nu=8.5{\rm GHz}}$) are taken from McClintock et al. (2009); Jonker et al. (2010); Coriat et al. (2011); Miller-Jones et al. (2012). For details of the radio observations used in the present study, see Appendix D. We note that not all _RXTE_ /PCA observations in this study have quasi-simultaneous radio measurements. However, for clearly showing the radio dependence of the two different branches, we plot the entire QPO data set from our timing analysis in gray in the top two panels of Fig. 6, while coloring these data points which have radio observations. Furthermore, since 2003 outburst is much brighter than the other outbursts, we present the radio data of 2003 outburst in Fig. 6a alone, while those of the other outbursts are shown in Fig. 6b. The dependence of the coronal temperature ($kT_{\rm e}$) and photon index ($\Gamma$) are presented in the bottom two panels of Fig. 6, respectively. The shade of the data points in each panel indicate the measured values of the parameters. Radio emission during the outburst rise phase of 2003 outburst is the strongest ($S_{\nu=8.5{\rm GHz}}>10\ {\rm mJy}$), while that of the other outbursts is relatively weaker ($S_{\nu=8.5{\rm GHz}}\sim 2.5\ {\rm mJy}$). For an individual outburst, there is a marginally decreasing trend of the radio flux density from low QPO frequency ($f_{\rm QPO}<2$ Hz) to high frequency ($f_{\rm QPO}\sim 8$ Hz). Additionally, radio emission of the decay phase ($S_{\nu=8.5{\rm GHz}}<1\ {\rm mJy}$) is significantly weaker than that of the rise branch. In contrast to the radio flux, the electron temperature is higher in the decay branch. These triangles plotted in Fig. 6c indicate that $kT_{\rm e}$ are too high to be constrained using _RXTE_ /PCA energy spectra, hence fixed at 300 keV (see Section 3.2 for details). However, for the photon index, there are no apparent differences between the two branches in the similar frequency range. Figure 6: The QPO intrinsic rms plotted as a function of frequency. The shade of the data points indicate the quasi-simultaneous measurements of the radio flux density (a and b), $kT_{\rm e}$ (c) and $\Gamma$ (d), respectively. In panels (a) and (b), we plot the entire QPO data set from our timing analysis in gray, only these data points with radio observations are colored. Furthermore, since the outburst 2003 is much brighter than the other outbursts, we present the radio flux density of the 2003 outburst in panel (a) alone, while those of the other outbursts are shown in panel (b). In panel (c), these triangles indicate that $kT_{\rm e}$ is too high to be constrained using _RXTE_ /PCA energy spectra, hence fixed at 300 keV. ### 3.5 Phenomenological Analysis with the L-T Precession Model The phenomenon that the QPO rms is dependent on the frequency and outburst stage (see Fig. 5) indicates that the intrinsic properties of QPOs changed during the outburst. If type C QPOs are produced by the L-T precession of the entire corona (see Ingram et al., 2009; You et al., 2018), a simplified geometrical model presented as follows can be used to estimate the intrinsic rms semi-quantitatively. Figure 7: Schematic diagram illustrating the cross-section of the corona. $\hat{\bm{n}}$, $\hat{\bm{o}}$ and $\hat{\bm{a}}$ represent the normal of the corona, unit vector pointing from the black hole to the observer and unit vector of black hole spin. respectively. $\beta$ is the angle between binary orbit and black hole spin, and $\theta$ is the included angle between $\hat{n}$ and $\hat{o}$. The coronal shape is described by the height scale $h/r$, the ratio of the semiminor and semimajor axis of the ellipse. We note this presented representative geometry is a case of $\Phi=0$. Using the coordinate system described by Veledina et al. (2013); Ingram et al. (2015), the unit vector pointing from the central black hole to the observer is given by $\hat{\bm{o}}=\left(\sin{i}\cos{\Phi},\sin{i}\sin{\Phi},\cos{i}\right),$ (4) where $i$ is the binary inclination and $\Phi$ is the azimuth of the observer. The instantaneous normal of the corona is $\hat{\bm{n}}$, changing with the precession phase angle $\omega$: $\displaystyle\hat{\bm{n}}=($ $\displaystyle\sin{\beta}\cos{\beta}\left(1+\cos{\omega}\right),$ $\displaystyle\sin{\beta}\sin{\omega},\cos^{2}{\beta}-\sin^{2}{\beta}\cos{\omega}),$ (5) where $\beta$ is the angle between binary orbit and black hole spin. We note that $\omega$ changes from 0 to $2\pi$ on the precession period, hence leads to the changes of $\hat{\bm{n}}$. Then cosine of the included angle ($\theta$) between $\hat{\bm{n}}$ and $\hat{\bm{o}}$ can be written by $\displaystyle\cos{\theta}=$ $\displaystyle\hat{\bm{n}}\cdot\hat{\bm{o}}$ $\displaystyle=$ $\displaystyle\sin{\beta}\cos{\beta}\sin{i}\cos{\Phi}\left(1+\cos{\omega}\right)$ $\displaystyle+\sin{\beta}\sin{\omega}\sin{i}\sin{\Phi}$ $\displaystyle+\cos{i}\left(\cos^{2}{\beta}-\sin^{2}{\beta}\cos{\omega}\right).$ (6) We consider a simplified problem with assuming that the coronal shape is a crushed sphere, i.e. viewed from the coronal normal ($\hat{\bm{n}}$) is a circle while the cross-section is an ellipse (see Fig. 7). The shape of the corona can be described by defining the scale height, $h/r$, where $h$ is the semiminor axis and $r$ is the semimajor axis of the ellipse, respectively. In the geometry described above, the projected area of the corona to the observer is $\displaystyle S_{\rm ob}$ $\displaystyle=\pi r^{2}\left[\sin^{2}{\theta}\cdot\left(h/r\right)^{2}+\cos^{2}{\theta}\right]^{1/2}$ $\displaystyle=\pi r^{2}\left\\{\left(h/r\right)^{2}+\left[1-(h/r)^{2}\right]\cos^{2}{\theta}\right\\}^{1/2}.$ (7) For simplicity, we assume the radiation of the corona is homogeneous and isotropic, then the observed flux from the corona is a function of $\theta$ and $h/r$: $\displaystyle F$ $\displaystyle=\bar{I}\cdot\Delta\Omega=\bar{I}\cdot S_{\rm ob}/D^{2}$ $\displaystyle\propto\left\\{\left(h/r\right)^{2}+\left[1-(h/r)^{2}\right]\cos^{2}{\theta}\right\\}^{1/2}\left(1-e^{-\tau}\right),$ (8) where $\bar{I}$ is the average radiation intensity, $\Delta\Omega$ is the solid angle of the corona in the view of the observer, $D$ is the distance from the source to the observer, and $\tau$ is the effective optical depth, also a function of $\theta$ and $h/r$. We estimate $\tau$ as a simple case: $\tau=\tau_{0}\left\\{\left(h/r\right)^{2}+\left[1-\left(h/r\right)^{2}\right]\cos^{2}{\theta}\right\\}^{-1/2},$ (9) where $\tau_{0}$ is the minimum optical depth of the corona, i.e. viewed from the coronal normal (see Appendix B for details). Combining Equations 3.5 and 9, we find that the flux from the corona changes with $\theta$ on the precession period. On the basis of the above, the minimum and maximum coronal flux could be calculated in the $\omega$ value range $0-2\pi$. Then the QPO rms can be estimated by ${\rm rms}=\frac{F_{\rm max}-F_{\rm min}}{F_{\rm max}+F_{\rm min}},$ (10) where $F_{\rm max}$ and $F_{\rm min}$ are the maximum and minimum value of the coronal flux, respectively, on the precession period. Figure 8: The calculated results using the geometrical model. In the calculation, we assume a translucent corona with $\tau_{0}=1$ which is consistent with the previous hypotheses and simulations (see You et al., 2018; Ingram & Motta, 2019; You et al., 2020). Panel (a) displays the intrinsic rms dependence on the entire range of viewing angles with a fixed $h/r$ value (0.2), panel (b) displays the dependence of intrinsic rms on $h/r$ with different inclinations ($\Phi$ is set to $90^{\circ}$). The calculated results using the simplified geometrical model are presented in Fig. 8. In our calculation, we assume a translucent corona with $\tau_{0}=1$ which is consistent with the previous hypotheses and simulations (see Ingram et al., 2015; You et al., 2018, 2020). Fig. 8a displays the intrinsic rms dependence on the entire range of viewing angles ($h/r=0.2$), where it shows that rms increases with the inclination ($i$), while gets the peak values at $\Phi=180^{\circ}$. However, $i$ and $\Phi$ are constant for a specific black hole source, so the scale height ($h/r$) is the only parameter to influence intrinsic rms during the outburst evolution. In Fig. 8b, we find that QPO rms is highly dependent on $h/r$, especially in the high inclination sources ($i\sim 60^{\circ}$–$90^{\circ}$). We note that the general relativity effects, e.g. light bending, are not considered in the simplified L-T precession model, but the results presented in Fig. 8 are consistent with previous studies which took the general relativity into account (see Veledina et al., 2013; Ingram et al., 2015; You et al., 2018). This fact indicates that GR effects, e.g. light bending, could be the second-order effects to affect the QPO rms in the L-T precession model. Figure 9: Fitting results of the relations between QPO intrinsic rms and frequency that presented in Fig. 5b with the simplified L-T precession model, where the trend lines and colored shaded regions represent the median and 3$\sigma$ confidence intervals of the fittings: the rise phase is displayed in red and decay phase is display in blue. The data points plotted in the left panel are same as that presented in Fig. 5b. Calculating with different values of $h/r$ can obtain different intrinsic rms, which allows us to fit the relations between QPO intrinsic rms and frequency presented in Fig. 5b with the model. From a phenomenological motivation, we assume $h/r$ as a quadratic function of $f_{\rm QPO}$: $h/r=k_{1}\cdot f_{\rm QPO}^{2}+k_{2}\cdot f_{\rm QPO}+k_{3},$ (11) for both of the rise and decay phases, where $f_{\rm QPO}$ is the QPO frequency, $k_{1}$, $k_{2}$ and $k_{3}$ are the model parameters. In this fitting, we fix the inclination, $i$, to $75^{\circ}$ which is consistent with our spectral analysis and the misalignment angle, $\beta$, to $10^{\circ}$ following Veledina et al. (2013) and Ingram et al. (2015). Since the azimuth of observer, $\Phi$, is constant for a specific BH source, it only affects our estimation for the value of $h/r$, not the evolution trend, we fix it to a median ($90^{\circ}$). The MCMC technique is used to carry out the parameter estimation with the uniform prior distribution. The posterior probability distributions of the model parameters are presented in Appendix C. From the fitting, we get $k_{1}=(5.4\pm 0.3)\times 10^{-3}$, $k_{2}=(-1.96\pm 0.27)\times 10^{-2}$, $k_{3}=(3.39\pm 0.05)\times 10^{-1}$ and $\chi^{2}/{\rm d.o.f}=65.90/53$ for the rise phase, while $k_{1}=(7.5\pm 2.9)\times 10^{-3}$, $k_{2}=(-5.0\pm 1.9)\times 10^{-2}$, $k_{3}=(4.70\pm 0.27)\times 10^{-1}$ and $\chi^{2}/{\rm d.o.f}=28.02/47$ for the decay phase. The fitting results are displayed in Fig. 9, where the trend lines and colored shaded regions represent the median and 3$\sigma$ confidence intervals of the fitting, and the data points shown in the left panel are the same as those presented in Fig. 5b. In Fig. 9b, we display the $h/r$ dependence on $f_{\rm QPO}$ (described by Eq. 11). The evolution trends of $h/r$ are consistent between the two phases at high frequencies ($f_{\rm QPO}>4$ Hz), which show the increase with the increasing frequency. At low frequencies, $h/r$ of both the two phases are negatively correlated to the frequency, but $h/r$ values and the gradient of the decay phase are significantly larger than those of the rise phase. ## 4 Discussion We have systematically investigated the properties of type C QPOs by analysing one hundred and six _RXTE_ /PCA observations across multiple outbursts of H 1743–322 from 2003 to 2011. Fig. 4 shows large variances of non-thermal fluxes among different outbursts, which distort the correlation between the non- thermal flux and the QPO fractional rms. However, the QPO fractional rms of different outbursts is positively correlated to the non-thermal fraction ($F_{\rm nthcomp}/F_{\rm total}$) consistently. Additionally, the co-evolution between the QPO intrinsic rms and frequency keeps similar traces among different outbursts (see Fig. 5). Since different outbursts could have very different luminosity levels, these consistent behaviours across outbursts indicate that the QPO intrinsic rms (hereafter QPO rms) is independent on the individual outburst brightness. However, the dependence of the QPO rms on frequency can be classified into two branches, where QPO rms in the outburst rise phase is significantly higher than that in the decay phase at low frequencies. Radio observations and X-ray spectral analyses reveal more differences between the two branches. ### 4.1 Trace the Coronal Geometry with the Simplified L-T Precession Model This phenomenon that the QPO rms is independent on the outburst brightness has been also reported in other sources (e.g. GX 339–4, Shui et al., 2021). If type C QPOs are produced by L-T precession of the corona, the X-ray QPO variability owes to the geometric wobble of the corona, which changes the projection area of the corona with respect to the observer to modulate the X-ray flux (Ingram et al., 2009, 2015; You et al., 2018, 2020). Calculations presented in Section 3.5 give that the intrinsic amplitude is highly dependent on the coronal shape, $h/r$, for a specific BH source (assuming a translucent corona, $\tau\sim 1$). Accordingly, the similar QPO rms amplitudes with different outburst intensities indicate the coronal shape may not depend on the individual outburst brightness. The left panel of Fig. 9 shows that, in the rise phase, the QPO rms increases slightly at low frequencies, while decreases sharply after reaching the peak value at $\sim 2$ Hz. In the dynamic part of the L-T precession model, $f_{\rm QPO}$ is negatively correlated to the outer radius, $r$, of the corona (Ingram et al., 2009), so the increasing $f_{\rm QPO}$ indicates a decreasing $r$. A possible explanation for the dependence of the QPO rms on frequency is that when the coronal outer radius ($r$) evolves, the coronal shape, i.e. $h/r$, changes synchronously. The fitting results of the rms-$f_{\rm QPO}$ relation with the simplified L-T precession model show that $h/r$ of the rise and decay phases are consistent at high frequencies ($f_{\rm QPO}>3$ Hz), while $h/r$ of the decay phase is larger than that of the rise phase in the low frequency range (see Fig. 9). The spectral analysis shows that the coronal temperature ($kT_{\rm e}$) of the decay phase is obviously higher than that of the rise phase (see Fig. 6c). If the corona is a hot accretion flow (e.g. Advection- Dominated Accretion Flow, ADAF), it is mainly supported by the gas pressure (Narayan & Yi, 1994; Yuan & Narayan, 2014; Liu & Qiao, 2022), and the height ($h$) could hence be lager with the higher coronal temperature. Accordingly, the corona in the decay phase could have relatively larger $h/r$ at a specific frequency, and then precesses with a lower variability amplitude. This is because the variability of the coronal projection area with respect to the observer is smaller with a higher $h/r$ value (see Fig. 8). ### 4.2 Qualitative Interpretation with the Time-dependent Comptonization Model Time-dependent Comptonization models can explain quantitatively the rms spectrum and the phase lag spectrum of QPOs, by requiring coupled oscillations of the physical quantities: coronal temperature, $kT_{\rm e}$, temperature of the source of seed photons, $kT_{\rm s}$ and the external heating rate, $\dot{H}_{\rm ext}$ (see Karpouzas et al., 2020; Bellavita et al., 2022). Although the QPO dynamic origin is not specified in these models, the recent proposal of Mastichiadis et al. (2022) that the QPO frequency arises from a resonance between the hot Comptonizing corona and the colder accretion disc via the coupling of the energy gains and losses in the system, could provide the dynamic part to the time-dependent Comptonization models, since the disc- corona coupling is also the mechanism suggested by these models to explain the QPO radiative properties. In the time-dependent Comptonization models, the fractional variability amplitude of QPOs (QPO fractional rms) is normalized by the variability amplitude of the external hating rate, $\delta\dot{H}_{\rm ext}$, which is a fitting parameter. The energy-averaged rms in a specific energy range is therefore dependent on both the shape and the normalization ($\delta\dot{H}_{\rm ext}$) of the rms spectrum. Karpouzas et al. (2021) and García et al. (2022) fitted the rms spectra and phase lag spectra of type C QPO observations of GRS 1915+105 and found $\delta\dot{H}_{\rm ext}$ is dependent on QPO frequency, where $\delta\dot{H}_{\rm ext}$ is positively correlated strongly to $f_{\rm QPO}$ at low frequencies ($f_{\rm QPO}<1$ Hz), and negatively correlated to $f_{\rm QPO}$ in the narrow frequency range of 1–1.8 Hz, then decreases slightly from $f_{\rm QPO}\sim 2$ Hz to $f_{\rm QPO}\sim 6$ Hz. Since the rms dependence on $f_{\rm QPO}$ of H 1743–322 in the rise phase is similar to that of GRS 1915+105, the $\delta\dot{H}_{\rm ext}$ dependence on $f_{\rm QPO}$ could also affect the rms-$f_{\rm QPO}$ relation of H 1743–322 in the rise phase. It may also work in the decay phase of H1743–322 if one considers in decay phase the positive and negative correlations at lower and higher frequencies, respectively. In Fig. 6, we show that the coronal temperature of the decay phase is higher than that of the rise phase, while there are no significant differences in photon index ($\Gamma$). In this case, as shown in Fig. 3 of Bellavita et al. (2022), the energy-averaged QPO rms in the energy range of 3–30 keV could be relatively lower in the decay phase. We propose the possible reason is that for the escaping photons in the specific energy range (3–30 keV), the higher coronal temperature indicates the less scatterings of the photons before escaping the Comptonizing medium, hence the effect of the variability amplitude amplification is weaker, if a balance between Compton cooling and external heating is at work for QPO amplification within the corona. We note that the above proposal is only qualitative, while the detailed investigation of the radiative properties of QPOs requires fitting the rms, phase lag, and time- averaged spectra simultaneously with the time-dependent Comptonization model (see Karpouzas et al., 2021; Méndez et al., 2022; García et al., 2022; Zhang et al., 2022). Figure 10: The hardness-intensity diagram (HID) of the 2018 failed-transition outburst monitored by HXMT/LE (the top panel) and correlation between QPO rms and frequency using data from both of _RXTE_ /PCA and HXMT/ME (the bottom panel). The data of _Insight_ -HXMT are taken from Wang et al. (2022). The top panel displays the HID with separating the rise and decay phases: the rise data are plotted as red hollow squares, while the decay data are plotted as green hollow triangles. The bottom panel displays the correlations with separating data sets from different instruments and outburst phases. ### 4.3 Possible Relations Between Rms Differences and Hysteresis Effect The most recent failed-transition outburst of H 1743–322, monitored by _Insight_ -HXMT in 2018, has been reported by Wang et al. (2022). Different from the outbursts presented above, this outburst remained in the LHS and never showed the significant hysteresis trace in the HID throughout the observed stage (see Fig. 10a). The _Insight_ -HXMT data of the 2018 outburst are taken from Wang et al. (2022). We plot the fractional rms dependence on $f_{\rm QPO}$ using both _Insight_ -HXMT/ME and _RXTE_ /PCA data in Fig. 10b. As one can see, in the 2018 outburst, the QPO fractional rms remains roughly constant with a value $\sim 12\%$ without significant differences between the rise and decay phases. The 2008b outburst is also classified into failed- transition outbursts, however, the system experienced the LHS-HIMS transition and a hysteresis trace in the HID during this outburst (see Coriat et al., 2011), with the rms differences between rise and decay phases which are consistent with those of other complete outbursts (see Fig. 5). On the basis of the above, the different rms-$f_{\rm QPO}$ relations between the rise and decay phases seem to be associated with the hysteresis trace in the HID, because these two phenomena accompany each other in H 1743–322. However, details about the relations between the rms differences and hysteresis effect need further investigations using more observational samples from more sources. ### 4.4 Possible Scenario of the Corona-jet Coupling in H 1743–322 Radio emission of BHXRBs is thought to be strongly related to relativistic jets (Fender, 2001), which can also produce the Comptonization of soft photons from the disc (see Band & Grindlay, 1986; Georganopoulos et al., 2002; Reig & Kylafis, 2021). The jet is believed to be coupled to the accretion flow, but the nature of this connection is still not well understood. Based on a large dataset of _RXTE_ observations, frequent radio observations and the time- dependent Comptonization model, a series of studies have revealed a possible picture of the corona-jet coupling in GRS 1915+105 (see Zhang et al., 2020; Karpouzas et al., 2021; Méndez et al., 2022; García et al., 2022). In H 1743–322, although quasi-simultaneous radio observations are not as abundant as that in GRS 1915+105, for each individual outburst, there is a marginally decreasing trend of the radio flux density from low QPO frequency ($<2$ Hz) to high QPO frequency ($\sim 8$ Hz) in the rise phase (see Fig. 6), which indicates a quenching compact jet during the hard-to-soft state transition. van den Eijnden et al. (2017) found that the QPO hard phase lag became negative at high frequencies. Following the idea of the time-dependent Comptonization model (see Karpouzas et al., 2020; Bellavita et al., 2022), the nature of the hard lag (positive lag) is Comptonization, in which hard photons could experience more scatterings than soft ones before escaping the medium. The phase lag becomes negative due to the _feedback_ mechanism: the hard Comptonized photons may return back to the disc and be re-emitted later, so the softer photons could arrive later (Lee & Miller, 1998). Based on the above, we propose that the scenario of the corona-jet coupling in GRS 1915+105 can be applied to the rise phase of H 1743–322. At low QPO frequencies where the radio emission is strong, the Comptonizing medium is dominated by the jet- like corona, where the phase lag is positive (see van den Eijnden et al., 2017). However, at high QPO frequencies where the radio emission is weak, the jet is quenched and replaced by an extended corona, which covers the inner parts of the thin disc. Since the disc has a large solid angle to receive the returning back photons, the _feedback_ effect could be strong enough to produce the observed negative phase lag (van den Eijnden et al., 2017). We refer readers to Méndez et al. (2022) for details about the physical picture of the coupling between the corona and the jet. In the outburst decay phase, the coronal temperature is relatively higher and the radio emission is weaker. If the corona is powered by the magnetic energy (Merloni & Fabian, 2001), especially as the case that the jet and the coronal power are tapped from the common magnetic energy reservoir (see Malzac et al., 2004), the higher coronal temperature in the decay phase could be associated with the relatively weaker radio emission. On the basis of the above, we propose for H 1743–322 a similar scenario as the one proposed for GRS 1915+105 (Méndez et al., 2022) which, as in that case, would account for the differences of the radio emission and coronal temperature between the two outburst stages. During the outburst decay phase, the magnetic field lines are disorganized and the magnetic energy is mainly dissipated stochastically in the corona, probably via magnetic reconnection, hence the jet is quenched and the coronal temperature is high. However, the magnetic configuration could be very different during the earlier stage of the rise phase ($f_{\rm QPO}<2$ Hz and the phase lag is positive), where the magnetic field lines are spatially coherent with a large scale poloidal component, which could channel materials out of the corona and collimate them in the direction perpendicular to the disc, then the Comptonization mainly occurs in the jet. During the hard-to-soft transition, the magnetic field lines vary from the coherent to the disorganized configuration, the jet is quenched and replaced by the extended corona. Such a different journey experienced by the accretion flow in the rise and decay phases may play a role in the difference seen in the rms-frequency relation (see Fig. 5). ## 5 Conclusions We performed systematic analyses of type C QPO observations from seven outbursts of H 1743–322 caught in the _RXTE_ era. With a number of observational samples, we confirm the independence of the type C QPO intrinsic rms on the individual outburst brightness which has been reported as well in GX 339–4. However, the dependence of QPO rms on frequency shows two branches in the outburst rise and decay phases, where the radio flux and coronal temperature are also different between the two phases. Both of the L-T precession model and the time-dependent Comptonization model can account for the rms difference, where the former needs a variable coronal geometric shape. Combining the recent _Insight_ -HXMT observations of this source during its failed-transition outburst, we suggest such the rms difference between the two outburst stages could be also related to the hysteresis effect in the HID. The co-evolution among the radio flux, coronal temperature and phase lags indicates there could be corona-jet transitions in H 1743–322 which have been recently reported in GRS 1915+105. We are grateful to the anonymous referee for constructive comments that helped us improve this paper. This research has made use of data obtained from the High Energy Astrophysics Science Archive Research Center (HEASARC), provided by NASA’s Goddard Space Flight Center, and the _Insight_ -HXMT mission, a project funded by China National Space Administration (CNSA) and the Chinese Academy of Sciences (CAS). This work is supported by the National Key R&D Program of China (2021YFA0718500) and the National Natural Science Foundation of China under grants, U1838201, U1838202, 12173103, U2038101 and U1938103. This work is partially supported by International Partnership Program of Chinese Academy of Sciences (Grant No.113111KYSB20190020). ## Appendix A The distribution of reduced chi-square in the Spectral fitting In this section, we show the reduced chi-square ($\chi^{2}_{\rm red}$) distribution of the spectral fitting with the Model 3 presented in Section 3.2 (see Fig. 11). There are one hundred and six observational samples in our spectral analysis. Figure 11: The reduced chi-square ($\chi^{2}_{\rm red}$) distribution of the spectral fitting with the Model 3 presented in Section 3.2. ## Appendix B Calculation details of the coronal optical depth We consider a simplified coronal geometry with assuming the coronal shape is a crushed sphere (see Section 3.5). As a simplified model, the general relativity effects and Comptonization processes are neglected in our calculations. The effective optical depth of the corona is $\tau=\alpha L,$ (B1) where $\alpha$ is the absorption coefficient, $L$ is the average size of the corona in the view of the observer. We can define the three-dimensional Cartesian coordinates $xyz$, where the $z$-axis is consistent with the normal of the corona ($\hat{\bm{n}}$), and the unit vector of the observer, $\hat{\bm{o}}$, is in the $yz$ plane. Then the ellipsoidal equation of the coronal surface is written by $\frac{x^{2}+y^{2}}{r^{2}}+\frac{z^{2}}{h^{2}}=1,$ (B2) where $r$ is the coronal radius and $h$ is the coronal height. We consider that the distance between the source and observer is far larger than the coronal radius (i.e. $D\gg r$), so the photon trajectories from the corona to the observer can be described as a cluster of parallel lines which are perpendicular to the $x$-axis. Then the linear equation of a representative photon trajectory to the observer is $\displaystyle z$ $\displaystyle=\cot{\theta}\cdot y+z_{0}$ $\displaystyle x$ $\displaystyle=x_{0},$ (B3) where $\theta$ is the include angle between $\hat{\bm{n}}$ and $\hat{\bm{o}}$, $z_{0}$ is the $z$-intercept of the projection of the line in the $yz$ plane, and $x_{0}$ is the $x$-intercept of the projection of the line in the $xy$ plane, respectively. Then the length of the corona which contributes to the flux in the direction of the representative photon trajectory is $l=\sqrt{1+\cot^{2}{\theta}}\frac{2hr\sqrt{\left(1-\frac{x_{0}^{2}}{r^{2}}\right)h^{2}+\left(1-\frac{x_{0}^{2}}{r^{2}}\right)r^{2}\cot^{2}{\theta}-z_{0}^{2}}}{h^{2}+r^{2}\cot^{2}{\theta}},$ (B4) i.e. the length of the line cut by the ellipsoidal surface. The average size of the corona in the view of the observer can be calculated by $\displaystyle L$ $\displaystyle=\frac{\int_{-r}^{r}\int_{-s}^{s}l(x_{0},z_{0}){\rm d}z_{0}{\rm d}x_{0}}{\int_{-r}^{r}\int_{-s}^{s}{\rm d}z_{0}{\rm d}x_{0}}$ $\displaystyle=\frac{4}{3}h\left\\{\left(h/r\right)^{2}+\left[1-\left(h/r\right)^{2}\right]\cos^{2}{\theta}\right\\}^{-1/2},$ (B5) where $s\equiv\sqrt{\left[1-(x_{0}/r)^{2}\right]h^{2}+\left[1-(x_{0}/r)^{2}\right]r^{2}\cot^{2}{\theta}}$. When $z_{0}=\pm s$, the line of Eq. B is tangent to the ellipsoidal surface. Based on Eq. B1 and B, the optical depth can be written by $\tau=\frac{4}{3}\alpha h\left\\{\left(h/r\right)^{2}+\left[1-\left(h/r\right)^{2}\right]\cos^{2}{\theta}\right\\}^{-1/2}.$ (B6) We define $\tau_{0}\equiv\frac{4}{3}\alpha h$, where $\tau_{0}$ is the minimum optical depth of the corona, i.e. viewed from the coronal normal ($\theta=0$), then the optical depth can be written by $\tau=\tau_{0}\left\\{\left(h/r\right)^{2}+\left[1-\left(h/r\right)^{2}\right]\cos^{2}{\theta}\right\\}^{-1/2}.$ (B7) ## Appendix C MCMC Parameter Probability Distributions In this section, we show the contour maps and probability distributions for the set of model parameters derived using the MCMC analysis of the relation between the QPO intrinsic rms and frequency. The MCMC analysis is preformed using `emcee` package (Foreman-Mackey et al., 2013), and the contour maps and probability distributions are plotted using `corner` package (Foreman-Mackey, 2016). For each map, we show the 0.16, 0.5, and 0.84 quantiles (see Fig. 12). Figure 12: One- and two-dimensional projections of the posterior probability distributions, and the 0.16, 0.5 and 0.84 quantile contours derived from the MCMC analysis for the model parameters $k_{1}$, $k_{2}$ and $k_{3}$ described in Eq. 11. The left panel is plotted for the rise phase and the right panel is plotted for the decay phase. ## Appendix D QPO parameters and radio observations In this study, we preform a systematic analysis of type C QPOs with one hundred and six _RXTE_ /PCA observational samples of black hole X-ray binary H 1743–322. This source also exhibit significant radio emission, so we take the quasi-simultaneous radio observational results from previously published studies to add to our joint analysis. In this section, we present the QPO parameters of our analysis and the quasi-simultaneous radio flux measurements (if present) at $\sim 8.5$ GHz. The parameters, QPO frequency ($f_{\rm QPO}$), full width at half maximum (FWHM), fractional rms amplitude and radio flux density at $\sim 8.5$ GHz ($S_{\nu=8.5{\rm GHz}}$), etc., are presented in Table 2. Table 2: QPO Parameters and the Quasi-simultaneous Radio Flux Density ObsIDa | Outburst Phaseb | X-ray MJDc | $f_{\rm QPO}$ | FWHMd | Fractional rmse | Radio MJDf | $S_{\nu=8.5{\rm GHz}}^{\rm g}$ | Notes ---|---|---|---|---|---|---|---|--- | | | (Hz) | (Hz) | (%) | | (mJy) | 80138-01-06-00 | Rise | 52739.66 | $3.22_{-0.01}^{+0.01}$ | $0.35_{-0.02}^{+0.02}$ | $12.66_{-0.21}^{+0.22}$ | 52739.46 | $20.68\pm 0.06$ | VLAh 80138-01-07-00 | Rise | 52741.83 | $7.17_{-0.03}^{+0.03}$ | $1.13_{-0.07}^{+0.07}$ | $5.49_{-0.13}^{+0.13}$ | 52741.56 | $7.71\pm 0.12$ | VLAh 80146-01-01-00 | Rise | 52743.22 | $8.51_{-0.02}^{+0.02}$ | $0.56_{-0.04}^{+0.04}$ | $3.71_{-0.09}^{+0.09}$ | 52742.52 | $3.87\pm 0.13$ | VLAh 80146-01-02-00 | Rise | 52744.20 | $5.62_{-0.02}^{+0.02}$ | $0.85_{-0.04}^{+0.04}$ | $8.11_{-0.13}^{+0.13}$ | 52745.43 | $37.15\pm 0.13$ | VLAh 80146-01-03-00 | Rise | 52746.18 | $4.74_{-0.01}^{+0.01}$ | $0.55_{-0.03}^{+0.03}$ | $10.52_{-0.21}^{+0.22}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-03-01 | Rise | 52747.61 | $7.01_{-0.03}^{+0.03}$ | $1.19_{-0.08}^{+0.08}$ | $5.52_{-0.14}^{+0.14}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-29-00 | Rise | 52766.56 | $5.60_{-0.02}^{+0.02}$ | $0.74_{-0.04}^{+0.05}$ | $7.87_{-0.15}^{+0.15}$ | 52765.42 | $11.12\pm 0.13$ | VLAh 80146-01-30-00 | Rise | 52767.81 | $4.43_{-0.01}^{+0.01}$ | $0.53_{-0.02}^{+0.02}$ | $10.92_{-0.13}^{+0.13}$ | 52767.51 | $23.02\pm 0.12$ | VLAh 80146-01-31-00 | Rise | 52768.53 | $5.41_{-0.02}^{+0.02}$ | $0.83_{-0.05}^{+0.05}$ | $9.41_{-0.18}^{+0.19}$ | 52768.49 | $30.14\pm 0.16$ | VLAh 80146-01-32-00 | Rise | 52769.72 | $4.90_{-0.01}^{+0.01}$ | $0.55_{-0.03}^{+0.03}$ | $10.74_{-0.21}^{+0.23}$ | 52769.51 | $23.80\pm 0.14$ | VLAh 80146-01-33-01 | Rise | 52770.37 | $6.02_{-0.03}^{+0.03}$ | $0.85_{-0.06}^{+0.07}$ | $7.66_{-0.20}^{+0.20}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-33-00 | Rise | 52770.65 | $6.22_{-0.02}^{+0.02}$ | $0.85_{-0.05}^{+0.06}$ | $7.16_{-0.15}^{+0.15}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-34-00 | Rise | 52771.74 | $2.82_{-0.01}^{+0.01}$ | $0.33_{-0.02}^{+0.02}$ | $13.22_{-0.17}^{+0.19}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-35-00 | Rise | 52771.97 | $2.27_{-0.01}^{+0.01}$ | $0.19_{-0.03}^{+0.03}$ | $12.06_{-0.52}^{+0.52}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-36-00 | Rise | 52772.67 | $1.84_{-0.01}^{+0.01}$ | $0.25_{-0.01}^{+0.01}$ | $13.92_{-0.26}^{+0.26}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-37-00 | Rise | 52773.66 | $1.90_{-0.01}^{+0.01}$ | $0.26_{-0.02}^{+0.02}$ | $13.71_{-0.25}^{+0.25}$ | 52773.36 | $35.76\pm 0.23$ | VLAh 80146-01-39-00 | Rise | 52775.57 | $2.15_{-0.01}^{+0.01}$ | $0.29_{-0.02}^{+0.02}$ | $14.11_{-0.24}^{+0.24}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-40-00 | Rise | 52776.62 | $1.72_{-0.01}^{+0.01}$ | $0.32_{-0.02}^{+0.02}$ | $16.21_{-0.27}^{+0.27}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-41-00 | Rise | 52777.61 | $2.50_{-0.01}^{+0.01}$ | $0.31_{-0.02}^{+0.02}$ | $13.83_{-0.23}^{+0.22}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-42-00 | Rise | 52778.46 | $3.21_{-0.01}^{+0.01}$ | $0.46_{-0.02}^{+0.02}$ | $13.83_{-0.18}^{+0.19}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-43-01 | Rise | 52779.53 | $3.82_{-0.02}^{+0.02}$ | $0.41_{-0.03}^{+0.04}$ | $11.87_{-0.33}^{+0.34}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-43-00 | Rise | 52779.58 | $3.81_{-0.01}^{+0.01}$ | $0.37_{-0.02}^{+0.02}$ | $12.79_{-0.21}^{+0.22}$ | 52779.44 | $11.99\pm 0.17$ | VLAh 80146-01-44-00 | Rise | 52780.57 | $3.82_{-0.01}^{+0.01}$ | $0.37_{-0.02}^{+0.02}$ | $12.59_{-0.21}^{+0.22}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-45-00 | Rise | 52781.55 | $3.64_{-0.01}^{+0.01}$ | $0.45_{-0.02}^{+0.02}$ | $13.31_{-0.17}^{+0.18}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-46-00 | Rise | 52782.67 | $4.70_{-0.01}^{+0.01}$ | $0.45_{-0.03}^{+0.03}$ | $11.18_{-0.19}^{+0.21}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-47-00 | Rise | 52783.46 | $6.99_{-0.02}^{+0.02}$ | $1.40_{-0.05}^{+0.06}$ | $6.13_{-0.10}^{+0.10}$ | $\cdots$ | $\cdots$ | $\cdots$ 80146-01-50-00 | Rise | 52786.29 | $9.43_{-0.04}^{+0.04}$ | $0.85_{-0.08}^{+0.09}$ | $2.09_{-0.08}^{+0.08}$ | 52786.36 | $16.06\pm 0.11$ | VLAh 80137-01-25-00 | Decay | 52937.02 | $7.11_{-0.15}^{+0.21}$ | $0.91_{-0.41}^{+1.51}$ | $6.89_{-1.17}^{+1.18}$ | 52939.99 | $0.14\pm 0.04$ | VLAh 80137-01-26-00 | Decay | 52938.00 | $5.93_{-0.39}^{+0.44}$ | $1.89_{-0.77}^{+1.32}$ | $9.05_{-1.66}^{+1.93}$ | $\cdots$ | $\cdots$ | $\cdots$ 80137-02-01-00 | Decay | 52944.11 | $2.48_{-0.11}^{+0.15}$ | $0.97_{-0.38}^{+0.68}$ | $10.03_{-1.75}^{+2.11}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-03-01 | Decay | 54492.18 | $7.03_{-0.10}^{+0.10}$ | $0.67_{-0.26}^{+0.41}$ | $5.81_{-0.87}^{+1.19}$ | 54493.32 | $0.44\pm 0.09$ | ATCAi 93427-01-04-00 | Decay | 54498.84 | $3.90_{-0.08}^{+0.08}$ | $0.65_{-0.22}^{+0.33}$ | $8.00_{-1.10}^{+1.71}$ | 54499.74 | $0.52\pm 0.06$ | VLAj 93427-01-04-02 | Decay | 54500.80 | $2.60_{-0.10}^{+0.13}$ | $1.50_{-0.52}^{+0.82}$ | $13.42_{-1.77}^{+1.98}$ | 54501.64 | $0.48\pm 0.08$ | VLAj 93427-01-04-03 | Decay | 54502.83 | $2.36_{-0.16}^{+0.14}$ | $0.77_{-0.35}^{+0.39}$ | $8.86_{-1.44}^{+1.74}$ | 54502.56 | $0.45\pm 0.09$ | VLAj 93427-01-09-00 | Rise | 54742.98 | $0.33_{-0.01}^{+0.01}$ | $0.03_{-0.01}^{+0.01}$ | $15.19_{-1.67}^{+1.68}$ | 54744.21 | $1.74\pm 0.07$ | ATCAi 93427-01-09-01 | Rise | 54746.51 | $0.38_{-0.01}^{+0.01}$ | $0.03_{-0.01}^{+0.01}$ | $15.69_{-0.89}^{+0.91}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-09-03 | Rise | 54747.49 | $0.43_{-0.01}^{+0.01}$ | $0.03_{-0.01}^{+0.01}$ | $15.00_{-0.89}^{+0.91}$ | 54747.44 | $2.54\pm 0.08$ | ATCAi 93427-01-09-02 | Rise | 54748.21 | $0.47_{-0.01}^{+0.01}$ | $0.04_{-0.01}^{+0.01}$ | $14.31_{-0.98}^{+0.99}$ | 54748.44 | $2.43\pm 0.09$ | ATCAi 93427-01-10-00 | Rise | 54750.37 | $0.58_{-0.01}^{+0.01}$ | $0.06_{-0.02}^{+0.02}$ | $14.19_{-1.31}^{+1.31}$ | 54749.36 | $2.38\pm 0.11$ | ATCAi 93427-01-10-01 | Rise | 54752.26 | $0.68_{-0.01}^{+0.01}$ | $0.07_{-0.01}^{+0.01}$ | $15.59_{-0.91}^{+0.92}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-10-02 | Rise | 54755.23 | $0.75_{-0.02}^{+0.02}$ | $0.12_{-0.02}^{+0.03}$ | $13.88_{-1.24}^{+1.22}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-11-00 | Rise | 54756.25 | $0.86_{-0.01}^{+0.01}$ | $0.10_{-0.01}^{+0.01}$ | $15.44_{-0.67}^{+0.68}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-11-01 | Rise | 54758.15 | $1.02_{-0.01}^{+0.01}$ | $0.12_{-0.01}^{+0.02}$ | $14.89_{-0.62}^{+0.62}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-11-03 | Rise | 54762.14 | $5.60_{-0.06}^{+0.06}$ | $0.49_{-0.17}^{+0.23}$ | $5.43_{-0.66}^{+0.75}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-12-04 | Decay | 54767.84 | $3.87_{-0.03}^{+0.03}$ | $0.47_{-0.08}^{+0.09}$ | $10.48_{-0.62}^{+0.65}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-12-02 | Decay | 54768.64 | $3.30_{-0.03}^{+0.03}$ | $0.35_{-0.07}^{+0.09}$ | $11.45_{-0.81}^{+0.84}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-12-05 | Decay | 54769.14 | $2.91_{-0.04}^{+0.04}$ | $0.38_{-0.08}^{+0.10}$ | $12.88_{-1.05}^{+1.10}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-13-00 | Decay | 54770.12 | $2.61_{-0.02}^{+0.02}$ | $0.42_{-0.06}^{+0.06}$ | $12.40_{-0.61}^{+0.63}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-13-05 | Decay | 54770.40 | $2.26_{-0.02}^{+0.02}$ | $0.25_{-0.04}^{+0.05}$ | $11.90_{-0.68}^{+0.70}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-13-04 | Decay | 54771.76 | $2.38_{-0.02}^{+0.02}$ | $0.29_{-0.05}^{+0.05}$ | $11.89_{-0.68}^{+0.70}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-13-01 | Decay | 54772.15 | $2.25_{-0.03}^{+0.04}$ | $0.36_{-0.12}^{+0.15}$ | $12.44_{-1.27}^{+1.37}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-13-02 | Decay | 54773.27 | $2.27_{-0.03}^{+0.03}$ | $0.33_{-0.07}^{+0.08}$ | $12.21_{-0.86}^{+0.90}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-13-06 | Decay | 54774.64 | $2.12_{-0.02}^{+0.02}$ | $0.38_{-0.06}^{+0.07}$ | $12.26_{-0.75}^{+0.79}$ | 54774.43 | $0.94\pm 0.12$ | ATCAi 93427-01-13-03 | Decay | 54775.56 | $1.83_{-0.03}^{+0.03}$ | $0.32_{-0.06}^{+0.07}$ | $11.92_{-0.92}^{+0.96}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-14-00 | Decay | 54777.86 | $1.89_{-0.04}^{+0.04}$ | $0.29_{-0.07}^{+0.08}$ | $10.59_{-1.04}^{+1.10}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-14-01 | Decay | 54778.77 | $1.82_{-0.03}^{+0.04}$ | $0.29_{-0.09}^{+0.12}$ | $10.99_{-1.21}^{+1.32}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-14-02 | Decay | 54779.04 | $1.52_{-0.02}^{+0.02}$ | $0.21_{-0.05}^{+0.06}$ | $11.68_{-1.09}^{+1.14}$ | 54779.35 | $0.94\pm 0.08$ | ATCAi 93427-01-14-03 | Decay | 54780.02 | $1.48_{-0.03}^{+0.03}$ | $0.20_{-0.06}^{+0.07}$ | $9.89_{-1.16}^{+1.22}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-14-04 | Decay | 54781.78 | $1.72_{-0.02}^{+0.02}$ | $0.17_{-0.05}^{+0.07}$ | $9.88_{-1.01}^{+1.06}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-14-05 | Decay | 54782.89 | $2.21_{-0.04}^{+0.04}$ | $0.32_{-0.08}^{+0.10}$ | $11.15_{-1.11}^{+1.16}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-14-06 | Decay | 54783.81 | $2.07_{-0.05}^{+0.06}$ | $0.42_{-0.14}^{+0.20}$ | $11.16_{-1.44}^{+1.73}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-15-00 | Decay | 54784.45 | $1.79_{-0.04}^{+0.05}$ | $0.26_{-0.09}^{+0.12}$ | $9.74_{-1.22}^{+1.36}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-15-01 | Decay | 54785.70 | $1.95_{-0.08}^{+0.09}$ | $0.52_{-0.19}^{+0.27}$ | $10.83_{-1.74}^{+2.16}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-15-02 | Decay | 54786.48 | $1.76_{-0.03}^{+0.04}$ | $0.28_{-0.08}^{+0.11}$ | $9.31_{-1.07}^{+1.18}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-15-03 | Decay | 54787.73 | $1.54_{-0.06}^{+0.06}$ | $0.27_{-0.11}^{+0.16}$ | $9.15_{-1.57}^{+1.77}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-15-04 | Decay | 54788.64 | $0.97_{-0.03}^{+0.04}$ | $0.18_{-0.07}^{+0.14}$ | $9.29_{-1.67}^{+2.06}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-15-06 | Decay | 54788.84 | $0.99_{-0.07}^{+0.08}$ | $0.23_{-0.14}^{+0.19}$ | $8.09_{-2.08}^{+2.52}$ | $\cdots$ | $\cdots$ | $\cdots$ 93427-01-15-05 | Decay | 54789.49 | $0.88_{-0.05}^{+0.05}$ | $0.22_{-0.11}^{+0.24}$ | $8.20_{-1.82}^{+2.68}$ | $\cdots$ | $\cdots$ | $\cdots$ 94413-01-02-00 | Rise | 54980.40 | $0.91_{-0.01}^{+0.01}$ | $0.04_{-0.01}^{+0.01}$ | $17.96_{-0.96}^{+0.96}$ | 54978.38 | $2.24\pm 0.03$ | VLAk 94413-01-02-02 | Rise | 54980.85 | $1.00_{-0.01}^{+0.01}$ | $0.03_{-0.01}^{+0.01}$ | $17.47_{-0.75}^{+0.75}$ | 54978.38 | $2.24\pm 0.03$ | VLAk 94413-01-02-01 | Rise | 54981.95 | $1.19_{-0.01}^{+0.01}$ | $0.03_{-0.01}^{+0.01}$ | $17.16_{-1.07}^{+1.08}$ | 54981.95 | $2.73\pm 0.10$ | VLAk 94413-01-02-05 | Rise | 54982.28 | $1.28_{-0.01}^{+0.01}$ | $0.04_{-0.01}^{+0.01}$ | $17.47_{-0.81}^{+0.82}$ | $\cdots$ | $\cdots$ | $\cdots$ 94413-01-02-04 | Rise | 54983.33 | $2.02_{-0.01}^{+0.01}$ | $0.09_{-0.01}^{+0.01}$ | $17.43_{-0.66}^{+0.67}$ | $\cdots$ | $\cdots$ | $\cdots$ 94413-01-02-03 | Rise | 54984.37 | $3.58_{-0.01}^{+0.01}$ | $0.33_{-0.04}^{+0.04}$ | $14.97_{-0.69}^{+1.21}$ | 54984.35 | $1.8\pm 0.3$ | VLBAl 94413-01-07-00 | Decay | 55016.32 | $4.94_{-0.12}^{+0.06}$ | $0.26_{-0.26}^{+0.45}$ | $6.22_{-1.46}^{+2.10}$ | $\cdots$ | $\cdots$ | $\cdots$ 94413-01-07-01 | Decay | 55019.45 | $3.40_{-0.10}^{+0.09}$ | $0.39_{-0.18}^{+0.29}$ | $7.49_{-1.54}^{+2.04}$ | 55019.46 | $0.592\pm 0.055$ | VLAk 94413-01-07-02 | Decay | 55021.42 | $3.81_{-0.11}^{+0.17}$ | $0.62_{-0.29}^{+0.85}$ | $10.16_{-2.10}^{+3.87}$ | 55021.42 | $0.410\pm 0.074$ | VLAk 95405-01-02-06 | Decay | 55223.39 | $3.82_{-0.04}^{+0.04}$ | $0.43_{-0.10}^{+0.13}$ | $9.55_{-0.84}^{+0.91}$ | $\cdots$ | $\cdots$ | $\cdots$ 95405-01-02-02 | Decay | 55224.67 | $3.05_{-0.04}^{+0.04}$ | $0.39_{-0.09}^{+0.12}$ | $10.54_{-0.95}^{+1.04}$ | $\cdots$ | $\cdots$ | $\cdots$ 95405-01-03-00 | Decay | 55226.53 | $2.02_{-0.03}^{+0.04}$ | $0.30_{-0.08}^{+0.10}$ | $11.23_{-1.11}^{+1.19}$ | $\cdots$ | $\cdots$ | $\cdots$ 95405-01-03-04 | Decay | 55227.77 | $2.24_{-0.04}^{+0.03}$ | $0.30_{-0.07}^{+0.09}$ | $11.36_{-1.07}^{+1.12}$ | $\cdots$ | $\cdots$ | $\cdots$ 95405-01-03-01 | Decay | 55228.61 | $2.20_{-0.03}^{+0.03}$ | $0.26_{-0.06}^{+0.07}$ | $10.21_{-0.86}^{+0.89}$ | $\cdots$ | $\cdots$ | $\cdots$ 95405-01-03-05 | Decay | 55229.54 | $1.71_{-0.04}^{+0.04}$ | $0.24_{-0.07}^{+0.09}$ | $10.10_{-1.22}^{+1.31}$ | $\cdots$ | $\cdots$ | $\cdots$ 95405-01-03-02 | Decay | 55230.59 | $1.35_{-0.04}^{+0.04}$ | $0.32_{-0.10}^{+0.14}$ | $10.75_{-1.40}^{+1.62}$ | $\cdots$ | $\cdots$ | $\cdots$ 95405-01-04-01 | Decay | 55233.40 | $0.90_{-0.03}^{+0.03}$ | $0.15_{-0.09}^{+0.17}$ | $8.15_{-1.69}^{+2.44}$ | $\cdots$ | $\cdots$ | $\cdots$ 95360-14-01-00 | Rise | 55418.41 | $1.00_{-0.01}^{+0.01}$ | $0.05_{-0.01}^{+0.01}$ | $16.56_{-0.60}^{+0.60}$ | $\cdots$ | $\cdots$ | $\cdots$ 95360-14-02-01 | Rise | 55419.09 | $1.04_{-0.01}^{+0.01}$ | $0.04_{-0.01}^{+0.01}$ | $17.38_{-1.45}^{+1.46}$ | $\cdots$ | $\cdots$ | $\cdots$ 95360-14-02-00 | Rise | 55420.23 | $1.17_{-0.01}^{+0.01}$ | $0.04_{-0.01}^{+0.01}$ | $16.66_{-0.61}^{+0.61}$ | $\cdots$ | $\cdots$ | $\cdots$ 95360-14-03-00 | Rise | 55421.28 | $1.48_{-0.01}^{+0.01}$ | $0.05_{-0.01}^{+0.01}$ | $17.02_{-0.64}^{+0.64}$ | $\cdots$ | $\cdots$ | $\cdots$ 95360-14-02-03 | Rise | 55422.02 | $1.76_{-0.01}^{+0.01}$ | $0.08_{-0.01}^{+0.01}$ | $17.30_{-0.89}^{+0.89}$ | $\cdots$ | $\cdots$ | $\cdots$ 95360-14-02-02 | Rise | 55423.20 | $2.96_{-0.01}^{+0.02}$ | $0.20_{-0.03}^{+0.04}$ | $14.88_{-0.80}^{+0.87}$ | $\cdots$ | $\cdots$ | $\cdots$ 95360-14-03-01 | Rise | 55424.06 | $4.80_{-0.04}^{+0.04}$ | $0.37_{-0.07}^{+0.09}$ | $9.44_{-0.70}^{+0.71}$ | $\cdots$ | $\cdots$ | $\cdots$ 95360-14-23-00 | Decay | 55456.67 | $3.34_{-0.07}^{+0.08}$ | $0.58_{-0.22}^{+0.38}$ | $11.00_{-1.77}^{+2.30}$ | $\cdots$ | $\cdots$ | $\cdots$ 95360-14-23-01 | Decay | 55457.12 | $2.57_{-0.04}^{+0.05}$ | $0.25_{-0.11}^{+0.18}$ | $9.34_{-1.47}^{+1.77}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-01-00 | Rise | 55663.67 | $0.43_{-0.01}^{+0.01}$ | $0.03_{-0.01}^{+0.01}$ | $14.61_{-0.87}^{+0.88}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-02-00 | Rise | 55667.59 | $0.67_{-0.01}^{+0.01}$ | $0.05_{-0.01}^{+0.01}$ | $15.72_{-0.53}^{+0.54}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-02-01 | Rise | 55668.98 | $0.89_{-0.01}^{+0.01}$ | $0.08_{-0.01}^{+0.02}$ | $16.05_{-0.96}^{+0.99}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-02-02 | Rise | 55670.62 | $1.36_{-0.01}^{+0.01}$ | $0.14_{-0.01}^{+0.01}$ | $15.80_{-0.52}^{+0.53}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-02-05 | Rise | 55671.53 | $1.82_{-0.01}^{+0.01}$ | $0.21_{-0.03}^{+0.03}$ | $15.61_{-0.67}^{+0.67}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-02-03 | Rise | 55672.84 | $3.61_{-0.03}^{+0.03}$ | $0.29_{-0.06}^{+0.07}$ | $12.21_{-0.86}^{+0.87}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-05-01 | Decay | 55690.13 | $2.98_{-0.07}^{+0.07}$ | $0.54_{-0.17}^{+0.27}$ | $10.99_{-1.35}^{+1.49}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-05-02 | Decay | 55691.51 | $2.01_{-0.03}^{+0.03}$ | $0.24_{-0.11}^{+0.16}$ | $11.00_{-1.31}^{+1.51}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-05-03 | Decay | 55693.00 | $1.81_{-0.04}^{+0.04}$ | $0.28_{-0.10}^{+0.14}$ | $10.88_{-1.34}^{+1.47}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-06-00 | Decay | 55694.00 | $1.47_{-0.03}^{+0.04}$ | $0.24_{-0.11}^{+0.16}$ | $9.48_{-1.39}^{+1.59}$ | $\cdots$ | $\cdots$ | $\cdots$ 96425-01-06-01 | Decay | 55695.42 | $1.20_{-0.04}^{+0.04}$ | $0.19_{-0.08}^{+0.14}$ | $8.43_{-1.47}^{+1.64}$ | $\cdots$ | $\cdots$ | $\cdots$ Note. — a The power-law index of the emissivity profile ($\epsilon\propto r^{-q}$). b Normalization of $diskbb$ model. c Normalization of $nthcomp$ model. d Log of the ionization parameter ($\xi$) of the accretion disc, where $\xi=L/nR^{2}$, with $L$ as the ionizing luminosity, $n$ as the gas density, and $R$ as the distance to the ionizing source. e Normalization of $relxillcp$ model. Note. — a Observational ID of _RXTE_ data. b Outburst phase (rise phase or decay phase). c Modified Julian Day of the _RXTE_ observation. d Full width at half maximum of the QPO. e The QPO fractional rms calculated by Eq. 1. f Modified Julian Day of the radio observation. g The radio flux density at $\sim$8.5 GHz. h The radio flux measurements of Very Large Array (VLA) are taken from McClintock et al. (2009). i The radio flux measurements of Australia Telescope Compact Array (ATCA) are taken from Coriat et al. (2011). j The radio flux measurements of Very Large Array (VLA) are taken from Jonker et al. (2010). k The radio flux measurements of Very Large Array (VLA) are taken from Coriat et al. (2011). l The radio flux measurements of Very Long Baseline Array (VLBA) are taken from Miller-Jones et al. (2012). ## References * Aneesha & Mandal (2020) Aneesha, U., & Mandal, S. 2020, A&A, 637, A47, doi: 10.1051/0004-6361/202037577 * Aneesha et al. (2019) Aneesha, U., Mandal, S., & Sreehari, H. 2019, MNRAS, 486, 2705, doi: 10.1093/mnras/stz1000 * Axelsson & Done (2016) Axelsson, M., & Done, C. 2016, MNRAS, 458, 1778, doi: 10.1093/mnras/stw464 * Axelsson et al. (2013) Axelsson, M., Hjalmarsdotter, L., & Done, C. 2013, MNRAS, 431, 1987, doi: 10.1093/mnras/stt315 * Band & Grindlay (1986) Band, D. L., & Grindlay, J. E. 1986, ApJ, 311, 595, doi: 10.1086/164799 * Bellavita et al. (2022) Bellavita, C., García, F., Méndez, M., & Karpouzas, K. 2022, MNRAS, 515, 2099, doi: 10.1093/mnras/stac1922 * Belloni & Hasinger (1990) Belloni, T., & Hasinger, G. 1990, A&A, 230, 103 * Belloni et al. (2005) Belloni, T., Homan, J., Casella, P., et al. 2005, A&A, 440, 207, doi: 10.1051/0004-6361:20042457 * Bhattacharjee et al. (2017) Bhattacharjee, A., Banerjee, I., Banerjee, A., Debnath, D., & Chakrabarti, S. K. 2017, MNRAS, 466, 1372, doi: 10.1093/mnras/stw3117 * Bisnovatyi-Kogan & Blinnikov (1976) Bisnovatyi-Kogan, G. S., & Blinnikov, S. I. 1976, Soviet Astronomy Letters, 2, 191. https://arxiv.org/abs/astro-ph/0003275 * Blandford & Königl (1979) Blandford, R. D., & Königl, A. 1979, ApJ, 232, 34, doi: 10.1086/157262 * Bu et al. (2015) Bu, Q.-c., Chen, L., Li, Z.-s., et al. 2015, ApJ, 799, 2, doi: 10.1088/0004-637X/799/1/2 * Capitanio et al. (2009) Capitanio, F., Belloni, T., Del Santo, M., & Ubertini, P. 2009, MNRAS, 398, 1194, doi: 10.1111/j.1365-2966.2009.15196.x * Casella et al. (2005) Casella, P., Belloni, T., & Stella, L. 2005, ApJ, 629, 403, doi: 10.1086/431174 * Chand et al. (2020) Chand, S., Agrawal, V. K., Dewangan, G. C., Tripathi, P., & Thakur, P. 2020, ApJ, 893, 142, doi: 10.3847/1538-4357/ab829a * Chen et al. (2010) Chen, Y. P., Zhang, S., Torres, D. F., et al. 2010, A&A, 522, A99, doi: 10.1051/0004-6361/201014017 * Coriat et al. (2011) Coriat, M., Corbel, S., Prat, L., et al. 2011, MNRAS, 414, 677, doi: 10.1111/j.1365-2966.2011.18433.x * Dauser et al. (2014) Dauser, T., Garcia, J., Parker, M. L., Fabian, A. C., & Wilms, J. 2014, MNRAS, 444, L100, doi: 10.1093/mnrasl/slu125 * Dauser et al. (2010) Dauser, T., Wilms, J., Reynolds, C. S., & Brenneman, L. W. 2010, MNRAS, 409, 1534, doi: 10.1111/j.1365-2966.2010.17393.x * Done et al. (2007) Done, C., Gierliński, M., & Kubota, A. 2007, A&A Rev., 15, 1, doi: 10.1007/s00159-007-0006-1 * Esin et al. (1997) Esin, A. A., McClintock, J. E., & Narayan, R. 1997, ApJ, 489, 865, doi: 10.1086/304829 * Fabian et al. (1989) Fabian, A. C., Rees, M. J., Stella, L., & White, N. E. 1989, MNRAS, 238, 729, doi: 10.1093/mnras/238.3.729 * Fender & Belloni (2004) Fender, R., & Belloni, T. 2004, ARA&A, 42, 317, doi: 10.1146/annurev.astro.42.053102.134031 * Fender (2001) Fender, R. P. 2001, MNRAS, 322, 31, doi: 10.1046/j.1365-8711.2001.04080.x * Fender et al. (2004) Fender, R. P., Belloni, T. M., & Gallo, E. 2004, MNRAS, 355, 1105, doi: 10.1111/j.1365-2966.2004.08384.x * Fender et al. (2009) Fender, R. P., Homan, J., & Belloni, T. M. 2009, MNRAS, 396, 1370, doi: 10.1111/j.1365-2966.2009.14841.x * Foreman-Mackey (2016) Foreman-Mackey, D. 2016, The Journal of Open Source Software, 1, 24, doi: 10.21105/joss.00024 * Foreman-Mackey et al. (2013) Foreman-Mackey, D., Hogg, D. W., Lang, D., & Goodman, J. 2013, PASP, 125, 306, doi: 10.1086/670067 * García et al. (2022) García, F., Karpouzas, K., Méndez, M., et al. 2022, MNRAS, 513, 4196, doi: 10.1093/mnras/stac1202 * García et al. (2021) García, F., Méndez, M., Karpouzas, K., et al. 2021, MNRAS, 501, 3173, doi: 10.1093/mnras/staa3944 * García et al. (2013) García, J., Dauser, T., Reynolds, C. S., et al. 2013, ApJ, 768, 146, doi: 10.1088/0004-637X/768/2/146 * García et al. (2014) García, J., Dauser, T., Lohfink, A., et al. 2014, ApJ, 782, 76, doi: 10.1088/0004-637X/782/2/76 * García et al. (2015a) García, J. A., Dauser, T., Steiner, J. F., et al. 2015a, ApJ, 808, L37, doi: 10.1088/2041-8205/808/2/L37 * García et al. (2015b) García, J. A., Steiner, J. F., McClintock, J. E., et al. 2015b, ApJ, 813, 84, doi: 10.1088/0004-637X/813/2/84 * Georganopoulos et al. (2002) Georganopoulos, M., Aharonian, F. A., & Kirk, J. G. 2002, A&A, 388, L25, doi: 10.1051/0004-6361:20020567 * Homan & Belloni (2005) Homan, J., & Belloni, T. 2005, Ap&SS, 300, 107, doi: 10.1007/s10509-005-1197-4 * Homan et al. (2005) Homan, J., Miller, J. M., Wijnands, R., et al. 2005, ApJ, 623, 383, doi: 10.1086/424994 * Homan et al. (2001) Homan, J., Wijnands, R., van der Klis, M., et al. 2001, ApJS, 132, 377, doi: 10.1086/318954 * Huang et al. (2018) Huang, Y., Qu, J. L., Zhang, S. N., et al. 2018, ApJ, 866, 122, doi: 10.3847/1538-4357/aade4c * Ingram et al. (2009) Ingram, A., Done, C., & Fragile, P. C. 2009, MNRAS, 397, L101, doi: 10.1111/j.1745-3933.2009.00693.x * Ingram et al. (2015) Ingram, A., Maccarone, T. J., Poutanen, J., & Krawczynski, H. 2015, ApJ, 807, 53, doi: 10.1088/0004-637X/807/1/53 * Ingram & van der Klis (2015) Ingram, A., & van der Klis, M. 2015, MNRAS, 446, 3516, doi: 10.1093/mnras/stu2373 * Ingram et al. (2017) Ingram, A., van der Klis, M., Middleton, M., Altamirano, D., & Uttley, P. 2017, MNRAS, 464, 2979, doi: 10.1093/mnras/stw2581 * Ingram et al. (2016) Ingram, A., van der Klis, M., Middleton, M., et al. 2016, MNRAS, 461, 1967, doi: 10.1093/mnras/stw1245 * Ingram & Motta (2019) Ingram, A. R., & Motta, S. E. 2019, New A Rev., 85, 101524, doi: 10.1016/j.newar.2020.101524 * Jonker et al. (2010) Jonker, P. G., Miller-Jones, J., Homan, J., et al. 2010, MNRAS, 401, 1255, doi: 10.1111/j.1365-2966.2009.15717.x * Kaluzienski & Holt (1977) Kaluzienski, L. J., & Holt, S. S. 1977, IAU Circ., 3099, 3 * Kara et al. (2019) Kara, E., Steiner, J. F., Fabian, A. C., et al. 2019, Nature, 565, 198, doi: 10.1038/s41586-018-0803-x * Karpouzas et al. (2021) Karpouzas, K., Méndez, M., García, F., et al. 2021, MNRAS, 503, 5522, doi: 10.1093/mnras/stab827 * Karpouzas et al. (2020) Karpouzas, K., Méndez, M., Ribeiro, E. M., et al. 2020, MNRAS, 492, 1399, doi: 10.1093/mnras/stz3502 * Kato (1990) Kato, S. 1990, PASJ, 42, 99 * Kong et al. (2020) Kong, L. D., Zhang, S., Chen, Y. P., et al. 2020, Journal of High Energy Astrophysics, 25, 29, doi: 10.1016/j.jheap.2020.01.003 * Kumar & Misra (2014) Kumar, N., & Misra, R. 2014, MNRAS, 445, 2818, doi: 10.1093/mnras/stu1946 * Lee & Miller (1998) Lee, H. C., & Miller, G. S. 1998, MNRAS, 299, 479, doi: 10.1046/j.1365-8711.1998.01842.x * Liu & Qiao (2022) Liu, B. F., & Qiao, E. 2022, arXiv e-prints, arXiv:2201.06198. https://arxiv.org/abs/2201.06198 * Lynden-Bell & Pringle (1974) Lynden-Bell, D., & Pringle, J. E. 1974, MNRAS, 168, 603, doi: 10.1093/mnras/168.3.603 * Maccarone (2003) Maccarone, T. J. 2003, A&A, 409, 697, doi: 10.1051/0004-6361:20031146 * Malzac et al. (2004) Malzac, J., Merloni, A., & Fabian, A. C. 2004, MNRAS, 351, 253, doi: 10.1111/j.1365-2966.2004.07772.x * Markoff et al. (2005) Markoff, S., Nowak, M. A., & Wilms, J. 2005, ApJ, 635, 1203, doi: 10.1086/497628 * Mastichiadis et al. (2022) Mastichiadis, A., Petropoulou, M., & Kylafis, N. D. 2022, A&A, 662, A118, doi: 10.1051/0004-6361/202243397 * McClintock et al. (2009) McClintock, J. E., Remillard, R. A., Rupen, M. P., et al. 2009, ApJ, 698, 1398, doi: 10.1088/0004-637X/698/2/1398 * Méndez et al. (2022) Méndez, M., Karpouzas, K., García, F., et al. 2022, Nature Astronomy, 6, 577, doi: 10.1038/s41550-022-01617-y * Merloni & Fabian (2001) Merloni, A., & Fabian, A. C. 2001, MNRAS, 321, 549, doi: 10.1046/j.1365-8711.2001.04060.x * Miller et al. (2006) Miller, J. M., Raymond, J., Homan, J., et al. 2006, ApJ, 646, 394, doi: 10.1086/504673 * Miller-Jones et al. (2012) Miller-Jones, J. C. A., Sivakoff, G. R., Altamirano, D., et al. 2012, MNRAS, 421, 468, doi: 10.1111/j.1365-2966.2011.20326.x * Miyamoto et al. (1991) Miyamoto, S., Kimura, K., Kitamoto, S., Dotani, T., & Ebisawa, K. 1991, ApJ, 383, 784, doi: 10.1086/170837 * Molteni et al. (1996) Molteni, D., Sponholz, H., & Chakrabarti, S. K. 1996, ApJ, 457, 805, doi: 10.1086/176775 * Motta et al. (2010) Motta, S., Muñoz-Darias, T., & Belloni, T. 2010, MNRAS, 408, 1796, doi: 10.1111/j.1365-2966.2010.17246.x * Motta et al. (2015) Motta, S. E., Casella, P., Henze, M., et al. 2015, MNRAS, 447, 2059, doi: 10.1093/mnras/stu2579 * Narayan & Yi (1994) Narayan, R., & Yi, I. 1994, ApJ, 428, L13, doi: 10.1086/187381 * Nathan et al. (2022) Nathan, E., Ingram, A., Homan, J., et al. 2022, MNRAS, 511, 255, doi: 10.1093/mnras/stab3803 * Parmar et al. (2003) Parmar, A. N., Kuulkers, E., Oosterbroek, T., et al. 2003, A&A, 411, L421, doi: 10.1051/0004-6361:20031140 * Plant et al. (2014) Plant, D. S., Fender, R. P., Ponti, G., Muñoz-Darias, T., & Coriat, M. 2014, MNRAS, 442, 1767, doi: 10.1093/mnras/stu867 * Reig & Kylafis (2021) Reig, P., & Kylafis, N. D. 2021, A&A, 646, A112, doi: 10.1051/0004-6361/202039903 * Remillard & McClintock (2006) Remillard, R. A., & McClintock, J. E. 2006, ARA&A, 44, 49, doi: 10.1146/annurev.astro.44.051905.092532 * Shakura & Sunyaev (1973) Shakura, N. I., & Sunyaev, R. A. 1973, A&A, 24, 337 * Shidatsu et al. (2014) Shidatsu, M., Ueda, Y., Yamada, S., et al. 2014, ApJ, 789, 100, doi: 10.1088/0004-637X/789/2/100 * Shui et al. (2021) Shui, Q. C., Yin, H. X., Zhang, S., et al. 2021, MNRAS, 508, 287, doi: 10.1093/mnras/stab2521 * Smith et al. (2009) Smith, A., Kaaret, P., Holder, J., et al. 2009, ApJ, 693, 1621, doi: 10.1088/0004-637X/693/2/1621 * Sobolewska & Życki (2006) Sobolewska, M. A., & Życki, P. T. 2006, MNRAS, 370, 405, doi: 10.1111/j.1365-2966.2006.10489.x * Steiner et al. (2012) Steiner, J. F., McClintock, J. E., & Reid, M. J. 2012, ApJ, 745, L7, doi: 10.1088/2041-8205/745/1/L7 * Stella & Vietri (1998) Stella, L., & Vietri, M. 1998, ApJ, 492, L59, doi: 10.1086/311075 * Stella et al. (1999) Stella, L., Vietri, M., & Morsink, S. M. 1999, ApJ, 524, L63, doi: 10.1086/312291 * Stiele & Yu (2016) Stiele, H., & Yu, W. 2016, MNRAS, 460, 1946, doi: 10.1093/mnras/stw821 * Sunyaev & Titarchuk (1980) Sunyaev, R. A., & Titarchuk, L. G. 1980, A&A, 86, 121 * Tagger & Pellat (1999) Tagger, M., & Pellat, R. 1999, A&A, 349, 1003. https://arxiv.org/abs/astro-ph/9907267 * Titarchuk (1994) Titarchuk, L. 1994, ApJ, 434, 570, doi: 10.1086/174760 * Uttley et al. (2011) Uttley, P., Wilkinson, T., Cassatella, P., et al. 2011, MNRAS, 414, L60, doi: 10.1111/j.1745-3933.2011.01056.x * van den Eijnden et al. (2017) van den Eijnden, J., Ingram, A., Uttley, P., et al. 2017, MNRAS, 464, 2643, doi: 10.1093/mnras/stw2634 * van der Klis (1989) van der Klis, M. 1989, ARA&A, 27, 517, doi: 10.1146/annurev.aa.27.090189.002505 * van Doesburgh & van der Klis (2020) van Doesburgh, M., & van der Klis, M. 2020, MNRAS, 496, 5262, doi: 10.1093/mnras/staa1867 * Veledina et al. (2013) Veledina, A., Poutanen, J., & Ingram, A. 2013, ApJ, 778, 165, doi: 10.1088/0004-637X/778/2/165 * Wagoner (1999) Wagoner, R. V. 1999, Phys. Rep., 311, 259, doi: 10.1016/S0370-1573(98)00104-5 * Wang et al. (2022) Wang, P. J., Kong, L. D., Chen, Y. P., et al. 2022, MNRAS, 512, 4541, doi: 10.1093/mnras/stac773 * Weng et al. (2021) Weng, S.-S., Cai, Z.-Y., Zhang, S.-N., et al. 2021, ApJ, 915, L15, doi: 10.3847/2041-8213/ac0a7b * Wijnands et al. (1999) Wijnands, R., Homan, J., & van der Klis, M. 1999, ApJ, 526, L33, doi: 10.1086/312365 * Williams et al. (2020) Williams, D. R. A., Motta, S. E., Fender, R., et al. 2020, MNRAS, 491, L29, doi: 10.1093/mnrasl/slz152 * Wilms et al. (2000) Wilms, J., Allen, A., & McCray, R. 2000, ApJ, 542, 914, doi: 10.1086/317016 * You et al. (2018) You, B., Bursa, M., & Życki, P. T. 2018, ApJ, 858, 82, doi: 10.3847/1538-4357/aabd33 * You et al. (2020) You, B., Życki, P. T., Ingram, A., Bursa, M., & Wang, W. 2020, ApJ, 897, 27, doi: 10.3847/1538-4357/ab9838 * You et al. (2021) You, B., Tuo, Y., Li, C., et al. 2021, Nature Communications, 12, 1025, doi: 10.1038/s41467-021-21169-5 * Yuan & Narayan (2014) Yuan, F., & Narayan, R. 2014, ARA&A, 52, 529, doi: 10.1146/annurev-astro-082812-141003 * Zdziarski et al. (2004) Zdziarski, A. A., Gierliński, M., Mikołajewska, J., et al. 2004, MNRAS, 351, 791, doi: 10.1111/j.1365-2966.2004.07830.x * Zdziarski et al. (1996) Zdziarski, A. A., Johnson, W. N., & Magdziarz, P. 1996, MNRAS, 283, 193, doi: 10.1093/mnras/283.1.193 * Zdziarski et al. (1999) Zdziarski, A. A., Lubiński, P., & Smith, D. A. 1999, MNRAS, 303, L11, doi: 10.1046/j.1365-8711.1999.02343.x * Zhang et al. (2017) Zhang, L., Wang, Y., Méndez, M., et al. 2017, ApJ, 845, 143, doi: 10.3847/1538-4357/aa8138 * Zhang et al. (2020) Zhang, L., Méndez, M., Altamirano, D., et al. 2020, MNRAS, 494, 1375, doi: 10.1093/mnras/staa797 * Zhang et al. (2022) Zhang, Y., Méndez, M., García, F., et al. 2022, MNRAS, 512, 2686, doi: 10.1093/mnras/stac690 * Zhou et al. (2013) Zhou, J. N., Liu, Q. Z., Chen, Y. P., et al. 2013, MNRAS, 431, 2285, doi: 10.1093/mnras/stt326 * Życki et al. (1999) Życki, P. T., Done, C., & Smith, D. A. 1999, MNRAS, 309, 561, doi: 10.1046/j.1365-8711.1999.02885.x
# High-Throughput Ab Initio Design of Atomic Interfaces using InterMatch Eli Gerber School of Applied and Engineering Physics, Cornell University, Ithaca, NY 14853, USA Steven B. Torrisi Department of Physics, Harvard University, Cambridge, MA 02138, USA Energy & Materials Division, Toyota Research Institute, Los Altos, CA 94022, USA Sara Shabani Department of Physics, Columbia University, New York, NY, USA Eric Seewald Department of Physics, Columbia University, New York, NY, USA Jordan Pack Department of Physics, Columbia University, New York, NY, USA Jennifer E. Hoffman Department of Physics, Harvard University, Cambridge, MA 02138, USA John A. Paulson School of Engineering and Applied Sciences,Harvard University, Cambridge, MA 02138, USA Cory R. Dean Department of Physics, Columbia University, New York, NY, USA Abhay N. Pasupathy Department of Physics, Columbia University, New York, NY, USA Eun-Ah Kim Department of Physics, Cornell University, Ithaca, NY 14853, USA ( [file: main]) ###### Abstract Forming a hetero-interface is a materials-design strategy that can access an astronomically large phase space. However, the immense phase space necessitates a high-throughput approach for an optimal interface design. Here we introduce a high-throughput computational framework, InterMatch, for efficiently predicting charge transfer, strain, and superlattice structure of an interface by leveraging the databases of individual bulk materials. Specifically, the algorithm reads in the lattice vectors, density of states, and the stiffness tensors for each material in their isolated form from the Materials Project. From these bulk properties, InterMatch estimates the interfacial properties. We benchmark InterMatch predictions for the charge transfer against experimental measurements and supercell density-functional theory calculations. We then use InterMatch to predict promising interface candidates for doping transition metal dichalcogenide MoSe2. Finally, we explain experimental observation of factor of 10 variation in the supercell periodicity within a few microns in graphene/$\alpha$-RuCl3 by exploring low energy superlattice structures as a function of twist angle using InterMatch. We anticipate our open-source InterMatch algorithm accelerating and guiding ever-growing interfacial design efforts. Moreover, the interface database resulting from the InterMatch searches presented in this paper can be readily accessed through MPContribs. Figure 1: (a) Role of InterMatch in the materials discovery process. (b) Input from the bulk database. Lattice vectors $\vec{a}_{\alpha}$, $\vec{b}_{\alpha}$, density-of-states $g_{\alpha}$ of systems $\alpha=1,2$. Elastic tensors are additional inputs. (c) $E_{F}^{\alpha}$ are bulk Fermi levels and $E_{F}^{\prime}$ is new equilibrium Fermi level. (d) $\Delta n$ is the transferred charge density and $d$ is the interlayer separation between the two systems, taken to be the sum of the largest van der Waals radii of the species in each system 1 and 2. (e) Superlattice vectors $\vec{v_{i}}$ (orange arrows) and their near-equivalent vectors $\vec{u_{i}}$ (magenta arrows). Candidate supercells, formed in each basis by combining $\\{(\vec{v}_{i},\vec{u}_{i}),(\vec{v}_{j},\vec{u}_{j})\\}$ pairs specify the strain $\varepsilon_{i}$ (blue arrows). (f) Optimal supercell minimizes the elastic energy and the number of atoms in the cell. With increasing control in interface fabrication, interfacial systems form an arena of limitless possibilitiesGeim and Grigorieva (2013). Recent developments with moiré heterostructuresKennes et al. (2021) further enlarged the phase space to include the twist angle. However, the vast space of possibilities also implies it is crucial to go beyond serendipitous discoveries and empirical explorations to effectively harness the intrinsic potential of interfacial systems. The traditional approach to theoretically studying interfaces is to carry out density-functional theory (DFT) calculations on a supercell system consisting of two materialsKomsa and Krasheninnikov (2013); Terrones and Terrones (2014); Ebnonnasir et al. (2014); Bokdam et al. (2014); Zhou et al. (2015). While such approaches are rigorous, computational limitations regularly require imposing unnatural strain to form a periodic structure. Moreover, the $\mathcal{O}(N^{3})$ scaling of DFT in the number of electrons $N$ for each such calculation prohibits a comprehensive exploration. Some of us recently proposed an “intermediate scale” approach called Mismatched INterface Theory (MINT)Gerber et al. (2020), which can predict charge transfer and natural strain approximating one layer of the interface using finite-size scaling of atomic clusters. While MINT calculations are much more computationally affordable, they are not fast enough for an exhaustive survey in real-time. Therefore, a comprehensive and fast approach to scanning the relevant phase space of interfacial combinations is greatly needed. With advancements in widely available comprehensive materials databasesHorton et al. (2021); Levin (2020); Landis et al. (2012); Curtarolo et al. (2012); Jain et al. (2013); Saal et al. (2013); de Jong et al. (2015); Ashton et al. (2017); Haastrup et al. (2018); Zhou et al. (2019); Draxl and Scheffler (2019); Talirz et al. (2020); Choudhary et al. (2020a), it is timely to establish a high-throughput approach to interface design that can leverage the information contained in these databases to make predictions of interface physics. While it is possible to look up a pair of materials in the database and try to match their properties, the phase space of all possible combinations amounts easily to $>\mathcal{O}(10^{6})$ possibilities, which is beyond the scale of what a manual search can reasonably accomplish. At the same time, it is desirable to benefit from extensive existing databases and search exhaustively for ideal combinations to reach a given objective. Indeed, efforts to make interfacial predictions using bulk material databases are beginning to emergeMathew et al. (2016); Ding et al. (2016); Choudhary et al. (2020b); Boland and Singh (2022). However, so far, the existing approaches aid growth and calculation decisions for a specific pair of materials rather than allowing for a comprehensive query to yield fast, approximate predictions over a wide range of possible interfaces. In this paper, we introduce InterMatch, which uses information readily available from preexisting materials databases such as the Materials Project and 2DMatPedia databases to predict charge transferMilnes and Feucht (1972), strainLandau et al. (1986), stabilityMuller (1994), and optimal superlatticeKaxiras (2003) of an atomic interface. Using these predictions, InterMatch can narrow the candidate pool from $C>\mathcal{O}(10^{6})$ to $\mathcal{O}(10)$ that can then be investigated in greater detail using MINT or supercell DFT (see Fig. 1(a)). We first illustrate how two branches of InterMatch predict the charge transfer and optimal superlattice after querying the entries of the Materials Project for each of the constituents of the interface. We then benchmark InterMatch predictions for the charge transfer against experimental measurements and supercell DFT predictions. We then employ InterMatch’s branches to address two bottleneck problems obstructing design of interfaces towards the goal of discovering new physics: the problem of doping transition metal dichalcogenides and the problem of predicting stable interface structure, applied to the graphene/$\alpha$-RuCl3 system. We comment on many other classes of interfaces that can be optimized using InterMatch. Starting with ab initio materials data, InterMatch performs high-throughput screening of possible heterostructures via pairwise calculation of desired interface properties including charge transfer $\Delta n$, strain tensor $\tilde{\varepsilon}$, optimized superlattice vectors $\vec{v_{1}},$ $\vec{v_{2}}$, and number of atoms $N$. Once InterMatch identifies a promising pool of candidate combinations, one can make a more in-depth analysis of the smaller pool using MINT or supercell DFT (See Fig 1(a)). The InterMatch algorithm has two branches to predict two key electronic and mechanical characteristics of candidate interfaces: charge transfer and optimized superlattice structure (See Fig. 1). One branch is devoted to calculating charge transfer (Fig. 1(c)-(d)), and the other is devoted to optimizing supercell structure by minimizing the number of atoms and elastic energy (Fig. 1(e)-(f)). For the first branch that estimates the direction and magnitude of charge transfer, we use a simple model to describe the Fermi level shifts occurring in each material when they are brought together in proximityRuan and Ching (1987). Fig. 1(c) shows how the Fermi level shifts are determined: systems 1 and 2 are designated as “donor” or “acceptor” based on their relative Fermi levels $E_{F}^{1}$ and $E_{F}^{2}$, and the difference of integrals over $g_{1}(E)$ and $g_{2}(E)$ $\int_{E_{F}^{{}^{\prime}}}^{E_{F}^{1}}dE\;g_{1}(E)=\int_{E_{F}^{2}}^{E_{F}^{{}^{\prime}}}dE\;g_{2}(E)$ (1) is minimized to determine the equilibrium Fermi level $E_{F}^{{}^{\prime}}$. We take interaction between the two systems at the interface into account in the estimation of the charge transfer $\Delta n$ using a simple capacitor modelLowell and Rose-Innes (1980). Specifically, we model the interface as a parallel plate capacitor with the separation $d$ given by the sum of the largest van der Waals radii of the species in each system 1 and 2 (See Fig. 1(d)). The charge transfer depends on the equilibrium Fermi level $E_{F}^{\prime}$ and the distance $d$ as $e\Delta n=\varepsilon_{0}E_{F}^{{}^{\prime}}/d$. Figure 2: Benchmarking InterMatch. (a) Comparison of charge transfer predicted by InterMatch with measured experimental values for interfaces in RefsAnnadi et al. (2013); Wang et al. (2020); Sutter et al. (2009); Zheng et al. (2021a). (b) InterMatch screening of over 10,000 2D materials in heterostructure with monolayer MoSe2, ranked in descending order of charge transfer $\absolutevalue{\Delta n}$. (c) Substrate selection based on InterMatch screening results from red box in (a) according to $\Delta n$, elastic energy $\mathcal{E}$, and energy above-hull. (d) Comparison of InterMatch predictions for $\Delta n$ (solid symbols) with supercell DFT calculations (open symbols). The second branch of the InterMatch algorithm sketched in Fig. 1(e-f) constructs optimal supercells from a pair of queried systems by calculating strain and elastic energy at their interface over a series of supercell configurations. Given the lattice vectors of system 1, $\vec{a}_{1}$, $\vec{b}_{1}$, and those of system 2, $\vec{a}_{2}$, $\vec{b}_{2}$, the algorithm searches for pairs of near-equivalent superlattice vectors $\\{(\vec{u}_{1},\vec{v}_{1}),(\vec{u}_{2},\vec{v}_{2})\\}$: $\displaystyle\vec{u}_{i}$ $\displaystyle=\mathbf{M}^{i}_{11}\vec{a}_{i}+\mathbf{M}^{i}_{12}\vec{b}_{i}$ (2) $\displaystyle\vec{v}_{i}$ $\displaystyle=\mathbf{M}^{i}_{21}\vec{a}_{i}+\mathbf{M}^{i}_{22}\vec{b}_{i}$ where $\mathbf{M}^{i}$ is a $2\times 2$ matrix of integer coefficients for the system $i$ and the “near-equivalence” is defined by $\mathbf{M}^{1}=(\tilde{\varepsilon}^{2}+1)\mathcal{R}_{\theta}\mathbf{M}^{2}.$ (3) Here $\mathcal{R}_{\theta}$ is an in-plane rotation matrix by angle $\theta$ and $\tilde{\varepsilon}^{2}$ is the strain tensor resulting from straining the Bravais lattice of system 2 to match that of system 1. We choose system 2 to be the material with the smallest elements of the stiffness tensor $\mathbf{C}$ (queried from the Materials Project) in the strain direction. InterMatch then computes the elastic energy $\mathcal{E}=\frac{1}{2}C_{ijkl}\varepsilon_{ij}\varepsilon_{kl}$ for the superlattice candidate according to classical elastic plate theoryLandau et al. (1986). The optimal supercell is determined by simultaneously minimizing the elastic energy $\mathcal{E}$ and the cell area $\absolutevalue{\vec{u}_{i}\crossproduct\vec{v}_{i}}$. The use of elastic energy goes beyond previous approaches for finding the superlatticeStradi et al. (2017); Lazić (2015) which only consider geometric strain. Figure 3: Superlattice structure prediction optimizing elastic energies. (a)-(b) (Top view) Diagonal stiffness tensor components $C_{11}$ and $C_{22}$ of primitive MoSe2 and ZrTe3 unit cells in Voigt notation. (c)-(d) (Top view) ZrTe3 layer of two candidate MoSe2/ZrTe3 supercells with the same number of atoms and average strain $\varepsilon_{av}^{\textnormal{ZrTe}_{3}}$. The solid black boxes denote the strained ZrTe3 unit cells and the dashed red boxes are the original unstrained primitive cells. (e) Average strain values $\varepsilon_{av}$ of the MoSe2/ZrTe3 interfaces in Cells 1 and 2. (f) Elastic energies $\mathcal{E}$ of the interfaces. Now we demonstrate how elastic energy considerations can make a difference in optimization of the superlattice. Consider the MoSe2/ZrTe3 interface. The primitive unit cells of MoSe2 and ZrTe3 are shown in Fig. 3 (a) and (b) respectively, along with the diagonal components of their stiffness tensors $C_{ii}$ in Voigt notation. The anisotropy of the ZrTe3 stiffness tensor is such that the energetic cost to deforming ZrTe3 along the direction of $C_{22}$ far exceeds the cost of an equivalent deformation along the direction of $C_{11}$. Cells 1 and 2 in Fig. 3 (c) and (d) result from an InterMatch search for low-area, low-strain MoSe2/ZrTe3 supercells. The two cells are identical in number of atoms and geometric strain $\varepsilon_{av}$ (shown in Fig. 3 (e)), however, cell 1 is favored energetically due to the different strains required to make each supercell commensurate with MoSe2 (Fig. 3 (f)). We now benchmark charge transfer predictions by InterMatch against experimentally measured charge transfer in known interfaces. Fig. 2 (a) shows a comparison of InterMatch predictions of charge transfer with experimentally obtained values for several interfaces: LaAlO3/SrTiO3(1 1 0)Annadi et al. (2013), GR/$\alpha$-RuCl3Wang et al. (2020), GR/Pt$(1\,1\,1)$Sutter et al. (2009), and MoS2/MgAl2O4Zheng et al. (2021a). The magnitudes of the $\Delta n$ predicted with InterMatch are at the same order of magnitude as the measured values, especially given experimental error bars, with the exception of GR/$\alpha$-RuCl3. However, spin-orbit coupling effects (absent from our calculations) are known to affect the band structure of $\alpha$-RuCl3Kim et al. (2015), altering the band alignment with GR and the resulting charge transfer. Next we turn to the application of charge transfer prediction to the problem of doping transition metal dichalcogenides (TMDs). TMDs have emerged as an exciting van der Waals material platform at the intersection of semiconductor physics and strong correlation physicsManzeli et al. (2017). Due to the spin- valley locking Ising spin-orbit coupling, an exotic $p$-wave superconducting state was proposed for hole-doped TMDsHsu et al. (2017). Recent developments in TMD moiré systems have further extended the phase space of possibilities. However, a major bottleneck against testing these proposals is the difficulty of establishing a good contact. Empirically, it has been established that doping the contact area can significantly improve the contact resistanceZheng et al. (2021b). However, gate-based doping does not scale well. While successful modulation doping using work function difference was established in graphene/$\alpha$-RuCl3 heterostructuresWang et al. (2020), it is desirable to perform an exhaustive search of interface possibilities. We seek 2D substrates for controlling carrier concentration in MoSe2. We screen all entries of the 2DMatPedia database and 3000 entries from the Materials Project for stable 2D materials composed of elements making up the majority of commercially available semiconductors, semimetals, and metals. We use InterMatch to down-sample from 10,000 candidate 2D substrates based on the magnitude of the predicted charge transfer $\absolutevalue{\Delta n}$ to MoSe2 in the desired range $\gtrsim\mathcal{O}(10^{13})$ cm-2 (Fig. 2 (b)). We then select from these the compounds with maximum $\absolutevalue{\Delta n}$, minimal strain, and minimal above-hull energy (Fig. 2 (c)). Finally, we choose a small subset of the top interfaces (2H-TaS2, $\beta$-GaSe, and ZrTe3, in the case of our example) and benchmark InterMatch predictions against supercell DFT calculations using the optimized supercells generated by InterMatch. Fig. 2 (d) shows a comparison of InterMatch predictions with the results from DFT for the top three interfaces (for computational details, see the Supplemental Material). The magnitude of the $\Delta n$ prediction from the two approaches are within $10^{13}$ cm-2. Moreover, both approaches find consistent relative magnitude of charge transfer. Given the high-throughput nature of InterMatch, these agreements encourage using InterMatch as the first pass in searches for optimal heterostructures. As an example of the power of Intermatch to understand superlattice structure, we consider the graphene/$\alpha$-RuCl3 heterostructure (GR/$\alpha$-RuCl3). This system has attracted great interest due to the presence of strong modulation dopingWang et al. (2020) and enhancement of $\alpha$-RuCl3’s proximity to the Kitaev spin liquid phase. However, relatively little attention has been paid to the atomic scale structure of the heterostructure and the possible influence on electronic properties. In order to study this experimentally, we used scanning tunneling microscopy (STM) to investigate the properties of GR/$\alpha$-RuCl3 heterostructures created by mechanical exfoliation and colamination, as shown in Fig. 4 (a). The angle between the $\alpha$-RuCl3 substrate and graphene was not intentionally controlled. Shown in Fig. 4 (b)-(d) are a set of STM topographs taken at various locations of the GR/$\alpha$-RuCl3 heterostructure. The locations are within a few microns of each other on the sample shown in Fig. 4 (a). Intriguingly, all three of the regions show moiré patterns with large wavelengths - 2.7 nm in Fig. 4 (b), 11.7 nm in Fig. 4 (c) and 25.7 nm in Fig. 4 (d). All three of these wavelengths are much larger than the wavelength set by the difference in lattice constants. Using InterMatch, we perform a comprehensive mapping of the space of superlattice configurations spanned by $(\theta,L,\mathcal{E})$ where $\theta$ is the twist angle, $L$ is the moiré period, and $\mathcal{E}$ is the elastic energy of the interface. The resulting spectrum of low-energy superlattice configurations for $0^{\circ}\leq\theta\leq 30^{\circ}$ and $0\;\textnormal{nm}\leq L\leq 30\;\textnormal{nm}$ is shown in Fig. 4 (e). We identify four prominent moiré length scales (blue boxes in Fig. 4 (e)) occurring within a $5^{\circ}$ range between $15^{\circ}$-$20^{\circ}$. Three of the four length scales coincide with those observed in STM at $L=2.7,\;11.7,\;25.7$ nm, shown in Fig. 4 (b)-(d). Correctly identifying energetically favorable GR/$\alpha$-RuCl3 superlattices over a narrow range of twist angles showcases InterMatch’s capability to predict interfacial structure of complex (e.g extremely lattice-mismatched) systems. The presence of an atomic reconstruction at the interface of GR/$\alpha$-RuCl3 can have dramatic consequences for the spectroscopic properties of the material. Shown in Fig. 4 (f)-(h) are scanning tunneling spectra averaged over the regions shown in Fig. 4 (b)-(d). These spectra show dramatic differences from the simple expectation for a doped Dirac spectrum as might be expected from charge transfer alone. Instead, we observe strong resonances in all three regions, with the spacing between resonances following the expectation from Landau levels on a Dirac spectrum. Previously, such spectra have been observed when graphene has a periodic bucklingMao et al. (2020), where it was ascribed to periodic strain in the material. In our case, apart from the strain associated with the moiré latticeShabani et al. (2021), we expect that there will also be strong periodic variations in the dopingRizzo et al. (2022) that contribute to the formation of resonances. Figure 4: (a) Optical image of the measured device of GR/$\alpha$-RuCl3 contacted with bismuth indium tin for the STM measurements. The blue and green dashed lines show the boundary of $\alpha$-RuCl3 and graphene, respectively. (b)-(d) STM topographic images (in pm) of GR/$\alpha$-RuCl3 on 2.7 nm (set points of -100 mV and -100 pA), 11.7 nm and 25.7nm (set points of -1 V and -50 pA) moiré patterns due to atomic reconstruction. (e) InterMatch predictions for low-energy GR/$\alpha$-RuCl3 superlattice configurations as a function of period $L$, twist angle $\theta$, and elastic energy $\mathcal{E}$. Left panel is a projection onto the $L$-axis, bottom panel is a projection onto the $\theta$-axis. Dashed box indicates interval of $\theta$ containing largest range of stable superlattice periods, shaded blue boxes indicate regions of likely superlattice configurations. Purple stars denote the periodicities extracted from the experiment. (f)-(h) dI/dV measurements corresponding to the three moiré patterns in (b)-(d) showing strong resonances dependent on moire wavelengths. In summary, we introduce and demonstrate InterMatch, a high-throughput computational framework and database for predicting charge transfer, strain, and superlattice of an interface between two arbitrary materials. Charge transfer allows heterostructure-based modulation dopingWang et al. (2020), which can guide device fabrication and contact design Zheng et al. (2021b). Efficiently determining the smallest energetically favorable commensurate supercells from a wide variety of interface configurations is crucial for accelerating ab initio studies. We showcase the use of InterMatch by identifying high-charge transfer substrates for doping TMDs, and by predicting equilibrium moiré superlattice configurations for the lattice-mismatched GR/$\alpha$-RuCl3 interface that are validated by STM measurements. The presence of such long-wavelength superlattice modulations at van der Waals interfaces present new opportunities to tailor bandstructure using materials that do not have a close match in lattice constants. The evolving interface database provides open access to InterMatch results which we hope will help guide future exploration of interfacial systems. To broadly benefit the community, we made the InterMatch code openly accessible at https://doi.org/10.5281/zenodo.6823973Gerber and Torrisi (2022). Moreover, we tabulate InterMatch results in an open-access “interface database” directly integrated with the Materials Project via the MPContribs platform. At the time of writing, the database contains $\sim 200,000$ interfaces (and counting) in simple JavaScript Object Notation (JSON) that are queryable and sortable according to the chemical composition of either constituent system, charge transfer, strain, and optimized supercell size. In addition, we generate crystallographic information files (CIF) of interface supercells with InterMatch which may be readily accessed from the database and used as inputs for DFT or other first principles studies. ###### Acknowledgements. The authors thank Kin Fai Mak, Jie Shan, Stephen Carr, Patrick Huck, Matthew Horton, Jason Munro, and Vidya Madhavan for helpful discussions. EAK and JH were supported by MURI grant FA9550-21-1-0429. EG was supported by the Cornell Center for Materials Research with funding from the NSF MRSEC program (DMR-1719875). SBT was supported by the Department of Energy Computational Science Graduate Fellowship under grant DE-FG02-97ER25308. The computation was done using the high powered computing cluster WALLE2 that was established through the support of Gordon and Betty Moore Foundation’s EPiQS Initiative, Grant GBMF10436 to EAK and the New Frontier Grant from Cornell University’s College of Arts and Sciences and hosted and maintained by Cornell Center for Advanced Computing. STM experiments were supported by NSF DMR-2004691 (ANP) and AFOSR via grant FA9550-21-1-0378 (SS, ES). Sample synthesis for STM measurements was supported by the NSF MRSEC program through Columbia in the Center for Precision-Assembled Quantum Materials (PAQM), grant number DMR-2011738. ## Supplemental Material In Section I we present additional InterMatch results identifying optimal substrate candidates for forming high-charge-transfer interfaces with group-VI transition metal dichalcogenides and putative spin liquid materials $\alpha$-RuCl3 and TbInO3. In Section II we provide details of the InterMatch code. In Section III we provide computational details of the ab initio density-functional theory (DFT) calculations of MoSe2 supercells in Fig. 2(d) of the main text. ## I Charge transfer substrates for WTe2, WSe2, and $\alpha$-RuCl3 We further demonstrate InterMatch by applying it to two types of materials: group-VI TMDs, and putative spin liquid $\alpha$-RuCl3Plumb et al. (2014). TMDs’ unique combination of properties makes them highly attractive for nanoelectronics applications and fundamental studies of novel physical phenomenaManzeli et al. (2017). However, the realization of many such high- performance devices and exotic phases is limited by the availability of systems with high carrier mobility and low contact resistances between metal contacts and the semiconductor. Quantum spin liquids (QSLs) are interacting quantum systems in which spins do not order at low temperatures, and have been theorized to offer insights into high-temperature superconductivity upon doping. $\alpha$-RuCl3, for example, has been intensively discussed as a possible candidate for Kitaev physics; however, it orders antiferromagnetically at low temperatures due to the presence of additional magnetic couplings extending beyond the pure Kitaev interaction. Doping $\alpha$-RuCl3 with charge carriers has been predicted to enhance Kitaev interactions and push $\alpha$-RuCl3 closer to the spin liquid phase. We use InterMatch to identify stable, high-charge-transfer interfaces for electron- and hole-doping the TMDs WTe2, WSe2, and MoSe2, and the putative QSL $\alpha$-RuCl3. Fig. S1 shows sample InterMatch results of substrate candidates for interfaces with WTe2, WSe2, and $\alpha$-RuCl3, highlighting those that minimize elastic energy $\mathcal{E}$ and maximize charge transfer $\Delta n$. In Fig. S1 (a)-(b), we survey all $\sim 70,000$ oxides from the Materials Project. Oxides have a wide range of charge neutrality levels and therefore constitute a powerful addition to electrostatic gating or chemical doping for controlling carrier concentration in heterostructures. In Fig. S1 (c) we survey all entries in the 2DMatpedia database, and determine two- dimensional (2D) substrate candidates to maximally dope $\alpha$-RuCl3. Figure S1: Sample InterMatch results for different charge transfer interfaces. The solid green line indicates the pareto-optimal frontier of substrate candidates that minimize elastic energy $\mathcal{E}$ and maximize charge transfer $\absolutevalue{\Delta n}$. Calculated quantities in (a)-(b) use ab initio data from the Materials Project as inputs, while those in (c) use data from the 2DMatpedia database. ## II InterMatch Code The InterMatch code is written in Python 3.7 and makes extensive use of pymatgenOng et al. (2013), an open-source Python package of the Materials Project, for the manipulation and analysis of various structures of interest. The code is continuously being developed, and the latest version can be obtained at https://doi.org/10.5281/zenodo.6823973Gerber and Torrisi (2022). We aim to provide an efficient scheme for computing interface properties capable of screening a significant fraction of combinations of existing Materials Project structure entries, returning the results in real-time (typical run time for the calculation of a single interface is $10\pm 5$ seconds on a 1.7 GHz Intel Core i5 processor at 1333 MHz using 4 GB of RAM, running macOS Sierra 10.12.6). ## III Computational Details All ab initio DFT calculations were carried out within the total-energy plane wave density-functional pseudopotential approach, using Perdew-Burke-Ernzerhof generalized gradient approximation functionalsPerdew et al. (1996) and optimized norm-conserving Vanderbilt pseudopotentials in the SG15 familySchlipf and Gygi (2015). Plane wave basis sets with energy cutoffs of 30 hartree were used to expand the electronic wave functions. We used fully periodic boundary conditions and a $8\times 8\times 1$ $k$-point mesh to sample the Brillouin zone. Electronic minimizations were carried out using the analytically continued functional approach starting with a LCAO initial guess within the DFT$++$ formalismFreysoldt et al. (2009), as implemented in the open-source code JDFTxSundararaman et al. (2017) using direct minimization via the conjugate gradients algorithmPayne et al. (1992). All unit cells were constructed to be inversion symmetric about $z=0$ with a distance of $\sim 60$ bohr between periodic images of the MoSe2 surface, using coulomb truncation to prevent image interaction. ## References * Geim and Grigorieva (2013) A. K. Geim and I. V. Grigorieva, Nature 499, 419 (2013). * Kennes et al. (2021) D. M. Kennes, M. Claassen, L. Xian, A. Georges, A. J. Millis, J. Hone, C. R. Dean, D. N. Basov, A. N. Pasupathy, and A. Rubio, Nature Physics 17, 155 (2021), ISSN 1745-2481. * Komsa and Krasheninnikov (2013) H.-P. Komsa and A. V. Krasheninnikov, Physical Review B 88, 085318 (2013), publisher: American Physical Society. * Terrones and Terrones (2014) H. Terrones and M. Terrones, Journal of Materials Research 29, 373 (2014), ISSN 2044-5326. * Ebnonnasir et al. (2014) A. Ebnonnasir, B. Narayanan, S. Kodambaka, and C. V. Ciobanu, Applied Physics Letters 105, 031603 (2014), ISSN 0003-6951, publisher: American Institute of Physics. * Bokdam et al. (2014) M. Bokdam, T. Amlaki, G. Brocks, and P. J. Kelly, Phys. Rev. B 89, 201404 (2014). * Zhou et al. (2015) S. Zhou, J. Han, S. Dai, J. Sun, and D. J. Srolovitz, Physical Review B 92, 155438 (2015). * Gerber et al. (2020) E. Gerber, Y. Yao, T. A. Arias, and E.-A. Kim, Physical Review Letters 124, 106804 (2020). * Horton et al. (2021) M. K. Horton, S. Dwaraknath, and K. A. Persson, Nature Computational Science 1, 3 (2021), ISSN 2662-8457. * Levin (2020) I. Levin, _NIST Inorganic Crystal Structure Database (ICSD)_ (2020). * Landis et al. (2012) D. D. Landis, J. S. Hummelshøj, S. Nestorov, J. Greeley, M. Dułak, T. Bligaard, J. K. Nørskov, and K. W. Jacobsen, Computing in Science Engineering 14, 51 (2012). * Curtarolo et al. (2012) S. Curtarolo, W. Setyawan, S. Wang, J. Xue, K. Yang, R. H. Taylor, L. J. Nelson, G. L. Hart, S. Sanvito, M. Buongiorno-Nardelli, et al., Computational Materials Science 58, 227 (2012), ISSN 0927-0256. * Jain et al. (2013) A. Jain, S. P. Ong, G. Hautier, W. Chen, W. D. Richards, S. Dacek, S. Cholia, D. Gunter, D. Skinner, G. Ceder, et al., APL Materials 1, 011002 (2013). * Saal et al. (2013) J. E. Saal, S. Kirklin, M. Aykol, B. Meredig, and C. Wolverton, JOM 65, 1501 (2013). * de Jong et al. (2015) M. de Jong, W. Chen, T. Angsten, A. Jain, R. Notestine, A. Gamst, M. Sluiter, C. Krishna Ande, S. van der Zwaag, J. J. Plata, et al., Scientific Data 2, 150009 (2015). * Ashton et al. (2017) M. Ashton, J. Paul, S. B. Sinnott, and R. G. Hennig, Physical Review Letters 118, 106101 (2017). * Haastrup et al. (2018) S. Haastrup, M. Strange, M. Pandey, T. Deilmann, P. S. Schmidt, N. F. Hinsche, M. N. Gjerding, D. Torelli, P. M. Larsen, A. C. Riis-Jensen, et al., 2D Materials 5, 042002 (2018), ISSN 2053-1583. * Zhou et al. (2019) J. Zhou, L. Shen, M. D. Costa, K. A. Persson, S. P. Ong, P. Huck, Y. Lu, X. Ma, Y. Chen, H. Tang, et al., Scientific Data 6, 86 (2019), ISSN 2052-4463. * Draxl and Scheffler (2019) C. Draxl and M. Scheffler, Journal of Physics: Materials 2, 036001 (2019). * Talirz et al. (2020) L. Talirz, S. Kumbhar, E. Passaro, A. V. Yakutovich, V. Granata, F. Gargiulo, M. Borelli, M. Uhrin, S. P. Huber, S. Zoupanos, et al., Scientific Data 7, 299 (2020). * Choudhary et al. (2020a) K. Choudhary, K. F. Garrity, A. C. E. Reid, B. DeCost, A. J. Biacchi, A. R. Hight Walker, Z. Trautt, J. Hattrick-Simpers, A. G. Kusne, A. Centrone, et al., npj Computational Materials 6, 173 (2020a). * Mathew et al. (2016) K. Mathew, A. K. Singh, J. J. Gabriel, K. Choudhary, S. B. Sinnott, A. V. Davydov, F. Tavazza, and R. G. Hennig, Computational Materials Science 122, 183 (2016), ISSN 0927-0256. * Ding et al. (2016) H. Ding, S. S. Dwaraknath, L. Garten, P. Ndione, D. Ginley, and K. A. Persson, ACS Applied Materials & Interfaces 8, 13086 (2016), ISSN 1944-8244. * Choudhary et al. (2020b) K. Choudhary, K. F. Garrity, G. Pilania, and F. Tavazza (2020b). * Boland and Singh (2022) T. M. Boland and A. K. Singh, Computational Materials Science 207, 111238 (2022). * Milnes and Feucht (1972) A. G. Milnes and D. L. Feucht, _Heterojunctions and metal-semiconductor junctions_ (Academic Press, New York, 1972), ISBN 0-12-498050-3. * Landau et al. (1986) L. D. Landau, E. M. Lifshitž, J. B. Sykes, and W. H. Reid, in _Course of Theoretical Physics: Theory of Elasticity_ (Butterworth-Heinemann, 1986), chap. 2, pp. 46–58. * Muller (1994) P. Muller, Pure and Applied Chemistry 66, 1077 (1994). * Kaxiras (2003) E. Kaxiras, _Atomic and Electronic Structure of Solids_ (Cambridge University Press, 2003). * Ruan and Ching (1987) Y.-C. Ruan and W. Y. Ching, Journal of Applied Physics 62, 2885 (1987), ISSN 0021-8979. * Lowell and Rose-Innes (1980) J. Lowell and A. C. Rose-Innes, Advances in Physics 29, 947 (1980), ISSN 0001-8732. * Annadi et al. (2013) A. Annadi, Q. Zhang, X. Renshaw Wang, N. Tuzla, K. Gopinadhan, W. M. Lü, A. Roy Barman, Z. Q. Liu, A. Srivastava, S. Saha, et al., Nature Communications 4, 1838 (2013). * Wang et al. (2020) Y. Wang, J. Balgley, E. Gerber, M. Gray, N. Kumar, X. Lu, J.-Q. Yan, A. Fereidouni, R. Basnet, S. J. Yun, et al., Nano Lett. 20, 8446 (2020), ISSN 1530-6984. * Sutter et al. (2009) P. Sutter, J. T. Sadowski, and E. Sutter, Phys. Rev. B 80, 245411 (2009). * Zheng et al. (2021a) X. Zheng, E. Gerber, J. Park, D. Werder, O. Kigner, E.-A. Kim, S. Xie, and D. G. Schlom, Applied Physics Letters 118, 093103 (2021a), ISSN 0003-6951. * Stradi et al. (2017) D. Stradi, L. Jelver, S. Smidstrup, and K. Stokbro, Journal of Physics: Condensed Matter 29, 185901 (2017), ISSN 0953-8984. * Lazić (2015) P. Lazić, Computer Physics Communications 197, 324 (2015), ISSN 0010-4655. * Kim et al. (2015) H.-S. Kim, V. S. V., A. Catuneanu, and H.-Y. Kee, Phys. Rev. B 91, 241110 (2015). * Manzeli et al. (2017) S. Manzeli, D. Ovchinnikov, D. Pasquier, O. V. Yazyev, and A. Kis, Nature Reviews Materials 2, 17033 (2017). * Hsu et al. (2017) Y.-T. Hsu, A. Vaezi, M. H. Fischer, and E.-A. Kim, Nature Communications 8, 14985 (2017). * Zheng et al. (2021b) Y. Zheng, J. Gao, C. Han, and W. Chen, Cell Reports Physical Science 2, 100298 (2021b), ISSN 2666-3864. * Mao et al. (2020) J. Mao, S. P. Milovanović, M. Andjelković, X. Lai, Y. Cao, K. Watanabe, T. Taniguchi, L. Covaci, F. M. Peeters, A. K. Geim, et al., Nature 584, 215 (2020). * Shabani et al. (2021) S. Shabani, D. Halbertal, W. Wu, M. Chen, S. Liu, J. Hone, W. Yao, D. N. Basov, X. Zhu, and A. N. Pasupathy, Nature Physics 17, 720 (2021). * Rizzo et al. (2022) D. J. Rizzo, S. Shabani, B. S. Jessen, J. Zhang, A. S. McLeod, C. Rubio-Verdú, F. L. Ruta, M. Cothrine, J. Yan, D. G. Mandrus, et al., Nano letters 22, 1946 (2022). * Gerber and Torrisi (2022) E. Gerber and S. B. Torrisi, _Basic intermatch scripts_ , Zenodo (2022). * Plumb et al. (2014) K. W. Plumb, J. P. Clancy, L. J. Sandilands, V. V. Shankar, Y. F. Hu, K. S. Burch, H.-Y. Kee, and Y.-J. Kim, Physical Review B 90, 041112 (2014). * Ong et al. (2013) S. P. Ong, W. D. Richards, A. Jain, G. Hautier, M. Kocher, S. Cholia, D. Gunter, V. L. Chevrier, K. A. Persson, and G. Ceder, Computational Materials Science 68, 314 (2013), ISSN 0927-0256. * Perdew et al. (1996) J. P. Perdew, K. Burke, and M. Ernzerhof, Phys. Rev. Lett. 77, 3865 (1996). * Schlipf and Gygi (2015) M. Schlipf and F. Gygi, Computer Physics Communications 196, 36 (2015), ISSN 0010-4655. * Freysoldt et al. (2009) C. Freysoldt, S. Boeck, and J. Neugebauer, Phys. Rev. B 79, 241103 (2009). * Sundararaman et al. (2017) R. Sundararaman, K. Letchworth-Weaver, K. A. Schwarz, D. Gunceler, Y. Ozhabes, and T. Arias, SoftwareX 6, 278 (2017), ISSN 2352-7110. * Payne et al. (1992) M. C. Payne, M. P. Teter, D. C. Allan, T. A. Arias, and J. D. Joannopoulos, Rev. Mod. Phys. 64, 1045 (1992).
# Computationally Efficient Reinforcement Learning: Targeted Exploration leveraging Simple Rules Loris Di Natale, Bratislav Svetozarevic, Philipp Heer, and Colin N. Jones This research was supported by the Swiss National Science Foundation under NCCR Automation, grant agreement 51NF40_180545, and in part by the Swiss Data Science Center, grant agreement C20-13.L. Di Natale, B. Svetozarevic, and P. Heer are with the Urban Energy Systems Lab, Empa, Dübendorf, Switzerland. {loris.dinatale, bratislav.svetozarevic<EMAIL_ADDRESS>Di Natale and C. N. Jones are with the Laboratoire d’Automatique, EPFL, Lausanne, Switzerland<EMAIL_ADDRESS> ###### Abstract Model-free Reinforcement Learning (RL) generally suffers from poor sample complexity, mostly due to the need to exhaustively explore the state-action space to find well-performing policies. On the other hand, we postulate that expert knowledge of the system often allows us to design simple rules we expect good policies to follow at all times. In this work, we hence propose a simple yet effective modification of continuous actor-critic frameworks to incorporate such rules and avoid regions of the state-action space that are known to be suboptimal, thereby significantly accelerating the convergence of RL agents. Concretely, we saturate the actions chosen by the agent if they do not comply with our intuition and, critically, modify the gradient update step of the policy to ensure the learning process is not affected by the saturation step. On a room temperature control case study, it allows agents to converge to well-performing policies up to $6-7\times$ faster than classical agents without computational overhead and while retaining good final performance. ## I Introduction Despite its success in many applications [1, 2], model-free Reinforcement Learning (RL), and in particular deep RL (DRL), usually suffer from data inefficiency, i.e., they require a significant number of interactions with the environment to converge [3, 4]. This stems from the necessity to explore the state-action space to find optimal policies and leads to significant computational costs. It also limits the deployment of DRL methods on physical systems without pretraining in simulation: learning a building temperature control policy from scratch can for example take years of data [5, 6]. To speed up the training of DRL agents, researchers have for example investigated how to leverage expert demonstrations [7, 8], but this requires access to an expert policy which is not always available in practice. Instead, we postulate that prior knowledge of physical systems often allows us to design simple rules that agents should follow a priori, such as “Do not heat the room if it is already $26\text{\,}\mathrm{\SIUnitSymbolCelsius}$”; we indeed know this action will always be suboptimal in that state, there is no need for agents to explore its consequences. In this paper, we hence propose modifications of actor-critic algorithms to encode simple rules in RL agents, introducing state-dependent constraints on the agents’ actions to restrict exploration to interesting regions of the state-action space. In other words, the key idea is to avoid visiting state- action pairs that are known to be suboptimal by the expert to accelerate the convergence towards meaningful solutions and thus increase the efficiency of (D)RL. Note that while state-dependent bounds were concurrently introduced in [9], they are enforced a posteriori in the environment instead of directly modifying the agent’s behavior and do not necessarily improve convergence. In another line of work, prior knowledge successfully accelerated learning in [10], but relying on fuzzy rather than direct rule integration. ### I-A Constraining RL agents To bound the decisions taken by RL agents, one typically defines some constrained set of actions and either project the actions of the agents on this set at each time step or switch to a fallback controller when needed [11, 12, 13, 14]. The main challenge with these operations is that they are usually not differentiable and hence cannot be learned by RL agents, with the notable exceptions of [15, 16, 17], who leveraged differentiable optimization layers [15], modified the policy updates to account for projections [16], or derived a closed-form solution of the projection step [17]. However, these methods either entail additional computational burden [15, 16] or rely on a learned linearization of the constraints [17]. Alternatively, one could also apply tools from the safe RL literature [18, 19, 20], typically relying on constrained policy optimization [21, 22]. However, this would again introduce both engineering and computational overhead. The complexity of the methods discussed above often stems from the fact that they are designed to impose state constraints on DRL agents, which is a more challenging problem in general since it leads to complex action bounds for the agent at each step. Here, however, we argue that prior knowledge can straightforwardly be used to accelerate the training of DRL agents through simple state-dependent box constraints on their actions, which allows us to leverage less computationally expensive tools. To alleviate the issue of non-differentiability without increasing either the engineering or the computational burden, Reward Shaping (RS) heuristics might be used in various forms to penalize agents when constraints are violated, let them know when a fallback controller was used or they were saturated, or introduce prior knowledge about the task to solve [14, 23, 24]. While such methods might accelerate the learning process to some extent, they are however indirect, i.e., they only influence the learned policies through the reward function that the agent will learn to optimize over time. Moreover, shaping the reward function simultaneously impacts the learning process of both the actor and the critic. ### I-B Contribution In this work, we propose to clip DRL agents’ actions according to simple expert-designed state-dependent bounds and then modify the gradient update step of the actor to let agents learn from their mistakes and accelerate their convergence to expected actions. To explain the effectiveness of the gradient modification, we provide an intuitive analytical analysis of its impact on the learning of DRL agents. Importantly, contrary to RS, our modifications only affect the actor, allowing the critic to learn the true Q-values. Remarkably, our method bypasses the need for complex projection steps and does not require access to a fallback controller or an expert policy. Moreover and critically, the proposed modifications do not impact the computational complexity of the algorithm, are straightforward to design and implement, and can be coupled with any actor-critic algorithms. Note that, in contrast with [10], where the expert knowledge is potentially overridden by the policy, our method enforces the wanted behaviors on agents at all times. The effectiveness of the proposed Efficient Agents (EAs) is demonstrated in simulation on a room temperature control case study, where they converge up to $6$ – $7$ times faster than classical ones and $2$ – $3$ times faster than RS- based agents while retaining good final performance. This hints at how the proposed modifications can provide a simple yet effective and computationally inexpensive mean to leverage expert knowledge to accelerate DRL algorithms. ## II Preliminaries ### II-A Reinforcement Learning At each time step $t$, given an observation $s_{t}$ of the state of the environment, an RL agent chooses an action $a_{t}$. The environment then transitions to $s_{t+1}$ according to the transition probabilities $P(s_{t},a_{t})$ and sends the new state and the reward signal $r(s_{t},a_{t})$ to the agent. The objective of any deterministic RL algorithm111While the presented analyses deal with deterministic actor-critic agents for clarity, the results can easily be extended to the stochastic case. is to find a policy $\pi(s_{t})$ that maximizes the expected discounted cumulative returns: $\displaystyle J(\pi)=\mathbb{E}_{a_{t}\sim\pi(s_{t}),s_{t+1}\sim P(s_{t},a_{t})}\left[\sum_{t=0}^{\infty}{\gamma^{t}r(s_{t},a_{t})}\right],$ (1) where $\gamma$ is the discount factor trading off near- and long-term rewards, and the initial state $s_{0}\sim\rho$ is sampled from the corresponding initial distribution. With a slight abuse of notation on the expectation for clarity, we can define the Q-function of any state-action pair $(s,a)$: $\displaystyle Q^{\pi}(s,a)=\mathbb{E}_{\pi}\left[\sum_{t=0}^{\infty}{\gamma^{t}r(s_{t},a_{t})|s_{0}=s,a_{0}=a}\right],$ (2) which captures the expected returns when action $a$ is chosen in state $s$ and the policy $\pi$ is followed thereafter. In our simulations, we let agents explore the environment with the $\epsilon$-greedy exploration strategy, which means we apply the following action to the environment: $a(s)=\text{clip}(\pi(s)+\epsilon,a^{\textit{low}},a^{\textit{up}}),\qquad\epsilon\sim\mathcal{N}(0,\sigma),$ (3) where the noisy actions are clipped elementwise between $a^{\textit{low}}$ and $a^{\textit{up}}$, the predefined action bounds from the environment, and $\epsilon$ is the Gaussian exploration noise with a standard deviation of $\sigma$. All the transition tuples $(s,a,r,s^{\prime})$ observed by the agent are stored in a replay buffer. ### II-B Actor-critic algorithms In practice, policies and Q-functions are often parametrized with Neural Networks (NNs) as $\pi_{\theta}$ and $Q_{\phi}$, respectively, leading to DRL, and numerous algorithms have been developed to maximize (1) [25]. In this work, we are interested in deterministic actor-critic methods stemming from [26], where both the actor $\pi_{\theta}$ (also referred to as the policy) and the critic $Q_{\phi}$ are optimized in parallel leveraging gradient descent. While different flavors exist, most algorithms compute the gradient of the critic using the Temporal Difference (TD) loss [26]: $\displaystyle\hat{\nabla}_{\phi}Q_{\phi}$ $\displaystyle=\nabla_{\phi}\left[\frac{1}{|B|}\sum_{b=(s,a,r,s^{\prime})\in B}{\left(Q_{\phi}(s,a)-y(b)\right)^{2}}\right],$ (4) with $y(b)=\left(r+\gamma\max_{a^{\prime}}Q_{\phi}(s^{\prime},a^{\prime})\right)$ and where a batch $B$ of past transitions is sampled from the replay buffer and used to estimate expectations. Leveraging the policy gradient theorem [27], one can similarly use the critic to estimate the actor gradient as: $\displaystyle{\color[rgb]{0,0,0}\hat{\nabla}_{\theta}\pi_{\theta}}$ $\displaystyle=-{\color[rgb]{0,0,0}\nabla_{\theta}\left[\frac{1}{|B|}\sum_{s\in B}{Q_{\phi}(s,\pi_{\theta}(s))}\right].}$ (5) Note that these gradients are easily computed using automatic differentiation when the actor and the critic are parametrized with NNs. In this paper, we rely on the Twin Delayed Deep Deterministic (TD3) policy gradient algorithm, which introduces a few modifications to limit the well- known overestimation bias of Q-functions plaguing vanilla actor-critic algorithms [28]. Remarkably, however, these adjustments do not impact the actor gradient in (5), allowing us to seamlessly integrate the proposed modifications detailed in Section III. ## III Methods ### III-A State-dependent action saturation In many cases, prior knowledge allows us to design state-dependent upper and lower bounds $a^{\textit{max}}(s)$ and $a^{\textit{min}}(s)$, respectively, on the actions we expect well-performing control policies to take in a given state $s$, with: $\displaystyle\quad a^{\textit{low}}\leq a^{\textit{min}}(s)\leq a^{\textit{max}}(s)\leq a^{\textit{up}}.$ (6) To limit the exploration of known suboptimal state-action pairs, we can then modify (3) accordingly to: $\displaystyle a(s)$ $\displaystyle=\text{clip}(\pi_{\theta}(s)+\epsilon,a^{\textit{min}}(s),a^{\textit{max}}(s)).$ (7) Note that these bounds stemming from prior knowledge are also enforced at test time when $\epsilon=0$ to ensure an agent would never turn left if whenever there is a wall in that direction, for example, neither during the training nor the deployment phase. ### III-B Actor gradient modification The major problem with the clipping operation in (7) is its non- differentiability. Worse yet, its subdifferentials go to zero whenever agents are saturated (see (9) in Section III-C), making any backward flow of information on the overriding process impossible. As a countermeasure, to let agents learn from their mistakes, we also modify the actor gradient (5) to: $\displaystyle\hat{\nabla}^{\textit{EA}}_{\theta}\pi_{\theta}=-\nabla_{\theta}\Bigg{(}$ $\displaystyle\frac{1}{|B|}\sum_{(s,a)\in B}\bigg{[}{Q_{\phi}(s,\pi_{\theta}(s))}$ $\displaystyle-\frac{\lambda}{2}\left(\pi_{\theta}(s)-a(s)\right)^{2}\bigg{]}\Bigg{)},$ (8) where $\lambda$ is a hyperparameter. The last term in (8) penalizes actions chosen by the policy $\pi_{\theta}(s)$ if they deviate from the constrained action $a(s)$ that was applied to the environment, thus steering the agent’s decisions towards expected actions.222In another line of work, this penalty was also used in [15] to improve the robustness of differentiable layer-based RL for state-constrained problems. Alternatively, we note here that one could instead modify the reward function to include this penalty (RS) and then maximize (1). Remarkably, however, the latter also impacts the learning process of the critic in (4) when applied to actor-critic frameworks, contrary to our method. We will show empirical benefits of the proposed modification (8) over RS-based penalties in terms of convergence speed in Section V. ### III-C Implications of the modified gradients Let $C(s)=\left\\{a\in\mathbb{R}:a^{\textit{min}}(s)\leq a\leq a^{\textit{max}}(s)\right\\}$ for any given state $s$.333Without loss of generality, we assume that $a\in\mathbb{R}$ in this section for clarity. This assumption can easily be lifted for multi-dimensional problems. Grouping all the parameters $\theta$ in a vector and recalling the definition of the action $a(s)$ applied to the environment in state $s$ from (7), we can define its subgradient $\nabla_{\theta}a(s)$ as: $\displaystyle a(s)$ $\displaystyle=\begin{cases}a^{min}(s),&\text{if }\pi(s)<a^{min}(s),\\\ \pi_{\theta}(s)+\epsilon,&\text{if }\pi_{\theta}(s)\in C(s),\\\ a^{max}(s),&\text{if }\pi(s)>a^{max}(s).\end{cases}$ $\displaystyle\implies\nabla_{\theta}a(s)$ $\displaystyle=\begin{cases}\nabla_{\theta}\pi_{\theta}(s),&\text{if }\pi_{\theta}(s)\in C(s),\\\ 0,&\text{else,}\end{cases}$ (9) where $\nabla_{\theta}\pi_{\theta}(s)$ is the actor gradient. We can then rewrite the gradient of EAs (8) as: $\displaystyle\hat{\nabla}^{\textit{EA}}_{\theta}\pi_{\theta}$ $\displaystyle=-\frac{1}{|B|}\sum_{(s,a)\in B}\bigg{[}{\nabla_{\theta}Q_{\phi}(s,\pi_{\theta}(s))}$ $\displaystyle\qquad\quad-\nabla_{\theta}\left(\frac{\lambda}{2}\left(\pi_{\theta}(s)-a(s)\right)^{2}\right)\bigg{]}$ $\displaystyle=-\frac{1}{|B|}\sum_{(s,a)\in B}\bigg{[}\nabla_{\theta}Q_{\phi}(s,\pi_{\theta}(s))$ $\displaystyle\qquad\quad-\left(\lambda\left(\nabla_{\theta}\pi_{\theta}(s)-\nabla_{\theta}a(s)\right)^{\top}\left(e_{\theta}(s)\right)\right)\bigg{]},$ where we introduce the error term $e_{\theta}(s)=\pi_{\theta}(s)-a(s)$. We hence get the following modified actor gradient, where we omit $(s,a)\in B$ for clarity: $\displaystyle\hat{\nabla}^{\textit{EA}}_{\theta}\pi_{\theta}$ $\displaystyle=\begin{cases}-\frac{1}{|B|}\sum_{B}\Big{[}\nabla_{\theta}Q_{\phi}(s,\pi_{\theta}(s))\Big{]},\ \text{if }\pi_{\theta}(s)\in C(s),\\\ -\frac{1}{|B|}\sum_{B}\Big{[}\nabla_{\theta}Q_{\phi}(s,\pi_{\theta}(s))\\\ \qquad\quad\ -\lambda\nabla_{\theta}\pi_{\theta}(s)^{\top}e_{\theta}(s)\Big{]},\ \text{else.}\end{cases}$ Remarkably, the additional penalty term in (8) hence allows us to solve the issue of the subdifferentials of the clipping operator being zero when actions are saturated, modifying the gradients only when the constraints are not met. Indeed, as long as the action chosen by the agent respects the constraints provided by the expert, the classical gradient (5) is used. On the other hand, as soon as the constraints are not met, the gradient is modified in the direction $e_{\theta}(s)$ to accelerate the convergence of $\pi_{\theta}(s)$ to $C{\color[rgb]{0,0,0}(s)}$ despite the subdifferential of the clipped action being zero, confirming the graphical intuition from [15, Fig. 2]. This allows EAs to learn from their mistakes and — we hypothesize — helps them rapidly converge to meaningful policies. ## IV Room temperature control case study To assess the effectiveness of the proposed method, we apply it to a temperature control case study, where the objective is to minimize the energy consumption of a room while maintaining the comfort of the occupants, represented by predefined temperature bounds that should not be exceeded. ### IV-A Reinforcement Learning framework The continuous action space of the agents corresponds to how much heating or cooling power, should be applied at each time step, normalized between $a^{\textit{low}}=-1$ and $a^{\textit{up}}=1$. During the heating season, $a^{\textit{low}}$ corresponds to the heating being turned off and $a^{\textit{up}}$ to full heating, and the contrary in the cooling case. Physically Consistent Neural Networks (PCNNs) [29] are used to simulate one bedroom in the NEST building [30] and $s_{t}$ gathers time, weather, temperature, and comfort bound information (see [6] for details). The reward function is defined as the negative weighted sum of energy consumption $E_{t}$ and comfort violations, i.e. how far from the designed bounds the temperature inside the room is: $\displaystyle r(s_{t},a_{t})$ $\displaystyle=-\max{\\{L_{t}-T_{t},T_{t}-U_{t},0\\}}-\alpha E_{t},$ (10) $\displaystyle E_{t}$ $\displaystyle=\begin{cases}\frac{a_{t}+1}{2}\ E^{\textit{max}}_{\textit{heat}},&\text{in the heating season},\\\ \frac{1-a_{t}}{2}\ E^{\textit{max}}_{\textit{cool}},&\text{in the cooling season}.\end{cases}$ where $L_{t}$ and $U_{t}$ represent the lower and upper comfort bounds on the temperature $T_{t}$ at time $t$, respectively, $\alpha$ is a weighting factor, and $E^{\textit{max}}_{\textit{heat}}$ and $E^{\textit{max}}_{\textit{cool}}$ stand for the maximal heating and cooling power, respectively. ### IV-B Design of the saturation rules In the context of room temperature control, we intuitively know that an optimal policy should gradually stop heating when the temperature reaches the upper comfort bound and gradually start heating as soon as the lower bound is not met (and vice versa for cooling). To encode these simple rules, we design state-dependent action bounds as follows: $\displaystyle a^{\textit{min}}(s_{t})$ $\displaystyle=\text{clip}\left(\frac{(L_{t}-m)-T_{t}}{n-m},\ 0,\ 1\right)^{2}*2-1$ (11) $\displaystyle a^{\textit{max}}(s_{t})$ $\displaystyle=1-2*\text{clip}\left(\frac{T_{t}-(U_{t}+m)}{n-m},\ 0,\ 1\right)^{2},$ (12) with $n\geq m\geq 0$ representing design parameters to leave more or less freedom to the agents. In words, we start constraining the action of the agents as soon as the temperature deviates from the bounds for more than $m$ degrees and then quadratically increase the constraint until $n$ degrees have been reached, where the agent is forced to use the maximum or minimum power, as pictured in Fig. 1. Figure 1: Representation of the action bounds used in this work. ## V Results To investigate the influence of $m$ and $n$, which measure how much prior knowledge is transmitted to DRL agents, we train different EAs (EA $m$ / $n$). For comparison purposes, we also train agents with the classical actor gradient (5), introducing the additional squared penalty $\frac{\lambda}{2}\left(\pi_{\theta}(s)-a(s)\right)^{2}$ in the reward function instead as another computationally inexpensive means to incorporate prior knowledge in agents (RS $m$ / $n$). Finally, we also analyze two classical DRL agents with different random seeds (Classical $1$ and $2$).444The code and data are available on https://gitlab.nccr- automation.ch/loris.dinatale/efficient-drl. ### V-A Final performance All the agents were trained on up to three-day-long episodes randomly sampled from three years of data. They were evaluated after each $96$ steps of $15\text{\,}\mathrm{min}$, i.e. one day’s worth of data, hereafter also referred to as an epoch, on a testing set of $50$ unseen sequences of three days. They all use the same hyperparameters as in [6]. We manually set $\lambda=100$ for EAs to ensure the constraints are enforced as fast as possible and $\lambda=10$ for RSs since larger penalties led to instability. While we empirically observed more robust performance of EAs with respect to $\lambda$ compared to RSs, a complete sensitivity analysis is left for future work. The best reward obtained by all the trained agents over the first $500$ epochs can be found in Table I, and the corresponding trade-off between energy consumption and comfort violations is plotted in Fig. 2. These results illustrate how tighter parameters $m$ and $n$, i.e., higher levels of prior knowledge, allow EAs and RSs to converge to better solutions in this limited training regime. In particular, it allows EAs to reduce the amount of comfort violations without significantly increasing the energy consumption. Classical DRL agents on the other hand usually use less energy at the cost of additional comfort violations in this early phase of learning before converging to near- optimal solutions after longer training times [6]. TABLE I: Best reward obtained by each agent on the test set over the first $500$ epochs. Agent | Rew. | Agent | Rew. | Agent | Rew. ---|---|---|---|---|--- Classical 1 | -2.64 | Classical 2 | -2.75 | | RS $0$ / $1$ | -2.58 | EA $0$ / $1$ | -2.85 | EA $0.5$ / $1$ | -2.83 RS $0$ / $0.5$ | -2.44 | EA $0$ / $0.5$ | -2.58 | EA $0.25$ / $0.5$ | -2.74 RS $0$ / $0.25$ | -2.37 | EA $0$ / $0.25$ | -2.51 | EA $0.2$ / $0.25$ | -2.62 RS $0$ / $0.1$ | -3.69 | EA $0$ / $0.1$ | -2.46 | EA $0.075$ / $0.1$ | -2.46 Figure 2: Average energy consumption (En.) and comfort violations (Vio.) over the test set corresponding to each agent in Table I. The performance of two industrial baselines, of an agent trained for 125,000 epochs (Best Agent), and the optimal performance achievable (Optimum), all computed as in [6], are reported in gray. ### V-B Visualization of the impact of prior knowledge Figure 3: Behavior of a classical agent and EAs with various $m$ and $n$ parameters (ours) minimizing the heating power consumption (bottom) while maintaining the temperature in the grey dotted bounds (top). Left: Performance before training, where EAs are saturated once they exceed the bounds by $n$ degrees. Right: Performance after training, showing how all agents converged to similar solutions (Table II). To intuitively understand the effect of action saturation, we visualize its impact on some EAs in Fig. 3, where the behavior of all agents is plotted before training on the left, and after on the right, for the same three days during the heating season. Focusing on the left plot, we see the untrained classical DRL agent in black letting the temperature diverge to an uncomfortably high range (out of the bounds of the plot) as it starts exploring the state space using roughly constant heating power. On the other hand, all the EAs are forced to stop heating once they are $n$ degrees out of bounds. Consequently, even before training, such agents will not overheat the room and keep it at acceptable temperatures for the occupants, corresponding to what we expect from good controllers. However, note that EAs can present control input oscillations due to the impact of external disturbances, mainly the solar gains around noon, triggering the saturation mechanism on and off. On the right plot, after training, one can observe that all EAs generally take comparable decisions — still being sometimes saturated, which ensures compliance with prior expert knowledge — leading to similar temperature patterns. On the other hand, the classical agent presents a slightly different behavior, with smoother decision patterns. Interestingly, this agent is the only one heating in the early afternoon, while the EAs wait until the end of the afternoon to heat the room with high power and meet the comfort bound tightening at $8$pm. This allows the classical agent to use less energy than EAs over these three days but can incur additional comfort violations (Table II), as expected from Fig. 2. TABLE II: Reward, sum of comfort violations (Vio.), and energy consumption (En.) of each agent over the three days depicted on the right of Fig. 3. Agent | Classical 1 | EA (ours) ---|---|--- $m$ / $n$ | - | $0$ / $1$ | $0$ / $0.5$ | $0$ / $0.25$ Reward | -0.68 | -0.69 | -0.89 | -0.60 Vio. [Kh] | 1.28 | 1.18 | 2.23 | 0.65 En. [kWh] | 5.03 | 5.38 | 5.54 | 5.46 ### V-C Data efficiency of the proposed gradient modification Figure 4: Convergence speed of various EAs with different $m$ and $n$ parameters (ours), compared to agents using RS and two classical agents in black (one in the top plots, one in the bottom ones). The vertical lines and annotations specify the number of days of data required to obtain a reward of $-2.95$ for each agent, which corresponds to the performance of two industrial rule-based baselines. A comparison of the convergence speed of various agents over the first $300$ epochs is plotted in Fig. 4, where the vertical lines and annotations illustrate the number of days required to attain performance on par with rule- based on-off industrial baselines from [6]. In general, we observe that all the EAs attain returns on par with the baselines significantly earlier than classical DRL agents, in as little as $29$ days instead of roughly $200$, an improvement of almost an order of magnitude. In particular, the smaller $n$ is chosen (from left to right in Fig. 4), the faster the convergence of the EAs in green and blue. Intuitively, this makes sense, as tighter constraints introduce more prior knowledge to the EAs, thereby allowing them to find interesting solutions faster, without losing time exploring suboptimal state- action pairs. On the other hand, the influence of $m$ is less marked, with $m\neq 0$ (blue) and $m=0$ (green) leading to very similar convergence patterns in the bottom row of plots in Fig. 4. Remarkably, RS does not seem to drastically speed the training up in this case study (red). While RS $0$ / $0.25$ does converge twice as fast as the classical DRL agents, RS $0$ / $0.1$ does not converge at all, hinting at the fragility of RS in general. Even when they converge, RSs are still two to three times slower than their EA counterparts. On the other hand, RS seems to lead to more consistent performance than classical agents and EAs after a few hundred epochs, which is confirmed by their good final performance in Fig. 2. Overall, these results support our claim that, as long as the rules provided to the agents are well-defined and correspond to expected behaviors, the proposed modifications can indeed greatly accelerate the convergence of DRL agents. Critically, this does not significantly impact the quality of the final solution (Sec. V-A). Interestingly, incorporating more specific expert knowledge in EAs — through smaller $m$ and $n$ — further accelerates their convergence. This corresponds to our intuition: better-defined rules help agents more. Remarkably, the modifications proposed in Sec. III provide the desired speedup for a wide variety of parameters $m$ and $n$, contrary to RS, hinting at the robustness of the proposed scheme. ## VI Conclusion Starting from the postulate that prior expert knowledge often gives us an intuition of how good control policies should behave, we presented a scheme to encode it in actor-critic frameworks through simple rules to accelerate learning and decrease the associated computational load. These rules take the form of bounds on the agent’s actions that can directly be enforced during training and online operations. To ensure agents learn from their mistakes, we also modified the actor gradients to steer control policies towards expected behaviors, limiting the exploration of known suboptimal state-action pairs. Critically, both these operations are computationally inexpensive, ensuring the gains in sample complexity positively impact the training time of the agents. On a room temperature control case study, this scheme allowed us to accelerate the convergence speed of DRL agents by up to $6$ – $7\times$. Furthermore, modifying actor gradients proved to be $2$ – $3$ times more effective and more robust than the widespread reward shaping method. Remarkably, this was done without suffering from a significant drop in the final performance of the control policies, illustrating how prior knowledge can help alleviate the computational burden of DRL. This represents an interesting first step towards efficient agents that can be deployed and learned from scratch on physical systems, potentially bypassing the need for complex simulators. In future work, it would be interesting to investigate annealing strategies on $\lambda$ or leverage primal-dual optimization tools to adaptively tune the influence of the additional penalty in the actor gradient and let agents learn more expressive policies after the initial exploration phase. ## References * [1] A. Coronato, M. Naeem, G. De Pietro, and G. Paragliola, “Reinforcement learning for intelligent healthcare applications: A survey,” _Artificial Intelligence in Medicine_ , vol. 109, p. 101964, 2020. * [2] Z. Zhang, D. Zhang, and R. C. Qiu, “Deep reinforcement learning for power system applications: An overview,” _CSEE Journal of Power and Energy Systems_ , vol. 6, no. 1, pp. 213–225, 2019. * [3] Y. Duan, J. Schulman, X. Chen, P. L. Bartlett, I. Sutskever, and P. Abbeel, “RL2: Fast reinforcement learning via slow reinforcement learning,” _arXiv preprint arXiv:1611.02779_ , 2016. * [4] M. Schwarzer, N. Rajkumar, M. Noukhovitch, A. Anand, L. Charlin, R. D. Hjelm, P. Bachman, and A. C. Courville, “Pretraining representations for data-efficient reinforcement learning,” _Advances in Neural Information Processing Systems_ , vol. 34, 2021. * [5] Z. Wang and T. Hong, “Reinforcement learning for building controls: The opportunities and challenges,” _Applied Energy_ , vol. 269, p. 115036, 2020\. * [6] L. Di Natale, B. Svetozarevic, P. Heer, and C. N. Jones, “Near-optimal Deep Reinforcement Learning Policies from Data for Zone Temperature Control,” _arXiv preprint arXiv:2203.05434_ , 2022. * [7] A. Nair, B. McGrew, M. Andrychowicz, W. Zaremba, and P. Abbeel, “Overcoming exploration in reinforcement learning with demonstrations,” in _2018 IEEE international conference on robotics and automation (ICRA)_. IEEE, 2018, pp. 6292–6299. * [8] T. Hester, M. Vecerik, O. Pietquin, M. Lanctot, T. Schaul, B. Piot, D. Horgan, J. Quan, A. Sendonaris, I. Osband _et al._ , “Deep q-learning from demonstrations,” in _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 32, no. 1, 2018. * [9] B. De Cooman, J. Suykens, and A. Ortseifen, “Enforcing Hard State-Dependent Action Bounds on Deep Reinforcement Learning Policies,” in _Machine Learning, Optimization, and Data Science: 8th International Workshop, LOD 2022, Certosa di Pontignano, Italy, September 19–22, 2022, Revised Selected Papers, Part II_. Springer, 2023, pp. 193–218. * [10] P. Zhang, J. Hao, W. Wang, H. Tang, Y. Ma, Y. Duan, and Y. Zheng, “KoGuN: accelerating deep reinforcement learning via integrating human suboptimal knowledge,” _arXiv preprint arXiv:2002.07418_ , 2020. * [11] R. Wang, X. Zhang, X. Zhou, Y. Wen, and R. Tan, “Toward Physics-Guided Safe Deep Reinforcement Learning for Green Data Center Cooling Control,” in _2022 ACM/IEEE 13th International Conference on Cyber-Physical Systems (ICCPS)_. IEEE, 2022, pp. 159–169. * [12] H. Mao, M. Schwarzkopf, H. He, and M. Alizadeh, “Towards safe online reinforcement learning in computer systems,” in _33rd conference on neural information processing systems (NeurIPS 2019)_ , 2019. * [13] R. Bautista-Montesano, R. Galluzzi, K. Ruan, Y. Fu, and X. Di, “Autonomous navigation at unsignalized intersections: A coupled reinforcement learning and model predictive control approach,” _Transportation Research Part C: Emerging Technologies_ , vol. 139, p. 103662, 2022. * [14] H. H. Goh, Y. Huang, C. S. Lim, D. Zhang, H. Liu, W. Dai, T. A. Kurniawan, and S. Rahman, “An Assessment of Multi-Stage Reward Function Design for Deep Reinforcement Learning-Based Microgrid Energy Management,” _IEEE Transactions on Smart Grid_ , 2022. * [15] B. Chen, P. L. Donti, K. Baker, J. Z. Kolter, and M. Bergés, “Enforcing policy feasibility constraints through differentiable projection for energy optimization,” in _Proceedings of the Twelfth ACM International Conference on Future Energy Systems_ , 2021, pp. 199–210. * [16] S. Gros, M. Zanon, and A. Bemporad, “Safe reinforcement learning via projection on a safe set: How to achieve optimality?” _IFAC-PapersOnLine_ , vol. 53, no. 2, pp. 8076–8081, 2020. * [17] G. Dalal, K. Dvijotham, M. Vecerik, T. Hester, C. Paduraru, and Y. Tassa, “Safe exploration in continuous action spaces,” _arXiv preprint arXiv:1801.08757_ , 2018. * [18] P. Osinenko, D. Dobriborsci, and W. Aumer, “Reinforcement learning with guarantees: a review,” _IFAC-PapersOnLine_ , vol. 55, no. 15, pp. 123–128, 2022. * [19] S. Gu, L. Yang, Y. Du, G. Chen, F. Walter, J. Wang, Y. Yang, and A. Knoll, “A Review of Safe Reinforcement Learning: Methods, Theory and Applications,” _arXiv preprint arXiv:2205.10330_ , 2022. * [20] J. Garcıa and F. Fernández, “A comprehensive survey on safe reinforcement learning,” _Journal of Machine Learning Research_ , vol. 16, no. 1, pp. 1437–1480, 2015. * [21] T. D. Simão, N. Jansen, and M. T. Spaan, “AlwaysSafe: Reinforcement learning without safety constraint violations during training,” in _Proceedings of the 20th International Conference on Autonomous Agents and MultiAgent Systems_. International Foundation for Autonomous Agents and Multiagent Systems, 2021. * [22] J. Achiam, D. Held, A. Tamar, and P. Abbeel, “Constrained policy optimization,” in _International conference on machine learning_. PMLR, 2017, pp. 22–31. * [23] M. Alshiekh, R. Bloem, R. Ehlers, B. Könighofer, S. Niekum, and U. Topcu, “Safe reinforcement learning via shielding,” in _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 32, no. 1, 2018. * [24] Y. Hu, W. Wang, H. Jia, Y. Wang, Y. Chen, J. Hao, F. Wu, and C. Fan, “Learning to utilize shaping rewards: A new approach of reward shaping,” _Advances in Neural Information Processing Systems_ , vol. 33, pp. 15 931–15 941, 2020. * [25] OpenAI, “Part 2: Kinds of RL Algorithms,” 2018, accessed: 25.11.2022. [Online]. Available: https://spinningup.openai.com/en/latest/spinningup/rl_intro2.html#a-taxonomy-of-rl-algorithms * [26] T. P. Lillicrap, J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, D. Silver, and D. Wierstra, “Continuous control with deep reinforcement learning,” _arXiv preprint arXiv:1509.02971_ , 2015. * [27] D. Silver, G. Lever, N. Heess, T. Degris, D. Wierstra, and M. Riedmiller, “Deterministic policy gradient algorithms,” in _International conference on machine learning_. PMLR, 2014, pp. 387–395. * [28] S. Fujimoto, H. Hoof, and D. Meger, “Addressing function approximation error in actor-critic methods,” in _International Conference on Machine Learning_. PMLR, 2018, pp. 1587–1596. * [29] L. Di Natale, B. Svetozarevic, P. Heer, and C. N. Jones, “Physically consistent neural networks for building thermal modeling: theory and analysis,” _Applied Energy_ , vol. 325, p. 119806, 2022. * [30] Empa, “NEST,” 2022, accessed: 25.11.2022. [Online]. Available: https://www.empa.ch/web/nest/overview
# ACFlow: An open source toolkit for analytical continuation of quantum Monte Carlo data Li Huang Science and Technology on Surface Physics and Chemistry Laboratory, P.O. Box 9-35, Jiangyou 621908, China ###### Abstract The purpose of analytical continuation is to establish a real frequency spectral representation of single-particle or two-particle correlation function (such as Green’s function, self-energy function, and dynamical susceptibilities) from noisy data generated in finite temperature quantum Monte Carlo simulations. It requires numerical solutions of a family of Fredholm integral equations of the first kind, which is indeed a challenging task. In this paper, an open source toolkit (dubbed ACFlow) for analytical continuation of quantum Monte Carlo data is presented. We at first give a short introduction to the analytical continuation problem. Next, three primary analytical continuation algorithms, including maximum entropy method, stochastic analytical continuation, and stochastic optimization method, as implemented in this toolkit are reviewed. And then we elaborate major features, implementation details, and basic usage of this toolkit. Finally, four representative examples are shown to demonstrate usefulness and flexibility of the ACFlow toolkit. ###### keywords: Quantum Monte Carlo simulation , Analytical continuation problem , Maximum entropy method , Stochastic analytical continuation , Stochastic optimization method ††journal: Computer Physics Communications PROGRAM SUMMARY Program Title: ACFlow CPC Library link to program files: (to be added by Technical Editor) Developer’s repository link: https://github.com/huangli712/ACFlow Code Ocean capsule: (to be added by Technical Editor) Licensing provisions (please choose one): GPLv3 Programming language: Julia Supplementary material: Journal reference of previous version:* Does the new version supersede the previous version?:* Reasons for the new version:* Summary of revisions:* Nature of problem (approx. 50-250 words): Most of the quantum Monte Carlo methods work on imaginary axis. In order to extract physical observables and compare them with the experimental results, analytical continuation must be done in the post-processing stage to convert the quantum Monte Carlo simulated data from imaginary axis to real axis. Solution method (approx. 50-250 words): Three established analytical continuation methods, including maximum entropy method, stochastic analytical continuation, and stochastic optimization method, have been implemented in the ACFlow toolkit. Additional comments including restrictions and unusual features (approx. 50-250 words): The ACFlow toolkit is written by pure Julia language. It is highly optimized and parallelized. It can be executed interactively in a Jupyter notebook environment. ## 1 Introduction It is well-known that quantum Monte Carlo (QMC) method is a powerful and exact numerical approach, and has been widely used in many research fields, such as nuclear physics [1], condense matter physics [2], and many-body physics [3]. In this paper, we just focus on the finite temperature QMC algorithms, which are used to solve the interacting lattice models or quantum impurity models [4]. Generally speaking, the simulated results of QMC methods are some sorts of single-particle or two-particle correlation functions, which are usually defined on imaginary time axis ($\tau\equiv-it$) or Matsubara frequency axis ($i\omega_{n}$). Therefore, they can’t be compared directly with the correspondingly experimental results, including but not limited to the electronic density of states $A(\omega)$, optical conductivity $\sigma(\omega)$, dynamical structure factor $S(\mathbf{q},\omega)$, and so on. It is necessary to convert the QMC simulated results from imaginary time axis or Matsubara frequency axis to real axis (i.e. $\tau\to\omega$ or $i\omega_{n}\to\omega$), which is the origin of the analytical continuation problem. Let’s concentrate on the following Fredholm integral equation of the first kind: $g(y)=\int K(y,x)f(x)~{}dx.$ (1) Here, $K(y,x)$ is the known kernel function, $f(x)$ is the model function, and $g(y)$ denotes the raw data. Given $f(x)$, it is quite easy to get $g(y)$ via numerical integration. However, given $g(y)$, solving the Fredholm integral equation reversely to get $f(x)$ is not as easy as expected. There is no universal solution. Notice that the so-called analytical continuation problem can be reformulated in terms of the Fredholm integral equation. Thus, its objective is to seek a reasonable $f(x)$ to satisfy the above equation. The QMC simulated data $g(y)$ are noisy and the kernel function $K(y,x)$ is ill conditioned, which make analytical continuation of QMC simulated data a huge challenge. In order to solve this problem, peoples have developed numerous methods in the past decades. These methods include the least square fitting method, singular value decomposition [5, 6], Padé approximation [7, 8, 9, 10], Tikhonov-Philips regularization method, maximum entropy method [11, 12], stochastic analytical continuation [13, 14], stochastic optimization method [15, 16], sparse modelling method [17], and machine learning method [18, 19, 20], etc. However, each method has its pros and cons. None of these methods can override the others. The analytical continuation problem is still far away from being completely solved. In recent years, quite a few analytical continuation codes have been released, including maxent (by Mark Jarrell) [12], maxent (in ALPSCore) [21], $\Omega$Maxent [22], ana_cont [23], SOM (in TRIQS) [24, 25], Stoch (in ALF) [26], just to name a few. We note that the maximum entropy method has dominated this field for quite a long time. Thus most of these codes only implement the maximum entropy method [12, 21, 22, 23]. It is rather difficult to crosscheck the simulated results obtained by various analytical continuation methods. In addition, the features of the available codes are quite limited and hard to examine new algorithms. In order to fill in this gap, we would like to present a new open source toolkit, called ACFlow, for analytical continuation. This toolkit implements three primary analytical continuation methods, including the maximum entropy method, stochastic analytical continuation, and stochastic optimization method, within an united framework. It provides an easy-to-used library and application interface. Some diagnostic and analytical tools are also available. With ACFlow, the users can easily setup and execute analytical continuation calculations, and validate the obtained results. We believe that this toolkit will play a vital role in solving analytical continuation problems. The rest of this paper is organized as follows. In section 2, background of the analytical continuation problem is introduced. In section 3, basic principles and key ingredients of the three analytical continuation methods as implemented in the ACFlow toolkit are summarized. Section 4 gives a brief overview about ACFlow’s main features and structures. Section 5 is the major part of this paper, it explains basic usage, input and output files of ACFlow. In order to demonstrate usefulness of this toolkit, four typical examples are illustrated in section 6. Finally, section 7 serves as a short conclusion. ## 2 Problem ### 2.1 Finite temperature Green’s functions Under the Wick’s rotation $t\to i\tau$, the time evolution operator in the Heisenberg picture $e^{itH}$ will be replaced by $e^{-\tau H}$. Such a transformation will increase efficiency of QMC random walking and suppress numerical oscillation (when $t$ is large, the periodic oscillation of $e^{itH}$ is quite obvious). This is an important reason why most of the finite temperature QMC algorithms are formulated in imaginary time axis. The outputs of finite temperature QMC simulations are usually single-particle or two-particle correlation functions. For example, the single-particle Green’s function $G(\tau)$ is defined as follows: $G(\tau)=\langle\mathcal{T}_{\tau}d(\tau)d^{\dagger}(0)\rangle,$ (2) where $\tau$ denotes imaginary time, $\mathcal{T}_{\tau}$ denotes time-ordered operator, and $d$ and $d^{\dagger}$ are annihilation and creation operators, respectively. The Matsubara Green’s function $G(i\omega_{n})$ can be measured by QMC simulations or constructed from $G(\tau)$ via direct Fourier transformation: $G(i\omega_{n})=\int^{\beta}_{0}d\tau~{}e^{-i\omega_{n}\tau}G(\tau),$ (3) $G(\tau)=\frac{1}{\beta}\sum_{n}e^{i\omega_{n}\tau}G(i\omega_{n}).$ (4) Here, $\beta$ means the inverse temperature ($\beta\equiv 1/T$) and $\omega_{n}$ is the Matsubara frequency. Note that $\omega_{n}$ is equal to $(2n+1)\pi/\beta$ for fermions and $2n\pi/\beta$ for bosons ($n$ is an integer). ### 2.2 Spectral density Clearly, neither $G(\tau)$ nor $G(i\omega_{n})$ can be observed experimentally. We have to extract dynamical response function, i.e., the spectral density $A(\omega)$, from them. $A(\omega)$ is indeed an observable quantity. It is related to $G(\tau)$ via the following Laplace transformation: $G(\tau)=\int^{+\infty}_{-\infty}d\omega\frac{e^{-\tau\omega}}{1\pm e^{-\beta\omega}}A(\omega),$ (5) where +(-) in the denominator is for fermionic (bosonic) system. $G(i\omega_{n})$ and $A(\omega)$ manifest similar relation: $G(i\omega_{n})=\int^{+\infty}_{-\infty}d\omega\frac{A(\omega)}{i\omega_{n}-\omega}.$ (6) It is obvious that Eq. (5) and Eq. (6) are indeed two special forms of the Fredholm integral equation of the first kind [see Eq. (1)]. So, the central problem of analytical continuation is to search optimal $A(\omega)$ for given $G(\tau)$ or $G(i\omega_{n})$. Sometimes the spectral density $A(\omega)$ is called spectral function in the references. It is tied to the imaginary part of real frequency Green’s function $G(\omega)$: $A(\omega)=-\frac{1}{\pi}\rm{Im}G(\omega).$ (7) From Im$G(\omega)$, Re$G(\omega)$ could be calculated via the Kramers-Kronig transformation: $\mathrm{Re}G(\omega)=\frac{1}{\pi}\mathcal{P}\int_{-\infty}^{\infty}d\omega^{\prime}~{}\frac{\mathrm{Im}G(\omega^{\prime})}{\omega^{\prime}-\omega},$ (8) where $\mathcal{P}$ means Cauchy principal value. Besides Eq. (5) and Eq. (6), $A(\omega)$ has to obey some additional constraints or sum-rules. For fermionic systems, the spectral functions must be positive: $A(\omega)\geq 0.$ (9) While for bosonic systems, the constraint becomes: $\text{sign}(\omega)A(\omega)\geq 0.$ (10) In addition, the spectral function $A(\omega)$ is always bounded, $\int^{+\infty}_{-\infty}d\omega~{}A(\omega)<\infty.$ (11) It can be utilized to normalize the resulting spectral function. ### 2.3 Kernel functions Eq. (5) and Eq. (6) can be reformulated as follows: $G(\tau)=\int^{+\infty}_{-\infty}d\omega~{}K(\tau,\omega)A(\omega),$ (12) and $G(i\omega_{n})=\int^{+\infty}_{-\infty}d\omega~{}K(\omega_{n},\omega)A(\omega),$ (13) where $K(\tau,\omega)$ and $K(\omega_{n},\omega)$ are the so-called kernel functions. Their definitions are as follows: $K(\tau,\omega)=\frac{e^{-\tau\omega}}{1\pm e^{-\beta\omega}},$ (14) and $K(\omega_{n},\omega)=\frac{1}{i\omega_{n}-\omega},$ (15) where +(-) in the denominator of Eq. (14) stands for fermions (bosons). As mentioned above, the kernel function is quite strange. The values of $K(\tau,\omega)$ could change by tens of orders of magnitude. Especially, at large positive and negative frequencies, $K(\tau,\omega)$ is exponentially small. It implies that at large $|\omega|$ the features of $A(\omega)$ are sensitive to the fine structures of $G(\tau)$. However, the data of $G(\tau)$ provided by QMC simulations are always fluctuant and noisy [27]. Tiny deviations in $G(\tau)$ from its expected values can lead to enormous changes in $A(\omega)$. Thus, analytical continuation is often characterized as an ill-posed problem [12]. In principle, for incomplete and noise $G(\tau)$ or $G(i\omega_{n})$, the number of spectral functions $A(\omega)$ that satisfy Eq. (12) and Eq. (13) is infinite. So the question becomes which $A(\omega)$ should be chosen. Now there are two different strategies to solve this problem. The first one is to choose the most likely $A(\omega)$. The second one is to evaluate the average of all the candidate spectral functions. In next section, we will introduce three primary analytical continuation methods that follow the two strategies and have been implemented in the ACFlow toolkit. For the sake of simplicity, we will concentrate on analytical continuation of imaginary time Green’s functions in main text. ## 3 Methods ### 3.1 Maximum entropy method Perhaps the maximum entropy method is the most frequently used approach for analytical continuation problems in the last decades [11, 12] because of its high computational efficiency. Next, we will discuss the basic principle and several variants of it. #### 3.1.1 Bayesian inference Bayes’s theorem is the cornerstone of the maximum entropy method. Given two events $a$ and $b$, Bayes’s theorem says: $P[a|b]P[b]=P[b|a]P[a],$ (16) where $P[a]$ is the probability of event $a$, $P[a|b]$ is the conditional probability of event $a$ with given event $b$. In the scenario of analytical continuation problem, $\bar{G}(\tau)$ and $A(\omega)$ are treated as two events, where $\bar{G}(\tau)$ denotes the measured value of $G(\tau)$. So the best solution for $A(\omega)$ is of course the one that maximizes $P[A|\bar{G}]$, which is called the posterior probability. According to the Bayes’s theorem, we get $P[A|\bar{G}]=\frac{P[\bar{G}|A]P[A]}{P[\bar{G}]},$ (17) where $P[\bar{G}|A]$ is the likelihood function, $P[A]$ is the prior probability, and $P[\bar{G}]$ is the evidence. Since the evidence is a normalization constant depending on the prior probability and the likelihood function only, it is ignored in the following discussions. Thus, $P[A|\bar{G}]\propto P[\bar{G}|A]P[A].$ (18) #### 3.1.2 Posterior probability In the maximum entropy method, the likelihood function $P[\bar{G}|A]$ is assumed to be in direct proportion to $e^{-\chi^{2}/2}$. Here, $\chi^{2}$ is named as goodness-of-fit function. It measures the distance between $\bar{G}(\tau)$ and reconstructed imaginary time Green’s function $\tilde{G}(\tau)$: $\chi^{2}=\sum^{L}_{i=1}\left[\frac{\bar{G}_{i}(\tau)-\tilde{G}_{i}(\tau)}{\sigma_{i}}\right]^{2},$ (19) $\tilde{G}_{i}=\sum_{j}K_{ij}A_{j}.$ (20) Here, $L$ is number of imaginary time points, $\sigma$ denotes the error bar (standard deviation) of $\bar{G}(\tau)$. $K_{ij}$ and $A_{j}$ are discrete kernel and spectral functions, respectively. On the other hand, the prior probability $P[A]$ is supposed to be in direct proportion to $e^{\alpha S}$, where $\alpha$ is a regulation parameter and $S$ means entropy. Sometimes $S$ is also known as the Kullback-Leibler distance. Its formula is as follows: $S=\int d\omega\left(A(\omega)-m(\omega)-A(\omega)\log\left[\frac{A(\omega)}{m(\omega)}\right]\right),$ (21) where $m(\omega)$ is the default model function. According to the Bayes’s theorem, the posterior probability $P[A|\bar{G}]\propto e^{Q}$ and $Q=\alpha S-\frac{\chi^{2}}{2}.$ (22) #### 3.1.3 Algorithms of maximum entropy method Now the original analytical continuation problem becomes how to figure out the optimal $A(\omega)$ that maximizes $Q$. In other words, we have to solve the following equation: $\frac{\partial Q}{\partial A}\bigg{|}_{A=\hat{A}}=0,$ (23) where $\hat{A}(\omega)$ is the optimal $A(\omega)$. Eq. (23) can be easily solved by using standard Newton method. However, the obtained $\hat{A}(\omega)$ is $\alpha$-dependent. That is to say, for a given $\alpha$, there is always a $\hat{A}(\omega)$ that satisfies Eq. (23). So, new problem arises because we have to figure out a way to construct the final spectral function from these $\alpha$-resolved $\hat{A}(\omega)$. Now there exist four algorithms, namely “historic”, “classic”, “bryan”, and “$\chi^{2}$kink”. Next we will introduce them one by one. _Historic algorithm_. The historic algorithm is quite simple. The $\alpha$ parameter will be adjusted iteratively to meet the following criterion: $\chi^{2}=N,$ (24) where $N$ is the number of mesh points for spectral density $A(\omega)$. _Classic algorithm_. The basic equation for the classic algorithm reads: $-2\alpha S(A_{\alpha})=\text{Tr}\left[\frac{\Lambda(A_{\alpha})}{\alpha I+\Lambda(A_{\alpha})}\right],$ (25) where $I$ is an identity matrix. The elements of $\Lambda$ matrix are calculated as follows: $\Lambda_{ij}=\sqrt{A_{i}}\left(\sum_{kl}K_{ki}[C^{-1}]_{kl}K_{lj}\right)\sqrt{A_{j}},$ (26) where $C$ is the covariance matrix. Eq. (25) will be iteratively solved until the optimal $\alpha$ and $\hat{A}(\omega)$ are determined. _Bryan algorithm_. In both historic and classic algorithms, the spectral function $\hat{A}(\omega)$ is always related to an optimal $\alpha$ parameter. However, the spirit of the bryan algorithm [28] is completely different. It tries to generate a series of $\alpha$ parameters and yield the corresponding $A_{\alpha}(\omega)$. Then the final spectral function $A(\omega)$ is obtained by evaluating the following integration: $\overline{A(\omega)}=\int d\alpha~{}A_{\alpha}(\omega)P[\alpha|\bar{G}].$ (27) _$\chi^{2}$ kink algorithm_. This algorithm was proposed by Bergeron and Tremblay [22] recently. The first step is to generate a series of $\alpha$ parameters, and evaluate the corresponding spectral functions $A_{\alpha}(\omega)$ and the goodness-of-fit functions $\chi^{2}[A_{\alpha}]$. Then we plot $\log_{10}(\chi^{2})$ as a function of $\log_{10}(\alpha)$. Usually this plot is split into three different regions: (1) Default model region. In the limit of $\alpha\to\infty$, $\chi^{2}$ goes to a constant high value. It means that the likelihood function $e^{-\chi^{2}/2}$ has negligible weight, such that the prior probability $e^{\alpha S}$ becomes dominant and minimizes $Q[A]$. At that time, the calculated $A(\omega)$ resembles the default model function $m(\omega)$. (2) Noise-fitting region. In the limit of $\alpha\to 0$, $\chi^{2}$ is relatively flat and approaches its global minimum. In this region, the minimization algorithm tends to fit the noise in $G(\tau)$. (3) Information-fitting region. $\alpha S$ is comparable with $\chi^{2}/2$, so that $\chi^{2}$ is strongly dependent on $\alpha$. Bergeron _et al._ suggested that the optimal $\alpha$ parameter situates in the crossover between noise-fitting region and information-fitting region [22]. So the second derivative of $\chi^{2}$ with respect to $\alpha$ is calculated, and the maximum value in the resulting curve indicates the optimal value of $\alpha$. Quite recently, Kaufmann and Held proposed a more numerically stable and flexible approach to compute the optimal $\alpha$ [23]. They use the following empirical function to fit dataset $\\{\log_{10}(\alpha),\log_{10}(\chi^{2})\\}$: $\phi(x;a,b,c,d)=a+\frac{b}{1+e^{-d(x-c)}},$ (28) where $a$, $b$, $c$, and $d$ are fitting parameters. Then the optimal $\alpha$ is approximated by $10^{c-f/d}$, where $f$ is a numerical constant (Its favorite value lies in $[2,2.5]$). ### 3.2 Stochastic analytical continuation In principle, for given Green’s function $G$, there exists infinitely many spectral densities $A(\omega)$ that can be used to reconstruct $G$ via Eq. (12) and Eq. (13). The maximum entropy method tries to pick up the most likely spectral function which maximizes $P[A|\bar{G}]$ (It actually maximizes $Q$) [11, 12]. Here, we would like to introduce an alternative approach, namely the stochastic analytical continuation [13, 14, 29, 30, 31, 32, 33, 34]. It is argued that the weights for all the possible spectral densities are the same if they can give rise to the same $\chi^{2}$. At first, a sequence of spectral densities will be generated by stochastic method. Then an unbiased thermal average of all possible spectra, Boltzmann weighted according to goodness-of- fit function $\chi^{2}$, produces an average spectrum. Thus sometimes the method was named as average spectrum method or stochastic sampling method in the references [35, 36, 37, 38]. There are several variants for the stochastic analytical continuation. Next we will introduce two representative algorithms as proposed by A. W. Sandvik [13] and K. S. D. Beach [29], respectively. #### 3.2.1 Sandvik’s algorithm Figure 1: Typical Monte Carlo field configurations for the stochastic analytical continuation (A. W. Sandvik’s algorithm) [13]. Here, the $\delta$ functions reside at unrestricted frequencies $\\{\omega_{i}\\}$, but their amplitudes $\\{\gamma_{i}\\}$ are equal and fixed. Note that different parameterizations are also possible [34]. It was early on realized that a different way to achieve a smooth spectrum is to average over many solutions with reasonable $\chi^{2}$ values [33]. Several years later, A. W. Sandvik introduced the stochastic analytical continuation in a slightly different form [13]. He suggested that the spectral function $A(\omega)$ can be parameterized using $N$ $\delta$ functions (Please see Figure 1 for a schematic diagram): $A(\omega)=\sum^{N}_{i=1}\gamma_{i}\delta(\omega-\omega_{i}),$ (29) where $\gamma_{i}$ and $\omega_{i}$ denote the amplitude and position of the $i$-th $\delta$ function, respectively. Next, the Metropolis important sampling algorithm is employed to sample the configuration space $\mathcal{C}=\\{\omega_{i},\gamma_{i}\\}$. In practice, there are two elementary Monte Carlo updates. One is to change the amplitudes of a pair of $\delta$ functions under the constraint $\sum_{i}\gamma_{i}=1$. Another one is to shift position of a randomly chosen $\delta$ function. Of course, block or global updates can be implemented to improve ergodicity and sampling efficiency [35, 36]. The transition probability of Monte Carlo updates reads: $p(\mathcal{C}\to\mathcal{C}^{\prime})=\exp\left(-\frac{\Delta\chi^{2}}{2\Theta}\right),$ (30) where the goodness-of-fit function $\chi^{2}$ can be evaluated by Eq. (19), $\Theta$ is a regulation parameter which is similar to the $\alpha$ parameter appeared in the maximum entropy method. Well, the remaining problem is how to fix $\Theta$. Sandvik suggested to measure the following entropic term for a series of $\Theta$: $S(\Theta)=-\sum^{N}_{i=1}\gamma_{i}\log(\gamma_{i})K(0,\omega_{i}),$ (31) where $K$ is the kernel function as defined above [13]. Then make a plot of $S$ with respect to $\log(\Theta^{-1})$. Overall, when $\Theta$ is large, $S$ exhibits large fluctuations. When $\Theta$ is small, $S$ will approach its global minimum steadily. A sharp drop in $S$ before the approach to a constant value has been observed, and there is a local maximum at some $\Theta=\hat{\Theta}$ preceding the drop. Thus, Sandvik postulated that $\hat{\Theta}$ was the optimum value at which to accumulate and average the spectral function. Syljuåsen _et al._ [38] suggested that let $\Theta=1$. Fuchs _et al._ tried to fix $\Theta$ by using Bayesian inference. Such that their approach was named as stochastic analytical inference [14]. Very recently, Shao and Sandvik _et al._ proposed a smart method to determine the optimal value of $\Theta$ [32, 30]. $\Theta$ is adjusted so that $\langle\chi^{2}(\Theta)\rangle\approx\chi^{2}_{\text{min}}+c\sqrt{2\chi^{2}_{\text{min}}},$ (32) where $c$ is a constant of order 1, $\chi^{2}_{\text{min}}$ is the minimum value of $\chi^{2}$ at given $\Theta$. Note that $\chi^{2}_{\text{min}}$ can be obtained in a simulated annealing process [39] to very low $\Theta$. #### 3.2.2 Beach’s algorithm Figure 2: Typical Monte Carlo field configurations for the stochastic analytical continuation (K. S. D. Beach’s algorithm) [29]. Note that the amplitudes $\\{\gamma_{i}\\}$ of all the $\delta$ functions are not identical. Both amplitudes $\\{\gamma_{i}\\}$ and positions $\\{r_{i}\\}$ ($0.0<r_{i}<1.0$) can be sampled by using Monte Carlo method. K. S. D. Beach proposed another variant of stochastic analytical continuation in 2004 [29]. In his proposal, the analytical continuation problem is mapped into a system of interacting classic fields at first. Then the classic field is sampled using Monte Carlo method to obtain the final solution. He concluded that the maximum entropy method is simply the mean field limit of the stochastic analytical continuation. Next, this algorithm will be explained concisely. _Classic fields_. Recalled that the goodness-of-fit functional $\chi^{2}[A]$ measures how closely the Green’s function generated from $A(\omega)$ matches the raw input data. Its expression is rewritten as follows: $\chi^{2}[A]=\int^{\beta}_{0}\frac{1}{\sigma(\tau)^{2}}\left|\int d\omega~{}K(\tau,\omega)A(\omega)-\bar{G}(\tau)\right|^{2}d\tau.$ (33) At first, a new variable $x$ is introduced. The relation between $x$ and $\omega$ is: $x=\phi(\omega)=\int^{\omega}_{-\infty}d\omega^{\prime}~{}m(\omega^{\prime}),$ (34) where $m(\omega)$ denotes the default model function. Clearly, the $\phi(\omega)$ function defines a smooth mapping from $\mathbf{R}\to[0,1]$. Since $\omega=\phi^{-1}(x)$, a dimensionless classic field $n(x)$ is created: $n(x)=\frac{A(\phi^{-1}(x))}{m(\phi^{-1}(x))}.$ (35) It is easy to prove that both $n(x)$ and $A(\omega)$ obey similar normalization condition: $\int d\omega~{}A(\omega)=\int^{1}_{0}dx~{}n(x)=1.$ (36) Next, in analogy with the goodness-of-fit functional $\chi^{2}[A]$, the Hamiltonian for the system of classic field $\\{n(x)\\}$ can be defined as follows: $H[n(x)]=\int^{\beta}_{0}\frac{d\tau}{\sigma(\tau)^{2}}\left|\int^{1}_{0}dx~{}K(\tau,x)n(x)-\bar{G}(\tau)\right|.$ (37) Supposing $\alpha$ is an inverse temperature of the system, then the partition function $Z$ is: $Z=\int\mathcal{D}n~{}e^{-\alpha H[n]},$ (38) where $\int\mathcal{D}n=\int^{\infty}_{0}\left[\prod_{x}dn(x)\right]\delta\left(\int^{1}_{0}dx~{}n(x)-1\right).$ (39) The thermally averaged value of the classic field is: $\langle n(x)\rangle=\frac{1}{Z}\int\mathcal{D}n~{}n(x)e^{-\alpha H[n]}.$ (40) Finally, according to the definition of the classic field, the averaged spectral density $\langle A(\omega)\rangle$ can be expressed as: $\langle A(\omega)\rangle=\langle n(\phi(\omega))\rangle m(\omega).$ (41) So, by introducing the classic field $\\{n(x)\\}$, the analytical continuation problem is converted into a statistical sampling of the classic field, which is easily solved by using Monte Carlo method. _Monte Carlo sampling_. Next we clarify how to sample the classic field. Similar to Sandvik’s algorithm [13, 34], $n(x)$ is parameterized as a superposition of many $\delta$ functions (see Figure 2 for a schematic diagram): $n_{\mathcal{C}}(x)=\sum_{i}\gamma_{i}\delta(x-r_{i}),$ (42) where $\gamma_{i}$ and $r_{i}$ denote amplitude (weight) and position of the $i$-th $\delta$ function, respectively. And $\mathcal{C}$ means a configuration space formed by a set of $r_{i}$ and $\gamma_{i}$, $\mathcal{C}=\\{r_{i},\gamma_{i}\\}.$ (43) Note that $\gamma_{i}$ and $r_{i}$ satisfy the following constraints: $\forall i,~{}\gamma_{i}>0,~{}\sum_{i}\gamma_{i}=1,~{}0\leq r_{i}\leq 1.$ (44) Supposed that there is a transition from $\mathcal{C}$ to $\mathcal{C}^{\prime}$ ($\\{r_{i},\gamma_{i}\\}\to\\{r^{\prime}_{i},\gamma^{\prime}_{i}\\}$): $r_{i}\to r^{\prime}_{i}=r_{i}+\sum_{\lambda\in\Lambda}\delta_{i\lambda}\Delta r_{\lambda},$ (45) $\gamma_{i}\to\gamma^{\prime}_{i}=\gamma_{i}+\sum_{\lambda\in\Lambda}\delta_{i\lambda}\Delta\gamma_{\lambda},$ (46) where $\Lambda$ means a subset of the $\delta$ functions, then the Hamiltonian of the system is changed from $H_{\mathcal{C}}$ to $H_{\mathcal{C}^{\prime}}$. According to Eq. (37), $H_{\mathcal{C}}$, $H_{\mathcal{C}^{\prime}}$, and their difference $\Delta H$ can be calculated by: $H_{\mathcal{C}}=\int^{\beta}_{0}d\tau~{}h_{\mathcal{C}}(\tau)^{2},$ (47) $H_{\mathcal{C}^{\prime}}=\int^{\beta}_{0}d\tau\left[h_{\mathcal{C}}(\tau)+\Delta h(\tau)\right]^{2},$ (48) $\Delta H=H_{\mathcal{C}^{\prime}}-H_{\mathcal{C}}=\int^{\beta}_{0}d\tau~{}\Delta h(\tau)[2h_{\mathcal{C}}(\tau)+\Delta h(\tau)].$ (49) Here, $h(\tau)=\frac{1}{\sigma(\tau)}\left[\int^{1}_{0}dx~{}K(\tau,x)n(x)-\bar{G}(\tau)\right],$ (50) and $\Delta h(\tau)=\frac{1}{\sigma(\tau)}\sum_{\lambda\in\Lambda}\left[\gamma^{\prime}_{\lambda}K(\tau,r^{\prime}_{\lambda})-\gamma_{\lambda}K(\tau,r_{\lambda})\right].$ (51) Finally, the transition probability from $\mathcal{C}$ to $\mathcal{C}^{\prime}$ reads $p(C\to C^{\prime})=\exp(-\alpha\Delta H).$ (52) _Parallel tempering_. The parallel tempering trick [40] is adopted to improve the Monte Carlo algorithm as described above. It is possible to proceed multiple simulations simultaneously for a sequence of inverse temperature parameters $\\{\alpha_{1},\alpha_{2},\cdots,\alpha_{N}\\}$. The ratio for two adjacent $\alpha$ parameters is a constant: $\alpha_{p+1}/\alpha_{p}=R$. Note that the field configurations in all simulations evolve in parallel but not independently. We can swap the field configurations between two adjacent layers. Of course, the detailed balance is always preserved, and each simulation will eventually settle into thermal equilibrium at given $\alpha$. The transition probability of such a global Monte Carlo update is: $p(\mathcal{C}\to\mathcal{C}^{\prime})=\exp[(\alpha_{p}-\alpha_{q})(H_{p}-H_{q})],$ (53) where $p$ and $q$ are layer indices, and $p=q\pm 1$. Parallel tempering eliminates the need for an initial annealing stage. Another advantage of parallel tempering is that it yields a complete temperature profile of all the important thermodynamic variables (such as specific heat and internal energy), which can be used to estimate the critical $\alpha$ and the final spectral function $\langle A(\omega)\rangle$. _Critical inverse temperature_. Clearly, $\langle n(x)\rangle$ strongly depends on the inverse temperature $\alpha$. How to use these $\alpha$-dependent $\langle n(x)\rangle$ to construct the final spectral function? Beach suggested a novel method [29]. During parallel tempering process, the internal energy of the system is also measured in addition to $\langle n(x)\rangle$: $U(\alpha_{p})=\langle H[n]\rangle_{\alpha_{p}}.$ (54) Let us plot $\log_{10}[U(\alpha)]$ as a function of $\log_{10}(\alpha)$. We find that $\log_{10}[U(\alpha)]$ drops quickly at first when $\log_{10}(\alpha)$ increases, and then it approaches to a constant value slowly. The knee in $\log_{10}[U(\alpha)]$ function, occurring in the vicinity of $\alpha=\alpha^{*}$ (the corresponding layer index $p=p^{*}$), signals a jump in specific heat (a thermodynamic phase transition). Then the averaged spectral function is constructed by: $\langle\langle n(x)\rangle\rangle=\frac{\sum^{N-1}_{p=p*}[U(\alpha_{p})-U(\alpha_{p+1})]\langle n(x)\rangle_{\alpha_{p}}}{U(\alpha_{p*})-U(\alpha_{N})},$ (55) where $N$ is the total number of $\alpha$, and $\alpha_{p*}$ ($\equiv\alpha^{*}$) is the critical inverse temperature. _Likelihood function_. Neither of the Sandvik’s and Beach’s algorithms needs extra entropic term to regulate the spectral densities [13, 29]. All the stochastically generated spectra are treated on the same footing. Thus, the calculated spectral function retains more subtle structures than that obtained by the maximum entropy method. Actually, in the stochastic analytical continuation, $\langle A\rangle=\int\mathcal{D}A~{}P[A|\bar{G}]A.$ (56) The weight of the candidate spectral function $A$ is given by the likelihood function $P[A|\bar{G}]$. Eq. (30) and Eq. (52) can be viewed as likelihood functions in the stochastic analytical continuation. ### 3.3 Stochastic optimization method Figure 3: Typical Monte Carlo field configurations for the stochastic optimization method [15]. The spectral function is parameterized by multiple rectangle functions. Here, $c_{i}$, $w_{i}$, and $h_{i}$ denote the center, width, and height of the $i$-th rectangle, respectively. A. O. Mishchenko _et al._ [15] proposed the stochastic optimization method. Though it looks like the stochastic analytical continuation [13, 29], their differences are quite apparent. The stochastic optimization method does not need any likelihood function or Boltzmann distribution to weight the candidate spectral functions. It generates a lot of spectral functions through Monte Carlo samplings. For each candidate spectral function, the deviation $D$ between the reconstructed Green’s function $\tilde{G}$ and original Green’s function $\bar{G}$ is measured. Those spectral functions with small deviations $D$ are selected and averaged. Such that the desired spectral function is obtained. _Deviation function_. In the stochastic optimization method, the deviation between reconstructed data $\tilde{G}$ and input data $\bar{G}$ is described by: $D[A]=\sum^{M}_{m=1}|\Delta(m)|,$ (57) where $M$ is the number of input data, and $\Delta(m)$ is the deviation function, $\Delta(m)=\frac{\bar{G}(m)-\tilde{G}(m)}{S(m)}.$ (58) Here, $S(m)=|G(m)|^{d}$ (where $0\leq d\leq 1$). Recently, Krivenko _et al._ suggested that it would be better to use the goodness-of-fit functional $\chi^{2}[A]$ to replace $D[A]$ [24, 25]. _Spectral density_. The stochastic optimization method will try to accumulate the candidate spectral functions that manifest small $D[A]$. Supposed the Monte Carlo simulations are repeated for $L$ times. For the $i$-th Monte Carlo simulation, the spectral density $A_{i}(\omega)$ and deviation $D[A_{i}]$ are recorded. The minimum value of deviation is $\min\\{D[A_{i}]\\}$. Thus, the final spectral density reads: $A(\omega)=\frac{1}{L_{\text{good}}}\sum^{L}_{i=1}\theta(\alpha_{\text{good}}\min\\{D[A_{i}]\\}-D[A_{i}])A_{i}(\omega).$ (59) Here, $\theta(x)$ is the Heaviside step function, and $\alpha_{\text{good}}$ is a adjustable parameter. $L_{\text{good}}$ denotes the number of “good” spectral functions: $L_{\text{good}}=\sum^{L}_{i=1}\theta(\alpha_{\text{good}}\min\\{D[A_{i}]\\}-D[A_{i}]).$ (60) That is to say, only those spectral functions who satisfy the following condition will be selected: $D[A_{i}]\leq\alpha_{\text{good}}\min\\{D[A_{i}]\\}.$ (61) Clearly, the larger $\alpha_{\text{good}}$ is, the more spectral functions are included. It is usually set to 2. _Rectangle representation_. Similar to the stochastic analytical continuation [13, 29], the stochastic optimization method usually employs a few rectangle functions to parameterize the spectral function: $A(\omega)=\sum_{i}R_{\\{c_{i},w_{i},h_{i}\\}}(\omega),$ (62) where $i$ is the index of rectangle function. The definition of rectangle function $R_{\\{c_{i},w_{i},h_{i}\\}}(\omega)$ reads: $R_{\\{c_{i},w_{i},h_{i}\\}}(\omega)=h_{i}\theta[\omega-(c_{i}-w_{i}/2)]\theta[(c_{i}+w_{i}/2)-\omega],$ (63) where $c_{i}$, $w_{i}$, $h_{i}$ denote the center, width, and height of the $i$-th rectangle, respectively. Pay attention to that the area of all rectangles must be normalized to 1: $\sum_{i}h_{i}w_{i}=1.$ (64) _Monte Carlo sampling_. The parameters of all rectangle functions create a configuration space: $\mathcal{C}=\\{c_{i},w_{i},h_{i}\\}.$ (65) Then the Metropolis algorithm is utilized to sample this configuration space. Mishchenko _et al._ introduces seven Monte Carlo updates [15, 24], including: (a) Insert a new rectangle, change width and height of another rectangle; (b) Remove an existing rectangle, change width and height of another rectangle; (c) Shift position of any rectangles; (d) Change widths of any two rectangles; (e) Change heights of any two rectangles; (f) Split a rectangle into two new rectangles; (g) Merge two adjacent rectangles into a new rectangle. The transition probability of these Monte Carlo updates is: $p(\mathcal{C}\to\mathcal{C}^{\prime})=\left(\frac{D[A_{\mathcal{C}}]}{D[A_{\mathcal{C}^{\prime}}]}\right)^{1+d}$ (66) As compared to the maximum entropy method [11, 12], the likelihood function, entropic term, and model function are absent in the stochastic optimization method. As compared to the stochastic analytical continuation [13, 29], there are no adjustable parameters, such as $\Theta$ in Sandvik’s algorithm and $\alpha$ in Beach’s algorithm. Thus, the simulated results of the stochastic optimization method are less affected by artificial parameters. ## 4 Overview ### 4.1 Major features Now the ACFlow toolkit supports three analytical continuation methods as introduced above. It includes four different analytical continuation solvers, namely MaxEnt, StochAC, StochSK, and StochOM. Just as their names suggested, the MaxEnt solver implements the maximum entropy method [12]. The StochAC and StochSK solvers implement the K. S. D. Beach’s algorithm [29] and A. W. Sandvik’s algorithm [13] of the stochastic analytical continuation, respectively. The StochOM solver implements the stochastic optimization method [15]. The ACFlow toolkit also provides a convenient library, which can be used to prepare and carry out analytical continuation calculations flexibly. The major features of the ACFlow toolkit are summarized in Table 1. Features | MaxEnt | StochAC | StochSK | StochOM ---|---|---|---|--- Matrix-valued Green’s function | Y | N | N | N Imaginary time grid | Y | Y | Y | Y Matsubara frequency grid | Y | Y | Y | Y Linear mesh | Y | Y | Y | Y Nonlinear mesh | Y | Y | Y | Y Fermionic kernel | Y | Y | Y | Y Bosonic kernel | Y | Y | Y | Y Self-defined model function | Y | N | N | N Constrained analytical continuation | N | Y | Y | Y Regeneration of input data | Y | Y | Y | Y Kramers-Kronig transformation | Y | Y | Y | Y Parallel computing | N | Y | Y | Y Parallel tempering | N | Y | N | N Interactive mode | Y | Y | Y | Y Script mode | Y | Y | Y | Y Standard mode | Y | Y | Y | Y Table 1: Major features of the ACFlow toolkit. MaxEnt, StochAC, StochSK, and StochOM are the four analytical continuation solvers as implemented in this toolkit. In Table 1, “Y” means yes while “N” means no. “Interactive mode”, “Script mode”, and “Standard model” are the three running modes supported by the ACFlow toolkit. We will introduce them in next section. The MaxEnt solver supports the “historic”, “classic”, “bryan”, and “chi2kink” algorithms to determine the $\alpha$ parameter. The StochAC solver is only compatible with a flat model function, while the StochSK and StochOM solvers don’t rely on any default model functions. The StochOM solver does not support analytical continuation of fermionic imaginary time Green’s function for the moment. ### 4.2 Implementations The ACFlow toolkit is developed with pure Julia language. Thanks to powerful type system and multiple dispatch paradigm of the Julia language, the four different analytical continuation solvers are integrated into an united software architecture. Redundant codes are greatly reduced. It is quite easy to implement new analytical continuation solver or add new features to the existing solvers in the future. Distributed computing is a built-in feature of Julia. So, it is straightforward to realize parallel calculations in the ACFlow toolkit. Now except for the MaxEnt solver, all the other solvers are parallelized. Filename | Description ---|--- ACFlow.jl | Entry of the ACFlow module. maxent.jl | Maximum entropy method. sac.jl | Stochastic analytical continuation (K. S. D. Beach’s algorithm). san.jl | Stochastic analytical continuation (A. W. Sandvik’s algorithm). som.jl | Stochastic optimization method. global.jl | Numerical and physical constants. types.jl | Basic data structures and computational parameters. base.jl | Driver for analytical continuation simulation. inout.jl | Read input data and write calculated results. config.jl | Parse configuration file and extract computational parameters. math.jl | Root finding, numerical integration, interpolation, Einstein summation, and curve fitting. util.jl | Some utility functions. mesh.jl | Meshes for spectral density. grid.jl | Grids for input data. model.jl | Default model functions. kernel.jl | Kernel functions. Table 2: List of source codes of the ACFlow toolkit. The source codes of the ACFlow toolkit are placed in the acflow/src folder. Their functions are summarized in Table 2. The documentation of the ACFlow toolkit is written by using the Markdown language and Documenter.jl package. The source codes are placed in the acflow/docs folder. The users can build documentation by themselves. Please see section 5 for how to do that. Or they can read the latest documentation in the following website: https://huangli712.github.io/projects/acflow/index.html Ten tests and four tutorials are also shipped with the ACFlow toolkit. Their source codes are placed in the acflow/test folder. See acflow/test/test.md and acflow/test/tutor.md for more details. The code repository of the ACFlow toolkit is: https://github.com/huangli712/ACFlow ## 5 Getting started In this section, we will discuss how to install and use the ACFlow toolkit. ### 5.1 Installation It is an easy task to install the ACFlow toolkit. First, since it is written in pure Julia language, it is necessary to install the Julia runtime environment at first. The newest version of Julia is always preferred (version $>$ 1.60). Since the core codes only rely on Julia’s built-in standard library, no the third-party packages are needed. Second, just download source codes of the ACFlow toolkit from its github repository. It should be a compressed file, such as acflow.zip or acflow.tar.gz. Please uncompress it in your favorite directory by using the following commands: $ unzip acflow.zip or $ tar xvfz acflow.tar.gz Third, the users have to declare a new environment variable ACFLOW_HOME. Supposed that the root directory of the ACFLow toolkit is /home/your_home/acflow, then ACFLOW_HOME should be setup as follows: $ export ACFLOW_HOME=/home/your_home/acflow/src Finally, in order to generate the documentation, the users should type the following commands in the terminal: $ pwd /home/your_home/acflow $ cd docs $ julia make.jl After a few seconds, the documentation is built and saved in the acflow/docs/build directory if everything is OK. The home page of the documentation is acflow/docs/build/index.html. We can open it with any web browsers. ### 5.2 Run The ACFlow toolkit is designed to be flexible and easy-to-use. It provides three running modes to facilitate analytical continuation calculations, namely the interactive, script, and standard modes. _Interactive mode_. With the ACFlow toolkit, the users can setup and carry out analytical continuation simulations interactively in Julia’s REPL (Read-Eval- Print Loop) environment. For example, julia> push!(LOAD_PATH, ENV["ACFLOW_HOME"]) julia> using ACFlow julia> setup_args("ac.toml") julia> read_param() julia> mesh, Aout, Gout = solve(read_data()) Here, ac.toml is a configuration file, which contains essential computational parameters. The return values of the solve() function (i.e., mesh, Aout, and Gout) are mesh at real axis $\omega$, spectral density $A(\omega)$, and reproduced Green’s function $\tilde{G}$, respectively. They can be further analyzed or visualized by the users. _Script mode_. The core functionalities of the ACFlow toolkit are exposed to the users via a simple application programming interface. So, the users can write Julia scripts easily by themselves to perform analytical continuation simulations. A minimal Julia script (acrun.jl) is listed as follows: #!/usr/bin/env julia push!(LOAD_PATH, ENV["ACFLOW_HOME"]) using ACFlow setup_args("ac.toml") read_param() mesh, Aout, Gout = solve(read_data()) Of course, this script can be extended to finish complex tasks. In section 6.1, a realistic example is provided to show how to complete an analytical continuation of Matsubara self-energy function via the script mode. _Standard mode_. In the standard mode, the users have to prepare the input data manually. In addition, a configuration file must be provided. Supposed that the configuration file is ac.toml, then the analytical continuation calculation is launched as follows: $ /home/your_home/acflow/util/acrun.jl ac.toml or $ /home/your_home/acflow/util/Pacrun.jl ac.toml Noted that the acrun.jl script runs sequentially, while the Pacrun.jl script supports parallel and distributed computing. As we can conclude from the filename extension of configuration file (ac.toml), it adopts the TOML specification. The users may edit it with any text-based editors. Next we will introduce syntax and format of the input data files and configuration files. ### 5.3 Input files The input files for the ACFlow toolkit can be divided into two groups: data files and configuration files. _Data files_. The input data should be store in CSV-like text files. For imaginary time Green’s function, the data file should contain three columns. They represent $\tau$, $\bar{G}(\tau)$, and standard deviation of $\bar{G}(\tau)$. For fermionic Matsubara Green’s function, the data file should contain five columns. They represent $\omega_{n}$, Re$G(i\omega_{n})$, Im$G(i\omega_{n})$, standard deviation of Re$G(i\omega_{n})$, and standard deviation of Im$G(i\omega_{n})$. For bosonic correlation function $\chi(i\omega_{n})$, the data file should contain four columns. They represent $\omega_{n}$, Re$\chi(i\omega_{n})$, and standard deviation of Re$\chi(i\omega_{n})$. _Configuration files_. The configuration file adopts the TOML format. It is used to setup the computational parameters. It consists of one or more blocks. Possible blocks (or sections) of the configuration file include [BASE], [MaxEnt], [StochAC], [StochSK], and [StochOM]. The [BASE] block is mandatory, while the other blocks are optional. A schematic configuration file (ac.toml) is listed as follows: ⬇ [BASE] finput = "giw.data" solver = "StochOM" ... [MaxEnt] method = "chi2kink" ... [StochAC] nfine = 10000 ... [StochSK] method = "chi2min" ... [StochOM] ntry = 100000 ... In the [BASE] block, the analytical continuation problem is defined. The solver used to solve the problem must be assigned. The types of mesh, grid, default model function, and kernel function are also determined. The [MaxEnt], [StochAC], [StochSK], and [StochOM] blocks are used to customize the corresponding analytical continuation solvers further. In Table 3-Table 6, all the possible input parameters for these blocks are collected and summarized. As for detailed explanations of these parameters, please refer to the user guide of the ACFlow toolkit. The uses can find it in the acflow/docs directory. [BASE] block --- Parameter | Type | Default | Description finput | string | “green.data” | Filename for input data. solver | string | “MaxEnt” | Solver for the analytical continuation problem. ktype | string | “fermi” | Type of kernel function. mtype | string | “flat” | Type of default model function. grid | string | “ffreq” | Grid for input data (imaginary axis). mesh | string | “linear” | Mesh for output data (real axis). ngrid | integer | 10 | Number of grid points. nmesh | integer | 501 | Number of mesh points. wmax | float | 5.0 | Right boundary (maximum value) of mesh. wmin | float | -5.0 | Left boundary (minimum value) of mesh. beta | float | 10.0 | Inverse temperature. offdiag | bool | false | Treat the off-diagonal part of matrix-valued function? pmodel | array | N/A | Additional parameters for customizing the default model. pmesh | array | N/A | Additional parameters for customizing the mesh. exclude | array | N/A | Restriction of energy range of the spectrum. Table 3: Possible parameters for the [BASE] block. [MaxEnt] block --- Parameter | Type | Default | Description method | string | “chi2kink” | How to determine the optimized $\alpha$ parameter? nalph | integer | 12 | Total number of the chosen $\alpha$ parameters. alpha | float | 1e9 | Starting value for the $\alpha$ parameter. ratio | float | 10.0 | Scaling factor for the $\alpha$ parameter. blur | float | -1.0 | Shall we preblur the kernel and spectrum? Table 4: Possible input parameters for the [MaxEnt] block, which are used to setup the solver based on the maximum entropy method [12, 11]. [StochAC] block --- Parameter | Type | Default | Description nfine | integer | 10000 | Number of points of a very fine linear mesh. ngamm | integer | 512 | Number of $\delta$ functions. nwarm | integer | 4000 | Number of Monte Carlo thermalization steps. nstep | integer | 4000000 | Number of Monte Carlo sweeping steps. ndump | integer | 40000 | Intervals for monitoring Monte Carlo sweeps. nalph | integer | 20 | Total number of the chosen $\alpha$ parameters. alpha | float | 1.0 | Starting value for the $\alpha$ parameter. ratio | float | 1.2 | Scaling factor for the $\alpha$ parameter. [StochSK] block Parameter | Type | Default | Description method | string | “chi2min” | How to determine the optimized $\Theta$ parameter? nfine | integer | 100000 | Number of points of a very fine linear mesh. ngamm | integer | 1000 | Number of $\delta$ functions. nwarm | integer | 1000 | Number of Monte Carlo thermalization steps. nstep | integer | 20000 | Number of Monte Carlo sweeping steps. ndump | integer | 200 | Intervals for monitoring Monte Carlo sweeps. retry | integer | 10 | How often to recalculate the goodness-of-fit function. theta | float | 1e6 | Starting value for the $\Theta$ parameter. ratio | float | 0.9 | Scaling factor for the $\Theta$ parameter. Table 5: Possible input parameters for the [StochAC] and [StochSK] blocks, which are used to setup the two solvers based on the stochastic analytical continuation (Beach’s and Sandvik’s algorithms) [13, 29]. [StochOM] block --- Parameter | Type | Default | Description ntry | integer | 2000 | Number of attempts to figure out the solution. nstep | integer | 1000 | Number of Monte Carlo steps per try. nbox | integer | 100 | Number of boxes to construct the spectrum. sbox | float | 0.005 | Minimum area of the randomly generated rectangles. wbox | float | 0.02 | Minimum width of the randomly generated rectangles. norm | float | -1.0 | Is the norm calculated? Table 6: Possible input parameters for the [StochOM] block, which are used to setup the solver based on the stochastic optimization method [15]. ### 5.4 Output files Once the analytical continuation simulation is finished, the final spectral function $A(\omega)$ is outputted to Aout.data. As is shown in Eq. (7), $A(\omega)$ is equivalent to the imaginary part of real frequency Green’s function Im$G(\omega)$. Then the ACFlow toolkit will automatically calculate the corresponding real part Re$G(\omega)$ via the Kramers-Kronig transformation [see Eq. (8)]. The full Green’s function at real axis $G(\omega)$ is stored in Gout.data. The spectral function is also used to reconstruct the imaginary time or Matsubara Green’s functions [$\tilde{G}(\tau)$ or $\tilde{G}(i\omega_{n})$], which is stored in repr.data. Besides the three output files, the ACFlow toolkit will generate quite a few output files, which can be used to analyze and diagnose the calculated results. All of the possible output files of the ACFlow toolkit are collected and explained in Table 7. Filename | Description ---|--- Aout.data | Final spectral function $A(\omega)$. Gout.data | Full Green’s function at real axis $G(\omega)$. repr.data | Reproduced Green’s function $\tilde{G}$ at imaginary time or frequency axis. model.data | Default model function $m(\omega)$. chi2.data | $\log_{10}(\chi^{2})$ vs $\log_{10}(\alpha)$. prob.data | $P[\alpha|\bar{G}]$ vs $\alpha$ for the MaxEnt solver (bryan algorithm). Aout.data.alpha_$i$ | $\alpha$-resolved spectral function $A_{\alpha}(\omega)$ for the StochAC solver. hamil.data | $U(\alpha)$ vs $\alpha$ for the StochAC solver. goodness.dat | $\log_{10}(\chi^{2})$ vs $\log_{10}(\Theta)$ for the StochSK solver. stat.data | Monte Carlo statistical information for stochastic sampling method. Table 7: Possible output files of the ACFlow toolkit. ## 6 Examples In order to demonstrate usefulness of the ACFlow toolkit, four examples are illustrated in this section. These examples cover typical application scenarios of the ACFlow toolkit, including analytical continuations of Matsubara self-energy function, Matsubara Green’s function, imaginary time Green’s function, and current-current correlation function within the script mode or standard mode. All of the necessary source codes and data files, which can be used to reproduce the results as shown in this section, are placed in the /home/your_home/acflow/test/T* folders. ### 6.1 Matsubara self-energy function Now let us consider the following single-band Hubbard model on a Bethe lattice at first: $H=-t\sum_{\langle ij\rangle\sigma}c^{\dagger}_{i\sigma}c_{j\sigma}-\mu\sum_{i}n_{i}+U\sum_{i}n_{i\uparrow}n_{i\downarrow},$ (67) where $t$ is the hopping parameter, $\mu$ is the chemical potential, $U$ is the Coulomb interaction, $n$ is the occupation number, $\sigma$ denotes the spin, $i$ and $j$ are site indices. This model is solved by using the dynamical mean-field theory (dubbed DMFT) [41] with the hybridization expansion continuous-time quantum Monte Carlo solver (dubbed CT-HYB) [3] as implemented in the $i$QIST package [42, 43]. The parameters used in the DMFT + CT-HYB calculation are $t=0.5$, $U=2.0$, $\mu=1.0$, and $\beta=10.0$. Once the DMFT self-consistent calculation is finished, the Matsubara self-energy function $\Sigma(i\omega_{n})$ is obtained. We are going to convert it to real frequency self-energy function $\Sigma(\omega)$. The data of Matsubara self- energy function $\Sigma(i\omega_{n})$ have been preprocessed and stored in siw.data. This file contains five columns, which are used to record the Matsubara frequency $\omega_{n}$, Re$\Sigma(i\omega_{n})$, Im$\Sigma(i\omega_{n})$, error bar of Re$\Sigma(i\omega_{n})$, error bar of Im$\Sigma(i\omega_{n})$, respectively. Only the first twenty Matsubara frequency points are kept, because the high-frequency data are somewhat noisy. The purpose of this example is to demonstrate usage of the MaxEnt solver and the script mode of the ACFlow toolkit. Next we will explain the key steps in detail. As for the complete Julia script, please refer to sigma.jl and gendata.jl in the /home/your_home/acflow/test/T01/ folder. First, we have to load the essential Julia packages. Both the DelimitedFiles and Printf packages belong to Julia’s standard library. They are used to read input data and write calculated results, respectively. ⬇ #!/usr/bin/env julia push!(LOAD_PATH, ENV["ACFLOW_HOME"]) using DelimitedFiles using Printf using ACFlow welcome() # Print welcome message only Next, the data of Matsubara self-energy function are read from siw.data. The Hartree term $\Sigma_{H}$ should be subtracted from its real part: $\Sigma(i\omega_{n})\to\Sigma(i\omega_{n})-\Sigma_{H}.$ (68) Note that $\Sigma_{H}$ is approximately equal to the asymptotic value of real part of $\Sigma(i\omega_{n})$ when $n$ goes to infinite. ⬇ # Deal with self-energy function # # Read self-energy function dlm = readdlm("siw.data") # # Get grid grid = dlm[:,1] # # Get self-energy function Sinp = dlm[:,2] + im * dlm[:,3] # Value Serr = dlm[:,4] + im * dlm[:,5] # Error bar # # Subtract hartree term Sh = 1.0 @. Sinp = Sinp - Sh Next, the computational parameters are encapsulated into two dictionaries. The dictionary B is for the [BASE] block, while the dictionary S is for the MaxEnt solver. Then the setup_param() function is called, so that these parameters take effect. Here, the MatEnt solver [12, 11] is employed to tackle the analytical continuation problem. But the other stochastic sampling solvers are also applicable. The default model function is gaussian. The mesh for spectral density is non-uniform (A tangent mesh). The number of used $\alpha$ parameters is 15, and the optimal $\alpha$ parameter is determined by the $\chi^{2}$kink algorithm [22]. ⬇ # Setup parameters # # For [BASE] block # See types.jl/_PBASE for default setup B = Dict{String,Any}( "solver" => "MaxEnt", # Choose MaxEnt solver "mtype" => "gauss", # Default model function "mesh" => "tangent", # Mesh for spectral density "ngrid" => 20, # Number of input points "nmesh" => 801, # Number of output points "wmax" => 8.0, # Right boundary of mesh "wmin" => -8.0, # Left boundary of mesh "beta" => 10.0, # Inverse temperature ) # # For [MaxEnt] block # See types.jl/_PMaxEnt for default setup S = Dict{String,Any}( "nalph" => 15, # Number of alpha "alpha" => 1e12, # Starting value of alpha "blur" => -1.0, # Enable preblur or not ) # # Let the parameters take effect setup_param(B, S) It is quite easy to start the analytical continuation calculation. Just call the solve() function and pass the grid, input data, and error bar data to it. The return values of this function call are real frequency mesh, spectral density, and reconstructed Matsubara self-energy function. ⬇ # Call the solver mesh, Aout, Sout = solve(grid, Sinp, Serr) Finally, the real frequency self-energy function must be supplemented with the Hartree term, and then the final results are written into sigma.data. ⬇ # Calculate final self-energy function on real axis # # Add hartree term @. Sout = Sout + Sh # # Write self-energy function to sigma.data open("sigma.data", "w") do fout for i in eachindex(mesh) z = Sout[i] @printf(fout, "%20.16f␣%20.16f␣%20.16f\n", mesh[i], real(z), imag(z)) end end Figure 4: Analytical continuation of Matsubara self-energy function by using the maximum entropy method. (a) Real part of real frequency self-energy function. (b) Imaginary part of real frequency self-energy function. (c) $\chi^{2}$ as a function of $\alpha$. The vertical bar indicates the optimal $\alpha$ parameter chosen by the $\chi^{2}$kink algorithm. (d) Reproduced and original data for imaginary part of the Matsubara self-energy functions. The calculated results are displayed in Fig. 4. Fig. 4(a) and (b) show the real and imaginary parts of the real frequency self-energy function, respectively. Near the Fermi level, Re$\Sigma(\omega)$ exhibits quasi-linear behavior, with which the quasiparticle weight $Z$ and effective mass of electron $m^{*}$ can be easily evaluated [41]. As for the imaginary part, Im$\Sigma(0)$ is finite, which indicates that the electron-electron scattering is not trivial. Fig. 4(c) shows the $\alpha$-dependent $\chi^{2}$. The vertical bar in this figure indicates the optimal $\alpha$ is around 102.154. In Fig. 4(d), the reproduced and raw Matsubara self-energy functions are compared. It is apparent that they are consistent with each other. ### 6.2 Matsubara Green’s function The purpose of the second example is to treat the Matsubara Green’s function by using the StochOM solver. At first, please consider the following spectral density with two gaussian peaks: $A(\omega)=A_{1}\exp\left[\frac{-(\omega-\epsilon_{1})^{2}}{2\Gamma^{2}_{1}}\right]+A_{2}\exp\left[\frac{-(\omega-\epsilon_{2})^{2}}{2\Gamma^{2}_{2}}\right],$ (69) with $A_{1}=1.0$, $A_{2}=0.3$, $\epsilon_{1}=0.5$, $\epsilon_{2}=-2.5$, $\Gamma_{1}=0.2$, and $\Gamma_{2}=0.8$. Then the Matsubara Green’s function $G(i\omega_{n})$ is evaluated by using Eq. (13) with $\beta=10.0$. Random noises, generated by the formula $0.0001r_{1}\exp(i2\pi r_{2})$ where $r_{1}$ and $r_{2}$ are pseudo random numbers in ($0.0<r_{1}$, $r_{2}<1.0$), are added to $G(i\omega_{n})$. The error bar of $G(i\omega_{n})$ is fixed to 1e-4. The generated data for $G(i\omega_{n})$ are written in giw.data. Next, we are going to use the standard mode, such that a configure file (ac.toml) must be prepared. It is listed as follows. Since the StochOM solver is chosen, the [BASE] and [StochOM] blocks must be present. ⬇ [BASE] finput = "giw.data" solver = "StochOM" ktype = "fermi" mtype = "flat" grid = "ffreq" mesh = "linear" ngrid = 10 nmesh = 501 wmax = 5.0 wmin = -5.0 beta = 10.0 offdiag = false [StochOM] ntry = 100000 nstep = 1000 nbox = 100 sbox = 0.005 wbox = 0.02 norm = -1.0 Figure 5: Analytical continuation of Matsubara Green’s function by using the stochastic optimization method. (a) Simulated and exact spectral functions. (b) Reconstructed and synthetic Matsubara Green’s functions. Only the imaginary parts are presented in this figure. Then we use the acrun.jl or Pacrun.jl script to perform analytical continuation simulation. The calculated results are shown in Fig. 5. As is seen in Fig. 5(a), both the sharp peak around 0.5 eV and the broad peak around -2.5 eV are correctly reproduced by the StochOM solver. In Fig. 5(b), the reconstructed Matsubara Green’s function agrees quite well with the raw input data. ### 6.3 Imaginary time Green’s function In this example, analytical continuation of imaginary time Green’s function will be tested. Note that this example is borrowed from Reference [29] directly. The exact spectral function reads: $A(\omega)=\begin{cases}\frac{1}{W}\frac{|\omega|}{\sqrt{\omega^{2}-\Delta^{2}}},~{}\quad&\text{if}~{}\Delta<|\omega|<W/2.\\\ 0,&\text{otherwise}.\end{cases}$ (70) Here, $W$ denotes bandwidth, and $\Delta$ is used to control size of the energy gap. Let $W=6$ and $2\Delta=1$. This spectrum should exhibit flat shoulders, steep peaks, and sharp gap edges at the same time. Actually, it is a typical spectrum of a BCS superconductor. First, the imaginary time Green’s function $G(\tau)$ is generated using Eq. (12). Then a normally-distributed random noise is add to $G(\tau)$. Maximum amplitude of the noise is 1e-4. The error bar of $G(\tau)$ is fixed to 1e-3. The data are written in gtau.data. Next, we try to prepare the configure file (ac.toml). In this case, we would like to benchmark the StochAC solver, so the solver parameter is set to “StochAC” and the grid parameter is set to “ftime”. Furthermore, the exclude parameter is enabled to impose some _a priori_ constraints to the spectrum. The full ac.toml is listed as follows: ⬇ [BASE] finput = "giw.data" solver = "MaxEnt" ktype = "fermi" mtype = "flat" grid = "ffreq" mesh = "linear" ngrid = 10 nmesh = 501 wmax = 5.0 wmin = -5.0 beta = 10.0 offdiag = false exclude = [[-5.0,-3.0], [-0.5,0.5], [3.0,5.0]] [StochAC] nfine = 10000 ngamm = 512 nwarm = 4000 nstep = 10000000 ndump = 40000 nalph = 40 alpha = 1.00 ratio = 1.20 We perform analytical continuation simulation by using the acrun.jl or Pacrun.jl script. In order to obtain smooth spectral density, it is useful to increase number of $\delta$ functions (See ngamm parameter) and number of Monte Carlo sampling steps (See nstep parameter). Figure 6: Analytical continuation of imaginary time Green’s function by using the stochastic analytical continuation (Beach’s algorithm). (a) Simulated and exact spectral functions. (b) $\alpha$-dependent spectral functions. (c) Internal energy $U$ as a function of $\alpha$. The vertical bar indicates the optimal $\alpha$ parameter. (d) Simulated and exact imaginary time Green’s functions. Figure 6 shows the calculated results. In Fig. 6(a), the exact spectral function is compared with the simulated spectrum. Note that besides the StochAC solver, the other three solvers are also tested. Their results are also plotted in this figure for a direct comparison. It is remarkable that the StochAC and StochSK solvers do a superior job of modelling the spectrum. The major characteristics of the spectrum, including flat regions, steep peaks, and sharp gap edges, are well captured by the two solvers. Especially, we have finished more tests without any constraints on the spectral density. The gap in the spectrum can be reproduced as well. On the other hand, the spectra obtained by the MaxEnt and StochOM solvers are much too smooth, and show extra shoulder peaks around $\pm$2.0. Figure 6(b) shows $\alpha$-resolved spectral functions $A_{\alpha}(\omega)$ for selected $\alpha$ parameters. Fluctuation in the flat regions of the calculated spectral density grows when $\alpha$ increases. Figure 6(c) shows internal energy $U$ as a function of $\alpha$. From this figure, the critical $\alpha$ is estimated, which is indicated by the vertical bar. Finally, the reproduced Green’s function $\tilde{G}(\tau)$ agrees quite well with the raw input data, as is shown in Fig. 6(d). ### 6.4 Current-current correlation function The previous three examples only involve fermionic correlators. How about bosonic correlation functions? In this example, we will demonstrate how to perform analytical continuation simulation for a typical bosonic correlator, the current-current correlation function $\Pi(\tau)$, to obtain the optical conductivity $\sigma(\omega)$. Note that this example is taken from Reference [5] directly. The exact optical conductivity $\sigma(\omega)$ reads: $\sigma(\omega)=\left\\{\frac{W_{1}}{1+(\omega/\Gamma_{1})^{2}}+\frac{W_{2}}{1+[(\omega-\epsilon)/\Gamma_{2}]^{2}}+\frac{W_{2}}{1+[(\omega+\epsilon)/\Gamma_{2}]^{2}}\right\\}\frac{1}{1+(\omega/\Gamma_{3})^{6}},$ (71) where $W_{1}=0.3$, $W_{2}=0.2$, $\Gamma_{1}=0.3$, $\Gamma_{2}=1.2$, $\Gamma_{3}=4.0$, and $\epsilon=3.0$. The current-current correlation function $\Pi(\tau)$ can be evaluated from $\sigma(\omega)$ by using the following equation: $\Pi(\tau)=\int^{\infty}_{-\infty}K(\tau,\omega)\sigma(\omega)~{}d\omega,$ (72) where the kernel function $K(\tau,\omega)$ is different from Eq. (14). It reads: $K(\tau,\omega)=\frac{1}{\pi}\frac{\omega e^{-\tau\omega}}{1-e^{-\beta\omega}}.$ (73) In this case, $\beta$ is fixed to be 20.0. At first, we use Eq. (71) $\sim$ Eq. (73) to prepare $\Pi(\tau)$. A normally- distributed random noise is add to $\Pi(\tau)$. Maximum amplitude of the noise is 1e-4. The error bar of $\Pi(\tau)$ is fixed to 1e-4. The data of $\Pi(\tau)$ are written in chit.data. Next, we conduct analytical continuation simulation as usual. The used configuration file is attached as follows. Here, the StochSK solver is adopted, so the solver parameter is “StochSK” and the grid parameter is “btime”. And the Shao-Sandvik algorithm [30] is applied to seek optimal $\Theta$, so the method parameter is “chi2min”. The users can further increase the values of nfine, ngamm, and nstep parameters to improve computational accuracy. ⬇ [BASE] finput = "chit.data" solver = "StochSK" ktype = "bsymm" mtype = "flat" grid = "btime" mesh = "linear" ngrid = 501 nmesh = 801 wmax = 8.0 wmin = 0.0 beta = 20.0 offdiag = false [StochSK] method = "chi2min" nfine = 40000 ngamm = 1000 nwarm = 1000 nstep = 1000000 ndump = 200 retry = 10 theta = 1e+6 ratio = 0.90 Figure 7: Analytical continuation of current-current correlation function by using the stochastic analytical continuation (Sandvik’s algorithm). (a) Simulated and exact optical conductivities $\sigma(\omega)$. (b) Simulated and exact current-current correlation functions $\Pi(\tau)$. The calculated results are illustrated in Fig. 7. From Fig. 7(a), it is clear that the main features of optical conductivity are successfully captured by the StochSK solver. Both the sharp Drude peak at $\omega=0$ and a broad satellite peak around $\omega=3.0$ are well reproduced. As is seen in Fig. 7(b), the reconstructed $\tilde{\Pi}(\tau)$ coincides with the original $\Pi(\tau)$. ## 7 Concluding remarks In this paper, a full-fledged analytical continuation toolkit named ACFlow is presented. It implements several primary analytical continuation methods, including the maximum entropy method, stochastic analytical continuation, and stochastic optimization method. It provides quite a few validation and diagnostic tools. It can be used with great flexibility for the analytical continuations of arbitrary fermionic and bosonic correlation functions generated by finite-temperature quantum Monte Carlo simulations. Note that analytical continuation problem is a hotspot in computational physics and many-body physics all the time. Many efforts have been devoted to solve it in recent years. Noticeable achievements include maximum quantum entropy method [44], Nevanlinna analytical continuation [45, 46], blocked-mode sampling and grid point sampling in stochastic analytical continuation [35, 36], constrained stochastic analytical continuation [31, 34], machine learning assisted analytical continuation [18, 19], and so on. We would like to incorporate these new progresses into the ACFlow toolkit in the near future. Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. Data availability The data that support the findings of this study will be made available upon reasonable requests to the corresponding author. Acknowledgement This work is supported by the CAEP Foundation (under Grant No. CX100000) and the National Natural Science Foundation of China (under Grants No. 11874329 and No. 11934020). ## References * [1] J. Carlson, S. Gandolfi, F. Pederiva, S. C. Pieper, R. Schiavilla, K. E. Schmidt, R. B. Wiringa, Quantum monte carlo methods for nuclear physics, Rev. Mod. Phys. 87 (2015) 1067–1118. doi:10.1103/RevModPhys.87.1067. * [2] W. M. C. Foulkes, L. Mitas, R. J. Needs, G. Rajagopal, Quantum monte carlo simulations of solids, Rev. Mod. Phys. 73 (2001) 33–83. doi:10.1103/RevModPhys.73.33. * [3] E. Gull, A. J. Millis, A. I. Lichtenstein, A. N. Rubtsov, M. Troyer, P. Werner, Continuous-time monte carlo methods for quantum impurity models, Rev. Mod. Phys. 83 (2011) 349–404. doi:10.1103/RevModPhys.83.349. * [4] J. Gubernatis, N. Kawashima, P. Werner, Quantum Monte Carlo Methods: Algorithms for Lattice Models, Cambridge University Press, 2016. doi:10.1017/CBO9780511902581. * [5] O. Gunnarsson, M. W. Haverkort, G. Sangiovanni, Analytical continuation of imaginary axis data for optical conductivity, Phys. Rev. B 82 (2010) 165125. doi:10.1103/PhysRevB.82.165125. * [6] C. E. Creffield, E. G. Klepfish, E. R. Pike, S. Sarkar, Spectral weight function for the half-filled hubbard model: A singular value decomposition approach, Phys. Rev. Lett. 75 (1995) 517–520. doi:10.1103/PhysRevLett.75.517. * [7] H. J. Vidberg, J. W. Serene, Solving the eliashberg equations by means ofn-point padéapproximants, J. Low Temp. Phys. 29 (3) (1977) 179–192. doi:10.1007/BF00655090. * [8] K. S. D. Beach, R. J. Gooding, F. Marsiglio, Reliable padé analytical continuation method based on a high-accuracy symbolic computation algorithm, Phys. Rev. B 61 (2000) 5147–5157. doi:10.1103/PhysRevB.61.5147. * [9] i. c. v. Osolin, R. Žitko, Padé approximant approach for obtaining finite-temperature spectral functions of quantum impurity models using the numerical renormalization group technique, Phys. Rev. B 87 (2013) 245135. doi:10.1103/PhysRevB.87.245135. * [10] J. Schött, I. L. M. Locht, E. Lundin, O. Grånäs, O. Eriksson, I. Di Marco, Analytic continuation by averaging padé approximants, Phys. Rev. B 93 (2016) 075104. doi:10.1103/PhysRevB.93.075104. * [11] J. E. Gubernatis, M. Jarrell, R. N. Silver, D. S. Sivia, Quantum monte carlo simulations and maximum entropy: Dynamics from imaginary-time data, Phys. Rev. B 44 (1991) 6011–6029. doi:10.1103/PhysRevB.44.6011. * [12] M. Jarrell, J. Gubernatis, Bayesian inference and the analytic continuation of imaginary-time quantum monte carlo data, Phys. Rep. 269 (3) (1996) 133–195. doi:https://doi.org/10.1016/0370-1573(95)00074-7. * [13] A. W. Sandvik, Stochastic method for analytic continuation of quantum monte carlo data, Phys. Rev. B 57 (1998) 10287–10290. doi:10.1103/PhysRevB.57.10287. * [14] S. Fuchs, T. Pruschke, M. Jarrell, Analytic continuation of quantum monte carlo data by stochastic analytical inference, Phys. Rev. E 81 (2010) 056701. doi:10.1103/PhysRevE.81.056701. * [15] A. S. Mishchenko, N. V. Prokof’ev, A. Sakamoto, B. V. Svistunov, Diagrammatic quantum monte carlo study of the fröhlich polaron, Phys. Rev. B 62 (2000) 6317–6336. doi:10.1103/PhysRevB.62.6317. * [16] O. Goulko, A. S. Mishchenko, L. Pollet, N. Prokof’ev, B. Svistunov, Numerical analytic continuation: Answers to well-posed questions, Phys. Rev. B 95 (2017) 014102. doi:10.1103/PhysRevB.95.014102. * [17] J. Otsuki, M. Ohzeki, H. Shinaoka, K. Yoshimi, Sparse modeling approach to analytical continuation of imaginary-time quantum monte carlo data, Phys. Rev. E 95 (2017) 061302. doi:10.1103/PhysRevE.95.061302. * [18] R. Fournier, L. Wang, O. V. Yazyev, Q. Wu, Artificial neural network approach to the analytic continuation problem, Phys. Rev. Lett. 124 (2020) 056401. doi:10.1103/PhysRevLett.124.056401. * [19] H. Yoon, J.-H. Sim, M. J. Han, Analytic continuation via domain knowledge free machine learning, Phys. Rev. B 98 (2018) 245101. doi:10.1103/PhysRevB.98.245101. * [20] L.-F. Arsenault, R. Neuberg, L. A. Hannah, A. J. Millis, Projected regression method for solving fredholm integral equations arising in the analytic continuation problem of quantum physics, Inverse Problems 33 (11) (2017) 115007\. doi:10.1088/1361-6420/aa8d93. * [21] R. Levy, J. LeBlanc, E. Gull, Implementation of the maximum entropy method for analytic continuation, Comput. Phys. Commun. 215 (2017) 149–155. doi:https://doi.org/10.1016/j.cpc.2017.01.018. * [22] D. Bergeron, A.-M. S. Tremblay, Algorithms for optimized maximum entropy and diagnostic tools for analytic continuation, Phys. Rev. E 94 (2016) 023303. doi:10.1103/PhysRevE.94.023303. * [23] J. Kaufmann, K. Held, ana_cont: Python package for analytic continuation, Comput. Phys. Commun. 282 (2023) 108519. doi:https://doi.org/10.1016/j.cpc.2022.108519. * [24] I. Krivenko, M. Harland, Triqs/som: Implementation of the stochastic optimization method for analytic continuation, Comput. Phys. Commun. 239 (2019) 166–183. doi:https://doi.org/10.1016/j.cpc.2019.01.021. * [25] I. Krivenko, A. S. Mishchenko, Triqs/som 2.0: Implementation of the stochastic optimization with consistent constraints for analytic continuation, Comput. Phys. Commun. 280 (2022) 108491. doi:https://doi.org/10.1016/j.cpc.2022.108491. * [26] F. F. Assaad, M. Bercx, F. Goth, A. Götz, J. S. Hofmann, E. Huffman, Z. Liu, F. P. Toldin, J. S. E. Portela, J. Schwab, The ALF (Algorithms for Lattice Fermions) project release 2.0. Documentation for the auxiliary-field quantum Monte Carlo code, SciPost Phys. Codebases (2022) 1doi:10.21468/SciPostPhysCodeb.1. * [27] L. Boehnke, H. Hafermann, M. Ferrero, F. Lechermann, O. Parcollet, Orthogonal polynomial representation of imaginary-time green’s functions, Phys. Rev. B 84 (2011) 075145. doi:10.1103/PhysRevB.84.075145. * [28] R. K. Bryan, Maximum entropy analysis of oversampled data problems, Eur. Biophys. J. 18 (3) (1990) 165–174. doi:10.1007/BF02427376. * [29] K. S. D. Beach, Identifying the maximum entropy method as a special limit of stochastic analytic continuation (2004). arXiv:0403055. * [30] H. Shao, Y. Q. Qin, S. Capponi, S. Chesi, Z. Y. Meng, A. W. Sandvik, Nearly deconfined spinon excitations in the square-lattice spin-$1/2$ heisenberg antiferromagnet, Phys. Rev. X 7 (2017) 041072. doi:10.1103/PhysRevX.7.041072. * [31] A. W. Sandvik, Constrained sampling method for analytic continuation, Phys. Rev. E 94 (2016) 063308. doi:10.1103/PhysRevE.94.063308. * [32] Y. Q. Qin, B. Normand, A. W. Sandvik, Z. Y. Meng, Amplitude mode in three-dimensional dimerized antiferromagnets, Phys. Rev. Lett. 118 (2017) 147207\. doi:10.1103/PhysRevLett.118.147207. * [33] S. R. White, D. J. Scalapino, R. L. Sugar, N. E. Bickers, Monte carlo calculation of dynamical properties of the two-dimensional hubbard model, Phys. Rev. Lett. 63 (1989) 1523–1526. doi:10.1103/PhysRevLett.63.1523. * [34] H. Shao, A. W. Sandvik, Progress on stochastic analytic continuation of quantum monte carlo data (2022). arXiv:2202.09870. * [35] K. Ghanem, E. Koch, Average spectrum method for analytic continuation: Efficient blocked-mode sampling and dependence on the discretization grid, Phys. Rev. B 101 (2020) 085111. doi:10.1103/PhysRevB.101.085111. * [36] K. Ghanem, E. Koch, Extending the average spectrum method: Grid point sampling and density averaging, Phys. Rev. B 102 (2020) 035114. doi:10.1103/PhysRevB.102.035114. * [37] K. Vafayi, O. Gunnarsson, Analytical continuation of spectral data from imaginary time axis to real frequency axis using statistical sampling, Phys. Rev. B 76 (2007) 035115. doi:10.1103/PhysRevB.76.035115. * [38] O. F. Syljuåsen, Using the average spectrum method to extract dynamics from quantum monte carlo simulations, Phys. Rev. B 78 (2008) 174429. doi:10.1103/PhysRevB.78.174429. * [39] S. Kirkpatrick, C. D. Gelatt, M. P. Vecchi, Optimization by simulated annealing, Science 220 (1983) 671–680. doi:10.1126/science.220.4598.671. * [40] E. Marinari, Optimized monte carlo methods (1996). arXiv:9612010. * [41] A. Georges, G. Kotliar, W. Krauth, M. J. Rozenberg, Dynamical mean-field theory of strongly correlated fermion systems and the limit of infinite dimensions, Rev. Mod. Phys. 68 (1996) 13–125. doi:10.1103/RevModPhys.68.13. * [42] L. Huang, Y. Wang, Z. Y. Meng, L. Du, P. Werner, X. Dai, iqist: An open source continuous-time quantum monte carlo impurity solver toolkit, Comput. Phys. Commun. 195 (2015) 140–160. doi:https://doi.org/10.1016/j.cpc.2015.04.020. * [43] L. Huang, iqist v0.7: An open source continuous-time quantum monte carlo impurity solver toolkit, Comput. Phys. Commun. 221 (2017) 423–424. doi:https://doi.org/10.1016/j.cpc.2017.08.026. * [44] J.-H. Sim, M. J. Han, Maximum quantum entropy method, Phys. Rev. B 98 (2018) 205102\. doi:10.1103/PhysRevB.98.205102. * [45] J. Fei, C.-N. Yeh, E. Gull, Nevanlinna analytical continuation, Phys. Rev. Lett. 126 (2021) 056402. doi:10.1103/PhysRevLett.126.056402. * [46] J. Fei, C.-N. Yeh, D. Zgid, E. Gull, Analytical continuation of matrix-valued functions: Carathéodory formalism, Phys. Rev. B 104 (2021) 165111. doi:10.1103/PhysRevB.104.165111.
# A fully asymptotic preserving decomposed multi-group method for the frequency-dependent radiative transfer equations Xiaojiang Zhang<EMAIL_ADDRESS>Peng Song<EMAIL_ADDRESS>Institute of Applied Physics and Computational Mathematics, Beijing, 100094, P.R. China. HEDPS, Center for Applied Physics and Technology, College of Engineering, Peking University, Beijing, 100871, P.R. China. Yi Shi <EMAIL_ADDRESS>School of Mathematics, Shandong University, Jinan, 250100, P.R. China. Min Tang<EMAIL_ADDRESS> ###### Abstract The opacity of FRTE depends on not only the material temperature but also the frequency, whose values may vary several orders of magnitude for different frequencies. The gray radiation diffusion and frequency-dependent diffusion equations are two simplified models that can approximate the solution to FRTE in the thick opacity regime. The frequency discretization for the two limit models highly affects the numerical accuracy. However, classical frequency discretization for FRTE considers only the absorbing coefficient. In this paper, we propose a new decomposed multi-group method for frequency discretization that is not only AP in both gray radiation diffusion and frequency-dependent diffusion limits, but also the frequency discretization of the limiting models can be tuned. Based on the decomposed multi-group method, a full AP scheme in frequency, time, and space is proposed. Several numerical examples are used to verify the performance of the proposed scheme. ## 1 Introduction The frequency-dependent radiative transfer equation (FRTE) describes the time evolution of the probability density function of photons that transport and interact with the background material, which plays an important role in astrophysics, inertial confinement fusion (ICF), and high energy physics. Due to the complexity of the FRTE and its wide applications, FRTE has attracted the attention of many physicists and mathematicians. The numerical simulation of FRTE is challenging due to the following difficulties: 1) The high dimensionality. The photon probability distribution function depends on space, time, moving velocity and photon frequency, which is a 7 dimensional function. For the simplest slab geometry, the dimension is 4, which is hard to find an efficient solver due to the ”curse of dimensionality”; 2) Multiscale parameters. Absorbing and scattering coefficients are two important material properties that influence the photons’ movements. The absorbing coefficient $\sigma_{a}$ (scattering coefficient $\sigma_{s}$) gives the probability that photons will be absorbed (scattered) by the background material. Both $\sigma_{a}$ and $\sigma_{s}$ may depend on the material type, material temperature and the photon frequency. When $\sigma_{a}+\sigma_{s}$ is large, the mean free path (the average distance between two successive scattering or absorption) of the photon is small, the material is referred to as ”optical thick”, while when $\sigma_{a}+\sigma_{s}$ events is at $O(1)$, the material is ”optical thin”. For photons with different frequencies, the values of $\sigma_{a}$ and $\sigma_{s}$ may vary several orders of magnitude, which induces the multiscale property of the system. To guarantee both accuracy and stability, classical schemes require very fine space and time steps, which leads to high computational costs. 3) The strong nonlinearity. The material opacity and the Planck function which gives the equilibrium radiation frequency distribution depend nonlinearly on the material temperature. If one needs an efficient solver that allows for large time steps, some terms have to be treated implicitly and thus nonlinear iterations are required. However, the nonlinear iterations for high dimensional equations are extremely expensive. One popular strategy for the FRTE simulation is the stochastic method [steinberg2022multi, densmore2012hybrid] based on the Fleck-Cummings’ implicit Monte Carlo (IMC) method [FC71]. However, it requires a large number of sampling particles and suffers from unavoidable statistical fluctuations. On the other hand, there are deterministic solvers for the radiative transfer equation (RTE), which is a simplified version of FRTE that ignores the frequency dependence. Deterministic solvers include Discontinuous Galerkin method [lcs, bailey2008piecewise], photon free method [chang2007incorporation, chang2007deterministic], diffusion-synthetic acceleration method (DSA) [H1993, Larsen1988, ALarsen2002] etc. and some of them have been extended to FRTE. In order to deal with the multiscale parameters in FRTE, two strategies are proposed recently. One is designing asymptotic preserving (AP) schemes. The AP scheme provides a general framework of solving the difficulty coming from multiscale parameters. A scheme is AP when the asymptotic limit of the discretization becomes a stable solver for the limit model as the scaling parameter in the microscopic equation goes to zero [jin20101, hu2017asymptotic, 2018Tang]. In last decade, research on the AP schemes has been developed rapidly for various applications and we refer interesting readers to a recent review paper [jin_2022]. The other approach is the moment- based acceleration scheme which is called High-Order/Low-Order (HOLO) algorithm [chacon2017multiscale, yee2017stable, park2019multigroup]. In the HOLO algorithm, a low-order system which consists of the first two moments of the RTE and the material temperature equation is solved. Then the RTE is solved by determining the Planckian emission source term using the temperature obtained in the lower-order system. The benefit of HOLO algorithms is nonlinear elimination, but their long-term accuracy and nonlinear stability require discrete consistency between HO and LO formulations [chacon2017multiscale]. Both approaches are well developed for RTE simulations. Designing AP schemes for the nonlinear radiation hydrodynamic system is a popular topic recently. For the gray radiative transfer equations (GRTE) in the gray radiative diffusion limit, several AP schemes can be found in the literature. An AP scheme is constructed in [2001Klar] by decomposing the distribution function into the equilibrium and non-equilibrium parts; in [2020park], the authors developed an AP-HOLO algorithm based on the linear discontinuous Galerkin HO transport solver and the corner-balance LO solver; in [sun2015asymptotic], an unified gas kinetics scheme (AP-UGKS) was proposed. However, in order to use an opacity independent time step, a linearized iterative solver involving the radiation equation has to be employed for the time-implicit part. In [tang2021accurate], an AP accurate front capturing scheme was proposed that allows both space and time steps being independent of the opacity and only requires a scalar Newton’s solver. Later on, in order to get AP schemes for more complex applications, UGKS has been extended to FRTE [sunj] and the radiation magnetohydrodynamics (RMHD) system [jin2022spatial, sun2020multiscale]; The AP linear-discontinuous spatial differencing scheme has been extended to FRTE in [MWTS]; The idea of intensity decomposition used in [tang2021accurate] has been extended to the RMHD system in [jin2022spatial]. However, none of the above mentioned work considered about the improvement of the frequency discretization. [terh] separates particles into several groups. It is assumed that particles in the same group have the same frequency and the same absorption and scattering coefficients. However, due to the limitation of computational resources, the number of groups is usually not enough to resolve all variances in the frequency distribution. On the other hand, when the opacity is thick, there are two approximating models, one is the gray radiation diffusion model where the particles are at the equilibrium Planck distribution in frequency; the other is the frequency-dependent diffusion limit (FDDL), where particles with different frequencies diffuse at different speeds. In order to use a small number of groups to provide a correct solution behavior, different frequency discretizations of the limit models are proposed in the literature, for example, the piece-constant approximation [chang2007deterministic, chang2007incorporation], the Rosseland mean approximation [MWTS, yee2017stable], the Planck approximation [sunj, sunj18, 2017An]. However, it is not clear how to construct a group discretization for the FRTE whose diffusion limits can coincide with the group discretization of the limiting models. In this paper, by decomposing the intensity into three parts, we obtain a new decomposed multi-group method whose asymptotic limit, when the mean free path goes to zero, can be consistent with the required frequency discretization of both limiting models. To the best of our knowledge, the decomposed multi-group method is the first work that gives an AP discretization in the frequency domain. Based on the new decomposed multi-group method, a full AP scheme in frequency, space, and time is proposed. The scheme accuracy is guaranteed in both optically thin and thick regimes when large spatial and time steps are used. Moreover, nonlinear iterations are employed to solve a system with only macroscopic quantities. Then linear transport equations with decoupled direction and energy groups are solved. Thus the computational cost is similar to HOLO [park2019multigroup]. The organization of this paper is as follows. Section 2 shows two different asymptotic limits of the FRTE, one is the gray radiative diffusion limit, and the other is the FDDL. The classical multi-group method and our decomposed multi-group method are compared in the frequency-dependent optical thick regime in Sections 3 and 4. The semi-discretization that is AP in frequency and time is proposed in section 5, and the fully discretized scheme that is AP in frequency, time, and space is displayed in section 6. Their AP properties in both gray radiation and frequency-dependent diffusion limits are shown by asymptotic analysis. In section 7, some numerical tests are given to illustrate the stability and accuracy of our proposed scheme. Finally, the paper is concluded with some discussions in Section 8. ## 2 The FRTE and its limit models ### 2.1 The model and nondimensionalization The FRTE under consideration is as in [2017An], that writes: $\left\\{\begin{aligned} &\frac{1}{c}\partial_{t}I+\Omega\cdot\nabla I=\sigma_{a}\left(B(\nu,T)-I\right)+\sigma_{s}\left(\rho-I\right),\qquad\rho=\frac{1}{4\pi}\int_{4\pi}I\text{d}\Omega:=\left\langle I\right\rangle,\\\ &C_{v}\partial_{t}T\equiv\partial_{t}U=\int\limits_{4\pi}\int\limits_{0}^{\infty}\sigma_{a}\left(I-B(\nu,T)\right)\text{d}\nu\text{d}\Omega,\ \ \\\ \end{aligned}\right.$ (2.1) where $x$ is the spatial variable; $\Omega$ is the angular variable; $\nu\in(0,+\infty)$ is the frequency variable; $I(t,x,\Omega,\nu)$ is the radiation intensity; $T(t,x)$ is the material temperature; $\sigma_{a}(x,\nu,T)$ and $\sigma_{s}(x,\nu,T)$ are the absorbing and scattering coefficient respectively; $c$ and $C_{v}$ represent respectively the light speed and the specific heat capacity; the Planck function $B(\nu,T)$ is defined by $B(\nu,T)=\frac{2h\nu^{3}}{c^{2}}\frac{1}{e^{h\nu/kT}-1},$ with Boltzmann’s constant $k$ and Planck’s constant $h$. The Planck function satisfies $\int\limits_{0}^{\infty}4\pi B\ \text{d}\nu=a_{r}cT^{4},\qquad\int\limits_{0}^{\infty}4\pi\frac{\partial B}{\partial T}\ \text{d}\nu=4a_{r}cT^{3},$ with the radiation constant $a_{r}$. The FRTE system (2.1) is composed of two equations, the first one describes the time evolution of radiation intensity and the second one gives the time evolution of background material temperature. If we integrate the first equation in system (2.1) with respect to the frequency $\nu$ from $0$ to $+\infty$ and assume that the absorbing and scattering coefficients are independent of the frequency, the GRTE can be obtained. GRTE is a good approximation to FRTE when the media is assumed to be gray, i.e. its properties are independent of the radiation spectrum [traugott1968grey]. However, in reality, media opacity varies a lot for different temperatures, densities and particle frequencies. FRTE has to be considered in some situations [vaytet2011numerical, vaytet2013influence]. We then consider the dimensionless form of (2.1). Similar as in [LREJ, HFJM], we consider the following nondimensionalization: $x=\widehat{x}\ell_{\infty},\quad t=\widehat{t}t_{\infty},\quad\nu=\widehat{\nu}\frac{kT_{\infty}}{h},\quad\sigma_{a}=\widehat{\sigma}_{a}/\ell_{\infty},\quad\sigma_{s}=\widehat{\sigma}_{s}/\ell_{\infty},$ $I=\widehat{I}\frac{a_{r}chT_{\infty}^{3}}{k},\quad B=\widehat{B}\frac{a_{r}chT_{\infty}^{3}}{k},\quad\rho=\widehat{\rho}\frac{a_{r}chT_{\infty}^{3}}{k},\quad U=\widehat{U}\frac{\ell_{\infty}^{2}}{t_{\infty}^{2}},$ where variables with a hat denote nondimensional quantity and variables with $\infty-$subscript are the characteristic value with unit. More precisely, $\ell_{\infty}$, $t_{\infty}$ and $T_{\infty}$ are respectively the reference length, time and temperature, and $a_{r}$ is the radiation constant. Then the full dimensionless FRTE becomes $\left\\{\begin{aligned} &\frac{1}{\mathcal{C}}\partial_{\widehat{t}}\widehat{I}+\Omega\cdot\nabla_{\widehat{x}}\widehat{I}=\widehat{\sigma}_{a}\mathscr{L}_{a}\left(\widehat{B}-\widehat{I}\right)+\widehat{\sigma}_{s}\mathscr{L}_{s}\left(\widehat{\rho}-\widehat{I}\right),\\\ &\widehat{C}_{v}\partial_{\widehat{t}}\widehat{T}\equiv\partial_{\widehat{t}}\widehat{U}=\mathcal{C}\mathcal{P}_{0}\int\limits_{4\pi}\int\limits_{0}^{\infty}\widehat{\sigma}_{a}\mathscr{L}_{a}\left(\widehat{I}-\widehat{B}\right)\text{d}\widehat{\nu}\text{d}\Omega,\ \ \\\ \end{aligned}\right.$ (2.2) where $\mathcal{C}=\frac{ct_{\infty}}{\ell_{\infty}}$, $\mathcal{P}_{0}=\frac{a_{r}t_{\infty}^{2}T_{\infty}^{4}}{\ell^{2}_{\infty}}$, $\widehat{C}_{v}$ is a positive constant and $\widehat{\sigma}_{a}\mathscr{L}_{a}$, $\widehat{\sigma}_{s}\mathscr{L}_{s}$ are respectively $\sigma_{a}$ and $\sigma_{s}$. According to [LMH], $\mathcal{P}_{0}$ is a nondimensional parameter that measures radiation effects on the matter, which is always considered as $\mathcal{P}_{0}=\mathcal{O}(1)$ when a moderate amount of radiation in the matter is considered. $\mathcal{C}=\mathcal{O}(1/\varepsilon)$ for nonrelativistic case, which indicates that the speed of light travels much faster than the material velocity when $\varepsilon\ll 1$. Thus, in this paper, we fix the values of $\mathcal{P}_{0}$ and $\mathcal{C}$ as $\mathcal{P}_{0}=1,\qquad\mathcal{C}=1/\varepsilon.$ (2.3) Depending on the scales of $\mathscr{L}_{a}$ and $\mathscr{L}_{s}$, the solution of system (2.1) tends to solutions of different limit systems in the diffusive regime [GPG]. How the different scales of $\mathscr{L}_{a}$ and $\mathscr{L}_{s}$ lead to different limits will be discussed in the subsequent part. For simplicity, the hat in (2.2) will be dropped later on. ### 2.2 The diffusion limits of the FRTE We focus on two different regimes for the FRTE, one is the gray radiation diffusion regime and the other is the frequency dependent diffusion regime, whose derivations are similar as in [GPG, LMH]. The gray radiation diffusion limit is valid when collisions are so frequent that the frequency distribution depends predominantly on the local temperature. In the FDDL, the determination of the frequency distribution becomes complicated, it depends on both the local temperature and the state the radiation field [Holland1969, manuel2001non]. Both limiting models are widely used in the simulations of RMHD problems [CHENG2020109724, CD20, jin2022spatial]. For the convenience of readers, the details of the derivation are in Appendix A. Using (2.3), the FRTE (2.1) becomes: $\displaystyle\varepsilon\partial_{t}I+\Omega\cdot\nabla I=\mathscr{L}_{a}\sigma_{a}\left(B(\nu,T)-I\right)+\mathscr{L}_{s}\sigma_{s}\left(\rho(t,x,\nu)-I\right),$ (2.4a) $\displaystyle\varepsilon C_{v}\partial_{t}T=\int\limits_{4\pi}\int\limits_{0}^{\infty}\mathscr{L}_{a}\sigma_{a}\left(I-B(\nu,T)\right)\text{d}\nu\text{d}\Omega.$ (2.4b) * • The gray radiation diffusion regime. In this regime, the radiation intensity $I$ adapts to the material temperature. The following three different scalings yield the gray radiation diffusion equation: $\displaystyle\mathscr{L}_{a}=1/\varepsilon,\quad\mathscr{L}_{s}=\varepsilon,$ (2.5a) $\displaystyle\mathscr{L}_{a}=1,\quad\mathscr{L}_{s}=1/\varepsilon,$ (2.5b) $\displaystyle\mathscr{L}_{a}=1/\varepsilon,\quad\mathscr{L}_{s}=1/\varepsilon.$ (2.5c) When $\varepsilon\to 0$ in (2.4), the material temperature $T$ can be approximated by the solution of the following gray radiation diffusion equation $\partial_{t}(T^{(0)})^{4}+C_{v}\partial_{t}T^{(0)}=\nabla\cdot\left(\frac{D_{d}}{\sigma}\nabla(T^{(0)})^{4}\right)\,,$ (2.6) with $D_{d}=\frac{1}{3}I_{d}$ (where $I_{d}$ denotes the 3 by 3 identity matrix) and the mean opacity $\sigma$ being given by $\frac{1}{\sigma(x,T)}\equiv\frac{\int_{0}^{\infty}\frac{4\pi\mathcal{C}}{\mathscr{L}_{a}\sigma_{a}(x,\nu,T)+\mathscr{L}_{s}\sigma_{s}(x,\nu,T)}\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}{\int_{0}^{\infty}4\pi\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}=\frac{\int_{0}^{\infty}\frac{4\pi\mathcal{C}}{\mathscr{L}_{a}\sigma_{a}(x,\nu,T)+\mathscr{L}_{s}\sigma_{s}(x,\nu,T)}\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}{4(T^{(0)})^{3}}.$ (2.7) The radiation intensity $I$ can then be approximated by $B(\nu,T^{(0)})-\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B(\nu,T^{(0)})}{\partial T}\Omega\cdot\nabla T^{(0)}.$ (2.8) * • The frequency dependent diffusion regime. In this regime, the radiation intensity $I$ tends to be independent of $\Omega$, but the frequency distribution can no longer be determined by the material temperature $T$. Similar as in [GPG], we consider the following scaling $\mathscr{L}_{a}=\varepsilon,\quad\mathscr{L}_{s}=1/\varepsilon.$ In this regime, the FRTE (2.4) writes: $\displaystyle\varepsilon\partial_{t}I+\Omega\cdot\nabla I=\varepsilon\sigma_{a}\left(B(\nu,T)-I\right)+\frac{\sigma_{s}}{\varepsilon}\left(\rho(t,x,\nu)-I\right),$ (2.9a) $\displaystyle C_{v}\partial_{t}T=\int\limits_{4\pi}\int\limits_{0}^{\infty}\sigma_{a}\left(I-B(\nu,T)\right)\text{d}\nu\text{d}\Omega.$ (2.9b) When $\varepsilon\to 0$ in (2.9), the solution $I(t,x,\Omega,\nu)\approx\rho^{(0)}(t,x,\nu)$ where $\rho^{(0)}$ and $T$ satisfies the following frequency dependent diffusion system: $\left\\{\begin{aligned} &\partial_{t}\rho^{(0)}-\nabla\cdot\left(\frac{D_{d}}{\sigma_{s}}\nabla\rho^{(0)}\right)=\sigma_{a}(B^{(0)}-\rho^{(0)}),\\\ &C_{v}\partial_{t}T^{(0)}=\int_{0}^{\infty}4\pi\sigma_{a}(\rho^{(0)}-B^{(0)})\text{d}\nu.\ \ \\\ \end{aligned}\right.$ (2.10) ## 3 Extension of the classical multi-group method and the diffusive limits ### 3.1 The classical multi-group method and its extension We first consider the frequency discretization of the system (2.2). In the multi-group method, the continuous frequency space $(0,\infty)$ is divided into $G$ groups, and each frequency interval is denoted by $(\nu_{g-1/2},\nu_{g+1/2})$ ($g=1,\cdots,G$), with $\nu_{\frac{1}{2}}=0$, $\nu_{G+\frac{1}{2}}=+\infty$. By integrating the first equation in system (2.2) over the frequency interval $(\nu_{g-1/2},\nu_{g+1/2})$ yields $\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\frac{1}{\mathcal{C}}\partial_{t}I+\Omega\cdot\nabla I\text{d}\nu=\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\mathscr{L}_{a}\sigma_{a}\left(B(\nu,T)-I\right)+\mathscr{L}_{s}\sigma_{s}\left(\rho-I\right)\text{d}\nu.\\\ $ (3.1) Then the multi-group discretization of FRTE becomes: $\displaystyle\frac{1}{\mathcal{C}}\partial_{t}I_{g}+\Omega\cdot\nabla I_{g}=\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\mathscr{L}_{a}\sigma_{a}\left(B-I\right)-\mathscr{L}_{s}\sigma_{s}\left(\rho-I\right)\text{d}\nu,\quad(g=1,\cdots,G),$ (3.2a) $\displaystyle C_{v}\partial_{t}T=\mathcal{C}\mathcal{P}_{0}\int\limits_{4\pi}\sum\limits_{g^{\prime}=1}^{G}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\left[\mathscr{L}_{a}\sigma_{a}\left(I-B\right)\right]\text{d}\nu\text{d}\Omega,$ (3.2b) in which the radiation intensity $I(t,x,\Omega,\nu)$ and Planck function $B(\nu,T)$ in different groups are given by $I_{g}=\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}I(t,x,\Omega,\nu)\text{d}\nu,\qquad B_{g}=\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}B(\nu,T)\text{d}\nu.$ (3.3) After frequency discretization, only $I_{g}$ in each group and the Planck function $B$ are known, while the full frequency distribution of $I$ is not known. Thus the main difficulty is how to approximate the two terms $\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{a}\left(B-I\right)\text{d}\nu\quad\mbox{and}\quad\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{s}\left(\rho-I\right)\text{d}\nu.$ (3.4) In order to approximate the two terms on the right hand side of (3.1) by $I_{g}$, $B_{g}$ and $T$, some approximations can be found in the literature. We review some classical multi-group frequency discretization for FRTE in the subsequent part. Most works in the literature consider only the case when $\sigma_{s}=0$. In this subsection, we extend similar idea of approximating the term $\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\mathscr{L}_{a}\sigma_{a}\left(B(\nu,T)-I\right)$ in [yee2017stable, MWTS] to the case when $\sigma_{s}\neq 0$. In the classical multi-group frequency discretization in [terh], by defining $\sigma_{a,g}$ and $\sigma_{s,g}$ as follows $\sigma_{a,g}=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{a}(B-I)\mathrm{d}\nu}{B_{g}-I_{g}},\quad\sigma_{s,g}=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{s}(\rho-I)\mathrm{d}\nu}{\rho_{g}-I_{g}},$ (3.5) the multi-group frequency discretization for FDDL writes $\displaystyle\frac{1}{\mathcal{C}}\partial_{t}I_{g}+\Omega\cdot\nabla I_{g}=\mathscr{L}_{a}\sigma_{a,g}\left(B_{g}-I_{g}\right)+\mathscr{L}_{s}\sigma_{s,g}\left(\rho_{g}-I_{g}\right),\qquad(g=1,\cdots,G),$ (3.6a) $\displaystyle C_{v}\partial_{t}T=\mathcal{C}\mathcal{P}_{0}\int\limits_{4\pi}\sum\limits_{g^{\prime}=1}^{G}\left[\mathscr{L}_{a}\sigma_{a,g}\left(I_{g}-B_{g}\right)\right]\text{d}\Omega.$ (3.6b) Here $\sigma_{a,g}$ and $\sigma_{s,g}$ are respectively the mean absorption and scattering coefficients that have to be approximated by $I_{g}$, $B_{g}$ and $T$. In the diffusive regime, different approximations yield different coefficients in the limit equations. Two classical approximations are discussed in the subsequent part. The most straight forward way is to use piece-wise constant approximation and let $B(\nu)\approx B_{g}$ and $\rho(\nu)\approx\rho_{g}$, $I(\nu)\approx I_{g}$ when $\nu\in(\nu_{g-\frac{1}{2}},\nu_{g+\frac{1}{2}})$. Then $\sigma_{a,g}\approx\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{a}\mathrm{d}\nu,\qquad\sigma_{s,g}\approx\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{s}\mathrm{d}\nu.$ (3.7) In [yee2017stable, MWTS] and some related work later on, the authors considered the Rosseland mean for which $\sigma_{a,g}$ are determined from the approximation in (2.8) which yields $B-I\approx\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}(T)\Omega\cdot\nabla T,$ and the Rosseland mean is $\sigma_{a,g}=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\sigma_{a}}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}(T)\mathrm{~{}d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}(T)\mathrm{~{}d}\nu}.$ (3.8) Extending similar idea to the approximation of $\sigma_{s,g}$, the most straight forward way is to use the following approximation $\rho-I\approx-\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\Omega\cdot\nabla\rho.$ If the intensity $I$ has Planck distribution in the frequency domain, i.e. $\rho\approx I\approx B(\nu,T_{r})$, where the radiation temperature $T_{r}$ is given by $T_{r}^{4}=\int_{4\pi}\int_{0}^{+\infty}I\mathrm{d}\Omega\mathrm{d}\nu=4\pi\int_{0}^{\infty}\rho\mathrm{d}\nu=4\pi\sum\limits_{g^{\prime}=1}^{G}\rho_{g},$ (3.9) the scattering coefficient $\sigma_{s,g}$ can then be approximated by $\displaystyle\sigma_{s,g}$ $\displaystyle=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{s}(\rho-I)\mathrm{d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}(\rho-I)\mathrm{~{}d}\nu}\approx\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\sigma_{s}}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\Omega\cdot\nabla\rho\mathrm{d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\Omega\cdot\nabla\rho\mathrm{~{}d}\nu}$ (3.10) $\displaystyle=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\sigma_{s}}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}(T_{r})\Omega\cdot\nabla T_{r}\mathrm{d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}(T_{r})\Omega\cdot\nabla T_{r}\mathrm{~{}d}\nu}=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\sigma_{s}}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}(T_{r})\mathrm{~{}d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{a}}\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}(T_{r})\mathrm{~{}d}\nu}.$ It is easy to see that the Rosseland mean is an approximation to the average of $\sigma_{a}$, $\sigma_{s}$ inside the interval $(\nu_{g+\frac{1}{2}},\nu_{g+\frac{1}{2}})$ as well. ### 3.2 The diffusion limit of the classical multi-group method #### 3.2.1 The gray radiation diffusion regime When $\mathscr{L}_{a}=1/\varepsilon$, $\mathscr{L}_{s}=1/\varepsilon$, the multi-group frequency discretization in (3.6) becomes $\displaystyle\varepsilon\partial_{t}I_{g}+\Omega\cdot\nabla I_{g}=\frac{\sigma_{a,g}}{\varepsilon}\left(B_{g}-I_{g}\right)+\frac{\sigma_{s,g}}{\varepsilon}\left(\rho_{g}-I_{g}\right),\qquad(g=1,\cdots,G),$ (3.11a) $\displaystyle\varepsilon C_{v}\partial_{t}T=4\pi\sum\limits_{g^{\prime}=1}^{G}\left[\frac{\sigma_{a,g^{\prime}}}{\varepsilon}\left(\rho_{g^{\prime}}-B_{g^{\prime}}\right)\right].$ (3.11b) We derive its gray radiation diffusion limit equation in the subsequent part. By using Chapman-Enskog expansion in (3.11a), one has $\displaystyle I_{g}$ $\displaystyle=\frac{\sigma_{a,g}}{\sigma_{a,g}+\sigma_{s,g}}B_{g}+\frac{\sigma_{s,g}}{\sigma_{a,g}+\sigma_{s,g}}\rho_{g}-\frac{\varepsilon}{\sigma_{a,g}+\sigma_{s,g}}\left[\varepsilon\partial_{t}I_{g}+\Omega\cdot\nabla I_{g}\right]$ (3.12) $\displaystyle=\frac{\sigma_{a,g}}{\sigma_{a,g}+\sigma_{s,g}}B_{g}+\frac{\sigma_{s,g}}{\sigma_{a,g}+\sigma_{s,g}}\rho_{g}-\frac{\varepsilon}{\sigma_{a,g}+\sigma_{s,g}}\Omega\cdot\nabla\left(\frac{\sigma_{a,g}}{\sigma_{a,g}+\sigma_{s,g}}B_{g}+\frac{\sigma_{s,g}}{\sigma_{a,g}+\sigma_{s,g}}\rho_{g}\right)+\mathcal{O}\left(\varepsilon^{2}\right).$ Taking the integral with respect to $\Omega$ of equation (3.11a), we can obtain $\rho_{g}=B_{g}+\mathcal{O}(\varepsilon)\,,$ which implies that $\rho_{g}\approx B_{g}$. Taking the integral with respect to $\Omega$ of the equation (3.11a), adding up all groups and (3.11b), one can get $4\pi\partial_{t}\left(\sum_{g=1}^{G}\rho_{g}\right)+\frac{4\pi}{\varepsilon}\sum_{g=1}^{G}\nabla\cdot\left\langle\Omega I_{g}\right\rangle+C_{v}\partial_{t}T=0\,.$ (3.13) Plugging (3.12) into the equation (3.13), using the condition $\rho_{g}=B_{g}+\mathcal{O}(\varepsilon)$, letting $\varepsilon\to 0$, then (3.13) reduces to $\partial_{t}\left(\sum_{g=1}^{G}4\pi B_{g}\right)+C_{v}\partial_{t}T=\sum_{g=1}^{G}\nabla\cdot\left(\frac{4\pi}{3(\sigma_{a,g}+\sigma_{s,g})}\nabla B_{g}\right)\,,$ (3.14) which implies that $\partial_{t}\left(\sum_{g=1}^{G}4\pi B_{g}\right)+C_{v}\partial_{t}T=\sum_{g=1}^{G}\nabla\cdot\left(\frac{4\pi}{3(\sigma_{a,g}+\sigma_{s,g})}\frac{\partial B_{g}}{\partial T}\frac{1}{4T^{3}}\nabla T^{4}\right)\,,$ (3.15) When the group discretization of the Planck equilibrium $B$ satisfies the following relations $4\pi\sum_{g=1}^{G}B_{g}=\int\limits_{0}^{\infty}4\pi B\ \text{d}\nu=T^{4},$ (3.16) (3.15) then reduces to $\partial_{t}T^{4}+C_{v}\partial_{t}T=\nabla\cdot\left(\frac{1}{3\widehat{\sigma}_{t}}\nabla T^{4}\right)\,,$ (3.17) When one approximates $\sigma_{a,g}$ and $\sigma_{s,g}$ as in (3.7), the diffusion coefficient of the above equation is determined by $\frac{1}{\widehat{\sigma}^{c}_{t}}=\frac{1}{4T^{3}}\left(\sum\limits_{g=1}^{G}\frac{4\pi}{\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int_{\nu_{g-1/2}}^{\nu_{g+1/2}}{\sigma_{a}+\sigma_{s}}\text{d}\nu}\frac{\partial B_{g}}{\partial T}\right).$ (3.18) On the other hand, when one uses the Rosseland mean as in (3.8) and (3.10), one has $\frac{1}{\sigma_{a,g}+\sigma_{s,g}}=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}(T)\mathrm{~{}d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T)\mathrm{~{}d}\nu},$ which implies $\frac{1}{\widehat{\sigma}^{r}_{t}}=\frac{1}{4T^{3}}\left(\sum\limits_{g=1}^{G}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\frac{4\pi}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\right).$ (3.19) Both $\widehat{\sigma}_{t}^{c}$ and $\widehat{\sigma}_{t}^{r}$ provide reasonable approximations to $\sigma$ in (2.7) when the $\sigma_{a}$, $\sigma_{s}$ are at $O(1)$ and when there are enough groups. However, we will see in the next subsection that $\widehat{\sigma}^{c}_{t}$ determined by piece-wise constant approximation as in (3.18) can not provide a good approximation to the the diffusion coefficient as in (2.7), while the Rosseland mean $\widehat{\sigma}^{r}_{t}$ can. Moreover, When $\sigma_{s}=0$, $\frac{1}{\widehat{\sigma}^{r}_{t}}$ becomes $\frac{1}{\widehat{\sigma}^{r}_{a}}=\frac{1}{4T^{3}}\left(\sum\limits_{g=1}^{G}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\frac{4\pi}{\sigma_{a}}\frac{\partial B}{\partial T}\text{d}\nu\right).$ which is the same as in [MWTS]. #### 3.2.2 The frequency dependent diffusion regime When $\mathscr{L}_{a}=\varepsilon$, $\mathscr{L}_{s}=1/\varepsilon$, the multi-group frequency discretization in (3.6) becomes $\displaystyle\varepsilon\partial_{t}I_{g}+\Omega\cdot\nabla I_{g}=\varepsilon\sigma_{a,g}\left(B_{g}-I_{g}\right)+\frac{\sigma_{s,g}}{\varepsilon}\left(\rho_{g}-I_{g}\right),\qquad(g=1,\cdots,G),$ (3.20a) $\displaystyle\varepsilon C_{v}\partial_{t}T=4\pi\sum\limits_{g^{\prime}=1}^{G}\left[\varepsilon\sigma_{a,g^{\prime}}\left(\rho_{g^{\prime}}-B_{g^{\prime}}\right)\right].$ (3.20b) We derive its radiation diffusion limit equation in the subsequent part. By using Chapman-Enskog expansion in (3.20a), one has $\displaystyle I_{g}$ $\displaystyle=\frac{\varepsilon^{2}\sigma_{a,g}}{\varepsilon^{2}\sigma_{a,g}+\sigma_{s,g}}B_{g}+\frac{\sigma_{s,g}}{\varepsilon^{2}\sigma_{a,g}+\sigma_{s,g}}\rho_{g}-\frac{\varepsilon}{\varepsilon^{2}\sigma_{a,g}+\sigma_{s,g}}\left[\varepsilon\partial_{t}I_{g}+\Omega\cdot\nabla I_{g}\right]$ (3.21) $\displaystyle=\frac{\sigma_{s,g}}{\varepsilon^{2}\sigma_{a,g}+\sigma_{s,g}}\rho_{g}-\frac{\varepsilon}{\varepsilon^{2}\sigma_{a,g}+\sigma_{s,g}}\Omega\cdot\nabla\left(\frac{\sigma_{s,g}}{\varepsilon^{2}\sigma_{a,g}+\sigma_{s,g}}\rho_{g}\right)+\mathcal{O}\left(\varepsilon^{2}\right)$ $\displaystyle=\rho_{g}-\frac{\varepsilon}{\sigma_{s,g}}\Omega\cdot\nabla\rho_{g}+\mathcal{O}\left(\varepsilon^{2}\right).$ Taking the integral with respect to $\Omega$ of the equation (3.20a), one can get $\partial_{t}\rho_{g}+\frac{1}{\varepsilon}\nabla\cdot\left\langle\Omega I_{g}\right\rangle=\sigma_{a,g}(B_{g}-\rho_{g})\,.$ (3.22) Plugging (3.21) into the equation (3.22), letting $\varepsilon\to 0$, then (3.22) reduces to $\partial_{t}\rho_{g}-\nabla\cdot\left(\frac{1}{3\sigma_{s,g}}\nabla\rho_{g}\right)=\sigma_{a,g}(B_{g}-\rho_{g})\,.$ (3.23) Moreover, by letting $\varepsilon\to 0$ in (3.20b), one has $C_{v}\partial_{t}T=4\pi\sum\limits_{g^{\prime}=1}^{G}\left[\sigma_{a,g^{\prime}}\left(\rho_{g^{\prime}}-B_{g^{\prime}}\right)\right],$ and the above two equations give a group discretization of the frequency dependent diffusion limit (2.10). When one approximates $\sigma_{a,g}$ and $\sigma_{s,g}$ as in (3.7), the coefficients in the frequency dependent diffusion limit are determined by $\sigma^{c}_{a,g}=\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{a}\mathrm{d}\nu,\quad\frac{1}{\sigma_{s,g}^{c}}=\frac{1}{\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{s}\mathrm{d}\nu}$ When (3.8) and (3.10) are used, $\sigma^{r}_{a,g}$ and $1/\sigma_{s,g}^{r}$ become $\sigma^{r}_{a,g}=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\sigma_{a}}{\sigma_{s}}\frac{\partial B}{\partial T}(T)\mathrm{~{}d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{s}}\frac{\partial B}{\partial T}(T)\mathrm{~{}d}\nu},\quad\frac{1}{\sigma^{r}_{s,g}}=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{s}}\frac{\partial B}{\partial T}(T)\mathrm{~{}d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T)\mathrm{~{}d}\nu}.$ We emphasize that, different approximations provide different coefficient values for different groups in the limiting models, which is crucial to get the correct solution behavior. ### 3.3 Frequency discretization of the FDDL From the above discussion, we can see that different group discretizations for FRTE yield different approximations of the group diffusion and absorption coefficients in the FDDL. In the frequency dependent diffusion regime, different group diffusion and absorption coefficients can affect the propagation speed of heat waves under different physical conditions. Moreover, which approximation is better depends on different physical conditions [yee2017stable, park2019multigroup]. In this subsection, we would like to compare several different approximations of the coefficients $\sigma$ in (2.6) and $\sigma_{s}$, $\sigma_{a}$ in (2.10). We do not discuss about which coefficient approximation is better in the limiting model, but to show that they can be quite different from each other. Three different approximations of $\sigma_{a,g}$ are considered: the piece- wise constant approximation $\sigma^{c}_{a,g}$ in (3.7), the Rosseland mean approximation $\sigma_{a,g}^{r}$ in (3.8) and the Planck approximation. In the Planck approximation as in [sunj], the group absorptions to discretize (2.10) are defined as follows $\sigma^{e}_{a,g}=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{a}B(\nu,T)\mathrm{~{}d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}B(\nu,T)\mathrm{~{}d}\nu},\quad\sigma^{a}_{a,g}=\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{a}\rho\mathrm{~{}d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\rho\mathrm{~{}d}\nu}.$ (3.24) It should be pointed out that $\sigma^{a}_{a,g}$ is a weighted integration with the unknown function $\rho$. Usually, the unknown function $\rho$ is replaced by the Planck function with the radiation temperature $T_{r}$, and the multi-group FDDL writes: $\displaystyle\partial_{t}\rho_{g}+\nabla\cdot\left(\frac{1}{3\sigma_{s,g}}\nabla\rho_{g}\right)=\sigma^{e}_{a,g}B_{g}-\sigma^{a}_{a,g}\rho_{g},\qquad(g=1,\cdots,G),$ (3.25a) $\displaystyle C_{v}\partial_{t}T=4\pi\sum\limits_{g^{\prime}=1}^{G}\left(\sigma^{a}_{a,g^{\prime}}\rho_{g^{\prime}}-\sigma^{e}_{a,g^{\prime}}B_{g^{\prime}}\right).$ (3.25b) Similar as in [densmore2012hybrid, sunj], we consider a simple dependence of the frequency such that $\sigma_{a}=\nu^{-3}$ and $\sigma_{s}=\nu^{-3}$. Here we ignore the scale parameters $\mathscr{L}_{a}$ and $\mathscr{L}_{s}$ since they not affect the results of the comparison. The frequency domain is from $0.1\ eV$ to $100\ keV$ and divided into 30 frequency groups, and the groups are logarithmically spaced. The following quantities are compared: * • The two different approximations of $1/\sigma_{t}$ in (3.18) and (3.19) at different temperatures $T$. The results are displayed in Table 1 and Figure. 1, and the reference solution is calculated by dividing the frequency domain into 600 groups. * • The diffusion coefficients $1/\sigma_{s}$ in the FDDL (2.10) are approximated by the $1/\sigma^{c}_{s,g}$ and two different $1/\sigma^{r}_{s,g}$ at $T=1\ keV$ and $T=16\ keV$. In Figure. 1, From Figure. 1, the Rosseland mean approximation is closer to the reference. * • For the absorbing coefficient $\sigma_{a}$, we compare the piece-wise constant approximation $\sigma^{c}_{a,g}$, the Rosseland mean approximation $\sigma_{a,g}^{r}$ and the Plank approximation $\sigma^{e}_{a,g}$ in Figure. 2. We observe that they can be very different. Table 1: The different approximations to $1/\sigma$ in the gray radiation diffusion limit (2.7). The reference solution is calculated by dividing the frequency domain into 600 groups. T | Reference | $1/\sigma_{t}^{r}$ | relative error | $1/\sigma_{t}^{c}$ | relative error ---|---|---|---|---|--- $1\ keV$ | 25.3901 | 25.3920 | $7.48\times 10^{-5}$ | 61.7458 | 1.43 $2\ keV$ | 203.1204 | 203.1042 | $7.97\times 10^{-5}$ | 493.8879 | 1.43 $4\ keV$ | $1.6250\times 10^{3}$ | $1.6251\times 10^{3}$ | $6.15\times 10^{-5}$ | $3.9516\times 10^{3}$ | 1.43 $8\ keV$ | $1.2574\times 10^{4}$ | $1.2284\times 10^{4}$ | $2.31\times 10^{-2}$ | $2.8811\times 10^{4}$ | 1.29 $16\ keV$ | $5.1674\times 10^{4}$ | $5.4364\times 10^{4}$ | $5.21\times 10^{-2}$ | $9.2914\times 10^{4}$ | $7.98\times 10^{-1}$ Figure 1: Left: comparison of the different approximations of the mean free path $1/\sigma$ in the gray radiation diffusion limit (2.7) at the different temperatures; Right: comparison of the different approximations of the diffusion coefficient $1/\sigma_{s}$ in the FDDL (2.10) at $T=1\ keV$ and $T=16\ keV$. Figure 2: (a)comparison of the different approximations of $\sigma_{a}$ at $T=1\ keV$ and $T=16\ keV$; (b)zoom in part of figure (a), here the some red dots and green dots overlap; (c)zoom in, here the blue dots and green dots overlap. ## 4 The decomposed multi-group method and its asymptotic limits ### 4.1 The decomposed multi-group method A new decomposed multi-group method is proposed in this section. We decompose the radiation intensity $I(t,x,\Omega,\nu)$ into three parts: $I(t,x,\Omega,\nu)=\left\langle I\right\rangle+3\Omega\cdot\left\langle\Omega I\right\rangle+Q(t,x,\Omega,\nu),$ (4.1a) $\hskip 91.04872pt:=\rho(t,x,\nu)+\Omega\cdot R(t,x,\nu)+Q(t,x,\Omega,\nu),$ (4.1b) with $\left\langle\cdot\right\rangle=\frac{1}{4\pi}\int_{4\pi}\cdot\ \text{d}\Omega$ and 3$\left\langle\Omega I\right\rangle=R$. Then taking $\left\langle\Omega\cdot\right\rangle$ on both sides of (4.1a) and using the condition $\left\langle\Omega\Omega\right\rangle:=D_{d}=\frac{1}{3}I_{d}$ (where $I_{d}$ denotes the 3 by 3 identity matrix), one gets $\left\langle\Omega I\right\rangle=\left\langle\Omega I\right\rangle+\left\langle\Omega Q\right\rangle,$ which indicates that $\left\langle\Omega Q\right\rangle=0.$ (4.2) Moreover, taking $\left\langle\cdot\right\rangle$ on both sides of (4.1), one gets $\left\langle I\right\rangle=\left\langle\rho+\Omega\cdot R+Q\right\rangle=\left\langle I\right\rangle+\left\langle Q\right\rangle,$ which implies $\left\langle Q\right\rangle=0.$ (4.3) (4.2) and (4.3) are two properties that $Q$ has to satisfy. We reconstruct the intensity $I$ in each interval $(\nu_{g-1/2},\nu_{g+1/2})$ in the frequency domain as follows: $I(t,x,\Omega,\nu)=\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}{\color[rgb]{0,0,0}\Big{(}\omega_{1}\rho_{g}+\Omega\cdot R_{g}+Q_{g}+s_{\rho_{g}}+s_{\Omega\cdot R_{g}}\Big{)}}$ (4.4) where $\rho_{g}=\left\langle I_{g}\right\rangle,\qquad R_{g}=3\left\langle\Omega I_{g}\right\rangle,\qquad Q_{g}=I_{g}-\rho_{g}-\Omega\cdot R_{g},$ and $s_{\rho_{g}}={\color[rgb]{0,0,0}(\nu_{g+1/2}-\nu_{g-1/2})B(\nu,T)-\omega_{2}B_{g}},\qquad s_{\Omega\cdot R_{g}}=\frac{\frac{\nu_{g+1/2}-\nu_{g-1/2}}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu}\Omega\cdot R_{g}-\Omega\cdot R_{g}.$ (4.5) Here $\omega_{1}$ and $\omega_{2}$ are two given weight functions that depend on $\nu$. We will see later on that by tuning $\omega_{1}$, $\omega_{2}$, one can get the required approximations of the group absorption coefficients in FDDL. The first three terms on the right hand side of (4.4) $\rho_{g}$, $\Omega\cdot R_{g}$, $Q_{g}$ depend only on $g$ not in $\nu$. Moreover, the zeroth and first moments of $Q_{g}$ equal to 0, which is the same as $Q$. The two additional terms $s_{\rho_{g}}$, $s_{\Omega\cdot R_{g}}$ depend on $\nu$ but satisfy $\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}s_{\rho_{g}}\text{d}\nu=\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}s_{\Omega\cdot R_{g}}\text{d}\nu=0$. Thus the reconstructed $I$ in (4.4) satisfies $\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}I\text{d}\nu=I_{g}$. The two additional terms $s_{\rho_{g}}$ and $s_{\Omega\cdot R_{g}}$ are important to get the correct limits in the gray radiation and frequency dependent diffusion regime. The detailed derivations will be displayed in section 3.3. It is important to note that, as far as $I_{g},T$ are given, $\rho_{g}$, $R_{g}$, $Q_{g}$ and $s_{\rho_{g}}$, $s_{\Omega\cdot R_{g}}$ are known. Then $I(t,x,\Omega,\nu)$ in (4.4) can be determined by $I_{g}$ and the two terms in (3.4) are approximated by $\displaystyle\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{a}\left(B-I\right)\text{d}\nu=\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{a}\Big{[}B-\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}{\color[rgb]{0,0,0}\big{(}\omega_{1}\rho_{g}+\Omega\cdot R_{g}+Q_{g}+s_{\rho_{g}}+s_{\Omega\cdot R_{g}}\big{)}}\Big{]}\text{d}\nu$ $\displaystyle=\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{a}\Big{[}B-\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}{\color[rgb]{0,0,0}\Big{(}\omega_{1}\rho_{g}+(\nu_{g+1/2}-\nu_{g-1/2})B-\omega_{2}B_{g}+\frac{\frac{\nu_{g+1/2}-\nu_{g-1/2}}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu}\Omega\cdot R_{g}+Q_{g}\Big{)}}\Big{]}\text{d}\nu$ $\displaystyle{\color[rgb]{0,0,0}=\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{a}(\omega_{2}B_{g}-\omega_{1}\rho_{g})\text{d}\nu-\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{a}\text{d}\nu Q_{g}}$ $\displaystyle\hskip 199.16928pt-\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\sigma_{a}}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\frac{\Omega\cdot R_{g}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu},$ and $\displaystyle\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{s}\left(\rho-I\right)\text{d}\nu$ $\displaystyle=\frac{-1}{\nu_{g+1/2}-\nu_{g-1/2}}\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{s}\big{(}\Omega\cdot R_{g}+Q_{g}+s_{\Omega\cdot R_{g}}\big{)}\text{d}\nu$ $\displaystyle=\frac{-1}{\nu_{g+1/2}-\nu_{g-1/2}}\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\sigma_{s}\text{d}\nu Q_{g}-\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\sigma_{s}}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\frac{\Omega\cdot R_{g}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu}.$ Then we can get the decomposed multi-group radiative transfer equation: $\displaystyle\frac{1}{\mathcal{C}}\partial_{t}I_{g}+\Omega\cdot\nabla I_{g}=\mathscr{L}_{a}\sigma^{2}_{a,g}B_{g}-\mathscr{L}_{a}\sigma^{1}_{a,g}\rho_{g}-\big{(}\mathscr{L}_{a}+\mathscr{L}_{s}\big{)}\sigma_{t,g}\Omega\cdot R_{g}-(\mathscr{L}_{a}\sigma_{a,g}+\mathscr{L}_{s}\sigma_{s,g})Q_{g},$ $\displaystyle\hskip 341.43306pt(g=1,\cdots,G),$ (4.6a) $\displaystyle C_{v}\partial_{t}T=4\pi\mathcal{C}\mathcal{P}_{0}\sum\limits_{g^{\prime}=1}^{G}\left[\mathscr{L}_{a}\left(\sigma^{1}_{a,g^{\prime}}\rho_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B_{g^{\prime}}\right)\right],$ (4.6b) where $\displaystyle\sigma^{1}_{a,g}=\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\omega_{1}\sigma_{a}\text{d}\nu,\quad\sigma^{2}_{a,g}=\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\omega_{2}\sigma_{a}\text{d}\nu,$ (4.7) $\displaystyle\sigma_{a,g}=\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{a}\text{d}\nu,\quad\sigma_{s,g}=\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{s}\text{d}\nu,$ $\displaystyle\hskip 85.35826pt\sigma_{t,g}=\frac{1/(\mathscr{L}_{a}+\mathscr{L}_{s})}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\mathscr{L}_{a}\sigma_{a}+\mathscr{L}_{s}\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\big{/}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}\text{d}\nu}.$ It is important to use the reconstruction as in (4.4) to get the required absorption coefficients for group discretizations of the FDDL. One can consider the following possible $\omega_{1}$ and $\omega_{2}$: * • When $\omega_{1}=\omega_{2}=1$, then the absorbing coefficients are $\sigma^{1}_{a,g}=\sigma^{2}_{a,g}=\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{a}\text{d}\nu,$ which will yield piece-constant approximations of the group absorption coefficients in FDDL. * • When $\omega_{1}=\frac{\frac{\partial B}{\partial T}(T_{r})}{\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\frac{\partial B_{g}}{\partial T}(T_{r})},\quad\omega_{2}=\frac{\frac{\partial B}{\partial T}(T)}{\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}\frac{\partial B_{g}}{\partial T}(T)}$, then the absorbing coefficients are $\sigma^{1}_{a,g}=\frac{\int_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{a}\frac{\partial B}{\partial T}(T_{r})\text{d}\nu}{\frac{\partial B_{g}}{\partial T}(T_{r})},\quad\sigma^{2}_{a,g}=\frac{\int_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{a}\frac{\partial B}{\partial T}(T)\text{d}\nu}{\frac{\partial B_{g}}{\partial T}(T)},$ which gives the Rosseland mean approximations of the group absorption coefficients in FDDL. * • When $\omega_{1}=\frac{B(T_{r})}{\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}B_{g}(T_{r})},\quad\omega_{2}=\frac{B(T)}{\frac{1}{\nu_{g+1/2}-\nu_{g-1/2}}B_{g}(T)}$, then the absorbing coefficients are $\sigma^{1}_{a,g}=\frac{\int_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{a}B(T_{r})\text{d}\nu}{B_{g}(T_{r})},\quad\sigma^{2}_{a,g}=\frac{\int_{\nu_{g-1/2}}^{\nu_{g+1/2}}\sigma_{a}B(T)\text{d}\nu}{B_{g}(T)},$ which corresponds to the Planck approximation of the absorption coefficient in FDDL. When the different $\omega_{1}$ and $\omega_{2}$ are chosen, the different approximations of the absorbing coefficients in FDDL are determined. Comparing (4.6) with (3.6), we can see that the scattering and absorbing coefficients for $\rho_{g}$, $R_{g}$ and $Q_{g}$ are different in the decomposed multi- group method, while $\rho_{g}$, $R_{g}$ and $Q_{g}$ share the same values of scattering and absorbing coefficients in the classical multi-group method. ### 4.2 The decomposed multi-group method in the diffusion regime #### 4.2.1 The gray radiation diffusion regime For the scaling $\mathscr{L}_{a}=1/\varepsilon,\mathscr{L}_{s}=1/\varepsilon$, the decomposed multi-group radiative transfer system (4.6) can be rewritten into $\displaystyle\varepsilon\partial_{t}I_{g}+\Omega\cdot\nabla I_{g}=\frac{1}{\varepsilon}\left(\sigma_{a,g}^{2}B_{g}-\sigma_{a,g}^{1}\rho_{g}\right)-\frac{\Omega\cdot R_{g}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\varepsilon}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\big{/}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}\text{d}\nu}-\frac{\sigma_{a,g}+\sigma_{s,g}}{\varepsilon}Q_{g},$ $\displaystyle\hskip 298.75394pt(g=1,\cdots,G),$ (4.8a) $\displaystyle\varepsilon C_{v}\partial_{t}T=4\pi\sum\limits_{g^{\prime}=1}^{G}\left[\frac{1}{\varepsilon}\left(\sigma^{1}_{a,g^{\prime}}\rho_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B_{g^{\prime}}\right)\right].$ (4.8b) At the leading order of (4.8a), one has $\displaystyle\sigma^{2}_{a,g}B_{g}-\sigma^{1}_{a,g}\rho_{g}-\frac{\Omega\cdot R_{g}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\big{/}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}\text{d}\nu}-(\sigma_{a,g}+\sigma_{s,g})Q_{g}=\mathcal{O}(\varepsilon),$ (4.9) Integrating with respect to $\Omega$ in equation (4.9) and using $\left\langle Q_{g}\right\rangle=0$, we can obtain $\sigma^{1}_{a,g}\rho_{g}=\sigma^{2}_{a,g}B_{g}+\mathcal{O}(\varepsilon)\,.$ (4.10) Summing the above equation over the group index $g$, and by the definitions (3.5), one can get $\int\limits_{0}^{\infty}\sigma_{a}\rho\text{d}\nu=\int\limits_{0}^{\infty}\sigma_{a}B\text{d}\nu+\mathcal{O}(\varepsilon),$ (4.11) which implies that $\rho_{g}=B_{g}+\mathcal{O}(\varepsilon)\,.$ (4.12) Integrating (4.8a) with respect to $\Omega$ and adding up all groups and equation (4.8b), one gets $4\pi\partial_{t}\left(\sum_{g=1}^{G}\rho_{g}\right)+\frac{4\pi}{3\varepsilon}\sum_{g=1}^{G}\nabla\cdot R_{g}+C_{v}\partial_{t}T=0\,.$ (4.13) From (4.12) and the conditions that $B_{g}$ satisfies in (3.16), $4\pi\sum_{g=1}^{G}\rho_{g}=T^{4}+\mathcal{O}(\varepsilon)$. In order to get the diffusion limit equation, one has to determine the relationship between $\rho_{g}$ and $R_{g}$. Taking the first moment of (4.8a), and using $I_{g}=\rho_{g}+\Omega\cdot R_{g}+Q_{g}$, (4.12) and $\left\langle\Omega Q_{g}\right\rangle$=0, one has $-\frac{1}{\varepsilon}R_{g}=\frac{1}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}\text{d}\nu}\left[\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\nabla\rho_{g}}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu+\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\nabla\left\langle\Omega\Omega Q_{g}\right\rangle}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\right]+\mathcal{O}(\varepsilon).$ (4.14) Taking the second moment of $\Omega$ in equation (4.9) and noting (4.12) yields $\left\langle\Omega\Omega Q_{g}\right\rangle:=K_{Q_{g}}=\mathcal{O}(\varepsilon),$ which implies $-\frac{1}{\varepsilon}R_{g}=\frac{1}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}\text{d}\nu}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\nabla\rho_{g}}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu+\mathcal{O}(\varepsilon).$ (4.15) Plugging (4.12), (4.15) into the equation (4.13), when $\varepsilon\to 0$, (4.13) reduces to $\partial_{t}\left(\sum_{g=1}^{G}4\pi B_{g}\right)+C_{v}\partial_{t}T=\sum_{g=1}^{G}\nabla\cdot\left[\left(\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\frac{4\pi}{3(\sigma_{a}+\sigma_{s})}\frac{\partial B}{\partial T}\text{d}\nu\right)\nabla T\right]\,.$ (4.16) Due to (3.16), (4.16) becomes $\partial_{t}T^{4}+C_{v}\partial_{t}T=\nabla\cdot\left(\frac{1}{3\widehat{\sigma}_{R}}\nabla T^{4}\right)\,,$ (4.17) where $\frac{1}{\widehat{\sigma}_{R}}=\frac{1}{4T^{3}}\left(\sum\limits_{g=1}^{G}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\frac{4\pi}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\right),$ which is consistent with equation (2.6). #### 4.2.2 The frequency dependent diffusion regime In the frequency dependent diffusion regime, the scaling is $\mathscr{L}_{a}=\varepsilon$ and $\mathscr{L}_{s}=1/\varepsilon$, then the multi-group radiative transfer system (4.6) can be rewritten to $\displaystyle\varepsilon\partial_{t}I_{g}+\Omega\cdot\nabla I_{g}=\varepsilon\left(\sigma^{2}_{a,g}B_{g}-\sigma^{1}_{a,g}\rho_{g}\right)-\frac{\Omega\cdot R_{g}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\varepsilon\sigma_{a}+\frac{\sigma_{s}}{\varepsilon}}\frac{\partial B}{\partial T}\text{d}\nu\big{/}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}\text{d}\nu}-\left(\varepsilon\sigma_{a,g}+\frac{\sigma_{s,g}}{\varepsilon}\right)Q_{g},$ $\displaystyle\hskip 341.43306pt(g=1,\cdots,G),$ (4.18a) $\displaystyle\varepsilon C_{v}\partial_{t}T=4\pi\sum\limits_{g^{\prime}=1}^{G}\left[\varepsilon\left(\sigma^{1}_{a,g^{\prime}}\rho_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B_{g^{\prime}}\right)\right].$ (4.18b) The asymptotic analysis for the FDDL begins to find the relation between $\rho_{g}$ and $R_{g}$. From (4.18a), one has $\displaystyle-\frac{\Omega\cdot R_{g}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\big{/}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}\text{d}\nu}-\sigma_{s,g}Q_{g}=\mathcal{O}(\varepsilon).$ (4.19) Taking the second moment in $\Omega$ of equation (4.19) yields $K_{Q_{g}}=\mathcal{O}(\varepsilon).$ (4.20) Moreover, using the relation (4.20), the leading order of the first moment of equation (4.18a) is $-\frac{1}{\varepsilon}R_{g}=\frac{1}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}\text{d}\nu}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\nabla\rho_{g}}{\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu+\mathcal{O}(\varepsilon),$ (4.21) which gives the relation between $R_{g}$ and $\rho_{g}$. Integrating (4.18a) with respect to $\Omega$, one can get $\partial_{t}\rho_{g}+\frac{1}{3\varepsilon}\nabla\cdot R_{g}=\sigma^{2}_{a,g}B_{g}-\sigma^{1}_{a,g}\rho_{g}\,.$ (4.22) Plugging (4.21) into (4.22), and letting $\varepsilon\to 0$, (4.22) gives $\partial_{t}\rho_{g}-\nabla\cdot\left(\frac{1}{3\sigma^{\prime}_{s,g}}\nabla\rho_{g}\right)=\sigma^{2}_{a,g}B_{g}-\sigma^{1}_{a,g}\rho_{g}\,,$ (4.23) where $\frac{1}{\sigma^{\prime}_{s,g}}=\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{s}}\frac{\partial B}{\partial T}\text{d}\nu\big{/}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}\text{d}\nu.$ From equation (4.18b), one has $C_{v}\partial_{t}T=\sum\limits_{g^{\prime}=1}^{G}\left[4\pi(\sigma^{1}_{a,g^{\prime}}\rho_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B_{g^{\prime}})\right].$ In summary, the diffusion limit of the decomposed multi-group radiative transfer system (4.18) is $\displaystyle\partial_{t}\rho_{g}-\nabla\cdot\left(\frac{1}{3\sigma^{\prime}_{s,g}}\nabla\rho_{g}\right)=\sigma^{2}_{a,g}B_{g}-\sigma^{1}_{a,g}\rho_{g}\,,$ (4.24a) $\displaystyle C_{v}\partial_{t}T=\sum\limits_{g^{\prime}=1}^{G}\left[4\pi(\sigma^{1}_{a,g^{\prime}}\rho_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B_{g^{\prime}})\right].$ (4.24b) The decomposed multi-group radiative transfer equations (3.2) is AP in both gray radiation diffusion limit and the FDDL. Moreover, one can choose different approximations of $\sigma_{a,g}^{1}$, $\sigma_{a,g}^{2}$ in the limiting frequency discretization. ## 5 AP time discretization of the decomposed multi-group method ### 5.1 The time discretization Based on the decomposition in (4.1) for radiation intensity $I$ and the decomposed multi-group method in (4.6), we will present a time discretization for the FRTE and show its AP property. Let $t^{n}=n\Delta t,\qquad n=0,1,2,\cdots,$ and $I^{n}\approx I(t^{n},x,\Omega,\nu),\qquad\rho^{n}\approx\rho(t^{n},x,\nu),\qquad T^{n}\approx T(t^{n},x),\qquad B^{n}\approx B(\nu,T(t^{n},x)),$ with $I_{g}^{n}\approx\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}I(t^{n},x,\Omega,\nu)\text{d}\nu,\qquad\rho_{g}^{n}\approx\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\rho(t^{n},x,\nu)\text{d}\nu,\qquad B_{g}^{n}\approx\int\limits_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}B(\nu,T(t^{n},x))\text{d}\nu.$ We use the following semi-discrete scheme in time for (3.2) in each group: $\displaystyle\frac{1}{\mathcal{C}}\frac{I_{g}^{n+1}-I_{g}^{n}}{\Delta t}+\Omega\cdot\nabla(\rho^{n+1}_{g}+\Omega\cdot R^{n+1}_{g}+Q^{n}_{g})=$ $\displaystyle\hskip 14.22636pt\mathscr{L}_{a}\left(\sigma^{2}_{a,g}B^{n+1}_{g}-\sigma^{1}_{a,g}\rho^{n+1}_{g}\right)-(\mathscr{L}_{a}+\mathscr{L}_{s})\sigma_{t,g}\Omega\cdot R^{n+1}_{g}-(\mathscr{L}_{a}\sigma_{a,g}+\mathscr{L}_{s}\sigma_{s,g})Q^{n}_{g},$ (5.1a) $\displaystyle\frac{C_{v}}{\mathcal{C}\mathcal{P}_{0}}\frac{T^{n+1}-T^{n}}{\Delta t}=\sum\limits_{g^{\prime}=1}^{G}\left[4\pi\mathscr{L}_{a}\left(\sigma^{1}_{a,g^{\prime}}\rho^{n+1}_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B^{n+1}_{g^{\prime}}\right)\right],$ (5.1b) Instead of solving system (5.1) directly, we take the zeroth and first moments of (5.1a), and then couple them together with (5.1b). More precisely, $\displaystyle\frac{1}{\mathcal{C}}\frac{\rho_{g}^{n+1}-\rho_{g}^{n}}{\Delta t}+\left\langle\Omega\cdot\nabla(\Omega\cdot R^{n+1}_{g})\right\rangle=\mathscr{L}_{a}\left(\sigma^{2}_{a,g}B^{n+1}_{g}-\sigma^{1}_{a,g}\rho^{n+1}_{g}\right),$ (5.2a) $\displaystyle\frac{1}{3\mathcal{C}}\frac{R_{g}^{n+1}-R_{g}^{n}}{\Delta t}+\left\langle\Omega\Omega\cdot\nabla(\rho^{n+1}_{g}+Q^{n}_{g})\right\rangle=-\frac{(\mathscr{L}_{a}+\mathscr{L}_{s})\sigma_{t,g}R^{n+1}_{g}}{3},$ (5.2b) $\displaystyle\frac{C_{v}}{\mathcal{C}\mathcal{P}_{0}}\frac{T^{n+1}-T^{n}}{\Delta t}=\sum\limits_{g^{\prime}=1}^{G}\left[4\pi\mathscr{L}_{a}\left(\sigma^{1}_{a,g^{\prime}}\rho^{n+1}_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B^{n+1}_{g^{\prime}}\right)\right].$ (5.2c) The macroscopic quantities $\rho_{g},\ R_{g}$ and $T$ can be updated by solving equation (5.2), then $Q_{g}$ can be updated implicitly by equation (5.1a), i.e., $\displaystyle\frac{1}{\mathcal{C}}\frac{\rho_{g}^{n+1}-\rho_{g}^{n}}{\Delta t}+\frac{\Omega}{\mathcal{C}}\cdot\frac{R_{g}^{n+1}-R_{g}^{n}}{\Delta t}+\frac{1}{\mathcal{C}}\frac{Q_{g}^{n+1}-Q_{g}^{n}}{\Delta t}+\Omega\cdot\nabla(\rho^{n+1}_{g}+\Omega\cdot R^{n+1}_{g}+Q^{n+1}_{g})=$ (5.3) $\displaystyle\hskip 28.45274pt\mathscr{L}_{a}\left(\sigma^{2}_{a,g}B^{n+1}_{g}-\sigma^{1}_{a,g}\rho^{n+1}_{g}\right)-(\mathscr{L}_{a}+\mathscr{L}_{s})\sigma_{t,g}\Omega\cdot R^{n+1}_{g}-(\mathscr{L}_{a}\sigma_{a,g}+\mathscr{L}_{s}\sigma_{s,g})Q^{n+1}_{g},$ in which only a linear transport equation is solved with all direction and energy group being decoupled. Next, we will discuss the diffusion limit of system (5.2). ### 5.2 The gray radiation diffusion regime of (5.2) The asymptotic analysis of the time discretization is similar to the analysis in Section 3.3. In the gray radiation diffusion regime, $\mathscr{L}_{a}=1/\varepsilon$ and $\mathscr{L}_{s}=1/\varepsilon$. At the leading order of (5.2a), one gets $\sigma^{1}_{a,g}\rho^{n+1}_{g}=\sigma^{2}_{a,g}B^{n+1}_{g}+\mathcal{O}(\varepsilon)\,,$ (5.4) Summing the above equation over the group index $g$, and by the definitions (3.5), one can get $\int\limits_{0}^{\infty}\sigma_{a}\rho^{n+1}\text{d}\nu=\int\limits_{0}^{\infty}\sigma_{a}B^{n+1}\text{d}\nu+\mathcal{O}(\varepsilon),$ (5.5) which implies that $\rho^{n+1}_{g}=B^{n+1}_{g}+\mathcal{O}(\varepsilon)\,.$ (5.6) In (5.2a), adding up all the equations in each group and (5.2c), we have $4\pi\frac{\sum_{g=1}^{G}\rho^{n+1}_{g}-\sum_{g=1}^{G}\rho^{n}_{g}}{\Delta t}+C_{v}\frac{T^{n+1}-T^{n}}{\Delta t}=\frac{4\pi}{3\varepsilon}\sum_{g=1}^{G}\left(\nabla\cdot R_{g}^{n+1}\right),$ (5.7) From (5.6) and the conditions that $B_{g}$ satisfies in (3.16), $4\pi\sum_{g=1}^{G}\rho^{n+1}_{g}=(T^{n+1})^{4}+\mathcal{O}(\varepsilon)$. To get the diffusion limit of equation (5.7), the relation between $\rho_{g}^{n+1}$ and $R^{n+1}_{g}$ has to be analyzed. By using Champan-Enskog expansion in equation (5.3), one can get $\displaystyle-\frac{\Omega\cdot R^{n+1}_{g}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu\big{/}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu}-(\sigma_{a,g}+\sigma_{s,g})Q^{n+1}_{g}=\mathcal{O}(\varepsilon).$ (5.8) Taking the second moment in $\Omega$ of equation (5.8) yields $\left\langle\Omega\Omega Q^{n+1}_{g}\right\rangle:=K^{n+1}_{Q_{g}}=\mathcal{O}(\varepsilon).$ (5.9) Taking the first moment of (5.1a), and using the conditions $\rho^{n+1}_{g}\approx B^{n+1}_{g}$, $K^{n}_{Q_{g}}=\mathcal{O}(\varepsilon)$ and $\left\langle\Omega Q^{n}_{g}\right\rangle$=0, one has $-\frac{1}{\varepsilon}R^{n+1}_{g}=\frac{1}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\nabla\rho^{n+1}_{g}}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu+\mathcal{O}(\varepsilon),$ (5.10) which gives the relation between $R_{g}$ and $\rho_{g}$, then putting it into (5.7), and using (5.6), when $\varepsilon\to 0$, (5.7) reduces to $4\pi\frac{\sum\limits_{g=1}^{G}B^{n+1}_{g}-\sum\limits_{g=1}^{G}B^{n}_{g}}{\Delta t}+C_{v}\frac{T^{n+1}-T^{n}}{\Delta t}=\sum_{g=1}^{G}\nabla\cdot\left[\left(\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\frac{4\pi}{3(\sigma_{a}+\sigma_{s})}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu\right)\nabla T^{n+1}\right]\,.$ (5.11) Simultaneously, we can obtain $\frac{(T^{n+1})^{4}-(T^{n})^{4}}{\Delta t}+C_{v}\frac{T^{n+1}-T^{n}}{\Delta t}=\nabla\cdot\left(\frac{1}{3\widehat{\sigma}_{R}}\nabla(T^{n+1})^{4}\right)\,,$ (5.12) with $\frac{1}{\widehat{\sigma}_{R}}=\frac{1}{4(T^{n+1})^{3}}\left(\sum\limits_{g=1}^{G}\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}\frac{4\pi}{\sigma_{a}+\sigma_{s}}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu\right),$ which is a semi-discretization for equation (4.17). ### 5.3 The frequency dependent diffusion regime of (5.2) In the frequency dependent diffusion regime, $\mathscr{L}_{a}=\varepsilon$ and $\mathscr{L}_{s}=1/\varepsilon$. By using Champan-Enskog expansion in equation (5.3), one can get $\displaystyle-\frac{\Omega\cdot R^{n+1}_{g}}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{s}}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu\big{/}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu}-\sigma_{s,g}Q^{n+1}_{g}=\mathcal{O}(\varepsilon).$ (5.13) Taking the second moment for $\Omega$ in equation (5.13) yields $\left\langle\Omega\Omega Q^{n+1}_{g}\right\rangle=\mathcal{O}(\varepsilon).$ (5.14) From the second equation of system (5.2), the leading order is $-\frac{1}{\varepsilon}R^{n+1}_{g}=\frac{1}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\nabla\rho^{n+1}_{g}}{\sigma_{s}}\frac{\partial B}{\partial T}(T^{n+1})\text{d}\nu+\mathcal{O}(\varepsilon),$ (5.15) then putting it into the zeroth moment of (5.1a), and sending $\varepsilon\to 0$, one gets $\frac{\rho_{g}^{n+1}-\rho_{g}^{n}}{\Delta t}-\nabla\cdot\left(\frac{1}{3\sigma^{\prime}_{s,g}}\nabla\rho^{n+1}_{g}\right)=\sigma^{2}_{a,g}B_{g}^{n+1}-\sigma^{1}_{a,g}\rho_{g}^{n+1},$ (5.16) where $\frac{1}{\sigma^{\prime}_{s,g}}=\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{s}}\frac{\partial B}{\partial T}(T_{r}^{n+1})\text{d}\nu\big{/}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T_{r}^{n+1})\text{d}\nu.$ Moreover, sending $\varepsilon\to 0$ in equation (5.1b) gives $C_{v}\frac{T^{n+1}-T^{n}}{\Delta t}=\sum\limits_{g^{\prime}=1}^{G}\left[4\pi\left(\sigma^{1}_{a,g^{\prime}}\rho^{n+1}_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B^{n+1}_{g^{\prime}}\right)\right].$ (5.17) In summary, the diffusion limit of system (5.2) in the frequency dependent diffusion regime is $\left\\{\begin{aligned} &\frac{\rho_{g}^{n+1}-\rho_{g}^{n}}{\Delta t}-\nabla\cdot\left(\frac{1}{3\sigma^{\prime}_{s,g}}\nabla\rho^{n+1}_{g}\right)=\sigma^{2}_{a,g}B_{g}^{n+1}-\sigma^{1}_{a,g}\rho_{g}^{n+1}\,,\\\ &C_{v}\frac{T^{n+1}-T^{n}}{\Delta t}=\sum\limits_{g^{\prime}=1}^{G}\left[4\pi\left(\sigma^{1}_{a,g^{\prime}}\rho^{n+1}_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B^{n+1}_{g^{\prime}}\right)\right],\end{aligned}\right.$ which is a semi-discretization for system (4.24). ## 6 AP full discretization for FRTE For the ease of exposition, we will explain our spatial discretion in 1D. That is, $x\in[0,L]$, $\Omega\in[-1,1]$, and $\left\langle f(\Omega)\right\rangle=\frac{1}{2}\int_{-1}^{1}f(\Omega)\mathrm{d}\Omega$, and the boundary conditions for the FRTE are $\displaystyle I_{g}(t,0,\Omega)=b_{g,\text{L}}(t,\Omega)=\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}b_{\text{L}}(t,\Omega,\nu)\text{d}\nu,~{}\text{ for }~{}\Omega>0;$ (6.1) $\displaystyle I_{g}(t,L,\Omega)=b_{g,\text{R}}(t,\Omega)=\int\limits_{\nu_{g-1/2}}^{\nu_{g+1/2}}b_{\text{R}}(t,\Omega,\nu)\text{d}\nu,~{}\text{ for }~{}\Omega<0\,.$ Higher dimensions can be treated in the dimension by dimension manner. Let $\Delta x=L/N_{x}$, and we consider the uniform mesh as follows $x_{i}=(i-1)\Delta x,\quad i=1,2\cdots,N_{x}+1,$ and let $x_{i+\frac{1}{2}}=\left(x_{i}+x_{i+1}\right)/2,\qquad\mbox{for $i=1,\cdots,N_{x}.$}$ ### 6.1 Even-odd space discretization To get a consistent stencil in spatial discretization, we use the even-odd parity method [k2016asymptotic, tang2021accurate, jin2000uniformly]. Let $E_{g}(\Omega)=\frac{1}{2}\big{(}I_{g}(\Omega)+I_{g}(-\Omega)\big{)},\qquad O_{g}(\Omega)=\frac{1}{2}(I_{g}(\Omega)-I_{g}(-\Omega)),\qquad\Omega>0$ (6.2) be respectively the even and odd part of $I_{g}$, and we consider the even part on half spatial grid and odd part on regular grid, i.e., $\displaystyle E_{g,i+1/2}(\Omega)\approx E_{g}(x_{i+1/2},\Omega),\quad i=1,\cdots,N_{x}\,;$ (6.3) $\displaystyle O_{g,i}(\Omega)\approx O_{g}(x_{i},\Omega),\quad i=1,\cdots,N_{x}+1\,.$ (6.4) Moreover, plugging the decomposition (4.1) into (6.2), we have $E_{g}=\rho_{g}+\big{(}Q_{g}(\Omega)+Q_{g}(-\Omega)\big{)}:=\rho_{g}+E_{Q_{g}},\quad O_{g}=\Omega\cdot R_{g}+\frac{1}{2}\big{(}Q_{g}(\Omega)-Q_{g}(-\Omega)\big{)}:=\Omega\cdot R_{g}+O_{Q_{g}}\,.$ (6.5) Then the full discretization for the macroscopic equation (5.2) and the microscopic equation (5.3) are described below. Space discretization for the first term (5.2). System (5.1) can be written into a system for $\rho_{g},\ R_{g}$ and $E_{Q_{g}},\ O_{Q_{g}}$: $\left\\{\begin{aligned} &\frac{1}{\mathcal{C}}\partial_{t}(\Omega\cdot R_{g}+O_{Q_{g}})+\Omega\cdot\nabla_{x}(\rho_{g}+E_{Q_{g}})=-(\mathscr{L}_{a}+\mathscr{L}_{s})\sigma_{t,g}\Omega\cdot R_{g}-(\mathscr{L}_{a}\sigma_{a,g}+\mathscr{L}_{s}\sigma_{s,g})O_{Q_{g}}\,;\\\ &\frac{1}{\mathcal{C}}\partial_{t}(\rho_{g}+E_{Q_{g}})+\Omega\cdot\nabla_{x}(\Omega\cdot R_{g}+O_{Q_{g}})=\mathscr{L}_{a}\left(\sigma^{2}_{a,g}B_{g}-\sigma^{1}_{a,g}\rho_{g}\right)-(\mathscr{L}_{a}\sigma_{a,g}+\mathscr{L}_{s}\sigma_{s,g})E_{Q_{g}}\,;\\\ &\frac{C_{v}}{\mathcal{C}\mathcal{P}_{0}}\partial_{t}T=\sum\limits_{g^{\prime}=1}^{G}\left[4\pi\mathscr{L}_{a}\left(\sigma^{1}_{a,g^{\prime}}\rho_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B_{g^{\prime}}\right)\right]\,.\end{aligned}\right.$ (6.6) In terms of $\rho_{g},\ R_{g}$ and $E_{Q_{g}},\ O_{Q_{g}}$, the system (5.2) becomes $\left\\{\begin{aligned} &\frac{1}{3\mathcal{C}}\partial_{t}R_{g}+\left\langle\Omega\Omega\cdot\nabla(\rho_{g}+E_{Q_{g}})\right\rangle=-\frac{(\mathscr{L}_{a}+\mathscr{L}_{s})\sigma_{t,g}R_{g}}{3},\\\ &\frac{1}{\mathcal{C}}\partial_{t}\rho_{g}+\left\langle\Omega\cdot\nabla(\Omega\cdot R_{g})\right\rangle=\mathscr{L}_{a}\left(\sigma^{2}_{a,g}B_{g}-\sigma^{1}_{a,g}\rho_{g}\right),\\\ &\frac{C_{v}}{\mathcal{C}\mathcal{P}_{0}}\partial_{t}T=\sum\limits_{g^{\prime}=1}^{G}\left[4\pi\mathscr{L}_{a}\left(\sigma^{1}_{a,g^{\prime}}\rho_{g^{\prime}}-\sigma^{2}_{a,g^{\prime}}B_{g^{\prime}}\right)\right].\end{aligned}\right.$ (6.7) The spatial discretization then takes the following form: $\displaystyle\frac{1}{3\mathcal{C}}\frac{R_{g,i}^{n+1}-R_{g,i}^{n}}{\Delta t}+\frac{1}{3\Delta x}(\rho^{n+1}_{g,i+1/2}-\rho^{n+1}_{g,i-1/2})+\frac{1}{\Delta x}\left(\int_{0}^{1}\Omega\Omega E^{n}_{Q_{g},i+1/2}\text{d}\Omega-\int_{0}^{1}\Omega\Omega E^{n}_{Q_{g},i-1/2}\text{d}\Omega\right)$ $\displaystyle\hskip 213.39566pt=-\frac{(\mathscr{L}_{a}+\mathscr{L}_{s})\sigma_{t,g,i}R^{n+1}_{g,i}}{3},\quad 2\leq i\leq N_{x}\,;$ (6.8a) $\displaystyle\frac{1}{\mathcal{C}}\frac{\rho_{g,i+1/2}^{n+1}-\rho_{g,i+1/2}^{n}}{\Delta t}+\frac{R^{n+1}_{g,i+1}-R^{n+1}_{g,i}}{3\Delta x}=\mathscr{L}_{a}\left(\sigma^{2}_{a,g,i+1/2}B^{n+1}_{g,i+1/2}-\sigma^{1}_{a,g,i+1/2}\rho^{n+1}_{g,i+1/2}\right),$ $\displaystyle\hskip 341.43306pt\quad 1\leq i\leq N_{x}\,;$ (6.8b) $\displaystyle\frac{C_{v}}{\mathcal{C}\mathcal{P}_{0}}\frac{T_{i+1/2}^{n+1}-T_{i+1/2}^{n}}{\Delta t}=\sum\limits_{g^{\prime}=1}^{G}\left[4\pi\mathscr{L}_{a}\left(\sigma^{1}_{a,g^{\prime},i+1/2}\rho^{n+1}_{g^{\prime},i+1/2}-\sigma^{2}_{a,g^{\prime},i+1/2}B^{n+1}_{g^{\prime},i+1/2}\right)\right],1\leq i\leq N_{x}\,.$ (6.8c) Space discretization for the second term (5.3). The variables $\rho_{g}^{n+1}$, $R_{g}^{n+1}$ and $T^{n+1}$ will be updated by solving system (6.8), and this is a nonlinear system that has to be solved by Newton iteration. Then the full discretization of (5.3) to update $O_{Q_{g}}$, $E_{Q_{g}}$ is: $\displaystyle\frac{1}{\mathcal{C}}\frac{O^{n+1}_{Q_{g},i}-O^{n}_{Q_{g},i}}{\Delta t}+\frac{\Omega}{\mathcal{C}}\frac{R^{n+1}_{g,i}-R^{n}_{g,i}}{\Delta t}+\Omega\frac{\rho^{n+1}_{g,i+1/2}-\rho^{n+1}_{g,i-1/2}}{\Delta x}+\Omega\frac{E^{n+1}_{Q_{g},i+1/2}-E^{n+1}_{Q_{g},i-1/2}}{\Delta x}$ $\displaystyle\hskip 42.67912pt=-(\mathscr{L}_{a}+\mathscr{L}_{s})\sigma_{t,g,i}\Omega\cdot R^{n+1}_{g,i}-(\mathscr{L}_{a}\sigma_{a,g,i}+\mathscr{L}_{s}\sigma_{s,g,i})O^{n+1}_{Q_{g},i},\quad 2\leq i\leq N_{x}\,;$ (6.9a) $\displaystyle\frac{1}{\mathcal{C}}\frac{\rho_{g,i+1/2}^{n+1}-\rho_{g,i+1/2}^{n}}{\Delta t}+\frac{1}{\mathcal{C}}\frac{E^{n+1}_{Q_{g},i+1/2}-E^{n}_{Q_{g},i+1/2}}{\Delta t}+\Omega\Omega\frac{R^{n+1}_{g,i+1}-R^{n+1}_{g,i}}{\Delta x}+\Omega\frac{O^{n+1}_{Q_{g},i+1}-O^{n+1}_{Q_{g},i}}{\Delta x}$ $\displaystyle\hskip 28.45274pt=\mathscr{L}_{a}\left(\sigma^{2}_{a,g,i+1/2}B^{n+1}_{g,i+1/2}-\sigma^{1}_{a,g,i+1/2}\rho^{n+1}_{g,i+1/2}\right)-(\mathscr{L}_{a}\sigma_{a,g,i+1/2}+\mathscr{L}_{s}\sigma_{s,g,i+1/2})E^{n+1}_{Q_{g},i+1/2},$ $\displaystyle\hskip 312.9803pt\quad 1\leq i\leq N_{x}\,.$ (6.9b) Here only a linear transport equation is solved with all direction and energy group being decoupled. Boundary conditions To cope with (6.1), using the relation (6.2), we have $E_{g,3/2}^{n+1}+\frac{1}{2}\left(O_{g,2}^{n+1}+O_{g,1}^{n+1}\right)=b_{\text{g,L}}(\Omega),~{}\Omega>0;\quad E_{g,N_{x}+1/2}^{n+1}-\frac{1}{2}\left(O_{g,N_{x}}^{n+1}+O_{g,N_{x}+1}^{n+1}\right)=b_{\text{g,R}}(-\Omega),~{}\Omega>0\,.$ (6.10) Hence, $O_{g,1}^{n+1}(\Omega)=2\left(b_{\text{g,L}}(\Omega)-E_{g,3/2}^{n+1}\right)-O_{g,2}^{n+1},\quad O_{g,N_{x}+1}^{n+1}(\Omega)=2\left(E_{g,N_{x}+\frac{1}{2}}^{n+1}-b_{\text{g,R}}(-\Omega)\right)-O_{g,N_{x}}^{n+1}.$ (6.11) To summarize, we have the following one time step update of the fully discrete version of FRTE. Input: $T_{i+1/2}^{n}$, $E_{g,i+1/2}^{s}$,$O_{g,i}^{n}$ Output: $T_{i+1/2}^{n+1}$,$E_{g,i+1/2}^{n+1}$,$O_{g,i}^{n+1}$ 1 get $R_{g,i}^{n+1}$, $\rho_{g,i+1/2}^{n+1}$, $T_{i+1/2}^{n+1}$ from the system (6.8); 2 get $O_{Q_{g},i}^{n+1}$ and $E_{Q_{g},i+1/2}^{n+1}$ from the system (6.9); obtain $E_{g,i+1/2}^{n+1}$,$O_{g,i}^{n+1}$ from (6.5). Algorithm 1 one step of fully discrete update for FRTE ### 6.2 The gray radiation diffusion regime of (6.8) In this subsection, the gray radiation diffusion regime is considered, i.e., $\mathscr{L}_{a}=1/\varepsilon$ and $\mathscr{L}_{s}=1/\varepsilon$. From (6.8b), one can get $\sigma^{1}_{a,g,i+1/2}\rho^{n+1}_{g,i+1/2}=\sigma^{2}_{a,g,i+1/2}B^{n+1}_{g,i+1/2}+\mathcal{O}(\varepsilon),$ (6.12) Summing the above equation over the group index $g$, and by the definitions (3.5), one can get $\int\limits_{0}^{\infty}\sigma_{a}\rho_{g,i+1/2}^{n+1}\text{d}\nu=\int\limits_{0}^{\infty}\sigma_{a}B_{g,i+1/2}^{n+1}\text{d}\nu+\mathcal{O}(\varepsilon),$ (6.13) which implies that $\rho^{n+1}_{g,i+1/2}=B^{n+1}_{g,i+1/2}+\mathcal{O}(\varepsilon).$ (6.14) Next, we will find the relation between $\rho^{n+1}_{g,i+1/2}$ and $R^{n+1}_{g,i}$. By using Champan-Enskog expansion in equation (6.9b), and using (6.14), one can get $\displaystyle-\sigma_{a,g,i+1/2}E_{Q_{g},i+1/2}^{n+1}=\mathcal{O}(\varepsilon).$ (6.15) Taking the second moment for $\Omega$ in equation (6.15) yields $\left\langle\Omega\Omega E_{Q_{g},i+1/2}^{n+1}\right\rangle=\mathcal{O}(\varepsilon).$ (6.16) By using (6.16), and letting $\varepsilon\to 0$, equation (6.8a) reduces to $-\frac{1}{\varepsilon}R^{n+1}_{g,i}=\frac{1}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T_{i}^{n+1})\text{d}\nu}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{a,i}+\sigma_{s,i}}\frac{\partial B}{\partial T}(T_{i}^{n+1})\text{d}\nu\frac{\rho^{n+1}_{g,i+1/2}-\rho^{n+1}_{g,i-1/2}}{\Delta x}.$ (6.17) Substituting (6.17) into equation (6.8b), adding up all the equations in each group, then adding it to equation (6.8c), using the equation (6.14) and letting $\varepsilon\to 0$, one can get $\displaystyle 4\pi\frac{\sum_{g=1}^{G}B^{n+1}_{g,i+1/2}-\sum_{g=1}^{G}B^{n}_{g,i+1/2}}{\Delta t}+C_{v}\frac{T_{i+1/2}^{n+1}-T_{i+1/2}^{n}}{\Delta t}=\sum_{g=1}^{G}\frac{1}{3\Delta x}$ (6.18) $\displaystyle\left[\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{4\pi}{\sigma_{a,i+1}+\sigma_{s,i+1}}\frac{\partial B}{\partial T}(T_{i+1}^{n+1})\text{d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T_{i+1}^{n+1})\text{d}\nu}\frac{B^{n+1}_{g,i+3/2}-B^{n+1}_{g,i+1/2}}{\Delta x}-\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{4\pi}{\sigma_{a,i}+\sigma_{s,i}}\frac{\partial B}{\partial T}(T_{i}^{n+1})\text{d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T_{i}^{n+1})\text{d}\nu}\frac{B^{n+1}_{g,i+1/2}-B^{n+1}_{g,i-1/2}}{\Delta x}\right]\,.$ Moreover, we have $4\pi\sum_{g=1}^{G}B^{n+1}_{g,i+1/2}=\int_{0}^{\infty}B(\nu,T^{n+1}_{i+1/2})\text{d}\nu=(T^{n+1}_{i+1/2})^{4},$ which is is consistent with equation (4.17), and the calculations of the other two gray radiation diffusion states are similar. ### 6.3 The frequency dependent diffusion regime of (6.8) In the frequency dependent diffusion regime, $\mathscr{L}_{a}=\varepsilon$ and $\mathscr{L}_{s}=1/\varepsilon$. By using Champan-Enskog expansion in equation (6.9b), one can get $\displaystyle-\sigma_{s,g,i+1/2}E_{Q_{g},i+1/2}^{n+1}=\mathcal{O}(\varepsilon).$ (6.19) Taking the second moment for $\Omega$ in equation (6.19) yields $\left\langle\Omega\Omega E_{Q_{g},i+1/2}^{n+1}\right\rangle=\mathcal{O}(\varepsilon).$ (6.20) By using (6.20), and letting $\varepsilon\to 0$, equation (6.8a) reduces to $-\frac{1}{\varepsilon}R^{n+1}_{g,i}=\frac{1}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T_{r,i}^{n+1})\text{d}\nu}\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{s,i}}\frac{\partial B}{\partial T}(T_{r,i}^{n+1})\text{d}\nu\frac{\rho^{n+1}_{g,i+1/2}-\rho^{n+1}_{g,i-1/2}}{\Delta x}.$ (6.21) Substituting (6.21) into equation (6.8b), and letting $\varepsilon\to 0$, one can get $\displaystyle\frac{\rho_{g,i+1/2}^{n+1}-\rho_{g,i+1/2}^{n}}{\Delta t}$ $\displaystyle-\frac{1}{3\Delta x}\left[\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{s,i+1}}\frac{\partial B}{\partial T}(T_{r,i+1}^{n+1})\text{d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T_{r,i+1}^{n+1})\text{d}\nu}\frac{\rho^{n+1}_{g,i+3/2}-\rho^{n+1}_{g,i+1/2}}{\Delta x}\right.$ (6.22) $\displaystyle\left.-\frac{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{1}{\sigma_{s,i}}\frac{\partial B}{\partial T}(T_{r,i}^{n+1})\text{d}\nu}{\int_{\nu_{g-\frac{1}{2}}}^{\nu_{g+\frac{1}{2}}}\frac{\partial B}{\partial T}(T_{r,i}^{n+1})\text{d}\nu}\frac{\rho^{n+1}_{g,i+1/2}-\rho^{n+1}_{g,i-1/2}}{\Delta x}\right]=\sigma^{2}_{a,g,i+\frac{1}{2}}B^{n+1}_{g,i+\frac{1}{2}}-\sigma^{1}_{a,g,i+\frac{1}{2}}\rho^{n+1}_{g,i+\frac{1}{2}}.$ Moreover, letting $\varepsilon\to 0$, equation (6.8c) reduces to $C_{v}\frac{T_{i+1/2}^{n+1}-T_{i+1/2}^{n}}{\Delta t}=\sum\limits_{g^{\prime}=1}^{G}4\pi\left(\sigma^{1}_{a,g^{\prime},i+\frac{1}{2}}\rho^{n+1}_{g^{\prime},i+\frac{1}{2}}-\sigma^{2}_{a,g^{\prime},i+\frac{1}{2}}B^{n+1}_{g^{\prime},i+\frac{1}{2}}\right).$ (6.23) which is consistent with equation (4.24). ## 7 Numerical Examples In this section, we conduct several numerical examples to test the performance of the AP scheme for the FRTE. The units of the length, time, temperature and energy are respectively $cm,\ ns,\ keV$ and $Jk$ ($1\ Jk=10^{9}\ J$). With the above units, let the speed of light $c$ be 29.98 $cm\ ns^{-1}$, the specific heat capacity $C_{v}$ be 0.1 $Jk\ keV^{-1}\ cm^{-3}$ and the radiation constant $a_{r}$ be 0.01372 $Jk\ keV^{-4}\ cm^{-3}$. As in [densmore2012hybrid, sunj], in all of these examples, the opacities take the following forms : $\sigma_{a}=\frac{\sigma_{a0}(x)}{(h\nu)^{3}\sqrt{kT}},\qquad\sigma_{s}=\frac{\sigma_{s0}(x)}{(h\nu)^{3}\sqrt{kT}}.$ (7.1) In all numerical examples in this section, similar as in [densmore2012hybrid, sunj], we divide the frequency domain from 0.1 $eV$ to 100 $keV$ into 30 frequency groups, and the groups are logarithmically spaced. In each group, we determine the group absorption, scattering and total cross section by using (4.7) and (7.1). Moreover, from Figure 2, the piece-wise constant approximations $\sigma_{a,g}^{c}$ are closer to the reference solutions, thus we use $\omega_{1}=\omega_{2}=1$ in all examples. ### 7.1 Example 1 In this example, we will show the convergence order and stability of the proposed scheme. The thickness of the computational domain is $2\ cm$ and the initial temperature is $T(0,x)=\max\big{\\{}1-20(x-1)^{2},10^{-3}\big{\\}}\ keV.$ (7.2) The initial radiation intensity is set to be a Planck distribution evaluated at the temperature (7.2). Zero Dirichlet boundary condition is used in this example. Four different sets of $\sigma_{a0}$, $\sigma_{s0}$ are tested: 1) $\sigma_{a0}=1\ keV^{7/2}\ cm^{-1}$, $\sigma_{s0}=1000\ keV^{7/2}\ cm^{-1}$; 2) $\sigma_{a0}=1000\ keV^{7/2}\ cm^{-1}$, $\sigma_{s0}=1\ keV^{7/2}\ cm^{-1}$; 3) $\sigma_{a0}=1000\ keV^{7/2}\ cm^{-1}$, $\sigma_{s0}=1000\ keV^{7/2}\ cm^{-1}$; 4) $\sigma_{a0}=1\ keV^{7/2}\ cm^{-1}$, $\sigma_{s0}=1\ keV^{7/2}\ cm^{-1}$. In the four cases, the first case belongs to the frequency dependent diffusion regime, and the second case and the third case are the gray radiation diffusion regime, and $\sigma_{a}$ and $\sigma_{s}$ are both optical thin in the last case. In Fig. 3, we plot the errors $\text{error}_{T}=\|T_{\Delta x}(\cdot,t_{\text{max}})-T_{\Delta x/2}(\cdot,t_{\text{max}})\|_{l_{1}}$ (7.3) with $\Delta x=2*10^{-2}\ cm$, $10^{-2}\ cm$, $5*10^{-3}\ cm$, $2.5*10^{-3}\ cm$, $1.25*10^{-3}\ cm$ and $\Delta t=2*10^{-2}\ ns$, $10^{-2}\ ns$, $5*10^{-3}\ ns$, $2.5*10^{-3}\ ns$, $1.25*10^{-3}\ ns$, and $t_{\text{max}}=1\ ns$. A uniform first order accuracy can be observed for all four different cases. Moreover, it is important to point out that the scheme is stable when the time step $\Delta t=8\Delta x$ for all the cases. Figure 3: Example 1. Errors of different $\Delta x$, $\Delta t$ for the four cases. Here different $\Delta x=2*10^{-2}\ cm$, $10^{-2}\ cm$, $5*10^{-3}\ cm$, $2.5*10^{-3}\ cm$, $1.25*10^{-3}\ cm$ are tested and $\Delta t$ are chosen to be $\Delta t=2*10^{-2}\ ns$, $10^{-2}\ ns$, $5*10^{-3}\ ns$, $2.5*10^{-3}\ ns$, $1.25*10^{-3}\ ns$ respectively. At last, we compare the results in the gray radiation diffusion regime and frequency dependent diffusion regime to the reference results. For the four cases, we use a uniform spatial mesh and the cell size $\Delta x=0.01\ cm$, and the time step $\Delta t=0.04\ ns$. Moreover, the reference solution of all the cases are computed with $\Delta x=0.001\ cm$ and $\Delta t=0.00001\ ns$. The numerical results at $t=2\ ns$ are displayed in Fig. 4. The material temperature from our AP method agrees well with the reference solution in all the four cases. (a) $\sigma_{a0}=1\ keV^{7/2}\ cm^{-1},\ \sigma_{s0}=1000\ keV^{7/2}\ cm^{-1}$ (b) $\sigma_{a0}=1000\ keV^{7/2}\ cm^{-1},\ \sigma_{s0}=1\ keV^{7/2}\ cm^{-1}$ (c) $\sigma_{a0}=1000\ keV^{7/2}\ cm^{-1},\sigma_{s0}=1000\ keV^{7/2}\ cm^{-1}$ (d) $\sigma_{a0}=1\ keV^{7/2}\ cm^{-1},\sigma_{s0}=1\ keV^{7/2}\ cm^{-1}$ Figure 4: Example 1. Comparison of the material temperature using our AP scheme for equilibrium regime and the reference solution at $t=2\ ns$. ### 7.2 Example 2 We test an example in [densmore2012hybrid, sunj, steinberg2022multi] where particles are injected from the left boundary. Three space homogeneous problems with $\sigma_{a0}=10\ keV^{7/2}\ cm^{-1},\ 100\ keV^{7/2}\ cm^{-1}$, $1000\ keV^{7/2}\ cm^{-1}$, and $\sigma_{s0}=0\ keV^{7/2}\ cm^{-1}$ are tested. The thickness of the computational domain is $5\ cm$, the initial temperature is $10^{-3}\ keV$ and the initial radiation intensity is a Planck distribution evaluated at the initial temperature. On the left boundary, the incident radiation intensity is given by a Planck distribution with the temperature being 1 $keV$, and we use a reflective boundary condition on the right boundary. For all three tests, we use a uniform spatial mesh with $\Delta x=0.005\ cm$ and $\Delta t=0.005\ $ns to ensure the scheme stable. The reference solution is calculated using a finer temporal mesh such that $\Delta t=0.00001\ $ns. Moreover, in the example , the time size $\Delta t=0.005$ ns used in our AP method is much bigger than the time step $\Delta t=0.004/c$ ns used in [sunj]. The numerical results at $t=1\ ns$ are displayed in Fig. 5 (Note that we only show the part where the material temperature differs significantly from the initial value in the second and third figure). The good agreement between our solution and the reference solution indicates that our method works well in the optical thin and thick regime. (a) $\sigma_{a0}=10\ keV^{7/2}\ cm^{-1}$ (b) $\sigma_{a0}=100\ keV^{7/2}\ cm^{-1}$ (c) $\sigma_{a0}=1000\ keV^{7/2}\ cm^{-1}$ Figure 5: Example 2. Plot of three different cases at $t=1\ ns$. The red curves are reference solutions computed in the finer mesh. Here $\Delta x=0.005\ cm$, in our AP scheme $\Delta t=0.005\ ns$, and in the reference solution, $\Delta t=0.00001\ ns$. ### 7.3 Example 3 A 1D Marshak wave problem similar as in [steinberg2022multi, densmore2012hybrid, sunj] is tested, where both optical thin and thick regions coexist. The opacities take the forms of (7.1) with $\sigma_{a0}=\begin{cases}10\quad\mathrm{keV}^{7/2}\ cm^{-1},&0<x<2.0\ cm,\\\ 1000\quad\mathrm{keV}^{7/2}\ cm^{-1},&2.0<x<3.0~{}cm,\end{cases}$ and $\sigma_{s0}=0$. Moreover, the width of the computational domain is $3$cm and we divide it uniformly into $150$ spatial cells. The time step is $\Delta t=0.02\ ns$. The problem begins with flat material and radiation temperature profiles at 1 $eV$. A boundary source of 1 $keV$ is placed at the left boundary, while the right boundary is vacuum. In this example, the reference solution is computed with $\Delta x=0.001\ cm$ and $\Delta t=0.00001\ ns$. The numerical results at $t=1\ ns$ are displayed in Fig. 6. Moreover, the time size $\Delta t=0.02$ ns of our AP method in the example is much bigger than the time step $\Delta t=0.016/c$ ns used in [sunj]. The material temperature from our AP method agrees well with the reference solution in the co-exit regime. Figure 6: Example 3. Plot of multi-region problem at $t=1\ ns$. The red curves are reference solutions computed in the finer mesh. Here $\Delta x=0.02\ cm$, $\Delta t=0.02\ ns$ in our AP scheme, and in the reference solution, $\Delta x=0.005\ cm$ and $\Delta t=0.00001\ ns$. ## 8 Conclusion In this paper, we propose an AP decomposed multi-group scheme for the FRTE in the gray radiation diffusion regime and the frequency-dependent diffusion regime. The idea is to decompose the intensity into three parts: the zeroth moment of the intensity, the first moment of the intensity, and the residual term. The main contribution is the new decomposed multi-group energy discretization, which allows tuning of the approximations of the coefficients in the radiation diffusion regime. In the full AP discretization, the two macroscopic moments and the material temperature are updated first by treating them implicitly with the explicit residual term. After the zeroth and first-order moments are obtained, we update the residual term by solving implicitly a linear transport equation for each decoupled direction and energy group. The AP property is proved by asymptotic analysis. Numerical results of both optically thin and thick regions are presented. We can see that coarse meshes in energy, space, and time can be used to get the correct results. We will focus on the extension of the algorithm to high-dimensional problems and curvilinear geometrical problems in the future. Acknowledgement: X. J. Zhang and M. Tang were supported by NSFC12031013, Shanghai pilot innovation project 21JC1403500 and the Strategic Priority Research Program of Chinese Academy of Sciences Grant No.XDA25010401; X. J. Zhang was supported by China Postdoctoral Science Foundation No.2022M722107; P. Song was supported by NSFC12031001; Y. Shi was supported by NSFC12001052. ## References ## Appendix A A The derivation of the two limiting models In the scaling of (2.5a), we assume that the intensity $I$ and material temperature $T$ can be expanded as in [MWTS]: $\displaystyle I=I^{(0)}+\varepsilon I^{(1)}+\varepsilon^{2}I^{(2)}+\cdots,$ (A.1) $\displaystyle T=T^{(0)}+\varepsilon T^{(1)}+\varepsilon^{2}T^{(2)}+\cdots,$ and the Planck function follows that $B=B^{(0)}+\varepsilon B^{(1)}+\cdots,$ (A.2) where $B^{(0)}=B|_{\varepsilon=0}=B|_{T=T^{{(0)}}},$ $B^{(1)}=\frac{\partial B}{\partial\varepsilon}\Bigg{|}_{\varepsilon=0}=\frac{\partial B}{\partial T}\frac{\partial T}{\partial\varepsilon}\Bigg{|}_{\varepsilon=0}=\frac{\partial B}{\partial T}\Bigg{|}_{T=T^{{(0)}}}T^{(1)},$ By substituting ansatzes (A.1) and (A.2) into equation (2.4a) and collecting the terms of the same order in $\varepsilon$, we have $\displaystyle\mathcal{O}\left(\frac{1}{\varepsilon}\right)$ $\displaystyle:I^{(0)}=B^{(0)},$ (A.3a) $\displaystyle\mathcal{O}\left(1\right)$ $\displaystyle:\Omega\cdot\nabla I^{(0)}=\sigma_{a}\left(B^{(1)}-I^{(1)}\right).$ (A.3b) Taking (2.4a) integral with respect to $\nu$ from 0 to $\infty$, and taking its integral with respect to $\Omega$, then adding the obtained equation to (2.4b), and dividing both sides of the equation by $\varepsilon$, one has $4\pi\partial_{t}\int\limits_{0}^{\infty}\rho\text{d}\nu+\frac{1}{\varepsilon}\int\limits_{4\pi}\int\limits_{0}^{\infty}\Omega\cdot\nabla I\text{d}\nu\text{d}\Omega+C_{v}\partial_{t}T=0.$ By using (A.3a), (A.3b) and condition $\left\langle\Omega\Omega\right\rangle:=D_{d}=\frac{1}{3}I_{d}$ (where $I_{d}$ denotes the 3 by 3 identity matrix), and sending $\varepsilon\to 0$, which give $4\pi\partial_{t}\int\limits_{0}^{\infty}B^{(0)}\text{d}\nu+C_{v}\partial_{t}T^{(0)}=\int\limits_{0}^{\infty}\nabla\cdot\left(\frac{4\pi D_{d}}{\sigma_{a}}\nabla B^{(0)}\right)\text{d}\nu.$ (A.4) by using the following relations $\int\limits_{0}^{\infty}4\pi B\ \text{d}\nu=T^{4},\qquad\int\limits_{0}^{\infty}4\pi\frac{\partial B}{\partial T}\ \text{d}\nu=4T^{3},$ then (A.4) reduces to $\partial_{t}(T^{(0)})^{4}+C_{v}\partial_{t}T^{(0)}=\nabla\cdot\left(\frac{D_{d}}{\sigma_{R}}\nabla(T^{(0)})^{4}\right)\,,$ (A.5) with the Rosseland mean opacity $\sigma_{R}$ which is given by $\frac{1}{\sigma_{R}(x,T)}\equiv\frac{\int_{0}^{\infty}\frac{4\pi}{\sigma_{a}(x,\nu,T)}\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}{\int_{0}^{\infty}4\pi\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}=\frac{\int_{0}^{\infty}\frac{4\pi}{\sigma_{a}(x,\nu,T)}\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}{4(T^{(0)})^{3}}.$ In the scaling of (2.5b), the same expansions (A.1) and (A.2) are used, moreover, substitute (A.1) and (A.2) into the radiation transfer equation and collect the terms of the same order in $\varepsilon$, which gives $\displaystyle\mathcal{O}\left(\frac{1}{\varepsilon}\right)$ $\displaystyle:I^{(0)}=\rho^{(0)},$ (A.6a) $\displaystyle\mathcal{O}\left(1\right)$ $\displaystyle:\Omega\cdot\nabla I^{(0)}=\sigma_{a}\left(B^{(0)}-I^{(0)}\right)+\sigma_{s}\left(\rho^{(1)}-I^{(1)}\right).$ (A.6b) Taking (A.6b) integral with respect to $\Omega$, one has $B^{(0)}=\rho^{(0)},$ by similar calculations in the first part, one can get the following equilibrium system: $\partial_{t}(T^{(0)})^{4}+C_{v}\partial_{t}T^{(0)}=\nabla\cdot\left(\frac{D_{d}}{\sigma_{S}}\nabla(T^{(0)})^{4}\right)\,,$ (A.7) and the mean opacity $\sigma_{S}$ is given by $\frac{1}{\sigma_{S}}\equiv\frac{\int_{0}^{\infty}\frac{4\pi}{\sigma_{s}}\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}{\int_{0}^{\infty}4\pi\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}=\frac{\int_{0}^{\infty}\frac{4\pi}{\sigma_{s}}\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}{4(T^{(0)})^{3}}.$ In the scaling of (2.5c), the same expansions (A.1) and (A.2) are used, moreover, substitute (A.1) and (A.2) into the radiation transfer equation and collect the terms of the same order in $\varepsilon$, which gives $\displaystyle\mathcal{O}\left(\frac{1}{\varepsilon}\right)$ $\displaystyle:(\sigma_{a}+\sigma_{s})I^{(0)}=\sigma_{a}B^{(0)}+\sigma_{s}\rho^{(0)},$ (A.8a) $\displaystyle\mathcal{O}\left(1\right)$ $\displaystyle:\Omega\cdot\nabla I^{(0)}=\sigma_{a}\left(B^{(1)}-I^{(1)}\right)+\sigma_{s}\left(\rho^{(1)}-I^{(1)}\right).$ (A.8b) Taking (A.8a) and (A.8b) integral with respect to $\Omega$, one has $B^{(0)}=\rho^{(0)},\qquad B^{(1)}=\rho^{(1)},$ by similar calculations in the first part, one can get the following equilibrium system: $\partial_{t}(T^{(0)})^{4}+C_{v}\partial_{t}T^{(0)}=\nabla\cdot\left(\frac{D_{d}}{\sigma_{T}}\nabla(T^{(0)})^{4}\right)\,,$ (A.9) and the mean opacity $\sigma_{T}$ is given by $\frac{1}{\sigma_{T}}\equiv\frac{\int_{0}^{\infty}\frac{4\pi}{(\sigma_{a}+\sigma_{s})}\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}{\int_{0}^{\infty}4\pi\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}=\frac{\int_{0}^{\infty}\frac{4\pi}{(\sigma_{a}+\sigma_{s})}\frac{\partial B^{(0)}(\nu,T)}{\partial T}\text{d}\nu}{4(T^{(0)})^{3}}.$ In the frequency-dependent radiative diffusion regime, the same expansions (A.1) and (A.2) are used, moreover, substitute (A.1) and (A.2) into equation (2.9a) and collect the terms of the same order in $\varepsilon$, which gives $\displaystyle\mathcal{O}\left(\frac{1}{\varepsilon}\right)$ $\displaystyle:I^{(0)}=\rho^{(0)},$ (A.10a) $\displaystyle\mathcal{O}\left(1\right)$ $\displaystyle:\Omega\cdot\nabla I^{(0)}=\sigma_{s}\left(\rho^{(1)}-I^{(1)}\right).$ (A.10b) Taking (2.9a) integral with respect to $\Omega$, and dividing both sides of the equation by $\varepsilon$, one has $4\pi\partial_{t}\rho+\frac{4\pi}{\varepsilon}\left\langle\Omega\cdot\nabla I\right\rangle=4\pi\sigma_{a}(B-\rho),$ by similar calculations in the first part, one can get the following non- equilibrium system: $\left\\{\begin{aligned} &\partial_{t}\rho^{(0)}-\nabla\cdot\left(\frac{D_{d}}{\sigma_{s}}\nabla\rho^{(0)}\right)=\sigma_{a}(B^{(0)}-\rho^{(0)}),\\\ &C_{v}\partial_{t}T^{(0)}=\int_{0}^{\infty}4\pi\sigma_{a}(\rho^{(0)}-B^{(0)})\text{d}\nu.\ \ \\\ \end{aligned}\right.$ (A.11)
# Automated anomaly-aware 3D segmentation of bones and cartilages in knee MR images from the Osteoarthritis Initiative Boyeong Woo<EMAIL_ADDRESS>Craig Engstrom William Baresic Jurgen Fripp Stuart Crozier Shekhar S. Chandra School of Information Technology and Electrical Engineering, The University of Queensland, Australia School of Human Movement and Nutrition Sciences, The University of Queensland, Australia Australian e-Health Research Centre, Commonwealth Scientific and Industrial Research Organisation, Australia ###### Abstract In medical image analysis, automated segmentation of multi-component anatomical structures, which often have a spectrum of potential anomalies and pathologies, is a challenging task. In this work, we develop a multi-step approach using U-Net-based neural networks to initially detect anomalies (bone marrow lesions, bone cysts) in the distal femur, proximal tibia and patella from 3D magnetic resonance (MR) images of the knee in individuals with varying grades of osteoarthritis. Subsequently, the extracted data are used for downstream tasks involving semantic segmentation of individual bone and cartilage volumes as well as bone anomalies. For anomaly detection, the U-Net- based models were developed to reconstruct the bone profiles of the femur and tibia in images via inpainting so anomalous bone regions could be replaced with close to normal appearances. The reconstruction error was used to detect bone anomalies. A second anomaly-aware network, which was compared to anomaly- naïve segmentation networks, was used to provide a final automated segmentation of the femoral, tibial and patellar bones and cartilages from the knee MR images containing a spectrum of bone anomalies. The anomaly-aware segmentation approach provided up to 58% reduction in Hausdorff distances for bone segmentations compared to the results from the anomaly-naïve segmentation networks. In addition, the anomaly-aware networks were able to detect bone lesions in the MR images with greater sensitivity and specificity (area under the receiver operating characteristic curve [AUC] up to 0.896) compared to the anomaly-naïve segmentation networks (AUC up to 0.874). ###### keywords: Anomaly detection, Segmentation, U-Net, Knee osteoarthritis, MRI ††journal: Medical Image Analysis ## 1 Introduction Deep learning methods, particularly convolutional neural networks (CNNs), have shown promising results in image recognition tasks and are a rapidly evolving area of research in the field of automated medical image analysis. Common deep learning-based computer vision tasks that have been applied to medical imaging include classification, detection, and segmentation [24]. For example, U-Net, proposed by Ronneberger et al. [23], has become a popular CNN model for medical image segmentation. There are now several publicly available medical image databases, such as the Osteoarthritis Initiative [19], which provide the basis for deep learning research in medical imaging. Major challenges in automated analysis of medical images using deep learning include lack of annotated data and variable presence of anomalies. Manual annotation of tomographic images from techniques such as computed tomography (CT) and magnetic resonance (MR) imaging is typically expertise- and time- intensive. This makes efficient automated volumetric image processing particularly desirable in medical imaging but also presents a major challenge to deep learning research in medical imaging because there are often not enough annotated data for training a deep learning model. In addition, incidental anomalies (e.g. lesions, anatomical variations, imaging artefacts, etc.) apart from any “primary” pathoanatomy of interest are frequently present in clinical settings, making automated segmentation of multi-component anatomical structures difficult. A method to identify and quantify “secondary” anomalies will be required for robust segmentation performance. This work is an extension to preliminary experiments presented in Woo et al. [26] involving segmentation of the distal femur and proximal tibia using U-Net-based networks. In the current work, we develop an approach based on CNN to detect bone anomalies (bone marrow lesions, bone cysts) in the distal femur, proximal tibia and patella from knee MR images from individuals with varying levels of osteoarthritis. The rational was that the output from the anomaly detector can also be utilized for improving segmentation of the knee with osteoarthritis containing bone lesions. The main contributions of this work include: 1. 1. The use of 3D U-Net-based CNNs for anomaly detection via inpainting in medical imaging. While there have been previous works that used U-Nets for inpainting and visual anomaly detection [15, 29], they have focused on natural images or 2D image slices. This work extends such ideas to 3D medical images using fully 3D inpainting. 2. 2. Using information learned from anomaly detection to improve segmentation of the distal femur, proximal tibia and patella from knee MR images using an anomaly-aware CNN approach. The current work shows that the proposed approach is capable of significantly improving the segmentation of femur and tibia on osteoarthritic knee MR images with bone anomalies. 3. 3. Demonstrating that the anomaly-aware approach has an advantage when the size of the training dataset is limited. We added additional labels to the segmentation task, including the patella and bone lesions, which are relatively more difficult to segment due to their smaller volumes and greater variability. The proposed anomaly-aware approach is able to detect and segment these new labels with moderate accuracy despite the limited size of the training dataset. 4. 4. Evaluation of the effect of anomaly-aware approach on segmentation performance using two different categories of CNNs: 3D U-Net and 3D context aggregation network (CAN). The current work shows that the anomaly-aware approach can be readily applied to different segmentation CNNs to improve their segmentation performance as well as transfer learning. ## 2 Background In recent years, several machine learning algorithms have been proposed for automatic anomaly detection. Unsupervised methods using generative models were shown to be promising [3, 4] and are particularly useful in medical imaging for which labeled data can be extremely difficult to obtain. A recent work by Pinaya et al. [20] showed that leading-edge techniques such as transformers can also be utilized for anomaly detection and segmentation. However, application of transformers to 3D images is currently challenging due to their very high demands on data and computational resources. Convolutional autoencoders such as U-Net are computationally less intensive than transformers. Pinaya et al. [20] also used vector quantized variational autoencoder (VQ-VAE) for dimensionality reduction which allowed them to apply transformers. U-Net, developed by Ronneberger et al. [23], is a popular CNN model for semantic segmentation. It is a fully convolutional network with an autoencoder structure, consisting of a contracting path (encoder) and an expansive path (decoder). The distinguishing feature of U-Net is the skip connections that transfer feature maps at each resolution from the encoder to the decoder. The skip connections recover spatial information lost during downsampling, which is critical for segmentation tasks. The original U-Net proposed was a 2D convolutional network for segmentation of 2D images. Çiçek et al. [5] extended this work to dense volumetric segmentation using 3D U-Net. V-Net [16] is another autoencoder-based model similar to U-Net that is trained end-to-end on image volumes. Some later works on volumetric segmentation using 3D CNNs (typically U-Net- like) have employed a technique referred to as “deep supervision” [12, 10, 22]. The main idea of deep supervision is to provide integrated direct supervision to the hidden layers, rather than providing supervision only at the output layer [14]. Deep supervision has been found to speed up convergence likely because it encourages deeper layers to produce improved segmentation results [12]. The skip connections in U-Net force aggregation only at the same-scale feature maps, and it was shown that redesigned skip connections and deep supervision in U-Net enable a significantly higher level of segmentation performance [30]. In addition to segmentation, CNNs with U-Net architecture have been applied for different tasks, such as image-to-image translation [11]. More recently, Liu et al. [15] adopted a U-Net-based model to perform image inpainting and generate synthetic brain tissue intensities for a tumor region. Zavrtanik et al. [29] proposed an anomaly detection method whereby a U-Net-based model was used to reconstruct images from partial inpaintings which were then used to localize visual anomalies in the images. The current work is similar to this idea and perform unsupervised anomaly detection on unlabeled 3D knee MR images through inpainting and lossy reconstruction. The approach is to “erase” a region of interest (ROI) from an image, which may or may not contain anomalies, and then let the network generate synthetic tissue intensities in the region without the potential anomalies. The difference between the original image and the reconstructed image can then be used to detect the anomalies. Here, we use a version of 3D U-Net for efficient volumetric image processing. Fig. 1: Overall pipeline for our method including anomaly detection and the downstream segmentation task. Components 1 and 2 are the anomaly detection models (Sections 3.1 and 3.2). Component 3 is the downstream segmentation task using the anomaly-aware segmentation approach (Section 3.3). As an alternative to autoencoder-based CNNs, context aggregation network (CAN) was proposed by Yu and Koltun [27] for semantic segmentation. It is structurally different from U-Net in that it is not based on an encoder- decoder architecture. In U-Net, the progressive downsampling achieves the effect of integrating contextual information at multi-scale, and the lost resolution is recovered through upsampling. However, since semantic segmentation task requires full-resolution output, there is the question of whether such downsampling and upsampling are truly necessary, which is why Yu and Koltun [27] proposed using dilated convolutions rather than downsampling for multi-scale context aggregation. As with U-Net, CAN was proposed as a 2D network, but Dai et al. [6] extended it to a 3D version for volumetric segmentation. Accurate segmentation of knee bones and cartilages is an important step in processing and analyzing knee MR images in the context of osteoarthritis. The use of CNNs has grown in knee tissue segmentation to automatically learn image features [7]. The first published work in this area is Prasoon et al. [21], which used a triplanar patch-based approach where three CNNs were trained to classify the central voxel as cartilage or background. Although this triplanar approach has now become obsolete, the finding that it achieved a better performance than the previous state-of-the-art showed that a CNN is capable of learning image features from knee MR images. In our preliminary experiments using 3D U-Net and 3D CAN for knee MR image segmentation, the CNNs were able to achieve mean Dice similarity coefficients (DSCs) for bones and cartilages in the knee joint similar to Ambellan et al. [2] for the OAI ZIB dataset (dataset described in Section 4.1). However, it was observed that the CNN-based segmentation of images was difficult for cases with visible coexisting abnormalities. Without shape regularization, the surface distance errors tended to be high because CNN outputs often contained “holes” and “noises” (false negatives/positives) due to the localized nature of the CNN-based classification. Indeed, Ambellan et al. [2] used a combination of U-Nets and statistical shape models (SSMs), and the authors explicitly state that SSM regularization of the CNN outputs was needed to attain “anatomically plausible” segmentations, which is consistent with our preliminary finding. In particular, it was difficult to achieve a good segmentation with a U-Net or CAN when there was a large anomaly in the image (see Figure 6). Therefore, we also propose an anomaly-aware segmentation mechanism for CNNs which is more robust in the presence of anomalies. ## 3 Methods The overall pipeline for the current work has 3 major components (Figure 1). In Component 1, the anatomical regions of interest—here, the distal femur and proximal tibia profiles—were erased from the images using the reference segmentation masks, and then these regions were inpainted using a 3D U-Net- based model ($G$). In Component 2, the outputs from Component 1 and another 3D U-Net-based model ($A$) were used to change anomalous bone regions in the original images to close to normal appearances. In Component 3, a 3D CNN-based segmentation network ($S$), which utilizes the information extracted from Component 2, was used to guide the automated segmentation of bone and cartilage volumes compared to the vanilla segmentation networks, specifically aiming to improve segmentation of the structures from images containing visible bone anomalies. ### 3.1 Anomaly detection using masked images Using the reference segmentation masks available in the OAI ZIB MR image dataset (see Section 4.1), the profiles of the femur and tibia were erased from the MR images. To ensure that the boundaries are not missed and to include some periarticular areas, the masks were dilated by 50 pixels in all directions before being applied to the images. The masked images were then used as input to the 3D U-Net-based model referred to as $G$, which was trained to reconstruct the original image, i.e. to inpaint the erased area. (a) (b) Fig. 2: The two anomaly detection networks with a 3D U-Net-based architecture. Blue boxes represent feature maps, with the number of channels denoted above each box. (a) Network $G$ regenerates the original images from masked images through inpainting and decoding of compressed images. The compressed images are provided by a small network $C$ trained concurrently with $G$. (b) Network $A$ is trained using the images generated by $G$ as the target output given the original images. The compressed images from $C$ are added to the decoder. IN: Instance Normalization. Since the erased profile was relatively large, an image compressor $C$ was added here to assist with the inpainting. This $C$ was trained simultaneously with $G$ to compress the original MR image, and the compressed image was then fed into the decoder part of $G$. The compressor $C$ had to be a very small network because the bone anomalies must not be recovered in the output; a larger network is likely to restore the anomalies in the inpainted images. Figure 2(a) shows the overall structure of the model. The loss function for training $G$ and $C$ was the mean squared error (MSE) between the original image $x$ and the regenerated image $G(x)$: $\mathcal{L}_{G}=||x-G(x)||_{2}^{2}.$ (1) It was expected that the model $G$ would recover most of the image but not the anomalous bone regions, and therefore, the squared differences between the original image and the inpainted image $\mathrm{E}(G(x))=(x-G(x))^{2}$ were used to highlight the anomalies. The region with a larger difference is more likely to be an anomalous region. ### 3.2 Anomaly detection using the original images A main limitation of the above model $G$ is that it requires a segmentation mask to start with, in order to generate input for the network. Therefore, another network $A$ was trained, which takes the original images (without any masking) as its input and the output images of $G$ as its target output. The network architecture of $A$ was the same as $G$, and the image compressor $C$, trained previously with $G$, was added here as well. Figure 2(b) shows the overall structure of the model. The loss function for training $A$ was the mean squared error (MSE) between the output from the previous network $G(x)$ and the output from the current network $A(x)$, plus the MSE between the respective error images $\mathrm{E}(G(x))=(x-G(x))^{2}$ and $\mathrm{E}(A(x))=(x-A(x))^{2}$ to further guide the model: $\mathcal{L}_{A}=||G(x)-A(x)||_{2}^{2}+||\mathrm{E}(G(x))-\mathrm{E}(A(x))||_{2}^{2}.$ (2) Since the outputs of $G$ were used as the target output, it was expected that the model $A$ would produce outputs that are very similar to the outputs from $G$, but it has the main advantage of not requiring segmentation masks to generate the input. This model can be used to detect bone anomalies when given the original images only. Again, the squared differences between the original image and the generated image $\mathrm{E}(A(x))=(x-A(x))^{2}$ were used to highlight the bone anomalies. ### 3.3 Downstream task: Anomaly-aware segmentation (a) (b) Fig. 3: The anomaly-aware segmentation network $S$ based on (a) 3D U-Net and (b) 3D CAN with deep supervision. The information extracted from the anomaly detector $A$ was utilized to inform the segmentation of the distal femur and proximal tibia from the knee MR images containing bone abnormalities. To demonstrate the utility of the model $A$, the error images $\mathrm{E}(A(x))$ were utilized to construct segmentation models which can manage anomalies. The anomaly-aware mechanism is a generalized method that can be applied to a CNN-based segmentation network. In this study, the proposed method was tested on two different types of CNNs: 3D U-Net [5] and 3D CAN [6]. Figure 3(a) shows the 3D U-Net architecture, with 4 downsamplings and upsamplings. Figure 3(b) shows the modified 3D CAN. Ideally, CAN would have no downsampling, but it was impossible to use a reasonable number of filters (at least 32) in the full-resolution (160 slices $\times$ 384 $\times$ 384) layers due to the memory limitation of the graphics card, so two downsampling (and upsampling) blocks were added as a trade-off; the CAN module is applied after the two downsamplings. To help stabilize convergence, both networks were modified with deep supervision (see Section 2) by producing secondary segmentation maps at deeper levels of the network and combining them with the final segmentation map via upsampling and element-wise summation. Although the original U-Net [23] used the conventional categorical cross- entropy as the loss function, a multi-class Dice loss is now often used in medical image segmentation because it intrinsically addresses the class imbalance problem commonly seen with medical images [10]. The multi-class Dice loss function is defined as: $\mathcal{L}_{DSC}=1.0-\frac{2}{|K|}\sum_{k\in K}{\frac{\sum_{i}{u_{i,k}v_{i,k}}}{\sum_{i}{u_{i,k}}+\sum_{i}{v_{i,k}}}}.$ (3) Here, $u$ is the softmax output of the network and $v$ is the one-hot encoded ground truth segmentation map; $K$ is the number of classes, and $u_{i,k}$ and $v_{i,k}$ denote the softmax output and ground truth label, respectively, for class $k$ at voxel $i$. While the conventional U-Net is able to produce plausible segmentation results for most images without anomalies, it can easily fail when there are anomalies in the images since anomalies were not taken into account. To address this problem, the error images $\mathrm{E}(A(x))$ were added as an additional input (Figure 3) along with an additional loss function. Inspired by the work of Nie and Shen [17], which used a difficulty-aware attention mechanism, a focal cross-entropy loss was added to the loss function, where the focal weights were given by the error images: $\mathcal{L}_{FCE}=-\sum_{i}{\sum_{k\in K}{F_{i}v_{i,k}\log{u_{i,k}}}},$ (4) where $F=1.0+\beta\mathrm{E}(A(x))$ and $F_{i}$ denotes the focal weight at voxel $i$. Here, a weighting factor of $\beta=99.0$ was used so that the values of $F$ ranges from 1 to 100. (If $\beta=0.0$, Equation 4 would be the same as the usual categorical cross-entropy.) The total loss for the segmentation network was then: $\mathcal{L}_{S}=\mathcal{L}_{DSC}+\alpha\mathcal{L}_{FCE},$ (5) where $\alpha$ is another weighting factor. Here, $\alpha=10.0$ was used. Using this loss, the segmentation network can be trained to pay more attention to the bone voxels that were found to be anomalous by the network $A$ and hence likely to be difficult for the segmentation network to classify. For a segmentation task, it is assumed that segmentation masks are not available for test images, so only the outputs from $A$ (not $G$), which did not require segmentation masks, would be used for training and testing the segmentation network. Table 1: Summary of the datasets used in the current study. | OAI ZIB | OAI ZIB–UQ | OAI AKOA ---|---|---|--- Number of subjects | 507 | 20 | 24 Timepoints | baseline | baseline | baseline & last follow-up Manual segmentations | FB, FC, TB, TC | FB, FC, TB, TC, PB, PC, FL, TL PL Used for | training $G$ | training $S^{T}$ | testing $G$ & $A$ training & testing (CV) $A$ and $S$ | training & testing (CV) $S^{T}$ FB: femoral bone; FC: femoral cartilage; TB: tibial bone; TC: tibial cartilage; PB: patellar bone; PC: patellar cartilage; FL: femoral lesion; TL: tibial lesion; PL: patellar lesion; CV: cross-validation In this paper, we will refer to our 3D U-Net and 3D CAN modified with deep supervision and anomaly-aware mechanism as $UNET$-$S$ and $CAN$-$S$, respectively. ## 4 Experiments We trained and tested our method on subsets of knee MR images from the publicly available Osteoarthritis Initiative (OAI) database [19]. Three different subsets were used: OAI ZIB, OAI ZIB–UQ, and OAI AKOA (summarized in Table 1). ### 4.1 MR image datasets This study used the publicly available knee MR image dataset OAI ZIB, generated by researchers at Zuse Institute Berlin (ZIB) [2]. The dataset consists of 507 MR examinations from the OAI for which manual reference segmentations of femoral and tibial bones and cartilages were produced by experienced analysts starting from a model-based auto-segmentation. The images are 160 sagittal slices $\times$ 384 $\times$ 384 voxels, and they are all images of the right knee at baseline. The MR imaging sequence is 3D DESS (double-echo steady state) with water excitation. Segmentation labels consist of the background (0), femoral bone (1), femoral cartilage (2), tibial bone (3), and tibial cartilage (4). (Background refers to all voxels that were not labeled specifically.) The dataset covers the full spectrum of osteoarthritis grades. Further details of the dataset can be found in Ambellan et al. [2]. The public OAI ZIB dataset is relatively large, but it has segmentation labels for femoral and tibial bones and cartilages only. As a pilot study, we added some additional segmentation labels to 20 MR examinations from the OAI ZIB dataset with varying osteoarthritis severity. The additional segmentation labels included patellar bone (5) and patellar cartilage (6). In addition, bone marrow lesions (BMLs) or subchondral cysts for each bone—i.e. femoral lesion (7), tibial lesion (8), and patellar lesion (9)—were also segmented. We will refer to these 20 MR examinations as the OAI ZIB–UQ dataset. A new dataset called OAI AKOA with 9 segmentation labels (10 including image background) was also produced. The dataset contains images from the OAI database with “accelerated” knee osteoarthritis, defined as patients who had Kellgren–Lawrence (KL) grade $\leq$ 1 at baseline but progressed to KL grade $\geq$ 2 within the follow-up period. The OAI AKOA dataset consists of 48 MR examinations from 24 patients acquired at 2 timepoints, (baseline and last imaging follow-up at a maximum of 96 months). The segmentations for the OAI ZIB–UQ and OAI AKOA datasets were carried out manually using ITK-SNAP [28] by an analyst (WB) with supervision from another analyst (CE) with expertise in segmentation of the human musculoskeletal system. The manual segmentations in the OAI ZIB–UQ and OAI AKOA datasets as well as the cross-validation data splits (vide infra) are released in https://github.com/wooboyeong/Anomaly-Aware-3D-Segmentation. Ethics approval for collection of human data was obtained by the OAI and the participating clinical sites [19]. ### 4.2 Experimental setup Fig. 4: Transfer learning for further segmentation of knee MR images. Here, $N$ refers to the number of MR examinations while $K$ refers to the number of segmentation classes (including the background) in the dataset. See Table 1 for the list of segmentation classes. Fig. 5: Example outputs from the bone anomaly detection networks $G$ and $A$. Figures (a) and (b) are images from the OAI ZIB dataset, and Figure (c) is an image from the OAI AKOA dataset. The last column shows the error images (color-mapped and overlaid on the input images) highlighting the difference between the input image and the output from $A$. Regions of BMLs (blue arrows) in the femur and patella and part of an osteophyte (yellow arrow) on the femur had high reconstruction errors. Fig. 6: Example segmentation outputs for the femoral and tibial bones (purple) and cartilages (yellow) generated by the individual network implementation with the OAI ZIB dataset. The examples show (a) images with little to no visible bone anomalies where all networks produced good segmentation masks and (b,c) images with visible bone anomalies where segmentation networks tended to fail to produce plausible segmentation masks. The anomaly-aware networks, especially $UNET$-$S$, were better able to correctly segment the images with anomalies. To evaluate the anomaly-aware segmentation method, the segmentation networks $UNET$-$S$ and $CAN$-$S$ were compared with the standard 3D U-Net and 3D CAN without any modification, as well as the models with deep supervision only (without the anomaly-aware mechanism). The loss function for these networks was multi-class Dice loss only and there was no additional input. The network architecture and the training setup were otherwise the same as $UNET$-$S$ and $CAN$-$S$. For convenience, the models without deep supervision will be referred to as $UNET$ and $CAN$, and those with deep supervision will be referred to as $UNET+$ and $CAN+$. See A for implementation and training details. Table 2: Mean DSC, ASD, and HD values for segmentations of the femoral and tibial bone and cartilage volumes from the proposed anomaly-aware method ($UNET$-$S$ and $CAN$-$S$) with their baseline networks, evaluated using 5-fold cross-validation on the OAI ZIB dataset ($N$ = 507). Class | Metric | UNET | UNET+ | UNET-S | CAN | CAN+ | CAN-S ---|---|---|---|---|---|---|--- | DSC (%) | 98.6$\pm$0.33 | 98.7$\pm$0.33 | 98.7$\pm$0.30 | 98.6$\pm$0.37† | 98.6$\pm$0.35† | 98.7$\pm$0.30 | ASD (mm) | 0.23$\pm$0.06 | 0.24$\pm$0.08† | 0.22$\pm$0.05 | 0.25$\pm$0.14†‡ | 0.25$\pm$0.09†‡ | 0.23$\pm$0.07 FB | HD (mm) | 7.05$\pm$6.01†‡ | 9.96$\pm$7.06†‡ | 3.40$\pm$2.67 | 10.07$\pm$7.53†‡ | 11.33$\pm$7.67†‡ | 4.22$\pm$3.96 HD∗ (mm) | 55.54$\pm$35.29†‡ | 24.18$\pm$19.96†‡ | 4.05$\pm$5.94 | 71.78$\pm$26.49†‡ | 28.28$\pm$19.94†‡ | 5.82$\pm$8.36 | HD0 (mm) | 5.97$\pm$5.39†‡ | 9.46$\pm$7.33†‡ | 2.82$\pm$1.63 | 8.98$\pm$7.14†‡ | 10.24$\pm$7.76†‡ | 3.46$\pm$2.71 | HD1 (mm) | 7.74$\pm$6.29†‡ | 10.29$\pm$6.87†‡ | 3.78$\pm$3.10 | 10.76$\pm$7.71†‡ | 12.03$\pm$7.54†‡ | 4.71$\pm$4.52 | DSC (%) | 89.6$\pm$2.79‡ | 89.7$\pm$2.91‡ | 89.5$\pm$2.67 | 89.2$\pm$2.71 | 89.3$\pm$2.69 | 89.0$\pm$2.46 | ASD (mm) | 0.26$\pm$0.07 | 0.26$\pm$0.07‡ | 0.26$\pm$0.07 | 0.27$\pm$0.07† | 0.27$\pm$0.07 | 0.27$\pm$0.07 FC | HD (mm) | 7.44$\pm$4.49†‡ | 6.71$\pm$3.99†‡ | 5.29$\pm$2.39 | 8.46$\pm$4.89†‡ | 7.17$\pm$4.15†‡ | 5.57$\pm$2.63 HD∗ (mm) | 14.38$\pm$14.85†‡ | 14.06$\pm$15.04†‡ | 5.58$\pm$3.69 | 34.82$\pm$30.24†‡ | 16.74$\pm$16.18†‡ | 6.05$\pm$4.31 | HD0 (mm) | 6.51$\pm$4.25†‡ | 5.59$\pm$3.06† | 4.52$\pm$1.60 | 7.48$\pm$4.59†‡ | 5.92$\pm$3.51†‡ | 4.85$\pm$1.95 | HD1 (mm) | 8.04$\pm$4.54†‡ | 7.43$\pm$4.35†‡ | 5.78$\pm$2.68 | 9.09$\pm$4.98†‡ | 7.98$\pm$4.33†‡ | 6.03$\pm$2.89 | DSC (%) | 98.7$\pm$0.35 | 98.7$\pm$0.33 | 98.7$\pm$0.32 | 98.6$\pm$0.35† | 98.6$\pm$0.36† | 98.7$\pm$0.33 | ASD (mm) | 0.22$\pm$0.10 | 0.22$\pm$0.06 | 0.21$\pm$0.06 | 0.23$\pm$0.08† | 0.24$\pm$0.10†‡ | 0.22$\pm$0.08 TB | HD (mm) | 6.29$\pm$5.72†‡ | 7.03$\pm$5.23†‡ | 3.23$\pm$1.86 | 9.26$\pm$7.03†‡ | 10.09$\pm$7.23†‡ | 3.85$\pm$3.16 HD∗ (mm) | 76.04$\pm$45.14†‡ | 27.63$\pm$26.23†‡ | 3.82$\pm$5.63 | 65.28$\pm$37.62†‡ | 37.06$\pm$24.76†‡ | 5.90$\pm$10.58 | HD0 (mm) | 5.59$\pm$4.97†‡ | 5.85$\pm$4.37†‡ | 3.01$\pm$1.82 | 8.44$\pm$6.81†‡ | 9.94$\pm$7.64†‡ | 3.12$\pm$1.94 | HD1 (mm) | 6.73$\pm$6.11†‡ | 7.78$\pm$5.59†‡ | 3.38$\pm$1.87 | 9.78$\pm$7.13†‡ | 10.19$\pm$6.96†‡ | 4.32$\pm$3.67 | DSC (%) | 85.9$\pm$4.21 | 85.8$\pm$4.16 | 86.0$\pm$4.00 | 85.3$\pm$4.11 | 85.2$\pm$4.24† | 85.3$\pm$4.17 | ASD (mm) | 0.27$\pm$0.10 | 0.27$\pm$0.10 | 0.26$\pm$0.09 | 0.28$\pm$0.10† | 0.28$\pm$0.10 | 0.27$\pm$0.11 TC | HD (mm) | 5.83$\pm$3.23†‡ | 5.43$\pm$2.83† | 4.74$\pm$2.03 | 7.00$\pm$4.15†‡ | 6.09$\pm$3.36†‡ | 4.89$\pm$2.20 HD∗ (mm) | 11.47$\pm$16.26†‡ | 8.75$\pm$12.13†‡ | 4.74$\pm$2.03 | 16.38$\pm$17.96†‡ | 14.23$\pm$17.93†‡ | 4.99$\pm$2.81 | HD0 (mm) | 4.70$\pm$2.18 | 4.69$\pm$2.29 | 4.19$\pm$1.40 | 6.01$\pm$3.42†‡ | 5.22$\pm$2.78†‡ | 4.32$\pm$1.52 | HD1 (mm) | 6.56$\pm$3.58†‡ | 5.91$\pm$3.04† | 5.09$\pm$2.27 | 7.63$\pm$4.45†‡ | 6.65$\pm$3.58†‡ | 5.26$\pm$2.48 Bold with underline represents the best value within each metric. Bold without underline represents the second best value. HD∗ refers to Hausdoff distances before post-processing. All other metrics refer to results after post-processing. HD0 refers to HDs (after post-processing) for cases with mild to no osteoarthritis (radiographic grade $\leq$ 2; $N_{0}$ = 198). HD1 refers to HDs (after post-processing) for cases with moderate to severe osteoarthritis (radiographic grade $\geq$ 3; $N_{1}$ = 309). † represents significant difference (p-value $<$ 0.05) compared to UNET-S with Tukey’s HSD test. ‡ represents significant difference (p-value $<$ 0.05) compared to CAN-S with Tukey’s HSD test. The OAI ZIB dataset was randomly and evenly split into 5-fold cross-validation sets, i.e. 102/102/101/101/101 MR examinations in each test set, and the split was maintained over all parts of Components 2 ($A$) and 3 ($S$) in Figure 1, as well as over all different segmentation models being evaluated. ($S$ refers to either $UNET$–$S$ or $CAN$–$S$.) Segmentation performance was evaluated using Dice similarity coefficient (DSC), average surface distance (ASD), and Hausdorff distance (HD) values. See B for the definitions of these evaluation metrics. ### 4.3 Transfer learning for further segmentation As an extension to the above experiments, the segmentation networks were trained further with the expanded number of segmentation labels in the OAI ZIB–UQ and OAI AKOA datasets. Since these datasets are relatively small, the segmentation networks were first trained on the entire OAI ZIB dataset, and then the learned weights were transferred to new networks (Figure 4). The new networks have 10 channels in the output layer instead of 5 since there are now 10 segmentation classes (including image background). The new networks will be denoted with the superscript T. The Dice loss function (see Equation 3) was modified so that higher weights were given to the new classes that the network now has to learn: $\mathcal{L}_{wDSC}=\frac{1}{|K|}\sum_{k\in K}w_{k}\left(1.0-\frac{2\sum_{i}{u_{i,k}v_{i,k}}}{\sum_{i}{u_{i,k}}+\sum_{i}{v_{i,k}}}\right),$ (6) where $w_{k}=1.0$ for the first 5 classes that were already learned, and $w_{k}=10.0$ for the new 5 classes. In addition, since many images do not contain lesions, $w_{k}$ was changed to 0 if $\sum_{i}{u_{i,k}}+\sum_{i}{v_{i,k}}=0.0$; this means the image did not contain label $k$ and the network successfully predicted that there is no label $k$, so the loss in this case would be 0. The loss function for the baseline networks ($UNET^{T}$, $UNET+^{T}$, $CAN^{T}$, $CAN+^{T}$) was this weighted multi-class Dice loss. The focal cross-entropy loss for the anomaly-aware networks (see Equation 4) was the same as before. Therefore, the total loss for $UNET$-$S^{T}$ and $CAN$-$S^{T}$ was: $\mathcal{L}_{S^{T}}=\mathcal{L}_{wDSC}+\alpha\mathcal{L}_{FCE}.$ (7) The new networks were then trained on the entire OAI ZIB–UQ dataset to learn to segment the 10 classes. Lastly, in order to test the pipeline $A\rightarrow S^{T}$, the anomaly information for images in the OAI AKOA dataset was obtained by running the images through the network $A$ that was trained on the OAI ZIB dataset. Since $A$ only requires the original images for reconstruction, it did not need any further training or transfer learning. However, since the segmentation networks only had 20 images (the OAI ZIB–UQ dataset) for learning the new labels, it required some further training, so the OAI AKOA dataset was randomly and evenly split into 5-fold cross- validation sets on a patient basis, i.e. 5$\times$2/5$\times$2/5$\times$2/5$\times$2/4$\times$2 testing images in each set. Due to the limited size of the datasets, online data augmentation was applied to reduce overfitting. Figure 4 shows a summary of the transfer learning process. For all of the networks, weights for the first two convolutional layers were frozen during training (see Figure 10 in the Appendix). See A for implementation and training details. ## 5 Results ### 5.1 Anomaly detection Figure 5 shows example output images from the anomaly detection networks $G$ and $A$. The input MR images of the knee have some visible bone anomalies including BMLs and osteophytes. The network outputs are lossy reconstructions of the input images with the bright signal bone anomalies within the cancellous bone mostly removed from the images. Some of the osteophytes were incompletely reconstructed. The network $A$ only had the original images as inputs, but the outputs from $A$ still have most of the anomalies blurred out. The last column of Figure 5 shows the reconstruction error images from $A$ in which the anomalous regions detected within the cancellous bones are highlighted. Note that the MR image in Figure 5(c) is from the OAI AKOA dataset, which was not used for training either $G$ or $A$, but only the masks of femur and tibia were used to generate input for $G$ since the model was only trained with the OAI ZIB dataset. However, since the masks had been dilated before erasing the images (see Section 3.1), the subchondral patellar lesions were also detected, as can be seen in Figure 5(c). Also note that masks are unnecessary for $A$ in any case. ### 5.2 Anomaly-aware segmentation on OAI ZIB dataset Figure 6 shows example outputs from the anomaly-aware segmentation networks $UNET$-$S$ and $CAN$-$S$ and also the outputs from their baseline networks. As noted in Section 2, the CNNs performed well in terms of DSCs for most of the OAI ZIB images (Figure 6(a)), but failed on some cases with severe abnormalities. The anomaly-aware method was found to be more robust against these difficult cases, resulting in a noticeable improvement in the quality of the segmentation of bone volume for the femur and tibia (Figures 6(b) and 6(c)). The image in Figure 6(b) was the most difficult from the OAI ZIB dataset for the segmentation CNNs due to the presence of a large femoral BML, but the segmentation error was fixed with $UNET$-$S$ and partially fixed with $CAN$-$S$. The image in Figure 6(c) also has a notable femoral BML, but both $UNET$-$S$ and $CAN$-$S$ were able to correctly segment the femoral bone. Table 2 shows the quantitative results for the segmentation task. Tukey’s honestly significant difference (HSD) test was used to compare all possible pairs of means for each metric; this test is is similar to the t-test, except it corrects for family-wise error rate. There was not a significant improvement in the mean DSCs, but HDs were substantially reduced for all of the segmentation classes for both $UNET$-$S$ and $CAN$-$S$ compared to their baselines. The HDs for cases with moderate to severe osteoarthritis (HD1 in Table 2) were higher than the HDs for cases with mild to no osteoarthritis (HD0) for all models, but the HD1 was still relatively low with the anomaly- aware models compared to the other models. Note that the DSCs, ASDs, and HDs were calculated after post-processing the CNN outputs. The post-processing was performed because most of the CNN outputs for non-anomaly-aware models contained random stray voxels which resulted in extremely high HDs (see HD∗ in Table 2). As mentioned in Section 2, this is one of the main limitations of CNN-based segmentations. The random stray voxels were removed by extracting the largest components and removing smaller objects that are distant from the main structures. See C for description of the post-processing method. The mean HDs were significantly reduced after post-processing for the non-anomaly-aware models. The CNN outputs from the anomaly-aware models did not contain as many stray voxels to start with, so the effect of post-processing was much less obvious. Even after the post-processing, the mean HDs for the anomaly-aware models were significantly less than the mean HDs for the non-anomaly-aware models. See also the boxplots in Figure 7 to assess the distributions of HDs (calculated after post-processing). It can be seen that the HDs for the anomaly-aware method are much smaller overall. Fig. 7: Boxplots of Hausdorff distance (HD) values for the proposed anomaly- aware segmentation approach ($UNET$-$S$ and $CAN$-$S$) and baseline networks, evaluated on the OAI ZIB dataset using 5-fold cross-validation. Note that these HDs are results after post-processing. ### 5.3 Anomaly-aware segmentation on OAI AKOA dataset with transfer learning Fig. 8: Example outputs from the segmentation networks for images from the OAI AKOA dataset, segmenting the patella and visible bone lesions in addition to the femur and tibia. Note that the images are zoomed in to view the patella more closely. The masks are overlaid on the input images (purple: bones; yellow: cartilages; red: bone lesions). (a) When the images had no visible anomalies, all networks except $UNET^{T}$ produced good segmentation of the patella. The network $UNET^{T}$ failed to converge for the patellar cartilage label. (b,c) The anomaly-aware networks were able to detect and segment most of the visible lesions along with the anatomical structures on images that the other segmentation networks had difficulty with. Figure 8 shows example outputs from the anomaly-aware segmentation networks $UNET$-$S^{T}$ and $CAN$-$S^{T}$ for the OAI AKOA test sets and also the outputs from their baseline networks with transfer learning. As with Section 5.2, images with noticeable abnormalities tended to be more difficult to segment. For images with no visible anomalies, all networks except $UNET^{T}$ produced good segmentation of the patella (Figure 8(a)). The network $UNET^{T}$ did not converge for the patellar cartilage label for any of the training sets in 5-fold cross-validation. For images with visible abnormalities, $UNET$-$S^{T}$ and $CAN$-$S^{T}$ correctly detected most of the visible lesions and produced acceptable segmentations while the other networks occasionally failed (Figures 8(b) and 8(c)). Table 3 shows the quantitative results for segmentation of the bones and cartilages. Again, there were statistically significant improvements in HDs for the anomaly-aware method compared to the baseline. In this case, the DSCs of the femoral bone and tibial bone volumes were also significantly improved. Note that the same post-processing method as in Section 5.2 was applied here as well. The boxplots in Figure 9 shows the distributions of HDs (calculated after post-processing). It can be observed that the HDs for the anomaly-aware method are much smaller than their baseline networks. Table 3: Mean DSC, ASD, and HD values for segmentation of the femoral, tibial, and patellar bone and cartilage volumes from the proposed anomaly-aware method ($UNET$-$S^{T}$ and $CAN$-$S^{T}$) with their baseline networks with transfer learning, evaluated using 5-fold cross-validation on the OAI AKOA dataset ($N$ = 24$\times$2). Class | Metric | UNETT | UNET+T | UNET-ST | CANT | CAN+T | CAN-ST ---|---|---|---|---|---|---|--- | DSC (%) | 96.9$\pm$1.63†‡ | 97.3$\pm$1.04†‡ | 98.4$\pm$0.41 | 96.5$\pm$1.75†‡ | 97.0$\pm$1.30†‡ | 98.4$\pm$0.55 FB | ASD (mm) | 0.55$\pm$0.39†‡ | 0.47$\pm$0.26 | 0.24$\pm$0.08 | 0.97$\pm$0.91†‡ | 0.54$\pm$0.36†‡ | 0.26$\pm$0.12 HD (mm) | 16.36$\pm$9.49† | 18.21$\pm$6.56†‡ | 8.96$\pm$6.08 | 41.78$\pm$22.94†‡ | 15.84$\pm$8.06† | 10.24$\pm$7.62 | HD∗ (mm) | 41.16$\pm$24.56†‡ | 34.00$\pm$15.91†‡ | 10.70$\pm$9.96 | 87.78$\pm$5.74†‡ | 30.71$\pm$21.26†‡ | 16.28$\pm$16.02 | DSC (%) | 85.3$\pm$3.13 | 85.9$\pm$2.67 | 86.6$\pm$2.48 | 84.5$\pm$3.10†‡ | 85.0$\pm$3.22 | 86.4$\pm$2.53 FC | ASD (mm) | 0.32$\pm$0.07†‡ | 0.29$\pm$0.05 | 0.26$\pm$0.04 | 0.34$\pm$0.10†‡ | 0.31$\pm$0.06†‡ | 0.26$\pm$0.04 HD (mm) | 19.75$\pm$5.58†‡ | 9.43$\pm$5.17†‡ | 5.73$\pm$2.25 | 7.64$\pm$3.66 | 10.41$\pm$4.64†‡ | 6.09$\pm$3.96 | HD∗ (mm) | 29.98$\pm$16.98†‡ | 14.07$\pm$10.51† | 5.73$\pm$2.25 | 15.99$\pm$17.43†‡ | 20.52$\pm$15.33†‡ | 7.22$\pm$6.70 | DSC (%) | 97.4$\pm$0.93†‡ | 97.6$\pm$0.73†‡ | 98.4$\pm$0.31 | 97.4$\pm$1.04†‡ | 97.3$\pm$1.40†‡ | 98.3$\pm$0.31 TB | ASD (mm) | 0.51$\pm$0.27†‡ | 0.37$\pm$0.14 | 0.24$\pm$0.10 | 0.42$\pm$0.20†‡ | 0.46$\pm$0.58†‡ | 0.24$\pm$0.08 HD (mm) | 26.50$\pm$17.89†‡ | 12.55$\pm$7.12 | 6.98$\pm$7.13 | 16.61$\pm$8.97†‡ | 9.64$\pm$6.48 | 7.06$\pm$5.42 | HD∗ (mm) | 107.47$\pm$6.61†‡ | 46.55$\pm$23.83†‡ | 7.96$\pm$7.96 | 99.87$\pm$21.99†‡ | 39.56$\pm$25.70†‡ | 15.94$\pm$17.27 | DSC (%) | 84.5$\pm$4.24 | 84.7$\pm$3.66 | 85.0$\pm$3.16 | 83.2$\pm$4.55 | 83.6$\pm$4.50 | 84.7$\pm$3.44 TC | ASD (mm) | 0.31$\pm$0.14 | 0.30$\pm$0.11 | 0.28$\pm$0.11 | 0.43$\pm$0.74 | 0.32$\pm$0.13 | 0.29$\pm$0.12 HD (mm) | 11.08$\pm$7.35†‡ | 6.07$\pm$3.47 | 4.84$\pm$2.53 | 13.59$\pm$6.91†‡ | 5.66$\pm$2.97 | 5.19$\pm$2.36 | HD∗ (mm) | 30.21$\pm$23.06†‡ | 19.61$\pm$18.21†‡ | 4.84$\pm$2.53 | 36.27$\pm$21.11†‡ | 12.97$\pm$17.77 | 9.19$\pm$13.59 | DSC (%) | 96.0$\pm$1.29 | 96.2$\pm$0.86 | 96.6$\pm$0.80 | 96.0$\pm$1.48 | 95.8$\pm$1.29† | 96.3$\pm$1.02 PB | ASD (mm) | 0.33$\pm$0.12 | 0.30$\pm$0.07 | 0.26$\pm$0.07 | 0.41$\pm$0.59 | 0.42$\pm$0.31 | 0.32$\pm$0.18 HD (mm) | 10.22$\pm$9.78† | 8.27$\pm$5.86† | 3.69$\pm$1.97 | 7.64$\pm$7.17 | 11.86$\pm$8.50†‡ | 7.43$\pm$6.69 | HD∗ (mm) | 78.45$\pm$20.18†‡ | 75.03$\pm$19.35†‡ | 10.33$\pm$21.00‡ | 94.30$\pm$27.55†‡ | 72.00$\pm$26.59†‡ | 29.04$\pm$34.58† | DSC (%) | 0.0$\pm$0.00§ | 85.1$\pm$5.48 | 85.7$\pm$5.02 | 83.9$\pm$6.02 | 84.5$\pm$5.43 | 84.5$\pm$6.15 PC | ASD (mm) | N/A§ | 0.29$\pm$0.11 | 0.27$\pm$0.08 | 0.33$\pm$0.13 | 0.30$\pm$0.12 | 0.31$\pm$0.12 HD (mm) | N/A§ | 4.20$\pm$3.90 | 3.49$\pm$2.36 | 4.93$\pm$3.99 | 5.13$\pm$4.44 | 4.29$\pm$3.70 | HD∗ (mm) | N/A§ | 24.23$\pm$30.61†‡ | 7.54$\pm$16.23 | 38.96$\pm$36.00†‡ | 28.89$\pm$33.77†‡ | 5.98$\pm$12.17 Bold with underline represents the best value within each metric. Bold without underline represents the second best value. HD∗ refers to Hausdoff distances before post-processing. All other metrics refer to results after post-processing. † represents significant difference (p-value $<$ 0.05) compared to UNET-ST with Tukey’s HSD test. ‡ represents significant difference (p-value $<$ 0.05) compared to CAN-ST with Tukey’s HSD test. § UNETT failed to converge, so it was excluded from Tukey’s HSD test (for the PC label only). Fig. 9: Boxplots of Hausdorff distance (HD) values for the proposed anomaly- aware segmentation approach ($UNET$-$S^{T}$ and $CAN$-$S^{T}$) and baseline networks with transfer learning, evaluated on the OAI AKOA dataset using 5-fold cross-validation. These HDs are results after post-processing. Note also that $UNET^{T}$ failed to converge for the PC label, so it was excluded from the plot. Performance for segmentation of bone lesions was less straightforward to evaluate because (a) there are a variable number of lesions (some images have none while some images have many) and (b) some lesions are too small to be reliably detected. In this study, we assumed “bone-wise” lesion detection for simplicity, in which each bone was classified as either a positive or a negative case: * 1. For positive cases (bone has a lesion), a prediction was considered to be “true positive” if the segmentation DSC $\geq$ 5% and “false negative” if DSC $<$ 5%. * 2. For negative cases (bone has no lesion), a prediction was considered to be “true negative” if the network successfully predicted that there is no lesion in the bone and “false positive” if the network predicted that there is one. Note there are 3 bones (i.e. 3 “cases”) in each image: femur, tibia, and patella. In the 48 images from the OAI AKOA dataset, there were 76 positive cases and 68 negative cases in total. Using these criteria, sensitivity (true positive rate; TPR) was $>$ 90% for most models while specificity (true negative rate; TNR) was $<$ 50% (Table 4). The specificity is low mainly due to the presence of subtle lesions in many images. See Figure 11 in the Appendix showing example images with very small lesions that often resulted in “false positives” or “false negatives”. In most cases, the neural networks were highly sensitive and detected those small lesions, resulting in high false positive rates. However, such lesions are not consistently detected by human observers either and often are not clinically important. Therefore, we applied various size thresholds to the lesion masks to see if the specificity is higher for larger lesions. Using progressively increasing thresholds up to 6.0 mm3, isolated lesions smaller than the threshold were removed from the output masks, and the same post-processing was applied to the manual masks for consistency. See C for details on the post- processing for bone lesion detection. Table 4 shows the bone lesion detection and segmentation performance of the different models in terms of accuracy and mean DSC. By progressively increasing the threshold, sensitivity was decreased but specificity was increased. The highest accuracy achieved with the post-processing was 84.7% using $CAN$-$S^{T}$. The highest mean DSC (segmentation performance for positive cases only) was 53.6% using $UNET$-$S^{T}$. The area under the receiver operating characteristic curve (AUC) was 0.896 for $CAN$-$S^{T}$ and 0.892 for $UNET$-$S^{T}$, both of which were higher than their baselines. Note in Table 4 that $UNET^{T}$ did not converge for the patellar lesion label. Its specificity is paradoxically high because it did not detect any patellar lesion, but the accuracy is still low due to low sensitivity. Another notable finding is that while $UNET$-$S^{T}$ performed the best in the segmentation tasks overall (Tables 2 and 3), $CAN$-$S^{T}$ performed the best in the bone lesion detection task (Table 4). ## 6 Discussion ### 6.1 Anomaly detection The current study presented a method to apply 3D U-Net-based CNNs for visual anomaly detection in volumetric medical images. There were two steps to this end, involving two separate networks $G$ and $A$. Network $G$ (the first step) could be sufficient if you already have the appropriate segmentation masks and you are only interested in highlighting anomalies in the images. However, segmentation masks are initially unavailable in most cases, so network $A$ (the second step) was added to carry out the same task without having to obtain segmentation masks first. The last column of Figure 5 shows that the anomaly detection network $A$ can visually highlight anomalies on MR images of the knee. These autoencoder-based networks also reconstruct the images with most of the visible lesions removed and therefore, they can theoretically be used to show what the images likely looked like if the bones had no lesions. The main limitation for this is that the quality of the images reconstructed from the current networks are rather poor. This is complex because it is actually easy for convolutional autoencoders to reconstruct images with minimal reconstruction error, but for anomaly detection, the reconstruction needs to be lossy since the anomalies should be removed, which means we might need to compromise on image quality. Nevertheless, a useful future work would include improving the model to make the images look more anatomically realistic. A possible approach would be to modify the image compressor, for example by using a more sophisticated model such as a conditional encoder or vector quantization [18]. ### 6.2 Anomaly-aware segmentation This study also presented a generalized method for CNNs in which the information from the anomaly detectors is utilized to improve image segmentation. Although the main purpose was to compare the segmentation performance of the anomaly-aware networks to that of the baseline networks, there were also some incidental findings from the experiments such as the effect of deep supervision and the difference between U-Net and CAN. First of all, it was demonstrated that the anomaly-aware mechanism is capable of improving the segmentation of overall bone volume on MR images of osteoarthritic knees with visible anomalies. The networks $UNET$ and $CAN$ were already capable of achieving high DSCs, but it was difficult to achieve good HDs without any shape regularization or post-processing. While deep supervision ($UNET+$ and $CAN+$) helped improve convergence and overall performance, it was found that the anomaly-aware method ($UNET$-$S$ and $CAN$-$S$) was able to further improve these for both U-Net and CAN. The anomaly-aware attention mechanism provided a substantial improvement in HDs in addition to a visible improvement in the quality of segmentation (Figures 6–9). Although none of the CNNs were completely error-proof, as can be seen with the few outliers in HDs (Figures 7 and 9), the significant drop in mean HDs with the anomaly-aware models indicates that these models make errors much less frequently. This is probably because the additional information provided to the network guided the network to focus its attention to the ROI, and the model was less vulnerable to anomalies and noise in the images. The differences between HD∗s before post-processing and HDs after post- processing (Tables 2 and 3) are also notable. The post-processing decreased HDs for most images for the non-anomaly-aware models, whereas it had an effect only for some images for the anomaly-aware models. Indeed, the post-processing made no difference in HDs for 94% of all images (524/555) for $UNET$-$S$ and for 84% (464/555) for $CAN$-$S$. In contrast, the post-processing made no difference in HDs for no images (0/555) for $UNET$ and $CAN$ and for 20% (110/555) and 10% (58/555), respectively, for $UNET+$ and $CAN+$. The inference time for the CNNs (without post-processing) was about 1 second per image, with the post-processing taking additional 4 seconds per image. An additional finding from Table 2 is that the U-Net-based models performed slightly better than the CAN-based models in the segmentation of femur and tibia on the OAI ZIB dataset. However, $UNET^{T}$ failed to learn the patellar cartilage and patellar lesion labels on the OAI AKOA dataset (Tables 3 and 4) despite using weighted Dice loss. Some possible reasons for the difficulty in learning these labels include (a) not enough training images, (b) highly imbalanced classes, (c) spatial heterogeneity, and (d) larger proportional impact of BMLs in the patella. An interesting observation is that the other models were still able to converge for all segmentation classes. Since the only difference between $UNET^{T}$ and $UNET+^{T}$ was deep supervision, it can be inferred that the deep supervision helped stabilize convergence. The more stable convergence behavior in $CAN^{T}$ compared to $UNET^{T}$ was initially unexpected but explainable; two possible explanations are as follows: 1. 1. The CAN-based models are relatively “simple”, having a much less number of parameters than the U-Net-based models. For example, $CAN$ has 852,565 learnable parameters while $UNET$ has 5,887,765 learnable parameters. Having many parameters is not necessarily helpful because it can increase the chance of overfitting and the model might converge slower due to the model complexity. 2. 2. Since the CAN-based models used minimal downsampling, they might have been better than $UNET^{T}$ at detecting very small structures such as the patellar cartilage and lesions. The U-Net-based models, on the other hand, rely on downsampling for aggregating multi-scale contextual information, which can be advantageous when segmenting larger anatomical structures but less desirable when having to analyze smaller structures. This is also consistent with the finding that although $UNET$-$S^{T}$ performed the best as a segmentation method, $CAN$-$S^{T}$ gave the best performance when used as a bone lesion detection method (Table 4). Indeed, all CAN-based methods performed better than the U-Net-based methods on average in terms of accuracy and AUC. Table 4: Bone lesion detection and segmentation performance on the OAI AKOA dataset in terms of accuracy and mean DSC. Here, Acc. refers to the accuracy with no post-processing while $\lceil$Acc.$\rceil$ refers to the highest accuracy achieved with post-processing. Both are reported with the corresponding sensitivity (TPR) and specificity (TNR). $\lceil$DSC$\rceil$ is the highest mean DSC (averaged over all bone lesions) achieved with post-processing. AUC is the area under the receiver operating characteristic (ROC) curve. Note that $UNET^{T}$ failed to converge for the patellar lesion label. Results for each bone can be found in the supplementary material. Model | Acc. (TPR, TNR) | $\lceil$Acc.$\rceil$ (TPR, TNR) | $\lceil$DSC$\rceil$ | AUC ---|---|---|---|--- UNETT | 0.597 (0.632, 0.559) | 0.729 (0.592, 0.882) | 0.349 | 0.713 UNET+T | 0.563 (0.921, 0.162) | 0.764 (0.750, 0.779) | 0.501 | 0.795 UNET-ST (Ours) | 0.681 (0.947, 0.382) | 0.819 (0.803, 0.838) | 0.536 | 0.892 CANT | 0.681 (0.921, 0.412) | 0.833 (0.750, 0.926) | 0.462 | 0.871 CAN+T | 0.611 (0.934, 0.250) | 0.806 (0.697, 0.926) | 0.464 | 0.874 CAN-ST (Ours) | 0.694 (0.921, 0.441) | 0.847 (0.855, 0.838) | 0.493 | 0.896 Bold with underline represents the best value within each metric. Bold without underline represents the second best value. Lastly, it can also be noted that the models with deep supervision ($UNET+^{T}$ and $CAN+^{T}$) did not outperform those without deep supervision ($UNET^{T}$ and $CAN^{T}$) in terms of detection accuracy; the models with deep supervision had higher sensitivity but also lower specificity, so the overall accuracy was not improved (Table 4). The benefit of the anomaly-aware method is more apparent since the two models ($UNET$-$S^{T}$ and $CAN$-$S^{T}$) performed better in terms of accuracy, mean DSC, as well as AUC when compared to their baselines. The anomaly-aware models were able to maintain both high sensitivity and specificity. This is likely because these models already had information about where the lesions are likely to be whereas the baseline networks had to learn the information from scratch. According to these findings, our proposed method is expected to be helpful in the analysis of medical images with visible anomalies, but the main limitation is that the current pipeline is rather involved. In Figure 1, it can be noted that the models $A$ and $S$ actually form a linear pipeline. Therefore, future work could investigate combining $A$ and $S$ into a single multi-task model to perform anomaly detection and segmentation simultaneously. In addition, since we can detect and segment pathologies, a future work may also include another downstream task such as classification of osteoarthritis grades. Osteophytes are also a major feature of osteoarthritis, but detection of osteophytes was not included in the current work because manual segmentation of osteophytes was challenging. A method to evaluate osteophyte detection may help with the development of the automated classification of osteoarthritis grades. Finally, since the anomaly-aware approach is a generalized method for CNNs, one could look into combining it with another CNN-based method such as nnU-Net [9] to further enhance the model. ## 7 Conclusion In summary, this work demonstrated how simple U-Net-like neural networks can be used for detecting bone lesions in knee MR images through reconstruction via inpainting. Moreover, it showed how the detected anomalies can be further utilized for downstream tasks such as segmentation. The anomaly-aware networks gave a better performance on average than their baseline networks in the segmentation tasks as well as in the detection of bone lesions. The stable convergence behavior and performance with the new labels in the OAI ZIB–UQ and OAI AKOA datasets are promising and suggest that the proposed method has an advantage when there are relatively few training images and/or the classes are highly imbalanced. It is hoped that future works will show additional improvements and further applications of the anomaly detection and anomaly- aware segmentation models in medical imaging. ## Acknowledgments We would like to thank Dr Jessica Bugeja for curating the OAI AKOA dataset. ## Appendix A Implementation details All of the neural networks in this study were implemented using Tensorflow [1] version 2.4 with Keras API (http://tensorflow.org/guide/keras) and were trained on a high-performance computer with NVIDIA Tesla V100-SXM2-32GB. The two anomaly detection networks $G$ and $A$ (Sections 3.1 and 3.2; Figure 2) were both based on 3D U-Net [5], consisting of a contracting path (encoder) and an expansive path (decoder) with skip connections. Overall, the networks had 5 levels of resolution with 4 progressive downsamplings followed by 4 progressive upsamplings using strided convolutions. In the contracting path, there were two convolution blocks at each level, where each convolution block was a 3D convolution layer with a kernel size of 3 $\times$ 3 $\times$ 3 followed by an instance normalization [25] and leaky rectified linear unit (ReLU) activation function with a negative slope coefficient of 0.1. Instance normalization was used because the large image size only allowed for a batch size of 1. The number of feature maps was progressively increased as the resolution was decreased. (a) (b) Fig. 10: The anomaly-aware segmentation network $S^{T}$ for transfer learning based on (a) 3D U-Net and (b) 3D CAN. The network $S^{T}$ is a slight modification from $S$ (Figure 3) where 5 more channels were added to the output layer. During training, the first two convolution blocks were frozen. The image compressor $C$ was a very small network, again with 4 progressive downsamplings, but with the number of feature maps progressively decreasing as the resolution was decreased. To create a bottleneck, the downsampled image was flattened, and then a dense layer with 100 units was applied. Then, after another dense layer with 5760 units, the flattened image was reshaped into a 3D image of 10 $\times$ 24 $\times$ 24 to be added to the decoder part of the U-Net. In the expansive path of U-Net, the resolution was progressively recovered using transposed convolutions, and the feature maps from the encoder at the corresponding level were transferred using element-wise summation. The combined feature maps then passed through two more convolution blocks before another transposed convolution for upsampling. After reaching the original image resolution, the final output convolution layer with a kernel size of 1 $\times$ 1 $\times$ 1 and sigmoid activation function was applied. The MR images were Z-normalized (centered to 0 mean, unit standard deviation), clipped at [-5, 5], and subsequently rescaled to [0, 1] for network input. The models were trained using the Adam optimizer [13] with a learning rate of 0.0005 for 50 epochs for each training set. For the downstream segmentation task (Section 3.3; Figure 3), two different types of CNNs were used: 3D U-Net and 3D CAN. The architecture of the U-Net- based models ($UNET$, $UNET+$ and $UNET$-$S$) was similar to $G$ and $A$, except that the final activation function was softmax and that $UNET+$ and $UNET$-$S$ were modified with deep supervision. The CAN-based models ($CAN$, $CAN+$ and $CAN$-$S$) had 3 levels of resolution, where the first two levels were basically the same as those of U-Net. After two downsamplings, the CAN module was applied instead of further downsampling. The CAN module consisted of convolution blocks with progressively increasing dilation rates and then a non-dilated convolution block as the final block of the CAN module. A convolution block was almost the same as that defined above for U-Net except its dilation rates and kernel initializer. While non-dilated convolution layers were initialized using the Glorot uniform initializer [8], dilated convolutions were initialized with the identity initializer which was found to be more effective for context aggregation [27]. After the CAN module, the resolution was recovered with transposed convolutions and skip connections as with U-Net, and then the final output layer was applied. For the segmentation networks, the “mixed precision” policy in Keras API was used to enable the use of a larger number of feature maps in 3D CNNs. Mixed precision refers the use of 16-bit floating-point type in parts of the model during training to make it use less memory. The output layer (softmax layer) was kept in the 32-bit type for numerical stability. The MR images were Z-normalized for network input, and the models were trained using the Adam optimizer with a learning rate of 0.0005 for 50 epochs for each training set. A few modifications were made to all segmentation networks for transfer learning (Section 4.3; Figure 10). Firstly, 5 more channels were added to the output layer to segment the additional classes. In addition, since the initial layers in CNNs tend to extract general image features, weights for the first two convolution blocks were frozen during training, and the rest of the network was trained with a slightly lower learning rate of 0.00025 for 50 epochs for each training set. Due to the limited size of the datasets, online data augmentation was applied to reduce overfitting. The data augmentation consisted of random scaling (scaling factor within [0.9, 1.1]), random rotation (angle within [-10, 10] degrees in a random direction), and random translation (within [-10, 10] pixels in each direction). ## Appendix B Evaluation metrics Segmentation performance was evaluated using Dice similarity coefficient (DSC), which is defined as: $DSC=\frac{2|B\cap A|}{|B|+|A|}.$ (8) Here, $A$ and $B$ denote the set of positive voxels in the ground truth segmentation map and the predicted segmentation map, respectively. DSC is a volumetric measure usually expressed as a percentage. Although it is useful for general assessment of the overall segmentation results, it provides limited sensitivity to errors on the boundaries of the segmentation if the segmented volume is large. We therefore also included surface distance measures for the anatomical structures, to evaluate segmentation errors on their boundaries. The average surface distance (ASD) and Hausdorff distance (HD; also known as maximum surface distance) are defined as: $ASD=\frac{1}{n_{\partial(A)}+n_{\partial(B)}}\left(\sum_{a\in\partial(A)}d(a,B)+\sum_{b\in\partial(B)}d(b,A)\right),$ (9) $HD=\max\left(\max_{a\in\partial(A)}d(a,B),\max_{b\in\partial(B)}d(b,A)\right).$ (10) Here, $\partial(\cdot)$ denotes the boundary of the segmentation set, $n_{\partial(\cdot)}$ denotes the number of voxels on the boundary $\partial(\cdot)$, and $d(p,Q)=\min_{q\in\partial(Q)}||p-q||_{2}$ (11) is the minimum distance from point $p$ to the boundary of a segmentation set $Q$. ## Appendix C Post-processing of CNN outputs For the bone and cartilage labels, random stray voxels in the CNN outputs were removed by extracting the largest component for each label and removing smaller components that are more than 50 voxels away from the largest component. Simply extracting the largest component only (i.e. removing all smaller components) was invalid because sometimes that also removed correctly segmented voxels. For example, in some images with very severe cartilage degeneration, there might be smaller fragments of cartilages, so removing all smaller components sometimes makes the segmentation worse. Therefore, some allowance had to be made for the stray voxels that are relatively close to the main structures even though that means not all stray voxels could be removed. The post-processing still removed most of the very extremely stray voxels, significantly reducing mean HDs for the non-anomaly-aware models. For the bone lesion labels, a different post-processing method was used since there are a variable number of lesions and many of the lesions are quite small in size. Figure 11 shows example images with very small lesions that often resulted in “false positives” or “false negatives”, decreasing the specificity of the segmentation networks for the bone lesion detection task (Section 5.3). To see if the specificity is higher for larger lesions, various size thresholds were applied to the lesion masks. Using progressively increasing size thresholds of 0.0 (no post-processing), 0.5, 1.0, …, 6.0 mm3, isolated lesions smaller than the threshold were removed from the output masks, with the same post-processing applied to the manual masks for consistency. Fig. 11: Example of (a) “false positive” and (b) “false negative” case, which mostly occurred with very small lesions. The orange boxes highlight the small lesions. For $>$ 6.0 mm3 size threshold, some positive cases changed to negative cases, paradoxically increasing the sensitivity of the detection method. (It means all positive cases had an isolated lesion at least 6.0 mm3.) At 6.0 mm3 threshold, the specificities of the models ranged from 68–88%. In order to complete the receiver operating characteristic (ROC) curves, the size threshold was fixed at 6.0 mm3, and then the specificity of the detection method was increased by increasing the softmax output threshold. The voxels with softmax output (probability) less than the threshold were removed from the output masks, using thresholds of 0.5, 0.6, 0.7, …, 1.0. The areas under the ROC curves (AUCs) along with the highest accuracies and the highest mean DSCs achieved with post-processing are reported in Table 4. ## References * Abadi et al. [2015] Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Corrado, G.S., Davis, A., Dean, J., Devin, M., Ghemawat, S., Goodfellow, I., Harp, A., Irving, G., Isard, M., Jia, Y., Jozefowicz, R., Kaiser, L., Kudlur, M., Levenberg, J., Mané, D., Monga, R., Moore, S., Murray, D., Olah, C., Schuster, M., Shlens, J., Steiner, B., Sutskever, I., Talwar, K., Tucker, P., Vanhoucke, V., Vasudevan, V., Viégas, F., Vinyals, O., Warden, P., Wattenberg, M., Wicke, M., Yu, Y., Zheng, X., 2015. TensorFlow: Large-scale machine learning on heterogeneous systems. Software available from tensorflow.org. * Ambellan et al. [2019] Ambellan, F., Tack, A., Ehlke, M., Zachow, S., 2019\. Automated segmentation of knee bone and cartilage combining statistical shape knowledge and convolutional neural networks: Data from the osteoarthritis initiative. Med. Image Anal. 52, 109–118. doi:https://doi.org/10.1016/j.media.2018.11.009. * Baur et al. [2018] Baur, C., Wiestler, B., Albarqouni, S., Navab, N., 2018\. Deep autoencoding models for unsupervised anomaly segmentation in brain MR images, in: International MICCAI Brainlesion Workshop, Springer. pp. 161–169. doi:10.1007/978-3-030-11723-8_16. * Chen et al. [2020] Chen, X., You, S., Tezcan, K.C., Konukoglu, E., 2020. Unsupervised lesion detection via image restoration with a normative prior. Med. Image Anal. 64, 101713\. doi:https://doi.org/10.1016/j.media.2020.101713. * Çiçek et al. [2016] Çiçek, Ö., Abdulkadir, A., Lienkamp, S.S., Brox, T., Ronneberger, O., 2016. 3D U-Net: Learning dense volumetric segmentation from sparse annotation, in: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer. pp. 424–432. doi:10.1007/978-3-319-46723-8_49. * Dai et al. [2022] Dai, W., Woo, B., Liu, S., Marques, M., Engstrom, C., Greer, P.B., Crozier, S., Dowling, J.A., Chandra, S.S., 2022\. CAN3D: Fast 3D medical image segmentation via compact context aggregation. Med. Image Anal. 82, 102562\. doi:https://doi.org/10.1016/j.media.2022.102562. * Ebrahimkhani et al. [2020] Ebrahimkhani, S., Jaward, M.H., Cicuttini, F.M., Dharmaratne, A., Wang, Y., de Herrera, A.G.S., 2020\. A review on segmentation of knee articular cartilage: from conventional methods towards deep learning. Artif. Intell. Med. 106, 101851\. doi:https://doi.org/10.1016/j.artmed.2020.101851. * Glorot and Bengio [2010] Glorot, X., Bengio, Y., 2010\. Understanding the difficulty of training deep feedforward neural networks, in: Teh, Y.W., Titterington, M. (Eds.), Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, PMLR, Chia Laguna Resort, Sardinia, Italy. pp. 249–256. * Isensee et al. [2021] Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H., 2021. nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18, 203–211. doi:10.1038/s41592-020-01008-z. * Isensee et al. [2017] Isensee, F., Kickingereder, P., Wick, W., Bendszus, M., Maier-Hein, K.H., 2017. Brain tumor segmentation and radiomics survival prediction: Contribution to the BRATS 2017 challenge, in: International MICCAI Brainlesion Workshop, Springer. pp. 287–297. doi:10.1007/978-3-319-75238-9_25. * Isola et al. [2017] Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A., 2017\. Image-to-image translation with conditional adversarial networks, in: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5967–5976. doi:10.1109/CVPR.2017.632. * Kayalibay et al. [2017] Kayalibay, B., Jensen, G., van der Smagt, P., 2017. CNN-based segmentation of medical imaging data. arXiv preprint arXiv:1701.03056 arXiv:1701.03056. * Kingma and Ba [2014] Kingma, D.P., Ba, J.L., 2014\. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 arXiv:1412.6980. * Lee et al. [2015] Lee, C.Y., Xie, S., Gallagher, P., Zhang, Z., Tu, Z., 2015. Deeply-Supervised Nets, in: Lebanon, G., Vishwanathan, S.V.N. (Eds.), Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics, PMLR, San Diego, California, USA. pp. 562--570. * Liu et al. [2020] Liu, X., Xing, F., Yang, C., Kuo, C.C.J., Fakhri, G.E., Woo, J., 2020. Symmetric-constrained irregular structure inpainting for brain MRI registration with tumor pathology, in: International MICCAI Brainlesion Workshop, Springer. pp. 80--91. doi:10.1007/978-3-030-72084-1_8. * Milletari et al. [2016] Milletari, F., Navab, N., Ahmadi, S.A., 2016. V-Net: Fully convolutional neural networks for volumetric medical image segmentation, in: 2016 Fourth International Conference on 3D Vision (3DV), pp. 565--571. doi:10.1109/3DV.2016.79. * Nie and Shen [2020] Nie, D., Shen, D., 2020. Adversarial confidence learning for medical image segmentation and synthesis. Int. J. Comput. Vis. 128, 2494--2513. doi:10.1007/s11263-020-01321-2. * van den Oord et al. [2017] van den Oord, A., Vinyals, O., Kavukcuoglu, K., 2017. Neural discrete representation learning, in: Proceedings of the 31st International Conference on Neural Information Processing Systems, Curran Associates Inc., Red Hook, NY, USA. p. 6309–6318. * Peterfy et al. [2008] Peterfy, C., Schneider, E., Nevitt, M., 2008. The osteoarthritis initiative: report on the design rationale for the magnetic resonance imaging protocol for the knee. Osteoarthr. Cartil. 16, 1433--1441. doi:https://doi.org/10.1016/j.joca.2008.06.016. * Pinaya et al. [2022] Pinaya, W.H., Tudosiu, P.D., Gray, R., Rees, G., Nachev, P., Ourselin, S., Cardoso, M.J., 2022. Unsupervised brain imaging 3D anomaly detection and segmentation with transformers. Med. Image Anal. 79, 102475\. doi:https://doi.org/10.1016/j.media.2022.102475. * Prasoon et al. [2013] Prasoon, A., Petersen, K., Igel, C., Lauze, F., Dam, E., Nielsen, M., 2013\. Deep feature learning for knee cartilage segmentation using a triplanar convolutional neural network, in: Mori, K., Sakuma, I., Sato, Y., Barillot, C., Navab, N. (Eds.), Medical Image Computing and Computer-Assisted Intervention -- MICCAI 2013, Springer Berlin Heidelberg, Berlin, Heidelberg. pp. 246--253. * Raj et al. [2018] Raj, A., Vishwanathan, S., Ajani, B., Krishnan, K., Agarwal, H., 2018. Automatic knee cartilage segmentation using fully volumetric convolutional neural networks for evaluation of osteoarthritis, in: 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), pp. 851--854. doi:10.1109/ISBI.2018.8363705. * Ronneberger et al. [2015] Ronneberger, O., Fischer, P., Brox, T., 2015. U-Net: Convolutional networks for biomedical image segmentation, in: Medical Image Computing and Computer-Assisted Intervention (MICCAI), Springer. pp. 234--241. doi:10.1007/978-3-319-24574-4_28. (available on arXiv:1505.04597 [cs.CV]). * Soffer et al. [2019] Soffer, S., Ben-Cohen, A., Shimon, O., Amitai, M.M., Greenspan, H., Klang, E., 2019\. Convolutional neural networks for radiologic images: A radiologist’s guide. Radiology 290, 590--606. doi:10.1148/radiol.2018180547. pMID: 30694159\. * Ulyanov et al. [2016] Ulyanov, D., Vedaldi, A., Lempitsky, V., 2016. Instance normalization: The missing ingredient for fast stylization. arXiv preprint arXiv:1607.08022 arXiv:1607.08022. * Woo et al. [2022] Woo, B., Engstrom, C., Fripp, J., Crozier, S., Chandra, S.S., 2022. Anomaly-aware 3D segmentation of knee magnetic resonance images, in: Medical Imaging with Deep Learning. * Yu and Koltun [2016] Yu, F., Koltun, V., 2016. Multi-scale context aggregation by dilated convolutions, in: Bengio, Y., LeCun, Y. (Eds.), 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings, pp. 1--13. * Yushkevich et al. [2006] Yushkevich, P.A., Piven, J., Cody Hazlett, H., Gimpel Smith, R., Ho, S., Gee, J.C., Gerig, G., 2006. User-guided 3D active contour segmentation of anatomical structures: Significantly improved efficiency and reliability. NeuroImage 31, 1116--1128. doi:10.1016/j.neuroimage.2006.01.015. * Zavrtanik et al. [2021] Zavrtanik, V., Kristan, M., Skočaj, D., 2021. Reconstruction by inpainting for visual anomaly detection. Pattern Recognit. 112, 107706\. doi:https://doi.org/10.1016/j.patcog.2020.107706. * Zhou et al. [2020] Zhou, Z., Siddiquee, M.M.R., Tajbakhsh, N., Liang, J., 2020\. UNet++: Redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imaging 39, 1856--1867. doi:10.1109/TMI.2019.2959609.
# SGDraw: Scene Graph Drawing Interface Using Object-Oriented Representation Tianyu Zhang JAIST Ishikawa,Japan <EMAIL_ADDRESS>Xusheng Du JAIST Ishikawa,Japan <EMAIL_ADDRESS>Chia-Ming Chang The University of Tokyo Tokyo,Japan <EMAIL_ADDRESS>Xi Yang Jilin University Jilin,China <EMAIL_ADDRESS>Haoran Xie JAIST Ishikawa,Japan <EMAIL_ADDRESS> ###### Abstract Scene understanding is an essential and challenging task in computer vision. To provide the visually fundamental graphical structure of an image, the scene graph has received increased attention due to its powerful semantic representation. However, it is difficult to draw a proper scene graph for image retrieval, image generation, and multi-modal applications. The conventional scene graph annotation interface is not easy to use in image annotations, and the automatic scene graph generation approaches using deep neural networks are prone to generate redundant content while disregarding details. In this work, we propose SGDraw, a scene graph drawing interface using object-oriented scene graph representation to help users draw and edit scene graphs interactively. For the proposed object-oriented representation, we consider the objects, attributes, and relationships of objects as a structural unit. SGDraw provides a web-based scene graph annotation and generation tool for scene understanding applications. To verify the effectiveness of the proposed interface, we conducted a comparison study with the conventional tool and the user experience study. The results show that SGDraw can help generate scene graphs with richer details and describe the images more accurately than traditional bounding box annotations. We believe the proposed SGDraw can be useful in various vision tasks, such as image retrieval and generation. ## 1 Introduction Scene graph is a common and popular way to describe scene understanding, first proposed for image retrieval tasks to search for images with similar descriptions in image datasets [8]. In addition, scene graphs are used for a wide range of vision applications, such as image retrieval [8, 16], image captioning [3], visual reasoning [17], visual answering [5], and robotics [1]. Among these previous works, the generation of scene graphs from images has seldom been discussed. In this work, we focus on the user interface for generating the scene graph. The task of scene graph generation is to generate a corresponding graph- structure representation from an image, and then to abstract the objects and object relationships from the images. Scene graphs can facilitate the understanding of scenes in images, especially complex ones. The conventional approaches for scene graph creation have been mainly based on a manually annotated image detection dataset, and the annotation tasks are usually time- consuming and laborious. Currently, the visual genome dataset [10] is the most common and popular dataset for scene graphs, with around $10^{5}$ annotated images. However, the scene graph drawing with a bounding box may lose accuracy due to missing detection results. To solve this issue, we aim to provide a novel scene graph drawing interface with users in the loop. Figure 1: Illustration of the structural unit (a) as an object-oriented representation and (b) its example in this work. Taking two objects as an example, the attributes as objects can be directly manipulated, and attached to the objects with user operations. The relationships can then be constructed among the objects. In this work, we provide the interactive drawing user interface for scene graphs using object-oriented representation, as shown in Figure 1, which can achieve more comprehensive and detailed results than using the conventional annotation with a bounding box. Therefore, we propose SGDraw, an object- oriented approach for scene graph annotation by considering objects, object relationships, and object attributes. SGDraw can take advantage of the user’s cognitive ability in the drawing process from target images. To verify the proposed interface, we conducted a comparison study between the previous annotation method using bounding boxes and our proposed approach. The evaluation results demonstrated the superiority of the proposed approach in generating comprehensive and detailed scene graphs. We list our main contributions as follows: * • We propose an object-oriented annotation approach of scene graphs, which adopts the structural unit of the object, its properties, and the relationship among objects. In this way, we can describe the image in detail and generate a comprehensive and detailed scene graph. * • We design a web-based lightweight annotation tool that is easy to deploy and operate so that the proposed SGDraw can generate scene graphs interactively. * • We conduct user experiments to verify the performance of our proposed object- oriented annotation approach for annotation, especially the effectiveness of annotation on complex images. ## 2 Related works ### 2.1 Scene Graph Generation Earlier works on scene graph generation have constructed several datasets, such as RW-SGD [8] and VRD [13]. Visual Genome (VG) [10] became the standard scene graph dataset due to its large size and challenging nature. Scene graphs can be generated in two different approaches [11]. The conventional methods of scene graph generation adopt both object detection and pairwise predicate estimation. The objects are first detected given a bounding box, and then the predicates are predicted using conditional random fields [4, 8] or a classification approach [9, 15]. The other approach is to jointly infer the objects and their relationships based on the suggestions of object regions [24]. These approaches focus on objects and their relations without attention to object attributes that play a significant weight of the image description. We consider that equalizing attributes over objects and their relations can improve the quality of scene graph generation. ### 2.2 User Interface for Scene Graph The scene graph can be drawn with an designed annotation tool with the colleced data from Amazon’s Mechanical Turk [8]. This tool can obtain instances of objects and their relationships by users drawing bounding boxes to identify objects and describing relationships between pairs of objects through text box input. GeneAnnotator [23] provided a semi-automatic annotation tool for scene graphs, which provides rule-based relationship recommendation algorithms that can reduce the annotation effort. SceneSketcher [12] generated scene graphs from sketches to retrieve scene images satisfying the user’s specific requirements via a freehand sketch. In this work, we aim to provide the interactive drawing interface for scene graphs. ### 2.3 Interactive Drawing Interface Interactive interface design can make drawing activities easy and effective to common users. Dualface [7] proposed a drawing interface with global and local guidance to enable novices to draw face portraits. DualSmoke [20] provided a two-stage global-to-local generation framework for the interactive smoke illustration design, which guides ordinary users without knowledge in the field of fluid simulation to interactively complete various smoke illustration designs through sketches. An interactive user interface [6] was proposed to support sketch-based fluid design with a perceptual understanding of human sketches. In contrast to these drawing interfaces, we focus on the user interface for scene graphs. ### 2.4 Object-Oriented Interface Object-oriented methods have been widely used in design tasks. An object- oriented interface [14] was proposed to generate and manipulate polyhedral data flow graphs. An integrated set of object-oriented pixel-based vectorization algorithms [18] was then presented for various classes of graphic objects in engineering drawings. These interfaces performed interactions that were previously tedious or even impossible with a coherent and consistent interaction experience throughout the entire interface. Next, an object-oriented drawing approach [19] was proposed to represent attributes as objects that can be manipulated directly. Considering the basic components of the scene graph, the objects, object properties, and their relationships could be manipulated directly by the user in this work. ## 3 Scene Graph Drawing In this work, we aim to propose a drawing interface for scene graph annotation and generation. The basic structural unit of the proposed approach is represented in Figure 1. Figure 2: Workflow of the proposed interface. (A) With a task image inputted to the SGDraw, the user starts (B) adding objects, and then they can (C) add attributes or (D) relationships. After that, they can choose to (E) use the auxiliary operations (like cloning, as shown in the figure) to change the structure of the graph. Repeat these steps until (F) obtaining the desired scene graph. ### 3.1 Scene Graph In this work, scene graph generation is achieved by objects without bounding boxes, and we provide the open-ended and free-form manipulation of scene graph nodes. The scene graph from an image is defined as follows by a tuple of $(O^{\prime},R)$: $P(G|I)=P(O^{\prime},R|I),$ (1) where $I$ represents the input image, and $G$ is the desired target scene graph containing objects with attributes $O^{\prime}$ and relations $R$. Specifically, the image is partitioned into a set of objects $O^{\prime}=\\{o^{\prime}_{1},...,o^{\prime}_{n}\\}$ corresponding to each of the $n$ objects with attributes in the image. $R=\\{r_{1},...,r_{l}\\}$ is the set of relations between pairs of objects. $P(o^{\prime})=P(o,A),$ (2) where $o^{\prime}$ is the target object with attributes to be generated, containing the input target object $o$ and its attribute set $A$. Each object comes with its unique set of attributes $A=\\{a_{1},...,a_{m}\\}$ corresponding to the $m$ attributes of the target object. In this work, we handle the object and its attributes as a manipulable object unit. By linking the pairwise relationships between different object units, we can perform the task of generating the original image to the scene graph. In contrast to the conventional scene graph generation task that focuses on the objects and their relationships, we also focus on the object attributes that are easily ignored and that accounts for a large proportion of the images. The generated scene graph can be more consistent with the image content and represent the image scene in more details. ### 3.2 Object-Oriented Representation Inspired by the object-oriented drawing approach [19] [22], we aim to propose an object-oriented representation of scene graphs for image annotation and graph editing tasks. In this representation, we can extend attributes to objects directly for manipulation, as shown in Figure 2. Specifically, the proposed object-oriented approach has the following features: Abstraction. The input images may contain much repetitive content, such as many people. For repetitive objects, the concrete objects may contain similar properties. Therefore, we can abstract them into a class, such as the class of people. Through the process of class abstraction, we can reuse the scene graph data to enhance the versatility and scalability of the proposed system. Uniqueness. Each object has its unique identifier by which the corresponding object can be located. The identity of the annotated objects will not change during the whole task. For example, different human objects in the category are given different identifiers, such as Man1 and Man2. By assigning the identifiers to different objects of the same category, we can correspond scene graph nodes to unique objects in the image to generate detailed and higher- quality scene graphs. Polymorphism. Polymorphism refers to the fact that objects have diverse properties. Multiple different attributes can be assigned to the same object at the same time, and the same attribute can be assigned to multiple different objects at the same time. For example, in the case of several men and women, different attributes like black color for clothes and blue color for pants can be assigned to Man1, and attributes like short hair and black hair can be assigned to Man2 and Woman1. The diversity of attributes allows each object to correspond to the same or different attributes , which enhances the flexibility and reusability of the system. Inheritance. Inheritance refers to the ability to define and implement objects on the basis of objects that already exist. The properties and relationships associated with existing objects can be used as their own content, and new properties and relationships can be added. Specifically, the user can select objects to store after drawing the scene graph. All properties and relationships of the selected object can be stored in the drop-down box in the left corner of the interactive interface. In subsequent operations, the user can select the object at any time to inherit all previously stored properties and relationships. The use of inheritance provides a hierarchical structure between objects, attributes, and relationships. Inheritance makes it possible to share common features and increase the reusability of the system. Figure 3: Screenshot of the SGDraw interface. On the left is the input image display area, and on the right is the scene graph operation and real-time generation area. Meanwhile, on the right side of the interactive interface, we provide some common attributes and relationships for users to choose and reference. ### 3.3 SGDraw Interface In order to explore the effectiveness of the proposed SGDraw, we developed a web-based user interface, as shown in Figure 3 (Please refer to the introduction video of SGDraw demos111https://youtu.be/acy0SNLfahg). The objects, attributes, and pairwise relationships between objects are considered as objects that can be manipulated directly. We also provide an open and fully free annotation tool that takes full advantage of the user’s cognitive ability to perform the task of annotating any kind of images and generating their corresponding scene graphs in real-time. In contrast to the previous approach that requires drawing the bounding box on the image first and then describing the objects inside the box, SGDraw can help users easily add/delete and drag nodes depending on their own knowledge of the image content until a satisfactory scene graph is generated. Specifically, we provide the SGDraw with the following interaction features: Figure 4: Examples of scene graphs completed by SGDraw with (a) an image input, and (b) a completed scene graph. (c) shows the scene graph freely designed from the text input “a couple drive to the forest to have a picnic.” Direct operation of the scene graph. In previous interfaces [8, 23], the users cannot observe the visualized scene graphs in real time and modify the input content. To solve this issue, we consider the input of the scene graph as drawing. We provide three kinds of nodes: red nodes for objects; yellow nodes for relationships; blue nodes for attributes. The user can add a new object and attribute by simply right-clicking on the background and then the object respectively. A relationship node can be generated when clicking two different objects. We also provided common attributes and relationships for reference, which were derived from the common ranking in the scene dataset [21]. We belive that common attributes and relationships can simplify users’ input and facilitate creative thinking. Various auxiliary operations. We designed the auxiliary operations to modify the graph structure and simplify the input process, including removing, cloning, undoing, zooming, collapsing, and dragging (see supplementary material for details). We implemented the undo function to reduce the loss caused by user error operations, and the removing function to help users change the graph structure freely. For complex scene graph drawing, we implemented the zooming and collapsing functions to help users understand the overall graph structure. The dragging function allows users to optimize the layout to achieve visual results that are easier to operate and observe. We also added the cloning function for similar attributes and relationships to increase the efficiency of the drawing process. These operations are directly conducted on the scene graphs with real-time feedback to help users clearly understand the changes in the scene graphs. Flexible data archive. SGDraw has designed two ways to save the scene graph: JSON files to facilitate subsequent research, and SVG files so that the visual scene graphs becomes available to study and understand the scene images more intuitively. For SVG files, users can draw the images directly and make the image interactive by changing part of the code, and then inserting it into HTML while viewing through the browser. SGDraw allows the users to load the graph and modify it for further usage, which can improve the reusability and maintainability of the scene graph data. ## 4 User study In the user study, we confirmed SGDraw for scene graphs generations and collected feedback from potential users about the effectiveness and usability of the interface. We asked 14 participants (aged 25 to 30, eight males and six females) who are graduate students with knowledge of machine learning. Especially, we designed two experiments: a comparison study and a user experience study. We reproduced the previous interface for scene graphs [8] in the comparison study, and seven task images used in the experiments were selected from the Visual Genome dataset [10]. The evaluation process consisted of three stages: 1) Introduction and Training (10–15 minutes). We first introduced the background of the scene graphs. The experimenter then guided the participants to explore the functions and workflow of the previous interface and SGDraw, and we intentionally provided a simple image for participants as an example. This was done for two reasons: First, is ensured that participants had a full understanding of the interfaces, and second, it allowed participants the opportunity to explore and demonstrate understanding. 2) Experience and Usage (20–30 minutes). In order to avoid the proficiency effect caused by the interface order for usage, participants were divided into two groups. One group used the previous interface first and then used SGDraw. The other group used SGDraw first and then used the previous interface. Each participant was asked to draw the scene graphs in two interfaces based on two different images. The system automatically counted time without user knowledge. We provided seven images for participants to complete the scene graph. When the participants thought they had finished describing the content of the images, the experimenter collected the results. We divided the results into seven tasks based on seven provided images. Each task included four results for an image, two results from the SGDraw, and two results from the previous interface. We collected the objects, attributes, and relationships that participants described on the scene graphs. After collecting the data, we calculated the instances (including objects, attributes, and relationships) per minute for each interface to compare the efficiency of the two interfaces. 3) Questionnaire and Interview (10–15 minutes). The participants next completed questionnaires about the system. The questionnaires were composed of a 5-point Likert scale (1 for strongly disagree and 5 for strongly agree) to collect the participants’ experience of SGDraw. The questions on the first questionnaire, based on the System Usability Scale(SUS), were used for the usability measurement of an interface. The second questionnaire collected usability feedback on every function. The interview then consisted of open- ended questions that were asked to gain the users’ feedback on workflow, utility, and usability. ## 5 Results ### 5.1 Generated Scene Graph Figure 4 shows that SGDraw could help users complete the scene graph drawing task. Without the given ground truth image, the users would use SGDraw to express the image they desired. Users were not limited to using the scene graphs in the dataset for subsequent tasks such as generation and retrieval. In addition, compared with the results of previous research that did not visualize [8] or whose layout was messy [23], the generated scene graphs in SGDraw are constructed based on an automatic tree layout, which enhances the readability and aesthetic of the users’ results. The automatic tree layout helps but does not limits users. The results will be expanded automatically and hierarchically, and users are allowed to change the graph structure with any node or link. In the results, red nodes represent objects, blue nodes represent attributes, and yellow nodes represent relationships. The nodes classified by colors strengthen the user’s understanding of specific details and further enhance the comprehensibility of the results on top of the clear layout. The joint storage of the visual scene graphs and the text results are more conducive to the user’s subsequent use and modification. Figure 5: We conducted the comparison study on Pre-UI [8] and the propsed SGDraw. (a) shows the Pre-UI that we reproduced, and (b) shows the visualization result of the graph. (c) is the screen shot of SGDraw, and (d) is the drawn result. ### 5.2 Comparison Study Figure 6: Scene graph generation results from user study. One participant used SGDraw to complete the complex scene graph for the task image by the auxiliary tool. The similar attributes and relationships of objects in the image made our interface facilitate the scene graph drawing process. Figure 7: The time cost of tasks. The system counts time automatically and implicitly. In the comparison study, participants are asked to draw scene graphs with the previous interface (pre-UI) and SGDraw respectively, with two given images, as shown in Figure 5. We recorded the number of objects, attributes, and relationships and the time cost. As shown in Figure 7, SGDraw does not achieve a time advantage on all drawings, but it does achieve the advantage of instance quantity(including objects, attributes, and relationships.) as shown in Figure 8. To verify efficiency, we calculated the number of instances that could be completed for each task per minute. In all experiments, SGDraw achieved the advantage of efficiency. For Task 1, SGDraw had a larger time cost but better performance in instances quantity (4.9 instances per minute, and 2.1 instances per minute for pre-UI). As shown in Figure 6, the objects of the task have the same attributes and relationships. After completing the scene graph of one person, the participants only needed to use the cloning function with a few modifications using SGDraw to complete the whole task. For Task 7, SGDraw achieved higher scores in both time cost and instance quantity than pre-UI. After the interview and analysis, we found that the recommendation part played a significant role. For images with blurry objects, participants had to think about and describe the scene for a long time. The common part helped participants facilitate the input that similar attributes can be added. This part enlightened the participants’ minds and reduced their thinking pressure. Task 6 had the smallest difference in the efficiency of the two interfaces. In this task, the image composition was simple and clear; thus, the participants understood the main objects explicitly in the image. Based on less similar attributes and relationships, scene graph was completely dependent on the participants to finish it, so that SGDraw achieved fair efficiency for this task. ### 5.3 User Experience Study Workflow. Participants were interviewed about how the SGDraw integrated with their workflow. Almost all responded positively; the specific feedback from the participants included the following: “After learning the simple operations, my workflow in this interface was fluid.” (P2). “In traditional painting, the overall composition of the picture is firstly carried out to determine the position of the objects. In addition, some people have the habit of proceeding to the next object after describing one object. Either of these two workflows can be well implemented in this interface, and both have specific functional assistance.” (P5). Participants also pointed out that there is some disadvantage to the SGDraw. “The system is smooth and can complete tasks well, but a lot of use of the right click will be unfriendly to people who are used to the left click.” (P10). According to the interviews, SGDraw obtained a good performance in the workflow. The participants agreed that SGDraw is smooth and helpful for users with different backgrounds. We also found that few function operations may cause discomfort for users with specific computer habits. Figure 8: For each task, experimenters calculated the instances (including objects, attributes, and relationships) in two interfaces. The number of instances that are completed by the SGDraw is more than pre-UI [8]. Utility. Participants were also asked to rate the usefulness of each of the functions. The functions of SGDraw received a high agreement, as shown in Figure 9. All of the functions obtained more than four points, which verified that the proposed functions could achieve simple operation and good user experience. Among the various functions, SGDraw’s ability to delete the link and node with their subtrees was strongly praised by participants. In addition, the scores of some functions were underrated due to the limitations of usage fields. For example, in the scenes with complex and repetitive objects (such as stadiums, and parking), the cloning function always brought a good experience to users. However, in other simple scenes, cloning was not so important. Figure 9: Result of five-point Likert-scale responses to “Do you agree that technique is useful for your scene graph drawing tasks?” Based on the various functions, that SGDraw could save and reuse scene graphs was strongly favored by participants. Users could thus build their datasets in turn to facilitate later operations. Some specific feedback included the following: “I like the Cloning function; it frees me from repetitive typing work. Being able to replicate attributes and relationships together is simply awesome.” (P5). “The Removing function combines delete and rollback, which is a sensible approach. I can easily do what I want, and reduce the cost of my mistakes.” (P7). The functions in SGDraw have high approval that indicates that the various functions are valued and desired. Users can achieve the desired results through the functions we provided in SGDraw, which makes the input process more convenient and quick. Usability. Participants were asked to complete questionnaires about interface usability. The questionnaire was designed with the question items using SUS (System Usability Scale), which provides an overall usability assessment measure consisting of 10 items. We use a five-Point Likert scale ( one for strongly disagree and five for strongly agree). Figure 10: The result of the SUS questionnaire. As shown in Figure 10, SGDraw achieved a good performance with high scores in positive items and low scores in negative items. Almost all participants agreed that SGDraw is easy to use (Scored 4.36) and that most people learned to use it quickly (Scored 4.57). They strongly disagreed that there were too many inconsistencies in SGDraw (Scored 1.29) and many things to learn before usage (Scored 1.57). The total SUS availability score of SGDraw was 78.9 out of 100. Based on the results of previous research [2], this total score shows that SGDraw should be judged to be acceptable, and the Adjective Rating should be “excellent.” All participants commented that the interface is intuitive and easy to learn. The participants noted the following: “The common attributes part is great! It helps me spread my mind when I don’t know what to write, and it’s usually hidden without making me rely on it.” (P11). “No need for more knowledge, it seems friendly for most people.” (P1). “It is interesting to operate directly on the results. You can intuitively see the changes in the graphs. Compared with the textual results, the graphs are always easier to understand.” (P4). Several participants pointed out that more functions are needed, such as selecting the common attributes by themselves: “Quite smart and consistent functions it has created. For people with different tasks, the common attributes required are also different. Perhaps it is a better way to let users set the common attributes themselves.”(P6). Based on the interviews, the visualization results and direct operations on the graphs have received favorable reviews. We noted that the common area on the right side of SGDraw sparked discussion. Participants agreed that it can simplify operations and inspire users’ minds, but it should also be able to be smarter. At the moment, we only display some common attributes fixedly. However, different users need different common attributes. Allowing users to change the common area by themselves will improve the interface’s interactivity. ## 6 Conclusion This work presented SGDraw, a scene graph drawing interface based on the object-oriented method, to help users draw scene graphs succinctly and conveniently. We designed a set of complete functions and presented common tools to ease user operations. Finally, we conducted a user study to verify our proposed interface. We found that the proposed interface can especially work well with complex scene graphs. For limitations of this work, the frequent right clicks may make users who are accustomed to left clicks uncomfortable. In addition, the function design can be improved to help users obtain free and better experiences. As future work, we envision combining SGDraw with the scene graph generation algorithms, which can help users semi-automatically complete the scene graph drawing. ## References * [1] Saeid Amiri, Kishan Chandan, and Shiqi Zhang. Reasoning with scene graphs for robot planning under partial observability. IEEE Robotics and Automation Letters, 7(2):5560–5567, 2022. * [2] Aaron Bangor, Philip T Kortum, and James T Miller. An empirical evaluation of the system usability scale. Intl. Journal of Human–Computer Interaction, 24(6):574–594, 2008\. * [3] Shizhe Chen, Qin Jin, Peng Wang, and Qi Wu. Say as you wish: Fine-grained control of image caption generation with abstract scene graphs. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9962–9971, 2020. * [4] Bo Dai, Yuqi Zhang, and Dahua Lin. Detecting visual relationships with deep relational networks. In Proceedings of the IEEE conference on computer vision and Pattern recognition, pages 3076–3086, 2017. * [5] Marcel Hildebrandt, Hang Li, Rajat Koner, Volker Tresp, and Stephan Günnemann. Scene graph reasoning for visual question answering. arXiv preprint arXiv:2007.01072, 2020. * [6] Zhongyuan Hu, Haoran Xie, Tsukasa Fukusato, Takahiro Sato, and Takeo Igarashi. Sketch2vf: Sketch-based flow design with conditional generative adversarial network. Computer Animation and Virtual Worlds, 30(3-4):e1889, 2019. * [7] Zhengyu Huang, Yichen Peng, Tomohiro Hibino, Chunqi Zhao, Haoran Xie, Tsukasa Fukusato, and Kazunori Miyata. dualface: Two-stage drawing guidance for freehand portrait sketching. Computational Visual Media, 8(1):63–77, 2022. * [8] Justin Johnson, Ranjay Krishna, Michael Stark, Li-Jia Li, David Shamma, Michael Bernstein, and Li Fei-Fei. Image retrieval using scene graphs. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3668–3678, 2015. * [9] Alexander Kolesnikov, Alina Kuznetsova, Christoph Lampert, and Vittorio Ferrari. Detecting visual relationships using box attention. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, pages 0–0, 2019. * [10] Ranjay Krishna, Yuke Zhu, Oliver Groth, Justin Johnson, Kenji Hata, Joshua Kravitz, Stephanie Chen, Yannis Kalantidis, Li-Jia Li, David A Shamma, et al. Visual genome: Connecting language and vision using crowdsourced dense image annotations. International journal of computer vision, 123(1):32–73, 2017. * [11] Yikang Li, Wanli Ouyang, Bolei Zhou, Jianping Shi, Chao Zhang, and Xiaogang Wang. Factorizable net: an efficient subgraph-based framework for scene graph generation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 335–351, 2018. * [12] Fang Liu, Changqing Zou, Xiaoming Deng, Ran Zuo, Yu-Kun Lai, Cuixia Ma, Yong-Jin Liu, and Hongan Wang. Scenesketcher: Fine-grained image retrieval with scene sketches. In European Conference on Computer Vision, pages 718–734. Springer, 2020. * [13] Cewu Lu, Ranjay Krishna, Michael Bernstein, and Li Fei-Fei. Visual relationship detection with language priors. In European conference on computer vision, pages 852–869. Springer, 2016. * [14] Tobi Popoola, Ravi Shankar, Anna Rift, Shivani Singh, Eddie C Davis, Michelle Mills Strout, and Catherine Olschanowsky. An object-oriented interface to the sparse polyhedral library. In 2021 IEEE 45th Annual Computers, Software, and Applications Conference (COMPSAC), pages 1825–1831. IEEE, 2021. * [15] Mengshi Qi, Weijian Li, Zhengyuan Yang, Yunhong Wang, and Jiebo Luo. Attentive relational networks for mapping images to scene graphs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3957–3966, 2019. * [16] Mengshi Qi, Yunhong Wang, and Annan Li. Online cross-modal scene retrieval by binary representation and semantic graph. In Proceedings of the 25th ACM international conference on Multimedia, pages 744–752, 2017. * [17] Jiaxin Shi, Hanwang Zhang, and Juanzi Li. Explainable and explicit visual reasoning over scene graphs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8376–8384, 2019. * [18] Jiqiang Song, Feng Su, Chiew-Lan Tai, and Shijie Cai. An object-oriented progressive-simplification-based vectorization system for engineering drawings: model, algorithm, and performance. IEEE transactions on pattern analysis and machine intelligence, 24(8):1048–1060, 2002. * [19] Haijun Xia, Bruno Araujo, Tovi Grossman, and Daniel Wigdor. Object-oriented drawing. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, pages 4610–4621, 2016. * [20] Haoran Xie, Keisuke Arihara, Syuhei Sato, and Kazunori Miyata. Dualsmoke: Sketch-based smoke illustration design with two-stage generative model. arXiv preprint arXiv:2208.10906, 2022. * [21] Danfei Xu, Yuke Zhu, Christopher B Choy, and Li Fei-Fei. Scene graph generation by iterative message passing. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5410–5419, 2017. * [22] Tianyu Zhang, Xusheng Du, Chia-Ming Chang, Xi Yang, and Haoran Xie. Interactive drawing interface for editing scene graph. In 2022 International Conference on Cyberworlds (CW), pages 171–172, 2022. * [23] Zhixuan Zhang, Chi Zhang, Zhenning Niu, Le Wang, and Yuehu Liu. Geneannotator: A semi-automatic annotation tool for visual scene graph. arXiv preprint arXiv:2109.02226, 2021. * [24] Guangming Zhu, Liang Zhang, Youliang Jiang, Yixuan Dang, Haoran Hou, Peiyi Shen, Mingtao Feng, Xia Zhao, Qiguang Miao, Syed Afaq Ali Shah, and Mohammed Bennamoun. Scene graph generation: A comprehensive survey. CoRR, abs/2201.00443, 2022.
# AirCon: Over-the-Air Consensus for Wireless Blockchain Networks Xin Xie,Cunqing Hua, Pengwenlong Gu, and Wenchao Xu, X. Xie and C. Hua are with the School of Cyber Science and Engineering, Shanghai Jiao Tong University, Shanghai 200240, China. E-mail<EMAIL_ADDRESS><EMAIL_ADDRESS>P. Gu is with the Department INFRES, LTCI, Telecom Paris, Institut Polytechnique de Paris, France. E-mail<EMAIL_ADDRESS>Xu is with the Department of Computing, The Hong Kong Polytechnic University, Hong Kong, China. E-mail: <EMAIL_ADDRESS> ###### Abstract Blockchain has been deemed as a promising solution for providing security and privacy protection in the next-generation wireless networks. Large-scale concurrent access for massive wireless devices to accomplish the consensus procedure may consume prohibitive communication and computing resources, and thus may limit the application of blockchain in wireless conditions. As most existing consensus protocols are designed for wired networks, directly apply them for wireless users may exhaust their scarce spectrum and computing resources. In this paper, we propose AirCon, a byzantine fault-tolerant (BFT) consensus protocol for wireless users via the over-the-air computation. The novelty of AirCon is to take advantage of the intrinsic characteristic of the wireless channel and automatically achieve the consensus in the physical layer while receiving from the end users, which greatly reduces the communication and computational cost that would be caused by traditional consensus protocols. We implement the AirCon protocol integrated into an LTE system and provide solutions to the critical issues for over-the-air consensus implementation. Experimental results are provided to show the feasibility of the proposed protocol, and simulation results to show the performance of the AirCon protocol under different wireless conditions. ###### Index Terms: Consensus protocol, Over-the-air computation, Wireless blockchain network, Lattice coding ## 1 Introduction The next-generation wireless networks are expected to provide ubiquitous access to heterogeneous devices with ultra-high throughput, reliability, and extremely low latency [1]. This will bring great challenges to the security management in current 4G/5G mobile systems. Blockchain, a generic distributed ledger technology (DLT), has received extensive attention due to the significant advantages from decentralization, immutability, and security in recent years. The blockchain technology enables registering and updating transactions in a decentralized fashion via consensus among participants, which has become the foundation of new security architecture for future wireless networks. The consensus protocol plays an important role in the blockchain system. The efficiency of the consensus protocol determines blockchain system security bounds (fault tolerances) and performance such as transaction throughput, delay, and node scalability[2]. Among many consensus protocols, the proof- based algorithm (PoX) is often used in a public chain, such as proof-of-work (PoW)[3], proof-of-stake (PoS)[4], whereby participants can join/leave the network without authentication. These PoX-based consensus protocols achieve good scalability with the cost of high resource consumption. For instance, the PoW protocol employed in Bitcoin consumes a huge amount of power resources to compute a meaningless hash value. And the confirmation delay and throughput of these consensus protocols also limit the application scenario. The voting- based consensus protocols, such as PBFT[5], RAFT[6], relies heavily on inter- participant communications to achieve consensus. Therefore, the communication resource is critical for the voting-based protocols, which can only be applicable for small/middle scale networks. Most of the existing consensus protocols are primarily designed over wired communication networks, some new challenges may arise when they are deployed in wireless networks. First, wireless devices usually have limited resources (energy, storage, computation, etc.), so the resource-consuming PoX protocols are not suitable. Second, the wireless communication channel condition varies dynamically, not to mention that users often suffer from limited bandwidth resources, which significantly affects the consensus performance, especially for the voting-based protocols. As a solution, the wireless channel features can be utilized to offer a solution to this dilemma. For the voting-based protocols, a block is generated based on the decision from the majority participants[7]. For instance, in the PBFT protocol, each participant determines the consensus results by counting the number of messages with consistent hash from other participants. For traditional “communicate-then-consensus” solutions, the consensus process can be conducted only after all messages from other participants are successful received and decoded. Instead of collecting messages from all participants, the consensus process only require the result of a function of these messages, e.g., the number of messages with consistent hash from other participants, rather than the details bits of each message. Therefore, a “communicate-and- consensus” scheme can be adopted to reduce the communication and computational complexity. Specifically, the communicating and consensus tasks can be accomplished simultaneously by leveraging the over-the-air computation (AirComp) technology[8] in wireless networks, which is a novel technology that exploits the superposition property of wireless multiple access channel (MAC) to aggregate signals from multiple participants over the same wireless channel. In this way, the destructive interferences from different participants are turned into a constructive one by appropriately matching the structure of the wireless channel and the function of the signals can be computed from the aggregated signal. In this paper, we propose “AirCon”, an over-the-air consensus protocol for wireless networks, which significantly reduces the communication and computational overhead in a wireless network by using the AirComp technique. The novelty of this protocol is two-folded. First, the hash symbols of all participating users are encoded using lattice codes and transmitted to the base station (BS) simultaneously over the same wireless spectrum via the AirComp technique, which greatly reduces the wireless resource usage. Second, by leveraging the structural property of the lattice codes, each user can verify the consistency of its hash value with respect to the aggregated hash symbols without decoding the signal, whereby the consensus can be achieved with extreme low computational complexity. To show the feasibility of the AirCon protocol, we implement AirCon on an LTE system, where two key issues are solved: 1) Synchronization problem, which is the key to ensure that all participants transmit data in the same frequency simultaneously such that the signals can be accurately aligned at the BS. We will discuss the details of LTE synchronization mechanisms and show that these intrinsic mechanisms are sufficient for satisfying the synchronization requirement for the implementation of AirComp in an LTE system. 2) Uplink channel estimation and feedback problem. The participants need to learn the uplink channel state information (CSI) so that the channel fading can be pre- compensated. To this end, we propose a flexible reference symbol assignment scheme, which achieves a better tradeoff between estimation accuracy and estimation latency. The main contributions of the paper are summarized as follows: * • We propose a novel AirComp-based consensus protocol (AirCon), which enables all participants to transmit their hash symbols to the BS simultaneously using the same wireless channel, and thus significantly reduces the communication cost of consensus messages. * • We propose a hash consistency verification scheme in the physical layer, wherein each participant determines whether its hash value is consistent with the aggregated hash symbols or not without decoding the signal, which significantly reduces the complexity of consensus computation. * • We implement AirCon based on srsLTE111The latest version of srsLTE has been renamed as srsRAN and to support 5G standards.[14], an open-source LTE soft defined radio (SDR) platform. Extensive experiments are conducted to validate the practical performance of AirCon in a real-world LTE testbed. We also provide simulation results to evaluate the performance of the proposed schemes under more general network conditions. The rest of this paper is organized as follows. In Section 2, we discuss the research progress in blockchain-enabled wireless networks and the AirComp technique. In Section 3, we introduce the system model and propose a modified PBFT-based over-the-air consensus (AirCon) procedure for wireless blockchain systems. The detailed design of AirCon protocol is presented in Section 4 and the details for implementation of AirCon in the LTE system are discussed in Section 5. Experimental and simulation results are provided in Section 6. Finally, we summarize this work in Section 7. ## 2 Related Works In this section, we give a brief introduction to the research progress on blockchain-enabled wireless networks and AirComp technique. ### 2.1 Blockchain-enabled wireless networks Blockchain has been studied extensively in literature as a security and privacy protection scheme for various application scenarios in wireless networks. In [15], the authors examined the application of blockchain in the Internet of things (IoT). They concluded that the blockchain-IoT combination is powerful and can lead to significant transformations across several industries. Blockchain technology also can be applied in the internet of vehicles[16] to provide cybersecurity protection for vehicular communications, including the dynamic control of source reliability, and the integrity and validity of the information exchanged. The incorporation of blockchain into the next-generation radio access network (RAN) also attracted great interest. A unified framework of the blockchain radio access network (B-RAN) was proposed in[17] as a trustworthy and secure paradigm for upcoming 6G networking. Some critical elements of B-RAN, such as the deployment of smart contract[18], trustworthy access[19, 20], mathematical modeling[21] were also explored. Due to the resource-limited feature of wireless devices, the blockchain system needs to be improved for the wireless environment. The authors in[17] proposed proof-of-devices (PoD) as a low-cost consensus protocol, which is based on the fact that B-RAN is comprised of a tremendous number of devices and attackers cannot control $51\%$ devices of the whole network. In an edge computing scenario, the authors in [22] consider constructing a collaborative mining network (CMN) to execute mining tasks for mobile blockchain. Miners can offload their mining tasks to non-mining devices within a CMN when the resources are insufficient. The authors in [23] proposed a lightweight blockchain system to reduce the computational cost and speed up the block generation rate in the industrial IoT (IIoT). On the other hand, some research works were concerned with the impact of a wireless channel on blockchain performance. In[24], the authors analyzed the trade-off between communication reliability and computing power in blockchain security and presented a lower bound to the computing power that is needed to conduct an attack with given communication reliability. Based on the widely used CSMA/CA mechanism, the impact of communication transmission delay on the confirmation delay, transaction per second, and transaction loss probability is analyzed in [25]. The authors in [26] analyzed the impact of node geographical distribution in the spatial domain and designed an algorithm to determine the optimal full-function node deployment. In [2], the authors evaluated the impact of scarce frequency spectrum resources on blockchain performance. ### 2.2 Over-the-air computation In this paper, we design a novel consensus protocol based on the AirComp technology, which exploits the intrinsical superposition property of the wireless channel to achieve efficient transmissions of multiple users. AirComp was proposed by B. Nazer and M. Gastpar in their cornerstone work[8], where an optimal joint source-channel strategy was developed to reliably reconstruct a function of sources over a multiple-access channel. Early research works of AirComp took on the information-theoretic view and focuses on the achievable computation rate under structured coding schemes. The transmission strategy under different source distribution was explored in [27, 28, 29]. The authors in [27, 28] employed lattice codes to efficiently compute the sum of source signals over MACs. Some recent research efforts have been made to study AirComp from other aspects. The authors in [30, 31] considered a more general power-limited scenario where some devices cannot fully compensate for channel fading, then a threshold-based power-control policy was proposed to minimize the mean square error (MSE) of the whole system. Multifunction computation was achieved by using AirComp in a multiple antenna system, beamforming strategies have been designed in [32, 33] to achieve optimal performance. In[34], the blind AirComp technique was proposed for low-complexity and low-latency wireless networks, which does not require channel state information for AirComp. The AirComp technique can be applied in a wide range of application scenarios. In[10, 11], AirComp was proposed for fast data fusion, whereby the fusion center (FC) attempts to compute a specific function from the data of all sensor nodes (e.g., average reading). AirComp also has been extended for model aggregation in the distributed edge learning system [35, 36, 37], which is shown to reduce the communication latency without significant loss of the learning accuracy[38]. In a multi-agent system, AirComp can be used for distributed consensus control[39, 40] to reach an agreement over a set of variables of common interest, such as velocity, acceleration, and trajectory in vehicular platooning. Our work extends the application of AirComp to the consensus protocol in wireless blockchain systems. Despite the extensive theoretical work in literature, the research on the implementation issues for AirComp is relatively scarce. In[10] and [11], the authors proposed to modulate the value to be computed through AirComp as the mean value or power of a random sequence, so the accurate synchronization and CSI is not necessary for implementing the AirComp technique. In [12], the authors proposed to achieve system synchronization via the “AirShare” scheme [41], and the uplink CSI was obtained by using channel reciprocity between uplink and downlink transmissions. In [13], the uplink CSI was obtained by feedback from the fusion center. These implementation schemes were designed for analog function computation (AFC) in a wireless network. In this paper, we implement the AirComp technique based on the universal LTE system, which demonstrate that AirComp can be realized in the current digital communication systems. ## 3 System model Figure 1: Blockchain-enabled wireless network model. Figure 2: Diagram of the over-the-air consensus protocol. Hash verification for the generated block is achieved through the BS using the multiple access and broadcast channels. We consider a blockchain-enabled wireless network as illustrated in Fig. 1, where a set of $K$ users are served by a BS. These users maintain the same blockchain. The transaction blocks are generated periodically based on a consensus protocol, and stored in a distributed manner in all users. In a blockchain system, the block appending procedure can be separated into three stages: Request, Consensus and Confirm, as shown in Fig. 2. Our work focus on the consensus step and the other two steps are the same as the traditional blockchain systems. Request: the client sends a request to the primary user to trigger a transaction. After collecting enough transactions, the primary user generates a candidate block and broadcasts it to other replica users and requests to start the consensus procedure 222We assume that the primary user is always honest. If other users or clients detect some malicious behavior from the primary user (such as broadcasting some illegal transactions), view change procedures[5] can be triggered to replace the primary user. Consensus: The consensus protocol in Fig. 2 is a revision of the PBFT protocol for wireless networks, which consists of four phases as follows: * • Pre-prepare: After receiving the candidate block and request from the primary user, all replica users verify the transactions of the candidate block, then send the hash symbols of candidate blocks to the BS in the same uplink channel and enter the prepare phase. Note that user-specific information (such as user ID) should not be included in the generation of the hash value. * • Prepare and Commit: The BS obtains the superposed hash symbols and broadcasts them to all users. All replica users adopt a threshold-based two-round hash verification method to determine the consensus result. * • Reply: Only those users satisfying the threshold return the reply message to the primary user through the BS. If more than $\lceil{K/2}\rceil$ users return the message, the primary user determines that the consensus has been achieved. Confirm: If the consensus has been achieved, the primary user sends a confirm message to the BS, which is broadcast to all replica users and the client, then the candidate block is appended in each blockchain which is distributively stored at each replica user. The bottleneck of the consensus protocol shown in Fig. 2 lies in the _Prepare_ and the _Commit_ phases. In the traditional hash cross-validation, the hash value generated by different users should be transmitted to each other via orthogonal wireless channels (such as TDMA, FDMA, or OFDMA in 4G/5G networks). Therefore, the wireless resource usage in these two phases increases with the number of participating users, which cannot guarantee reliable and low latency consensus for larger network size. ## 4 AirCon Protocol Design In this section, we propose a novel over-the-air consensus (AirCon) protocol to address the transmission bottleneck problem in the _Prepare_ and the _Commit_ phases based on the AirComp and the lattice coding techniques. The idea is to modulate the hash bits of different users using the lattice codes, which can be transmitted simultaneously to the BS over the same wireless channel. Due to the structural property of the lattice codes, the BS can adopt the AirComp technique to aggregate the linear combination of all codes (which is a codeword itself), and forwards it back to all users, which can be used by each user to make a consensus decision without decoding the signal. In the following, we first introduce the preliminary background on the AirComp and lattice coding techniques. We then present the hash consistency verification scheme based on these two techniques, with which a two-round consensus procedure for the AirCon protocol is designed. ### 4.1 Over-the-air computation We consider an OFDM-based wireless system, which consists of a single BS and a set of $K$ users. All users transmit signals to the BS through the same uplink channel. The received signal at the BS is expressed as: $\mathbf{y}=\sum_{k=0}^{K-1}\mathbf{H}_{k}\mathbf{B}_{k}\mathbf{x}_{k},$ (1) where $\mathbf{x}_{k}\in\mathbb{C}^{N\times 1}$ is the symbol vector transmitted by user $k$ with the covariance matrix $\mathbb{E}\\{\mathbf{x}_{k}\mathbf{x}_{k}^{H}\\}=\mathbf{I}$, where $\mathbf{I}$ is the identity matrix. $\mathbf{B}_{k}=\texttt{Diag}\\{b_{k0},b_{k1},\cdots,b_{kN-1}\\}\in\mathbb{C}^{N\times N}$ is the pre-processing matrix for user $k$. $\mathbf{H}_{k}=\texttt{Diag}\\{h_{k0},h_{k1},\cdots,h_{kN-1}\\}\in\mathbb{C}^{N\times N}$ is the channel matrix from user $k$ to the BS. $N$ is the total number of sub-carriers. $K$ is the total number of users. The superscript $(\cdot)^{H}$ denotes Hermitian transpose. Under the condition that each user $k$ has perfect knowledge of $\mathbf{H}_{k}$, and its transmit power is not bounded, then the channel fading can be pre-compensated by a simple channel-inversion operation as $\mathbf{B}_{k}=\mathbf{H}_{k}^{-1}$. In this way, the signals transmitted by all users will be aggregated at the BS as follows: $\mathbf{y}=\sum_{k=0}^{K-1}\mathbf{x}_{k},$ (2) which is equivalent to the computation of the summation of all $\mathbf{x}_{k}$s over the wireless channel. Motivated by this property, the AirComp technique can be generalized to support a wide range of mathematical operations, which are based on the property of nomographic function in the following form: $f(\mathbf{s}_{0},\mathbf{s}_{1},...,\mathbf{s}_{K-1})=\psi{\bigg{(}}\sum_{k=0}^{K-1}\phi{(\mathbf{s}_{k})}{\bigg{)}},$ (3) where $\phi(\cdot)$ and $\psi(\cdot)$ denote pre- and post-processing functions, respectively. Some well-known nomographic functions include: 1. 1. Arithmetic Mean: $f(s_{0},s_{1},...,s_{K-1})=\frac{1}{K}\sum_{k=0}^{K-1}s_{k}$, with $\phi(s_{k})=s_{k}$ and $\psi(y)=y/N$; 2. 2. Euclidean Norm: $f(s_{0},s_{1},...,s_{K-1})=\sqrt{\sum_{k=0}^{K-1}s_{k}^{2}}$, with $\phi(s_{k})=s_{k}^{2}$ and $\psi(y)=\sqrt{y}$; 3. 3. Number of Active Node: $f(s_{0},s_{1},...,s_{K-1})$ is the number of active node, with $\phi(s_{k})=1\,(\text{active})\,\text{or}\,0\,(\text{inactive)}$ and $\psi(y)=y$. ### 4.2 Nested lattice code A lattice is an infinite discrete set of points in the Euclidean space that are regularly arranged and are closed under addition[42]. As an important channel coding technique, the structural properties of nested lattice coding are well suited for multiple access channel in wireless networks, which allows multiple transmitters to effectively share the same radio resources and can protect against channel noise. Specifically, a $d$-dimensional lattice $\Lambda$ in the Euclidean space $\mathbb{R}^{d}$ can be generated as follows: $\Lambda=\\{\mathbf{Gu}\ :\ \textbf{u}\in\mathbb{Z}^{d}\\},$ (4) where $\mathbf{G}=[\mathbf{g}_{1},\mathbf{g}_{2},...,\mathbf{g}_{d}]$ is a full-rank generator matrix. A lattice $\Lambda_{C}$ is _nested_ in some lattice $\Lambda_{F}$ if $\Lambda_{C}\subseteq\Lambda_{F}$, i.e., $\Lambda_{C}$ is a sublattice of $\Lambda_{F}$. In this case, $\Lambda_{F}$ is denoted as the fine lattice, which defines the codewords, while $\Lambda_{C}$ is denoted as the coarse lattice, which is used for shaping. Specifically, the _nested lattice code_ $\mathcal{L}$ is the set of all points of a fine lattice $\Lambda_{F}$ that is within the fundamental Voronoi region $\mathcal{V}_{C}$ of a coarse lattice $\Lambda_{C}$: $\mathcal{L}=\Lambda_{F}\cap\mathcal{V}_{C}=\\{\textbf{x}\ :\ \textbf{x}=\lambda\ \text{mod}\ \Lambda_{C},\lambda\in\Lambda_{F}\\},$ (5) where the _fundamental Voronoi region_ , $\mathcal{V}_{C}$, of the lattice $\Lambda_{C}$, is the set of all points in $\mathbb{R}^{d}$ that are closest to the zero vector: $\mathcal{V}_{C}=\\{\textbf{z}\ :\ ||\textbf{z}||\leq||\textbf{z}-\lambda||,\forall\lambda\in\Lambda_{C},\textbf{z}\in\mathbb{R}^{d}\\},$ (6) For each user $k$, a $d$-dimension nested lattice codeword ${\bf{x}}_{k}\in\mathcal{L}$ can be generated based on its hash value ${\bf{s}}_{k}\in\mathbb{F}^{l}_{p}$ by the encoding function $\phi(\cdot)$ as follows: $\displaystyle\phi({\bf{s}}_{k}):\mathbb{F}^{l}_{p}$ $\displaystyle\rightarrow\mathbb{R}^{d}$ (7) $\displaystyle{\bf{s}}_{k}$ $\displaystyle\rightarrow{\bf{x}}_{k}$ where $\mathbb{F}_{p}=\\{0,1,...,p-1\\}$ forms a finite field under integer arithmetic modulo $p$. Then for a nested lattice codebook $\mathcal{L}$, the following property is held for all ${\bf{x}}_{k}\in\mathcal{L}$: ${\bigg{[}}\sum_{k=0}^{K-1}{\bf{x}}_{k}\mod{\Lambda}_{C}{\bigg{]}}\in\mathcal{L},$ (8) that is, the sum of lattice codewords modulo the shaping lattice is a codeword itself. Due to this linearity preserving characteristic, there exists a post- processing function $\psi(\cdot)$ that satisfies[9]: $\psi{\bigg{(}}\sum_{k=0}^{K-1}\phi({\bf{s}}_{k})\mod{\Lambda_{C}}{\bigg{)}}=\underset{k=0}{\overset{K-1}{\bigoplus}}{\bf{s}}_{k},$ (9) which is a nomographic function, that is, $f({\bf{s}}_{0},{\bf{s}}_{1},...,{\bf{s}}_{K-1})=\underset{k=0}{\overset{K-1}{\bigoplus}}{\bf{s}}_{k}$. Based on this property, AirComp can be used to transmit and compute the superposition of hash values from all users. Specifically, each user $k$ maps its hash value $\mathbf{s}_{k}$ to a lattice codeword $\mathbf{x}_{k}$ using the mapping operation in Eq.(7), then all users can transmit their lattice codes simultaneously to the BS using the AirComp technique. At the BS, the received signal is $\mathbf{y}=\sum_{k=0}^{K-1}\mathbf{x}_{k}+\mathbf{w}$, which is a $d$-dimension linear combination of the lattice codes of all users with noise of $\mathbf{w}$. In theory, there are many choices for the lattice codes with different codeword dimensions. However, in practice, the transmitted symbol can only represent information in a 2-dimensional space (in-phase and quadrature dimension). Therefore, for the high dimensional hash value, such as the 128-bits hash generated by the MD5 algorithm[43], it needs $2^{64}$ codewords at each dimension to represent the hash, which is impossible in a real communication system. Figure 3: Nested $\mathbb{Z}^{2}$ cubic lattice. In this work, we consider an $\mathbb{Z}^{2}$ cubic lattice $\tilde{\Lambda}_{F}$ as illustrated in Fig. 3, where the points are $\tilde{\Lambda}_{F}$ points and the Voronoi region of the coarse lattice $\tilde{\Lambda}_{C}$ is drawn in dashed lines. Only the eight codes in the outer tier of the Voronoi region are used (shown in red color), so the code rate is $B=\log_{2}8=3$ bits/symbol. Therefore, for an $L$-bits hash value, it can be partitioned into $N=\lceil{L/3}\rceil$ symbols. ### 4.3 Hash consistency verification Based on the $\mathbb{Z}^{2}$ cubic lattice in Fig. 3, each user $k$ can construct a hash symbol vector $\tilde{\bf{x}}_{k}=[\tilde{x}_{k0},\tilde{x}_{k1},...,\tilde{x}_{kN-1}]^{T}$, where $\tilde{x}_{kn}$ is mapped to one of the eight codes. Then each element of $\tilde{\bf{x}}_{k}$ is transmitted in an OFDM sub-carrier. The received symbol at the BS in sub-carrier $n$ is $\tilde{y}_{n}=\sum_{k=0}^{K-1}\tilde{x}_{kn}+w_{n}$. The received symbol at sub-carrier $n$ can be quantized to the nearest point $t_{n}\in\mathbb{Z}^{2}$ in the $\tilde{\Lambda}_{F}$ as follows: ${t_{n}}=\arg\underset{\lambda\in{\tilde{\Lambda}_{F}}}{\min}||\tilde{y}_{n}-\lambda||,\,n\in[0,N-1],$ (10) which are then broadcasted to all users through the downlink channel. A vector $\mathbf{t}=[t_{0},t_{1},...,t_{N-1}]^{T}\in\mathbb{Z}^{2N}$ is constructed by each user, which is the complete linear combination of the lattice codes of all users. Based on the received vector $\mathbf{t}$, we design a consensus algorithm by exploiting the geometric characteristics of lattice codes. Let $C_{m}$ denote a set of $m$ users that transmit the consistent hash symbol vector, and $C_{K-m}$ denote the rest $K-m$ users that transmit different hash symbol vectors. Then $\mathbf{t}$ can be re-written as follows: $\mathbf{t}=\sum_{k\in C_{m}}\tilde{\mathbf{x}}_{k}+\sum_{k\in C_{K-m}}\tilde{\mathbf{x}}_{k}+\mathbf{e},$ (11) where $\mathbf{e}$ is the random error pattern after symbols quantization in (10) with $\mathbb{E}\\{\mathbf{e}\\}=\mathbf{0}$. We assume that the hash symbol vector elements are uniformly mapped from the lattice codewords as shown in Fig. 3. For each user $k\in C_{m}$, we assume the transmitted hash symbol vector is $\bar{\mathbf{x}}$, i.e., $\tilde{\mathbf{x}}_{k}=\bar{\mathbf{x}},\forall k\in C_{m}$. For each user $k\in C_{K-m}$, we firstly consider the case that the hash symbol vector is independent of each other. Then for $k\neq j$, we have $\mathbb{E}\\{\tilde{\mathbf{x}}_{k}^{T}\tilde{\mathbf{x}}_{j}\\}=\begin{cases}N\sigma_{s}^{2},&{\text{If}~{}\forall k,j\in C_{m}}\\\ 0,&{\text{Otherwise}}\end{cases}$ (12) where $\sigma_{s}^{2}$ is the variance of codewords. Furthermore, we define $\tilde{I}_{k}$ as the hash consistency factor (HCF) of user $k$, which is the normalized inner product of its hash symbol vector $\tilde{\mathbf{x}}_{k}$ and vector $\mathbf{t}$, that is: $\displaystyle\tilde{I}_{k}=\frac{\mathbf{t}^{T}\tilde{\mathbf{x}}_{k}}{K|\tilde{\mathbf{x}}_{k}|^{2}},$ (13) From (12), we can obtain the expectation of $\tilde{I}_{k}$ as follows: $\mathbb{E}\\{\tilde{I}_{k}\\}=\begin{cases}\frac{m}{K},&{\forall k\in C_{m}}\\\ \frac{1}{K}.&{\forall k\in C_{K-m}}\end{cases}$ (14) Figure 4: $\mathbb{E}\\{\tilde{I}_{k^{\prime}}\\}$ and $\mathbb{E}\\{\tilde{I}_{k^{\prime\prime}}\\}$ vs. $m$ ($K=21$). In Fig. 4, we plot $\mathbb{E}\\{\tilde{I}_{k}\\}$ with 21 users ($K=21$) for $m$ varying from 1 to 20. It can be seen that for $m\geq 2$, $\mathbb{E}\\{\tilde{I}_{k^{\prime}}\\}>\mathbb{E}\\{\tilde{I}_{k^{\prime\prime}}\\},\forall k^{\prime}\in C_{m},k^{\prime\prime}\in C_{K-m}$. Thus $\tilde{I}_{k}$ can be used as a reliable metric for each user to determine whether its hash symbol vector is consistent with the majority of hash vectors in $\mathbf{t}$ or not, that is, it belongs to one of the $m$ users sending the same hash or not. Therefore, this metric can be used to check if the consensus is reached or not. Specifically, in the proposed AirCon protocol, the consensus can be reached if $m>\lfloor{K/2}\rfloor$ users generate the same hash vector. According to (14), we can set a threshold as $T_{h}=0.5$. After receiving the superposed hash vector $\mathbf{t}$, each user $k$ calculates $\tilde{I}_{k}$ as defined in (13). If $\tilde{I}_{k}\geq T_{h}$, the user assumes that the consensus can be achieved and sends the result to the primary user. However, if some users are controlled by an attacker, the consensus performance will be degraded based on this ideal threshold. For instance, consider the case where $m$ users are honest and the rest of $K-m$ users are controlled by a malicious attacker. The hash symbol vector transmitted by honest users is $\bar{\mathbf{x}}$, while the hash symbol vector transmitted by malicious users is $\hat{\mathbf{x}}=-\bar{\mathbf{x}}$, then the aggregated vector at the BS is $\mathbf{t}=m\bar{\mathbf{x}}+(K-m)\hat{\mathbf{x}}=(2m-K)\bar{\mathbf{x}}$. In this case, $\tilde{I}_{k}=(2m-K)/K$ for a honest user $k$, so the consensus cannot be achieved with a threshold of $T_{h}=0.5$ unless $m>\lfloor{3K/4}\rfloor$. Therefore, this one-round consensus procedure is vulnerable to this kind of conspiracy attack. As a solution, we propose a two-round consensus procedure to improve the security of the proposed consensus procedure. Specifically, in the first round, malicious users can be filtered by setting a proper threshold and only honest users enter the second round to decide the consensus results. In the following, we will analyze the optimal threshold for the first round and the corresponding fault tolerance of the proposed consensus protocol. Before getting into the details, we first make some assumptions about the capabilities of an attacker: 1. 1. _AS_ 1: An attacker knows the hash symbols transmitted by each honest user. This is feasible because the attacker can calculate the legal hash by acting as an honest user; 2. 2. _AS_ 2: An attacker cannot manipulate the parameters of the physical layer, such as the transmitting power. 3. 3. _AS_ 3: An attacker cannot manipulate the consensus protocol, such as the consensus threshold. We assume that an attack is successful if the consensus cannot be achieved even if the number of honest users is more than half of the total number of users (i.e., $m>\lfloor{K/2}\rfloor$). Based on the aforementioned attacking model, we assume all honest users transmit $\bar{\mathbf{x}}$ and all malicious users transmit $\hat{\mathbf{x}}$. Then $\mathbf{t}$ can be re-written as follows: $\mathbf{t}=m\bar{\mathbf{x}}+(K-m)\hat{\mathbf{x}}+\mathbf{e},$ (15) and the expectation of $\tilde{I}_{k}$ can be re-written by $\mathbb{E}\\{\tilde{I}_{k}\\}=\begin{cases}1-\alpha+\alpha\rho,&{\forall k\in C_{m}}\\\ \alpha+(1-\alpha)\rho,&{\forall k\in C_{K-m}}\end{cases}$ (16) where $\alpha=\frac{K-m}{K}$ is the percentage of malicious users and $\rho=\mathbb{E}\\{\frac{\hat{\mathbf{x}}^{T}\bar{\mathbf{x}}}{|\bar{\mathbf{x}}|^{2}}\\}=\mathbb{E}\\{\frac{\bar{\mathbf{x}}^{T}\hat{\mathbf{x}}}{|\hat{\mathbf{x}}|^{2}}\\}$ denote the correlation coefficient between $\hat{\mathbf{x}}$ and $\bar{\mathbf{x}}$. The attacker can change $\tilde{I}_{k}$ of the honest users by manipulating $\rho$ between $[-1,1]$ with properly setting of $\hat{\mathbf{x}}$. In the first round of consensus, a threshold should be set such that all honest users can get into the second round while the malicious users should be filtered as many as possible. To this end, the threshold should be set to $T_{h1}=\underset{\rho\in[-1,1]}{\text{min}}(1-\alpha+\alpha\rho)=1-2\alpha$ (17) From the attacker’s perspective, it must find a $\rho$ satisfying $\alpha+(1-\alpha)\rho>T_{h1}=1-2\alpha$ such that malicious users also can get into the second round. Otherwise, all malicious users will be filtered in the first round. Therefore, the range of $\rho$ can be given by $\rho>\frac{1-3\alpha}{1-\alpha}$ (18) In the second round of consensus, the HCF $\tilde{I}_{k}$ of honest users should be more than $T_{h2}=0.5$ such that the consensus can be achieved even if all malicious users satisfy the threshold in the first round of consensus by setting $\rho$ as in (18), that is $\tilde{I}_{k}=1-\alpha+\alpha\rho=1-\alpha+\alpha\frac{1-3\alpha}{1-\alpha}>\frac{1}{2},~{}\forall k\in C_{m}$ (19) Then we can obtain the range of $\alpha$ as follows: $\alpha<\frac{\sqrt{17}-1}{8}\approx 0.39:=\alpha^{*}$ (20) Substituting (20) into (17), we can obtain the optimal value of $T_{h1}$ as follows: $T_{h1}^{*}:=0.22$ (21) The range of $\alpha$ in (20) specifies the fault tolerance of the two-round consensus protocol. If the percentage of malicious users does not exceed 0.39, then the consensus can be achieved in the second round. Otherwise, the consensus cannot be achieved even if $m>\lfloor{K/2}\rfloor$. We show two examples in Fig. 5 and Fig. 6. In Fig. 5, we plot the HCF for all users with $\alpha=0.35$, which is under the threshold $\alpha^{*}$. The range of $\rho$ can be divided into two parts by the critical point $\rho_{0}$ that satisfies $\alpha+(1-\alpha)\rho_{0}=T_{h1}=1-2\alpha$. When $\rho<\rho_{0}$, the region is denoted as _Safety Region_ , all malicious users will be filtered by $T_{h1}$ and cannot enter the second round. When $\rho>\rho_{0}$, the region is denoted as _Consensus Region_ , malicious users can enter the second round, but they cannot affect the consensus results because the HCF of all honest users always exceeds $T_{h2}=0.5$ such that all honest users can enter _Reply Phase_ and send the reply message to the primary user. Note that in the second round, if $\rho_{0}<\rho<\rho^{\prime}_{0}$, the HCF of malicious users is less than $T_{h2}$ such that they cannot send the reply message. If $\rho>\rho^{\prime}_{0}$, malicious users also send a reply message. However, regardless of whether malicious users can send a reply message or not, it will not affect the consensus result because all honest users always send the reply message. Therefore, malicious users cannot affect consensus results whatever $\rho$ is set when $\alpha=0.35$. In Fig. 6, we plot the HCF of all users with $\alpha=0.45$, which is over the threshold $\alpha^{*}$. In this case, we can divide the range of $\rho$ into three parts by two critical point $\rho_{0}$ and $\rho_{1}$, where $\rho_{0}$ satisfies $\alpha+(1-\alpha)\rho_{0}=T_{h1}=1-2\alpha$ and $\rho_{1}$ satisfies $1-\alpha+\alpha\rho_{1}=T_{h2}=0.5$. Similarly, when $\rho<\rho_{0}$, the consensus can be achieved because malicious users cannot enter the second round. When $\rho>\rho_{1}$, the consensus also can be achieved because the HCF of honest users always exceeds $T_{h2}$. However, when $\rho_{0}<\rho<\rho_{1}$, we denote it as _Attacking Region_ , malicious users can enter the second round, and the HCF of all honest users does not exceed $T_{h2}$, so the consensus cannot be achieved and the attack is successful in this region. Figure 5: Consensus performance under fault tolerance, $\alpha=0.35$ Figure 6: Consensus performance over fault tolerance, $\alpha=0.45$ ### 4.4 Two-round consensus procedure Based on the hash consistency verification scheme discussed in the previous subsection, we propose a two-round consensus procedure as follows. In the first round (_Prepare phase_), all users send their hash symbols to the BS via the AirComp technique. The BS feeds back the superposed hash symbol vector $\mathbf{t}$ to all users. Each user $k$ calculates the HCF $\tilde{I}_{k}$ according to (13) and compares it with the threshold $T_{h1}=0.22$. If $\tilde{I}_{k}>T_{h1}$, then the user changes to the prepared state and enters the second round of the consensus process. In the second round (_Commit phase_), only users who are in the prepared state send the hash symbols to the BS via the AirComp technique. The BS feeds back the new superposed hash symbol vector $\mathbf{t}^{\prime}$ to all users. Each prepared user $k$ calculates the HCF $\tilde{I}_{k}$ once again using $\mathbf{t}^{\prime}$ and compares with the threshold $T_{h2}=0.5$. Only those users with $\tilde{I}_{k}>T_{h2}$ enter the _Reply Phase_ and return the reply message to the BS via the AirComp technique. The superposed reply message is forwarded to the primary user by the BS. The final consensus result is determined by the primary user based on the superposed reply message from the BS. The primary user also can calculate the HCF in (13) between its reply message and the superposed message. If the HCF exceeds $T_{h}=0.5$, which suggests that more than half of the total users return the consistent reply message, then the consensus is achieved. ## 5 AirCon implementation based on LTE system In this section, we consider the implementation of AirCon protocol based on the open-source srsLTE platform, which provides the standard LTE protocols. We present solutions to some of the critical issues for practical AirComp implementation. Specifically, two problems are considered for AirComp implementation: 1) Synchronization problem; 2) Uplink channel estimation and feedback problem. ### 5.1 Synchronization The first challenging problem for AirComp implementation is to achieve strict timing/frequency synchronization across different users so that accurate signal superposition can be obtained at the BS. Fortunately, the LTE system has a complete set of synchronization mechanisms that can meet the synchronization requirement of AirComp. Specifically, in the downlink of the LTE system, the BS broadcasts synchronization channels (including PSS and SSS) periodically. If a user wants to connect to the BS, the first step is to search for these two channels to get timing and frequency synchronization. As long as the user is connected to the BS, it will keep tracking the timing/frequency synchronization via PSS/SSS. The user also compensates for the frequency offset of uplink channels according to the estimated value from the downlink channels. In the uplink, a “timing advance” mechanism is adopted by all users for timing synchronization with the BS. Specifically, when the user connects to the BS, the BS measures the propagation delay and calculates a timing offset, named “Timing Advance (TA)”, which is fed back to the corresponding user. Whenever the user transmits data to the BS, it should transmit at the time with an advance of TA so that the signals from different users arrive at the BS without timing offset. In addition, the physical layer of the LTE system is based on the OFDM technique, which adopts a cyclic prefix (CP) to cope with the multipath interference problem. Due to the cyclic property, the CP can also be used for channel synchronization. That is, as long as the timing offset is within a CP, the impact of the timing offset can be treated as a channel phase shift and compensated as a part of channel fading. In summary, the existing synchronization mechanisms in the LTE system are sufficient for the implementation of AirComp, so there is no need to design a dedicated AirComp synchronization scheme in an LTE system, which is part of the reason we choose the LTE system as the implementation platform of the AirCon protocol. ### 5.2 Channel estimation and feedback The purpose of uplink channel estimation is to cope with the channel fading problem via pre-compensation on the user side. In practice, there are two different ways to obtain uplink CSI. One way is that the BS estimates the uplink channel and sends feedback to the corresponding user, the other way is that a user infers the uplink CSI from the downlink signals by leveraging the channel reciprocity of the TDD channel. In this work, we choose the first solution since it can be used for both FDD and TDD systems. To get uplink CSI at the BS, a user needs to transmit reference symbols in the uplink channel. Ideally, the reference symbols should be transmitted at each sub-carrier so that the channel can be estimated accurately for all sub- carriers. However, this will consume too many channel resources and take a longer time for channel estimation for a large number of participating users. To this end, we exploit the property that the sub-carriers within coherence bandwidth may have a similar fading coefficient, so the reference symbols are only needed for every $M$ sub-carriers333$M$ is a parameter depending on the coherence bandwidth. Note that this scheme is also adopted by the uplink sounding reference signal (SRS) in the LTE system.. As a result, at most $M$ users can share an OFDM symbol, therefore a total of $\lceil{K/M}\rceil$ OFDM symbols are needed for reference symbol transmissions. Fig. 7 illustrates the assignment of the reference symbols for different users. Figure 7: Assignment of reference symbols. Upon receiving the reference symbols, the BS extracts the received symbols and estimates the CSI in the order of user index one by one. The simplest estimation method is the least square (LS) method. Taking user 0 as a example, the BS firstly constructs the received symbol vector $\mathbf{y}=[y_{0},y_{M},\cdots,y_{\lfloor{N/M}\rfloor{M}}]$ from symbol 0, the corresponding transmitted reference signal for user 0 is $\mathbf{x}=[x_{0},x_{1},\cdots,x_{\lfloor{N/M}\rfloor}]$, then the LS estimation for uplink channel knowledge of user 0 can be expressed as $\tilde{\mathbf{h}}_{\text{LS}}=\left[\frac{y_{0}}{x_{0}},\frac{y_{M}}{x_{1}},\cdots,\frac{y_{\lfloor{N/M}\rfloor{M}}}{x_{\lfloor{N/M}\rfloor}}\right]^{T}$ (22) However, the LS method does not consider the noise in the reference symbols. The LMMSE method[46] can be employed to further reduce the noise impact by utilizing the cross-correlation between sub-carriers, which is given by [46]: $\tilde{\mathbf{h}}_{\text{LMMSE}}=\mathbf{R}_{\mathbf{h}\mathbf{h}}(\mathbf{R}_{\mathbf{h}\mathbf{h}}+\frac{\beta}{\text{SNR}}\mathbf{I})^{-1}\tilde{\mathbf{h}}_{\text{LS}},$ (23) where $\beta=\mathbb{E}\\{|x_{n}|^{2}\\}\mathbb{E}\\{|1/x_{n}|^{2}\\}$ is a constant depending on the reference symbol, $\mathbf{R}_{\mathbf{h}\mathbf{h}}=\mathbb{E}\\{\mathbf{h}\mathbf{h}^{H}\\}$ is the channel autocorrelation matrix444The matrix $\mathbf{R}_{\mathbf{h}\mathbf{h}}$ can be obtained from either a typical channel model[46] or the channel LS estimation.. Based on the channel estimation $\tilde{h}_{kn}$ for each subcarrier, the BS can set the coefficient of $\mathbf{B}_{k}$ as $b_{kn}={\tilde{h}^{\dagger}_{kn}}/{|\tilde{h}_{kn}|^{2}}$, which can compensate the channel fading and achieve the ideal signal aggregation as shown in Eq.(1) and Eq.(2). This parameter should be fed back to the users as shown in Fig. 8. Similar to the uplink channel estimation, two OFDM symbols are shared by every $M$ users, where the uplink CSI is transmitted in the _Data Symbol_ and downlink reference symbols are transmitted in the _Pilot Symbol_ for downlink channel estimation. Therefore, totally 2$\lceil{K/M}\rceil$ OFDM symbols are needed for downlink feedback transmissions. We still take user 0 as example: the coefficient vector $\mathbf{b}_{0}=[b_{00},b_{0M},\cdots,b_{0~{}\lfloor{N/M}\rfloor{M}}]$ is transmitted at the corresponding sub-carrier in the _Data Symbol_ 0\. In the _Pilot Symbol_ 0, the downlink reference symbol $s_{0}$ is transmitted at the corresponding sub-carrier. After receiving these two symbols, user $0$ firstly estimates the downlink channel using the received symbols from _Pilot Symbol_ 0\. Then the user can obtain the coefficient vector $\mathbf{b}_{0}$ from _Data Symbol_ 0 by canceling the impact of downlink channel fading. Since each user only utilizes $1/M$ sub-carriers, the coefficients for the rest of sub- carriers can be estimated by interpolation[45, 47] at the users. Figure 8: Assignment of downlink feedback symbols. In the low-SNR scenario, in addition to the channel estimation algorithm, a retransmission scheme (i.e., the uplink reference symbols and downlink feedback symbols are transmitted multiple times) can be adopted to further improve the performance. The impact of retransmission times on the consensus accuracy will be studied in section 6. ### 5.3 AirCon implementation The transmission procedure of the AirComp protocol is summarized in Fig. 9, which consists of all procedures discussed in previous subsections. Firstly, the synchronization between the users and the BS will be established. Then the BS estimates uplink CSI for all users based on the uplink reference signals. The coefficients of $\mathbf{B}_{k}$ are computed based on the channel estimation results and fed back to all users. Finally, all users transmit data using the AirComp technique. Figure 9: AirComp signaling and data transmission procedures. We implement the AirCon protocol based on the above AirComp procedures in the LTE system. The AirCon protocol is started when all users are in a stable state, such as the RRC IDLE state. In our implementation, we mainly focus on the feasibility of AirComp technology on hash verification of blockchain consensus, so only hash bits are generated in each user, rather than broadcasting an entire block from the primary user. In this case, we do not appoint a specific primary user. The consensus request is generated in the BS, which is a dedicated system information block (SIB) to notify all users that the BS is ready for consensus and users can transmit an uplink reference signal for AirComp uplink channel estimation. Since a user usually does not receive SIBs when it is in RRC IDLE state, the BS can set the _systemInfoModification_ field in the paging message (all users must receive the paging message periodically) to ensure all users receive the SIB (i.e., the consensus request). Similarly, the consensus result is determined by the BS rather than the primary user. ### 5.4 Complexity analysis of the AirCon protocol The complexity of the consensus protocol in wireless networks mainly comes from two aspects: communication complexity and computation complexity. As a benchmark, the traditional PBFT protocol is also illustrated in Fig. 10 for the convenience of complexity comparison. We mainly focus on the complexity in the _Prepare_ and _Commit_ phases. Figure 10: Traditional PBFT consensus diagram. The total number of messages in the _Prepare_ and _Commit_ phases of the traditional PBFT protocol is $\text{Num}^{\text{message}}_{\text{PBFT}}=(K-1)(K-1)+K(K-1)=(2K-1)(K-1)$ (24) Assuming each message consumes $N$ wireless resource blocks (RBs), then the number of wireless RBs consumed by the PBFT protocol is $\text{Num}^{\text{RBs}}_{\text{PBFT}}=2N\cdot\text{Num}^{\text{message}}_{\text{pbft}}=2N(2K-1)(K-1)$ (25) where the RE consumption in both the uplink and downlink directions is considered. For the AirCon protocol, the total number of wireless RBs consumed in the _Prepare_ and _Commit_ phases is $\text{Num}^{\text{REs}}_{\text{AirCon}}=4N$ (26) Therefore, the communication complexity of the traditional PBFT protocol is $\mathbb{O}(K^{2})$, while the communication complexity of the AirCon protocol is $\mathbb{O}(1)$. If the overhead of channel estimation (CE) and feedback are also considered, the extra overhead is $4\lceil{K/M}\rceil N$, which has a complexity of $\mathbb{O}(K)$. In this case, the total number of consumed RBs is $\text{Num}^{\text{REs}}_{\text{AirCon\\_CE}}=4N+4\lceil{K/M}\rceil N$ (27) Note that the complexity of channel estimation depends on the system. If channel reciprocity is utilized to estimate the uplink channel, the complexity of channel estimation can be reduced to $\mathbb{O}(1)$. The RB consumption in both consensus protocols under different users number is illustrated in Fig.11. Figure 11: Consensus protocol REs consuming. As for the computational complexity, the hash verification procedures in the traditional consensus protocol are carried out in the application layer. Therefore, demodulation, decoding in the physical layer, and decryption in the upper layer are required. However, in our AirCon protocol, the consensus procedures are conducted in the physical layer, and the demodulation/decoding procedures are not required using the proposed hash consistency verification method. Therefore, the computational complexity only comes from $\mathbb{O}(K)$ multiplication operations. In summary, compared with the traditional PBFT protocol, the AirCon protocol has much lower communication overhead and computational complexity. ## 6 Performance Evaluation In this section, we show the implementation of the AirCon testbed and provide experimental results to demonstrate the performance under the real-world testbed. Due to the limitation of testbed conditions, we also present simulation results to evaluate the performance of the AirCon protocol under more general channel conditions. ### 6.1 AirCon testbed design We implement the AirCon protocol based on srsLTE (version 20.04), an open- source LTE platform consisting of complete UE/eNodeB protocol stacks and a lightweight core network (CN) protocol stack. The software defines radio (SDR) board USRP B210[48] is used to transmit and receive RF signals. Based on the srsLTE platform, the following modules are re-used for AirComp implementation: (1) user attaching procedures, which guarantees all users are synchronized to the eNodeB; (2) The OFDM waveform generation, which can generate waveform based on the LTE frame structure. We then develop all other modules related to the AirCon protocol, which include: (1) reference symbols, hash bits generation, and lattice code modulation; (2) channel estimation module; (3) downlink feedback module; (4) hash consistency verification module and (5) AirCon protocol control module. All devices of the testbed and setup of the test environment are shown in Fig. 12. Totally eight USRP boards are used in our experiment, where one board is used as BS and the other seven boards are used as users. All these USRP boards are driven by four PCs via USB3.0 interfaces. For each experiment setup, the consensus procedures are repeated 1000 times. For a given number of $m$ users with consistent hash symbols, we define consensus error ratio ($\text{CER}(m)$) as a measure of consensus performance, which is given by: $\text{CER}(m)=\frac{\text{Number of consensus errors}}{\text{Total number of consensuses}}$ (28) whereby a consensus error corresponds to one of the following two cases: (1) The consensus is achieved when $m\leq\lfloor{\frac{K}{2}}\rfloor$; (2) The consensus is not achieved when $m>\lfloor{\frac{K}{2}}\rfloor$. Figure 12: Testbed components and environment ### 6.2 Testbed results Firstly, we demonstrate the symbol superposition results of the AirComp technique. As shown in Fig. 13, if all users transmit the same hash symbols, the received symbols at the BS are scaled with the number of users. On the contrary, as shown in Fig. 14, if all users transmit different hash symbols, the superposed symbols should be distributed on all lattice nodes randomly. From these two figures, it can be seen that the channel fading is effectively eliminated using the proposed AirComp implementation schemes (in particular the channel estimation, feedback, and pre-compensation schemes), and the desired channel superposition is achieved. Therefore, it is feasible to apply the AirComp technology in a digital communication system. Figure 13: Symbol superposition when all users transmit same hash symbols, the received symbols are scaled with the number of users Figure 14: Symbol superposition when three users transmit different hash symbols, the received symbols fall on all fine lattice points Secondly, we show the superposed hash symbols at the BS at different consensus stages. In Fig. 15, we show the case that all users transmit consistent hash symbols. It can be seen that all users can get into the Commit and Reply stages. In the second case, only four users transmit consistent hash symbols, while the other three users transmit random hash symbols. As shown in Fig. 16, these three users with random hash symbols are filtered in the Commit stage, and only four users with consistent hash symbols enter the Reply stage. (a) _Prepare_ stage (b) _Commit_ stage (c) _Reply_ stage Figure 15: Seven users transmit consistent hash symbols, all users determine the final consensus result (a) _Prepare_ stage (b) _Commit_ stage (c) _Reply_ stage Figure 16: Four users transmit consistent hash symbols, three users transmit random hash symbols, those users with random hash symbols are filtered The performance of the AirCon protocol are shown in Fig. 17 and Fig. 18 with different users number ($K$). For $K=5$, UE5 and UE6 in Fig. 12 are excluded in the experiments. It can be observed that when the number of users with consistent hash symbols ($m$) is close to the threshold for reaching a consensus ($m=3$ for $K=5$ and $m=4$ for $K=7$), it may lead to consensus error. However, even in the worst case, the consensus error is lower than $1\%$, which demonstrates that the proposed AirCon protocol is reliable in practice. By carefully examining the experimental results, we observe that the consensus error is mainly caused by inaccurate downlink feedback. Figure 17: Consensus error ratio for real testbed result with $K=5$ Figure 18: Consensus error ratio for real testbed result with $K=7$ ### 6.3 AirCon simulation setup We further evaluate the performance of the proposed AirCon protocol under more general conditions using simulations. In the simulation, the frame structure is the same as the practical LTE system with a 1.92MHz sampling rate. The channel gains for different users are generated independently. In our simulation, three wireless channels are considered: 1) AWGN; 2) flat fading; 3) EPA (Extended Pedestrian A model), which is a multi-path channel model defined in the 3GPP TS 36.104[50] for the typical pedestrian wireless environments. The path delay and the corresponding delay power of the EPA channel are shown in Table I. We introduce the average CER (ACER) as a metric to compare the performance difference among different simulation setups, which is given by $\text{ACER}=\frac{1}{K}\sum_{m=1}^{K}\text{CER}(m)$ (29) TABLE I: Parameters of EPA channel Delay (ns) | 0 | 30 | 70 | 90 | 110 | 190 | 410 ---|---|---|---|---|---|---|--- Power (dB) | 0 | -1 | -2 | -3 | -8 | -17.2 | -20.8 ### 6.4 Simulation results In Fig. 19, we show the ACER of the AirCon protocol under different channel conditions. Firstly, it can be seen that the consensus performance is degraded among all types of channels when SNR decreases. Secondly, the performance gap between different types of channels is increasing in the low-SNR region, which suggests that the channel estimation accuracy loss is the dominant factor of the consensus performance loss in the low-SNR region. If the channel fading is well pre-compensated (e.g., an AWGN case), the average consensus error ratio of AirCon is about $1\%$ even if the SNR is only 0 dB. Figure 19: Consensus error ratio under different SNR ($K=11$) In the low-SNR region, the retransmission scheme can help improve consensus accuracy. In Fig. 20, we show the consensus performance improvement under different retransmission numbers. It can be observed that the consensus error ratio decreases gradually with the increase of the retransmission number. Note that the retransmission will consume more wireless resources. Therefore, the trade-off between retransmission number and consensus performance needs to be considered in practice. Figure 20: Consensus error ratio under different repetition number (SNR = 0 dB) We also study the influence of total user number ($K$) on consensus accuracy. Due to space limitation, we only show the AWGN channel case. It can be observed from Fig. 21 that the AirCon protocol performs better when $K$ increases. The improvement of consensus performance comes from two aspects: (1) The consensus error only appears when $m$ is around the threshold for reaching the consensus (namely, $m=\lceil{\frac{K}{2}}\rceil$ and $m=\lceil{\frac{K}{2}}\rceil+1$ in our simulation). Therefore, the proportion of cases that will not incur consensus error increases when the total number of users increases. (2) Even in the cases that $m=\lceil{\frac{K}{2}}\rceil$ and $m=\lceil{\frac{K}{2}}\rceil+1$, the consensus error ratio also decreases, which is illustrated in Fig. 22. That is because when more users participate in the AirComp process, the average effect of noise on each user in the superposed signal is reduced, so the consensus performance becomes better. Figure 21: Averaging consensus error ratio under different $K$ (AWGN channel) Figure 22: Consensus error ratio around threshold under different $K$ (SNR = 0 dB) ## 7 Conclusion In this paper, we have proposed the AirCon to achieve low complexity consensus for blockchain-enabled wireless networks, which is novel in that the hash symbols of all users are transmitted to the BS simultaneously over the same wireless spectrum via the AirComp and lattice coding techniques, and the consensus can be done in the physical layer without decoding the hash symbols. We have shown that the AirCon protocol can significantly reduce the transmission and computational complexity during the consensus procedure. We have shown the design of AirCon based on the universal LTE system and provided solutions for the critical issues involved in the AirComp implementation, such as channel estimation and feedback. We also have implemented AirCon based on a srsLTE testbed, and real-world measurement results demonstrate that the proposed AirCon protocol is feasible in a practical LTE system. In addition, we have also conducted extensive simulation experiments and shown the accuracy and robustness of the proposed consensus protocol under different network conditions. The proposed AirCon is inspiring for designing efficient consensus schemes over wireless networks, where a more efficient channel status acquiring scheme can bring significant performance gain for AirCon. ## References * [1] J. Wang, X. Ling, Y. Le, Y. Huang, and X. You, “Blockchain-enabled wireless communications: a new paradigm towards 6g,” _Natl. Sci. Rev._ , no. 9, p. 9, 2021. * [2] L. Zhang, H. Xu, O. Onireti, M. A. Imran, and B. Cao, “How much communication resource is needed to run a wireless blockchain network?” _IEEE Network_ , vol. 36, no. 1, pp. 128–135, 2022. * [3] S. Nakamoto, “Bitcoin: A peer-to-peer electronic cash system,” [Online]. Available: http:// bitcoin.org/bitcoin.pdf, Tech. Rep., Oct. 2008. * [4] P. Vasin, “Blackcoin’s proof-of-stake protocol v2,” [Online]. Available: http:// blackcoin.org/blackcoin-pos-protocol-v2-whitepaper.pdf, Tech. Rep. * [5] Miguel, Castro, Barbara, and Liskov, “Practical byzantine fault tolerance and proactive recovery,” _ACM Trans. Comput. Syst._ , vol. 20, no. 4, pp. 398–461, 2002. * [6] D. Ongaro and J. Ousterhout, “In search of an understandable consensus algorithm,” in _2014 USENIX Ann. Tech. Conf., USENIX ATC 2014_ , vol. 22, no. 2, 2014, pp. 305–320. * [7] G. T. Nguyen and K. Kim, “A survey about consensus algorithms used in blockchain,” _J. Inf. Process. Syst._ , vol. 14, no. 1, pp. 101–128, 2018\. * [8] B. Nazer and M. Gastpar, “Computation over multiple-access channels,” _IEEE Trans. Inf. Theory_ , vol. 53, no. 10, pp. 3498–3516, 2007. * [9] ——, “Compute-and-forward: Harnessing interference through structured codes,” _IEEE Trans. Inf. Theory_ , vol. 57, no. 10, pp. 6463–6486, 2011\. * [10] S. Sigg, P. Jakimovski, and M. Beigl, “Calculation of functions on the rf-channel for iot,” in _2012 3rd IEEE Int. Conf. the Internet of Things_ , 2012, pp. 107–113. * [11] A. Kortke, M. Goldenbaum, and S. Stańczak, “Analog computation over the wireless channel: A proof of concept,” in _SENSORS, 2014 IEEE_ , 2014, pp. 1224–1227. * [12] O. Abari, H. Rahul, and D. Katabi, “Over-the-air function computation in sensor networks,” [Online]. Available: arXiv: http:// arxiv.org/abs/1612.02307. * [13] U. Altun, S. T. Başaran, H. Alakoca, and G. K. Kurt, “A testbed based verification of joint communication and computation systems,” in _2017 25th Telecommunication Forum (TELFOR)_ , 2017, pp. 1–4. * [14] “Open source sdr lte software suite,” [Online; accessed 18-April-2022]. Available: http:// github.com/srsRAN/. * [15] K. Christidis and M. Devetsikiotis, “Blockchains and smart contracts for the internet of things,” _IEEE Access_ , vol. 4, pp. 2292–2303, 2016. * [16] V. Ortega, F. Bouchmal, and J. F. Monserrat, “Trusted 5g vehicular networks: Blockchains and content-centric networking,” _IEEE Veh. Technol. Mag._ , vol. 13, no. 2, pp. 121–127, 2018. * [17] X. Ling, J. Wang, T. Bouchoucha, B. C. Levy, and Z. Ding, “Blockchain radio access network (b-ran): Towards decentralized secure radio access paradigm,” _IEEE Access_ , vol. 7, pp. 9714–9723, 2019. * [18] Y. Le, X. Ling, J. Wang, and Z. Ding, “Prototype design and test of blockchain radio access network,” in _2019 IEEE Int. Conf. Communications Workshops (ICC Workshops)_ , 2019, pp. 1–6. * [19] X. Ling, Y. Le, J. Wang, and Z. Ding, “Hash access: Trustworthy grant-free iot access enabled by blockchain radio access networks,” _IEEE Network_ , vol. 34, no. 1, pp. 54–61, 2020. * [20] B. Zhang, X. Ling, Y. Le, J. Wang, C. Cai, and Z. Tang, “Analysis and evaluation of hash access for blockchain radio access networks,” in _2020 Int. Conf. Wireless Communications and Signal Processing (WCSP)_ , 2020, pp. 62–67. * [21] X. Ling, Y. Le, J. Wang, Z. Ding, and X. Gao, “Practical modeling and analysis of blockchain radio access network,” _IEEE Trans. Commun._ , vol. 69, no. 2, pp. 1021–1037, 2021. * [22] S. Guo, Y. Dai, S. Guo, X. Qiu, and F. Qi, “Blockchain meets edge computing: Stackelberg game and double auction based task offloading for mobile blockchain,” _IEEE Trans. Veh. Technol._ , vol. 69, no. 5, pp. 5549–5561, 2020. * [23] Y. Liu, K. Wang, Y. Lin, and W. Xu, “Lightchain: A lightweight blockchain system for industrial internet of things,” _IEEE Trans. Ind. Inf._ , vol. 15, no. 6, pp. 3571–3581, 2019. * [24] H. Wei, W. Feng, Y. Chen, C.-X. Wang, and N. Ge, “Rethinking blockchains in the internet of things era from a wireless communication perspective,” _IEEE Network_ , vol. 34, no. 6, pp. 24–30, 2020. * [25] B. Cao, M. Li, L. Zhang, Y. Li, and M. Peng, “How does csma/ca affect the performance and security in wireless blockchain networks,” _IEEE Trans. Ind. Inf._ , vol. 16, no. 6, pp. 4270–4280, 2020. * [26] Y. Sun, L. Zhang, G. Feng, B. Yang, B. Cao, and M. A. Imran, “Blockchain-enabled wireless internet of things: Performance analysis and optimal communication node deployment,” _IEEE Internet Things J._ , vol. 6, no. 3, pp. 5791–5802, 2019. * [27] M. Gastpar, “Uncoded transmission is exactly optimal for a simple gaussian “sensor” network,” _IEEE Trans. Inf. Theory_ , vol. 54, no. 11, pp. 5247–5251, 2008. * [28] A. B. Wagner, S. Tavildar, and P. Viswanath, “Rate region of the quadratic gaussian two-encoder source-coding problem,” _IEEE Trans. Inf. Theory_ , vol. 54, no. 5, pp. 1938–1961, 2008. * [29] R. Soundararajan and S. Vishwanath, “Communicating linear functions of correlated gaussian sources over a mac,” _IEEE Trans. Inf. Theory_ , vol. 58, no. 3, pp. 1853–1860, 2012. * [30] X. Cao, G. Zhu, J. Xu, and K. Huang, “Optimized power control for over-the-air computation in fading channels,” _IEEE Trans. Wireless Commun._ , vol. 19, no. 11, pp. 7498–7513, 2020. * [31] W. Liu, X. Zang, Y. Li, and B. Vucetic, “Over-the-air computation systems: Optimization, analysis and scaling laws,” _IEEE Trans. Wireless Commun._ , vol. 19, no. 8, pp. 5488–5502, 2020. * [32] L. Chen, N. Zhao, Y. Chen, F. R. Yu, and G. Wei, “Over-the-air computation for iot networks: Computing multiple functions with antenna arrays,” _IEEE Internet Things J._ , vol. 5, no. 6, pp. 5296–5306, 2018. * [33] G. Zhu and K. Huang, “Mimo over-the-air computation for high-mobility multimodal sensing,” _IEEE Internet Things J._ , vol. 6, no. 4, pp. 6089–6103, 2019. * [34] J. Dong, Y. Shi, and Z. Ding, “Blind over-the-air computation and data fusion via provable wirtinger flow,” _IEEE Trans. Signal Process._ , vol. 68, pp. 1136–1151, 2020. * [35] K. Yang, T. Jiang, Y. Shi, and Z. Ding, “Federated learning via over-the-air computation,” _IEEE Trans. Wireless Commun._ , vol. 19, no. 3, pp. 2022–2035, 2020. * [36] G. Zhu, Y. Du, D. Gündüz, and K. Huang, “One-bit over-the-air aggregation for communication-efficient federated edge learning: Design and convergence analysis,” _IEEE Trans. Wireless Commun._ , vol. 20, no. 3, pp. 2120–2135, 2021. * [37] M. Mohammadi Amiri and D. Gündüz, “Machine learning at the wireless edge: Distributed stochastic gradient descent over-the-air,” _IEEE Trans. Signal Process._ , vol. 68, pp. 2155–2169, 2020. * [38] G. Zhu, J. Xu, K. Huang, and S. Cui, “Over-the-air computing for wireless data aggregation in massive iot,” _IEEE Wireless Commun._ , vol. 28, no. 4, pp. 57–65, 2021. * [39] F. Molinari, S. Stanczak, and J. Raisch, “Exploiting the superposition property of wireless communication for average consensus problems in multi-agent systems,” in _2018 European Control Conference (ECC)_ , 2018, pp. 1766–1772. * [40] F. Molinari, N. Agrawal, S. Stańczak, and J. Raisch, “Max-consensus over fading wireless channels,” _IEEE Trans. Control Network Syst._ , vol. 8, no. 2, pp. 791–802, 2021. * [41] O. Abari, H. Rahul, D. Katabi, and M. Pant, “Airshare: Distributed coherent transmission made seamless,” in _2015 IEEE Conference on Computer Communications (INFOCOM)_ , 2015, pp. 1742–1750. * [42] L. Natarajan, Y. Hong, and E. Viterbo, “Lattice codes achieve the capacity of common message gaussian broadcast channels with coded side information,” _IEEE Trans. Inf. Theory_ , vol. 64, no. 3, pp. 1481–1496, 2018. * [43] R. L. Rivest, “The MD5 Message-Digest Algorithm,” RFC 1321, Apr. 1992. * [44] 3GPP, “Evolved Universal Terrestrial Radio Access (E-UTRA); Physical layer procedures,” 3rd Generation Partnership Project (3GPP), Technical Specification (TS) 36.213, 03 2017, version 14.2.0. [Online]. Available: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=2427 * [45] Y. Zhao and A. Huang, “A novel channel estimation method for ofdm mobile communication systems based on pilot signals and transform-domain processing,” in _1997 IEEE 47th Vehicular Technology Conf. Technology in Motion_ , vol. 3, 1997, pp. 2089–2093. * [46] O. Edfors, M. Sandell, J.-J. van de Beek, S. Wilson, and P. Borjesson, “Ofdm channel estimation by singular value decomposition,” _IEEE Trans. Commun._ , vol. 46, no. 7, pp. 931–939, 1998. * [47] X. Dong, W.-s. Lu, and A. C. Soong, “Linear interpolation in pilot symbol assisted channel estimation for ofdm,” _IEEE Trans. Wireless Commun._ , vol. 6, no. 5, pp. 1910–1920, 2007. * [48] “Ettus research usrp b210,” [Online; accessed 18-April-2022]. Available: http:// www.ettus.com/all-products/UB210-KIT/. * [49] K. Baddour and N. Beaulieu, “Autoregressive modeling for fading channel simulation,” _IEEE Trans. Wireless Commun._ , vol. 4, no. 4, pp. 1650–1662, 2005. * [50] 3GPP, “Evolved Universal Terrestrial Radio Access (E-UTRA); Base Station (BS) radio transmission and reception,” 3rd Generation Partnership Project (3GPP), Technical Specification (TS) 36.104, 03 2017, version 14.3.0. [Online]. Available: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=2412
# A $3\times 3$ Lax form for the $q$-Painlevé equation of type $E_{6}$ Kanam Park National Institude of Technology, Toba College, 1-1, Ikegami, Toba, Mie, 517-0012, Japan Email<EMAIL_ADDRESS> ###### Abstract For the $q$-Painlevé equation with affine Weyl group symmetry of type $E_{6}^{(1)}$, a $2\times 2$ matrix Lax form and a second order scalar lax form were known. We give a new $3\times 3$ matrix Lax form and a third order scalar equation related to it. Continuous limit is also discussed. Key Words and Phrases : Lax formalism, $q$-Painlevé equation, 2020 Mathematics Subject Classification Numbers: 14H70, 34M56, 39A13. ## 1 Introduction The $q$-Painlevé equation with affine Weyl group symmetry of type $E_{6}^{(1)}$ was first discovered in [8]. The well-known form of it is as follows $\begin{array}[]{c}\cfrac{(\overline{f}g-1)(fg-1)}{f\overline{f}}=\cfrac{(g-1/a_{5})(g-1/a_{6})(g-1/a_{7})(g-1/a_{8})}{(g-a_{3})(g-a_{4})},\vspace{12pt}\\\ \cfrac{(\overline{f}g-1)(\overline{f}\overline{g}-1)}{g\overline{g}}=\cfrac{(\overline{f}-a_{5})(f-a_{6})(f-a_{7})(f-a_{8})}{(\overline{f}-a_{1}/q)(\overline{f}-a_{2}/q)},\vspace{12pt}\\\ q=\cfrac{a_{1}a_{2}}{a_{3}a_{4}a_{5}a_{6}a_{7}a_{8}},\end{array}$ ( 1.1) where $f$ and $g$ are dependent variables, $a_{1}$, $a_{2}$, $\cdots$, $a_{8}$ are parameters, $q$ is a constant and the overline symbol $"\ \bar{}\ "$ denotes the discrete time evolution. In previous works, the following Lax forms for the $q$-Painlevé equation for type $E_{6}^{(1)}$ has been obtained. In [11], a $2\times 2$ matrix Lax pair was derived as a reduction of the $q$-Garnier system. In [15], a second order scalar Lax form was obtained as a reduction from the $q$-Painlevé equation of type $E_{8}$. The relation between these Lax forms was given in [14]. In this article, we give a new Lax form with $3\times 3$ matrix Lax pair. We derive such a Lax form as a special case of the system investigated in our previous work [7]. As a result, we derive an equation which is equivalent to the $q$-Painlevé equation of type $E_{6}^{(1)}$ [3], [9]. In the previous work [7], we defined a nonlinear $q$-difference system as a connection preserving deformation of the following linear equation $\begin{array}[]{l}\Psi(qz)=\Psi(z)A(z),\quad A(z)=DX_{1}^{\varepsilon_{1}}(z)X_{2}^{\varepsilon_{2}}(z)\cdots X_{M}^{\varepsilon_{M}}(z),\end{array}$ ( 1.2) $\begin{array}[]{c}D=\mathrm{diag}[d_{1},d_{2},\cdots,d_{N}],\\\ X_{i}(z)=\mathrm{diag}[u_{1,i},u_{2,i},\cdots u_{N,i}]+\Lambda,\quad\Lambda=\begin{bmatrix}0&1&&{\bf O}\\\ &\ddots&\ddots&\\\ &&&1\\\ z&&&0\end{bmatrix},\end{array}$ ( 1.3) where the exponents are $\varepsilon_{i}=\pm 1$ $(1\leq i\leq M)$, $u_{j,i}$ ($1\leq j\leq N$) are dependent variables and $c_{i},d_{j}$ are parameters which satisfy $\prod_{j=1}^{N}u_{j,i}=c_{i}.$ ( 1.4) Since one can exchange the order of matrices $X_{i}^{\pm 1}$ by suitable rational transformations of variables $u_{j,i}$, the equation (1.2) essentially depends on $M_{+}$, $M_{-}$, where $M_{\pm}=\\#\\{\varepsilon_{i}|\varepsilon_{i}=\pm 1\\}$. The contents of this paper is as follows. In section 2, we set up a linear $q$-difference equation (2.1), which is a case of $(M_{+},M_{-},N)=(3,0,3)$ for the equation (1.2). And we discuss about its two deformations. One deformation gives rise to a well known form of the $q$-Painlevé equation of type $E_{6}^{(1)}$, and the other deformation gives an equation for a non- standard direction. In section 3, we derive a scalar equation from the $3\times 3$ matrix equation (2.1) and consider its characteristic properties. In section 4, we study continuous limit of our constructions and its relation to the Boalch’s Lax pair [1]. ###### Remark 1.1. The equation (1.2) in case $(M_{+},M_{-},N)=(2n+2,0,2)$ is known that it is equivalent to the linear $q$-difference equation related to the $2n$ dimensional $q$-Garnier system [10]. And in case of $(M_{+},M_{-},N)=(2,0,2n+2)$, the equation (1.2) is also equivalent to the linear $q$-difference equation related to the $2n$ dimensional system $q$-$P_{(n+1,n+1)}$ [12]. In both cases, when $n=1$, they give rise to the $q$-$P_{\rm VI}$ equation [4]. We note that the equation (2.1) is coincide to the $q$-difference linear equation of the Lax form in [13] in case of ($m$, $n$)=($3$, $1$). ## 2 A $3\times 3$ matrix Lax form In this section, we consider two types of deformations for the linear $q$-difference equation (1.2) in a case $(M_{+},M_{-},N)=(3,0,3)$. We consider the connection preserving deformation for the following $q$-difference equation for an unknown function $\Psi(z)=[\Psi_{1}(z),\Psi_{2}(z),\Psi_{3}(z)]$, $\begin{array}[]{c}\Psi(qz)=\Psi(z)A(z)=\Psi(z)A(z,t),\vspace{6pt}\\\ A(z)=DX_{1}(z)X_{2}(z)X_{3}(z)=\begin{bmatrix}b_{1}d_{1}&*&*\\\ 0&b_{2}d_{2}&*\\\ 0&0&b_{3}d_{3}\end{bmatrix}+\begin{bmatrix}d_{1}&0&0\\\ *&d_{2}&0\\\ *&*&d_{3}\end{bmatrix}z,\end{array}$ ( 2.1) where the matrices $D$ and $X_{i}(z)$ $(1\leq i\leq 3)$ stand for $\begin{array}[]{c}D=\mathrm{diag}[d_{1},d_{2},d_{3}],\vspace{6pt}\\\ X_{i}(z)=\mathrm{diag}[u_{1,i},u_{2,i},u_{3,i}]+\Lambda,\vspace{6pt}\\\ \Lambda=\begin{bmatrix}0&1&0\\\ 0&0&1\\\ z&0&0\end{bmatrix},\end{array}$ ( 2.2) and $u_{j,i}$ ($1\leq i$, $j\leq 3$) are dependent variables and $c_{i},d_{j}$ are parameters which satisfy $\begin{array}[]{c}\prod_{i=1}^{3}u_{j,i}=b_{j}\quad(1\leq j\leq 3),\quad\prod_{j=1}^{3}u_{j,i}=c_{i}\quad(1\leq i\leq 3),\vspace{6pt}\\\ b_{1}b_{2}b_{3}=c_{1}c_{2}c_{3}.\end{array}$ ( 2.3) The first equation in (2.3) is equivalent to that the characteristic exponents at $z=0$ of the equation (2.1) are $b_{j}d_{j}$. The second equation in (2.3) is equivalent to the following condition $|A(z)|=d_{1}d_{2}d_{3}(z+c_{1})(z+c_{2})(z+c_{3}).$ ( 2.4) Through a gauge transformation by a $3\times 3$ diagonal matrix we can take two components in (2.1) as 1. In the following, we use this kind of gauge fixings in case by case. By the condition (2.4), two of the remaining four components are determined by other components and parameters $b_{j},c_{i},d_{j}$. In this article, we will consider two deformations $T_{1}$ and $T_{2}$ for the equation (2.1) which act on parameters $b_{j},\ c_{i},\ d_{j}$ as: $\begin{array}[]{l}T_{1}:({b_{1}},{b_{2}},{b_{3}},{c_{1}},{c_{2}},{c_{3}},{d_{1}},{d_{2}},{d_{3}})\to(b_{1},qb_{2},\cfrac{b_{3}}{q},c_{1},c_{2},c_{3},qd_{1},d_{2},qd_{3}),\vspace{6pt}\\\ T_{2}:({b_{1}},{b_{2}},{b_{3}},{c_{1}},{c_{2}},{c_{3}},{d_{1}},{d_{2}},{d_{3}})\to(\cfrac{b_{1}d_{1}}{d_{2}},\cfrac{b_{2}d_{2}}{d_{3}},b_{3},c_{1},c_{2},c_{3}q,d_{2},d_{3},\cfrac{d_{1}}{q}).\end{array}$ ( 2.5) ### 2.1 The deformation $T_{1}$ We will show that the deformation $T_{1}$ gives rise to the standard form of the $q$-$E_{6}^{(1)}$. We consider the following $q$-difference linear equation $\begin{array}[]{c}\Psi(qz)=\Psi(z)A(z),\quad A(z)=\begin{bmatrix}b_{1}d_{1}&1&v_{1}\\\ 0&b_{2}d_{2}&1\\\ 0&0&b_{3}d_{3}\end{bmatrix}+\begin{bmatrix}d_{1}&0&0\\\ v_{2}&d_{2}&0\\\ v_{3}&v_{4}&d_{3}\end{bmatrix}z,\end{array}$ ( 2.6) $|A(z)|=d_{1}d_{2}d_{3}(z+c_{1})(z+c_{2})(z+c_{3}).$ ( 2.7) We take a deformation equation for the equation (2.6), (2.7) as follows: $\begin{array}[]{c}T_{1}\Psi(z)=\Psi(z)B(z),\quad B(z)=\begin{bmatrix}0&0&w_{1}\\\ 0&0&w_{2}\\\ 0&0&w_{3}\end{bmatrix}+\begin{bmatrix}w_{4}&0&0\\\ w_{5}&0&0\\\ w_{6}&1&w_{7}\end{bmatrix}z,\vspace{6pt}\\\ |B(z)|=(w_{1}w_{5}-w_{2}w_{4})z^{2},\end{array}$ ( 2.8) $T_{1}:({b_{1}},{b_{2}},{b_{3}},{c_{1}},{c_{2}},{c_{3}},{d_{1}},{d_{2}},{d_{3}};v_{1},v_{4})\to(b_{1},qb_{2},\cfrac{b_{3}}{q},c_{1},c_{2},c_{3},qd_{1},d_{2},qd_{3};\overline{v_{1}},\overline{v_{4}}),$ ( 2.9) ###### Theorem 2.1. Through a compatibility condition of the equations (2.6), (2.9), (2.8) $B(z)\overline{A(z)}=A(z)B(qz),$ ( 2.10) we obtain the following equations $\begin{array}[]{c}\cfrac{(fg-1)(\overline{f}g-1)}{f\overline{f}}=\cfrac{(g-1/c_{1})(g-1/c_{2})(g-1/c_{3})(g-d_{2}/b_{3}d_{3})}{(g-1/b_{2})(g-d_{2}/b_{1}d_{1})},\vspace{6pt}\\\ \cfrac{(\overline{f}g-1)(\overline{f}\overline{g}-1)}{g\overline{g}}=\cfrac{(\overline{f}-c_{1})(\overline{f}-c_{2})(\overline{f}-c_{3})(\overline{f}-b_{3}d_{3}/d_{2})}{(\overline{f}-b_{3}d_{3}/d_{1})(\overline{f}-b_{3}/q)},\end{array}$ ( 2.11) where, $\begin{array}[]{c}f=\cfrac{b_{3}d_{3}}{d_{3}-v_{1}v_{4}},\quad g=\cfrac{d_{2}v_{1}}{b_{2}d_{2}v_{1}-1},\end{array}$ ( 2.12) and $\overline{*}$ stands for $T_{1}(*)$. The equation (2.11) is the well known form of the $q$-Painlevé equation of type $E_{6}$ [3], [9] (See also [5]). ###### Proof. The result is obtained by a direct computation of the compatibility condition of (2.10). Since the computation is rather heavy, we will give a comment how to do it efficiently. Though the $2$ variables $v_{2}$, $v_{3}$ can be represented by the rational functions of the remaining two variables $v_{1}$, $v_{4}$ by the relation (2.7), it is more efficient to do this elimination after the calculation of the compatibility condition (2.10) in $4$ variables, and then reduce it to 2 variables. In this way, we get the following time evolutions for $v_{1}$, $v_{4}$ as rational functions of $v_{1}$, $v_{4}$: $\begin{array}[]{c}\overline{v_{1}}=\cfrac{C_{1}(v_{1},v_{4})C_{2}(v_{1},v_{4})}{D_{1}(v_{1},v_{4})D_{2}(v_{1},v_{4})},\quad\overline{v_{4}}=\cfrac{C_{3}(v_{1},v_{4})}{D_{3}(v_{1},v_{4})D_{4}(v_{1},v_{4})},\end{array}$ ( 2.13) where $\begin{array}[]{rl}C_{3}(v_{1},v_{4}),D_{2}(v_{1},v_{4})&:\text{polynomials in }v_{1},v_{4}\text{ of degree }(4,\ 3),\\\ C_{2}(v_{1},v_{4})&:\text{a polynomial in }v_{1},v_{4}\text{ of degree }(3,\ 1),\\\ C_{1}(v_{1},v_{4}),D_{3}(v_{1},v_{4}),D_{4}(v_{1},v_{4})&:\text{polynomials in }v_{1},v_{4}\text{ of degree }(2,\ 1),\\\ D_{1}(v_{1},v_{4})&:\text{a polynomial in }v_{1},v_{4}\text{ of degree }(1,\ 0).\end{array}$ ( 2.14) The remaining task is to rewrite the equation (2.13) to (2.11). A useful way to solve it is to look at the singularities of the equations [2]. Namely, we investigate the points at which the right hand side of the equations (2.13) are indeterminate. We focus on the equation $D_{3}(v_{1},\ v_{4})$ of them $D_{3}(v_{1},v_{4})=d_{3}\left(-b_{2}d_{2}v_{1}+b_{3}d_{2}v_{1}+1\right)+v_{1}v_{4}\left(b_{2}d_{2}v_{1}-1\right).$ ( 2.15) Investigating common zero points of the equation (2.15) and the other polynomials $C_{k}(v_{1},\ v_{4})$, $D_{l}(v_{1},\ v_{4})$, we find 4 indeterminate points as follows: $\begin{array}[]{c}(v_{1},v_{4})=\big{(}-\cfrac{1}{(u^{-1}-b_{2})d_{2}},-\cfrac{(u^{-1}-b_{3})(u^{-1}-b_{2})d_{3}d_{2}}{u^{-1}}\big{)},\vspace{6pt}\\\ (u=c_{1}^{-1},c_{2}^{-1},c_{3}^{-1},d_{2}/b_{3}d_{3}).\end{array}$ ( 2.16) In view of the form of the points in (2.16), we define the variables $f$, $g$ as follows $(v_{1},v_{1}v_{4})=\big{(}-\cfrac{1}{(g^{-1}-b_{2})d_{2}},\cfrac{(f-b_{3})d_{3}}{f}\big{)}.$ ( 2.17) By the transformation (2.17), the equation $D_{3}(v_{1},v_{4})=0$ is transformed to an equation $fg=1$. Through the correspondence (2.17) and the time evolution equations for the variables $v_{1}$, $v_{4}$ (2.13), we have time evolution equations for $f$, $g$ (2.11). ∎ ### 2.2 The deformation $T_{2}$ In this subsection, we take a deformation equation which is one of that considered in the previous work [7]. It corresponded to the permutations of the matrices $X_{i}(z)^{\pm 1}$. We consider the following Lax pair $\begin{array}[]{c}\Psi(qz)=\Psi(z)A(z),\quad A(z)=DX_{1}(z)X_{2}(z)X_{3}(z)=\begin{bmatrix}b_{1}d_{1}&1&*\\\ 0&b_{2}d_{2}&1\\\ 0&0&b_{3}d_{3}\end{bmatrix}+\begin{bmatrix}d_{1}&0&0\\\ *&d_{2}&0\\\ *&*&d_{3}\end{bmatrix}z,\end{array}$ ( 2.18) $\begin{array}[]{c}T_{2}{\Psi}(z)=\Psi(z)B(z),\quad B(z)=X_{3}(z/q)^{-1},\vspace{6pt}\\\ |B(z)|=\cfrac{q}{z+qc_{3}},\vspace{6pt}\\\ T_{2}:({b_{1}},{b_{2}},{b_{3}},{c_{1}},{c_{2}},{c_{3}},{d_{1}},{d_{2}},{d_{3}};x,y)\to(\cfrac{b_{1}d_{1}}{d_{2}},\cfrac{b_{2}d_{2}}{d_{3}},b_{3},c_{1},c_{2},c_{3}q,d_{2},d_{3},\cfrac{d_{1}}{q};\overline{x},\overline{y}),\end{array}$ ( 2.19) where we define variables $x$, $y$ and gauge freedom $w_{1}$, $w_{2}$ with the variables $u_{j,i}$ in (2.18) as follows: $x=\frac{u_{1,1}u_{1,2}u_{2,1}u_{2,2}}{u_{1,1}+u_{2,2}},\quad y=\frac{1}{u_{1,1}u_{1,2}\left(u_{2,1}+u_{3,2}\right)},\quad w_{1}=u_{1,1},\quad w_{2}=u_{1,3}.$ ( 2.20) ###### Theorem 2.2. Through a compatibility condition of the equations (2.18), (2.19), (2.20), $B(z)\overline{A(z)}=A(z)B(qz),$ ( 2.21) we obtain the following equations $\begin{array}[]{c}\cfrac{\overline{x}+c_{1}}{\overline{x}+c_{2}}=\cfrac{E_{1}E_{2}}{E_{3}E_{4}},\quad\cfrac{c_{1}\overline{y}+1}{c_{2}\overline{y}+1}=\cfrac{F_{1}F_{2}}{F_{3}F_{4}},\end{array}$ ( 2.22) where, $\begin{array}[]{l}E_{1}=b_{1}d_{1}\left(b_{2}d_{2}(1-xy)+c_{1}d_{3}y\left(c_{2}+x\right)\right)+c_{1}d_{2}d_{3}x\left(c_{2}y+1\right),\\\ E_{2}=b_{1}d_{1}\left(b_{2}d_{2}(1-xy)+c_{3}d_{3}qy\left(c_{1}+x\right)\right)+c_{3}d_{2}d_{3}qx\left(c_{1}y+1\right),\\\ E_{3}=b_{1}d_{1}\left(b_{2}d_{2}(1-xy)+c_{2}d_{3}y\left(c_{1}+x\right)\right)+c_{2}d_{2}d_{3}x\left(c_{1}y+1\right),\\\ E_{4}=b_{1}d_{1}\left(b_{2}d_{2}(1-xy)+c_{3}d_{3}qy\left(c_{2}+x\right)\right)+c_{3}d_{2}d_{3}qx\left(c_{2}y+1\right),\\\ F_{1}=E_{1},F_{2}=b_{1}d_{1}\left(b_{2}d_{2}(1-xy)+c_{2}d_{3}y\left(c_{1}+x\right)\right)+c_{3}d_{2}d_{3}qx\left(c_{1}y+1\right),\\\ F_{3}=E_{3},F_{4}=b_{1}d_{1}\left(b_{2}d_{2}(1-xy)+c_{1}d_{3}y\left(c_{2}+x\right)\right)+c_{3}d_{2}d_{3}qx\left(c_{2}y+1\right),\end{array}$ ( 2.23) and $\overline{*}$ stands for $T_{2}(*)$. ###### Proof. Solving a compatibility condition (2.21) with (2.20) for the variables $\overline{x}$, $\overline{y}$, we obtain the following equations $\overline{x}=\cfrac{(xy-1)G_{1}(x,y)}{H_{1}(x,y)},\quad\overline{y}=\cfrac{xG_{2}(x,y)}{H_{2}(x,y)},$ ( 2.24) where $G_{k}(x,y)$, $H_{k}(x,y)$ $(k=1,2)$ are polynomials in variables $x$, $y$. The polynomial $G_{1}(x,y)$ is of degree $(1,1)$, $G_{2}(x,y)$ is of degree $(1,2)$ and $H_{k}(x,y)$ are of degree $(2,2)$. Using a method [5] for finding point configuration, a configuration of points for the equation (2.22) is as follows: $\begin{array}[]{rl}(x,y)=&(-c_{1},-\cfrac{1}{c_{1}}),(-c_{2},-\cfrac{1}{c_{2}}),(-\cfrac{b_{1}d_{1}}{d_{2}},-\cfrac{d_{2}}{b_{1}d_{1}}),(-\cfrac{b_{1}b_{2}}{c_{3}},-\cfrac{c_{3}}{b_{1}b_{2}}),\\\ &(-b_{2},0),(-\cfrac{b_{1}b_{2}d_{1}}{c_{3}d_{3}q},0),\\\ &(0,-\cfrac{1}{b_{1}}),(0,-\cfrac{b_{2}d_{2}}{c_{1}c_{2}d_{3}}).\end{array}$ ( 2.25) Calculating $\cfrac{\overline{x}+c_{1}}{\overline{x}+c_{2}}$ and $\cfrac{c_{1}\overline{y}+1}{c_{2}\overline{y}+1}$ respectively, we have equations (2.22), (2.23). ∎ ###### Remark 2.1. Here, the time evolution equations $\overline{u_{j,i}}$ are derived by solving the compatibility condition (2.21). For another derivation using the transformations which correspond to permutations of the matrices $X_{i}^{\varepsilon_{i}}(z)$, see [7] Proposition 2.1. Before ending this subsection, we show a relation between pairs of the variables $(f,g)$ in (2.11) and $(x,y)$ in (2.22). ###### Proposition 2.1. The Lax equations (2.6) with (2.12) and (2.18) with (2.20) are equivalent with each other if the variables $f$, $g$ and $x$, $y$ are related as: $\begin{array}[]{l}f=\cfrac{c_{1}c_{2}c_{3}\left(b_{2}+x\right)(xy-1)}{b_{1}b_{2}y\left(c_{1}+x\right)\left(c_{2}+x\right)+c_{3}x\left(b_{2}(1-xy)+c_{2}xy+c_{1}y\left(c_{2}+x\right)+x\right)},\vspace{6pt}\\\ g=-\cfrac{y\left(b_{1}b_{2}+c_{3}x\right)}{c_{3}\left(b_{2}(1-xy)+x\right)+b_{1}b_{2}xy}\end{array}$ ( 2.26) ###### Proof. Comparing the coefficient matrix $A(z)$ of the equation (2.6), (2.12) to the coefficient matrix $A(z)$ of the equation (2.18), (2.20), we can solve for the variables $x$, $y$ and gauge freedom $w_{1}$, $w_{2}$ in terms of variables $f$, $g$ and parameters $b_{j}$, $c_{i}$, $d_{j}$ ($1\leq i,j\leq 3$). Then we have the desired relation between pairs of variables $(f,g)$ and $(x,y)$ (2.26). ∎ ## 3 A scalar equation related to the $q$-$E_{6}^{(1)}$ In this section, we derive a scalar $q$-difference equation from the matrix $q$-difference equation (2.18), (2.20) for an unknown function $\Psi(z)=[\Psi_{1}(z),\Psi_{2}(z),\Psi_{3}(z)]$ and its geometric properties. Eliminating functions $\Psi_{2}(z)$, $\Psi_{3}(z)$ in the equation (2.18), (2.20), we obtain the following third linear $q$-difference equation for $\Psi_{1}(z)=:\Phi(z)$ $L(z):=P_{3}(z)\Phi(q^{3}z)+P_{2}(z)\Phi(q^{2}z)+P_{1}(z)\Phi(qz)+P_{0}(z)\Phi(z)=0,$ ( 3.1) where $\begin{array}[]{l}P_{3}(z)=p_{31}(z-u),\\\ P_{2}(z)=p_{22}z^{2}+p_{21}z+p_{20},\\\ P_{1}(z)=p_{13}z^{3}+p_{12}z^{2}+p_{11}z+p_{10},\\\ P_{0}(z)=-P_{3}(qz)d_{1}d_{2}d_{3}(z+c_{1})(z+c_{2})(z+c_{3}).\end{array}$ ( 3.2) Here, the coefficients $p_{k,l}$ $(1\leq k\leq 3,0\leq l\leq 3)$ in the polynomials $P_{k}(z)$ (3.2) depend on parameters $b_{j}$, $c_{i}$, $d_{j}$, and a variable $u$ defined as the zero of $P_{3}(z)$. The variable $u$ is expressed in terms of $x$, $y$ as follows $\begin{array}[]{l}u=\cfrac{y\left(b_{1}\left(c_{2}xy+c_{1}y\left(c_{2}+x\right)+x\right)+c_{1}c_{2}(1-xy)\right)}{b_{2}\left(b_{1}y+1\right)(xy-1)}\vspace{6pt}\\\ \cfrac{\left(b_{2}d_{3}y\left(b_{1}\left(c_{2}xy+c_{1}y\left(c_{2}+x\right)+x\right)+c_{1}c_{2}(1-xy)\right)+b_{2}^{2}d_{2}\left(b_{1}y+1\right)(1-xy)+c_{3}d_{3}x\left(c_{1}y+1\right)\left(c_{2}y+1\right)\right)}{\left(d_{2}\left(b_{2}\left(b_{1}y+1\right)(xy-1)-x\left(c_{1}y+1\right)\left(c_{2}y+1\right)\right)-d_{3}y\left(b_{1}\left(c_{2}xy+c_{1}y\left(c_{2}+x\right)+x\right)+c_{1}c_{2}(1-xy)\right)\right)}.\end{array}$ ( 3.3) Explicit forms of the polynomials $P_{j}(z)$ ($0\leq j\leq 3$) (3.2) are given in Appendix. Then we have ###### Lemma 3.1. The equation $L(z)=0$ (3.1) has the following properties: 1. (i) It is a linear four term equation between $\Phi(q^{j}z)$ ($0\leq j\leq 3$) and its coefficients $P_{j}(z)$ are polynomials for $z$ of degree $4-j$, 2. (ii) A polynomial $P_{0}(z)$ has four zero points at $z=-c_{i}$ ($1\leq i\leq 3$), $u/q$, 3. (iii) The exponents of solutions $\Phi(z)$ are $b_{1}d_{1}$, $qb_{2}d_{2}$, $qb_{3}d_{3}$ (at $z=0$) and $d_{1}$, $d_{2}$, $d_{3}$ (at $z=\infty$), 4. (iv) A point $z=u$ such that $P_{3}(z)=0$ is an apparent singularity, namely we have $v:=\cfrac{P_{0}(u)}{P_{1}(u/q)}=\cfrac{P_{1}(u)}{P_{2}(u/q)}=\cfrac{P_{2}(u)}{P_{3}(u/q)}.$ ( 3.4) Conversely, the equation $L(z)=0$ (3.1) is uniquely characterized by these properties (i)-(iv) up to normalization. ###### Proof. The properties (i)-(iv) follows by computation through eliminating $\Psi_{2}(z)$, $\Psi_{3}(z)$ in (2.18), (2.20). The converse can be confirmed that coefficients $P_{j}(z)$ are defined uniquely by (i)-(iv) up to a normalization. To see this, we consider the following equation which satisfies the properties (i), (ii): $L^{\prime}(z)=P_{3}^{\prime}(z)\Phi(q^{3}z)+P_{2}^{\prime}(z)\Phi(q^{2}z)+P_{1}^{\prime}(z)\Phi(qz)+P_{0}^{\prime}(z)\Phi(z)=0,$ ( 3.5) $\begin{array}[]{l}P_{3}^{\prime}(z)=p_{31}^{\prime}(z-u),\\\ P_{2}^{\prime}(z)=p_{22}^{\prime}z^{2}+p_{21}^{\prime}z+p_{20}^{\prime},\\\ P_{1}^{\prime}(z)=p_{13}^{\prime}z^{3}+p_{12}^{\prime}z^{2}+p_{11}^{\prime}z+p_{10}^{\prime},\\\ P_{0}^{\prime}(z)=p_{04}^{\prime}(z-u/q)(z+c_{1})(z+c_{2})(z+c_{3}).\end{array}$ ( 3.6) From the property (iii), the condition of the exponents of solutions $\Phi(z)$ at $z=0$ determines the coefficients $p_{31}^{\prime}$, $p_{20}^{\prime}$, $p_{10}^{\prime}$ and the condition of the exponents of solutions $\Phi(z)$ at $z=\infty$ determines the coefficients $p_{22}^{\prime}$, $p_{13}^{\prime}$. The remaining coefficients except for $p_{04}^{\prime}$ are determined by the property (iv). If we put the normalization factor $p_{04}^{\prime}$ as $q^{5}uvc_{1}c_{2}d_{1}d_{2}d_{3}$, the function $L^{\prime}(z)$ (3.5) equals to the function $L(z)$ (3.1). ∎ In the following, viewing $\Phi(q^{i}z)$ $(0\leq i\leq 3)$ as parameters, we regard the scalar $q$-difference equation $L(z)=0$ (3.1) as an algebraic curve in variables $(u,v)$ $\in$ $\mathbb{P}^{1}\times\mathbb{P}^{1}$. We represent as the curve as $P(u,v)=0$. The features of the curve are the following. ###### Lemma 3.2. The algebraic curve $P(u,v)=0$ has the following properties: 1. (i) The polynomial $P(u,v)$ has the following form $P(u,v)=\displaystyle\sum_{0\leq j\leq 3\atop 0\leq i+j\leq 4}c_{i,j}u^{i}v^{j},\quad c_{0,0}:=c_{0}z^{2}\Phi(qz).$ ( 3.7) The coefficients $c_{i,j}$ depend on $b_{i}$, $c_{i}$, $d_{i}$, $q$, $z$, $\Phi(q^{i}z)$. 2. (ii) It passes the following $8$ points $\begin{array}[]{c}(u,v)=(0,qb_{1}d_{1}),\quad(0,q^{2}b_{2}d_{2}),\quad(0,q^{2}b_{3}d_{3}),\quad(qz,\infty),\vspace{6pt}\\\ (z,0),\quad(-c_{1},0),\quad(-c_{2},0),\quad(-c_{3},0),\end{array}$ ( 3.8) and 3 points in the coordinate $(r,s)=(u,v/u)$ $(r,s)=(\infty,q^{2}d_{1}),\quad(\infty,q^{2}d_{2}),\quad(\infty,q^{2}d_{3}).$ ( 3.9) 3. (iii) At $u=z$ the equation $P(u,v)=0$ has the following property $\begin{array}[]{c}\displaystyle\big{(}\sum_{0\leq i\leq 3-j}c_{i,j+1}z^{i}\big{)}\Big{|}_{z\to qz,\Phi(q^{k}z)\to\Phi(q^{k+1}z)}=\sum_{0\leq i\leq 4-j}c_{i,j}(qz)^{i}\quad(j=0,1,2).\end{array}$ ( 3.10) Conversely, the equation $P(u,v)=0$ is uniquely characterized by these properties (i)-(iii) up to normalization factor $c_{0}$. ###### Proof. The properties (i)-(iii) follow for the polynomial $P(u,v)$. Conversely, we consider a polynomial $P^{\prime}(u,v):=\displaystyle\sum_{0\leq j\leq 3\atop 0\leq i+j\leq 4}c_{i,j}^{\prime}u^{i}v^{j}$, $c_{0,0}^{\prime}(z)=c_{0}^{\prime}\Phi(qz)$. The polynomial $P^{\prime}(u,v)$ has $14$ coefficients. From the property (ii), $10$ coefficients are described in terms of parameters $b_{j}$, $c_{j}$, $d_{j}$ and the coefficient $c_{0}^{\prime}$. $\begin{array}[]{l}P^{\prime}(u,v)=-\cfrac{c_{0}^{\prime}z(qz-u)\Phi(qz)}{b_{1}b_{2}b_{3}d_{1}d_{2}d_{3}q^{6}}v^{3}\vspace{6pt}\\\ +\left(uc_{1,2}^{\prime}+\cfrac{c_{0}^{\prime}z\Phi(qz)\left(b_{2}d_{2}qz+b_{3}d_{3}qz+b_{1}d_{1}z-d_{1}u^{2}-d_{2}u^{2}-d_{3}u^{2}\right)}{b_{1}b_{2}b_{3}d_{1}d_{2}d_{3}q^{4}}\right)v^{2}\vspace{6pt}\\\ +\left(u^{2}c_{2,1}^{\prime}+uc_{1,1}^{\prime}-\cfrac{c_{0}^{\prime}z\Phi(qz)\left(b_{2}b_{3}d_{2}d_{3}qz+b_{1}b_{2}d_{1}d_{2}z+b_{1}b_{3}d_{1}d_{3}z-d_{1}d_{2}u^{3}-d_{1}d_{3}u^{3}-d_{2}d_{3}u^{3}\right)}{b_{1}b_{2}b_{3}d_{1}d_{2}d_{3}q^{2}}\right)v\vspace{6pt}\\\ +\cfrac{c_{0}^{\prime}z\left(c_{1}+u\right)\left(c_{2}+u\right)(z-u)\left(c_{3}+u\right)\Phi(qz)}{b_{1}b_{2}b_{3}}.\end{array}$ ( 3.11) The remaining $3$ parameters $c_{1,1}^{\prime}$, $c_{1,2}^{\prime}$, $c_{2,1}^{\prime}$ are determined in terms of parameters $b_{i}$, $c_{i}$, $d_{i}$, $q$, $z$, $\Phi(q^{i}z)$, $c_{0}^{\prime}$ by the property (iii). Namely the property (iii) gives $3$ linear inhomogeneous equations among $c_{1,1}^{\prime}$, $c_{1,2}^{\prime}$, $c_{2,1}^{\prime}$, $c_{1,1}^{\prime}\big{|}_{z\to qz}$, $c_{1,2}^{\prime}\big{|}_{z\to qz}$, $c_{2,1}^{\prime}\big{|}_{z\to qz}$. Though these relations are apparently $q$-difference equations, we can solve them algebraically. For example, in the equation (3.10), we solve $c_{12}^{\prime}$ when $j=2$. Then when $j=0$, we solve $\left.c_{21}^{\prime}\right|_{z\to qz,\Phi(q^{k}z)\to\Phi(q^{k+1}z)}$. And finally solving $c_{11}^{\prime}$ when $j=1$, they algebraically can be solved. ∎ Explicit forms of the coefficients $c_{i,j}(z)$ of the polynomial $P(u,v)$ are in Appendix. ### 3.1 Relations among pairs of variables $(f,g)$, $(x,y)$ and $(u,v)$ ###### Proposition 3.1. Under the relations among variables ($f$, $g$), ($x$, $y$) and ($u$, $v$): $\begin{array}[]{l}(f,g)=\vspace{6pt}\\\ \left(\cfrac{c_{1}c_{2}c_{3}\left(b_{2}+x\right)(xy-1)}{b_{1}b_{2}y\left(c_{1}+x\right)\left(c_{2}+x\right)+c_{3}x\left(b_{2}(1-xy)+c_{2}xy+c_{1}y\left(c_{2}+x\right)+x\right)},-\cfrac{y\left(b_{1}b_{2}+c_{3}x\right)}{c_{3}\left(b_{2}(1-xy)+x\right)+b_{1}b_{2}xy}\right),\end{array}$ ( 3.12) $\begin{array}[]{l}(x,y)=\vspace{6pt}\\\ \Big{(}\Big{(}b_{2}v\left(-b_{2}d_{2}q^{2}-d_{3}q^{2}u+v\right)\\\ \quad\quad\left(c_{1}c_{2}d_{3}q^{2}\left(c_{3}\left(b_{2}d_{2}q^{2}-v\right)+b_{2}d_{2}q^{2}u\right)+b_{1}b_{2}\left(d_{2}q^{2}\left(-b_{2}v+d_{3}q^{2}u\left(c_{3}+u\right)-uv\right)+v\left(v-d_{3}q^{2}u\right)\right)\right)\Big{)}\\\ \cdot\Big{(}q^{2}(b_{2}^{2}d_{2}(d_{2}q^{2}\left(d_{3}q^{2}\left(c_{1}+u\right)\left(c_{3}+u\right)-uv\right)\left(d_{3}q^{2}\left(c_{2}+u\right)\left(c_{3}+u\right)-uv\right)\\\ \quad-v\left(-d_{3}q^{2}v\left(b_{1}\left(2c_{3}+u\right)+u\left(c_{3}+2u\right)\right)+d_{3}^{2}q^{4}u\left(b_{1}+u\right)\left(c_{3}+u\right)+uv^{2}\right))\\\ \quad+b_{2}^{3}d_{2}^{2}\left(-q^{2}\right)v\left(d_{3}q^{2}\left(b_{1}+u\right)\left(c_{3}+u\right)-uv\right)-b_{2}c_{3}d_{3}v(b_{1}v\left(v-d_{3}q^{2}u\right)\\\ \quad+d_{2}q^{2}\left(c_{2}u\left(d_{3}q^{2}\left(c_{3}+u\right)-v\right)+c_{1}\left(d_{3}q^{2}\left(2c_{2}+u\right)\left(c_{3}+u\right)-uv\right)\right))+c_{1}c_{2}c_{3}^{2}d_{3}^{2}q^{2}v^{2})\Big{)}^{-1},\\\ \cfrac{b_{2}d_{2}q^{2}u\left(d_{3}q^{2}\left(c_{3}+u\right)-v\right)}{c_{1}c_{2}d_{3}q^{2}\left(c_{3}\left(b_{2}d_{2}q^{2}-v\right)+b_{2}d_{2}q^{2}u\right)+b_{1}b_{2}v\left(-b_{2}d_{2}q^{2}-d_{3}q^{2}u+v\right)}\Big{)},\end{array}$ ( 3.13) the corresponding Lax equations (2.6) with (2.17), (2.18) with (2.20) and (2.18) with (3.3), (3.4) are equivalent. Conversely, such relations among $(f,g)$, $(x,y)$ and $(u,v)$ are uniquely determined as (3.12), (3.13). ###### Proof. Using the relation (3.12), we can check that the equation (2.6) with (2.17) is equivalent to (2.18) with (2.20). Similarly, using the relation (3.13), we can check that the equation (2.18) with (2.20) is equivalent to (2.18) with with (3.3), (3.4). The converse is obvious from the form of the Lax matrix. ∎ ## 4 Continuous limit In this section, we describe a relation between our result and the result of Boalch [1]. in [1], a Lax pair for the additional-difference Painlevé equation with affine Weyl symmetry group of type $E_{6}$ was described. The linear differential equation of the Lax pair is as follows $\begin{array}[]{c}\cfrac{d}{dz}\Psi(z)=\Psi(z)\left(\cfrac{A_{1}^{b}}{z}+\cfrac{A_{2}^{b}}{z-1}\right),\quad A_{3}^{b}:=-(A_{1}^{b}+A_{2}^{b}),\\\ \end{array}$ ( 4.1) where the matrices $A_{i}^{b}$ ($1\leq i\leq 3$) are $3\times 3$ matrices with different eigenvalues. We show that the linear $q$-difference equation (2.1) reduces to the equation (4.1) via a continuous limit $q\to 1$. The equation (2.1) takes the following form after a scale transformation $z\to-z$ and gauge transformations $\begin{array}[]{c}(1-z)\Psi(qz)=\Psi(z)\mathcal{A}(z),\quad\mathcal{A}(z)=\begin{bmatrix}b_{1}d_{1}&k_{1}&v_{1}\\\ 0&b_{2}d_{2}&k_{2}\\\ 0&0&b_{3}d_{3}\end{bmatrix}-\begin{bmatrix}d_{1}&0&0\\\ v_{2}&d_{2}&0\\\ v_{3}&v_{4}&d_{3}\end{bmatrix}z,\vspace{6pt}\\\ |\mathcal{A}(z)|=d_{1}d_{2}d_{3}(z-c_{1})(z-c_{2})(z-c_{3}),\end{array}$ ( 4.2) where $k_{j}$ ($j=1$, $2$) are constants. We put $q=e^{h}$ and consider the limit $h\to 0$. We set $\begin{array}[]{c}b_{i}=q^{\beta_{i}},\quad c_{i}=q^{\gamma_{i}},\quad d_{i}=q^{\delta_{i}}\quad(1\leq i\leq 3),\vspace{6pt}\\\ k_{j}=hl_{j}\quad(j=1,2),\quad v_{m}=hu_{m}\quad(1\leq m\leq 4),\end{array}$ ( 4.3) where $l_{j}$ are constants. By using Taylor’s expansion for (4.2), (4.3) $\begin{array}[]{rl}(LHS)&=(1-z)\Psi(z)+h(1-z)z\cfrac{d}{dz}\Psi(z)+\mathcal{O}(h^{2}),\vspace{6pt}\\\ (RHS)&=(1-z)\Psi(z)+h\Psi(z)\begin{bmatrix}\beta_{1}-\delta_{1}(z-1)&l_{1}&u_{1}\\\ -u_{2}z&\beta_{2}-\delta_{2}(z-1)&l_{2}\\\ -u_{3}z&-u_{4}z&\beta_{3}-\delta_{3}(z-1)\end{bmatrix}+\mathcal{O}(h^{2}),\end{array}$ ( 4.4) we find the following limit as $h\to 0$ $\begin{array}[]{c}\cfrac{d}{dz}\Psi(z)=\Psi(z)\left(\cfrac{A_{1}}{z}+\cfrac{A_{2}}{z-1}\right),\vspace{6pt}\\\ A_{1}=\begin{bmatrix}\beta_{1}+\delta_{1}&l_{1}&u_{1}\\\ 0&\beta_{2}+\delta_{2}&l_{2}\\\ 0&0&\beta_{3}+\delta_{3}\end{bmatrix},\quad A_{2}=\begin{bmatrix}-\beta_{1}&-l_{1}&-u_{1}\\\ u_{2}&-\beta_{2}&-l_{2}\\\ u_{3}&u_{4}&-\beta_{3}\end{bmatrix},\vspace{6pt}\\\ A_{3}:=-(A_{1}+A_{2})=\begin{bmatrix}\delta_{1}&0&0\\\ u_{2}&\delta_{2}&0\\\ u_{3}&u_{4}&\delta_{3}\end{bmatrix},\end{array}$ ( 4.5) where eigenvelues of the matrix $A_{2}$ are $\gamma_{i}$ ($1\leq i\leq 3$) by the condition of the determinant of the matrix $\mathcal{A}(z)$ (4.2). Therefore the linear $q$-difference equation (2.1) reduces to the equation (4.1) via a continuous limit $q\to 1$. In the following, we consider a continuous limit $q\to 1$ of the result in §2.1. In the subsection §2.1, through a compatibility condition of the equations (2.10), we derived a standard $q$-Painlevé equation of type $E_{6}$. We take the following equation as a deformation equation for the differential equation (4.5) which is rewritten version of (2.9) $\begin{array}[]{c}T\Psi(z)=B(z)\Psi(z),\quad B(z)=\begin{bmatrix}w_{1}&w_{2}&w_{3}\\\ 0&0&0\\\ 0&0&0\end{bmatrix}+\begin{bmatrix}w_{4}&0&0\\\ 1&0&0\\\ w_{5}&w_{6}&w_{7}\end{bmatrix}z,\vspace{6pt}\\\ |B(z)|=(w_{3}w_{6}-w_{2}w_{7})z^{2},\vspace{6pt}\\\ T:(\beta_{1},\beta_{2},\beta_{3},\gamma_{1},\gamma_{2},\gamma_{3},\delta_{1},\delta_{2},\delta_{3})\to(\beta_{1}-1,\beta_{2}+1,\beta_{3},\gamma_{1},\gamma_{2},\gamma_{3},\delta_{1}+1,\delta_{2},\delta_{3}+1).\end{array}$ ( 4.6) Solving a compatibility condition for the equation (4.5), (4.6), we obtain the following additional-difference Painlevé equation of type $E_{6}$ [5], [9] $\begin{array}[]{c}(f+g)(\overline{f}+g)=\cfrac{\left(g+\gamma_{1}\right)\left(g+\gamma_{2}\right)\left(g+\gamma_{3}\right)\left(g+\beta_{1}+\delta_{1}-\delta_{2}\right)}{\left(g+\beta_{2}+1\right)\left(g+\beta_{3}-\delta_{2}+\delta_{3}+1\right)},\vspace{6pt}\\\ (\overline{f}+\overline{g})(\overline{f}+g)=\cfrac{(\overline{f}-\gamma_{1})(\overline{f}-\gamma_{2})(\overline{f}-\gamma_{3})(\overline{f}-\beta_{1}-\delta_{1}+\delta_{2})}{(\overline{f}-\beta_{1}+\beta_{2}-1)(\overline{f}-\beta_{1}-\delta_{1}+\delta_{3}+1)}.\end{array}$ ( 4.7) where, $f=\beta_{3}+\frac{u_{1}u_{4}}{l_{1}},\quad g=\frac{l_{1}l_{2}}{u_{1}}-\beta_{2},$ ( 4.8) and $\overline{*}$ stands for $T(*)$. From the above, we derive the additional-difference Painlevé equation of type $E_{6}$ solving a compatibility condition of the Lax pair via a continuous limit $q\to 1$. ## Acknowledgement The author would like to express her gratitude to Professor Yasuhiko Yamada for valuable suggestions and encouragement. She also thanks supports from JSPS KAKENHI Grant Numbers 17H06127 and 26287018 for the travel expenses in accomplishing this study. ## Appendix In the appendix, we give an explicit forms of $P_{j}(z)$ (3.2) $\begin{array}[]{l}P_{3}(z)=p_{31}(z-u),\\\ P_{2}(z)=p_{22}z^{2}+p_{21}z+p_{20},\\\ P_{1}(z)=p_{13}z^{3}+p_{12}z^{2}+p_{11}z+p_{10},\\\ P_{0}(z)=-P_{3}(qz)d_{1}d_{2}d_{3}(z+c_{1})(z+c_{2})(z+c_{3}),\end{array}$ ( 4.9) where the coefficients $p_{j,k}$ $(1\leq j\leq 3,0\leq k\leq 3)$ are $\begin{array}[]{l}p_{31}=-q^{2}uvc_{1}c_{2},\vspace{6pt}\\\ p_{22}=q^{4}uvc_{1}c_{2}\left(d_{1}+d_{2}+d_{3}\right),\vspace{6pt}\\\ p_{21}=-quvc_{1}c_{2}\left(ud_{3}q^{3}+\left(qu- b_{2}\right)d_{2}q^{2}-b_{3}d_{3}q^{2}-vq+v+\left(q^{3}u-qb_{1}\right)d_{1}\right),\vspace{6pt}\\\ p_{20}=-q^{2}u^{2}vc_{1}c_{2}\left(b_{1}d_{1}+qb_{2}d_{2}+qb_{3}d_{3}\right),\vspace{6pt}\\\ p_{13}=-q^{5}uvc_{1}c_{2}\left(d_{1}d_{2}+d_{2}d_{3}+d_{3}d_{1}\right),\vspace{6pt}\\\ p_{12}=-q(ub_{1}b_{2}b_{3}\left(u+c_{2}\right)d_{1}d_{2}d_{3}q^{5}+uc_{1}^{2}c_{2}\left(u+c_{2}\right)d_{1}d_{2}d_{3}q^{5}\vspace{6pt}\\\ \qquad+c_{1}(u^{2}c_{2}^{2}d_{1}d_{2}d_{3}q^{5}+ub_{1}b_{2}b_{3}d_{1}d_{2}d_{3}q^{5}\vspace{6pt}\\\ \qquad- c_{2}(qd_{1}(qu\left(qud_{2}\left(-ud_{3}q^{2}+vq+v\right)+v\left(q(q+1)ud_{3}-v\right)\right)\vspace{6pt}\\\ \qquad- b_{1}\left(v-q^{2}b_{2}d_{2}\right)\left(v-q^{2}b_{3}d_{3}\right))+v(d_{2}(u\left(q(q+1)ud_{3}-v\right)\vspace{6pt}\\\ \qquad+b_{2}\left(q^{2}b_{3}d_{3}-v\right))q^{2}+v\left(v-q^{2}\left(u+b_{3}\right)d_{3}\right))))),\vspace{6pt}\\\ p_{11}=u(ub_{1}b_{2}b_{3}\left(u+c_{2}\right)d_{1}d_{2}d_{3}q^{6}+uc_{1}^{2}c_{2}\left(u+c_{2}\right)d_{1}d_{2}d_{3}q^{6}\vspace{6pt}\\\ \qquad+c_{1}(u^{2}c_{2}^{2}d_{1}d_{2}d_{3}q^{6}+ub_{1}b_{2}b_{3}d_{1}d_{2}d_{3}q^{6}\vspace{6pt}\\\ \qquad- c_{2}(qd_{1}(b_{1}\left(b_{2}d_{2}\left(-b_{3}d_{3}q^{3}+vq+v\right)q^{2}+v\left(q^{2}(q+1)b_{3}d_{3}-v\right)\right)\vspace{6pt}\\\ \qquad- qu\left(v-q^{2}ud_{2}\right)\left(v-q^{2}ud_{3}\right))+v(d_{2}(u\left(q^{2}ud_{3}-v\right)\vspace{6pt}\\\ \qquad+b_{2}\left(q^{2}(q+1)b_{3}d_{3}-v\right))q^{2}+v\left(v-q^{2}\left(u+b_{3}\right)d_{3}\right))))),\vspace{6pt}\\\ p_{10}=q^{3}u^{2}vc_{1}c_{2}\left(qb_{2}b_{3}d_{2}d_{3}+b_{1}d_{1}\left(b_{2}d_{2}+b_{3}d_{3}\right)\right).\end{array}$ ( 4.10) We give also an explicit forms of the coefficients $c_{ij}$ $(0\leq j\leq 3,0\leq i+j\leq 4)$ of the polynomial $P(u,v)$ in variables $u$, $v$: $\begin{array}[]{l}c_{01}=c_{0}\cfrac{z^{2}}{q^{2}}\left(-\cfrac{q}{b_{1}d_{1}}-\cfrac{1}{b_{2}d_{2}}-\cfrac{1}{b_{3}d_{3}}\right)\Phi(qz),\vspace{6pt}\\\ c_{02}=c_{0}\cfrac{z^{2}\left(b_{1}d_{1}+qb_{2}d_{2}+qb_{3}d_{3}\right)\Phi(qz)}{q^{4}b_{1}b_{2}b_{3}d_{1}d_{2}d_{3}},\vspace{6pt}\\\ c_{03}=-c_{0}\cfrac{z^{2}\Phi(qz)}{q^{5}b_{1}b_{2}b_{3}d_{1}d_{2}d_{3}},\vspace{6pt}\\\ c_{10}=c_{0}\left(\cfrac{c_{1}c_{2}z^{2}}{b_{1}b_{2}b_{3}}+\cfrac{z^{2}}{c_{1}}+\cfrac{z^{2}}{c_{2}}-z\right)\Phi(qz),\vspace{6pt}\\\ c_{11}=-c_{0}\cfrac{z}{q^{4}b_{1}b_{2}b_{3}c_{1}c_{2}d_{1}d_{2}d_{3}}(b_{1}d_{1}(qb_{2}d_{2}\left(q^{2}b_{3}\left(z+c_{1}\right)\left(z+c_{2}\right)d_{3}\Phi(z)-(q+1)c_{1}c_{2}\Phi(qz)\right)\vspace{6pt}\\\ \quad\qquad+c_{1}c_{2}\left(\Phi\left(q^{2}z\right)-q(q+1)b_{3}d_{3}\Phi(qz)\right))\vspace{6pt}\\\ \quad\qquad+c_{1}c_{2}(zd_{3}\Phi\left(q^{2}z\right)q^{2}+zd_{1}\left(-qzd_{3}\Phi(qz)+qd_{2}\left(\left(z+c_{1}\right)\left(z+c_{2}\right)d_{3}\Phi(z)-z\Phi(qz)\right)+\Phi\left(q^{2}z\right)\right)q^{2}\vspace{6pt}\\\ \quad\qquad+b_{3}d_{3}\Phi\left(q^{2}z\right)q-d_{2}\left(q\left(qz^{2}+(q+1)b_{2}b_{3}\right)d_{3}\Phi(qz)-\left(qz+b_{2}\right)\Phi\left(q^{2}z\right)\right)q-\Phi\left(q^{3}z\right))),\vspace{6pt}\\\ c_{12}=c_{0}\cfrac{z}{q^{5}b_{1}b_{2}b_{3}d_{1}d_{2}d_{3}}(zd_{3}\Phi(qz)q^{2}+\left(qz- b_{2}\right)d_{2}\Phi(qz)q-b_{3}d_{3}\Phi(qz)q-\Phi\left(q^{2}z\right)q\vspace{6pt}\\\ \qquad\qquad+\left(q^{2}z-b_{1}\right)d_{1}\Phi(qz)+\Phi\left(q^{2}z\right)),\vspace{6pt}\\\ c_{13}=c_{0}\cfrac{z\Phi(qz)}{q^{6}b_{1}b_{2}b_{3}d_{1}d_{2}d_{3}},\vspace{6pt}\\\ c_{20}=c_{0}\left(\cfrac{c_{1}z\left(z-c_{2}\right)z+z^{2}c_{2}}{b_{1}b_{2}b_{3}}-\cfrac{-z^{2}+c_{1}z+c_{2}z}{c_{1}c_{2}}\right)\Phi(qz),\vspace{6pt}\\\ c_{21}=c_{0}\cfrac{b_{1}d_{1}}{q^{4}b_{1}b_{2}b_{3}c_{1}c_{2}d_{1}d_{2}d_{3}}(qb_{2}d_{2}\left(qb_{3}\left(z+c_{1}\right)\left(z+c_{2}\right)d_{3}\Phi(z)-c_{1}c_{2}\Phi(qz)\right)\vspace{6pt}\\\ \qquad\quad+c_{1}c_{2}\left(\Phi\left(q^{2}z\right)-qb_{3}d_{3}\Phi(qz)\right))-c_{0}c_{1}c_{2}(zd_{1}(-\left(z+c_{1}\right)\left(z+c_{2}\right)d_{2}d_{3}\Phi(z)+(q+1)zd_{2}\Phi(qz)\vspace{6pt}\\\ \qquad\quad+(q+1)zd_{3}\Phi(qz)-\Phi\left(q^{2}z\right))q^{2}-zd_{3}\Phi\left(q^{2}z\right)q^{2}-b_{3}d_{3}\Phi\left(q^{2}z\right)q+d_{2}(q\left((q+1)z^{2}+b_{2}b_{3}\right)d_{3}\Phi(qz)\vspace{6pt}\\\ \qquad\quad-\left(qz+b_{2}\right)\Phi\left(q^{2}z\right))q+\Phi\left(q^{3}z\right)),\vspace{6pt}\\\ c_{22}=-c_{0}\cfrac{z\left(d_{1}+d_{2}+d_{3}\right)\Phi(qz)}{q^{4}b_{1}b_{2}b_{3}d_{1}d_{2}d_{3}},\vspace{6pt}\\\ c_{30}=c_{0}\left(-\cfrac{-z^{2}+c_{1}z+c_{2}z}{b_{1}b_{2}b_{3}}-\cfrac{1}{c_{1}c_{2}}\right)\Phi(qz),\vspace{6pt}\\\ c_{31}=c_{0}\cfrac{\left(d_{1}d_{2}+d_{2}d_{3}+d_{3}d_{1}\right)z\Phi(qz)}{q^{2}b_{1}b_{2}b_{3}d_{1}d_{2}d_{3}},\vspace{6pt}\\\ c_{40}=-c_{0}\cfrac{z\Phi(qz)}{b_{1}b_{2}b_{3}}.\end{array}$ ( 4.11) ## References * [1] Boalch, P., ”Quivers and difference Painlevé equations”, (2009), CRM Proc. Lecture Notes, 47, 25–51. * [2] Dzhamay, A. and Takenawa, T. ”On some applications of Sakai’s geometric theory of discrete Painlevé equations”, (2018), SIGMA, 14, 075, 20pages. * [3] Grammaticos, B. and Ramani, A., ”The hunting for discrete Painlevé equations”, (2000), Regul. Chaotic Dyn., 5, (1), 53–66. * [4] Jimbo, M. and Sakai, H., ”A $q$-analog of the sixth Painlevé equation”, (1996), Lett. Math. Phys., 38, 145–154. * [5] Kajiwara, K., Noumi, M. and Yamada, Y., ”Geometry aspects of Painelvé equations”, (2017) J. Phys. A: Math. Theor. 50, 073001, 164pages * [6] Park, K., ”A certain generalization of $q$-hypergeometric functions and their related monodromy preserving deformation”, (2018), J. Int. Sys., 3, 1–14. * [7] Park, K., ”A certain generalization of $q$-hypergeometric functions and their related connection preserving deformation II”, (2022), Funkc. Ekvacioj., 65, 311–328. * [8] Ramani, A., Grammaticos, B. and Hietarinta, J., ”Discrete versions of the Painlevé equations”, (1991) Phys. Rev. Lett. 67, 1829–1832. * [9] Ramani, A., Grammaticos, B., Tamizhmani, T. and Tamizhmani, K. M., ”Special function solutions of the discrete Painlevé equations”, (2001), Comput. Math. with Appl. 42, 603–614 * [10] Sakai, H., ”A $q$-analog of the Garnier system”, (2005), Funkc. Ekvacioj., 48, 273–297. * [11] Sakai, H., ”Lax form of the $q$-Painlevé equation associated with the $A_{2}^{(1)}$ surface”, (2006), J. Phys. A: Math. Gen. 39, 12203–12210. * [12] Suzuki, T. ”A $q$-analogue of the Drinferd-Sokolov hierarchy of type $A$ and $q$-Painlevé system”, (2015), AMS Contemp. Math., 651, 25–38. * [13] Suzuki, T., ”A Lax Formulation of a generalized $q$-Garnier System”, (2021), Math. Phys. Anal. Geom., 24, 38. * [14] Witte, N. S. and Ormerod, C. M., ”Construction of a Lax pair for $E_{6}^{(1)}$ $q$-Painlevé system”, (2012), SIGMA, 8, 27pages. * [15] Yamada, Y., ”Lax formalism for $q$-Painlevé equations with affine Weyl group symmetry of type $E_{n}^{(1)}$”, (2011), Int. Math. Res. Notices, 17, 3823–3838.
2022 # Entropy fluctuation formulas of fermionic Gaussian states Youyi Huang Lu Wei Department of Computer Science Texas Tech University Lubbock, 79409, Texas, USA ###### Abstract We study the statistical behaviour of quantum entanglement in bipartite systems over fermionic Gaussian states as measured by von Neumann entropy. The formulas of average von Neumann entropy with and without particle number constrains have been recently obtained, whereas the main results of this work are the exact yet explicit formulas of variances for both cases. For the latter case of no particle number constrain, the results resolve a recent conjecture on the corresponding variance. Different than existing methods in computing variances over other generic state models, the key ingredient in proving the results of this work relies on a new simplification framework. The framework consists of a set of new tools in simplifying finite summations of what we refer to as dummy summation and re-summation techniques. As a byproduct, the proposed framework leads to various new transformation formulas of hypergeometric functions. ###### keywords: von Neumann entropy, fermionic Gaussian states, quantum entanglement, random matrix theory, orthogonal polynomials, special functions ## 1 Introduction and main results Quantum entanglement is the most important feature in quantum mechanics. The understanding of the phenomenon of entanglement is crucial in realizing the revolutionary advances of quantum science. In the emerging field of quantum information processing, quantum entanglement is also the resource and medium that enable the underlying quantum technologies. In this work, we study the statistical behaviour of entanglement over the fermionic Gaussian states. In the past decades, considerable effort has been devoted to investigating the degree of entanglement as measured by different entanglement entropies over the well-known Hilbert-Schmidt ensemble HLW06 ; Page93 ; Foong94 ; Ruiz95 ; VPO16 ; Wei17 ; Wei20 ; HWC21 ; Lubkin78 ; Sommers04 ; Giraud07 ; MML02 ; Wei19T . In particular, these studies focus on the statistical behaviour of entanglement entropies such as von Neumann entropy HLW06 ; Page93 ; Foong94 ; Ruiz95 ; VPO16 ; Wei17 ; Wei20 ; HWC21 , quantum purity Lubkin78 ; Sommers04 ; Giraud07 , and Tsallis entropy MML02 ; Wei19T . Driven by the recent breakthrough in probability theory on the Bures- Hall ensemble Bortola09 ; Bortola10 ; Bortola14 ; FK16 , considerable progress has been made in understanding the von Neumann entropy SK2019 ; Wei20BHA ; Wei20BH ; LW21 and quantum purity Sommers04 ; Osipov10 ; Borot12 ; LW21 over the Bures-Hall ensemble. Similar investigations are now being carried out over the fermionic Gaussian ensemble, which is a generic state model relevant for different quantum information processing tasks BHK21 ; BHKRV22 ; LRV20 ; OGMM19 ; ST21 . Very recently, the mean values of von Neumann entropy with and without particle number constrains over the fermionic Gaussian ensemble are obtained respectively in BHK21 and BHKRV22 . As an important step towards characterizing the statistical distribution of von Neumann entropy, we aim to derive the corresponding variances, which describe the fluctuation of the entropy around their mean values. The exact variance of von Neumann entropy over fermionic Gaussian states without particle number constrain has been conjectured in a previous work of the authors HW22 . In the current work, we prove the conjecture as well as derive a variance formula for the case of a fixed particle number. ### 1.1 Problem formulation We introduce the formulation that leads to the fermionic Gaussian states with and without particle number constrains. A system of $N$ fermionic degrees of freedom can be decomposed into two subsystems $A$ and $B$ of the dimensions $m$ and $n$, respectively, with $m+n=N$. Without loss of generality, we assume $m\leq n$ . In the present work, we consider two scenarios of fermionic Gaussian states – the fermionic Gaussian states with arbitrary number of particles and the fermionic Gaussian states with a fixed number of particles. #### Case A: Arbitrary number of particles A system of $N$ fermionic modes can be formulated in terms of a set of fermionic creation and annihilation operators $\hat{a}_{i}$ and $\hat{a}_{i}^{{\dagger}}$, $i=1,\dots,N$. Since the modes are fermionic, these operators obey the canonical anti-commutation relation OGMM19 ; BHKRV22 , $\\{\hat{a}_{i},\hat{a}_{j}^{{\dagger}}\\}=\delta_{ij}\mathbb{I},\qquad\\{\hat{a}_{i},\hat{a}_{j}\\}=0=\\{\hat{a}_{i}^{{\dagger}},\hat{a}_{j}^{{\dagger}}\\},$ (1) where $\\{\hat{A},\hat{B}\\}=\hat{A}\hat{B}+\hat{B}\hat{A}$ denotes the anti- commutation relation and $\mathbb{I}$ is an identity operator. Equivalently, one can also describe these fermionic modes in terms of the Majorana operators $\gamma_{l}$, $l=1,\dots,2N$, and $\hat{\gamma}_{2i-1}=\frac{\hat{a}_{i}^{\dagger}+\hat{a}_{i}}{\sqrt{2}},\qquad\hat{\gamma}_{2i}=\imath\frac{\hat{a}_{i}^{\dagger}+\hat{a}_{i}}{\sqrt{2}}$ (2) with $\imath=\sqrt{-1}$ being the imaginary unit. Note that the Majorana operators are Hermitian satisfying the anti-commutation relation $\\{\hat{\gamma}_{l},\hat{\gamma}_{k}\\}=\delta_{lk}\mathbb{I}.$ (3) By collecting the Majorana operators into a $2N$ dimensional operator-valued column vector ${\gamma}=(\hat{\gamma}_{1},\dots,\hat{\gamma}_{2N})^{{\dagger}}$, a fermionic Gaussian state is then written as the density operator of the form ST21 ; BHKRV22 $\rho(\gamma)=\frac{\mathrm{e}^{-\gamma^{\dagger}Q\gamma}}{\tr(\mathrm{e}^{-\gamma^{\dagger}Q\gamma})},$ (4) where the coefficient matrix $Q$ is a $2N\times 2N$ imaginary anti-symmetric matrix as the consequence of the anti-communication relation (3). There always exsits an orthogonal matrix $M$ that diagnoses the coefficient matrix $Q$ by transforming ${\gamma}$ into another Majorana basis $\mu=(\hat{\mu}_{1},\dots,\hat{\mu}_{2N})^{\dagger}=M\gamma$ BHKRV22 . A fermionic Gaussian state is labelled by its anti-symmetric covariance matrix BHKRV22 $J=-\imath\tanh(Q)=M^{T}J_{0}M,$ (5) where $\tanh(x)$ denotes the hyperbolic tangent function AS72 , the matrix $J_{0}$ takes the block diagonal form $J_{0}=\left(\begin{array}[]{ccc}\tanh(\lambda_{1})\mathbb{A}&\dots&0\\\ \vdots&\ddots&\vdots\\\ 0&\dots&\tanh(\lambda_{N})\mathbb{A}\\\ \end{array}\right),$ (6) and $\mathbb{A}=\left(\begin{array}[]{cc}0&1\\\ -1&0\\\ \end{array}\right).$ (7) We consider the von Neumann entropy as the measure of entanglement between the two subsystems. By restricting the matrix $J$ to the entries from subsystems $A$, the restricted matrix $J_{A}$ becomes the $2m\times 2m$ left-upper block of $J$. The von Neumann entropy of a fermionic Gaussian state of case A can be represented in terms of the real positive eigenvalues $x_{i},i=1,\dots,m$ of $\imath J_{A}$ as BHK21 ; BHKRV22 ; HW22 $S=-\sum_{i=1}^{m}v(x_{i}),$ (8) where $v(x)=\frac{1-x}{2}\ln\frac{1-x}{2}+\frac{1+x}{2}\ln\frac{1+x}{2}.$ (9) The resulting joint probability density of the eigenvalues $x_{i},i=1,\dots,m$ is proportional to BHK21 $\prod_{1\leq i<j\leq m}\left(x_{i}^{2}-x_{j}^{2}\right)^{2}\prod_{i=1}^{m}\left(1-x_{i}^{2}\right)^{n-m},\qquad x_{i}\in[0,1],$ (10) which is obtained by recursively applying the result (KFI19, , Proposition A.2). #### Case B: Fixed number of particles For a fermionic Gaussian state $\ket{F}$ with a fixed particle number $p$, $m\leq p\leq n$, the corresponding covariance matrix $H$ can be expressed via the commutator of fermionic creation and annihilation operators as BHKRV22 ; LRV20 $H_{ij}=-\imath\bra{F}\hat{a}_{i}^{{\dagger}}\hat{a}_{j}-\hat{a}_{j}\hat{a}_{i}^{{\dagger}}\ket{F}.$ (11) Recall the canonical anti-commutation relation (1), the entries of the matrix $H$ are then of the form $H_{ij}=-2\imath G_{ij}+\imath\delta_{ij}\mathbb{I},$ (12) where $G_{ij}=\bra{F}\hat{a}_{i}^{{\dagger}}\hat{a}_{j}\ket{F}$ denotes the entries of an $N\times N$ matrix $G$ of a fermionic system of $N$ modes. There exists a unitary transformation $U$ that diagonalizes $G$ into the form $U^{{\dagger}}GU$, where the first $p$ diagonal elements are equal to $1$ and the rest are $0$. Therefore, one can write $G=U_{N\times p}U_{N\times p}^{\dagger}.$ (13) A fermionic Gaussian state of dimension $N=m+n$ with $p$ particles can be fully characterized by the matrices $H$ and $G$. The von Neumann entropy of the fermionic system in the case B can be represented as BHKRV22 ; LRV20 $~{}S=-\sum_{i=1}^{m}v(2y_{i}-1),\qquad y_{i}\in[0,1],$ (14) where $y_{i}$, $i=1,\dots,m$ are the eigenvalues of the restricted $m\times m$ matrix $G_{A}=U_{m\times p}U_{m\times p}^{\dagger}$. The eigenvalue distribution of the random matrix $U_{m\times p}U_{m\times p}^{\dagger}$ is the well-known Jacobi unitary ensemble Mehta ; Forrester . We denote $x_{i}$, $i=1,\dots,m$ the eigenvalues of the $m\times m$ left-upper block of matrix $\imath H$. Changing the variables $x_{i}=2y_{i}-1$ in (14) leads to the von Neumann entropy (8) of case B. The resulting joint probability density of the eigenvalues $x_{i},i=1,\dots,m$ is proportional to BP21 $\prod_{1\leq i<j\leq m}\left(x_{i}-x_{j}\right)^{2}\prod_{i=1}^{m}\left({1+x_{i}}\right)^{p-m}\left(1-x_{i}\right)^{n-p},\qquad x_{i}\in[-1,1].$ (15) It is important to point out that the two joint probability densities (10) and (15) can be compactly represented by a single joint density as $f_{\mathrm{FG}}(x)\propto\prod_{1\leq i<j\leq m}\left(x_{i}^{\gamma}-x_{j}^{\gamma}\right)^{2}\prod_{i=1}^{m}\left(1-x_{i}\right)^{a}\left(1+x_{i}\right)^{b},$ (16) where for the case A we have $\gamma=2,~{}~{}~{}a=b=n-m\geq 0,~{}~{}~{}x\in[0,1],$ (17) and for the case B we have $\gamma=1,~{}~{}~{}a=n-p\geq 0,~{}~{}~{}b=p-m\geq 0,~{}~{}~{}x\in[-1,1].$ (18) We omit the normalizations in the density (16) as they will not be made use of in the subsequent calculations. Note that the variance computation is difficult for an arbitrary $\gamma$ in (16), and one has to consider the cases $\gamma=2$ and $\gamma=1$ separately. ### 1.2 Main results We now introduce the exact mean and variance formulas of von Neumann entropy for both case A and case B. The mean values have been recently computed BHK21 ; BHKRV22 as summarized in Proposition 1 and Proposition 2 for case A and case B, respectively. The corresponding variance formulas are presented in Proposition 3 and Proposition 4 below, which are the main results of the work. ###### Proposition 1 (BHK21 ). For subsystem dimensions $m\leq n$, the mean value of the von Neumann entropy (8) of fermionic Gaussian states with arbitrary number of particles (17) is given by $\displaystyle\mathbb{E}\\!\left[S\right]=$ $\displaystyle\left(m+n-\frac{1}{2}\right)\psi_{0}(2m+2n)+\left(\frac{1}{4}-m\right)\psi_{0}(m+n)+\left(\frac{1}{2}-n\right)\psi_{0}(2n)$ $\displaystyle-\frac{1}{4}\psi_{0}(n)-m,$ (19) where $\psi_{0}(x)=\frac{\differential\ln\Gamma(x)}{\differential x}$ (20) is the digamma function. ###### Proposition 2 (BHKRV22 ). For subsystem dimensions $m\leq n$, the mean value of the von Neumann entropy (8) of fermionic Gaussian states with a fixed particle number (18) is given by $\displaystyle\mathbb{E}\\!\left[S\right]=$ $\displaystyle-\frac{m(m+n-p)}{m+n}\psi_{0}(m+n-p)+(m+n)\psi_{0}(m+n+1)$ $\displaystyle-\frac{mp}{m+n}\psi_{0}(p+1)-n\psi_{0}(n+1)-m.$ (21) ###### Proposition 3. For subsystem dimensions $m\leq n$, the variance of the von Neumann entropy (8) of fermionic Gaussian states with arbitrary number of particles (17) is given by $\displaystyle\mathbb{V}\\!\left[S\right]=~{}\\!\\!$ $\displaystyle\\!\left(\frac{1}{2}-m-n\right)\psi_{1}(2m+2n)+\left(n-\frac{1}{2}\right)\psi_{1}(2n)+\left(\frac{m(2m+n-1)}{2m+2n-1}-\frac{1}{8}\right)$ $\displaystyle\times\psi_{1}(m+n)+\frac{\psi_{1}(n)}{8}-\frac{1}{2}(\psi_{0}(2m+2n)-\psi_{0}(2n)),$ (22) where $\psi_{1}(x)=\frac{\differential^{2}\ln\Gamma(x)}{\differential x^{2}}$ (23) is the trigamma function. ###### Proposition 4. For subsystem dimensions $m\leq n$, the variance of the von Neumann entropy (8) of fermionic Gaussian states with a fixed particle number (18) is given by $\displaystyle\mathbb{V}\\!\left[S\right]=~{}\\!\\!$ $\displaystyle c_{0}\psi_{1}(m+n-p)-(m+n)\psi_{1}(m+n)+n\psi_{1}(n)+c_{1}\psi_{1}(p)+c_{2}$ $\displaystyle\times(\psi_{0}(m+n-p)-\psi_{0}(p))^{2}+c_{3}\left(\psi_{0}(m+n-p)-\psi_{0}(p)\right)-\psi_{0}(m+n)$ $\displaystyle+\psi_{0}(n)+c_{4},$ (24) where the coefficients $c_{i}$ are summarized in Table 1 below with $(a)_{n}=\Gamma(a+n)/\Gamma(a)$ denoting the Pochhammer symbol. Table 1: Coefficients of von Neumann entropy variance in Proposition 4 $c_{0}~{}=$ | $\displaystyle\frac{m(m+n-p)\left(m^{2}+2mn+n^{2}-np-1\right)}{(m+n-1)_{3}}$ ---|--- $c_{1}~{}=$ | $\displaystyle\frac{mp\left(m^{2}+mn+np-1\right)}{(m+n-1)_{3}}$ $c_{2}~{}=$ | $\displaystyle\frac{mnp(m+n-p)}{(m+n)(m+n-1)_{3}}$ $c_{3}~{}=$ | $\displaystyle-\frac{m(m+1)(m+n-2p)}{(m+n)(m+n)_{2}}$ $c_{4}~{}=$ | $\displaystyle-\frac{m(2m+n+2)}{(m+n)(m+n)_{2}}$ The proof to Proposition 3 and Proposition 4 will be presented in Section 2. Note that a special case of equal subsystem dimensions ($m=n$) of the result in Proposition 3 has been established very recently HW22 by utilizing an existing simplification framework developed in Wei17 ; Wei20 ; HWC21 ; Wei22 ; Wei20BH ; LW21 ; HW22 . However, for the case of arbitrary subsystem dimensions, the existing framework is not sufficient to simplify some of the summations in the variance calculation, where a key technical contribution of the work is to develop a new simplification framework. Figure 1: Variance of von Neumann entropy: analytical results versus simulations. The black lines represent the obtained analytical result (3) for the cases $n=m$, $n=3m$, and $n=4m$. The red lines are drawn by the result (4) for the cases $p=m,n=3m$; $p=2m,n=3m$; and $p=2m,n=4m$. The diamond scatters represent numerical simulations. To illustrate the derived results (3) and (4), we plot in Figure 1 the exact variance of von Neumann entropy as compared with the simulations111The simulations performed in figures 1–3 utilize the Mathematica codes provided by Santosh Kumar based on the log-gas approach as discussed in (SK2019, , Appendix B).. In Figure 1, we observe that the variance in case A approaches to a constant when system dimensions increase with a fixed ratio $f_{1}=\frac{m}{n+m},$ (25) where the variance in case B follows the same behavior with fixed $f_{1}$ and $f_{2}=\frac{p}{n+m}.$ (26) This phenomenon can be analytically established by the asymptotic results of variances in the literature. For case A, in the asymptotic regime BHK21 $m\to\infty,\qquad n\to\infty,\qquad 0<f_{1}\leq\frac{1}{2},$ (27) one has BHK21 $\mathbb{V}\\!\left[S\right]=\frac{1}{2}\left(f_{1}+f_{1}^{2}+\ln(1-f_{1})\right)+o\left(\frac{1}{m+n}\right),$ (28) whereas for case B, in the asymptotic regime BHKRV22 $m\to\infty,\qquad p\to\infty,\qquad n\to\infty,\qquad 0<f_{1}\leq f_{2}\leq\frac{1}{2},$ (29) one has BHKRV22 $\displaystyle\mathbb{V}\\!\left[S\right]=\\!~{}$ $\displaystyle f_{1}+f_{1}^{2}+\ln\left(1-f_{1}\right)+f_{1}f_{2}\left(1-f_{1}\right)\left(1-f_{2}\right)\ln^{2}\frac{1-f_{2}}{f_{2}}$ $\displaystyle+f_{1}^{2}\left(2f_{2}-1\right)\ln\frac{1-f_{2}}{f_{2}}+o\left(\frac{1}{(m+n)^{2}}\right).$ (30) The above asymptotic variances (28) and (1.2) can be directly recovered by the results in Proposition 3 and Proposition 4, respectively. Moreover, the correction terms of any order can be simply obtained from our exact variance formulas upon using the asymptotic behaviour of polygamma functions $\displaystyle\psi_{0}(x)=$ $\displaystyle\ln(x)-\frac{1}{2x}-\sum_{l=1}^{\infty}\frac{B_{2l}}{2lx^{2l}},\qquad x\to\infty,$ (31) $\displaystyle\psi_{1}(x)=$ $\displaystyle\frac{1+2x}{2x^{2}}+\sum_{l=1}^{\infty}\frac{B_{2l}}{x^{2l+1}},\qquad x\to\infty,$ (32) where $B_{k}$ is the $k$-th Bernoulli number AS72 . For example, utilizing the next order of correction, the asymptotic result (1.2) is refined to $\displaystyle\mathbb{V}\\!\left[S\right]=$ $\displaystyle f_{1}^{2}+f_{1}+\ln\left(1-f_{1}\right)+f_{1}f_{2}\left(1-f_{1}\right)\left(1-f_{2}\right)\ln^{2}\frac{1-f_{2}}{f_{2}}+f_{1}^{2}$ $\displaystyle\times\left(2f_{2}-1\right)\ln\frac{1-f_{2}}{f_{2}}+\frac{1}{12(m+n)^{2}}\Bigg{(}\frac{f_{1}^{2}}{\left(f_{2}-1\right){}^{2}}+\frac{f_{1}^{2}}{f_{2}^{2}}+12f_{1}^{2}$ $\displaystyle-12f_{1}+\frac{1}{\left(f_{1}-1\right){}^{2}}+\frac{f_{1}-3f_{1}^{2}}{f_{2}-1}+\frac{3f_{1}^{2}-f_{1}}{f_{2}}-1$ $\displaystyle+\frac{2\left(f_{1}-1\right)f_{1}\left(12f_{2}^{3}-18f_{2}^{2}+4f_{2}+1\right)}{\left(f_{2}-1\right)f_{2}}\ln\frac{1-f_{2}}{f_{2}}$ $\displaystyle+12\left(f_{1}-1\right)f_{1}\left(f_{2}-1\right)f_{2}\ln^{2}\frac{1-f_{2}}{f_{2}}\\!\Bigg{)}+o\left(\frac{1}{(m+n)^{4}}\right).$ (33) Figure 2: Probability densities of standardized von Neumann entropy for case A in (17): a comparison of Gaussian density (35) to the simulation results. The dash-dot line in blue and the dashed line in red refer to the standardized von Neumann entropy (34) of subsystem dimensions $m=2$, $n=4$, and $m=16$, $n=32$, respectively. The solid black line represents the Gaussian density (35). Figure 3: Probability densities of standardized von Neumann entropy for case B in (18): a comparison of Gaussian density (35) to the simulation results. The dash-dot line in blue and the dashed line in red refer to the standardized von Neumann entropy (34) of dimensions $m=2$, $p=4$, $n=6$, and $m=6$, $p=12$, $n=18$, respectively. The solid black line represents the Gaussian density (35). To understand the distribution of the von Neumann entropy, simple approximations can now be constructed by using the obtained mean and variance formulas. We first standardize the von Neumann entropy as $X=\frac{S-\mathbb{E}\\!\left[S\right]}{\sqrt{\mathbb{V}\\!\left[S\right]}},$ (34) where the random variable $X$ is of zero mean and unit variance. We now compare the distribution of $X$ with a standard Gaussian distribution $\phi(x)=\frac{1}{\sqrt{2\pi}}\mathrm{e}^{-\frac{1}{2}x^{2}},\qquad x\in(-\infty,\infty).$ (35) In figures 2–3, we plot the simulation results of the standardized von Neumann entropy $X$ as compared with a standard Gaussian. Specifically, the ratios are fixed to $f_{1}=1/3$ for case A in Figure 2 and $f_{1}=1/4$, $f_{2}=1/2$ for case B in Figure 3. It is observed from the figures that the Gaussian density captures accurately the distribution of the standardized von Neumann entropy $X$ for moderately large dimensions. We also observe that the true distribution of $X$ is non-symmetric and appears to be left-skewed when the subsystem dimensions are small as seen from the dash-dot blue curves. In comparison, when the subsystem dimensions become larger, the distributions of $X$ appear to be closer to the Gaussian distribution. In fact, the Gaussian density as a limiting behavior of von Neumann entropy has been conjectured over different random matrix models of Hilbert-Schmidt ensemble Wei20 , Bures- Hall ensemble Wei20BH , and fermionic Gaussian ensemble of an arbitrary number of particles HW22 . Here, one is also attempting to conjecture that under the asymptotic regime (29), the standardized von Neumann entropy (34) of fermionic Gaussian states with a fixed particle number (18) converges in distribution to a standard Gaussian. The rest of the paper is organized as follows. The detailed calculations of the main results in Proposition 3 and Proposition 4 are provided in Section 2. In Appendix A, we list the summation representations of the integrals involved in the variance computations. Some additional finite sum identities utilized in the simplification are listed in Appendix B. ## 2 Variance calculation In this section, we prove the results in Proposition 3 and Proposition 4. In Section 2.1, we obtain the summation representations of the variances. Tools in simplifying these summations are introduced in Section 2.2, where the new simplification framework consisting of six lemmas is presented first. The detailed simplification procedures that lead to the claimed results (3) and (4) are discussed in Section 2.3. ### 2.1 Correlation functions and integral calculations Recall the definition (8) of von Neumann entropy $S=-\sum_{i=1}^{m}v(x_{i}),$ (36) with $v(x)=\frac{1-x}{2}\ln\frac{1-x}{2}+\frac{1+x}{2}\ln\frac{1+x}{2},$ (37) computing its variance requires one and two arbitrary eigenvalue densities of the fermionic Gaussian ensemble (16). Denoting $g_{l}(x_{1},\dots,x_{l})$ as the joint density of $l$ arbitrary eigenvalues, the variance of von Neumann entropy is written as $\mathbb{V}\\!\left[S\right]=\mathbb{E}\\!\left[S^{2}\right]-\mathbb{E}^{2}\\!\left[S\right],$ (38) where $\displaystyle\mathbb{E}\\!\left[S^{2}\right]$ $\displaystyle=m\int_{x}v^{2}(x)g_{1}(x)\differential x+m(m-1)\iint_{x,y}v(x)v(y)g_{2}\left(x,y\right)\differential x\differential y$ (39) $\displaystyle\mathbb{E}\\!\left[S\right]$ $\displaystyle=m\int_{x}v(x)g_{1}(x)\differential x.$ (40) In (39) and (40), the support is $x,y\in[0,1]$ for case A, and the support is $x,y\in[-1,1]$ for case B. For fermionic Gaussian ensemble (16), it is a well-known result in random matrix theory that the joint density $g_{l}(x_{1},\dots,x_{l})$ can be written in terms of an $l\times l$ determinant as Mehta ; Forrester $g_{l}(x_{1},\dots,x_{l})=\frac{(m-l)!}{m!}\det\left(K\left(x_{i},x_{j}\right)\right)_{i,j=1}^{l}.$ (41) The determinant in (41) is known as the $l$-point correlation function Forrester , where $K\left(x,y\right)=\sqrt{w(x)w(y)}\sum_{k=0}^{m-1}\frac{J_{k}^{(a,b)}(x)J_{k}^{(a,b)}(y)}{h_{k}}$ (42) is the correlation kernel with the weight function $w(x)=\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b}.$ (43) For convenience, we summarize in Table 2 the parameters $a$, $b$, and the order $k$ of the Jacobi polynomial $J_{k}^{(a,b)}(x)$ along with the normalization constants $h_{k}$ of case A and case B. Table 2: Parameters of correlation kernel (42) in case A and case B. | Case A: Arbitrary number of particle (17) | Case B: Fixed number of particle (18) ---|---|--- $a$ | $n-m$ | $n-p$ $b$ | $n-m$ | $p-m$ $k$ | $2k$ | $k$ $h_{k}$ | $\displaystyle\frac{(4k+2a+1)^{-1}\Gamma^{2}(2k+a+1)}{\Gamma(2k+2a+1)\Gamma(2k+1)}$ | $\displaystyle\frac{2\Gamma(k+a+1)\Gamma(k+b+1)}{(2k+a+b+1)\Gamma(k+1)\Gamma(k+a+b+1)}$ | | The constants $h_{k}$ in Table 2 are obtained as follows. For case B, the orthogonality relation Forrester $\displaystyle\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b}J^{(a,b)}_{k}(x)J^{(a,b)}_{l}(x)\differential x$ $\displaystyle=\frac{2\Gamma(k+a+1)\Gamma(k+b+1)}{(2k+a+b+1)\Gamma(k+1)\Gamma(k+a+b+1)}\delta_{kl},\quad\real(a,b)>-1,$ (44) directly leads to the normalization constant $h_{k}$ of Jacobi polynomials $J^{(a,b)}_{k}$. For case A, by using the parity property of Jacobi polynomials Szego $J^{(a,b)}_{k}(-x)=(-1)^{k}J^{(b,a)}_{k}(x),$ (45) the orthogonality relation (44) can also be written as $\displaystyle\int_{0}^{1}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{a}J^{(a,a)}_{2k}(x)J^{(a,a)}_{2l}(x)\differential x$ $\displaystyle=\frac{\Gamma(2k+a+1)\Gamma(2k+a+1)}{(4k+2a+1)\Gamma(2k+1)\Gamma(2k+2a+1)}\delta_{kl},\quad\real(a)>-1,$ (46) which gives the normalization constant $h_{k}$ of the polynomials $J^{(a,a)}_{2k}(x)$ as shown in Table 2. By using the joint density (41) and the results (39)–(40), the variance (38) now boils down to computing two integrals involving the $1$-point and $2$-point correlation functions as, cf. Wei17 ; Wei20BH ; BHK21 ; HW22 , $\mathbb{V}\\!\left[S\right]=\mathrm{I_{A}}-\mathrm{I_{B}},$ (47) where $\displaystyle\mathrm{I_{A}}$ $\displaystyle=\int_{x}v^{2}(x)K(x,x)\differential x$ (48) $\displaystyle\mathrm{I_{B}}$ $\displaystyle=\iint_{x,y}v(x)v(y)K^{2}\left(x,y\right)\differential x\differential y$ (49) with $x,y\in[0,1]$ for case A and $x,y\in[-1,1]$ for case B. We now compute the above two integrals $\mathrm{I_{A}}$ and $\mathrm{I_{B}}$ into the corresponding summation representations. Note that the subsequent calculations of the case A in (17) and case B in (18) are different, which will be performed separately in the following. #### Case A: Arbitrary number of particles By the definition of the correlation kernel (42) and keeping in mind the parity property (45) of Jacobi polynomials, the integral $\mathrm{I_{A}}=\int_{0}^{1}v^{2}(x)K(x,x)\differential x$ (50) of fermionic Gaussian states with arbitrary number of particles boils down to computing the two parts $\mathrm{I_{A}}=\mathrm{A_{1}}+\mathrm{A_{2}},$ (51) where $\displaystyle\mathrm{A_{1}}$ $\displaystyle=\sum_{k=0}^{m-1}\frac{1}{h_{k}}\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{a+2}\ln^{2}\frac{1+x}{2}J_{2k}^{(a,a)}(x)^{2}\differential x$ (52) $\displaystyle\mathrm{A_{2}}$ $\displaystyle=\sum_{k=0}^{m-1}\frac{1}{h_{k}}\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a+1}\left(\frac{1+x}{2}\right)^{a+1}\ln\frac{1-x}{2}\ln\frac{1+x}{2}J_{2k}^{(a,a)}(x)^{2}\differential x.$ Here, we recall that $a=n-m\geq 0$ denotes the subsystem difference of fermionic Gaussian states with arbitrary number of particles. Similarly, the integral $\mathrm{I_{B}}=\int_{0}^{1}\int_{0}^{1}v(x)v(y)K^{2}\left(x,y\right)\differential x\differential y$ (54) can be written in terms of the following two integrals $\mathrm{I_{B}}=\mathrm{B_{1}}+\mathrm{B_{2}},$ (55) where $\displaystyle\mathrm{B_{1}}=$ $\displaystyle\sum_{k=0}^{m-1}\frac{1}{h_{k}^{2}}\left(\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{a+1}\ln\frac{1+x}{2}J_{2k}^{(a,a)}(x)^{2}\differential x\right)^{2}$ (56) $\displaystyle\mathrm{B_{2}}=$ $\displaystyle\sum_{j=1}^{m-1}\sum_{k=0}^{m-j-1}\frac{2}{h_{k+j}h_{k}}\Bigg{(}\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{a+1}\ln\frac{1+x}{2}$ $\displaystyle\times J_{2k+2j}^{(a,a)}(x)J_{2k}^{(a,a)}(x)\differential x\Bigg{)}^{2}.$ (57) Computing the above integrals $\mathrm{A_{1}}$, $\mathrm{A_{2}}$, $\mathrm{B_{1}}$, and $\mathrm{B_{2}}$ requires the following two integral identities. The first one is $\displaystyle\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a_{1}}\left(\frac{1+x}{2}\right)^{c}J^{(a_{1},b_{1})}_{k_{1}}(x)J^{(a_{2},b_{2})}_{k_{2}}(x)\differential x$ $\displaystyle=\frac{2\left(k_{1}+1\right)_{a_{1}}}{\left(b_{2}+k_{2}+1\right)_{a_{2}}}\sum_{i=0}^{k_{2}}\frac{(-1)^{i+k_{2}}(i+1)_{c}\left(i+b_{2}+1\right)_{a_{2}+k_{2}}}{\Gamma\left(k_{2}-i+1\right)\Gamma\left(a_{1}+c+i+k_{1}+2\right)}$ $\displaystyle~{}~{}~{}\\!~{}\times\left(c+i-b_{1}-k_{1}+1\right)_{k_{1}},\quad\real(a_{1},a_{2},b_{1},b_{2},c)>-1.$ (58) To show this identity, we first note that the Jacobi polynomial $J_{k}^{(a,b)}(x)$ supported in $x\in[-1,1]$ admits different representations Szego ; Forrester $\displaystyle J^{(a,b)}_{k}(x)$ $\displaystyle=\frac{(-1)^{k}(b+1)_{k}}{k!}\sum_{i=0}^{k}\frac{(-k)_{i}(k+a+b+1)_{i}}{(b+1)_{i}\Gamma(i+1)}\left(\frac{1+x}{2}\right)^{i}$ (59) $\displaystyle=\sum_{i=0}^{k}\frac{(-1)^{i}\Gamma(a+k+1)(k+b-i+1)_{i}}{\Gamma(i+1)\Gamma(a+i+1)\Gamma(k-i+1)}\left(\frac{1-x}{2}\right)^{i}\left(\frac{1+x}{2}\right)^{k-i}.$ The identity (58) is then obtained by using the definition (59) for the polynomial $J^{(a_{2},b_{2})}_{k_{2}}$ before applying the well-known integral identity Szego ; Forrester $\displaystyle\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{c}J^{(a,b)}_{k}(x)\differential x$ $\displaystyle=\frac{2\Gamma(c+1)(k+1)_{a}(c-b-k+1)_{k}}{\Gamma(a+c+k+2)},\quad\real(a,b,c)>-1.$ (61) The second identity is $\displaystyle\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{d}\left(\frac{1+x}{2}\right)^{c}J^{(a_{1},b_{1})}_{k_{1}}J^{(a_{2},b_{2})}_{k_{2}}(x)\differential x$ $\displaystyle=\frac{2\Gamma\left(a_{2}+k_{2}+1\right)\Gamma\left(b_{2}+k_{2}+1\right)}{\Gamma\left(c+d+k_{1}+k_{2}+2\right)}\sum_{i=0}^{k_{2}}\frac{(-1)^{i}\Gamma\left(d-a_{1}+i+1\right)}{\Gamma(i+1)\Gamma\left(a_{2}+i+1\right)}$ $\displaystyle~{}~{}~{}\\!~{}\times\frac{\Gamma\left(c-b_{1}-i+k_{2}+1\right)}{\Gamma\left(k_{2}-i+1\right)\Gamma\left(b_{2}-i+k_{2}+1\right)}\sum_{j=0}^{k_{1}}\frac{(-1)^{j}\left(k_{1}-j+1\right)_{d+i}}{\Gamma(j+1)}$ $\displaystyle~{}~{}~{}\\!~{}\times\frac{\left(c-i+j-b_{1}-k_{1}+k_{2}+1\right)_{b_{1}+k_{1}}}{\Gamma\left(d-a_{1}+i-j+1\right)},\quad\real(a_{1},a_{2},b_{1},b_{2},c,d)>-1,$ (62) which is obtained by using the definition (LABEL:eq:J2) for the polynomial $J^{(a_{2},b_{2})}_{k_{2}}$ before applying the identity (HW22, , Equation (62)) $\displaystyle\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{d}\left(\frac{1+x}{2}\right)^{c}J^{(a,b)}_{k}(x)\differential x$ $\displaystyle=\frac{2\Gamma(c-b+1)\Gamma(d-a+1)}{\Gamma(c+d+k+2)}\sum_{i=0}^{k}\frac{(-1)^{i}\Gamma(c+i+1)\Gamma(d-i+k+1)}{\Gamma(i+1)\Gamma(k-i+1)}$ $\displaystyle~{}~{}~{}\\!~{}\times\frac{1}{\Gamma(d-a-i+1)\Gamma(c-b+i-k+1)},\quad\real(a,b,c,d)>-1.$ (63) The integral $\mathrm{A_{1}}$ is now calculated by applying the identity (58), where we need to assign $a_{1}=b_{1}=a_{2}=b_{2}=a,~{}~{}~{}k_{1}=k_{2}=2k,$ (64) and take twice derivatives of $c$ before setting $c=a+2$. Under the same specialization (64), the integral $\mathrm{A_{2}}$ is calculated by taking derivatives of both $c$ and $d$ of the identity (62) before setting $c=d=a+1$, while the integral $\mathrm{B_{1}}$ is calculated by taking derivative of $c$ of the identity (58) before setting $c=a+1$. According to the result (57), the integral $\mathrm{B_{2}}$ is calculated by specializing $a_{1}=b_{1}=a_{2}=b_{2}=a,~{}~{}~{}k_{1}=2k+2j,~{}~{}~{}k_{2}=2k,$ (65) in the identity (58), and taking derivative of $c$ before setting $c=a+1$. In writing down the summation forms of $\mathrm{A_{1}}$, $\mathrm{A_{2}}$, $\mathrm{B_{1}}$, and $\mathrm{B_{2}}$, one will have to resolve the indeterminacy by using the following asymptotic expansions of gamma and polygamma functions of negative arguments AS72 when $\epsilon\rightarrow 0$, $\displaystyle\Gamma(-l+\epsilon)$ $\displaystyle=\frac{(-1)^{l}}{l!\epsilon}\left(1+\psi_{0}(l+1)\epsilon+o\left(\epsilon^{2}\right)\right)$ (66) $\displaystyle\psi_{0}(-l+\epsilon)$ $\displaystyle=-\frac{1}{\epsilon}+\psi_{0}(l+1)+\left(2\psi_{1}(1)-\psi_{1}(l+1)\right)\epsilon+o\left(\epsilon^{2}\right)$ (67) $\displaystyle\psi_{1}(-l+\epsilon)$ $\displaystyle=\frac{1}{\epsilon^{2}}-\psi_{1}(l+1)+\psi_{1}(1)+\zeta(2)+o\left(\epsilon\right).$ (68) The resulting summation forms of $\mathrm{A_{1}}$, $\mathrm{A_{2}}$, $\mathrm{B_{1}}$, and $\mathrm{B_{2}}$ are summarized in (A.1)–(A.1) in Appendix A. #### Case B: Fixed number of particles By the definition of correlation kernel (42), the $\mathrm{I_{A}}$ integral (48) of case B boils down to computing the two parts $\mathrm{I_{A}}=\mathcal{A}_{1}+\mathcal{A}_{2},$ (69) where $\displaystyle\mathcal{A}_{1}=$ $\displaystyle\sum_{k=0}^{m-1}\frac{1}{h_{k}}\int_{-1}^{1}\left(\left(\frac{1+x}{2}\right)^{2}\ln^{2}\frac{1+x}{2}+\left(\frac{1-x}{2}\right)^{2}\ln^{2}\frac{1-x}{2}\right)$ $\displaystyle\times\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b}J_{k}^{(a,b)}(x)^{2}\differential x$ (70) $\displaystyle\mathcal{A}_{2}=$ $\displaystyle\sum_{k=0}^{m-1}\frac{2}{h_{k}}\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a+1}\left(\frac{1+x}{2}\right)^{b+1}\ln\frac{1-x}{2}\ln\frac{1+x}{2}J_{k}^{(a,b)}(x)^{2}\differential x.$ Due to the parity property (45), $\mathcal{A}_{1}$ admits a symmetric structure as $\mathcal{A}_{1}=\mathcal{A}_{1}^{(a,b)}+\mathcal{A}_{1}^{(b,a)},$ (72) where $\mathcal{A}_{1}^{(a,b)}=\sum_{k=0}^{m-1}\frac{1}{h_{k}}\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b+2}\ln^{2}\frac{1+x}{2}J_{k}^{(a,b)}(x)^{2}\differential x.$ (73) The summations in (LABEL:eq:fpA2) and (73) can be evaluated by using the following confluent form of Christoffel-Darboux formula Forrester $\sum_{k=0}^{m-1}\frac{J_{k}^{(a,b)}(x)^{2}}{h_{k}}=\alpha_{1}J_{m-1}^{(a+1,b+1)}(x)J_{m-1}^{(a,b)}(x)-\alpha_{2}J_{m-2}^{(a+1,b+1)}(x)J_{m}^{(a,b)}(x),$ (74) where $\displaystyle\alpha_{1}$ $\displaystyle=\frac{m(a+b+m)(a+b+m+1)}{h_{m-1}(a+b+2m-1)_{2}}$ (75) $\displaystyle\alpha_{2}$ $\displaystyle=\frac{m(a+b+m)^{2}}{h_{m-1}(a+b+2m-1)_{2}}.$ (76) Consequently, we have $\displaystyle\mathcal{A}_{1}^{(a,b)}=~{}$ $\displaystyle\alpha_{1}\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b+2}\ln^{2}\frac{1+x}{2}J_{m-1}^{(a+1,b+1)}(x)J_{m-1}^{(a,b)}(x)\differential x$ $\displaystyle\\!-\alpha_{2}\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b+2}\ln^{2}\frac{1+x}{2}J_{m-2}^{(a+1,b+1)}(x)J_{m}^{(a,b)}(x)\differential x$ $\displaystyle\mathcal{A}_{2}=~{}$ $\displaystyle 2\alpha_{1}\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a+1}\left(\frac{1+x}{2}\right)^{b+1}\ln\frac{1-x}{2}\ln\frac{1+x}{2}$ $\displaystyle\\!\times J_{m-1}^{(a+1,b+1)}(x)J_{m-1}^{(a,b)}(x)\differential x$ $\displaystyle\\!-2\alpha_{2}\int_{-1}^{1}\left(\frac{1-x}{2}\right)^{a+1}\left(\frac{1+x}{2}\right)^{b+1}\ln\frac{1-x}{2}\ln\frac{1+x}{2}$ $\displaystyle\\!\times J_{m-2}^{(a+1,b+1)}(x)J_{m}^{(a,b)}(x)\differential x.$ (78) Similarly, the $\mathrm{I_{B}}$ integral (49) can be written in terms of the following two integrals $\mathrm{I_{B}}=\mathcal{B}_{1}+\mathcal{B}_{2},$ (79) where $\displaystyle\mathcal{B}_{1}=$ $\displaystyle\sum_{k=0}^{m-1}\frac{1}{h_{k}^{2}}\Bigg{(}\int_{-1}^{1}\left(\frac{1-x}{2}\ln\frac{1-x}{2}+\frac{1+x}{2}\ln\frac{1+x}{2}\right)\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b}$ $\displaystyle\times J_{k}^{(a,b)}(x)^{2}\differential x\Bigg{)}^{2}$ (80) $\displaystyle\mathcal{B}_{2}=$ $\displaystyle\sum_{j=1}^{m-1}\sum_{k=0}^{m-j-1}\frac{2}{h_{k+j}h_{k}}\Bigg{(}\int_{-1}^{1}\left(\frac{1-x}{2}\ln\frac{1-x}{2}+\frac{1+x}{2}\ln\frac{1+x}{2}\right)~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}$ $\displaystyle\times\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b}J_{k+j}^{(a,b)}(x)J_{k}^{(a,b)}(x)\differential x\Bigg{)}^{2}.$ (81) Note that the integrals in $\mathcal{B}_{1}$ or $\mathcal{B}_{2}$ also admit the same symmetric structure as (72) due to the parity property of Jacobi polynomials. Because of the symmetry, the remaining computation process is similar to that of case A. Specifically, the integrals in $\mathcal{A}_{1}^{(a,b)}$ are calculated by taking twice derivative of $c$ of the identity (58), where one needs to assign $a_{1}=a,~{}~{}~{}b_{1}=b,~{}~{}~{}a_{2}=a+1,~{}~{}~{}b_{2}=b+1,~{}~{}~{}k_{1}=k_{2}=m-1,$ (82) and $a_{1}=a,~{}~{}~{}b_{1}=b,~{}~{}~{}a_{2}=a+1,~{}~{}~{}b_{2}=b+1,~{}~{}~{}k_{1}=m,~{}~{}~{}k_{2}=m-2,$ (83) respectively corresponding to the first and second integrals in (LABEL:eq:A1cd) before setting $c=b+2$. By the result (78), the two integrals in $\mathcal{A}_{2}$ are calculated by taking derivatives of $c$ and $d$ of identity (62) respectively with the specializations (82) and (83), before setting $c=b+1$, $d=a+1$. The integral $\int_{-1}^{1}\frac{1+x}{2}\ln\frac{1+x}{2}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b}J_{k}^{(a,b)}(x)^{2}\differential x$ (84) in $\mathcal{B}_{1}$ is calculated by specializing $a_{1}=a_{2}=a,~{}~{}~{}b_{1}=b_{2}=b,~{}~{}~{}k_{1}=k_{2}=k$ (85) in the identity (58), and taking derivative of $c$ before setting $c=b+1$. The integral $\int_{-1}^{1}\frac{1+x}{2}\ln\frac{1+x}{2}\left(\frac{1-x}{2}\right)^{a}\left(\frac{1+x}{2}\right)^{b}J_{k+j}^{(a,b)}(x)J_{k}^{(a,b)}(x)\differential x$ (86) in $\mathcal{B}_{2}$ is calculated by specializing $a_{1}=a_{2}=a,~{}~{}~{}b_{1}=b_{2}=b~{}~{}~{}k_{1}=k+j,~{}~{}~{}k_{2}=k$ (87) in the identity (58), and taking derivative of $c$ before setting $c=b+1$. After resolving the indeterminacy of gamma and polygamma functions by using (66)–(68), one arrives at the summation representations (281)–(A.2) as listed in Appendix A. ### 2.2 Tools for simplification of summations The major task in obtaining the variance formulas in propositions 3 and 4 is to simplify the summation representations (A.1)–(A.2) into the closed-form results (3) and (4). We summarize the necessary simplification tools in this section, where the existing simplification framework is briefly reviewed in Section 2.2.1 and the new simplification framework is introduced in Section 2.2.2. #### 2.2.1 Existing simplification framework The existing simplification tools have been utilized in various moments calculations over different ensembles, including the Hilbert-Schmidt ensemble Wei17 ; Wei20 ; HWC21 ; Wei22 , the Bures-Hall ensemble Wei20BH ; LW21 , and the fermionic Gaussian ensemble HW22 . In the existing framework, we mainly have two types of finite sum identities. The first type is of the form $\sum_{i=1}^{m}i^{c}\psi_{j_{1}}^{b_{1}}(i+a_{1})\psi_{j_{2}}^{b_{2}}(i+a_{2})\cdots\psi_{j_{m}}^{b_{m}}(i+a_{m}),$ (88) where $a$, $b$, $c$, $j$ are non-negative integers. The main idea in deriving the identities of this type of sums is to change the summation orders and make use of the obtained lower order summation formulas in a recursive manner. For example, the summation $\sum_{i=1}^{m}\psi_{0}^{2}(i+a)$ (89) will be derived by using the identity (287) of the lower order summation $\sum_{i=1}^{m}\psi_{0}(i+a)$. Specifically, by using the finite sum form of the digamma function $\psi_{0}(l)=-\gamma+\sum_{k=1}^{l-1}\frac{1}{k},$ (90) the summation (89) can be rewritten as $\psi_{0}(a)\sum_{i=1}^{m}\psi_{0}(i+a)+\sum_{j=1}^{m}\frac{1}{a+j-1}\sum_{i=j}^{m}\psi_{0}(i+a),$ (91) where we have changed the summation order of the double sum. The remaining sums can be simplified by using the lower order identity (287), leading to the result in (B.1). The second type of summations is of the form $S_{f}(m,n)=\sum_{i=1}^{m}\frac{(n-i)!}{(m-i)!}f(i),\qquad m\leq n,$ (92) where $f(i)$ can be the product of polygamma and rational functions in $i$. A well-known result of the second type summation is the identity $\sum_{i=1}^{m}\frac{(n-i)!}{(m-i)!}=\frac{n!}{(m-1)!(n-m+1)},$ (93) which is a special case of the Chu-Vandermonde identity Luke . In the existing simplification framework, the identity (93) is a fundamental result in obtaining several other second type summations. For example, when $f(i)=\frac{1}{i},$ (94) the summation $S_{1/i}(m,n)=\sum_{i=1}^{m}\frac{(n-i)!}{(m-i)!}\frac{1}{i}$ (95) is computed by first obtaining the recurrence relation $S_{1/i}(m,n)=\frac{n}{m}S_{1/i}(m-1,n-1)+\frac{n-m}{m}\sum_{i=1}^{m}\frac{(n-1-i)!}{(m-i)!}.$ (96) After recurring $m$ times, and using the existing result (93), one obtains the closed-form result of the summation (95) as listed in (305). #### 2.2.2 New simplification framework The existing simplification framework is useful in simplifying the summations in (A.1), (A.1), (281), and (A.2). However, the framework is insufficient to simplify some of the summations in (A.1), (A.1), (A.2), and (A.2). The main technique difficulty is that these summations, after exhausting all the possibilities of changing the order of summations, are not related to the first or second type summations of the existing framework. To convert these sums into the ones of the existing framework, the following new simplification framework is needed. The first technique in the new framework is what we refer to as “dummy summation". The idea of the dummy summation is as follows. For a summation $F$, $F=\sum_{i}f_{i},$ (97) where $\sum_{i}=\sum_{i_{1}}\sum_{i_{2}}\dots\sum_{i_{a}}$ (98) denotes a finite nested sum of $a$ indexes $i=\\{i_{1},i_{2},\dots,i_{a}\\}$. Each index $i_{k}$, $k=1,2,\dots,a$ may depend on its previous ones $i_{1},i_{2},\dots,i_{k-1}$. For the summation $F$ that is not simplifiable under the existing framework by any changes of the order of summations, one may interpret the summand $f_{i}$ into an additional nested finite sum over a set of new indexes $j=\\{j_{1},j_{2},\dots,j_{b}\\}$ as $f_{i}=\sum_{j}g_{i,j},$ (99) such that the resulting representation $F=\sum_{i,j}g_{i,j}$ (100) admits further simplifications by using the existing tools of first and second type sums (88) and (92) after appropriately changing some of the summation orders. Note that the indexes in the set $j$ in (99) may depend on the ones in the set $i$. To illustrate this new technique, we consider the following two examples that will be utilized in the simplifications in Section 2.3. The first example is the summation $F=\sum_{i_{1}=1}^{m}f_{i_{1}},$ (101) where $f_{i_{1}}=i_{1}\psi_{0}(i_{1}+a).$ (102) In this case, we have $i=\\{i_{1}\\}$ in the definition (97). By interpreting $i_{1}$ in the summand (102) as a dummy sum $i_{1}=\underbrace{1+1+\dots+1}_{i_{1}},$ (103) the original single sum in (101) now becomes a double sum $F=\sum_{i_{1}=1}^{m}\sum_{j_{1}=1}^{i_{1}}g_{i_{1},j_{1}},$ (104) where $g_{i_{1},j_{1}}=\psi_{0}(i_{1}+a)$ (105) with $j=\\{j_{1}\\}$ in the definition (100). After changing the summation order in (104) as $F=\sum_{j_{1}=1}^{m}\sum_{i_{1}=j_{1}}^{m}\psi_{0}(i_{1}+a),$ (106) the sum over $i_{1}$ can be simplified by using an existing identity (287), leading to $F=\frac{1-a}{2}\sum_{j_{1}=1}^{m}\psi_{0}\left(j_{1}+a\right)+\frac{m}{4}\left(2(a+m)\psi_{0}(a+m+1)-m-1\right).$ (107) Simplifying the single summation in (107) by the identity (287) directly gives the result listed in (B.1). In the above example, the dummy summation (103) converts the original sum (101) into a double sum (104), which appears more complicated but is simplifiable using an available identity (287) of the existing framework. Note also that the dummy summation technique may not be critical in this example, which may be derived in other ways. In the second example, we introduce a dummy summation (99) that is crucial as a subroutine in simplifying the summations in (A.1) and (A.2). The essential idea in creating this dummy summation is to interpret a Gamma ratio $\frac{\Gamma(i)}{\Gamma(c+i)}=\frac{1}{i(i+1)\cdots(i+c-1)},\qquad c,i\in\mathbb{Z}^{+}$ (108) as its partial fraction decomposition $a_{1}\frac{1}{i}+a_{2}\frac{1}{i+1}+\dots+a_{c}\frac{1}{i+c-1}$ (109) with $a_{j},j=1,2,\dots,c$ denoting the coefficients of the decomposition. The dummy summation that corresponds to the above interpretation is $\frac{\Gamma(i)}{\Gamma(c+i)}=\sum_{j=1}^{c}a_{j}\frac{1}{i+j-1},$ (110) where $a_{j}=\frac{(-1)^{j+1}}{\Gamma(j)\Gamma(c-j+1)}.$ (111) The coefficients $a_{j}$ are computed by evaluating a unit-argument hypergeometric function Brychkov08 $\sum_{j=1}^{c}a_{j}\frac{1}{i+j-1}=\frac{\\!{}_{2}F_{1}(-c,i-1\mathchar 24635\relax\;i\mathchar 24635\relax\;1)}{\Gamma(c+1)},$ (112) where one utilizes the well-known identity Brychkov08 $_{2}F_{1}(a,b\mathchar 24635\relax\;c\mathchar 24635\relax\;1)=\frac{\Gamma(c)\Gamma(c-a-b)}{\Gamma(c-a)\Gamma(c-b)},\quad\real(c)>\real(a+b).$ (113) The dummy summation in (110) is critical in proving the lemmas 2–4 as discussed later in this section. By utilizing these lemmas, the summations in (A.1) and (A.2) involving ratios of Gamma functions then become simplifiable under the existing simplification framework as will be shown in Section 2.3. The second technique in the new framework is what we refer to as the “re- summation" technique. The idea of this new technique is as follows. For a finite summation $G_{m}$ that appears not summable under the existing framework with $m$ denoting one of its parameters or the finite upper limit of the summation. The re-summation technique aims to find alternative forms of the summation $G_{m}$ to reveal the potential cancellations with other sums, such that the remaining terms can be simplified by using the existing tools. Specifically, the re-summation of $G_{m}$ can be generated by iterating a suitably chosen recurrence relation $G_{m}=c_{m-1}G_{m-1}+r_{m-1},$ (114) where $c_{i}$ denotes the coefficient and $r_{i}$ is the remainder of the recurrence relation. Here, the remainder can contain summations, see the example (121) below. Each iteration is to replace the term $G_{m-i}$ with its previous one $G_{m-i-1}$. Keep iterating until $G_{m-i}$ vanishes, we then obtain an alternative form of $G_{m}$, which is considered a useful one if it facilitates the cancellation with other sums. To illustrate the re-summation technique, we consider the difference of two summations $\mathcal{G}=G_{m}-G^{\prime}_{m},$ (115) where $\displaystyle G_{m}$ $\displaystyle=\sum_{i=1}^{m}\frac{(-1)^{i}\Gamma(a-i+m+1)}{\Gamma(m-i+1)(a-i+2m+1)}$ (116) $\displaystyle G^{\prime}_{m}$ $\displaystyle=\sum_{i=1}^{m}\frac{(-1)^{i}\Gamma(a-i+m+1)}{\Gamma(m-i+1)i}.$ (117) We first notice that the two summations in (116) and (117) above can not be simplified individually by using the identities of the existing framework. In addition, these two summations do not appear to cancel directly in the combination (115). By iterating $m$ times the following tailor made recurrence relations (in terms of the various choices of coefficients and reminders), cf. (114), $\displaystyle G_{m}$ $\displaystyle=c_{m-1}G_{m-1}+r_{m-1}$ (118) $\displaystyle G^{\prime}_{m}$ $\displaystyle=c_{m-1}G^{\prime}_{m-1}+r^{\prime}_{m-1},$ (119) where $\displaystyle c_{m-1}$ $\displaystyle=\frac{a+m}{m}$ (120) $\displaystyle r_{m-1}$ $\displaystyle=-\frac{a}{m}\sum_{i=1}^{m+1}\frac{(-1)^{i}\Gamma(a-i+m+1)}{\Gamma(m-i+2)}+\frac{a+m}{m}\left(\frac{\Gamma(a+m)}{\Gamma(m)(a+2m-1)}\right.$ $\displaystyle~{}~{}~{}\\!~{}\\!\left.-\frac{\Gamma(a+m+1)}{\Gamma(m+1)(a+2m)}\right)$ (121) $\displaystyle r^{\prime}_{m-1}$ $\displaystyle=\frac{a}{m}\sum_{i=1}^{m}\frac{(-1)^{i}\Gamma(a-i+m)}{\Gamma(m-i+1)},$ (122) we obtain the re-summation results $\displaystyle G_{m}=~{}\\!$ $\displaystyle\frac{a\Gamma(a+m+1)}{\Gamma(m+1)}\sum_{i=1}^{m}\sum_{j=1}^{i+1}\frac{(-1)^{j+1}\Gamma(i)\Gamma(a+i-j+1)}{\Gamma(i-j+2)\Gamma(a+i+1)}$ $\displaystyle+\frac{\Gamma(a+m+1)}{\Gamma(m+1)}\sum_{i=1}^{m}\left(\frac{1}{a+2i-1}-\frac{a+i}{i(a+2i)}\right)$ (123) $\displaystyle G^{\prime}_{m}=~{}\\!$ $\displaystyle\frac{a\Gamma(a+m+1)}{\Gamma(m+1)}\sum_{i=1}^{m}\sum_{j=2}^{i+1}\frac{(-1)^{j+1}\Gamma(i)\Gamma(a+i-j+1)}{\Gamma(i-j+2)\Gamma(a+i+1)}.$ (124) The relationship between the two summations now become obvious. After inserting the results (123) and (124) into (115), the cancellation occurs directly between the two double summations in (123) and (124). The remaining terms give the closed-form result $\displaystyle\sum_{i=1}^{m}\frac{(-1)^{i}\Gamma(a-i+m+1)}{\Gamma(m-i+1)}\left(\frac{1}{a-i+2m+1}-\frac{1}{i}\right)$ $\displaystyle=\frac{\Gamma(a+m+1)}{\Gamma(m+1)}\left(\psi_{0}(a+2m+1)-\psi_{0}(a+m+1)\right),$ (125) which is a special case of Lemma 5 when $c=0$. This result is also utilized in simplifying the double summations in (A.2). Note that the emphasis here is the inner cancellation among multiple sums when utilizing re-summation technique. On the other hand, this technique concerning only one summation has been studied in Wei17 ; Wei20 ; HWC21 . Using the two techniques in the new simplification framework, we obtain the following six lemmas. More precisely, proving the Lemma 1 utilizes the re- summation technique. The lemmas 2–4 are obtained by using the dummy summation together with the result of Lemma 1. The Lemma 5 and Lemma 6 are established by using the re-summation technique. In the six lemmas obtained by the new framework, each single sum is converted to another one, which will be applied in Section 2.3 to convert summations involving multiple ratios of Gamma functions into the ones that are simplifiable by the existing framework. ###### Lemma 1. For any complex numbers $a,b,c\notin\mathbb{Z}^{-}$, we have $\displaystyle\sum_{i=1}^{m}\frac{1}{\Gamma(i)\Gamma(a+i)\Gamma(m+1-i)\Gamma(m+b+1-i)(c+i)}$ $\displaystyle=\frac{1}{\Gamma(b+m)\Gamma(c+m+1)\Gamma(a+b+m)}\sum_{i=1}^{m}\frac{\Gamma(c-i+m+1)\Gamma(a+b-i+2m)}{\Gamma(m-i+1)\Gamma(a-i+m+1)}.$ (126) Proof: Proving the Lemma 1 uses the re-summation technique. The left side of the identity (1) can be rewritten as $G_{m}=\sum_{i=1}^{m}\frac{1}{\Gamma(i)\Gamma(a+i)\Gamma(m-i)\Gamma(m+b+1-i)(m-i)(c+i)}.$ (127) In (127), by performing a partial fraction decomposition $\frac{1}{(m-i)(c+i)}=\frac{1}{c+m}\left(\frac{1}{c+i}+\frac{1}{m-i}\right),$ (128) we obtain the recurrence relation, cf. (114), $G_{m}=c_{m-1}G_{m-1}+r_{m-1},$ (129) where $\displaystyle c_{m-1}=$ $\displaystyle\frac{1}{c+m}$ (130) $\displaystyle r_{m-1}=$ $\displaystyle\frac{1}{c+m}\sum_{i=1}^{m}\frac{1}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(m+b+1-i)}.$ (131) The summation in (131) is related to a hypergeometric function $\,{}_{2}F_{1}(1-b-m,1-m\mathchar 24635\relax\;a+1\mathchar 24635\relax\;1)$ that admits a closed-form representation by using the identity (113). Therefore, we have $\displaystyle\sum_{i=1}^{m}\frac{1}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(m+b+1-i)}$ $\displaystyle=\frac{\Gamma(a+b+2m-1)}{\Gamma(m)\Gamma(a+m)\Gamma(b+m)\Gamma(a+b+m)},$ (132) and $r_{m-1}=\frac{\Gamma(a+b+2m-1)}{(c+m)\Gamma(m)\Gamma(a+m)\Gamma(b+m)\Gamma(a+b+m)}.$ (133) Iterate $m$ times the recurrence relation (129), the desired identity (1) is established. This completes the proof of Lemma 1. ###### Lemma 2. For any complex numbers $a,b\notin\mathbb{Z^{-}}$, and any $c\in\mathbb{Z^{+}}$, we have $\displaystyle\sum_{i=1}^{m}\frac{1}{\Gamma(c+i)\Gamma(a+i)\Gamma(m+1-i)\Gamma(m+b+1-i)}$ $\displaystyle=\frac{1}{\Gamma(m+b)\Gamma(m+a+b)\Gamma(c)\Gamma(m+c)}$ $\displaystyle~{}~{}~{}~{}\times\sum_{i=1}^{m}\frac{\Gamma(m+a+b+i-1)\Gamma(m+c-i)}{\Gamma(a+i)\Gamma(m-i+1)}.$ (134) Proof: We prove Lemma 2 by using the dummy summation technique. Denote $F=\sum_{i=1}^{m}f_{i}$ (135) with $f_{i}=\frac{1}{\Gamma(c+i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b+m-i+1)}.$ (136) We first introduce an additional summation in (135) by replacing $\Gamma(i)/\Gamma(c+i)$ with the summation form in (110). The summation $F$ now becomes $F=\sum_{i=1}^{m}\sum_{j=1}^{c}g_{i,j}$ (137) with $g_{i,j}=\frac{(-1)^{j+1}}{\Gamma(j)\Gamma(c-j+1)(i+j-1)}\frac{1}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b+m-i+1)}.$ (138) In the double summation (137), we apply the Lemma 1 to first evaluate the sum over $i$, leading to $\displaystyle F=$ $\displaystyle\frac{1}{\Gamma(b+m)\Gamma(a+b+m)}\sum_{i=1}^{m}\frac{\Gamma(a+b-i+2m)}{\Gamma(m-i+1)\Gamma(a-i+m+1)}$ $\displaystyle\times\sum_{j=1}^{c}\frac{(-1)^{j+1}\Gamma(m-i+j)}{\Gamma(j)\Gamma(c-j+1)\Gamma(j+m)},$ (139) where it is now possible to evaluate the sum over j as $\displaystyle\sum_{j=1}^{c}\frac{(-1)^{j+1}\Gamma(m-i+j)}{\Gamma(j)\Gamma(c-j+1)\Gamma(j+m)}$ $\displaystyle=\frac{\Gamma(m-i+1)}{\Gamma(c)\Gamma(m+1)}\,_{2}F_{1}(1-c,m-i+1\mathchar 24635\relax\;m+1\mathchar 24635\relax\;1).$ (140) $\displaystyle=\frac{\Gamma(c+i-1)\Gamma(m-i+1)}{\Gamma(c)\Gamma(i)\Gamma(c+m)}.$ (141) The result (141) is obtained by using the identity (113) to evaluate the unit argument hypergeometric function in (140). Inserting (141) into (139) and shifting the index $i\to m-i+1$, Lemma 2 is proved. ###### Lemma 3. For any complex numbers $a,b\notin\mathbb{Z^{-}}$, and any $c\in\mathbb{Z^{+}}$, we have $\displaystyle\sum_{i=1}^{m}\frac{1}{\Gamma(c+i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)i}$ $\displaystyle=\frac{1}{\Gamma(a)\Gamma(a+m)\Gamma(1+b+m)\Gamma(b+c+m)}\sum_{i=1}^{m}\frac{\Gamma(a-i+m)\Gamma(b+c+i+m)}{\Gamma(c+i)\Gamma(m-i+1)i}$ $\displaystyle~{}~{}~{}\\!~{}+\frac{\psi_{0}(a)-\psi_{0}(a+m)}{\Gamma(a)\Gamma(c)\Gamma(m+1)\Gamma(b+m+1)}.$ (142) Proof: Similar to the proof of Lemma 2, we introduce an additional summation in $F=\sum_{i=1}^{m}\frac{1}{\Gamma(c+i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)i}$ (143) by replacing the term $\Gamma(i)/\Gamma(c+i)$ with the dummy summation (110). Consequently, one has $F=\sum_{i=1}^{m}\sum_{j=1}^{c}g_{i,j}$ (144) with $g_{i,j}=\frac{(-1)^{j+1}}{\Gamma(j)\Gamma(c-j+1)(i+j-1)}\frac{1}{\Gamma(i+1)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)},$ (145) which is the same form as the summand in (138). Therefore, one uses the results (137)–(140) to obtain the identity (142), and Lemma 3 is proved. ###### Lemma 4. For any complex numbers $a,b\notin\mathbb{Z^{-}}$, and any $c,d\in\mathbb{Z^{+}}$, we have $\displaystyle\sum_{i=1}^{m}\frac{1}{\Gamma(c+i)\Gamma(a+i)\Gamma(d+m-i+1)\Gamma(b+m-i+1)}$ $\displaystyle=\frac{1}{\Gamma(d)\Gamma(a+m)\Gamma(a+b+m)\Gamma(c+d+m)}\sum_{i=1}^{m}\frac{\Gamma(c+d+i-1)\Gamma(a+b-i+2m)}{\Gamma(c+i)\Gamma(b-i+m+1)}$ $\displaystyle~{}~{}~{}+\frac{1}{\Gamma(c)\Gamma(b+m)\Gamma(a+b+m)\Gamma(c+d+m)}\sum_{i=1}^{m}\frac{\Gamma(c+d+i-1)\Gamma(a+b-i+2m)}{\Gamma(d+i)\Gamma(a-i+m+1)}.$ Proof: Proving the Lemma 4 also utilizes the dummy summation technique. Denote the left side of the identity (LABEL:eq:lemma4) as $F=\sum_{i=1}^{m}f_{i},$ (147) with $f_{i}=\frac{1}{\Gamma(c+i)\Gamma(a+i)\Gamma(d+m-i+1)\Gamma(b+m-i+1)}.$ (148) By interpreting respectively the gamma ratios $\Gamma(i)/\Gamma(c+i)$ and $\Gamma(m-i+1)/\Gamma(d+m-i+1)$ as the dummy summations (110) and $\sum_{k=1}^{d}\frac{(-1)^{k+1}}{\Gamma(k)\Gamma(d-k+1)(m-i+k)},$ (149) the summation $F$ can be written as $F=\sum_{i=1}^{m}\sum_{j=1}^{c}\sum_{k=1}^{d}g_{i,j,k},$ (150) where $\displaystyle g_{i,j,k}=~{}\\!\\!$ $\displaystyle\frac{1}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)}$ $\displaystyle\times\frac{(-1)^{j+1}}{\Gamma(j)\Gamma(c-j+1)(i+j-1)}\frac{(-1)^{k+1}}{\Gamma(k)\Gamma(d-k+1)(m-i+k)}.$ (151) After taking a partial fraction decomposition $\frac{1}{(m-i+k)(i+j-1)}=\frac{1}{(i+j-1)(j+k+m-1)}+\frac{1}{(k+m-i)(j+k+m-1)}$ (152) in (2.2.2), the summation (150) is split into two sums $F=F_{1}+F_{2},$ (153) where $\displaystyle F_{1}=~{}\\!\\!$ $\displaystyle\sum_{j=1}^{c}\frac{(-1)^{j+1}}{\Gamma(j)\Gamma(c-j+1)}\sum_{k=1}^{d}\frac{(-1)^{k+1}}{\Gamma(k)\Gamma(d-k+1)(j+k+m-1)}$ $\displaystyle\times\sum_{i=1}^{m}\frac{1}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)(i+j-1)}$ (154) $\displaystyle F_{2}=~{}\\!\\!$ $\displaystyle\sum_{k=1}^{d}\frac{(-1)^{k+1}}{\Gamma(k)\Gamma(d-k+1)}\sum_{j=1}^{c}\frac{(-1)^{j+1}}{\Gamma(j)\Gamma(c-j+1)(j+k+m-1)}$ $\displaystyle\times\sum_{i=1}^{m}\frac{1}{\Gamma(i)\Gamma(b+i)\Gamma(m-i+1)\Gamma(a-i+m+1)(i+k-1)}.$ (155) Specialize $c=d$ and $i=j+m$ in the identity (110), one has $\sum_{k=1}^{d}\frac{(-1)^{k+1}}{\Gamma(k)\Gamma(d-k+1)(j+k+m-1)}=\frac{\Gamma(j+m)}{\Gamma(d+j+m)}.$ (156) Insert the above result into (154), and evaluate the resulting summation over $i$ by the identity (1), one arrives at $\displaystyle F_{1}=~{}\\!\\!$ $\displaystyle\frac{1}{\Gamma(b+m)\Gamma(a+b+m)}\sum_{j=1}^{c}\frac{(-1)^{j+1}}{\Gamma(j)\Gamma(c-j+1)\Gamma(d+j+m)}$ $\displaystyle\times\sum_{i=1}^{m}\frac{\Gamma(m+j-i)\Gamma(a+b-i+2m)}{\Gamma(m-i+1)\Gamma(a-i+m+1)},$ (157) where the summation over $j$ is simplified, by using the identity (113), to $\displaystyle\sum_{j=1}^{c}\frac{(-1)^{j+1}\Gamma(m-i+j)}{\Gamma(j)\Gamma(c-j+1)\Gamma(d+j+m)}$ $\displaystyle=\frac{\Gamma(m-i+1)}{\Gamma(c)\Gamma(d+m+1)}\,_{2}F_{1}(1-c,m-i+1\mathchar 24635\relax\;d+m+1\mathchar 24635\relax\;1)$ (158) $\displaystyle=\frac{\Gamma(m-i+1)\Gamma(c+d+i-1)}{\Gamma(c)\Gamma(d+i)\Gamma(c+d+m)}.$ (159) Consequently, we have $\displaystyle F_{1}=\\!\\!~{}$ $\displaystyle\frac{1}{\Gamma(c)\Gamma(b+m)\Gamma(a+b+m)\Gamma(c+d+m)}$ $\displaystyle\times\sum_{i=1}^{m}\frac{\Gamma(c+d+i-1)\Gamma(a+b-i+2m)}{\Gamma(d+i)\Gamma(a-i+m+1)}.$ (160) In the same manner, we obtain $\displaystyle F_{2}=\\!\\!~{}$ $\displaystyle\frac{1}{\Gamma(d)\Gamma(a+m)\Gamma(a+b+m)\Gamma(c+d+m)}$ $\displaystyle\times\sum_{i=1}^{m}\frac{\Gamma(c+d+i-1)\Gamma(a+b-i+2m)}{\Gamma(c+i)\Gamma(b-i+m+1)}.$ (161) Insert the results (160)–(161) into (153), we complete the proof of Lemma 4. ###### Lemma 5. Denote $\Phi_{a,b,c,d}^{(x)}=\frac{\Gamma(x+c+1)\Gamma(x+d+1)}{\Gamma(x+a+1)\Gamma(x+b+1)},$ (162) we have $\displaystyle\sum_{i=1}^{m}\frac{(-1)^{i}\Gamma(a-i+m+1)}{\Gamma(m-i+1)}\left(\frac{1}{a+c-i+2m+1}-\frac{1}{c+i}\right)$ $\displaystyle=\frac{\Gamma(a+c+m+1)}{\Gamma(c+m+1)}\sum_{i=1}^{m}\left(\frac{\Phi_{0,a+c,a,c}^{(m-i)}}{a+c-2i+2m+1}+a\Phi_{1,a+c+1,a,c}^{(m-i)}\right.$ $\displaystyle~{}~{}~{}~{}\left.\\!-\frac{\Phi_{1,a+c,a+1,c}^{(m-i)}}{a+c-2i+2m+2}\right),$ (163) where it is sufficient to consider $\real(a,c)\geq 0$. Proof: We prove the Lemma 5 by using the re-summation technique. We denote the left side of the identity (163) as $\mathcal{G}=G_{m}-G^{\prime}_{m},$ (164) where $\displaystyle G_{m}=$ $\displaystyle\sum_{i=1}^{m}\frac{(-1)^{i}\Gamma(a-i+m+1)}{\Gamma(m-i+1)(a+c-i+2m+1)}$ (165) $\displaystyle G^{\prime}_{m}=$ $\displaystyle\sum_{i=1}^{m}\frac{(-1)^{i}\Gamma(a-i+m+1)}{\Gamma(m-i+1)(c+i)}.$ (166) Here, we choose the recurrence relations $\displaystyle G_{m}=$ $\displaystyle c_{m-1}G_{m-1}+r_{m-1}$ (167) $\displaystyle G^{\prime}_{m}=$ $\displaystyle c_{m-1}G^{\prime}_{m-1}+r^{\prime}_{m-1},$ (168) where $\displaystyle c_{m-1}=~{}\\!\\!$ $\displaystyle\frac{a+c+m}{c+m}$ (169) $\displaystyle r_{m-1}=~{}\\!\\!$ $\displaystyle\frac{a}{c+m}\sum_{i=0}^{m}\frac{(-1)^{i}\Gamma(a-i+m)}{\Gamma(m-i+1)}+\frac{(a+c+m)\Gamma(a+m)}{(c+m)\Gamma(m+1)}$ $\displaystyle\times\left(\frac{m}{a+c+2m-1}-\frac{a+m}{a+c+2m}\right)$ (170) $\displaystyle r^{\prime}_{m-1}=~{}\\!\\!$ $\displaystyle\frac{a}{c+m}\sum_{i=1}^{m}\frac{(-1)^{i}\Gamma(a-i+m)}{\Gamma(m-i+1)}.$ (171) Iterating $m$ times the recurrence relations (167) and (168), we obtain $\displaystyle G_{m}=~{}\\!\\!$ $\displaystyle\frac{\Gamma(a+c+m+1)}{\Gamma(c+m+1)}\left(a\sum_{i=1}^{m}\\!\right.\sum_{j=0}^{m-i+1}(-1)^{j}\Phi_{a+c+1,1-j,c,a-j}^{(m-i)}$ $\displaystyle+\\!\left.\sum_{i=1}^{m}\left(\frac{\Phi_{0,a+c,c,a}^{(m-i)}}{a+c-2i+2m+1}-\frac{\Phi_{1,a+c,c,a+1}^{(m-i)}}{a+c-2i+2m+2}\right)\right)$ (172) $\displaystyle G^{\prime}_{m}=~{}\\!\\!$ $\displaystyle\frac{a\Gamma(a+c+m+1)}{\Gamma(c+m+1)}\sum_{i=1}^{m}\sum_{j=1}^{m-i+1}(-1)^{j}\Phi_{a+c+1,1-j,c,a-j}^{(m-i)}.$ (173) It is noticed that the double summation in (172) is the same form as the one in (173). By inserting the re-summations (172) and (173) into (164), we obtain the desired identity (163). This proves Lemma 5. ###### Lemma 6. Using the notation (162), we have $\displaystyle\sum_{i=1}^{m}\Phi_{0,a,b,a+b}^{(m-i)}\left(\frac{1}{a+b+c-i+2m+1}-\frac{1}{c+i}\right)$ $\displaystyle=\Phi_{0,a,b,a+b}^{(m+c)}\sum_{i=1}^{m}\Phi_{0,a,b,a+b}^{(i-1)}\Phi_{b,a+b,0,a}^{(c+i-1)}\left(\frac{1}{a+b+c+2i-1}-\frac{b(a+b)}{ai(a+b+c+i)}\right.$ $\displaystyle~{}~{}~{}~{}\\!\left.-\frac{b(a-b)}{a(a+i)(b+c+i)}+\frac{(b+i)(a+b+i)}{i(a+i)(a+b+c+2i)}-\frac{a+b+2i-2}{(b+i-1)(a+b+i-1)}\right)$ $\displaystyle~{}~{}~{}~{}+\frac{1}{b}\Phi_{0,a,b,a+b}^{(m-1)},$ (174) where it is sufficient to consider $a,b,c,a+b,a+b+c\notin\mathbb{Z}^{-}$. Proof: The proof of Lemma 6 uses the re-summation technique. The left side of the identity (174) can be written as $\mathcal{G}=G_{m}-G^{\prime}_{m},$ (175) where $\displaystyle G_{m}$ $\displaystyle=\sum_{i=1}^{m}\Phi_{0,a,b,a+b}^{(m-i)}\frac{1}{a+b+c-i+2m+1}$ (176) $\displaystyle G^{\prime}_{m}$ $\displaystyle=\sum_{i=1}^{m}\Phi_{0,a,b,a+b}^{(m-i)}\frac{1}{c+i}.$ (177) In relating the two summations above, we choose the following recurrence relations $\displaystyle G_{m}$ $\displaystyle=c_{m-1}G_{m-1}+r_{m-1}$ (178) $\displaystyle G^{\prime}_{m}$ $\displaystyle=c_{m-1}G^{\prime}_{m-1}+r^{\prime}_{m-1},$ (179) where $\displaystyle c_{m-1}=~{}\\!\\!$ $\displaystyle\frac{(b+c+m)(a+b+c+m)}{(c+m)(a+c+m)}$ (180) $\displaystyle r_{m-1}=~{}\\!\\!$ $\displaystyle c_{m-1}\left(\frac{\Phi_{0,a,b,a+b}^{(m-1)}}{a+b+c+2m-1}+\frac{\Phi_{0,a,b,a+b}^{(m)}}{a+b+c+2m}\right)-\frac{b}{a}c_{m-1}$ $\displaystyle\times\left(\frac{(a-b)}{b+c+m}\sum_{i=1}^{m}\Phi_{0,a+1,b,a+b}^{(m-i)}+\frac{(a+b)}{a+b+c+m}\sum_{i=1}^{m+1}\Phi_{1,a,b,a+b}^{(m-i)}\right)$ (181) $\displaystyle r^{\prime}_{m-1}=~{}\\!\\!$ $\displaystyle\frac{b(a-b)}{a(a+c+m)}\sum_{i=1}^{m-1}\Phi_{0,a+1,b,a+b}^{(m-i-1)}+\frac{b(a+b)}{a(c+m)}\sum_{i=1}^{m}\Phi_{1,a,b,a+b}^{(m-i-1)}.$ (182) Iterating $m$ times the above relations (178) and (179), one obtains $\displaystyle G_{m}=~{}\\!\\!$ $\displaystyle\Phi_{0,a,b,a+b}^{(m+c)}\left(\rule{0.0pt}{19.0633pt}\\!-\frac{b(a+b)}{a}\sum_{i=1}^{m}\Phi_{1,a,b,a+b}^{(i-2)}\sum_{j=1}^{m-i+1}\Phi_{a+b+1,b,0,a}^{(m-j+c)}-\frac{b(a-b)}{a}\right.$ $\displaystyle\times\sum_{i=1}^{m}\Phi_{0,a+1,b,a+b}^{(i-1)}\sum_{j=1}^{m-i+1}\Phi_{b+1,a+b,a,0}^{(m-j+c)}+\sum_{i=1}^{m}\Phi_{b,a+b,0,a}^{(m-i+c)}$ $\displaystyle\times\left(\frac{\Phi_{0,a,b,a+b}^{(m-i+1)}}{a+b+c-2i+2m+2}+\frac{\Phi_{0,a,b,a+b}^{(m-i)}}{a+b+c-2i+2m+1}\right.$ $\displaystyle\left.\\!-\frac{b(a+b)\Phi_{1,a,b,a+b}^{(m-i)}}{a(a+b+c-i+m+1)}\right)\\!\left.\rule{0.0pt}{19.0633pt}\right)$ (183) $\displaystyle G^{\prime}_{m}=~{}\\!\\!$ $\displaystyle\Phi_{0,a,b,a+b}^{(m+c)}\left(\frac{b(a+b)}{a}\sum_{i=1}^{m}\Phi_{1,a,b,a+b}^{(i-2)}\sum_{j=1}^{m-i+1}\Phi_{a+b+1,b+1,0,a+1}^{(m-j+c)}\right.$ $\displaystyle+\\!\left.\frac{b(a-b)}{a}\sum_{i=1}^{m}\Phi_{0,a+1,b,a+b}^{(i-2)}\sum_{j=1}^{m-i+1}\Phi_{b+1,a+b+1,a,1}^{(m-j+c)}\right).$ (184) Note that in order to reveal the potential cancellations between the two re- summations above, one will also need the identity $\displaystyle(c_{1}-b_{1})\sum_{j=1}^{m}\Phi_{a_{1},b_{1}+1,c_{1},d_{1}+1}^{(m-j)}+(d_{1}-a_{1}+1)\sum_{j=1}^{m}\Phi_{a_{1},b_{1},c_{1},d_{1}}^{(m-j)}$ $\displaystyle=\Phi_{a_{1}-1,b_{1},c_{1},d_{1}}^{(m)}-\Phi_{a_{1}-1,b_{1},c_{1},d_{1}}^{(0)},\qquad a_{1},b_{1},c_{1},d_{1}\notin\mathbb{Z}^{-}.$ (185) This identity is obtained by iterating $m$ times the following recurrence relation $H_{m,a_{1},d_{1}}=H_{m-1,a_{1}+1,d_{1}+1}+s_{m-1},$ (186) where $H_{m,a_{1},d_{1}}=\sum_{j=1}^{m}\Phi_{a_{1},b_{1},c_{1},d_{1}}^{(m-j)}$ (187) $s_{m-1}=H_{m,a_{1},d_{1}}-H_{m-1,a_{1}+1,d_{1}+1}=(c_{1}-b_{1})\sum_{j=1}^{m-1}\Phi_{a_{1},b_{1},c_{1}-1,d_{1}}^{(m-j)}+\Phi_{a_{1},b_{1},c_{1},d_{1}}^{(0)}.$ (188) We now use the identity (185) with the specializations $\displaystyle a_{1}=a+b+c+i,~{}~{}~{}b_{1}=b+c+i-1,~{}~{}~{}c_{1}=c+i-1,$ $\displaystyle d_{1}=a+c+i-1,~{}~{}~{}m\to m-i+1$ (189) and $\displaystyle a_{1}=b+c+i,~{}~{}~{}b_{1}=a+b+c+i-1,~{}~{}~{}c_{1}=a+c+i-1,$ $\displaystyle d_{1}=c+i-1,~{}~{}~{}m\to m-i+1,$ (190) the result (184) becomes $\displaystyle G^{\prime}_{m}=$ $\displaystyle~{}\Phi_{0,a,b,a+b}^{(c+m)}\Bigg{(}\\!-\frac{b(a+b)}{a}\sum_{i=1}^{m}\Phi_{1,a,b,a+b}^{(i-2)}\sum_{j=1}^{m-i+1}\Phi_{a+b+1,b,0,a}^{(m-j+c)}-\frac{b(a-b)}{a}$ $\displaystyle\\!~{}\times\sum_{i=1}^{m}\Phi_{0,a+1,b,a+b}^{(i-2)}\sum_{j=1}^{m-i+1}\Phi_{b+1,a+b,a,0}^{(m-j+c)}+\sum_{i=1}^{m}\left(\Phi_{a+b,b,0,a}^{(c+i-1)}-\Phi_{a+b,b,0,a}^{(c+m)}\right)$ $\displaystyle\\!~{}\times\left(\frac{a-b}{a}\Phi_{0,a+1,b,a+b}^{(i-2)}+\frac{a+b}{a}\Phi_{1,a,b,a+b}^{(i-2)}\right)\\!\Bigg{)},$ (191) where the double summations are of the same form as the ones in (183). Therefore, all the double summations are cancelled completely in (175). The remaining terms are the desired result (174). This completes the proof of Lemma 6. Note that the results in lemmas 2–4 are analytically continued to any complex number $c$ or $d$ except for negative integers. This fact allows us to take derivatives of the formulas (134), (142), and (LABEL:eq:lemma4) in obtaining the identities (B.2)–(B.2) listed in Appendix B.2. For example, the identity (B.2) is established by taking derivatives of $c$ and $d$ of the formula (LABEL:eq:lemma4) before setting $c=d=0$. The lemmas 1–4 along with the identities (B.2)–(B.2) are useful in simplifying the summations in (A.1) and (A.2). It is also worth mentioning that the identities of Lemma 5 and Lemma 6 admit closed-form representations for some special cases, e.g., when the parameter $c$ is fixed to non-negative integers, since all the gamma functions involved are reduced to rational functions. These identities as well as their derivatives of $c$ are the key tools in simplifying the summations in (A.1) and (A.2). Note also that in lemmas 1–4, by rewriting the summations involved as hypergeometric functions, the corresponding identities induce the following new transformation formulas of unit arguments (192)–(195), which may be of independent interest. $\,{}_{3}F_{2}(c+1,1-m,1-b-m\mathchar 24635\relax\;a+1,c+2\mathchar 24635\relax\;1)$ $\displaystyle=\frac{(c+1)\Gamma(a+1)\Gamma(a+b+2m-1)}{(c+m)\Gamma(a+m)\Gamma(a+b+m)}$ $\displaystyle~{}~{}~{}\times\,_{3}F_{2}(1,1-m,1-a-m\mathchar 24635\relax\;2-a-b-2m,1-c-m\mathchar 24635\relax\;1)$ (192) $\,{}_{3}F_{2}(1,1-m,1-b-m\mathchar 24635\relax\;a+1,c+1\mathchar 24635\relax\;1)$ $\displaystyle=\frac{c}{c+m-1}\,_{3}F_{2}(1,1-m,a+b+m\mathchar 24635\relax\;a+1,2-c-m\mathchar 24635\relax\;1)$ (193) $\,{}_{4}F_{3}(1,1,1-m,1-b-m\mathchar 24635\relax\;2,a+1,c+1\mathchar 24635\relax\;1)$ $\displaystyle=\frac{a(b+c+m)}{(a+m-1)(b+m)}\,_{4}F_{3}(1,1,1-m,b+c+m+1\mathchar 24635\relax\;2,c+1,2-a-m\mathchar 24635\relax\;1)$ $\displaystyle~{}~{}~{}+\frac{ac(\psi_{0}(a)-\psi_{0}(a+m))}{m(b+m)}$ (194) $\displaystyle\frac{\,{}_{3}F_{2}(1,1-b-m,1-d-m\mathchar 24635\relax\;a+1,c+1\mathchar 24635\relax\;1)}{\Gamma(a+1)\Gamma(c+1)\Gamma(b+m)\Gamma(d+m)}$ $\displaystyle=\frac{\,{}_{3}F_{2}(1,1-b,1-d\mathchar 24635\relax\;a+m+1,c+m+1\mathchar 24635\relax\;1)}{\Gamma(b)\Gamma(d)\Gamma(a+m+1)\Gamma(c+m+1)}-\frac{1}{a+b+m-1}$ $\displaystyle~{}~{}~{}\times\left(\frac{\,{}_{3}F_{2}(1,1-a,c+d+m\mathchar 24635\relax\;2-a-b-m,d+m+1\mathchar 24635\relax\;1)}{\Gamma(a)\Gamma(c)\Gamma(b+m)\Gamma(d+m+1)}\right.$ $\displaystyle~{}~{}~{}+\left.\\!\frac{\,{}_{3}F_{2}(1,1-b,c+d+m\mathchar 24635\relax\;2-a-b-m,c+m+1\mathchar 24635\relax\;1)}{\Gamma(b)\Gamma(d)\Gamma(a+m)\Gamma(c+m+1)}\right)$ $\displaystyle~{}~{}~{}+\frac{\Gamma(c+d)\Gamma(a+b+2m-1)}{\Gamma(c)\Gamma(d)\Gamma(a+m)\Gamma(b+m)\Gamma(a+b+m)\Gamma(c+d+m)}$ $\displaystyle~{}~{}~{}\times\left(\frac{1}{d}\,_{3}F_{2}(1,c+d,1-a-m\mathchar 24635\relax\;d+1,2-a-b-2m\mathchar 24635\relax\;1)\right.$ $\displaystyle~{}~{}~{}+\\!\left.\frac{1}{c}\,_{3}F_{2}(1,c+d,1-b-m\mathchar 24635\relax\;c+1,2-a-b-2m\mathchar 24635\relax\;1)\right).$ (195) ### 2.3 Simplification of summations In this section, we compute the integrals $\mathrm{I_{A}}$ and $\mathrm{I_{B}}$ in obtaining the variance formulas in Proposition 3 and Proposition 4. For convenience, we summarize in Table 3 below the corresponding integrals of the variance (47) for case A and case B. Table 3: Integrals of the variance (47) for case A and case B. | $\mathrm{I_{A}}$ | $~{}~{}~{}~{}\mathrm{I_{B}}$ ---|---|--- Case A: Arbitrary number of particle (17) | (51) | (55) Case B: Fixed number of particle (18) | (69) | (79) | | #### Case A: Arbitrary number of particles In case A, the variance calculation boils down to simplifying the summation representations of $\mathrm{A_{1}}$, $\mathrm{A_{2}}$, $\mathrm{B_{1}}$, and $\mathrm{B_{2}}$ in (A.1)–(A.1) as listed in Appendix A.1. The corresponding simplification procedures are shown below. We first discuss the simplifications of summation representations (A.1) and (A.1) in computing the integral $\mathrm{I_{A}}$ in (51). The summations in (A.1) can be simplified by appropriately changing the summation orders and using the first type identities (287)–(B.1) of the existing framework, where the closed-form identities are (287)–(294) and the semi closed-form ones are (295)–(B.1). Semi closed-form identities that we referred to represent the relation between two single summations. These sums are what we refer to as unsimplifiable basis, which will cancel completely in obtaining the closed- form results. On the other hand, whether the considered unsimplifiable basis can be computed into closed-form formulas is not of primary importance. The simplification of the summations in (A.1) is as follows. In (A.1), the first double summation is $\displaystyle\sum_{k=0}^{m-1}\sum_{j=2k-2}^{2k}\frac{2(-1)^{j}(2a+4k+1)(j+1)_{2}(a+j+1)_{2}}{\Gamma(2k-j+1)\Gamma(j-2k+3)(2a+j+2k+1)_{3}}\bigg{(}\\!(\psi_{0}(j+3)$ $\displaystyle-\psi_{0}(2a+j+2k+4)-\psi_{0}(j-2k+3)+\psi_{0}(a+j+3))^{2}$ $\displaystyle-\psi_{1}(2a+j+2k+4)+\psi_{1}(a+j+3)-\psi_{1}(j-2k+3)+\psi_{1}(j+3)\\!\bigg{)},$ (196) which is directly reduced to a single sum after evaluating the sum of $j$. By using the identities (287)–(B.1) along with the results AS72 $\displaystyle\psi_{0}(mk)$ $\displaystyle=\ln m+\frac{1}{m}\sum_{i=0}^{m-1}\psi_{0}\\!\left(k+\frac{i}{m}\right)$ (197) $\displaystyle\psi_{1}(mk)$ $\displaystyle=\frac{1}{m^{2}}\sum_{i=0}^{m-1}\psi_{1}\\!\left(k+\frac{i}{m}\right),\qquad m\in\mathbb{Z^{+}},$ (198) the remaining single sum is computed to the unsimplifiable basis of the form $\sum_{k=1}^{m}\frac{\psi_{0}(k+c)}{k+d},\qquad c\neq d.$ (199) The other double summation in (A.1) is $\displaystyle\sum_{k=0}^{m-1}\sum_{j=0}^{2k-3}\frac{4(2a+4k+1)(j+1)_{2}(a+j+1)_{2}}{(2k-j-2)_{3}(2a+j+2k+1)_{3}}\Psi(j),$ (200) where $\Psi(j)=\psi_{0}(2a+j+2k+4)-\psi_{0}(a+j+3)+\psi_{0}(2k-j-2)-\psi_{0}(j+3).$ (201) To process the summation (200), we first perform the partial fraction decomposition of the term $\displaystyle\frac{4(2a+4k+1)(j+1)_{2}(a+j+1)_{2}}{(2k-j-2)_{3}(2a+j+2k+1)_{3}}$ $\displaystyle=-\frac{2(a+k)(2a+2k+1)}{2a+j+2k+2}+\frac{2(a+k)(a+2k-1)(2a+2k-1)}{(2a+4k-1)(2a+j+2k+1)}$ $\displaystyle~{}~{}~{}\\!~{}+\frac{2(a+k+1)(a+2k+2)(2a+2k+1)}{(2a+4k+3)(2a+j+2k+3)}-\frac{2(k+1)(2k+1)(a+2k+2)}{(2a+4k+3)(j-2k)}$ $\displaystyle~{}~{}~{}\\!~{}-\frac{2k(2k-1)(a+2k-1)}{(2a+4k-1)(j-2k+2)}+\frac{2k(2k+1)}{j-2k+1}.$ (202) The sum (200) now boils down to $\displaystyle\sum_{k=0}^{m-1}\mathrm{p}_{c}(k)\sum_{j=0}^{2k-3}\frac{1}{j+2k+2a+c}\Psi(j)$ (203) $\displaystyle\sum_{k=0}^{m-1}\mathrm{p}^{\prime}_{c}(k)\sum_{j=0}^{2k-3}\frac{1}{j-2k-1+c}\Psi(j),$ (204) where the parameter $c$ takes the values $c=1,2,3$, $\mathrm{p}_{c}(k)$ and $\mathrm{p}^{\prime}_{c}(k)$ denote the polynomials in $k$ from the partial fraction decomposition (2.3). When considering $c=2$ and the term $\psi_{0}(j+3)$ in (201), the double summation $-\sum_{k=0}^{m-1}2(a+k)(2a+2k+1)\sum_{j=0}^{2k-3}\frac{\psi_{0}(j+3)}{j+2k+2a+2}$ (205) is simplified to a single sum as $\displaystyle\sum_{j=0}^{m-3}\Bigg{(}\psi_{0}(2j+3)((j+1)(2j+1)(\psi_{0}(1)-\psi_{0}(a+j+m+1))+(j-m+2)$ $\displaystyle\times(2a-j+m)+(j+1)(2j+1)(\psi_{0}(a+2j+3)-\psi_{0}(1)))+\psi_{0}(2j+4)$ $\displaystyle\times\left((j+1)(2j+3)\left(\psi_{0}(1)-\psi_{0}\\!\left(a+j+m+\frac{3}{2}\right)\right)+(j+1)(2j+3)\right.$ $\displaystyle\times\\!\left.\left(\psi_{0}\\!\left(a+2j+\frac{7}{2}\right)-\psi_{0}(1)\right)+(2a-j+m-1)(j-m+2)\right)\\!\Bigg{)}.$ (206) The above result is obtained as follows. In (205), it is noticed that directly evaluating the inner summation does not permit further simplifications. Instead, we first have to separate the inner summation that depends on an even index $2k$ into two as $-2\sum_{k=0}^{m-1}(a+k)(2a+2k+1)\left(\sum_{j=0}^{k-2}\frac{\psi_{0}(2j+3)}{2a+2k+2j+2}+\sum_{j=0}^{k-2}\frac{\psi_{0}(2j+4)}{2a+2k+2j+3}\right),$ (207) which allows us to change the summation order as $\displaystyle-2\sum_{j=0}^{m-3}\psi_{0}(2j+3)\sum_{k=j+2}^{m-1}\frac{(a+k)(2a+2k+1)}{2a+2k+2j+2}$ $\displaystyle-2\sum_{j=0}^{m-3}\psi_{0}(2j+4)\sum_{k=j+2}^{m-1}\frac{(a+k)(2a+2k+1)}{2a+2k+2j+3}.$ (208) Evaluating the inner summations over $k$ in (2.3) directly gives the result (2.3). Other double summations in (203) and (204) are similarly simplified into single sums. The resulting single sums, cf. (2.3), are further manipulated into the unsimplifiable basis of the form (199) by using the first type identities (287)–(B.1) along with the result (197). Different from the summations in (A.1), the simplification of the summations in (A.1) will need new simplification techniques before the existing framework is applicable. The summations in (A.1) include three double sums and one triple sum. We first simplify the inner summations over $j$ of the three double sums by using the identities (B.2) and (B.2) of the new simplification framework. Specifically, the inner summation $\sum_{j=0}^{2k}\frac{2(j+1)(2k-j+1)\psi_{0}(j+2)\psi_{0}(2k-j+2)}{\Gamma(j+1)\Gamma(a+j+1)\Gamma(2k-j+1)\Gamma(a-j+2k+1)}$ (209) of the first double sum in (A.1) is simplified to a semi-closed form representation that can be compactly written as $\displaystyle\frac{1}{\Gamma(2k)\Gamma(a+2k+1)\Gamma(2a+2k+1)}\Bigg{(}\\!\left(\frac{1-4a^{2}}{4a+8k-2}+a-8k^{2}-14k-\frac{15}{2}\right)$ $\displaystyle\times\sum_{j=1}^{2k}\frac{\Gamma(2a-j+4k)}{\Gamma(a-j+2k)j^{2}}+\left(\left(\frac{1-4a^{2}}{4a+8k-2}+a-8k^{2}-14k-\frac{15}{2}\right)\psi_{0}(2k)\right.$ $\displaystyle+\\!\left.\frac{1-4a^{2}}{2(2a+4k-1)^{2}}+\frac{3-4a^{2}}{4a+8k-2}+\frac{4(a-1)}{a+2k}+a-14k-\frac{4}{k}-19\right)$ $\displaystyle\times\sum_{j=1}^{2k}\frac{\Gamma(2a-j+4k)}{\Gamma(a-j+2k)j}\\!\Bigg{)}+\mathrm{CF},$ (210) where the shorthand notation $\mathrm{CF}$ denotes the closed-form terms omitted. The result (2.3) is obtained by using the identities (132), (B.2), and (B.2) after one rewrites the summation (209) as $\displaystyle 8\sum_{j=2}^{2k+1}\frac{1}{\Gamma(j)\Gamma(a+j)\Gamma(2k-j+2)\Gamma(a-j+2k+2)}$ $\displaystyle+8\sum_{j=1}^{2k+1}\frac{\psi_{0}(j)}{\Gamma(j)\Gamma(a+j)\Gamma(2k-j+2)\Gamma(a-j+2k+2)}$ $\displaystyle+8\sum_{j=1}^{2k}\frac{\psi_{0}(j)}{\Gamma(j)\Gamma(a+j+1)\Gamma(2k-j+1)\Gamma(a-j+2k+1)}$ $\displaystyle+2\sum_{j=1}^{2k+1}\frac{\psi_{0}(j)\psi_{0}(2k-j+2)}{\Gamma(j)\Gamma(a+j)\Gamma(2k-j+2)\Gamma(a-j+2k+2)}$ $\displaystyle+4\sum_{j=1}^{2k}\frac{\psi_{0}(j)\psi_{0}(2k-j+1)}{\Gamma(j)\Gamma(a+j)\Gamma(2k-j+1)\Gamma(a-j+2k+2)}$ $\displaystyle+2\sum_{j=1}^{2k-1}\frac{\psi_{0}(j)\psi_{0}(2k-j)}{\Gamma(j)\Gamma(a+j+1)\Gamma(2k-j)\Gamma(a-j+2k+1)}$ $\displaystyle-\frac{4}{\Gamma(a+1)\Gamma(a+2k+1)}\left(\frac{\psi_{0}(2k+1)}{\Gamma(2k+1)}+\frac{\psi_{0}(2k)}{\Gamma(2k)}\right).$ (211) Using the same approach, the other inner summations over $j$ of the first three double summations in (A.1) are simplified to a semi closed-form representation, cf. (2.3), where the resulting unsimplifiable basis are $\displaystyle\sum_{j=1}^{2k}\frac{\Gamma(2a-j+4k)}{\Gamma(a-j+2k)j}$ (212) $\displaystyle\sum_{j=1}^{2k}\frac{\Gamma(2a-j+4k)}{\Gamma(a-j+2k)j^{2}}.$ (213) Now we simplify the inner summations over the indexes $i$ and $j$ of the triple sum in (A.1). After the partial fraction decomposition $\frac{1}{(j)_{3}}=-\frac{1}{j+1}+\frac{1}{2(j+2)}+\frac{1}{2j},$ (214) the corresponding inner summation is written as $\displaystyle\sum_{j=1}^{2k-1}\sum_{i=1}^{2k-j}\frac{4i(2k-i+2)\Gamma(a+2k+j+3)\Gamma(a+2k-j+1)}{\Gamma(j+i+1)\Gamma(2k-j-i+1)\Gamma(a+i)\Gamma(a+2k-i+2)}$ $\displaystyle\times(\psi_{0}(a+2k+j+3)-\psi_{0}(2a+4k+4)+\psi_{0}(2k-i+3)-\psi_{0}(j+3))$ $\displaystyle\times\left(-\frac{1}{j+1}+\frac{1}{2(j+2)}+\frac{1}{2j}\right).$ (215) We first consider the simplification of the summation $\displaystyle\sum_{j=1}^{2k-1}\sum_{i=1}^{2k-j}\frac{4i(2k-i+2)\Gamma(a+2k+j+3)\Gamma(a+2k-j+1)}{(j+1)\Gamma(j+i+1)\Gamma(2k-j-i+1)\Gamma(a+i)\Gamma(a+2k-i+2)}$ $\displaystyle\times(\psi_{0}(a+2k+j+3)-\psi_{0}(2a+4k+4)+\psi_{0}(2k-i+3)-\psi_{0}(j+3)),$ (216) which involves the term $1/(j+1)$ from the decomposition (214), and the remaining parts in (2.3) can be simplified in the same manner. For convenience, we divide (2.3) into the three sums $\displaystyle\sum_{j=1}^{2k-1}\sum_{i=1}^{2k-j}\frac{4i(2k-i+2)\Gamma(a+2k+j+3)\Gamma(a+2k-j+1)}{(j+1)\Gamma(j+i+1)\Gamma(2k-j-i+1)\Gamma(a+i)\Gamma(a+2k-i+2)}$ $\displaystyle\times(-\psi_{0}(2a+4k+4)-\psi_{0}(j+3))$ (217) $\displaystyle\sum_{j=1}^{2k-1}\sum_{i=1}^{2k-j}\frac{4i(2k-i+2)\Gamma(a+2k+j+3)\Gamma(a+2k-j+1)}{(j+1)\Gamma(j+i+1)\Gamma(2k-j-i+1)\Gamma(a+i)\Gamma(a+2k-i+2)}$ $\displaystyle\times\psi_{0}(2k-i+3)$ (218) $\displaystyle\sum_{j=1}^{2k-1}\sum_{i=1}^{2k-j}\frac{4i(2k-i+2)\Gamma(a+2k+j+3)\Gamma(a+2k-j+1)}{(j+1)\Gamma(j+i+1)\Gamma(2k-j-i+1)\Gamma(a+i)\Gamma(a+2k-i+2)}$ $\displaystyle\times\psi_{0}(a+2k+j+3).$ (219) The digamma functions in the summation (217) is independent to the index $i$. This fact allows us to evaluate the summation over $i$ by using the identity (134) of Lemma 2. Specifically, we first rewrite (217) as $\displaystyle\sum_{j=1}^{2k-1}\frac{\Gamma(a-j+2k+1)\Gamma(a+j+2k+3)(-\psi_{0}(2a+4k+4)-\psi_{0}(j+3))}{j+1}$ $\displaystyle\times\sum_{i=1}^{2k-j}\frac{1}{\Gamma(a+i)\Gamma(i+j)\Gamma(a-i+2k+2)\Gamma(2k-i-j)}\frac{4i(2k-i+2)}{(i+j)(2k-j-i)}.$ (220) We now partial fraction decompose the rational function with respect to $i$ in (220). The summation (217) becomes $\displaystyle\sum_{j=1}^{2k-1}\frac{\Gamma(a-j+2k+1)\Gamma(a+j+2k+3)(-\psi_{0}(2a+4k+4)-\psi_{0}(j+3))}{j+1}$ $\displaystyle\times\left(4\sum_{i=1}^{2k-j-1}\frac{1}{\Gamma(a+i)\Gamma(i+j)\Gamma(a-i+2k+2)\Gamma(2k-j-i)}-\frac{j+2k+2}{k}\right.$ $\displaystyle\times 2j\sum_{i=1}^{2k-j-1}\frac{1}{\Gamma(a+i)\Gamma(i+j+1)\Gamma(a-i+2k+2)\Gamma(2k-j-i)}+\frac{2k-j}{k}$ $\displaystyle\times\\!\left.2(j+2)\sum_{i=1}^{2k-j}\frac{1}{\Gamma(a+i)\Gamma(i+j)\Gamma(a-i+2k+2)\Gamma(2k-j+1-i)}\right),$ (221) where the inner summations over $i$ can be evaluated by the identity (134) with the specializations $\displaystyle a=j,\qquad~{}~{}~{}~{}~{}~{}\\!b=a+j+2,\qquad c=a,\qquad m=2k-j-1$ (222) $\displaystyle a=j+1,\qquad b=a+j+2,\qquad c=a,\qquad m=2k-j-1$ (223) $\displaystyle a=j,\qquad~{}~{}~{}~{}~{}~{}\\!b=a+j+1,\qquad c=a,\qquad m=2k-j.$ (224) The summation (221) becomes $\displaystyle\sum_{j=1}^{2k-1}\left(j+1-\frac{(a+2k+1)^{2}}{j+1}\right)\frac{\psi_{0}(2a+4k+4)+\psi_{0}(j+3)}{\Gamma(a)\Gamma(a+2k+1)}$ $\displaystyle\times\left(4(a-j+2k-1)(a+j+2k+1)\sum_{i=1}^{2k-j-1}\frac{\Gamma(a+i-1)\Gamma(a-i+4k)}{\Gamma(i)\Gamma(2k-i)}\right.$ $\displaystyle-\frac{2j(j+2k+2)(a-j+2k-1)}{k}\sum_{i=1}^{2k-j-1}\frac{\Gamma(a+i-1)\Gamma(a-i+4k+1)}{\Gamma(i)\Gamma(2k-i+1)}$ $\displaystyle+\\!\left.\frac{2(j+2)(2k-j)(a+j+2k+1)}{k}\sum_{i=1}^{2k-j}\frac{\Gamma(a+i-1)\Gamma(a-i+4k+1)}{\Gamma(i)\Gamma(2k-i+1)}\right).$ (225) In (225), we further change the orders of summations to simplify the sums over $j$. These summations admit closed-form expressions by using the identities (B.1)–(B.1) and (293). The remaining summations only consist of single sums, which can be further simplified by using the identities (B.2)–(B.2) into the unsimplifiable basis (212) and $\displaystyle\sum_{j=1}^{2k}\frac{\Gamma(2a-j+4k-1)\psi_{0}(j)}{\Gamma(a-j+2k+1)}$ (226) $\displaystyle\sum_{j=1}^{2k}\frac{\Gamma(2a-j+4k-1)\psi_{0}(j)}{\Gamma(a-j+2k)j}$ (227) $\displaystyle\sum_{j=1}^{2k}\frac{\Gamma(2a-j+4k)\psi_{0}(2a-j+4k)}{\Gamma(a-i+2k)j}.$ (228) We now move on to simplifying the summation (218). In (218), we change the summation order to evaluate the sum over $j$ by the identity (142) in Lemma 3. The summation then becomes $\displaystyle\Gamma^{2}(a+2k+2)\sum_{i=1}^{2k-1}\sum_{j=1}^{2k-i}\frac{4(i+j+1)(a+i+j+1)(2k-i-j+1)}{j\Gamma(i+1)\Gamma(a+i+2)\Gamma(2k-i+1)\Gamma(a-i+2k+2)}$ $\displaystyle\times(a-i-j+2k+1)\psi_{0}(i+j+2)-\Gamma^{2}(a+2k+2)\sum_{i=1}^{2k-1}4i(2k-i+2)$ $\displaystyle\times\frac{\psi_{0}(2k-i+3)\left(\psi_{0}(a+i+1)-\psi_{0}(a+2k+2)\right)}{\Gamma(i)\Gamma(a+i)\Gamma(2k-i+2)\Gamma(a-i+2k+2)}-\Gamma(a+2k+1)$ $\displaystyle\times\Gamma(a+2k+3)\sum_{i=1}^{2k-1}\frac{4(2k-i+2)(2k-i+1)\psi_{0}(2k-i+3)}{\Gamma(i)\Gamma(a+i)\Gamma(2k-i+2)\Gamma(a-i+2k+2)}+\frac{1}{\Gamma(a+2)}$ $\displaystyle\times\frac{\Gamma(a+2k+2)}{\Gamma(2k+1)}\sum_{i=1}^{2k-1}\frac{4i(a+i)(2k-i+2)(a-i+2k+2)\psi_{0}(2k-i+3)}{2k-i+1}.$ (229) We now shift the index $i\to 2k-i$ of the three single sums in the above result (229). The first two single sums are simplified into the unsimplifiable basis (212) by using the identities (B.2)–(B.2). The last single sum in (229) is simplified directly into a closed-form representation by using the identities (B.1)–(B.1) and (293). For the double summation in (229), we take the partial fraction decomposition $\displaystyle\frac{4(i+j+1)(a+i+j+1)(2k-i-j+1)(a-i-j+2k+1)}{j}$ $\displaystyle=\frac{4(i+1)(a+i+1)(2k-i+1)(a-i+2k+1)}{j}+4j^{3}+16(i-k)j^{2}$ $\displaystyle~{}~{}~{}\\!~{}-4\left(a^{2}+2a(k+1)-6i^{2}+12ik-4k^{2}+4k+2\right)j$ $\displaystyle~{}~{}~{}\\!~{}+8(i-k)\left(-a^{2}-2a(k+1)+2(i+1)(i-2k-1)\right).$ (230) In (230), the polynomial part can be simplified similarly approach as in (225), while the rational part is simplified as follows. The corresponding summation is $\displaystyle\sum_{i=1}^{2k-1}\frac{4(i+1)(2k-i+1)\Gamma^{2}(a+2k+2)}{\Gamma(i+1)\Gamma(a+i+1)\Gamma(2k-i+1)\Gamma(a-i+2k+1)}\sum_{j=1}^{2k-i}\frac{\psi_{0}(i+j+2)}{j}.$ (231) To evaluate (231), the sum over $j$ is computed by using the identity (B.1) with the specialization $i=j,\qquad a=i+1,\qquad b=1,\qquad m=2k-i,$ (232) then we shift the summation index $i\to 2k-i$ of the outer sum. Consequently, (231) is simplified to $\displaystyle-\sum_{i=1}^{2k-1}\frac{4(i+1)(2k-i+1)\Gamma^{2}(a+2k+2)}{\Gamma(i+1)\Gamma(a+i+1)\Gamma(2k-i+1)\Gamma(a-i+2k+1)}$ $\displaystyle\times\Bigg{(}\sum_{j=1}^{2k-i+1}\frac{\psi_{0}(i+j+1)}{j}-\frac{1}{2}(\left(\psi_{0}(2k-i+2)+\psi_{0}(i+1)-2\psi_{0}(1)\right)$ $\displaystyle\times\left(\psi_{0}(2k-i+2)+\psi_{0}(i+1)\right)\left.-\psi_{1}(2k-i+2)-\psi_{1}(i+1)+2\psi_{1}(1)\right)\\!\Bigg{)}.$ (233) In (233), the double sum is the same form as in (231) but with a negative sign. Therefore, by adding up (231) and (233), and dividing the result by two, we reduce the double summation in (231) to a single sum as $\displaystyle\sum_{i=1}^{2k-1}\frac{4(i+1)(2k-i+1)\Gamma^{2}(a+2k+2)}{\Gamma(i+1)\Gamma(a+i+1)\Gamma(2k-i+1)\Gamma(a-i+2k+1)}\sum_{j=1}^{2k-i}\frac{\psi_{0}(i+j+2)}{j}$ $\displaystyle=\sum_{i=1}^{2k-1}\frac{2\Gamma^{2}(a+2k+2)}{\Gamma(i+1)\Gamma(a+i+1)\Gamma(2k-i+1)\Gamma(a-i+2k+1)}\Big{(}\frac{1}{2}(2k-i+1)$ $\displaystyle~{}~{}~{}\\!~{}\times(i+1)((\psi_{0}(2k-i+2)+\psi_{0}(i+1))(\psi_{0}(2k-i+2)+\psi_{0}(i+1)$ $\displaystyle~{}~{}~{}\\!~{}-2\psi_{0}(1))-\psi_{1}(2k-i+2)-\psi_{1}(i+1)+2\psi_{1}(1))+(2k-i+1)$ $\displaystyle~{}~{}~{}\\!~{}\times\left(\psi_{0}(2k-i+2)+\psi_{0}(i+2)-\psi_{0}(2k+3)-\psi_{0}(1)\right)-(i+1)\psi_{0}(2k+3)\Big{)},$ (234) which is further simplified into a semi closed-form expression involving the unsimplifiable basis (212)–(213) by using the identities (B.2), and (B.2). We now simplify the summation (219) as a last piece in (A.1). We first change the summation order in (219) to evaluate the sum over $j$ by the identity (B.2). As a result, (219) is simplified to $\displaystyle\Gamma^{2}(a+2k+2)\sum_{i=1}^{2k}4i(a+i)(2k-i+2)(a-i+2k+2)$ $\displaystyle\times\sum_{j=1}^{2k-i+1}\frac{\psi_{0}(a-i+2k+3)-\psi_{0}(a-i-j+2k+3)+\psi_{0}(a+2k+2)}{j\Gamma(i+j)\Gamma(a+i+j+1)\Gamma(2k-i-j+2)\Gamma(a-i-j+2k+3)}$ $\displaystyle-\Gamma^{2}(a+2k+2)\sum_{i=1}^{2k-1}\frac{4i(2k-i+2)\psi_{0}(a+2k+2)}{\Gamma(i)\Gamma(a+i)\Gamma(2k-i+2)\Gamma(a-i+2k+2)}$ $\displaystyle\times\left(\psi_{0}(a+i+1)-\psi_{0}(a+2k+2)\right)-\Gamma(a+2k+1)\Gamma(a+2k+3)$ $\displaystyle\times\sum_{i=1}^{2k-1}\frac{4(-i+2k+2)\psi_{0}(a+2k+3)}{\Gamma(i)\Gamma(a+i)\Gamma(-i+2k+1)\Gamma(a-i+2k+2)}-16(a+2)k(a+2k)$ $\displaystyle\times\frac{\Gamma(a+2k+2)}{\Gamma(a+2)\Gamma(2k+1)}\left(\psi_{0}(a+2k+2)-\psi_{0}(a+2)+\psi_{0}(a+3)\right),$ (235) where the single summations are simplified similarly to the ones in (229). To simplify the double summation in (235), we shift the summation index $i\to 2k-i-j+2$ as $\displaystyle\sum_{i=1}^{2k}\frac{\Gamma^{2}(a+2k+2)}{\Gamma(i)\Gamma(a+i+1)\Gamma(2k-i+2)\Gamma(a-i+2k+3)}$ $\displaystyle\times\sum_{j=1}^{2k-i+1}\frac{4(i+j)(a+i+j)(2k-i-j+2)(a-i-j+2k+2)}{j}$ $\displaystyle\times\left(\psi_{0}(a+i+j+1)-\psi_{0}(a+i+1)+\psi_{0}(a+2k+2)\right).$ (236) In (236), the inner summations over $j$ are simplified into a closed-form representation except for the sum $\sum_{j=1}^{2k-i+1}\frac{\psi_{0}(a+i+j+1)}{j}.$ (237) Therefore, the resulting summations consist of several single sums as well as a double sum. These single sums are simplified into closed-form expressions by using the identity (132) and its derivative of $a$. The double summation involves the sum in (237), and is written as $\displaystyle\sum_{i=1}^{2k}\frac{4i(2k-i+2)\Gamma^{2}(a+2k+2)}{\Gamma(i)\Gamma(a+i)\Gamma(2k-i+2)\Gamma(a-i+2k+2)}\sum_{j=1}^{2k-i+1}\frac{\psi_{0}(a+i+j+1)}{j}.$ (238) To process (238), we first evaluate the inner summation by the identity (B.1) with the specialization $i=j,\qquad a=a+1,\qquad b=i,\qquad m=2k-i+1,$ (239) and the sum (238) becomes $\displaystyle\sum_{i=1}^{2k}\frac{4i(2k-i+2)\Gamma^{2}(a+2k+2)}{\Gamma(i)\Gamma(a+i)\Gamma(2k-i+2)\Gamma(a-i+2k+2)}\Bigg{(}\sum_{j=1}^{2k-i+1}\frac{\psi_{0}(i+j)}{j}$ $\displaystyle-\sum_{j=1}^{a+1}\frac{\psi_{0}(j+2k+1)}{i+j-1}+\frac{1}{2}(-\psi_{1}(a+i+1)+\psi_{1}(i)+\left(\psi_{0}(a+i+1)-\psi_{0}(i)\right)$ $\displaystyle\times\left(\psi_{0}(a+i+1)+2\left(\psi_{0}(2k-i+2)-\psi_{0}(1)\right)+\psi_{0}(i)\right))\\!\Bigg{)}.$ (240) In comparison with the summations in (233), the summations in (240) can be simplified similarly except for a double sum $\displaystyle\sum_{i=1}^{2k}\frac{4i(2k-i+2)\Gamma^{2}(a+2k+2)}{\Gamma(i)\Gamma(a+i)\Gamma(2k-i+2)\Gamma(a-i+2k+2)}\sum_{j=1}^{a+1}\frac{\psi_{0}(j+2k+1)}{i+j-1},$ (241) where it does not permit further simplifications by directly evaluating the inner sum. To proceed further (241), we change the summation order and take the partial fraction decomposition $\frac{4i(2k-i+2)}{i+j-1}=-\frac{4(j-1)(j+2k+1)}{i+j-1}-4(i-1)+4(j+2k).$ (242) The summation (241) is now written as $\displaystyle-\Gamma^{2}(a+2k+2)\sum_{j=1}^{a+1}4(j-1)(j+2k+1)\psi_{0}(j+2k+1)$ $\displaystyle\times\sum_{i=1}^{2k}\frac{1}{\Gamma(i)\Gamma(a+i)\Gamma(2k-i+2)\Gamma(a-i+2k+2)}\frac{1}{i+j-1}$ $\displaystyle-4\sum_{j=1}^{a+1}\psi_{0}(j+2k+1)\sum_{i=2}^{2k}\frac{\Gamma^{2}(a+2k+2)}{\Gamma(i-1)\Gamma(a+i)\Gamma(2k-i+2)\Gamma(a-i+2k+2)}$ $\displaystyle+4\sum_{j=1}^{a+1}\psi_{0}(j+2k+1)\sum_{i=1}^{2k}\frac{(j+2k)\Gamma^{2}(a+2k+2)}{\Gamma(i)\Gamma(a+i)\Gamma(2k-i+2)\Gamma(a-i+2k+2)}.$ (243) In (243), the second and third double summations are simplified directly into closed-form expressions by using the identities (132), and (287)–(B.1). The first double summation in (243), after applying the identity (1) with the specialization $b=a,\qquad c=j-1,\qquad m=2k+1,$ (244) is simplified to $\displaystyle\frac{(a+2k+1)\Gamma(a+2k+2)}{\Gamma(a+1)\Gamma(2k+1)}\sum_{j=1}^{a+1}\frac{4(j-1)(j+2k+1)}{(j+2k)}\psi_{0}(j+2k+1)-$ $\displaystyle\frac{(a+2k+1)\Gamma(a+2k+2)}{\Gamma(2a+2k+1)}\sum_{j=1}^{a+1}\sum_{i=1}^{2k+1}4(j-1)(j+2k+1)\psi_{0}(j+2k+1)$ $\displaystyle\times\frac{\Gamma(2a-i+4k+2)}{\Gamma(2k-i+2)\Gamma(a-i+2k+2)}\frac{\Gamma(2k-i+j+1)}{\Gamma(j+2k+1)}.$ (245) The single sum over $j$ in (245) admits a closed-form representation by applying the identity (287), (B.1), and (293). The double sum in (245) is simplified by first using the identities (307)–(B.1) to compute the sum over $j$, before using the identities (304)–(B.1) to evaluate the sum over $i$. Consequently, we obtain a semi closed-form result of (245), and the corresponding unsimplifiable basis are (212) and $\sum_{i=1}^{2k}\frac{\psi_{0}(2a+i+2k)}{i}.$ (246) We have so far completed the simplification of the inner summations over indexes $i$ and $j$ in (A.1). Summing up these results, we observe the complete cancellation among the unsimplifiable basis (212)-(213) and (226)-(228). The only survived term is (246). In the resulting outer sum over index $k$ in (A.1), all the gamma functions are reduced to rational ones. Therefore, (A.1) can now be simplified into a similar form as (A.1) in terms of the unsimplifiable basis (199). Inserting the resulting summations of (A.1) and (A.1) into (51), we obtain $\displaystyle\mathrm{I_{A}}=$ $\displaystyle\sum_{k=1}^{m}\left(\left(\frac{2a-1}{2k}-\frac{2a+1}{2(a+k)}+\frac{2a+1}{2k+1}-\frac{2a-1}{2a+2k+1}\right)\psi_{0}(2a+4k)\right.$ $\displaystyle+\left(\frac{2(a+m)(2a+3m-1)}{2a+4m-1}\left(\frac{1}{a+2k+1}+\frac{1}{a+2k}\right)-\frac{2a}{2k+1}\right.$ $\displaystyle+\\!\left.\frac{1-2a}{2k}+\frac{1}{2(a+k)}\right)\psi_{0}(2a+2k)+\left(\frac{2am-2a+6m^{2}-6m+1}{(2k+1)(2a+4m-1)}\right.$ $\displaystyle+\\!\left.\frac{4am+2a+12m^{2}-1}{4k(2a+4m-1)}+\frac{1}{4(a+k)}\right)\psi_{0}(a+2k)-\left(\frac{2a+2m-1}{2(2k+1)}\right.$ $\displaystyle+\\!\left.\frac{a+m}{2k}\right)\psi_{0}(a+k+m)-\left(\frac{2a+2m-1}{4k}+\frac{a+m}{2k+1}\right)$ $\displaystyle\times\\!\left.\psi_{0}\\!\left(a+k+m+\frac{1}{2}\right)-\left(\frac{1}{2(2k+1)}+\frac{1}{4k}\right)\psi_{0}(a+k)\right)+\mathrm{CF}.$ (247) We remind the readers that the omitted closed-form terms are denoted by the abbreviation $\mathrm{CF}$, which in general is different in each use. Now we discuss the simplification of the summations in (A.1) and (A.1) in computing the integral $\mathrm{I_{B}}$ in (55). The single summation in (A.1) are computed into the unsimplifiable basis of the form (199) by using the identities (287)–(B.1) along with the result (197). The simplification of the double summation in (A.1) will utilize Lemma 6. By partial fraction decomposing the rational functions in $j$ and shifting the summation index $j\to m-k-j$, the simplification of (A.1) boils down to computing the summations $\displaystyle\sum_{k=1}^{m-1}\mathrm{p}_{c,\lambda}(k)\frac{\Gamma(2m-2k+1)}{\Gamma(2a-2k+2m+1)}\sum_{j=1}^{m-k}\frac{\Gamma(2m+2a-2j-2k+1)}{\Gamma(2m-2j-2k+1)}$ $\displaystyle\times\left(\frac{1}{\left(a-j-2k+2m+\frac{1}{2}+c\right)^{\lambda}}-\frac{1}{(j+c)^{\lambda}}\right),$ (248) where the parameter $c$ and $\lambda$ take the values $c=-1/2,0,1/2$, $\lambda=1,2$, and $\mathrm{p}_{c,\lambda}(k)$ denotes the rational functions in $k$. These summations are simplified by using Lemma 6 to evaluate the inner sums over $j$, which reduces the gamma ratio $\frac{\Gamma(2m-2k+1)}{\Gamma(2a-2k+2m+1)}$ (249) into a rational function. Specifically, in the case when $c=1/2$ and $\lambda=1$ in (248), the corresponding inner summation is $\sum_{j=1}^{m-k}\frac{\Gamma(2m+2a-2j-2k+1)}{\Gamma(2m-2j-2k+1)}\left(\frac{1}{1+a-j-2k+2m}-\frac{1}{j+\frac{1}{2}}\right).$ (250) By using the relation Brychkov08 $\Gamma(2k)=\frac{2^{2k-1}}{\sqrt{\pi}}\Gamma(k)\Gamma\\!\left(k+\frac{1}{2}\right),$ (251) the summation (250) is written as $2^{2a}\sum_{j=1}^{m-k}\Phi_{0,-\frac{1}{2},a,a-\frac{1}{2}}^{(m-k-j)}\left(\frac{1}{1+a-j-2k+2m}-\frac{1}{j+\frac{1}{2}}\right).$ (252) Here, we recall the notation (162) $\Phi_{a,b,c,d}^{(x)}=\frac{\Gamma(x+c+1)\Gamma(x+d+1)}{\Gamma(x+a+1)\Gamma(x+b+1)}.$ (253) By using Lemma 6 with the specialization $a=-\frac{1}{2},~{}~{}~{}b=a,~{}~{}~{}c=\frac{1}{2},~{}~{}~{}m\to m-k,$ (254) (250) is simplified to $\displaystyle\sum_{j=1}^{m-k}\frac{\Gamma(2m+2a-2j-2k+1)}{\Gamma(2m-2j-2k+1)}\left(\frac{1}{1+a-j-2k+2m}-\frac{1}{j+\frac{1}{2}}\right)$ $\displaystyle=\frac{\Gamma(2a-2k+2m+2)}{\Gamma(2m-2k+2)}\Bigg{(}\psi_{0}(a+1)-\psi_{0}(a-2k+2m+1)$ $\displaystyle~{}~{}~{}\\!~{}-\frac{2(2a+1)}{2a-2k+2m+1}+2\Bigg{)}.$ (255) To simplify the summation (248) for $c=1/2$ and $\lambda=2$, we again use the relation (251), and the corresponding inner summation $\sum_{j=1}^{m-k}\frac{\Gamma(2m+2a-2j-2k+1)}{\Gamma(2m-2j-2k+1)}\left(\frac{1}{(1+a-j-2k+2m)^{2}}-\frac{1}{\left(j+\frac{1}{2}\right)^{2}}\right)$ (256) becomes $2^{2a}\sum_{j=1}^{m-k}\Phi_{0,-\frac{1}{2},a,a-\frac{1}{2}}^{(m-k-j)}\left(\frac{1}{(1+a-j-2k+2m)^{2}}-\frac{1}{(j+\frac{1}{2})^{2}}\right),$ (257) which is evaluated by taking derivative of the parameter $c$ of the identity (174) with the specialization $a=-\frac{1}{2},~{}~{}~{}b=a,~{}~{}~{}m\to m-k,$ (258) before setting $c=1/2$. As a result, (256) is simplified to $\displaystyle\sum_{j=1}^{m-k}\frac{\Gamma(2m+2a-2j-2k+1)}{\Gamma(2m-2j-2k+1)}\left(\frac{1}{(1+a-j-2k+2m)^{2}}-\frac{1}{\left(j+\frac{1}{2}\right)^{2}}\right)$ $\displaystyle=\frac{\Gamma(2m+2a-2k+2)}{\Gamma(2m-2k+2)}\sum_{j=1}^{m-k}\Bigg{(}\\!\left(-\frac{4a^{2}-1}{a+j-1}-\frac{1-4a^{2}}{a+j}-\frac{4a^{2}-6a+2}{-2a-2j+3}\right.$ $\displaystyle~{}~{}~{}\\!~{}-\\!\left.\frac{2a(2a+1)}{2a+2j+1}+\frac{1}{a+2j}-\frac{2-8a}{2a+2j-1}-\frac{1}{-a-2j+1}\right)\left(\psi_{0}\\!\left(j+\frac{1}{2}\right)\right.$ $\displaystyle~{}~{}~{}\\!~{}-\psi_{0}\\!\left(a+j+\frac{1}{2}\right)+\psi_{0}(a-k+m+1)+\psi_{0}\\!\left(a-k+m+\frac{3}{2}\right)+\psi_{0}(j)$ $\displaystyle~{}~{}~{}\\!~{}-\\!\left.\psi_{0}(a+j)-\psi_{0}(m-k+1)-\psi_{0}\\!\left(m-k+\frac{3}{2}\right)\right)+\frac{a+j}{j(a+2j)^{2}}$ $\displaystyle~{}~{}~{}\\!~{}+\frac{1}{2a+2j-1}\left(\frac{a(2a-1)(2j-1)}{j(a+j)^{2}}-\frac{2a(2a+1)}{\left(a+j+\frac{1}{2}\right)^{2}}+\frac{2j-1}{(a+2j-1)^{2}}\right)\\!\Bigg{)}.$ (259) The other cases of $c$ and $\lambda$ combinations are obtained similarly. Inserting the results into (248), the gamma ratio (249) of the outer sum over $k$ is reduced to a rational function. The resulting summations are further computed into unsimplifiable basis of the form (199) by using the identities (287)–(B.1). Inserting the resulting summations of (A.1) and (A.1) into (55), we obtain $\displaystyle\mathrm{I_{B}}=$ $\displaystyle\sum_{k=1}^{m}\left(\left(\frac{2(a+m)(2a+3m-1)}{2a+4m-1}\left(\frac{1}{a+2k+1}+\frac{1}{a+2k}\right)+\frac{m}{k}+\frac{2m-1}{2k+1}\right)\right.$ $\displaystyle\times\psi_{0}(2a+2k)+\left(\frac{m}{2a+2k+1}+\frac{2m-1}{4(a+k)}-\frac{2m-1}{2(2k+1)}-\frac{m}{2k}\right)\times$ $\displaystyle\psi_{0}\\!\left(a+2k+\frac{1}{2}\right)+\left(\frac{m}{2(a+k)}+\frac{m}{2a+2k+1}+\frac{m(2m-1)}{2a+4m-1}\right.$ $\displaystyle\times\left.\left.\left(\frac{1}{2k+1}+\frac{1}{2k}\right)\right)\psi_{0}(a+2k)\right)+\mathrm{CF}.$ (260) Now inserting the $\mathrm{I_{A}}$ expression (247) and $\mathrm{I_{B}}$ expression (260) into (47), we obtain $\displaystyle\mathbb{V}\\!\left[S\right]=$ $\displaystyle\frac{-2a-2m+1}{4}\Omega_{1}^{\left(a,a+\frac{1}{2}\right)}+\frac{-2a-2m+1}{4}\left(\Omega_{2}^{\left(a+\frac{1}{2},0\right)}+\Omega_{2}^{\left(a+\frac{1}{2},\frac{1}{2}\right)}\right)$ $\displaystyle-\frac{a+m}{2}\left(\Omega_{2}^{\left(a,0\right)}+\Omega_{2}^{\left(a,\frac{1}{2}\right)}\right)+\mathrm{CF},$ (261) where $\displaystyle\Omega_{1}^{\left(a,b\right)}=$ $\displaystyle\sum_{k=1}^{m}\left(\frac{\psi_{0}(a+k)}{b+k}+\frac{\psi_{0}(b+k)}{a+k}\right)$ (262) $\displaystyle\Omega_{2}^{\left(a,b\right)}=$ $\displaystyle\sum_{k=1}^{m}\left(\frac{\psi_{0}(a+b+k+m)}{b+k}+\frac{\psi_{0}(a+k)}{b+k}+\frac{\psi_{0}(a+b+2k)}{a+k}\right.$ $\displaystyle-\\!\left.\frac{\psi_{0}(a+b+k)}{a+k}-\frac{\psi_{0}(a+b+2k)}{b+k}\right).$ (263) Simplifying the single sums in (261) by using the closed-form identities (B.1) and (B.1) directly leads to the desired result (3). This completes the proof of Proposition 3. #### Case B: Fixed number of particles In case B, the variance calculation boils down to simplifying the summation representations of $\mathcal{A}_{1}$, $\mathcal{A}_{2}$, $\mathcal{B}_{1}$, $\mathcal{B}_{2}$ as summarized in (281)–(A.2) in Appendix A.2. Note that the simplification procedure in case A also works for the majority of the summations in case B. The only new summation in case B is the double summation $\displaystyle\sum_{k=1}^{m-2}4(a+b-2k+2m+1)\frac{\Gamma(m-k+1)}{\Gamma(a+b-k+m+1)}\sum_{j=1}^{m-k-1}\frac{\Gamma(a+b-j-k+m)}{\Gamma(m-k-j)}$ $\displaystyle\times(-1)^{j}(2j+2k-2m-a-b+1)\left(\frac{1}{(j)_{3}(a+b-j-2k+2m-1)_{3}}\right)^{2}$ $\displaystyle\times\left((1-a)j(a+b-j-2k+2m-1)-2(a-k+m)(a+b-k+m)\right)$ $\displaystyle\times\left((1-b)j(a+b-j-2k+2m-1)-2(b-k+m)(a+b-k+m)\right),$ (264) which is obtained after opening the bracket of the double summation in (A.2) and shifting the index $k\to m-j-k$. To process (264), we take the partial fraction decomposition of the rational functions (starting from the second to the last line) in $j$, the summation (264) now boils down to $\displaystyle\sum_{k=1}^{m-2}\mathrm{p}_{c,\lambda}(k)\frac{\Gamma(m-k+1)}{\Gamma(a+b-k+m+1)}\sum_{j=1}^{m-k-1}\frac{(-1)^{j}\Gamma(m-k+a+b-j)}{\Gamma(m-k-j)}$ $\displaystyle\times\left(\frac{1}{(2m-2k-1+a+b-j+c)^{\lambda}}-\frac{1}{(j+c)^{\lambda}}\right),$ (265) where the parameters $c$ and $\lambda$ take the values $c=0,1,2$, $\lambda=1,2$, and $\mathrm{p}_{c,\lambda}(k)$ denotes the rational polynomials in $k$. The simplification of these summations will utilize the Lemma 5. For the case when $c=0$ and $\lambda=1$ in (265), the corresponding inner summation can be simplified into a closed-form expression by directly using the result (2.2.2), which is a special case of the identity (163) in Lemma 5. When $c=0$ and $\lambda=2$ in (265), the inner summation $\sum_{j=1}^{m-k-1}\frac{(-1)^{j}\Gamma(a+b-j-k+m)}{\Gamma(m-k-j)}\left(\frac{1}{(a+b-j-2k+2m-1)^{2}}-\frac{1}{j^{2}}\right)$ (266) is simplified to $\displaystyle\frac{\Gamma(a+b-k+m)}{\Gamma(m-k)}\sum_{j=1}^{m-k-1}\Bigg{(}\\!\left(\frac{1}{a+b-j-k+m}-\frac{1}{a+b-2j-2k+2m}\right.$ $\displaystyle-\\!\left.\frac{1}{a+b-2j-2k+2m-1}\right)(\psi_{0}(m-k)-\psi_{0}(a+b-j-k+m)$ $\displaystyle-\psi_{0}(a+b-k+m)-\psi_{0}(m-k-j))+\frac{1}{(a+b-2j-2k+2m-1)^{2}}\Bigg{)}.$ (267) The result (267) is obtained by taking derivative of $c$ of the identity (163) with the specialization $a\to a+b,~{}~{}~{}m\to m-k-1,$ (268) before setting $c=0$. For other combinations of $c$ and $\lambda$, the corresponding inner summations in (265) can be simplified similarly as the two cases above. Inserting the simplification results of the inner summations into (264), the resulting sums only consist of polygamma and rational functions, which are further computed into unsimplifiable basis of the form (199) by using the identities (287)–(B.1). For the integrals $\mathrm{I_{A}}$ and $\mathrm{I_{B}}$ in case B, the corresponding summations (281)–(A.2) are now simplified to the results shown below. For $\mathrm{I_{A}}$, one has $\displaystyle\mathrm{I_{A}}=~{}\\!\\!$ $\displaystyle d_{1}\sum_{k=1}^{m}\frac{\psi_{0}(a+k)}{k}+d_{2}\sum_{k=1}^{m}\frac{\psi_{0}(b+k)}{k}-2m\sum_{k=1}^{m}\frac{\psi_{0}(a+b+k+m)}{k}$ $\displaystyle+d_{3}\sum_{k=1}^{m}\frac{\psi_{0}(a+b+k+m)}{a+k}+\left(d_{3}+d_{4}\right)\sum_{k=1}^{m}\frac{\psi_{0}(a+b+k+m)}{b+k}+\mathrm{CF},$ (269) where the coefficients $d_{i}$ are $\displaystyle d_{1}=$ $\displaystyle\frac{2m(a+m)\left(a^{2}+a(b+3m)+2bm+3m^{2}-1\right)}{(a+b+2m-1)_{3}}$ (270) $\displaystyle d_{2}=$ $\displaystyle\frac{2m(b+m)\left(a(b+2m)+b^{2}+3bm+3m^{2}-1\right)}{(a+b+2m-1)_{3}}$ (271) $\displaystyle d_{3}=$ $\displaystyle\frac{2(b+m)(a+b+m)\left(m(3a+4b)+(a+b)^{2}+3m^{2}-1\right)}{(a+b+2m-1)_{3}}$ (272) $\displaystyle d_{4}=$ $\displaystyle\frac{2(a-b)(a+b+m)}{a+b+2m}.$ (273) For $\mathrm{I_{B}}$, one has $\displaystyle\mathrm{I_{B}}=~{}\\!\\!$ $\displaystyle d_{1}\sum_{k=1}^{m}\frac{\psi_{0}(a+k)}{k}+d_{2}\sum_{k=1}^{m}\frac{\psi_{0}(b+k)}{k}+2m\sum_{k=1}^{m}\frac{\psi_{0}(a+b+k)}{k}$ $\displaystyle-2m\sum_{k=1}^{m}\frac{\psi_{0}(a+b+2k)}{k}+d_{3}\sum_{k=1}^{m}\frac{\psi_{0}(a+b+k)}{b+k}+d_{4}\sum_{k=1}^{m}\frac{\psi_{0}(b+k)}{a+k}$ $\displaystyle- d_{4}\sum_{k=1}^{m}\frac{\psi_{0}(a+b+2k)}{a+k}+d_{4}\sum_{k=1}^{m}\frac{\psi_{0}(a+b+2k)}{b+k}$ $\displaystyle+\left(d_{3}+d_{4}\right)\sum_{k=1}^{m}\frac{\psi_{0}(a+b+k)}{a+k}+\mathrm{CF}.$ (274) Inserting the $\mathrm{I_{A}}$ expression (269) and $\mathrm{I_{B}}$ expression (274) into (47), we arrive at $\displaystyle\mathbb{V}\\!\left[S\right]=$ $\displaystyle-2m\Omega_{2}^{(a+b,0)}-d_{4}\Omega_{2}^{(b,a)}+\left(d_{3}+d_{4}\right)\Omega_{3}^{(a,b)}+\mathrm{CF},$ (275) where the summation $\Omega_{2}^{(a,b)}$ is defined in (263), and $\displaystyle\Omega_{3}^{(a,b)}=~{}\\!\\!$ $\displaystyle\sum_{k=1}^{m}\left(\frac{\psi_{0}(a+b+k+m)}{a+k}+\frac{\psi_{0}(a+b+k+m)}{b+k}-\frac{\psi_{0}(a+b+k)}{a+k}\right.$ $\displaystyle-\\!\left.\frac{\psi_{0}(a+b+k)}{b+k}\right).$ (276) By using the identities (B.1) and (B.1), the result (275) is simplified to the variance formula (4), which completes the proof of Proposition 4. ## 3 Conclusions In this work, we compute the exact yet explicit variance formulas of von Neumann entanglement entropy over fermionic Gaussian states with and without particle number constrains. The obtained formulas provide insights into the fluctuations of von Neumann entropy. An essential ingredient in obtaining the results is a new simplification framework of dummy summation and re-summation techniques. The new framework may also be useful in computing higher order moments of von Neumann entropy as well as other entanglement indicators over the fermionic Gaussian ensemble. Acknowledgments The work of Lu Wei is supported in part by the U.S. National Science Foundation ($\\#$2150486). ## Appendix A Summation representations In this appendix, we list the summation representations of the integrals in $\mathrm{I_{A}}$ and $\mathrm{I_{B}}$ as summarized in Table 3. The summation representations for case A are listed in Appendix (A.1), the ones for case B are listed in Appendix (A.2). ### A.1 Summation representations of case A 0pt $\displaystyle\mathrm{A_{1}}=$ $\displaystyle\sum_{k=0}^{m-1}\sum_{j=2k-2}^{2k}\frac{2(-1)^{j}(2a+4k+1)(j+1)_{2}(a+j+1)_{2}}{\Gamma(2k-j+1)\Gamma(j-2k+3)(2a+j+2k+1)_{3}}\bigg{(}(\psi_{0}(j+3)$ $\displaystyle-\psi_{0}(2a+j+2k+4)-\psi_{0}(j-2k+3)+\psi_{0}(a+j+3))^{2}$ $\displaystyle-\psi_{1}(2a+j+2k+4)+\psi_{1}(a+j+3)-\psi_{1}(j-2k+3)+\psi_{1}(j+3)\bigg{)}$ $\displaystyle+\sum_{k=0}^{m-1}\sum_{j=0}^{2k-3}\frac{4(2a+4k+1)(j+1)_{2}(a+j+1)_{2}}{(2k-j-2)_{3}(2a+j+2k+1)_{3}}(\psi_{0}(2a+j+2k+4)$ $\displaystyle-\psi_{0}(a+j+3)+\psi_{0}(2k-j-2)-\psi_{0}(j+3))$ (277) $\displaystyle\mathrm{A_{2}}=$ $\displaystyle\sum_{k=0}^{m-1}\frac{(2a+4k+1)\Gamma(2k+1)\Gamma(2a+2k+1)}{\Gamma(2a+4k+4)}\left(\rule{0.0pt}{19.0633pt}\sum_{j=0}^{2k}\frac{2(j+1)(2k-j+1)}{\Gamma(j+1)\Gamma(a+j+1)}\right.$ $\displaystyle\times\frac{\Gamma^{2}(a+2k+2)}{\Gamma(2k-j+1)\Gamma(a+2k-j+1)}((\psi_{0}(a+2k+2)-\psi_{0}(2a+4k+4)$ $\displaystyle-\psi_{0}(2)+\psi_{0}(2k-j+2))(\psi_{0}(a+2k+2)-\psi_{0}(2a+4k+4)+\psi_{0}(j+2)$ $\displaystyle-\psi_{0}(2))-\psi_{1}(2a+4k+4))-\sum_{j=0}^{2k}\frac{(j+1)\Gamma(a+2k+1)\Gamma(a+2k+3)}{\Gamma(j)\Gamma(a+j+1)\Gamma(a-j+2k+1)}$ $\displaystyle\times\frac{1}{\Gamma(2k-j+1)}((\psi_{0}(a+2k+1)-\psi_{0}(2a+4k+4)+\psi_{0}(2k-j+2)$ $\displaystyle-\psi_{0}(1))(\psi_{0}(a+2k+3)-\psi_{0}(2a+4k+4)+\psi_{0}(j+2)-\psi_{0}(3))$ $\displaystyle-\psi_{1}(2a+4k+4))-\sum_{j=0}^{2k}\frac{(2k-j+1)\Gamma(a+2k+1)\Gamma(a+2k+3)}{\Gamma(a+j+1)\Gamma(2k-j)\Gamma(2k-j+a+1)}$ $\displaystyle\times\frac{1}{\Gamma(j+1)}((\psi_{0}(a+2k+3)-\psi_{0}(2a+4k+4)+\psi_{0}(2k-j+2)$ $\displaystyle-\psi_{0}(3))(\psi_{0}(a+2k+1)-\psi_{0}(2a+4k+4)+\psi_{0}(j+2)-\psi_{0}(1))$ $\displaystyle-\psi_{1}(2a+4k+4))+\sum_{j=1}^{2k-1}\sum_{i=1}^{2k-j}\frac{4i(2k-i+2)\Gamma(a+2k-j+1)}{\Gamma(a+i)\Gamma(a+2k-i+2)}$ $\displaystyle\times\frac{\Gamma(a+2k+j+3)}{(j)_{3}\Gamma(j+i+1)\Gamma(2k-j-i+1)}(\psi_{0}(a+2k+j+3)$ $\displaystyle-\psi_{0}(2a+4k+4)+\psi_{0}(2k-i+3)-\psi_{0}(j+3))\left.\\!\\!\rule{0.0pt}{19.0633pt}\right)$ (278) $\displaystyle\mathrm{B_{1}}=$ $\displaystyle\sum_{k=0}^{m-1}\left(\psi_{0}(a+2k)+\psi_{0}(2a+2k)-2\psi_{0}(2a+4k)-\frac{1}{2}\left(\frac{a}{a+2k+1}\right.\right.~{}~{}~{}~{}~{}~{}~{}$ $\displaystyle\\!\left.\left.+\frac{a}{a+2k}+\frac{2}{2a+4k+1}\right)+1\right)^{2}$ (279) $\displaystyle\mathrm{B_{2}}=$ $\displaystyle\sum_{k=1}^{m-1}\sum_{j=1}^{m-k}\frac{\Gamma(2a+2k-1)\Gamma(2j+2k-1)}{2(2j-1)^{2}j^{2}(2j+1)^{2}\Gamma(2k-1)\Gamma(2a+2j+2k-1)}$ $\displaystyle\times\frac{(2a+4k-3)(2a+4j+4k-3)}{(a+j+2k-2)^{2}(a+j+2k-1)^{2}(2a+2j+4k-3)^{2}}\left(a^{2}(2j+1)\right.$ $\displaystyle+\\!\left.a(j+1)(2j+4k-3)+2j^{2}+j(4k-3)+4k^{2}-6k+2\right)^{2}$ (280) ### A.2 Summation representations of case B $\mathcal{A}_{1}=\mathcal{A}_{1}^{(a,b)}+\mathcal{A}_{1}^{(b,a)},$ (281) 0pt $\displaystyle\mathcal{A}_{1}^{(a,b)}=$ $\displaystyle-\frac{2(m(b+m))}{a+b+2m}\sum_{i=1}^{m-3}\frac{(b+i+1)(i)_{2}}{(m-i-2)_{3}(a+b+i+m+1)}(\psi_{0}(b+i+2)$ $\displaystyle-\psi_{0}(a+b+i+m+2)-\psi_{0}(m-i-2)+\psi_{0}(i+2))+\frac{a+b+m}{a+b+2m}$ $\displaystyle\times 2(a+m)\sum_{i=1}^{m-2}\frac{(b+i+1)(i)_{2}}{(m-i-1)(a+b+i+m)_{3}}(-\psi_{0}(a+b+i+m+3)$ $\displaystyle+\psi_{0}(b+i+2)-\psi_{0}(m-i-1)+\psi_{0}(i+2))-\frac{m(b+m)}{a+b+2m}$ $\displaystyle\sum_{i=m-3}^{m-1}\frac{(b+i+2)(-1)^{i+m}(i+1)_{2}}{\Gamma(m-i)\Gamma(i-m+4)(a+b+i+m+2)}\big{(}\psi_{1}(b+i+3)$ $\displaystyle+\psi_{1}(i+3)-\psi_{1}(i-m+4)-\psi_{1}(a+b+i+m+3)+(\psi_{0}(i+3)$ $\displaystyle-\psi_{0}(a+b+i+m+3)-\psi_{0}(i-m+4)+\psi_{0}(b+i+3))^{2}\big{)}$ $\displaystyle-\frac{(a+m)(a+b+m)(b+m)(m-1)_{2}}{(a+b+2m)(a+b+2m-1)_{3}}\big{(}\\!-\psi_{1}(a+b+2m+2)$ $\displaystyle+\psi_{1}(b+m+1)+\psi_{1}(m+1)-\psi_{1}(1)+\psi_{0}^{2}(1)+(\psi_{0}(b+m+1)$ $\displaystyle-\psi_{0}(a+b+2m+2)+\psi_{0}(m+1))(\psi_{0}(b+m+1)+\psi_{0}(m+1)$ $\displaystyle-\psi_{0}(a+b+2m+2)-2\psi_{0}(1))\big{)}$ (282) $\displaystyle\mathcal{A}_{2}=$ $\displaystyle-\frac{2(a+m)(b+m)\Gamma(m+1)}{(a+b+2m)(a+b+m+2)_{m}}\sum_{i=0}^{m-1}(-1)^{i}\frac{(i+1)(m-i)}{\Gamma(a+i+2)}$ $\displaystyle\times\sum_{j=i-1}^{i+1}(-1)^{j}\frac{\Gamma(a+i-j+m+1)(b-i+m+1)_{j}}{\Gamma(j+1)\Gamma(i-j+2)\Gamma(j-i+2)\Gamma(m-j)}$ $\displaystyle\times(\psi_{1}(a+b+2m+2)+(\psi_{0}(a+i-j+m+1)-\psi_{0}(a+b+2m+2)$ $\displaystyle-\psi_{0}(i-j+2)+\psi_{0}(i+2))(\psi_{0}(a+b+2m+2)+\psi_{0}(j-i+2)$ $\displaystyle-\psi_{0}(b-i+j+m+1)-\psi_{0}(m-i+1)))-2\Gamma(m+1)\Gamma(a+m+1)$ $\displaystyle\times\frac{(a+b+m)\Gamma(b+m+1)}{(a+b+2m)(a+b+m+1)_{m+1}}\sum_{i=0}^{m-2}\frac{1}{\Gamma(i+1)\Gamma(a+i+2)}$ $\displaystyle\times\frac{1}{\Gamma(m-i-1)\Gamma(b-i+m)}(\psi_{1}(a+b+2m+2)+(\psi_{0}(a+m+1)$ $\displaystyle-\psi_{0}(a+b+2m+2)+\psi_{0}(i+2)-\psi_{0}(1))(\psi_{0}(a+b+2m+2)$ $\displaystyle-\psi_{0}(b+m+1)-\psi_{0}(m-i)+\psi_{0}(1)))+\mathcal{A}_{2}^{(a,b)}+\mathcal{A}_{2}^{(b,a)},$ (283) $\displaystyle\mathcal{A}_{2}^{(a,b)}=$ $\displaystyle\frac{2\Gamma(m+1)\Gamma(a+b+m+1)}{\Gamma(a+b+2m+2)}\Bigg{(}\frac{(a+m)(b+m)(a+b+m+1)}{a+b+2m}$ $\displaystyle\times\sum_{i=1}^{m-2}\frac{i(m-i+1)}{\Gamma(b+i+1)\Gamma(a-i+m+2)}\sum_{j=1}^{m-i-1}\frac{\Gamma(a+j+m+2)}{(j)_{3}\Gamma(i+j+1)}$ $\displaystyle\times\frac{\Gamma(b-j+m)}{\Gamma(m-i-j)}(\psi_{0}(m-i+2)-\psi_{0}(a+b+2m+2)-\psi_{0}(j+3)$ $\displaystyle+\psi_{0}(a+j+m+2))-\frac{a+b+m}{a+b+2m}\sum_{i=1}^{m-1}\frac{i(m-i)}{\Gamma(b+i+1)}$ $\displaystyle\times\frac{1}{\Gamma(a-i+m+1)}(\psi_{0}(a+j+m+1)-\psi_{0}(a+b+2m+2)$ $\displaystyle+\psi_{0}(m-i+1)-\psi_{0}(j+1))\Bigg{)}$ (284) $\displaystyle\mathcal{B}_{1}=$ $\displaystyle\sum_{k=0}^{m-1}\left(\left(\frac{a^{2}-b^{2}}{4(a+b+2k)}+\frac{b^{2}-a^{2}}{4(a+b+2k+2)}+\frac{1}{2}\right)\psi_{0}(a+k+1)\right.$ $\displaystyle+\left(\frac{a^{2}-b^{2}}{4(a+b+2k+2)}+\frac{b^{2}-a^{2}}{4(a+b+2k)}+\frac{1}{2}\right)\psi_{0}(b+k+1)$ $\displaystyle+\psi_{0}(a+b+k+1)-2\psi_{0}(a+b+2k+2)-\frac{a+b}{2(a+b+2k)}$ $\displaystyle-\\!\left.\frac{a+b}{2(a+b+2k+2)}+\frac{1}{a+b+2k+1}+1\right)^{2}$ (285) $\displaystyle\mathcal{B}_{2}=$ $\displaystyle\sum_{k=1}^{m-1}\frac{k(a+b+k)}{2(a+k)(b+k)(a+b+2k)(a+b+2k-1)_{3}}\Bigg{(}2(a+k)(b+k)$ $\displaystyle\times(\psi_{0}(b+k+1)-\psi_{0}(a+k+1))+\frac{(k-1)(a-b)(a+b+2k+1)}{a+b+k}\\!\Bigg{)}^{2}$ $\displaystyle+\sum_{k=1}^{m-2}\sum_{j=1}^{m-k-1}\frac{2(a+b+2k-1)\Gamma(j+k+1)(a+b+2j+2k+1)}{\Gamma(k)\Gamma(a+k)\Gamma(b+k)\Gamma(a+j+k+1)\Gamma(b+j+k+1)}$ $\displaystyle\times\frac{\Gamma(a+b+k)}{\Gamma(a+b+j+k+1)}\Bigg{(}\frac{\Gamma(a+k)\Gamma(b+j+k+1)}{(j)_{3}(a+b+j+2k-1)_{3}}\left(a^{2}(j+2)+a(j+2)\right.$ $\displaystyle\times\left.\\!(b+j+2k)+j(b+2k+1)+2k(b+k)+j^{2}\right)$ $\displaystyle-\frac{(-1)^{j}\Gamma(b+k)\Gamma(a+j+k+1)}{(j)_{3}(a+b+j+2k-1)_{3}}(2(k-1)(a+(b+1)(j+2)+2k-2)$ $\displaystyle+(b+1)(j+2)(a+b+j+1))\\!\Bigg{)}^{2}$ (286) ## Appendix B Summation identities In this appendix, we list the finite sum identities useful in simplifying the summations in Appendix A that are not listed in the main part of the paper. The identities of the existing simplification framework are listed in Appendix B.1 and the identities of the new simplification framework are listed in Appendix B.2. ### B.1 Summation identities of the existing simplification framework We list below the identities of the existing simplification framework. Here, it is sufficient to assume that $a,b\geq 0,a\neq b$ in identities (287)–(293), (B.1)–(B.1), (B.1), (B.1)–(B.1), $a>m$ in (B.1), and $a,b\geq 1,n\geq m$ in (B.1), (304)–(B.1). 0pt $\displaystyle\sum_{i=1}^{m}\psi_{0}(i+a)=(m+a)\psi_{0}(m+a+1)-a\psi_{0}(a+1)-m$ (287) $\displaystyle\sum_{i=1}^{m}i\psi_{0}(i+a)=$ $\displaystyle-\frac{1}{2}(a-m-1)(a+m)\psi_{0}(a+m+1)$ $\displaystyle+\frac{1}{2}(a-1)a\psi_{0}(a+1)-\frac{1}{4}m(-2a+m+3)$ (288) $\displaystyle\sum_{i=1}^{m}i^{2}\psi_{0}(i+a)=$ $\displaystyle\frac{1}{6}\left(2a^{3}-3a^{2}+a+2m^{3}+3m^{2}+m\right)\psi_{0}(a+m+1)$ $\displaystyle-\frac{1}{6}a\left(2a^{2}-3a+1\right)\psi_{0}(a+1)$ $\displaystyle-\frac{1}{36}m\left(12a^{2}-6am-24a+4m^{2}+15m+17\right)$ (289) $\displaystyle\sum_{i=1}^{m}i^{3}\psi_{0}(i+a)=$ $\displaystyle-\frac{1}{4}\left(a^{4}-2a^{3}+a^{2}-m^{4}-2m^{3}-m^{2}\right)\psi_{0}(a+m+1)$ $\displaystyle+\frac{1}{4}(a-1)^{2}a^{2}\psi_{0}(a+1)-\frac{1}{48}m\left(-12a^{3}+6a^{2}m+30a^{2}\right.$ $\displaystyle-\\!\left.4am^{2}-18am-26a+3m^{3}+14m^{2}+21m+10\right)$ (290) $\displaystyle\sum_{i=1}^{m}\psi_{1}(i+a)=$ $\displaystyle(m+a)\psi_{1}(m+a+1)-a\psi_{1}(a+1)$ $\displaystyle+\psi_{0}(m+a+1)-\psi_{0}(a+1)$ (291) $\displaystyle\sum_{i=1}^{m}\psi_{0}^{2}(i+a)=$ $\displaystyle(a+m)\psi_{0}^{2}(a+m){}-(2a+2m-1)\psi_{0}(a+m)-a\psi_{0}^{2}(a){}$ $\displaystyle+(2a-1)\psi_{0}(a)+2m.$ (292) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(i+a)}{i+a}=\frac{1}{2}\left(\psi_{1}(m+a+1)-\psi_{1}(a+1)+\psi_{0}^{2}(m+a+1)-\psi_{0}^{2}(a+1)\right)$ (293) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(m+1-i)}{i}=\psi_{0}^{2}(m+1)-\psi_{0}(1)\psi_{0}(m+1)+\psi_{1}(m+1)-\psi_{1}(1)$ (294) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(m+1+i)}{i}=\psi_{0}^{2}(m+1)-\psi_{0}(1)\psi_{0}(m+1)-\frac{1}{2}\psi_{1}(m+1)+\frac{1}{2}\psi_{1}(1)$ (295) 0pt $\displaystyle\sum_{i=1}^{m}\psi_{0}(i+a)\psi_{0}(i+b)$ $\displaystyle=(b-a)\sum_{i=1}^{m-1}\frac{\psi_{0}(a+i)}{b+i}-a\psi_{0}(a+1)\psi_{0}(b+1)+(m+a)\psi_{0}(m+a)$ $\displaystyle~{}~{}~{}\\!~{}\times\psi_{0}(m+b)+a\psi_{0}(a+1)-(m+a-1)\psi_{0}(m+a)-(m+b)\psi_{0}(m+b)$ $\displaystyle~{}~{}~{}\\!~{}+(b+1)\psi_{0}(b+1)+2m-2$ (296) $\displaystyle\sum_{i=1}^{m}i\psi_{0}(i+a)\psi_{0}(i+b)$ $\displaystyle=\frac{1}{2}(b-a+1)(a-b)\sum_{i=1}^{m-1}\frac{\psi_{0}(a+i)}{b+i}-\frac{1}{4}a(a+2b-3)\psi_{0}(a+1)-\frac{1}{4}(b+1)$ $\displaystyle~{}~{}~{}\\!~{}\times(2a+b-2)\psi_{0}(b+1)+\frac{1}{2}(a-1)a\psi_{0}(a+1)\psi_{0}(b+1)+\frac{1}{4}(a+m-1)$ $\displaystyle~{}~{}~{}\\!~{}\times(a+2b-m-2)\psi_{0}(a+m)+\frac{1}{4}(b+m)(2a+b-m-1)\psi_{0}(b+m)-\frac{1}{2}$ $\displaystyle~{}~{}~{}\\!~{}\times\left(a^{2}-a-m(m+1)\right)\psi_{0}(a+m)\psi(b+m)-\frac{1}{4}(m-1)(3a+3b-m-4)$ (297) $\displaystyle\sum_{i=1}^{m}i^{2}\psi_{0}(i+a)\psi_{0}(i+b)$ $\displaystyle=\frac{1}{6}(a-b)\left(3a^{2}+2ab-4a-2b^{2}-b+1\right)\sum_{i=1}^{m-1}\frac{\psi_{0}(a+i)}{b+i}-\frac{1}{6}\left(-a^{2}(5b+2)\right.$ $\displaystyle~{}~{}~{}\\!~{}+\\!\left.3a^{3}+a(5b-1)-m\left(2m^{2}+3m+1\right)\right)\psi_{0}(a+m)\psi_{0}(b+m)+\frac{1}{6}(a-1)a$ $\displaystyle~{}~{}~{}\\!~{}\times(3a-5b+1)\psi_{0}(a+1)\psi_{0}(b+1)-\Bigg{(}\\!-\frac{1}{12}(2b-1)m^{2}+\frac{1}{36}m\left(24a^{2}\right.$ $\displaystyle~{}~{}~{}\\!~{}-\left.\\!24ab-24a+12b^{2}+12b-1\right)+\frac{1}{36}(a-1)\left(28a^{2}-18ab-5a+6b\right.$ $\displaystyle~{}~{}~{}\\!~{}+\\!\left.12b^{2}+6\right)+\frac{(a-1)(a-b)}{3(b+m-1)}+\frac{m^{3}}{9}\Bigg{)}\psi_{0}(a+m)-\frac{1}{36}\left(4m^{3}12a^{2}b\right.$ $\displaystyle~{}~{}~{}\\!~{}-3(2a-1)m^{2}+\left(12a^{2}-12a-1\right)m-30a^{2}+6ab^{2}-12ab+30a+4b^{3}$ $\displaystyle~{}~{}~{}\\!~{}-\\!\left.3b^{2}-b\right)\psi_{0}(b+m)+\frac{1}{36}a\left(28a^{2}-9a(2b+1)+12b^{2}-13\right)\psi_{0}(a+1)$ $\displaystyle~{}~{}~{}\\!~{}+\frac{1}{36}\left(6a^{2}(2b-3)+6a\left(b^{2}-2b+2\right)+4b^{3}-3b^{2}-b+6\right)\psi_{0}(b+1)$ $\displaystyle~{}~{}~{}\\!~{}+\frac{2m^{3}}{27}-\frac{5}{36}m^{2}(a+b-1)+\frac{1}{36}\left(-40a^{2}+12ab+51a-16b^{2}+3b-16\right)$ $\displaystyle~{}~{}~{}\\!~{}+\frac{1}{108}m\left(120a^{2}-36ab-138a+48b^{2}+6b+25\right)+\frac{a-1}{3(b+m-1)}$ $\displaystyle~{}~{}~{}\\!~{}-\frac{a-1}{3(a+m-1)}$ (298) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(a+1-i)}{i}=$ $\displaystyle-\sum_{i=1}^{m}\frac{\psi_{0}(i+a-m)}{i}+\frac{1}{2}\left(\psi_{1}(a+1)-\psi_{1}(a-m)\right)$ $\displaystyle+(\psi_{0}(a-m)+\psi_{0}(a+1))(\psi_{0}(m+1)-\psi_{0}(1))$ $\displaystyle+\frac{1}{2}(\psi_{0}(a-m)-\psi_{0}(a+1))^{2}$ (299) $\displaystyle\sum_{i=1}^{m}\left(\frac{\psi_{0}(i+b)}{i+a}+\frac{\psi_{0}(i+a)}{i+b}\right)$ $\displaystyle=\psi_{0}(m+a+1)\psi_{0}(m+b+1)-\psi_{0}(a+1)\psi_{0}(b+1)+\frac{1}{a-b}(\psi_{0}(m+a+1)$ $\displaystyle~{}~{}~{}-\psi_{0}(m+b+1)-\psi_{0}(a+1)+\psi_{0}(b+1))$ (300) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(a+b+i)}{i}=$ $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(b+i)}{i}-\sum_{i=1}^{a}\frac{\psi_{0}(b+i+m)}{b+i-1}+\frac{1}{2}\big{(}(\psi_{0}(a+b)$ $\displaystyle-\psi_{0}(b))\times(\psi_{0}(a+b)+\psi_{0}(b)+2(\psi_{0}(m+1)-\psi_{0}(1)))$ $\displaystyle-\psi_{1}(a+b)+\psi_{1}(b)\big{)}$ (301) $\displaystyle\sum_{i=1}^{m}\left(\frac{\psi_{0}(a+b+i+m)}{b+i}+\frac{\psi_{0}(a+i)}{b+i}+\frac{\psi_{0}(a+b+2i)}{a+i}-\frac{\psi_{0}(a+b+i)}{a+i}\right.$ $\displaystyle-\\!\left.\frac{\psi_{0}(a+b+2i)}{b+i}\right)$ $\displaystyle=\frac{\psi_{0}\left(\frac{a}{2}+\frac{b}{2}+m\right)}{b-a}-\frac{(a+b+m)\psi_{0}(a+b+m)}{b(a+m)}+\frac{\psi_{0}(a+b+2m)}{a+m}$ $\displaystyle~{}~{}~{}\\!~{}+\psi_{0}(a+m)\left(\psi_{0}(b+m+1)-\psi_{0}(b)-\frac{1}{b-a}\right)+\frac{a\psi_{0}(a)}{b(b-a)}+\frac{\psi_{0}(a+b)}{b}$ $\displaystyle~{}~{}~{}~{}\\!-\frac{\psi_{0}\left(\frac{a}{2}+\frac{b}{2}\right)}{b-a}$ (302) $\displaystyle\sum_{i=1}^{m}\left(\frac{\psi_{0}(a+b+i+m)}{a+i}+\frac{\psi_{0}(a+b+i+m)}{b+i}-\frac{\psi_{0}(a+b+i)}{a+i}\right.$ $\displaystyle-\\!\left.\frac{\psi_{0}(a+b+i)}{b+i}\right)$ $\displaystyle=\frac{(a+b+2m)\psi_{0}(a+b+2m)}{(a+m)(b+m)}-\left(\frac{1}{a+m}+\frac{1}{a}+\frac{1}{b+m}+\frac{1}{b}\right)\psi_{0}(a+b+m)$ $\displaystyle~{}~{}~{}~{}+\left(\frac{1}{a}+\frac{1}{b}\right)\psi_{0}(a+b)+\left(\psi_{0}(a+m)-\psi_{0}(a)\right)\left(\psi_{0}(b+m)-\psi_{0}(b)\right)$ (303) $\displaystyle\sum_{i=1}^{m}\frac{(n-i)!}{(m-i)!}=\frac{n!}{(m-1)!(n-m+1)}$ (304) $\displaystyle\sum_{i=1}^{m}\frac{(n-i)!}{(m-i)!i}=\frac{n!}{m!}(\psi_{0}(n+1)-\psi_{0}(n-m+1))$ (305) $\displaystyle\sum_{i=1}^{m}\frac{(n-i)!}{(m-i)!i^{2}}=$ $\displaystyle\frac{n!}{m!}\Bigg{(}\sum_{i=1}^{m}\frac{\psi_{0}(i+n-m)}{i}+\frac{1}{2}\big{(}\psi_{1}(n-m+1)-\psi_{1}(n+1)$ $\displaystyle+\psi_{0}(n-m+1){}^{2}-\psi_{0}(n+1){}^{2}\big{)}+\psi_{0}(n-m)$ $\displaystyle\times\left(-\psi_{0}(n-m+1)+\psi_{0}(n+1)-\psi_{0}(m+1)+\psi_{0}(1)\right)\\!\Bigg{)}$ (306) $\displaystyle\sum_{i=1}^{m}\frac{(n-i)!}{(m+a-i)!}=\frac{1}{n-m-a+1}\left(\frac{n!}{(a+m-1)!}-\frac{(n-m)!}{(a-1)!}\right)$ (307) $\displaystyle\sum_{i=1}^{m}\frac{(n-i)!}{(m+a-i)!}\psi_{0}(m+a-i+1)$ $\displaystyle=\frac{1}{1-a-m+n}\left(\frac{n!}{(a+m-1)!}\left(\psi_{0}(a+m)-\frac{1}{1-a-m+n}\right)\right.$ $\displaystyle~{}~{}~{}\\!~{}-\\!\left.\frac{(n-m)!}{(a-1)!}\left(\psi_{0}(a)-\frac{1}{1-a-m+n}\right)\right)$ (308) The derivation of the identities (287)–(B.1) and (304)–(B.1) can be found in Wei17 ; Wei20 ; Wei20BH ; HWC21 ; HW22 ; Milgram . In the same manner as obtaining (B.1) in Section 2.2.1, the three identities (B.1)–(B.1) are derived by first rewriting respectively the summations $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(a+b+i)}{i}=\sum_{i=1}^{m}\frac{\psi_{0}(b+i)}{i}+\sum_{j=1}^{a}\sum_{i=1}^{m}\frac{1}{i}\frac{1}{b+i+j-1}$ (309) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(a+b+2i)}{a+i}=\sum_{i=1}^{m}\frac{\psi_{0}(a+b+i)}{a+i}+\sum_{j=1}^{m}\sum_{i=j}^{m}\frac{1}{(a+i)(a+b+i+j-1)}$ (310) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(a+b+i+m)}{b+i}=\sum_{i=1}^{m}\frac{\psi_{0}(a+b+i)}{b+i}+\sum_{j=1}^{m}\sum_{i=1}^{m}\frac{1}{(b+i)(a+b+i+j-1)}.$ (311) For the identity (307), it is obtained by first considering $\displaystyle\sum_{i=1}^{m}\frac{(n-i)!}{(m+a-i)!}=\sum_{i=1}^{a+m}\frac{(n-i)!}{(m+a-i)!}-\sum_{i=1}^{a}\frac{(n-m-i)!}{(a-i)!}$ (312) before applying (304). Note that the identity (307) is analytically continued to any complex number $a$, where, by taking an derivative of $a$, the identity (B.1) is established. ### B.2 Additional summation identities of the new simplification framework Here, we list the additional summation identities of the new simplification framework useful in the simplification process in Section 2.3. These identities are obtained by taking appropriate derivatives of the formulas in lemmas 1–4. It is sufficient to assume that $a,b,c\geq 0$ in (B.2)–(B.2). $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}^{2}(a+i)-\psi_{1}(a+i)}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)}$ $\displaystyle=\frac{\Gamma(a+b+2m-1)}{\Gamma(m)\Gamma(a+m)\Gamma(b+m)\Gamma(a+b+m)}\big{(}\psi_{1}(a+b+2m-1)-\psi_{1}(a+b+m)$ $\displaystyle~{}~{}\\!~{}~{}-\psi_{1}(a+m)+\left(\psi_{0}(a+b+m)-\psi_{0}(a+b+2m-1)+\psi_{0}(a+m)\right){}^{2}\big{)}$ (313) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(i)}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)}$ $\displaystyle=-\frac{1}{\Gamma(m)\Gamma(b+m)\Gamma(a+b+m)}\sum_{i=1}^{m-1}\frac{\Gamma(a+b-i+2m-1)}{\Gamma(a-i+m)i}$ $\displaystyle~{}~{}~{}\\!~{}+\frac{\psi_{0}(m)\Gamma(a+b+2m-1)}{\Gamma(m)\Gamma(a+m)\Gamma(b+m)\Gamma(a+b+m)}$ (314) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(i)\psi_{0}(b-i+m+1)}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)}$ $\displaystyle=\frac{1}{\Gamma(m)\Gamma(b+m)\Gamma(a+b+m)}\sum_{i=1}^{m-1}\frac{\Gamma(a+b-i+2m-1)}{\Gamma(a-i+m)i}(-\psi_{0}(a+b+m)$ $\displaystyle~{}~{}~{}\\!~{}+\psi_{0}(a+b-i+2m-1)-\psi_{0}(b+m))+\frac{\Gamma(a+b+2m-1)\psi_{0}(m)}{\Gamma(m)\Gamma(a+m)\Gamma(b+m)}$ $\displaystyle~{}~{}\\!~{}~{}\times\frac{1}{\Gamma(a+b+m)}(\psi_{0}(a+b+m)-\psi_{0}(a+b+2m-1)+\psi_{0}(b+m))$ (315) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(i)\psi_{0}(a+i)}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)}$ $\displaystyle=\frac{1}{\Gamma(m)\Gamma(b+m)\Gamma(a+b+m)}\sum_{i=1}^{m-1}\frac{\Gamma(a+b-i+2m-1)}{\Gamma(a-i+m)i}(-\psi_{0}(a+b+m)$ $\displaystyle~{}~{}\\!~{}~{}+\psi_{0}(a+b-i+2m-1)-\psi_{0}(a-i+m))+\frac{\psi_{0}(m)\Gamma(a+b+2m-1)}{\Gamma(m)\Gamma(b+m)\Gamma(a+b+m)}$ $\displaystyle~{}~{}\\!~{}~{}\times\frac{\psi_{0}(a+b+m)-\psi_{0}(a+b+2m-1)+\psi_{0}(a+m)}{\Gamma(a+m)}$ (316) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}^{2}(i)-\psi_{1}(i)}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)}$ $\displaystyle=\frac{2}{\Gamma(m)\Gamma(b+m)\Gamma(a+b+m)}\sum_{i=1}^{m-1}\frac{\Gamma(a+b-i+2m-1)}{\Gamma(a-i+m)i}(\psi_{0}(i)-\psi_{0}(m)$ $\displaystyle~{}~{}~{}\\!~{}-\psi_{0}(1))+\frac{\Gamma(a+b+2m-1)}{\Gamma(m)\Gamma(a+m)\Gamma(b+m)\Gamma(a+b+m)}\left(\psi_{0}^{2}(m)-\psi_{1}(m)\right)$ (317) $\displaystyle\sum_{i=1}^{m}\frac{\psi_{0}(i)\psi_{0}(m-i+1)}{\Gamma(i)\Gamma(a+i)\Gamma(m-i+1)\Gamma(b-i+m+1)}$ $\displaystyle=\frac{1}{\Gamma(m)\Gamma(a+b+m)}\left(-\frac{1}{\Gamma(a+m)}\sum_{i=1}^{m-1}\frac{\Gamma(a+b-i+2m-1)}{\Gamma(b-i+m)i^{2}}-\frac{1}{\Gamma(b+m)}\right.$ $\displaystyle~{}~{}\\!~{}~{}\times\sum_{i=1}^{m-1}\frac{\Gamma(a+b-i+2m-1)}{\Gamma(a-i+m)i^{2}}-\frac{\psi_{0}(m)}{\Gamma(a+m)}\sum_{i=1}^{m-1}\frac{\Gamma(a+b-i+2m-1)}{\Gamma(b-i+m)i}$ $\displaystyle~{}~{}\\!~{}~{}-\left.\\!\frac{\psi_{0}(m)}{\Gamma(b+m)}\sum_{i=1}^{m-1}\frac{\Gamma(a+b-i+2m-1)}{\Gamma(a-i+m)i}\right)+\frac{\Gamma(a+b+2m-1)}{\Gamma(m)\Gamma(a+m)\Gamma(b+m)}$ $\displaystyle~{}~{}\\!~{}~{}\times\frac{1}{\Gamma(a+b+m)}\left(\psi_{0}^{2}(m)-\psi_{1}(m)+\psi_{1}(1)\right)$ (318) $\displaystyle\sum_{i=1}^{m}\frac{\Gamma(c-i+m)\Gamma(a+b+i+m)}{\Gamma(i)\Gamma(m-i+1)}\psi_{0}(i)$ $\displaystyle=-\frac{\Gamma(b+1)\Gamma(c)\Gamma(c+m)}{\Gamma(m)\Gamma(b+c+1)}\sum_{i=1}^{m-1}\frac{\Gamma(b+c-i+m)}{\Gamma(c-i+m)i}+\frac{\Gamma(b+1)\Gamma(c)\Gamma(b+c+m)}{\Gamma(m)\Gamma(b+c+1)}$ $\displaystyle~{}~{}~{}\\!~{}\times\left(\psi_{0}(b+c+m)-\psi_{0}(b+c+1)+\psi_{0}(m)\right)$ (319) $\displaystyle\sum_{i=1}^{m}\frac{\Gamma(c-i+m)\Gamma(a+b+i+m)}{\Gamma(a+i)\Gamma(m-i+1)}\left(\psi_{0}^{2}(i)-\psi_{1}(i)\right)$ $\displaystyle=\frac{2\Gamma(b+1)\Gamma(c)\Gamma(c+m)}{\Gamma(m)\Gamma(b+c+1)}\left(-\sum_{i=1}^{m-1}\frac{\Gamma(b+c-i+m)\psi_{0}(b+c-i+m)}{\Gamma(c-i+m)i}\right.$ $\displaystyle~{}~{}~{}\\!~{}+\sum_{i=1}^{m-1}\frac{\Gamma(b+c-i+m)\psi_{0}(i)}{\Gamma(c-i+m)i}-\left(-\psi_{0}(b+c+1)+\psi_{0}(m)+\psi_{0}(1)\right)$ $\displaystyle~{}~{}~{}\\!~{}\times\\!\left.\sum_{i=1}^{m-1}\frac{\Gamma(b+c-i+m)}{\Gamma(c-i+m)i}\right)+\frac{\Gamma(b+1)\Gamma(c)\Gamma(b+c+m)}{\Gamma(m)\Gamma(b+c+1)}\big{(}\psi_{1}(b+c+m)$ $\displaystyle~{}~{}~{}\\!~{}-\psi_{1}(b+c+1)-\psi_{1}(m)+\left(\psi_{0}(b+c+m)-\psi_{0}(b+c+1)+\psi_{0}(m)\right){}^{2}\big{)}$ (320) $\displaystyle\sum_{i=1}^{m}\frac{\Gamma(c-i+m)\Gamma(a+b+i+m)}{\Gamma(a+i)\Gamma(m-i+1)i}\psi_{0}(a+b+i+m)$ $\displaystyle=\Gamma(c+m)\Gamma(a+b+m)\Bigg{(}\\!\frac{\psi_{0}(a+b+m)}{\Gamma(a)\Gamma(m+1)}\left(\psi_{0}(c+m)-\psi_{0}(c)\right)+\Gamma(b+m+1)$ $\displaystyle~{}~{}~{}\\!~{}\times\Gamma(c)\sum_{i=1}^{m}\frac{\psi_{0}(a+b+m)-\psi_{0}(b-i+m+1)+\psi_{0}(b+m+1)}{\Gamma(a+i)\Gamma(c+i)\Gamma(m-i+1)\Gamma(b-i+m+1)i}\Bigg{)}$ (321) ## References * (1) Abramowitz, M., Stegun, I.A.: Handbook of Mathematical Functions with Formulas, Graphs, and Mathematical Tables. Dover, New York (1972) * (2) Bianchi, E., Hackl, L., Kieburg, M.: The Page curve for fermionic Gaussian states. Phys. Rev. B 103, L241118 (2021) * (3) Bianchi, E., Hackl, L., Kieburg, M., Rigol, M., Vidmar, L.: Volume-law entanglement entropy of typical pure quantum states. PRX Quantum 3(3), 030201 (2022) * (4) Bengtsson, I., Życzkowski, K.: Geometry of Quantum States: an Introduction to Quantum Entanglement. Cambridge University Press, Cambridge (2006) * (5) Bernard, D., Piroli, L.: Entanglement distribution in the quantum symmetric simple exclusion process. Phys. Rev. E 104, 014146 (2021) * (6) Bertola, M., Gekhtman, M., Szmigielski, J.: The Cauchy two-matrix model. Commun. Math. Phys. 287, 983–1014 (2009) * (7) Bertola, M., Gekhtman, M., Szmigielski, J.: Cauchy biorthogonal polynomials. J. Approx. Theory 162, 832-867 (2010) * (8) Bertola, M., Gekhtman, M., Szmigielski, J.: Cauchy-Laguerre two-matrix model and the Meijer-G random point field. Commun. Math. Phys. 326, 111-144 (2014) * (9) Borot, G., Nadal, C.: Purity distribution for generalized random Bures mixed states. J. Phys. A: Math. Theor. 45, 075209 (2012) * (10) Brychkov, Y.A.: Handbook of Special Functions: Derivatives, Integrals, Series and Other Formulas. CRC Press, Boca Raton (2008) * (11) Erdélyi, A., Magnus, W., Oberhettinger, F., Tricomi, F.G.: Tables of Integral Transforms. Vol. II. McGraw-Hill, New York (1954) * (12) Foong, S.K., Kanno, S.: Proof of Page’s conjecture on the average entropy of a subsystem. Phys. Rev. Lett. 72, 1148-1151 (1994) * (13) Forrester, P.: Log-gases and Random Matrices. Princeton University Press, Princeton (2010) * (14) Forrester, P., Kieburg, M.: Relating the Bures measure to the Cauchy two-matrix model. Commun. Math. Phys. 342, 151-187 (2016) * (15) Giraud, O.: Distribution of bipartite entanglement for random pure states. J. Phys. A: Math. Theor. 40, 2793 (2007) * (16) Hayden, P., Leung, D.W., Winter, A.: Aspects of generic entanglement. Commun. Math. Phys. 265, 95-117 (2006) * (17) Huang, Y., Wei, L., Collaku, B.: Kurtosis of von Neumann entanglement entropy. J. Phys. A: Math. Theor. 54, 504003 (2021) * (18) Huang, Y., Wei, L.: Second-order statistics of fermionic Gaussian states. J. Phys. A: Math. Theor. 55, 105201 (2022) * (19) Kieburg, M., Forrester, P., Ipsen, J.R.: Multiplicative convolution of real asymmetric and real anti-symmetric matrices. Adv. Pure Appl. Math. 10, 467 (2019) * (20) Li, S.-H., Wei, L.: Moments of quantum purity and biorthogonal polynomial recurrence. J. Phys. A: Math. Theor. 54, 445204 (2021) * (21) Lubkin, E.: Entropy of an n-system from its correlation with a k-reservoir. J. Maths. Phys. 19, 1028 (1978) * (22) Luke, Y.L.: The Special Functions and Their Approximations, vol.1. Academic, New York (1969) * (23) Lydżba, P., Rigol, M., Vidmar, L.: Entanglement in many-body eigenstates of quantum-chaotic quadratic Hamiltonians. Phys. Rev. Lett. 125, 180604 (2020) * (24) Mehta, M.L.: Random Matrices. Elsevier, Amsterdam (2004) * (25) Malacarne, L.C., Mendes, R.S., Lenzi, E.K.: Average entropy of a subsystem from its average Tsallis entropy. Phys. Rev. E 65, 046131 (2002) * (26) Milgram, M.: On some sums of digamma and polygamma functions. Preprint (2017). arXiv:0406338v3 * (27) Onuma-Kalu, M., Grimmer, D., Mann, R.B., Martín-Martínez, E.: A classification of Markovian fermionic Gaussian master equations. J. Phys. A: Math. Theor. 52, 435302 (2019) * (28) Osipov, V., Sommers H.-J., Życzkowski, K.: Random Bures mixed states and the distribution of their purity. J. Phys. A: Math. Theor. 43, 055302 (2010) * (29) Page, D.N.: Average entropy of a subsystem. Phys. Rev. Lett. 71(9), 1291-1294 (1993) * (30) Sánchez-Ruiz, J.: Simple proof of Page’s conjecture on the average entropy of a subsystem. Phys. Rev. E 52, 5653-5655 (1995) * (31) Sarkar, A., Kumar, S.: Bures-Hall ensemble: spectral densities and average entropies. J. Phys. A: Math. Theor. 52, 295203 (2019) * (32) Sommers, H.-J., Życzkowski, K.: Statistical properties of random density matrices. J. Phys. A: Math. Gen. 37, 35 (2004) * (33) Surace, J., Tagliacozzo, L.: Fermionic Gaussian states: an introduction to numerical approaches. SciPost Phys. Lect. Notes, 054 (2022) * (34) Szegő,G.: Orthogonal Polynomials. American Mathematical Society, Provindence (1975) * (35) Vivo, P., Pato, M.P., Oshanin, G.: Random pure states: Quantifying bipartite entanglement beyond the linear statistics. Phys. Rev. E 83, 052106 (2016) * (36) Wei, L.: Proof of Vivo-Pato-Oshanin’s conjecture on the fluctuation of von Neumann entropy. Phys. Rev. E 96, 022106 (2017) * (37) Wei, L.: On the exact variance of Tsallis entanglement entropy in a random pure state. Entropy 21, 539 (2019) * (38) Wei, L.: Skewness of von Neumann entanglement entropy. J. Phys. A: Math. Theor. 53, 075302 (2020) * (39) Wei, L.: Proof of Sarkar-Kumar conjectures on average entanglement entropies over the Bures-Hall ensemble. J. Phys. A: Math. Theor. 53, 235203 (2020) * (40) Wei, L.: Exact variance of von Neumann entanglement entropy over the Bures-Hall measure. Phys. Rev. E 102, 062128 (2020) * (41) Wei, L.: Average capacity of quantum entanglement. Preprint (2022). arXiv:2205.06343
# Extracting Semantic Knowledge from GANs with Unsupervised Learning Jianjin Xu, Zhaoxiang Zhang, Xiaolin Hu J. Xu is with the School of Mathematics and Computer Science, Panzhihua University and the Department of Computer Science and Technology, Tsinghua University. E-mail<EMAIL_ADDRESS>Z. Zhang is with with the Insitute of Automation, Chinese Academy of Sciences and the Centre for Artificial Intelligence and Robotics, Hong Kong Innovation and Science Insitute, Chinese Academy of Sciences. E-mail<EMAIL_ADDRESS>X. Hu is with the Department of Computer Science and Technology, Institute for Artificial Intelligence, State Key Laboratory of Intelligent Technology and Systems, THU-Bosch JCML Center, THBI, IDG/McGovern Institute for Brain Research, BNRist, Tsinghua University, Beijing, China. He is also with the Chinese Institute for Brain Research (CIBR), Beijing, China. E-mail<EMAIL_ADDRESS>work was supported in part by the National Natural Science Foundation of China (Nos. 62061136001, 61836014, and U19B2034) and THU-Bosch JCML center. Corresponding author: Xiaolin Hu. ###### Abstract Recently, unsupervised learning has made impressive progress on various tasks. Despite the dominance of discriminative models, increasing attention is drawn to representations learned by generative models and in particular, Generative Adversarial Networks (GANs). Previous works on the interpretation of GANs reveal that GANs encode semantics in feature maps in a linearly separable form. In this work, we further find that GAN’s features can be well clustered with the linear separability assumption. We propose a novel clustering algorithm, named KLiSH, which leverages the linear separability to cluster GAN’s features. KLiSH succeeds in extracting fine-grained semantics of GANs trained on datasets of various objects, _e.g_., car, portrait, animals, and so on. With KLiSH, we can sample images from GANs along with their segmentation masks and synthesize paired image-segmentation datasets. Using the synthesized datasets, we enable two downstream applications. First, we train semantic segmentation networks on these datasets and test them on real images, realizing unsupervised semantic segmentation. Second, we train image-to-image translation networks on the synthesized datasets, enabling semantic- conditional image synthesis without human annotations. ###### Index Terms: GAN, Unsupervised Learning, Semantic Segmentation, Conditional Image Synthesis ## 1 Introduction Representation is at the core of modern deep learning practice. Bengio _et al_. [1] propose that a good representation should have (1) _explanatory factors organized in hierarchy_ , and (2) _natural clustering of categories_. Chan _et al_. [2] claim that the objective of a deep network is to learn a linearly discriminative representation of the data. Researchers have found that the representation learned by GANs [3] not only explains the variation of images [4, 5, 6] but also is linearly discriminative [7], indicating that GANs learn a good representation. Given a dataset of images, a GAN learns to model the distribution of data and synthesizes images unseen in the dataset. The image synthesis process of GAN is typically a mapping from noise vectors to a series of intermediate feature maps and then to RGB images. From the perspective of causality, GAN’s representations contain all the information in the image. In other words, the variation factors (_e.g_., the attributes of objects) present in the image must also exist in GAN’s representation. Previous works have identified two types of variations captured by GANs. First, the image-level attributes (_e.g_., the gender of face images) are separable by linear hyperplanes in GAN’s latent space [4, 8, 5, 9, 10]. Second, the object-level semantics (_e.g_., the presence and location of a window in bedroom images) are encoded in GAN’s feature maps [6] in a linearly separable form [7]. We term the rich semantic information in GAN’s features as the semantic knowledge learned by the GANs. Figure 1: Semantic segmentation obtained by using the average feature of each class compared to using a trained linear classifier. The prediction from a pretrained DeepLabV3 [11] is regarded as the groundtruth. The linear classifier is supervised by a DeepLabV3 network pretrained on CelebAMask-HQ dataset [12] according to Xu _et al_. [7]. Existing works [13, 14, 15] often train few-shot learning classifiers to reveal GAN’s semantic knowledge. The classifiers predict each spatial location of GAN’s feature maps. Training the classifiers needs only a few labeled examples. The trained classifier is subsequently used for image editing [13, 14] and semantic segmentation [15]. However, the labeling cost for the few- shot learning method is still high. As reported in [13, 14], the number of annotations ranges from 16 to 40 and labeling each image takes around 10 minutes even for a skillful user. Therefore, if one can design an unsupervised segmentation algorithm based on GAN’s feature maps for image editing, it would enable users to upload their own images, train their own GANs, and edit those images without much labor cost. This is the aim of this study. Specifically, we study how to reveal GAN’s semantic knowledge with unsupervised learning methods. Existing unsupervised learning methods mostly apply conventional clustering algorithms on GAN’s features, _e.g_., K-means [16, 17] and Agglomerative Hierarchical Clustering (AHC) [18]. They use clustering results as an intermediate step for editing the attribute [17] or swapping the style [16] in a localized region. However, the obtained clusters are not satisfactory. Please refer to their results [16, 17, 18] and also our reproduction in Fig. 5. For example, for face images, we observed the oversegmentation of the hair class, and the missing of tiny classes like brow, upper lip, and lower lip (see Fig. 5, the last column). Figure 2: An illustration of clustering on linearly separable data. Top-left, the groundtruth cluster assignments, and the decision boundaries of the linear SVM trained on the groundtruth. Other panels, results of different clustering algorithms. Decision boundaries are visualized as solid lines, except for AHC which has no such boundary. “x” denotes the centroids of K-means. GAN | StyleGAN2 | StyleGAN | PGGAN ---|---|---|--- Dataset | FFHQ | CelebAHQ | CelebAHQ Linear Classifier [7] | 82.7 | 73.8 | 69.3 Centroids of Features | 49.6 | 46.6 | 41.8 TABLE I: The semantics segmentation performance (measured in mIoU) of linear classifier or centroids of features. Please see Sec. 4.2 for details. The reason might be that K-means and AHC require data to be centered around their class centroids, which is not well satisfied in GAN’s features. To illustrate this, we first took a pretrained GAN and segmented the generated images with a pretrained DeepLabV3 [11] model. The features from the GAN were averaged into a centroid for each semantic class. Then, we segmented another set of generated images by assigning the closest centroid to GAN’s features on each spatial location. It was observed that the segmentation results were rather coarse (Fig. 1). Interestingly, a previous work [7] found that GAN’s features of a semantic class were linearly separable from other classes. This observation is referred to as the GAN’s linear separability in semantics. We also found that the linear classifier produced better segmentation than using the centroid method (Tab. I). Therefore, we propose to cluster GAN’s features by leveraging their linear separability property, instead of the conventional centering assumption. We note that linearly separable data does not necessarily satisfy the centering assumption required by conventional clustering algorithms. To illustrate this, we generated three sets of points in a 2D plane, which were linearly separable from each other (Fig. 2, top-left). In this synthetic dataset, each cluster had a complex structure and its centroid had limited information about the structure. It was observed that both AHC and K-means assigned a part of the blue cluster to the red cluster. We propose K-means with Linear Separability Heuristic (KLiSH) to find semantically meaningful clusters in GAN’s features. In brief, KLiSH starts with a large number of initial K-means clusters and iteratively merges the clusters according to their degree of linear separability. KLiSH successfully clustered the three sets of points on the toy dataset described above (Fig. 2, bottom-right). We compared KLiSH to several clustering algorithms, AHC, K-means, and K-means based Approximate SPectral clustering (KASP) [19] on several GANs and datasets. Results showed that KLiSH outperformed all baselines in most cases. One can use KLiSH to obtain fine-grained semantic segmentation for the generated images. The process is straightforward. First, sample images from GANs and apply the classifier returned by KLiSH on GAN’s feature maps. Then, train downstream models on the resulting image-segmentation dataset. We demonstrate two applications in this paper: Unsupervised Fine-Grained Segmentation (UFGS) and Unsupervised Semantic-Conditional Synthesis (USCS). UFGS seeks to reduce the labeling cost of fine-grained semantic annotation on a new dataset. USCS aims to offer semantic-controllable image synthesis and editing on datasets without semantic annotation. Both UFGS and USCS can be realized easily with the semantic knowledge extracted from pretrained GANs. Figure 3: The pipeline of re-purposing pretrained GANs for downstream tasks, UFGS and USCS. First, we propose the K-means with Linear Separability Heuristic (KLiSH) to cluster GAN’s features. Then, we generate datasets with synthesized images and segmentation masks together from GANs. Finally, we train different models on the synthesized datasets and perform semantic segmentation on _real_ images and semantic-conditional image synthesis. Both applications require no human annotations at all. The pipeline of UFGS and USCS is illustrated in Fig. 3. Given a dataset of images, we first train a GAN on it and use KLiSH to parse the semantic clusters learned by GANs. Then, we generate a paired image-segmentation dataset by sampling images from GANs along with the segmentation masks. Finally, we train corresponding downstream models on the generated dataset. In summary, our contributions are as follows. First, we propose KLiSH and improve the quality of clustering on GAN’s features. Second, we provide evidence for GAN’s linear separability in semantics. Third, we propose two novel applications, UFGS and USCS, marking a successful attempt at fine- grained semantic segmentation and semantic-conditional synthesis in an unsupervised learning setting. ## 2 Related work #### Generative Adversarial Networks Originally put forward by Goodfellow _et al_. [3], GANs go through many improvements in training [20, 21], architecture [22, 23, 24, 25] and objective [26, 27, 28, 29]. Nowadays, GANs can map latent vectors to diverse and photo- realistic images on various datasets. Typical data domains are face [30, 31, 32], car [33, 34, 20], animal [35, 36], ImageNet [37, 38], _etc_.. Among various GANs, StyleGAN2 [34] is one of the state-of-the-art models. Another genre of GANs translates images to images [39, 40, 41]. When the input image is a semantic mask, the task is termed Semantic-Conditional Synthesis (SCS). Typical models capable of SCS include pix2pix [42], pix2pixHD [43], GauGAN [44] and so on [45, 12, 46, 47]. Most of them heavily rely on large-scale semantically-annotated datasets. Sun _et al_. [48] propose to synthesize objects in an image from bounding boxes, which lowers the label requirement, but still needs a large-scale dataset. In contrast, the USCS proposed in this work is fully unsupervised – it is realized by making use of the representations learned by pretrained GANs. #### GAN interpretation Increasing attention is placed on studying GAN’s representation. Most existing works focus on studying GAN’s latent vectors and feature maps. For the first aspect, prior works [4, 5] find that there are linear boundaries in latent space that separate the positive and negative attributes of the sampled images. Some other works focus on the unsupervised discovery of those boundaries [8, 10, 9]. For the second aspect, GAN Dissection [6] identifies convolution units that have causal relationships with semantics in the generated images. Xu and Zheng [7] propose and validate that GANs encode high- resolution semantics linearly in feature maps. Other works study the parameters of GANs [49, 50]. This work proposes an unsupervised approach to extract learned semantic classes from GANs, providing strong evidence that the GAN learns a naturally clustered and linearly separable semantic representation. #### Repurposing GAN for semantic segmentation Recently, increasing attention is directed towards semantic segmentation using GAN’s representation. Existing works either fall into few-shot learning or unsupervised learning. The few-shot learning methods typically apply classifiers on a pretrained GAN’s feature maps to get semantic segmentation for the generated image [7, 51, 15, 14]. These methods can segment the images well with a limited number of image annotations (ranging from 1 to 40 depending on the complexity of the target data). For unsupervised learning, the majority of methods [52, 53, 54, 55, 56, 57] focus on decoupling the foreground and background in GANs without external labeling. Though the segmentation is accurate, the practical usage is limited as only two clusters are segmented. Methods on obtaining more clusters mostly involve applying K-means++ [16] or AHC [18, 17] directly on GAN’s features. He _et al_. [58] modify K-means by re-weighting channels. Different from existing methods, we propose a new clustering algorithm, KLiSH, by leveraging a strong characteristic of GAN’s features – linear separability. We conducted quantitative evaluations and showed the superiority of KLiSH to previous clustering algorithms on GAN’s features. Aside from GAN, other works [59, 60, 61] also study unsupervised semantic segmentation on large-scale datasets, but mainly focus on object-level semantics instead of part-level semantics and their results are not directly comparable to GAN-based methods. ## 3 Clustering Generator’s Features ### 3.1 Linear Separability of GAN’s Features #### Classical linear separability KLiSH is based on GAN’s linear semantics [7], thus we first introduce this characteristic to make our paper self-contained. When the generator synthesizes an image, it also produces a set of feature maps. A $1\times 1$ convolutional layer is applied to the upsampled and concatenated feature maps, resulting in a score map of semantic classes for the generated image. As the extraction process is fully linear, the method is named Linear Semantic Extractor (LSE). The LSE is compared to two nonlinear semantic extractors (NSEs) with more complicated architectures. The LSE and NSEs are trained under similar settings and evaluated by the accuracy of the semantic masks they extracted compared to the pretrained segmentation network. Results show that the relative performance gaps between LSE and NSEs are within $3.5\%$ across various datasets and models. Therefore, it is shown that linear classifiers suffice to separate GAN’s representations of different semantics. For more detailed settings and results, we refer readers to the original paper [7]. The phenomenon that a linear classifier can classify a given dataset is usually described as linearly separable. Given a dataset $\mathcal{A}=\\{(\mathbf{x}_{i},y_{i})\\}_{i=1}^{N}$ where $\mathbf{x}_{i}\in\mathbb{R}^{D}$ and $y_{i}\in\\{1,\ldots,M\\}$, the classical linear separability states that there exists a set of hyperplanes $\mathbf{W}\in\mathbb{R}^{M\times D}$ and $\mathbf{b}\in\mathbb{R}^{D}$ such that $\forall(\mathbf{x},y)\in\mathcal{A},k\neq y,\mathbf{W}_{y}\mathbf{x}+\mathbf{b}_{y}>\mathbf{W}_{k}\mathbf{x}+\mathbf{b}_{k},$ (1) where $\mathbf{W}_{k}\in\mathbb{R}^{1\times D}$ is a row vector, and $\mathbf{W}_{y}$ denotes the classifier weight for the ground-truth class of a data point $\mathbf{x}$. #### Binary linear separability Instead of using classical linear separability, we find that binary linear separability is more useful for building our clustering algorithm, which is defined below. ###### Definition 1 Given a dataset $\mathcal{A}=\\{(\mathbf{x}_{i},y_{i})\\}_{i=1}^{N}$, we say $\mathcal{A}$ satisfies the binary linear separability if there exists $\mathbf{W}\in\mathbb{R}^{M\times D}$ and $\mathbf{b}\in\mathbb{R}^{D}$ such that $\forall(\mathbf{x}_{i},y_{i})\in\mathcal{A}$, $k\in\\{1,\dots,M\\}$, $\mathcal{I}(y_{i}=k)\cdot(\mathbf{W}_{k}\mathbf{x}_{i}+\mathbf{b}_{k})>1,$ (2) where $\mathcal{I}(y_{i}=k)$ returns 1 if $y_{i}=k$ and -1 otherwise. In other words, if for each class in the dataset, there is a two-class SVM separating itself from others, then the dataset satisfies the binary linearly separability. Unless otherwise specified, linear separability refers to Definition 1 throughout the rest of the paper. Clearly, Eq. 2 is harder to satisfy than Eq. 1. Clusters separable with a classical linear classifier might not be separable with binary SVMs. Do features in GAN satisfy the binary linear separability? We empirically found that the answer is yes. The experiment is presented in Sec. 4.2. ### 3.2 Clustering with Linear Separability Heuristic Now, we describe the algorithm for clustering GAN’s features by linear separability. #### Initialize from K-means We choose K-means for three reasons. First, K-means is easily scalable to large datasets with time complexity of $O(N)$. Second, the decision boundaries between clusters are linear and thus are suitable to implement linear separability. Third, K-means with a large number of clusters provides a good initialization for our algorithm. More specifically, most clusters found by K-means do not span across multiple semantic classes, see the K-means@100 column in Fig. 5. This simplifies our problem as we only need to merge these clusters to recover the ground-truth classes. #### Merge clusters Ideally, the merged clusters should belong to the same semantic class. However, we have no access to semantic information as the algorithm is fully unsupervised. We make a reasonable assumption: the more linearly separable a cluster is, the more likely it is a true semantic class. Under this assumption, our goal is to maximize the linear separability of clusters. How to quantify the degree of linear separability? According to Definition 1, it is implied that how accurately the SVM classifies a cluster indicates how well its linear separability is. Therefore, we use the performance of the SVM trained on the target dataset to indicate its linear separability. First, a pseudo-labeled dataset $\mathcal{A}$ with cluster assignments is constructed. Second, binary SVMs are trained on $\mathcal{A}$ using the L2-regularized L2-loss Support Vector Classification objective [62]: $\mathcal{L}=\frac{\lambda_{1}}{M|\mathcal{A}|}\sum_{(\mathbf{x},y)\in\mathcal{A}}\bigg{[}(1-\mathbf{W}_{y}\mathbf{x}-\mathbf{b}_{y})_{+}^{2}\\\ +\sum_{k\neq y}(1+\mathbf{W}_{k}\mathbf{x}+\mathbf{b}_{k})_{+}^{2}\bigg{]}+\frac{1}{2M}\left\lVert\mathbf{W}\right\rVert_{F}^{2},$ (3) where $(\cdot)_{+}$ denotes $max(0,\cdot)$ and $\left\lVert\cdot\right\rVert_{F}$ denotes the Frobenius norm. We use L-BFGS [63] to minimize $\mathcal{L}$ due to the convexity of the objective. The SVM is considered as converged if its $L_{\infty}$ norm of weight changing is below $10^{-4}$. Third, as in semantic segmentation, we measure the performance of SVM by IoU (Intersection-over-Union) defined as $\displaystyle\text{IoU}_{k}(\mathbf{W},\mathcal{A})=\frac{|\mathcal{S}_{k}\cap\mathcal{Y}_{k}|}{|\mathcal{S}_{k}\cup\mathcal{Y}_{k}|},$ (4) where $\mathcal{S}_{k}=\\{\mathbf{x}_{i}|\mathbf{W}_{k}\mathbf{x}_{i}+\mathbf{b}_{k}>0\\}$ is the prediction given by the SVM, $\mathcal{Y}_{k}=\\{\mathbf{x}_{i}|y_{i}=k\\}$ is the set of positive samples for class $k$, and $(\mathbf{x}_{i},y_{i})\in\mathcal{A}$. Therefore, we can use IoU to find the cluster with the lowest linear separability. Suppose that the identified cluster is $p$. The next step is to find another cluster $q$ to merge with $p$. As implied by Eq. 4, the cluster $p$ is not classified accurately by the SVM, _i.e_., the SVM might confuse $p$ with other clusters. Then a reasonable strategy is to merge $p$ with another cluster that confuses it the most. To quantify the degree of confusion, we propose the Effective Cosine Similarity (ECoS) metric as defined below. ###### Definition 2 Consider a dataset $\mathcal{A}=\\{(\mathbf{x}_{i},y_{i})\\}_{i=1}^{N}$. Suppose a set of binary SVMs with weight $\mathbf{W}\in\mathbb{R}^{M\times D}$ and bias $\mathbf{b}\in\mathbb{R}^{D}$ are trained on $\mathcal{A}$ until convergence. The ECoS between cluster $i$ and cluster $j$ is defined as $\begin{split}&\text{ECoS}_{i,j}(\mathbf{W},\mathcal{A})=\cos\langle\mathbf{s}_{i}^{\dagger},\mathbf{s}_{j}^{\dagger}\rangle\\\ &\mathbf{s}_{k}^{\dagger}=\text{CLAMP}_{[0,1]}\left(\frac{\mathbf{s}_{k}+1}{2}\right)\\\ &\mathbf{s}_{k}=\left[\mathbf{x}_{1}^{T}\mathbf{W}_{k}+\mathbf{b}_{k},\ldots,\mathbf{x}_{N}^{T}\mathbf{W}_{k}+\mathbf{b}_{k}\right]\end{split}$ (5) where $\text{CLAMP}_{[a,b]}$ denotes clamping the input within $[a,b]$. Here, $\mathbf{s}_{k}$ is the cluster scores predicted by the SVM. $\mathbf{s}_{k}^{\dagger}$ scales and truncates $\mathbf{s}_{k}$ within [0, 1], representing the confidence of SVM predicting positive data points. As each feature vector corresponds to a pixel, $\mathbf{s}_{k}^{\dagger}$ can be resized into a cluster confidence map. In other words, ECoS measures the cosine similarity between two cluster confidence maps. #### Filter initial clusters We observed in experiments that a small portion of initial clusters did span across multiple semantic classes. For example, see K-means@100 of StyleGAN2-FFHQ in Fig. 5, the boundary around ear (visualized as a yellow region) contains part of “ear”, “hair”, and “face”. This type of initial clusters would harm the accuracy of merged clusters. Fortunately, we empirically found that these clusters were not linearly separable, _i.e_., could be filtered by setting a threshold of IoU. The threshold could also be determined adaptively as described below. First, train SVMs on the initial clusters and calculate their IoU, resulting in a vector $\mathbf{m}$. Second, $\mathbf{m}$ is converted from [0, 1] to $(-\infty,\infty)$ by using the inverse sigmoid function $h^{-1}(\cdot)$. Third, any clusters below $\mu_{\mathbf{m}}-\sigma_{\mathbf{m}}$ are filtered, where $\mu$, $\sigma$ are the mean and standard deviation of $h^{-1}(\mathbf{m})$, respectively. Finally, K-means is run again with the filtered centroids as initialization. Input: $\mathcal{D}=\\{\mathbf{x}_{i}\\}_{i=1}^{N}$; $K_{0}$ Output: $\Theta=\\{(\mathbf{W}^{t},\mathbf{b}^{t})\\}_{t=1}^{K_{0}-1}$ $\mathbf{W}^{0},\mathbf{b}^{0},\mathcal{A}^{0}$ = K-means($\mathcal{D},K_{0}$) // Filter initial centriods $\mathbf{W}^{\prime},\mathbf{b}^{\prime}$ = TrainSVM($\mathbf{W}^{0},\mathbf{b}^{0},\mathcal{A}^{0}$) $\mathbf{m}=\left[h^{-1}(\text{IoU}_{p}(\mathbf{W}^{\prime},\mathbf{b}^{\prime},\mathcal{A}^{0}))\right]_{p=1}^{K_{0}}$ $\mathbf{W}^{0},\mathbf{b}^{0},\mathcal{A}^{0}\leftarrow\text{K-means}(\mathcal{D},\\{\mathbf{W}_{p}^{0}|\mathbf{m}_{p}>\mu_{\mathbf{m}}-\sigma_{\mathbf{m}}\\})$ for _$t=1,\ldots,K_{0}-1$_ do $\mathbf{W}^{t},\mathbf{b}^{t}$ = TrainSVM($\mathbf{W}^{t-1},\mathbf{b}^{t-1},\mathcal{A}^{t-1}$) $\Theta\leftarrow\Theta\cup\\{(\mathbf{W}^{t},\mathbf{b}^{t})\\}$ $p^{t},\phi^{t}=\min_{p}\text{IoU}_{p}(\mathbf{W}^{t},\mathcal{A}^{t-1})$ $q^{t},\psi^{t}=\max_{q}\text{ECoS}_{p^{t},q}(\mathbf{W}^{t},\mathcal{A}^{t-1})$ // Merge cluster $p^{t}$ into $q^{t}$ $\mathcal{A}^{t}=\\{(\mathbf{x}_{i},q^{t})|y_{i}=p^{t}\\}\cup\\{(\mathbf{x}_{i},y_{i})|y_{i}\neq p^{t}\\}$ $\mathbf{W}^{t}\leftarrow\text{cat}(\ldots,\mathbf{W}_{p^{t}-1}^{t},\mathbf{W}_{p^{t}+1}^{t},\ldots)$ $\mathbf{b}^{t}\leftarrow\text{cat}(\ldots,\mathbf{b}_{p^{t}-1}^{t},\mathbf{b}_{p^{t}+1}^{t},\ldots)$ Algorithm 1 The KLiSH clustering algorithm. #### KLiSH algorithm The full KLiSH algorithm is presented in Algorithm 1, where $\mathcal{D}$ is the dataset of GAN’s features, $K_{0}$ is the initial cluster number, and the output $\Theta=\\{(\mathbf{W}^{t},\mathbf{b},\phi^{t})\\}_{t=1}^{K_{0}-1}$ records the SVM and IoU at each merging step. The initialization stage has two steps. First, K-means is applied to the datasets. Second, filtering is performed as described above, resulting in the initial cluster assignments $\mathcal{A}^{0}$. Then, clusters are iteratively merged for $K_{0}-1$ steps. In step $t$, an SVM is first trained on dataset $\mathcal{A}^{t-1}$ , which is obtained from step $t-1$. Next, the IoU and ECoS are calculated as defined in Eq. 4 and Eq. 5. The class with the lowest IoU, denoted by $p^{t}$, will be merged with the class with the highest ECoS with it, denoted by $q^{t}$. The process is repeated until only one class is left. Without additional information, KLiSH cannot identify a proper number of classes, just like K-means. We can either slide through cluster numbers and select manually or set up a threshold $\phi^{*}$ and stop when the IoU reaches the threshold. ## 4 Evaluation of KLiSH ### 4.1 Experiment Setup #### Pretrained GANs We experimented on four widely used GANs, PGGAN [20], StyleGAN [33], StyleGAN2 [34], and StyleGAN2-ADA [35]. These GANs were pretrained on various datasets, including FFHQ [34], CelebAHQ [64], AFHQ [35], MetFace [35], and LSUN [65] Car, Bedroom and Church split. The pretrained models were obtained from the official release 111https://github.com/NVlabs/stylegan3. GANs | layer indices | channels ---|---|--- StyleGAN2-Bedroom | 9, 11, 13 | 896 StyleGAN2-Church PGGAN-[Bedroom, Church] | 7, 9, 11, 13 | 960 StyleGAN-[Bedroom, Church] StyleGAN2-Car | 9, 11, 13, 15 | 960 ADA-[Cat, Dog, Wildlife] PGGAN-CelebAHQ | 9, 11, 13, 15, 17 | 496 StyleGAN-CelebAHQ StyleGAN2-FFHQ | 9, 11, 13, 15, 17 | 992 ADA-MetFace TABLE II: The indices of GAN layers selected for clustering. #### GAN feature selection We only collected features from a subset of GAN’s layers mainly due to GPU memory constraints. To select the layers, we scanned through a generator’s layers from high resolution to low and stopped adding new layers if the accumulated number of channels exceeded 1,000. If multiple layers were having the same resolution, only the last layer of that resolution was selected. The selected layers for all GANs are shown in Tab. II. The feature maps from the selected layers are then bilinearly interpolated into the same resolution and concatenated along the channel axis. For clustering, we use $256\times 256$ resolution. For generating the segmentation (in UFGS and USCS), we use the full image resolution. #### Semantic segmentation networks For face images (CelebAHQ and FFHQ), we trained a DeepLabV3 model with ResNet50 [66] backbone on CelebAMask-HQ [12]. Following [7], we merged duplicate classes like “left eye” and “right eye”, resulting in 15 classes. The model was trained on the training split of CelebAMask-HQ for 20 epochs. We used Adam with learning rate $10^{-3}$, $\beta_{1}=0.9$, $\beta_{2}=0.999$ and batch size 32. The learning rate was lowered to $10^{-4}$ after 10 epochs. We found that 20 epochs were sufficient for the convergence of the DeepLabV3 model. For bedroom and church images, we used a DeepLabV3 model 222https://github.com/zhanghang1989/PyTorch-Encoding pretrained on ADE20K [67]. #### Baseline clustering algorithms We compared KLiSH with K-means, AHC, and KASP [19]. K-means and AHC have been used to cluster GAN’s features [16, 58, 18, 17], but they implicitly require the centroid assumption (described in Tab. I). Spectral clustering does not have such a constraint, however, it is not practical for our task as it has an $O(N^{3})$ time complexity [19]. Therefore, we consider using KASP, a fast approximate algorithm for spectral clustering. Interestingly, KASP also performs K-means first and then merges the initial clusters. Different from KLiSH, KASP uses spectral clustering on the initial centroids to merge them and does not use linear separability. We implemented a multi-GPU version of K-means++. KASP was implemented following the original paper [19]. For AHC, we made two modifications based on the scikit-learn implementation 333https://scikit- learn.org/stable/modules/generated/sklearn.cluster.AgglomerativeClustering.html. First, we trained a linear classifier on the cluster assignments obtained by AHC for prediction; otherwise, AHC would rebuild the whole merging tree to predict new data, which was impractical. Second, we gave AHC less number of features due to its expensive memory cost. More optimization of AHC might be possible, but would be beyond the scope of this work. #### Evaluation of clustering For each GAN, we sampled a total of 256 images, resulting in a block of features with shape $(256,256,256,D)$. The only exception is AHC, where we used 16 images and $64\times 64$ resolution, due to the expensive memory cost mentioned above. Then, we applied KLiSH, K-means, AHC, and KASP to the sampled features and evaluated them on the same set of 10k sampled images (different from those used for clustering). For evaluation, we consider two widely used metrics, Adjusted Mutual Information (AMI) and Adjusted Rand Index (ARI). Both AMI and ARI are designed for general-purpose clustering tasks and may not reflect the performance of segmentation tasks faithfully. In the context of (supervised) semantic segmentation, mIoU is the dominant metric. Therefore, we also propose the Maximum matching mIoU (MIoU) metric defined below, $\displaystyle\displaystyle\text{MIoU}(\\{\mathcal{Y}_{m}\\}_{m=1}^{M},\\{\mathcal{C}_{k}\\}_{k=1}^{K})$ $\displaystyle=\max_{\mathbf{a}\in\\{0,\ldots,M\\}^{K}}\frac{1}{M}J(\mathbf{a}),$ (6) $\displaystyle J(\mathbf{a};\\{\mathcal{Y}_{m}\\},\\{\mathcal{C}_{k}\\})$ $\displaystyle=\sum_{m=1}^{M}\text{IoU}(\mathcal{Y}_{m},\bigcup_{\\{k|\mathbf{a}_{k}=m\\}}\mathcal{C}_{k}),$ (7) where $M$ is the number of groundtruth classes, $K$ is the number of clusters, $\mathcal{Y}_{m}$ is the groundtruth class $m$, and $\mathcal{C}_{r}$ is the predicted cluster $r$. $\mathbf{a}$ is a $K$-dimensional vector that encodes the matched class for each cluster. If the $r$-th element of $\mathbf{a}$, $\mathbf{a}_{r}$ equals 0, then the cluster $r$ does not match any groundtruth class. Otherwise, $\mathbf{a}_{r}>0$ indicates cluster $r$ is matched to groundtruth class $\mathbf{a}_{r}$. In short, MIoU measures the segmentation performance of the predicted clusters after they have been matched correctly to the groundtruth classes. MIoU should only be compared when two clustering algorithms produce the same number of clusters, _i.e_., when the cost of label permutation is identical. This constraint is also required by other clustering metrics such as AMI and ARI. Figure 4: Demonstration of MIoU-permuted labels. The MIoU is calculated by a greedy algorithm as described in Algorithm 2, in which $(\mathbf{a}|_{\mathbf{a}_{k}\leftarrow m})$ denotes that the $k^{\text{th}}$ element of $\mathbf{a}$ is set to $m$. To demonstrate the effectiveness of MIoU, we tested Algorithm 2 on a few images. As shown in Fig. 4, the label permutations found by Algorithm 2 well matched the predicted clusters to the groundtruth classes. Input: $\\{\mathcal{Y}_{m}\\}_{m=1}^{M},\\{\mathcal{C}_{k}\\}_{k=1}^{K}$ Output: MIoU, $\mathbf{a}^{t}\in\\{0,1,\ldots,M\\}^{K}$ $\mathbf{a}^{0}=\mathbf{0}$ for _$t=1,\ldots,K$_ do $\mathcal{I}^{t}=\\{k|\mathbf{a}_{k}^{t-1}=0\\}\times\\{1,\ldots,M\\}$ $\mathcal{G}^{t}=\\{\mathbf{a}^{\prime}|\mathbf{a}^{\prime}=(\mathbf{a}^{t-1}|_{\mathbf{a}_{k}^{t-1}\leftarrow m}),(k,m)\in\mathcal{I}^{t}\\}$ $\mathbf{a}^{t}=\max_{\mathbf{a}^{\prime}\in\mathcal{G}^{t}}J(\mathbf{a}^{\prime};\\{\mathcal{Y}_{m}\\},\\{\mathcal{C}_{k}\\})$ MIoU = $J(\mathbf{a}^{t};\\{\mathcal{Y}_{m}\\},\\{\mathcal{C}_{k}\\})$ Algorithm 2 Maximum matching mIoU. All experiments were conducted using PyTorch [68]. KLiSH required 8 GeForce RTX 2080 Ti GPUs to run and took less than 30 minutes per GAN model. Figure 5: Results of K-means, AHC, KASP, and KLiSH on face images and car images. The first row shows the generated images and the other rows show the clustering results obtained with different cluster numbers, denoted by the number after “@”. “final” refers to the selected number of clusters for downstream tasks, which is 26, 11, 30, and 26 from left to right, respectively. GAN | StyleGAN2 | StyleGAN | PGGAN ---|---|---|--- Dataset | FFHQ | Bedroom | Church | CelebAHQ | Bedroom | Church | CelebAHQ | Bedroom | Church SVM | 81.7 | 43.5 | 33.7 | 72.2 | 30.9 | 31.4 | 66.4 | 27.7 | 45.4 LSE | 82.7 | 45.2 | 35.2 | 73.8 | 37.3 | 34.8 | 69.3 | 31.1 | 47.8 $\Delta\%$ | -1.3 | -3.8 | -4.3 | -2.2 | -9.8 | -17.2 | -4.2 | -10.9 | -5.0 TABLE III: The semantic extraction performance (in mIoU%) of SVM and LSE on various GANs and datasets. $\Delta$ denotes the relative difference, $\Delta=\frac{\text{SVM}-\text{LSE}}{\text{LSE}}$. ### 4.2 Binary Linear Separability Results As described in Sec. 3.1, KLiSH relies on the binary linear separability of GAN’s features (Definition 1). To test if this assumption is satisfied, we compared the performance of semantic segmentation using GAN’s features using either the SVM or the LSE. The LSE shared the same architecture with the SVM but was trained with cross-entropy loss, reflecting the classical linear separability. For each GAN, we sampled 50k images, collected their features, and obtained their segmentation masks using the pretrained DeepLabV3 model. Next, we trained LSE and SVM on the features and segmentation masks for 1 epoch using the Adam optimizer with $10^{-3}$ learning rate. The SVM used Eq. 3 with $\lambda_{1}=5000$, which was identical to the value used in KLiSH. For evaluation, we sampled another different 10k images and measured the mIoU between the segmentation of LSE or SVM and the segmentation of the DeepLabV3 model. Their performance is presented in Tab. III. SVM achieved close performance with LSE in most cases. The relative differences between SVM and LSE were within 5% on 6 models. Therefore, in general, the classical linear separability of GAN’s features can be relaxed to the binary linear separability. K | metric | K-means | KASP | AHC | KLiSH (ours) ---|---|---|---|---|--- euclidean | euclidean | ward | arccos | euclidean StyleGAN2-FFHQ 20 | AMI | 64.8 | 61.9 | 62.9 | 65.2 | 70.6 ARI | 57.1 | 54.4 | 60.7 | 65.1 | 70.5 MIoU | 50.2 | 40.3 | 46.4 | 37.7 | 52.7 $26^{\dagger}$ | AMI | 56.3 | 59.1 | 58.0 | 68.0 | 68.3 ARI | 25.1 | 37.2 | 37.7 | 69.2 | 65.4 MIoU | 52.6 | 48.9 | 46.9 | 43.8 | 56.3 30 | AMI | 56.3 | 59.3 | 57.7 | 65.9 | 67.6 ARI | 25.1 | 38.5 | 35.7 | 64.2 | 64.6 MIoU | 52.6 | 53.2 | 49.9 | 45.5 | 58.0 StyleGAN-CelebAHQ 20 | AMI | 53.0 | 55.5 | 53.1 | 62.5 | 69.0 ARI | 38.3 | 38.1 | 49.1 | 67.0 | 73.1 MIoU | 30.9 | 36.8 | 30.5 | 25.4 | 41.5 $30^{\dagger}$ | AMI | 51.9 | 49.7 | 43.4 | 57.9 | 63.6 ARI | 28.2 | 30.5 | 21.4 | 54.8 | 61.8 MIoU | 38.5 | 33.5 | 31.7 | 27.7 | 45.7 PGGAN-CelebAHQ 20 | AMI | 37.9 | 39.3 | 39.4 | 37.1 | 51.0 ARI | 18.5 | 19.4 | 20.5 | 30.1 | 53.5 MIoU | 15.2 | 16.9 | 16.4 | 19.1 | 23.1 $26^{\dagger}$ | AMI | 35.1 | 37.9 | 37.6 | 36.0 | 51.0 ARI | 11.5 | 15.3 | 15.6 | 28.7 | 53.5 MIoU | 17.4 | 19.0 | 16.4 | 19.2 | 29.0 30 | AMI | 35.1 | 36.2 | 38.0 | 38.1 | 50.2 ARI | 11.5 | 12.4 | 15.4 | 30.7 | 52.4 MIoU | 17.4 | 18.6 | 18.4 | 19.6 | 29.8 TABLE IV: Quantitative evaluation of clustering algorithms on face image generators. The final number of clusters used is indicated by a $\dagger$. For all metrics, the larger the better. The clustering results are reported as the best one among the 5 trials. ### 4.3 KLiSH Clustering Results #### Quantitative evaluations We evaluated the performance of clustering algorithms on GANs trained on face images. We report the metrics at cluster numbers K=20, K=30, and the finally selected number in Tab. IV. It was clear that KLiSH outperformed existing clustering algorithms. For AMI and MIoU, KLiSH was superior to all baselines in all cases. Notably, at the finally selected number of clusters, KLiSH achieved significantly higher MIoU scores. It indicated that KLiSH achieved the best segmentation performance at the cluster number that would be used by downstream tasks. For ARI, KLiSH also achieved the highest score in almost every case, except for K=26 in StyleGAN2-FFHQ, where AHC (arccos) achieved a better score. Even though, qualitative results showed that AHC was not better than KLiSH in this case (see row “final” of column “StyleGAN2-FFHQ” of Fig. 5). AHC failed to cluster “eye”, “nose”, “upper lip”, “lower lip”, but KLiSH successfully identified all those clusters. KLiSH’s lower ARI might result from leaving two clusters for “face” and “neck”. However, in this case, it would be rather easy to merge them manually. In contrast, dividing the undersegmented clusters obtained by AHC would be more difficult. In summary, KLiSH was better than K-means, AHC, and KASP on GAN’s features. #### Qualitative results We visualize the clusters obtained by all the cluster algorithms on various GANs and datasets in Fig. 5. The results were from the same trial (best out of 5 trials) as in Tab. IV. We had two main observations. First, KLiSH was significantly more accurate than K-means, AHC, and KASP. In StyleGAN2-FFHQ, K-means@30 failed to cluster “upper lip”, “lower lip”, “teeth”, “brow” classes and still undersegmented “background” and “hair”. AHC@30 already failed to cluster “nose” and the classes in mouth region. In contrast, KLiSH@30 had no undersegmentation of groundtruth classes. Similarly, for car images, KLiSH@final successfully segmented the “headlight”, “wheel”, and even two “side glasses”. In contrast, the most competitive baseline, AHC@final not only undersegmented “background” and “wheel” classes, but also completely missed the “headlight”. Second, the advantage of KLiSH over existing clustering algorithms was consistent across GANs and datasets. For StyleGAN-CelebAHQ, AHC@final missed “nose”, “lip”, and “teeth” classes and undersegmented “hair”, while KLiSH@final did not. For PGGAN-CelebAHQ, although all the clustering algorithms did not obtain good results – probably because PGGAN had learned the dataset well. Still, KLiSH@final successfully segmented “ear” and “hair” classes, but AHC@final missed “ear” and mixed “hair” with “background” (see the rightmost column). We provide more visualization in Fig. 7, which is also supportive of the observations described above. Figure 6: Demonstration of a merging step. The first two columns show the images and clusters. The confidence maps $\\{\mathbf{s}_{k}^{\dagger}\\}_{k=1}^{M}$ of the image are visualized (partially) in other columns. $\\{\mathbf{s}_{k}^{\dagger}\\}_{k=1}^{M}$ take value within [0, 1] and are visualized using the color bar on the right. The cluster in yellow box (say cluster $p$) has the smallest IoU (Eq. 4) with K=35. The cluster in green box (say cluster $q$) has the largest ECoS (Eq. 5) with cluster $p$. Cluster $p$ and $q$ are then merged in this step. Figure 7: Comparison of clustering algorithms on various datasets and GANs. “final” refers to 18, 9, 16, 15, and 28 from left to right. #### Visualization of a merging step To help readers gain a concrete picture of how KLiSH merged the clusters, we also visualized a single merging step in Fig. 6. See the cluster confidence map of “nose” and “brow” (Rows 2 and 3, Column 4 in the right figure), the red regions are dominated by deep red, indicating that the SVM predicts them with high confidence. On the contrary, the heatmap denoted by the yellow box has a large proportion of shallow red, indicating that the confidence of SVM to predict that cluster was low. This cluster had the lowest IoU at K=35 and was subsequently merged with another oversegmented cluster of hair, resulting in better cluster assignments. ## 5 Applications Method | #shots | #classes | mIoU(%) ---|---|---|--- UFGS (KLiSH) | 0 | 15 | 47.2 UFGS (AHC) | 0 | 15 | 36.6 LSE[7] | 1 | 15 | 47.1 $\pm$ 2.8 10 | 58.5 $\pm$ 0.9 RepurposeGAN [15] | 10 | 9 | 68.0 DatasetGAN [14] | 16 | 8 | 70.0 TABLE V: The performance of two UFGS methods and several few-shot learning methods. method | #samples | skin | nose | eye-g | eye | brow | ear | teeth | u-lip | l-lip | hair | hat | ear-r | neck | cloth | mouth* ---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|--- UFGS (KLiSH) | (K=26) | 82.2 | 76.0 | 25.7 | 50.4 | 60.2 | 50.3 | 32.1 | 52.2 | 51.0 | 76.5 | 0.0 | 0.0 | 70.1 | 34.0 | - UFGS (AHC) | (K=26) | 62.0 | 19.3 | 27.6 | 62.6 | 30.6 | 49.3 | 0.00 | 47.1 | 60.3 | 80.5 | 0.1 | 11.0 | 34.6 | 27.2 | - LSE | N=1 | 83.8 | 81.7 | 1.9 | 72.9 | 51.0 | 45.6 | 32.7 | 64.8 | 68.2 | 69.0 | 0.6 | 1.7 | 68.8 | 16.1 | - N=10 | 89.3 | 85.5 | 25.7 | 72.2 | 61.4 | 58.9 | 60.5 | 70.1 | 74.2 | 85.2 | 6.3 | 11.5 | 74.7 | 43.6 | - RepurposeGAN | N=1 | - | 76 | - | 58 | - | - | - | - | - | - | - | - | - | - | 71 N=10 | 90 | 85 | - | 75 | 68 | 37 | - | - | - | 84 | - | - | 73 | 16 | 87 TABLE VI: IoU(%) per class of DeepLabV3 trained on synthetic datasets generated with different annotation methods. The evaluation is conducted on real images from the test split of CelebAMask-HQ. “eye-g”, “u-lip”, “l-lip” denotes “eyeglasses”, “upper lip”, and “lower lip”, respectively. The mouth class is a union of “teeth”, “u-lip”, and “l-lip”. Our UFGS achieved close performance with one-shot LSE on most classes and outperformed one-shot LSE on hard classes like “eyeglasses” and “cloth”. Figure 8: UFGS results on _real_ images from CelebAMask-HQ (rows 1 and 2) and PASCAL VOC (rows 3 and 4). Figure 9: The qualitative results of USCS on various GANs and datasets. The semantic masks are drawn by users and the modifications on the masks are indicated by dashed boxes. The image on the right of each mask is generated by our USCS model. ### 5.1 Unsupervised Fine-Grained Segmentation Semantic segmentation labeling is very expensive. As described by Lee _et al_. in CelebAMask-HQ [12], annotating fine-grained semantic segmentation is not only time-consuming but also involves multiple iterations of refinement. To reduce the labeling cost, we propose an approach for Unsupervised Fine-grained Segmentation (UFGS). Existing methods on unsupervised segmentation [52, 54, 55, 56] only segment foreground from background, thus are not applicable to UFGS. A more closely related task to UFGS is few-shot learning-based segmentation [7, 15, 14]. Although they reduce the number of annotations, the cost of fine-grained semantic annotations is still significant. Zhang _et al_. [14] reports that annotating a single face image cost on average 20 minutes _per image_. In contrast, the only labeling cost involved in UFGS is naming and merging a few labels, which could take less than one minute for a _whole dataset_. #### Method The proposed method for UFGS is as follows: First, train a GAN on the target dataset, run KLiSH on the GAN, and select a proper number of semantic clusters. Next, generate an image-segmentation dataset by sampling images from the GAN and using the SVM returned by KLiSH to obtain the segmentation masks of the image. We apply the SVM directly on the feature maps upsampled to the resolution of images. Then, train a segmentation network, _e.g_., DeepLabV3 [11], on the synthetic dataset. Finally, segment _real_ images in the original dataset. #### Evaluation We evaluated the performance of our UFGS method on face images using the test split of the CelebAMask-HQ dataset. The cluster assignments were matched to the ground-truth classes by maximizing MIoU (Algorithm 2) using 100 images from the validation split. Note that the images and labels are not used to train the model. In practice, users can match clusters to labels with just several clicks. The performance of our method was compared to few-shot learning methods [7, 15, 14]. To the best of our knowledge, except for face images, there are no large-scale semantic annotations on the datasets described in Sec. 4.1. PASCAL VOC [69] contains car images, but they have a significant domain gap with the car images from the GAN’s dataset, LSUN-Car. In two relevant works, DatasetGAN[14] and RepurposeGAN [15], the authors make their own test sets by filtering PASCAL VOC and have not released them as of now. Therefore, we only present qualitative results on PASCAL VOC and report quantitative results on CelebAMask-HQ. #### Experiment setup We used StyleGAN2-FFHQ and StyleGAN2-Car for UFGS on CelebAMask-HQ and PASCAL VOC cars, respectively. The baseline for comparison was UFGS with AHC, as it was the most competitive method with KLiSH on StyleGAN2-FFHQ according to Tab. IV. The settings were identical to UFGS with KLiSH except that the datasets were generated with clusters found by AHC. We also compared UFGS with KLiSH to a few-shot semantic extractor, LSE. We reproduced LSE on StyleGAN2-FFHQ and StyleGAN2-Car following the settings in [7]. In brief, we randomly sampled 1 or 10 images (for one-shot or 10-shot experiments) and obtained their segmentation using the pretrained DeepLabV3 model. Then, we trained the LSE on the features and labels of those images for 6,400 iterations. To address the variation of few-shot experiments, we trained LSE five times, each time with a different set of image annotations. After training the LSE and obtaining the KLiSH clusters, we then compared their performances on benefiting the downstream segmentation task. We first sampled a synthesized image segmentation dataset for each of the LSE models and KLiSH results, which consisted of 50k images and segmentation masks. Then, we trained a DeepLabV3 model on each dataset following the training settings described in Sec. 4.1. Finally, we tested the DeepLabV3 models on the real images from the test split of the CelebAMask-HQ dataset. #### Results We present the qualitative results of our UFGS method in Fig. 8. As shown in Row 1 and 2, UFGS model segmented facial classes like “eye”, “brow”, “lips”, “teeth” and “nose” well. For car images in Row 3 and 4, our UFGS method also successfully segmented “headlight”, “body”, “glasses”, and “wheel”. However, we found there was a domain gap between synthesized images and real images. See Row 1, Col 3 and 4 in Fig. 8. The hand in front of the face is a rare case in the dataset. Therefore, the segmentation model cannot deal with this case well. The quantitative results are reported in Tab. V and Tab. VI. As shown in Tab. V, KLiSH surpassed UFGS with AHC by a large margin. In comparison with few- shot learning methods, UFGS achieved close performance with the one-shot method, LSE, but lagged behind the 10-shot methods. In summary, though the performance of our UFGS method was limited by the domain gap as a result of using GANs, we for the first time showed accurate fine-grained semantic segmentation in an unsupervised learning setting. ### 5.2 Unsupervised Semantic-Conditional Synthesis Enhancing controllability over the generated images is important for commercial applications of GANs. Most GANs do not allow the user to control the spatial structure of generated images precisely. Semantic-Controllable Synthesis (SCS) aims to provide controllability over precise spatial structures of generated images. Though achieving greater controllability for image synthesis, existing methods for SCS, _e.g_., pix2pix [42], SEAN [45] and GauGAN [44], rely heavily on large-scale human annotated semantic masks. Recently, the labeling requirements of SCS have been lowered to a few annotations [7, 13]. Yet, the user still has to label a few multi-class fine- grained segmentation masks, which limits the application in practice. For example, users may want to enable SCS on images of tigers. If there are no semantic segmentation masks on these images, the user needs to annotate 16 to 40 images. The annotation process would take around 2 to 5 hours, according to Zhang _et al_. [14]. To lower the label dependency of SCS methods, we propose the task of Unsupervised SCS, which aims to achieve semantic-controllable image synthesis on arbitrary datasets without human annotations. #### Method Our USCS application works as follows. First, the user shall provide a target image dataset. Then, we train an unconditional GAN and generate an image- segmentation dataset using the same method as described in Sec. 5.1. Finally, we train a GauGAN [44] on the generated dataset. The above process can be done offline. Then users can draw and edit on the semantic mask and the GauGAN will generate desired images. We developed a web-based interface to help users generate images with desired structures. Before the user starts painting, a few images and their semantic masks are shown as examples. Next, the user can click on a palette and paint with selected semantic classes. Then, the user can draw freely and submit the painting to the server, which will return an image synthesized given the semantic mask. #### Experiment setup We conducted the USCS experiment on StyleGAN2 models trained on FFHQ and Car datasets and StyleGAN2-ADA models trained on MetFaces, AFHQ [40] Cat, Dog, and Wildlife split. Similar to the setting of UFGS, we also synthesized a dataset with 50k paired images and segmentation masks from each GAN and trained a GauGAN on each dataset. The GauGAN was trained using the official release 444https://www.github.com/nvlabs/spade for 10 epochs on 8 GPUs. The training took around 100 hours for a GauGAN model. #### Results The editing results on different datasets and GAN models are presented in Fig. 9. We observed that USCS supported precise, diverse, and even out-of- distribution image editing operations. First, USCS respected the semantic mask well, accurately translating the user’s semantic input into corresponding images while maintaining good photo-realisticity. See the “big round eye” example in ADA-Cat. The enlarged eyes had a matched shape as specified by the user. Second, our USCS method well demonstrated the diversity and degree of freedom of editing operations. As shown in the left part of StyleGAN2-FFHQ, the shape of the sunglasses could even be changed to triangle. Third, the out- of-distribution editing operations could also lead to photorealistic synthesis results. See the “elf ear” examples in StyleGAN2-FFHQ and ADA-Wild, the synthesized sharp ear is not present in the training set of the model. Besides, the “blink eye” in ADA-MetFace and ADA-Wild, and “more wheels” in StyleGAN2-Car are all out-of-distribution semantic masks. To the best of our knowledge, SCS cannot be applied to datasets of wildlife and artistic portraits, _etc_., because so far no semantic annotations on such datasets are available. EditGAN [13] relies on 16 to 30 fine-grained segmentation masks to enable semantic editing on Cat, Dog, _etc_., but the cost of labeling is still high (around 10 minutes per image). In contrast, the USCS model is free from the constraint of annotations. Please find a demonstration of our USCS web application in Supplementary Video 1 and have a try on our online demonstration555https://atlantixjj.github.io/KLiSH/. ## 6 Limitations Our method is generally constrained by the performance of GANs. First, the clustering performance of KLiSH is dependent on the degree of linear separability of GAN’s features. If the GAN does not encode semantics into _one vs. rest_ linearly separable clusters, then the KLiSH might fail to find meaningful clusters. Second, the performance of UFGS and USCS is limited by the synthesis quality of the original GAN. When the GAN fails to generate realistic images, the UFGS model will have a domain gap when applied to real images and USCS will also generate unrealistic images. ## 7 Conclusion In this paper, we propose KLiSH to find semantic clusters in GAN’s features. By making use of GAN’s linear separability in semantics, KLiSH achieves better clustering results than conventional clustering algorithms including K-means, AHC, and KASP. The clusters found by KLiSH accurately correspond to the semantic parts of generated images. Therefore, for each pretrained GAN, we can generate an image-segmentation dataset for downstream tasks. One can easily enable semantics-based applications including UFGS and USCS with the dataset generated by GANs. The surprising performance of UFGS and USCS make us believe that GANs can play a more important role in representation learning. ## References * [1] Y. Bengio, A. Courville, and P. Vincent, “Representation learning: A review and new perspectives,” _PAMI_ , vol. 35, no. 8, pp. 1798–1828, 2013. * [2] K. H. R. Chan, Y. Yu, C. You, H. Qi, J. Wright, and Y. Ma, “Redunet: A white-box deep network from the principle of maximizing rate reduction,” _JMLR_ , vol. 23, no. 114, pp. 1–103, 2022. * [3] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in _NeurIPS_ , 2014, pp. 2672–2680. * [4] Y. Shen, C. Yang, X. Tang, and B. Zhou, “Interfacegan: Interpreting the disentangled face representation learned by gans,” _PAMI_ , vol. 44, no. 4, pp. 2004–2018, 2022. * [5] C. Yang, Y. Shen, and B. Zhou, “Semantic hierarchy emerges in deep generative representations for scene synthesis,” _IJCV_ , vol. 129, no. 5, pp. 1451–1466, 2021. * [6] D. Bau, J.-Y. Zhu, H. Strobelt, B. Zhou, J. B. Tenenbaum, W. T. Freeman, and A. Torralba, “Visualizing and understanding generative adversarial networks,” in _ICLR_ , 2019. * [7] J. Xu and C. Zheng, “Linear semantics in generative adversarial networks,” in _CVPR_ , 2021, pp. 9351–9360. * [8] A. Jahanian, L. Chai, and P. Isola, “On the” steerability” of generative adversarial networks,” in _ICLR_ , 2019. * [9] A. Voynov and A. Babenko, “Unsupervised discovery of interpretable directions in the gan latent space,” in _ICML_. PMLR, 2020, pp. 9786–9796. * [10] Y. Shen and B. Zhou, “Closed-form factorization of latent semantics in gans,” in _CVPR_ , 2021, pp. 1532–1540. * [11] L.-C. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinking atrous convolution for semantic image segmentation,” _arXiv preprint arXiv:1706.05587_ , 2017. * [12] C.-H. Lee, Z. Liu, L. Wu, and P. Luo, “Maskgan: Towards diverse and interactive facial image manipulation,” in _CVPR_ , 2020, pp. 5549–5558. * [13] H. Ling, K. Kreis, D. Li, S. W. Kim, A. Torralba, and S. Fidler, “Editgan: High-precision semantic image editing,” in _NeurIPS_ , 2021. * [14] Y. Zhang, H. Ling, J. Gao, K. Yin, J.-F. Lafleche, A. Barriuso, A. Torralba, and S. Fidler, “Datasetgan: Efficient labeled data factory with minimal human effort,” in _CVPR_ , 2021, pp. 10 145–10 155. * [15] N. Tritrong, P. Rewatbowornwong, and S. Suwajanakorn, “Repurposing gans for one-shot semantic part segmentation,” in _CVPR_ , 2021, pp. 4475–4485. * [16] E. Collins, R. Bala, B. Price, and S. Süsstrunk, “Editing in style: Uncovering the local semantics of GANs,” in _CVPR_ , 2020. * [17] O. Kafri, O. Patashnik, Y. Alaluf, and D. Cohen-Or, “Stylefusion: Disentangling spatial segments in stylegan-generated images,” _TOG_ , 2022\. * [18] Y. Zheng, Y.-K. Huang, R. Tao, Z. Shen, and M. Savvides, “Unsupervised disentanglement of linear-encoded facial semantics,” in _CVPR_ , 2021, pp. 3917–3926. * [19] D. Yan, L. Huang, and M. I. Jordan, “Fast approximate spectral clustering,” in _Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining_ , ser. KDD ’09. New York, NY, USA: Association for Computing Machinery, 2009, p. 907–916. * [20] T. Karras, T. Aila, S. Laine, and J. Lehtinen, “Progressive growing of gans for improved quality, stability, and variation,” _arXiv preprint arXiv:1710.10196_ , 2017. * [21] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, “Gans trained by a two time-scale update rule converge to a local nash equilibrium,” in _NeurIPS_ , 2017, pp. 6626–6637. * [22] E. L. Denton, S. Chintala, R. Fergus _et al._ , “Deep generative image models using a laplacian pyramid of adversarial networks,” in _NeurIPS_ , 2015, pp. 1486–1494. * [23] A. Odena, C. Olah, and J. Shlens, “Conditional image synthesis with auxiliary classifier gans,” in _ICML_. PMLR, 2017, pp. 2642–2651. * [24] X. Huang, Y. Li, O. Poursaeed, J. Hopcroft, and S. Belongie, “Stacked generative adversarial networks,” in _CVPR_ , 2017, pp. 5077–5086. * [25] T. R. Shaham, T. Dekel, and T. Michaeli, “Singan: Learning a generative model from a single natural image,” in _CVPR_ , 2019, pp. 4570–4580. * [26] M. Arjovsky, S. Chintala, and L. Bottou, “Wasserstein generative adversarial networks,” in _ICML_. PMLR, 2017, pp. 214–223. * [27] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. C. Courville, “Improved training of wasserstein gans,” in _NeurIPS_ , 2017, pp. 5767–5777. * [28] N. Kodali, J. Abernethy, J. Hays, and Z. Kira, “On convergence and stability of gans,” _arXiv preprint arXiv:1705.07215_ , 2017. * [29] X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, and P. Abbeel, “Infogan: Interpretable representation learning by information maximizing generative adversarial nets,” in _NeurIPS_ , 2016, pp. 2180–2188. * [30] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” _arXiv preprint arXiv:1511.06434_ , 2015. * [31] J. Zhao, M. Mathieu, and Y. LeCun, “Energy-based generative adversarial networks,” in _ICLR_ , 2017. * [32] D. Berthelot, T. Schumm, and L. Metz, “Began: Boundary equilibrium generative adversarial networks,” _arXiv preprint arXiv:1703.10717_ , 2017. * [33] T. Karras, S. Laine, and T. Aila, “A style-based generator architecture for generative adversarial networks,” in _CVPR_ , 2019, pp. 4401–4410. * [34] T. Karras, S. Laine, M. Aittala, J. Hellsten, J. Lehtinen, and T. Aila, “Analyzing and improving the image quality of stylegan,” in _CVPR_ , 2020, pp. 8110–8119. * [35] T. Karras, M. Aittala, J. Hellsten, S. Laine, J. Lehtinen, and T. Aila, “Training generative adversarial networks with limited data,” _arXiv preprint arXiv:2006.06676_ , 2020. * [36] T. Karras, M. Aittala, S. Laine, E. Härkönen, J. Hellsten, J. Lehtinen, and T. Aila, “Alias-free generative adversarial networks,” _arXiv preprint arXiv:2106.12423_ , 2021. * [37] H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena, “Self-attention generative adversarial networks,” in _ICML_. PMLR, 2019, pp. 7354–7363. * [38] A. Brock, J. Donahue, and K. Simonyan, “Large scale gan training for high fidelity natural image synthesis,” in _ICLR_ , 2018. * [39] Y. Choi, M. Choi, M. Kim, J.-W. Ha, S. Kim, and J. Choo, “Stargan: Unified generative adversarial networks for multi-domain image-to-image translation,” in _CVPR_ , 2018, pp. 8789–8797. * [40] Y. Choi, Y. Uh, J. Yoo, and J.-W. Ha, “Stargan v2: Diverse image synthesis for multiple domains,” in _CVPR_ , 2020, pp. 8188–8197. * [41] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in _CVPR_ , 2017, pp. 2223–2232. * [42] P. Isola, J. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation with conditional adversarial networks,” in _CVPR_ , 2017, pp. 5967–5976. * [43] T.-C. Wang, M.-Y. Liu, J.-Y. Zhu, A. Tao, J. Kautz, and B. Catanzaro, “High-resolution image synthesis and semantic manipulation with conditional gans,” in _CVPR_ , 2018, pp. 8798–8807. * [44] T. Park, M.-Y. Liu, T.-C. Wang, and J.-Y. Zhu, “Gaugan: semantic image synthesis with spatially adaptive normalization,” in _ACM SIGGRAPH 2019 Real-Time Live!_ , 2019, pp. 1–1. * [45] P. Zhu, R. Abdal, Y. Qin, and P. Wonka, “Sean: Image synthesis with semantic region-adaptive normalization,” in _CVPR_ , 2020, pp. 5104–5113. * [46] Y. Jo and J. Park, “Sc-fegan: Face editing generative adversarial network with user’s sketch and color,” in _CVPR_ , 2019, pp. 1745–1753. * [47] Y. Shi, X. Yang, Y. Wan, and X. Shen, “Semanticstylegan: Learning compositional generative priors for controllable image synthesis and editing,” in _CVPR_ , 2022, pp. 11 254–11 264. * [48] W. Sun and T. Wu, “Learning layout and style reconfigurable gans for controllable image synthesis,” _PAMI_ , pp. 1–1, 2021. * [49] D. Bau, S. Liu, T. Wang, J.-Y. Zhu, and A. Torralba, “Rewriting a deep generative model,” in _ECCV_ , 2020. * [50] S.-Y. Wang, D. Bau, and J.-Y. Zhu, “Sketch your own gan,” in _CVPR_ , 2021, pp. 14 050–14 060. * [51] D. Galeev, K. Sofiiuk, D. Rukhovich, M. Romanov, O. Barinova, and A. Konushin, “Learning high-resolution domain-specific representations with a gan generator,” in _Joint IAPR International Workshops on Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR)_. Springer, 2021, pp. 108–118. * [52] Y. Benny and L. Wolf, “Onegan: Simultaneous unsupervised learning of conditional image generation, foreground segmentation, and fine-grained clustering,” in _ECCV (26)_ , 2019, pp. 514–530. * [53] K. K. Singh, U. Ojha, and Y. J. Lee, “Finegan: Unsupervised hierarchical disentanglement for fine-grained object generation and discovery,” in _CVPR_ , 2019, pp. 6490–6499. * [54] R. Abdal, P. Zhu, N. J. Mitra, and P. Wonka, “Labels4free: Unsupervised segmentation using stylegan,” in _CVPR_ , 2021, pp. 13 970–13 979. * [55] L. Melas-Kyriazi, C. Rupprecht, I. Laina, and A. Vedaldi, “Finding an unsupervised image segmenter in each of your deep generative models,” in _ICLR_ , 2021. * [56] Y. Yang, H. Bilen, Q. Zou, W. Y. Cheung, and X. Ji, “Unsupervised foreground-background segmentation with equivariant layered gans.” _arXiv preprint arXiv:2104.00483_ , 2021. * [57] A. Voynov, S. Morozov, and A. Babenko, “Object segmentation without labels with large-scale generative models,” in _International Conference on Machine Learning_. PMLR, 2021, pp. 10 596–10 606. * [58] Y. He, Z. Zhang, J. Zhu, Y. Shen, and Q. Chen, “Interpreting class conditional gans with channel awareness,” _arXiv preprint arXiv:2203.11173_ , 2022. * [59] S. Gao, Z.-Y. Li, M.-H. Yang, M.-M. Cheng, J. Han, and P. H. S. Torr, “Large-scale unsupervised semantic segmentation.” _CVPR_ , 2021\. * [60] W. V. Gansbeke, S. Vandenhende, S. Georgoulis, and L. V. Gool, “Unsupervised semantic segmentation by contrasting object mask proposals,” in _CVPR_ , 2021, pp. 10 052–10 062. * [61] S. Liu, L. Zhang, X. Yang, H. Su, and J. Zhu, “Unsupervised part segmentation through disentangling appearance and shape,” in _CVPR_ , 2021, pp. 8355–8364. * [62] R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin, “Liblinear: A library for large linear classification,” _JMLR_ , vol. 9, no. Aug, pp. 1871–1874, 2008. * [63] D. C. Liu and J. Nocedal, “On the limited memory bfgs method for large scale optimization,” _Mathematical programming_ , vol. 45, no. 1, pp. 503–528, 1989. * [64] Z. Liu, P. Luo, X. Wang, and X. Tang, “Deep learning face attributes in the wild,” in _ICCV_ , December 2015. * [65] F. Yu, A. Seff, Y. Zhang, S. Song, T. Funkhouser, and J. Xiao, “Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop,” _arXiv preprint arXiv:1506.03365_ , 2015. * [66] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in _CVPR_ , 2016, pp. 770–778. * [67] B. Zhou, H. Zhao, F. X. P. Fernandez, S. Fidler, and A. Torralba, “Scene parsing through ade20k dataset,” in _CVPR_ , 2017. * [68] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, A. Desmaison, A. Kopf, E. Yang, Z. DeVito, M. Raison, A. Tejani, S. Chilamkurthy, B. Steiner, L. Fang, J. Bai, and S. Chintala, “Pytorch: An imperative style, high-performance deep learning library,” in _NeurIPS_. Curran Associates, Inc., 2019, pp. 8024–8035. * [69] M. Everingham, L. Van Gool, C. Williams, J. Winn, and A. Zisserman, “The pascal visual object classes (voc) challenge,” _IJCV_ , vol. 88, pp. 303–338, 06 2010. | Jianjin Xu (S’01-M’08-SM’13) received a B.E. degree in computer science from the Tsinghua University, Beijing, China, in 2019, and a M.S. degree in computer science from the Columbia University, New York City, NY, U.S., in 2021. He is currently an Assistant Research Scientist at the School of Mathematics and Computer Science, Panzhihua University, Sichuan, China. His current research interests include generative models, neural network interpretation, and computer vision. Previously he was a research assistant at the Department of Computer Science and Technology, Tsinghua University. ---|--- | Zhaoxiang Zhang received his bachelor’s degree in Circuits and Systems from the University of Science and Technology of China (USTC) in 2004. In 2004, he joined the National Laboratory of Pattern Recognition (NLPR), Institute of Automation, Chinese Academy of Sciences, under the supervision of Professor Tieniu Tan, and he received his Ph.D. degree in 2009. In October 2009, he joined the School of Computer Science and Engineering, Beihang University, as an Assistant Professor (2009-2011), an Associate professor (2012-2015) and the vise-director of the Department of Computer application technology (2014-2015). In July 2015, he returned to the Institute of Automation, Chinese Academy of Sciences. He is now a full Professor in the Center for Research on Intelligent Perception and Computing (CRIPAC) and the National Laboratory of Pattern Recognition (NLPR). His research interests include Computer Vision, Pattern Recognition, and Machine Learning. Recently, he specifically focuses on biologically inspired intelligent computing and its applications on human analysis and scene understanding. He has published more than 150 papers in the international journals and conferences, including reputable international journals such as IEEE TPAMI, IJCV, JMLR and top level international conferences like CVPR, ICCV, ECCV, ICLR, NeurIPS, AAAI and IJCAI. He served as the associated editor of IEEE TCSVT, PR, and Frontiers of Computer Science. He served as the area chair of international conferences like CVPR, ICCV, AAAI, IJCAI. He is a senior member of IEEE. ---|--- | Xiaolin Hu (S’01-M’08-SM’13) received B.E. and M.E. degrees in automotive engineering from the Wuhan University of Technology, Wuhan, China, in 2001 and 2004, respectively, and a Ph.D. degree in automation and computer-aided engineering from the Chinese University of Hong Kong, Hong Kong, in 2007. He is currently an Associate Professor at the Department of Computer Science and Technology, Tsinghua University, Beijing, China. His current research interests include deep learning and computational neuroscience. At present, he is an Associate Editor of the IEEE Transactions on Pattern Analysis and Machine Intelligence, IEEE Transactions on Image Processing, and Cognitive Neurodynamics. Previously he was an Associate Editor of the IEEE Transactions on Neural Networks and Learning Systems. ---|---
# Coordinating Cross-modal Distillation for Molecular Property Prediction ††thanks: * These authors contributed equally. Hao Zhang2,* Nan Zhang1,*, Ruixin zhang2, Lei Shen2, Yingyi Zhang2, Meng Liu3 1 Academy for Engineering and Technology, Fudan University, China 2 Tencent Youtu Lab 3 Shanghai Jiao Tong University ###### Abstract In recent years, molecular graph representation learning (GRL) has drawn much more attention in molecular property prediction (MPP) problems. The existing graph methods have demonstrated that 3D geometric information is significant for better performance in MPP. However, accurate 3D structures are often costly and time-consuming to obtain, limiting the large-scale application of GRL. It is an intuitive solution to train with 3D to 2D knowledge distillation and predict with only 2D inputs. But some challenging problems remain open for 3D to 2D distillation. One is that the 3D view is quite distinct from the 2D view, and the other is that the gradient magnitudes of atoms in distillation are discrepant and unstable due to the variable molecular size. To address these challenging problems, we exclusively propose a distillation framework that contains global molecular distillation and local atom distillation. We also provide a theoretical insight to justify how to coordinate atom and molecular information, which tackles the drawback of variable molecular size for atom information distillation. Experimental results on two popular molecular datasets demonstrate that our proposed model achieves superior performance over other methods. Specifically, on the largest MPP dataset PCQM4Mv2 served as an ”ImageNet Large Scale Visual Recognition Challenge” in the field of graph ML, the proposed method achieved a 6.9% improvement compared with the best works. And we obtained fourth place with the MAE of 0.0734 on the test-challenge set for OGB-LSC 2022 Graph Regression Task. We will release the code soon. ## I Introduction In recent years, molecular property prediction (MPP) has played a crucial role in new materials and drug discovery. Since atoms and bonds in a molecule are naturally in a graph structure, graph neural networks (GNN) [1] have shown great vitality in MPP. Many works [2, 3, 4] take 2D molecular structures as input of GNN for training and testing. Their performances are limited because they do not consider the 3D atom coordinates that determine certain chemical and physical functionalities of molecules. Methods [5, 6, 7] that exploit 3D structures exhibit advantages in accuracy, especially properties related to quantum mechanics, such as single-point energy, atomic forces, and dipole moments. However, obtaining accurate 3D structures is expensive and time- consuming. Calculating 3D information with DFT takes ${10^{5}}$ times longer than model feedforward [2], which seriously impedes the application of 3D-based methods. Some works attempt to augment 2D information with 3D information through training to enhance the prediction accuracy while keeping inference speed. One direction is developing 2D-3D contrastive tasks. Another way is to generate original conformations through reconstruction tasks. Usually, these tasks require several different 3D conformers to create positive and negative pairs for self-supervised learning, but obtaining different conformers is time- consuming for large-scale data. Also, different conformers from one SMILES[8] code may represent different property values, which lower the performance of contrastive-based methods. Lastly, some applications do not support the requirement of multi conformers. With the force field given, one SMILES code usually can only generate one 3D conformer structure. Distillation is another intuitive solution that utilizes 3D knowledge in training while keeping inference speed. Also, distillation methods do not have the limitation of multiple conformers requirements. But some challenging problems remain for distillation methods. On the one hand, the 3D view is quite distinct from the 2D view, making the top-level-feature-only distillation less effective. On the other hand, which tokens participate in distillation is still an open question for the MPP. As self-attention calculates information of the virtual token from all-atom features, it is natural to include all-atom tokens together in distillation. But simply distilling atom tokens and virtual token together leads to poor performance. We found the gradient magnitudes of atom tokens distillation are inconsistent and unstable due to the variable molecular size. So distilling local atom information without correction will cause degradation problems. Motivated by the above observation, we propose a novel coordinating cross- modal distillation (CCMD) framework. We formulate a coordinating weight, which dynamically scales and balances the gradients of global molecular token and local atom tokens according to the number of atoms. Also, the distillation of virtual token and atom tokens is carried in all layers to enhance the 2D-3D distillation performance. Lastly, we introduce absolute position encoding to integrate adjacent edge information, achieving better performance in both views of 2D and 3D. Experimental results on two popular molecular datasets demonstrate that our proposed model achieves state-of-the-art performance. In conclusion, our main contributions to this work can be summarized as follows: * • We exclusively develop an enhanced cross-modal distillation on all layers for molecular property prediction, which distills global molecular information and local atom information to transfer 3D geometric information to 2D view. * • For coordinating global molecular and local atom information, we formulate a novel weight according to the number of atoms in a molecular to scale gradient magnitudes, which boosts the distillation performance. Specific, we demonstrate the coordinating weight is $f(N)=\frac{1}{{N}^{2}}$ for transformer and $f(N)=\frac{1}{{N}}$ for GIN. N is the number of atoms in a molecular. * • We design an absolute position encoding integrated with adjacent edge information, which could distinguish atoms same atom id in the input layer and increase the performance of both 2D and 3D. * • CCMD can enable molecular attention to more atoms in the 2D view. Extensive experiments on the PCQM4Mv2 dataset [9] demonstrate that our proposed method consistently outperforms other state-of-the-art methods in most situations. ## II Related Work ### II-A Molecular Property Prediction According to different dimensions, molecular expressions contain 1D SMILES code, 2D graph, and 3D graph. Due to the natural capability for representing the molecular structure, graph neural networks (GNN) have attracted attention in molecular property prediction, especially with the input of 2D or 3D expressions. #### II-A1 2D Methods Many works focus on applying GNN to 2D graphs. For example, Gilmer et al.[2] proposed Message Passing Neural Networks, which combine message passing and aggregating algorithms with GNN, forming a successful framework for MPP. Liu et al.[3] introduced a simple unsupervised representation for molecules that embeds the vertices in the molecule graph and constructs a compact feature by assembling the vertex embeddings. Ying et al.[10] proposed Graphormer, which utilizes a Transformer [11] in the GNN by effectively encoding the graph structure information. Although these methods have shown promising capabilities, their potential is limited as they do not use 3D coordinates that better represent molecular structure and energy. #### II-A2 3D Methods 3D molecular structures contain spatial information important to molecular property prediction, such as bond angles and lengths. Many recent methods try to answer how to utilize this 3D information in the GNN model. For instance, Klicpera et al.[5] proposed to let message embeddings interact based on the distance between atoms and the angle between directions to improve quantum mechanical property prediction. Lu et al.[12] proposed a Multilevel Graph Convolutional neural Network that extracts features from the conformation and spatial information with multilevel interactions. Liu et al.[7] proposed SphereNet, a 3D graph network framework with spherical message passing. 3D-based methods achieve significant performance improvements. But obtaining 3D structures is time-costing, which reduces the value of 3D-based methods in large-scale applications. ### II-B Cross Modality Training Methods Since training is relatively insensitive to time-consuming, an intuitive approach is to use 2D and 3D information in training and only use 2D in prediction. #### II-B1 Contrastive Methods These methods use 3D conformers and 2D structures to construct positive and negative pairs while applying a contrastive loss during training. For example, Liu et al.[13] proposed the Graph Multi-View Pretraining framework, in which they perform self-supervised learning by leveraging the correspondence and consistency between 2D topological structures and 3D geometric views. Li et al.[14] developed GeomGCL, a dual-view geometric message passing network that utilizes the molecular geometry across 2D and 3D views. Hu et al.[15] proposed a strategy of pre-training GNN at the level of individual nodes as well as entire graph, so the GNN can learn useful local and global information. Stark et al.[16] proposed 3D Infomax, which pre-trains a GNN by maximizing the mutual information between its 2D and 3D embeddings. The drawbacks of contrastive methods lie in the requirements of multiple conformers. Large- scale conformer generation is costly, while one SMILES code usually generates only one 3D conformer with a specific force field. #### II-B2 Knowledge Distilling Methods Knowledge distilling (KD) is an intuitive solution for 3D to 2D cross-modality training. As our work relies on Transformer-based methods such as Graphormer, we only discuss Transformer related KD methods. Many works [17, 18, 19, 20, 21] have attempted to apply KD to the Transformer-based model. Upon molecular property prediction, Zhu et al.[22] proposed ST-KD, an end-to-end Transformer KD framework, bridging the knowledge transfer between graph-based and SMILES- based models. But ST-KD only KD the atom to atom attentions, neglecting the informative embeddings in atom tokens. Based on our experiment, gradient magnitudes of atom-token distillation are discrepant and unstable due to the variable molecular size. Distilling atom-token embeddings without correction will cause degradation problems. To address this issue, we propose a theoretical insight to justify how to coordinate atom token embeddings and molecular features and formulate an auto- correcting distilling weight to dynamically tune and balance the gradients according to the number of atoms. ## III Method Figure 1: The illustration of of our CCMD framework with global molecular and local atom information for molecular property prediction. $\mathcal{L}_{m}$ represents the global molecular distillation, $\mathcal{L}_{a}$ represents the local atom information distillation, and coordinating denotes the coordinating weight scaling gradient magnitudes. ### III-A Overview In this section, we present the key components of CCMD. As illustrated in Fig. 1, the whole graph transformer framework consists of a 3D model and a 2D model, and the backbone is based on the graphormer [10]. Firstly, the 3D model is trained with 3D information as a teacher model. Then, frozing the parameters of the teacher, we train the 2D student model with a distilling loss and a supervised target loss. The distilling module consists of the global molecular distillation and local atom distillation, which can further improve the performance of 2D molecular graphs. At last, for coordinating global and local information, we formulate a novel coordinating weight to scale and balance the gradients optimally. ### III-B Preliminary For convenience, let $\mathcal{D}=\\{(X,Y,E)\\}$ denotes a 2D molecular graph, the $X$, $Y$ and $E$ denote the input of atoms, property labels and edges respectively. In addition, we use $\widetilde{}$ mark the 3D view, so 3D molecular graph can be denoted as $\widetilde{\mathcal{D}}=\\{(\widetilde{X},\widetilde{Y},\widetilde{E})\\}$, where $\widetilde{E}=f(\widetilde{P})$, and $\widetilde{P}$ represents the 3D coordinates of atoms. For instance, the $i$-th atom embedding in $l$-th layer can be denoted as $X_{i}^{l}$, where $i\in[0,N+1]$. $i=0$ denotes the virtual token representing the molecular representation, $N$ denotes the total number of atoms. $l\in[0,L]$ denotes the order number of the layer. ### III-C Input Representations #### III-C1 Absolute Position Encoding Following the previous works (Park et al. 2022; Ying et al. 2021), we integrate an absolute position encoding (APE) into graphormer [10] illustrated in Fig.2. The APE is the adjacent edge information of the node, which could distinguish some atoms with the same atom id easily in the input layer. Let $X^{0}$ denote the input tokens which are integrates with the APE, $X^{0}$ is calculated as: $X^{0}=X\oplus E,$ (1) where $\oplus$ is used to gather the information from neighbors. In specific, the $i$-th input token is calculated as: $X_{i}^{0}=MLP(\sum_{i=1}^{N+1}e_{i,j}).$ (2) For the 2D view, ${e}_{i,j}$ is the feature of original Chemical bonds. For the 3D view, we utilize radial basis functions (RBF) [23] as the transformed function $\widetilde{e}_{i,j}=RBF(d_{ij})$ to obtain the edge information, where ${d}_{ij}$ is the distance between atom i and atom j. This 3D embedding can encode diverse geometric factors ### III-D Backbone The backbone used is graphormer [10]. All tokens in the $l+1$ layer is updated by layer normalization (LN) and multi-head attention (MHA) on the tokens in the $l$ layer $X^{{}^{\prime}(l+1)}=MHA(LN(X^{l}))+X^{l},$ (3) then the feed-forward blocks (FFN) is applied to the tokens in the $l+1$ layer: $X^{l+1}=FFN(LN(X^{l+1}))+X^{{}^{\prime}(l+1)}.$ (4) Moreover, the final molecular representation $R$ is obtained from the virtual token of the last layer by an MLP: $R=MLP(X_{0}^{L}).$ (5) At last, for the supervised task, we use $L1$ loss. So the supervised term $\mathcal{L}_{3D}$ is defined as: $\mathcal{L}_{3D}(\widetilde{R},\widetilde{Y})=L_{1}Loss(\widetilde{R},\widetilde{Y}),$ (6) The supervised term $\mathcal{L}_{2D}$ is defined as: $\mathcal{L}_{2D}(R,Y)=L_{1}Loss({R},{Y}),$ (7) ### III-E Distillation #### III-E1 Challenges Usually, for most of the existing distillation methods, the total loss is defined as: $\mathcal{L}(R,Y,\widetilde{R})={L}_{2D}(R,Y)+\mathcal{W}*L_{1}Loss({X},\widetilde{X}),$ (8) ${L}_{2D}(R,Y)$ is the task-related loss, and $L_{1}Loss({X},\widetilde{X})$ is the distilling loss on all tokens and ${W}$ is the hyper-parameter. However, the virtual token ${X}_{0}$ is the target, which represents the global representation of the molecular, and the other tokens ${X}_{i}(i>0)$ are the atom representations that are auxiliary to the molecular representation. So the strategy of distillation should be divided into two modules: global molecular distillation and local atom distillation. In addition, We should select a proper weight to balance the global molecular distillation and local atom distillation. #### III-E2 Global molecular distilling The virtual token ${X}_{0}$ represents the molecular embedding and usually be used to predict molecular property, so distilling the virtual token embedding provides a global view of learning on 3D geometry structural knowledge. Considering that 2D graphs are seriously distinct from 3D conformers, we conduct the global molecular distilling from 3D to 2D on all layers which could learn more geometric information at different levels step by step. At last, the target is conducted by minimizing the following loss function given any pairs of $\widetilde{X}_{l,0}$ and $X_{l,0}$ in the training batch: $\mathcal{L}_{m}=\sum_{l=1}^{L}L_{1}Loss(\widetilde{X}_{0}^{l},X_{0}^{l})$ (9) #### III-E3 Local Atom Distilling The virtual token ${X}_{0}$ is the global view of molecular, and the token ${X}_{i}(i>0)$ is the local view of molecular. The virtual token is calculated from atom tokens by self-attention with the updating process. Considering the relationship between the virtual token and atom tokens, we use the local atom distilling to further boost the performance of the virtual token. One possible way is to calculate the similarity (distance) of each pairwise atom embedding, then the local atom calibration distillation loss is defined as: $\mathcal{L}_{a}=\sum_{l=1}^{L}\sum_{j=1}^{N+1}L_{1}Loss(\widetilde{X}_{j}^{l},X_{j}^{l})$ (10) Figure 2: The illustration of the backbone with absolute position encoding and relative position encoding. #### III-E4 Coordinating weight For boosting the performance of distilling with global and local information, we formulate a novel coordinating weight according to the number of molecular size to scale gradient magnitudes. The analyzing and reducing process from two perspectives is as follows. Coordinating the total loss Comparing the global molecular loss $\mathcal{L}_{m}$ and local atom loss $\mathcal{L}_{a}$, we can draw the conclusion that $\mathcal{L}_{a}$ is unstable due to the sum operation on all tokens. This means that when the molecular is very large(with big N), the final loss ${L}$ is exploding. So, we should add the function $f(N)$ according to the variable number of atoms to scale and calibrate the local atom loss $\mathcal{L}_{a}$. The formulation is as follows: $\displaystyle\mathcal{L}$ $\displaystyle=\mathcal{L}_{2D}+(\sum_{l=1}^{L}L_{1}Loss(\widetilde{X}_{0}^{l},X_{0}^{l})$ (11) $\displaystyle+f(N)*\sum_{l=1}^{L}\sum_{j=1}^{N+1}L_{1}Loss(\widetilde{X}_{j}^{l},X_{j}^{l}),$ $L$ is fixed and $N$ is variable which represents the number of atoms. In order to eliminate the effect of sum operation on all tokens, it’s intuitive to set $f(N)=\frac{1}{N}$. In addition, we couldn’t add $\frac{1}{N}$ to the molecular loss, which will dissipate the loss of the virtual token as N increases, so it’s necessary to separate global and local loss, and only scale the atom loss ${{L}_{a}}$. Generally, to make the total loss not explode, we should add a function $f(N)\propto{O}(\frac{1}{N})$ or $f(N)\propto{o}(\frac{1}{N})$. ${O}$ means “is of the same order as” and ${o}$ means “is ultimately smaller than”. Coordinating the molecular gradient Since the virtual token is the target embedding, its gradient is much important. So we will calculate the relationship between its gradient and N. To simplify the forward of transformer, $X_{i}^{l+1}$ in $(l+1)$ layer is formed as: $X_{i}^{l+1}=MHA(X^{l})\\\ \&=\sum_{j=0}^{N+1}({W}^{l}_{ji}*{X}^{l}_{j})$ (12) The ${W}^{l}_{ji}$ is the attention weight. In the back-propagation, the gradient of the virtual token in $l$th layer is as formulated: $\displaystyle{grad(X_{0}^{l})}$ $\displaystyle=\nabla_{X_{0}^{l}}{L}$ (13) $\displaystyle=\sum_{k=0}^{N+1}((\nabla_{X_{k}^{l+1}}{L})*(\nabla_{X_{0}^{l}}{X_{k}^{l+1}}))$ Finally, for transformer, the virtual token gradient $grad(X_{0}^{l+1})$ in ${l}$ layer can be described in follow: $\displaystyle{grad(X_{0}^{l})}$ $\displaystyle\propto{{O}({N}^{2})},$ (14) Different from multi-head attention in transformer, a simple GNN always use ${MLP}$ to calculate ${X}^{l+1}_{i}$ with ${X}^{l}_{0}$ . So $(\nabla_{X_{0}^{l}}{X_{k}^{l+1}})$ is constant, then the result is: $\displaystyle{grad(X_{0}^{l})}$ $\displaystyle\propto{{O}({N})},$ (15) For detailed proof, please refer to the Section1 in Supplementary Material. For transformer, the formula indicates that the gradient of virtual token $grad(X_{0}^{l})$ will be scaled by ${N}^{2}$, and N is the number of atoms. So scaled with $\frac{1}{N}$ couldn’t solve the unstable gradient of the virtual token. However, for a simple GNN, $\frac{1}{N}$ is enough. Conclusion To this end, for transformer, we proved the weight $f(N)=\frac{1}{{N}^{2}}$ is suitable to scale the losses and balance the gradient magnitudes of the global molecular information and local atom information. So the total objective function is formulated as: $\mathcal{L}=\mathcal{L}_{2D}+(\mathcal{L}_{m}+\frac{1}{{N}^{2}}*\mathcal{L}_{a}).$ (16) ## IV Experiments To evaluate our method and investigate the effectiveness of the proposed components, we carry out extensive experiments on two popular molecular property prediction tasks, especially in PCQM4Mv2 [9]. In this section, we introduce the detailed configuration in our experiments and present the detailed experimental results. ### IV-A Datasets #### IV-A1 PCQM4Mv2 dataset PCQM4Mv2 dataset [9] served as an ”ImageNet Large Scale Visual Recognition Challenge” in the field of graph ML is a quantum chemistry dataset originally curated under the PubChemQC project. It contains 3.8 million molecular graphs. The 3D structure for training molecules is calculated by DFT, which is equipped with only one conformer for each SMILES code. #### IV-A2 MolHIV dataset MolHIV dataset [24] is adopted from the MOLECULENET [25] with 41K molecular graphs for molecular property prediction. All the molecules are pre-processed using RDKIT [26]. Each graph represents a molecule, where nodes are atoms, and edges are chemical bonds. Input node features are 9-dimensional, containing atomic number and chirality, as well as other additional atom features such as formal charge and whether the atom is in the ring. Input edge features are 3-dimensional, containing bond type, bond stereochemistry as well as an additional bond feature indicating whether the bond is conjugated. ### IV-B Experimental Setup Our experiments are conducted on eight NVIDIA Tesla V100 GPUs. The implementation is based on PyTorch [27]. We use the Graphormer architecture [10] as the based model (L = 12, d = 768) for both 2D and 3D branches. Our CCMD method contains improved baseline with APE, global molecular distillation, local atom distillation, searched weight and coordinating weight, denoted as baseline, $\mathcal{L}_{m}$, $\mathcal{L}_{a}$, ${W}_{search}$ and ${W}_{coordinating}$ respectively. In order to prevent the explosion of a total loss, the local atom distillation loss is calculated by means denoted as $\overline{\mathcal{L}_{a}}$, where $\overline{\mathcal{L}_{a}}=\frac{1}{N}*\mathcal{L}_{a}$. So we set the coordinating weight as $\frac{1}{N}$ for experiments on graphormer and $1$ for experiments on GIN [28]. For the settings and metrics on the PCQM4Mv2 dataset, the previous methods use the 2D view for training and validation. We add the 3D structure calculated by DFT for training and distillation, and also evaluate on the validation with 2D view (the test labels are no longer available and results are given over the validation set). And split of dataset is followed by previous work [10]. We set the batch size as 512 with 70 epochs and use the Adam algorithm [29] with an initialized learning rate of 0.0002 and a momentum of 0.9. The results are evaluated in terms of Mean Absolute Error (MAE), and a lower MAE indicates better performance. For the settings and metrics on the MolHIV dataset, we add the 3D structure calculated by RDKIT for 3D view training and distillation. And we evaluate on the test dataset with 2D view. And split of dataset is followed by previous work [10]. We pretrain 3D and 2D branches on the PCQM4Mv1 dataset following the previous works [10]. When finetune on the MolHIV dataset, we set the batch size as 64 with 10 epochs and use Adam algorithm [29] with an initialized learning rate of 0.0002 and a momentum of 0.9. The results are evaluated in terms of Area Under the ROC Curve (AUC), and a higher AUC indicates better performance. ### IV-C Experimental Results #### IV-C1 The comparison on PCQM4Mv2 dataset Table I presents the results for molecular property prediction task on the OGB-LSC PCQM4Mv2 datasets [9]. What’s worth mentioning is that contrastive methods are hard applicable because the PCQM4Mv2 is equipped with only one conformer for each SMILES code. Compared with established methods, our model is currently the best on the PCQM4Mv2 leaderboard. As shown in Table I, our approach achieves the significant MAE improvement by 6% compared with the state-of-the-art method, indicating the effectiveness of our proposed method. Figure 3: The learning curves of 2D supervised performance and distilling performance during training process. TABLE I: The comparison with state-of-the-art methods on the validation set of PCQM4Mv2 dataset. Our approach outperforms the best model on average by 6%. Method | MAE ---|--- GCN [30] | 0.1379 GCN-VN[2, 30] | 0.1153 Graphormer [10] | 0.0864 EGT [31] | 0.0869 GRPE-Standard [32] | 0.0890 GPS [33] | 0.0858 Ours | 0.0809 #### IV-C2 The comparison on MOLHIV dataset We further investigate our proposed method on the MOLHIV dataset. The comparison results are shown in Table II. From Table II, we can see that our method outperforms other methods, achieving prominent improvement by 1.39% on AUC compared with baseline. In addition, the origin MOLHIV dataset is not attached with 3D information, so we use the rdkit to generate the 3D position as the 3D coordinates of atoms. It should be noted that the generated 3D coordinates with randomly selected 3D conformers are not matched the real coordinates, so the generated force fields are different from the real force fields. The weakness of the 3D teacher model limits the performance of the student. Our approach can still obtain improvement with biased conformers, outperforming state-of-the-art methods and software (e.g.rdkit) generated 3D coordinates-based results. These phenomenon demonstrate the effectiveness of our distilling framework. TABLE II: The results of our method compared with other methods on the test set of MOLHIV dataset. Method | AUC ---|--- DeeperGCN-FLAG[30] | 79.42 ± 1.20 PNA [30] | 79.05 ± 1.32 DGN [28] | 79.70 ± 0.97 PHC-GNN[30] | 79.34 ± 1.16 (baseline)Graphormer [10] | 80.51 ± 0.53 GraphMVP [13] | 77.0 GeomGCL[14] | 80.6 ± 0.009 EGT [31] | 80.60 ± 0.65 GRPE-Standard [32] | 81.39 ± 0.49 GPS [33] | 78.80 ± 0.0101 Ours(baseline+$\mathcal{L}_{m}$) | 81.4±0.11 Ours(baseline+$\overline{\mathcal{L}_{a}}$) | 79.0±0.01 Ours(baseline+$\mathcal{L}_{m}$ +($\mathcal{W}_{search}*\overline{\mathcal{L}_{a}})$) | 81.47±0.03 Ours(baseline+$\mathcal{L}_{m}$ +($\mathcal{W}_{coordinating}*\overline{\mathcal{L}_{a}})$) | 81.9±0.01 ### IV-D Ablation Study We investigated the impact of these components of our proposed method on the validation set of PCQM4Mv2 dataset, and results are shown in Table III. TABLE III: The ablation study of our method on the validation set of the PCQM4Mv2 dataset. L denotes the conduction on the last layer and ”all” denotes distilling with the features, which are not divided into global and local respectively. Our contributions have a significant impact on performance. Method | MAE ---|--- Graphormer [10] | 0.0864 baseline(reproduce+APE) | 0.0845 baseline+3D | 0.040 (L)baseline+all | 0.0853 (L)baseline+$\mathcal{L}_{m}$ | 0.0842 (L)baseline+$\overline{\mathcal{L}_{a}}$ | 0.0874 (L)baseline $+$ $\mathcal{L}_{m}$ $+$ $\mathcal{W}_{coordinating}$ $*$($\overline{\mathcal{L}_{a}}$) | 0.0837 baseline+$\mathcal{L}_{m}$ | 0.0822 baseline+$\overline{\mathcal{L}_{a}}$ | 0.087 baseline$+$ $\mathcal{L}_{m}$ $+$ $\mathcal{W}_{search}*$($\overline{\mathcal{L}_{a}}$) | 0.0818 baseline$+$ $\mathcal{L}_{m}$ $+$ $\mathcal{W}_{coordinating}*$($\overline{\mathcal{L}_{a}}$) | 0.0809 ##### The validation of absolute position encoding (APE) We train our graphormer with APE for 80 epochs as the baseline in Table III. Compared to the original graphormer training with 300 epochs, our baseline obtains a 1.52% improvement in the second row. ##### The validation of dividing local atom and global molecular We observed that distilling the feature encoded with virtual token and atom token in the fourth row in Table III generates a negative gain decreased by 0.94% compared with baseline. The phenomenon indicates binding virtual token and atom tokens damages the performance of the distillation. ##### The validation of global molecular Compared to the baseline model, the global molecular distillation $\mathcal{L}_{m}$ slightly increases the performance by distilling the virtual token embedding which is encoded with the structure information, reaching 0.36% improvement in the fifth row in Table III. ##### The validation of layers In order to expand the effect of distilling molecular representation, we conduct the distillation on all layers, which generates 2.72% improvement compared with conducting on the last layer. This result demonstrates that distilling on all layers superiors to distilling on the last layer. ##### The validation of local atoms In addition, we employ local atom loss $\mathcal{L}_{a}$ on all layers. We have observed one interesting phenomenon: when only leveraging the local atom information distillation, the network decreased by 3.43% in sixth row. This phenomenon demonstrates that distilling local atom information without correction will cause degradation problems. ##### The validation of search weight In the seventh row, the network further increases 3.2% compared with baseline by distilling local atom and global molecular representation with coordinating weight searched manually denoted as $\mathcal{W}_{search}$. The result indicates that local atoms with coordinating weight can boost the performance of global molecular. But the optimal coordinating weight is hard to search manually due to the variable molecular size. ##### The validation of coordinating weight Finally, when we incorporate the global molecular and local atoms with coordinating weight $\mathcal{W}_{coordinating}$ to balance the grads according to the number of atoms, our method in the eleventh row achieves the best result with 0.809 MAE which achieves large improvement by 4.26% compared with baseline. These results validate that our proposed coordination distillation with global molecular and local atom information is effective for molecular property prediction. ### IV-E Visionlization of effectiveness To understand and analyze the availability of our approach, we show the variation of attention weight in Fig. 4. We observed that the correlation between points is more prominent and notable compared with the baseline model. Compared with the baseline that all atoms only focus on the 2-hop nearest neighbors, both the global molecular token and local atom tokens could focus on almost all the atoms with CCMD. Thus, CCMD enriched the student model with more structured information. This phenomenon promulgates the significance of distillation. To reveal the significance of coordinating weight, we certificate the validation of coordinating distilling strategy in Fig. 5, which shows that coordinating weight is superior to any manually searched weight. To analyze the behavior of the model after adding the distillation loss term during training, we present the learning curves of 2D supervised performance and distilling performance in Fig. 3. We can observe that the convergence of MAE is more stable compared with only 2D supervised loss. Figure 4: The visionlization of student model according to the attention weight. We compare the attention of atom tokens in yellow dashed fields and virtual token in the blue dashed boxes. Figure 5: The comparison between coordinating weight and manually searched weights. The purple points are the results with manually searched weights, and the green point is the result with coordinating weight according to the number of atoms. ### IV-F Effectiveness with Other Backbone To evaluate the effectiveness of our approach, we change the gragh transformer backbone of 2D branch into GIN [28] architecture (L = 3, d = 768) on the PCQM4Mv2 dataset in Table IV. The 3D model based on gragh transformer is a heavy model which has 12 blocks and 57.65 MB parameters, and the 2D model based on GIN is a light GNN model which has 12 blocks and 1.923 MB parameters. GIN and gragh transformer are very different in structure, making 2D-3D distilling much more challenging, while our method in the fourth row still increases MAE by 3.36% compared with baseline in the first row. What’s more, the distillation with coordinating weight as $\frac{1}{N}$ in the fourth row increases 0.67% compared with coordinating weight as $\frac{1}{{N}^{2}}$ in the third row. This phenomenon indicates that coordinating weight with $\frac{1}{N}$ is enough to scale the gradients for the GNN network. These results in Table IV confirm the scalability and effectiveness of the distilling strategy. TABLE IV: The results with GIN backbone on the validation set of PCQM4Mv2 dataset. Method | MAE ---|--- baseline | 0.122 baseline+$\mathcal{L}_{m}$ | 0.1198 baseline+$\mathcal{L}_{m}$\+ $\frac{1}{N^{2}}*(\mathcal{L}_{a})$ | 0.1187 baseline+$\mathcal{L}_{m}$ \+ $\frac{1}{N}*\mathcal{L}_{a}$ | 0.1179 ## V Conclusion In this work, we exclusively propose a novel CCMD framework for molecular property prediction. In particular, we integrate global molecular token distillation and local atom information distillation with a novel coordinating weight to balance and scale the gradient magnitudes. In addition, we establish an absolute position encoding to integrate adjacent edge information. Experimental results on two popular molecular property prediction tasks demonstrate the superiority of our model. In the future, we will further experiment on the datasets from other areas to demonstrate the generalization ability of our method and explore more superior distilling strategy. ## References * [1] Franco Scarselli, Marco Gori, Ah Chung Tsoi, Markus Hagenbuchner, and Gabriele Monfardini. The graph neural network model. IEEE transactions on neural networks, 20(1):61–80, 2008. * [2] Justin Gilmer, Samuel S Schoenholz, Patrick F Riley, Oriol Vinyals, and George E Dahl. Neural message passing for quantum chemistry. In International conference on machine learning, pages 1263–1272. PMLR, 2017. * [3] Shengchao Liu, Mehmet F Demirel, and Yingyu Liang. N-gram graph: Simple unsupervised representation for graphs, with applications to molecules. Advances in neural information processing systems, 32, 2019. * [4] Minghao Xu, Hang Wang, Bingbing Ni, Hongyu Guo, and Jian Tang. Self-supervised graph-level representation learning with local and global structure. In International Conference on Machine Learning, pages 11548–11558. PMLR, 2021. * [5] Johannes Klicpera, Janek Groß, and Stephan Günnemann. Directional message passing for molecular graphs. arXiv preprint arXiv:2003.03123, 2020. * [6] Zhuoran Qiao, Matthew Welborn, Animashree Anandkumar, Frederick R Manby, and Thomas F Miller III. c. The Journal of chemical physics, 153(12):124111, 2020. * [7] Yi Liu, Limei Wang, Meng Liu, Xuan Zhang, Bora Oztekin, and Shuiwang Ji. Spherical message passing for 3d graph networks. arXiv preprint arXiv:2102.05013, 2021. * [8] Andrey A Toropov, Alla P Toropova, Dilya V Mukhamedzhanoval, and Ivan Gutman. Simplified molecular input line entry system (smiles) as an alternative for constructing quantitative structure-property relationships (qspr). 2005\. * [9] Weihua Hu, Matthias Fey, Hongyu Ren, Maho Nakata, Yuxiao Dong, and Jure Leskovec. Ogb-lsc: A large-scale challenge for machine learning on graphs. arXiv preprint arXiv:2103.09430, 2021. * [10] Chengxuan Ying, Tianle Cai, Shengjie Luo, Shuxin Zheng, Guolin Ke, Di He, Yanming Shen, and Tie-Yan Liu. Do transformers really perform badly for graph representation? Advances in Neural Information Processing Systems, 34:28877–28888, 2021. * [11] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017. * [12] Chengqiang Lu, Qi Liu, Chao Wang, Zhenya Huang, Peize Lin, and Lixin He. Molecular property prediction: A multilevel quantum interactions modeling perspective. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 1052–1060, 2019. * [13] Shengchao Liu, Hanchen Wang, Weiyang Liu, Joan Lasenby, Hongyu Guo, and Jian Tang. Pre-training molecular graph representation with 3d geometry. arXiv preprint arXiv:2110.07728, 2021. * [14] Shuangli Li, Jingbo Zhou, Tong Xu, Dejing Dou, and Hui Xiong. Geomgcl: Geometric graph contrastive learning for molecular property prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 4541–4549, 2022. * [15] Weihua Hu, Bowen Liu, Joseph Gomes, Marinka Zitnik, Percy Liang, Vijay Pande, and Jure Leskovec. Strategies for pre-training graph neural networks. arXiv preprint arXiv:1905.12265, 2019. * [16] Hannes Stärk, Dominique Beaini, Gabriele Corso, Prudencio Tossou, Christian Dallago, Stephan Günnemann, and Pietro Liò. 3d infomax improves gnns for molecular property prediction. In International Conference on Machine Learning, pages 20479–20502. PMLR, 2022. * [17] Xiaoqi Jiao, Yichun Yin, Lifeng Shang, Xin Jiang, Xiao Chen, Linlin Li, Fang Wang, and Qun Liu. Tinybert: Distilling bert for natural language understanding. arXiv preprint arXiv:1909.10351, 2019. * [18] Siqi Sun, Yu Cheng, Zhe Gan, and Jingjing Liu. Patient knowledge distillation for bert model compression. arXiv preprint arXiv:1908.09355, 2019. * [19] Wenhui Wang, Furu Wei, Li Dong, Hangbo Bao, Nan Yang, and Ming Zhou. Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. Advances in Neural Information Processing Systems, 33:5776–5788, 2020. * [20] Yonglong Tian, Dilip Krishnan, and Phillip Isola. Contrastive representation distillation. arXiv preprint arXiv:1910.10699, 2019. * [21] Jin Dong, Marc-Antoine Rondeau, and William L Hamilton. Distilling structured knowledge for text-based relational reasoning. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6782–6791, 2020. * [22] Wenhao Zhu, Ziyao Li, Lingsheng Cai, and Guojie Song. Stepping back to smiles transformers for fast molecular representation inference. arXiv preprint arXiv:2112.13305, 2021. * [23] Martin D Buhmann. Radial basis functions. Acta numerica, 9:1–38, 2000. * [24] Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 33:22118–22133, 2020. * [25] Zhenqin Wu, Bharath Ramsundar, Evan N Feinberg, Joseph Gomes, Caleb Geniesse, Aneesh S Pappu, Karl Leswing, and Vijay Pande. Moleculenet: a benchmark for molecular machine learning. Chemical science, 9(2):513–530, 2018. * [26] Greg Landrum et al. Rdkit: A software suite for cheminformatics, computational chemistry, and predictive modeling. Greg Landrum, 2013. * [27] A. Paszke, S. Gross, S. Chintala, G. Chanan, E. Yang, Z. Devito, Z. Lin, A. Desmaison, L. Antiga, and A. Lerer. Automatic differentiation in pytorch. 2017\. * [28] Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. How powerful are graph neural networks? arXiv preprint arXiv:1810.00826, 2018. * [29] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014. * [30] Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016. * [31] Md Shamim Hussain, Mohammed J Zaki, and Dharmashankar Subramanian. Global self-attention as a replacement for graph convolution. 2022\. * [32] Wonpyo Park, Woong-Gi Chang, Donggeon Lee, Juntae Kim, et al. Grpe: Relative positional encoding for graph transformer. In ICLR2022 Machine Learning for Drug Discovery, 2022. * [33] Ladislav Rampášek, Mikhail Galkin, Vijay Prakash Dwivedi, Anh Tuan Luu, Guy Wolf, and Dominique Beaini. Recipe for a general, powerful, scalable graph transformer. arXiv preprint arXiv:2205.12454, 2022. ## VI section 1 In this part, we will prove the relation of gradient on the virtual token and atom number ${N}$. For transformer, to simplify the proof, $X_{i}^{l+1}$ in $l$ layer is formed as: $X_{i}^{l+1}=MHA(X^{l})=\sum_{j=0}^{N+1}({W}^{l}_{ji}*{X}^{l}_{j})$ (17) The ${W}^{l}_{ji}$ is the attention weight, and ${W}^{l}_{ji}$ is formulated as: ${W}^{l}_{ji}=(\frac{e^{{X}^{l}_{i}*{X}^{l}_{j}}}{\sum_{j=0}^{N+1}(e^{{X}^{l}_{i}*{X}^{l}_{j})}})$ (18) So, $\displaystyle{grad(X_{0}^{l})}$ $\displaystyle=\nabla_{X_{0}^{l}}{L}$ (19) $\displaystyle=\sum_{k=0}^{N+1}((\nabla_{X_{k}^{l+1}}{L})*(\nabla_{X_{0}^{l}}{X_{k}^{l+1}}))$ $\displaystyle=\sum_{k=0}^{N+1}((\nabla_{X_{k}^{l+1}}{L})*(\nabla_{X_{0}^{l}}(\sum_{j=0}^{N+1}{W}_{jk}*{X}^{l}_{j}))$ $\displaystyle=\sum_{k=0}^{N+1}((\nabla_{X_{k}^{l+1}}{L})*(\nabla_{X_{0}^{l}}((\sum_{j=1}^{N+1}{W}_{jk}*{X}^{l}_{j})+{W}_{0k}*{X}^{l}_{0}))$ $\displaystyle=\sum_{k=0}^{N+1}((\nabla_{X_{k}^{l+1}}{L})*((\sum_{j=1}^{N+1}\nabla_{X_{0}^{l}}({{W}_{jk})*{X}^{l}_{j}})+{W}_{0k})$ $\displaystyle\propto(\sum_{k=0}^{N+1}(\nabla_{X_{k}^{l+1}}{L})*(\sum_{j=1}^{N+1}\nabla_{X_{0}^{l}}({{W}_{jk})*{X}^{l}_{j}})+{O}({N})$ $\displaystyle\propto{O}(N)*((\sum_{j=1}^{N+1}\nabla_{{X}_{0}^{l}}{{W}_{jk}})+1)$ Now, we calculate ${\nabla_{{X}_{0}}{{W}_{jk}}}$ of the l-th layer, for convenience, ${a}$ and ${b}$ denote constants independent of the variable N and ${X}_{0}$: $\displaystyle\nabla_{{X}_{0}}{{W}_{jk}}$ $\displaystyle=\nabla_{{X}_{0}}(\frac{e^{{X}_{j}*{X}_{k}}}{\sum_{j=0}^{N+1}e^{{X}_{j}*{X}_{k}}})$ (20) $\displaystyle\propto(\nabla_{{X}_{0}}(\frac{a}{\sum_{j=0}^{N+1}e^{{X}_{j}*{X}_{k}}}))$ $\displaystyle\propto(\nabla_{{X}_{0}}(\frac{a}{\sum_{j=1}^{N+1}e^{{X}_{j}*{X}_{k}}+e^{{X}_{0}*{X}_{k}}}))$ $\displaystyle\propto(\nabla_{{X}_{0}}(\frac{a}{b+e^{{X}_{0}*{X}_{k}}}))$ $\displaystyle\propto({O}(1))$ Finally, the virtual token gradient $grad(X_{0}^{l})$ in $l$ layer can be described in follow: $\displaystyle{grad(X_{0}^{l})}$ $\displaystyle\propto{{O}({N})}*\sum_{j=1}^{N+1}({O}(1))$ (21) $\displaystyle\propto{{O}({N}^{2})},$ For a simple GNN like GIN: $\displaystyle{X}^{l+1}_{i}$ $\displaystyle={W}^{l+1}_{0i}*{X}^{l}_{0}$ (22) So, $\displaystyle{grad(X_{0}^{l})}$ $\displaystyle=\nabla_{X_{0}^{l}}{L}$ (23) $\displaystyle=\sum_{k=0}^{N+1}((\nabla_{X_{k}^{l+1}}{L})*(\nabla_{X_{0}^{l}}{X_{k}^{l+1}}))$ $\displaystyle=\sum_{k=0}^{N+1}((\nabla_{X_{k}^{l+1}}{L})*({W}^{l+1}_{0k}))$ $\displaystyle\propto{O}(N)$
# Uniformizations of compact Sasakian manifolds Hisashi Kasuya Department of Mathematics, Graduate School of Science, Osaka University, Osaka, Japan<EMAIL_ADDRESS>and Natsuo Miyatake Institute of Mathematics for Industry, Kyushu University 744 Motooka, Fukuoka 819-0395, Japan<EMAIL_ADDRESS><EMAIL_ADDRESS> ###### Abstract. We give a criterion for compact Sasakian manifolds to be deformed to Sasakian manifolds which are locally isomorphic to circle bundles of anti-canonical bundles over Hermitian symmetric spaces as a Sasakian analogue of Simpson’s uniformization results related to variations of Hodge structure and Higgs bundles. ###### Key words and phrases: Sasakian manifold, Uniformization, Variation of Hodge structures, Higgs bundle ###### 2010 Mathematics Subject Classification: 14D07, 32Q30, 32L05, 53C25,53C07 This work was partially supported by JSPS KAKENHI Grant Number 19H01787. ## 1\. Introduction We are interested in classifications of compact Sasakian manifolds. Sasakian geometry is an odd-dimensional counterpart of Kähler geometry. However, in contrast to classifications of compact Kähler manifolds up to biholomorphism, the existence of a finite dimensional local moduli space of a compact Sasakian manifold is not given by deformation theory. Sasakian structures are usually defined by contact CR-structures $(T^{1,0}_{M},\eta)$ on odd-dimensional manifolds $M$ satisfying the “normality” condition. It is known that a space of locally non-isomorphic CR structures on a compact strongly pseudo-convex CR manifold can have arbitrary many parameters (see [BSW]). For classifying $3$-dimensional compact Sasakian manifolds, it is a great idea to classify Sasakian structures up to deformations of certain types. In [Bel1, Bel2], Belgun proves that any $3$-dimensional compact Sasakian manifold is a deformation of a standard one; locally isomorphic to one of canonical homogeneous (left-invariant) Sasakian structures on $S^{3}$, $\widetilde{SL}_{2}({\mathbb{R}})$, $Nil_{3}$ (see also [Ge]). Belgun’s result is an analogue of uniformizations of compact Riemannian surfaces. It is natural to expect that we can prove a Sasakian version of uniformization results on higher dimensional compact Kähler manifolds in terms of Belgun’s arguments. The most important part of uniformizations of compact Riemannian surfaces is to prove that every Riemannian surface of genus $\geq 2$ admits a metric of constant negative curvature. In [Hi], Hitchin reproves this statement as a corollary of the Kobayashi-Hitchin correspondence of Higgs bundles. In [Si1], Simpson gives a uniformization theorem of higher dimensional compact Kähler manifolds by generalizing Hitchin’s idea. Precisely, Simpson shows that the existence of a uniformizing variation of Hodge structure is a criterion for uniformizing a compact Kähler manifold by a Hermitian symmetric space. By using this, he shows that the stability of a certain Higgs bundle is a criterion for uniformizing a compact Kähler manifold by the unit ball. The main result of this paper is to give a Sasakian version of Simpson’s uniformization theorem. For a Hermitian symmetric space $D$ of non-compact type, we can define the homogeneous Sasakian structure $(T_{D}^{1,0},\eta_{D})$ on the circle bundle $S^{1}(\bigwedge^{m}T^{1,0}_{D})$ of the determinant line bundle of the holomorphic tangent bundle of $D$. We give a criterion for uniformizing compact Sasakian manifolds by the homogeneous Sasakian structure $(T_{D}^{1,0},\eta_{D})$. ###### Theorem 1.1. Let $(M,T^{1,0}_{M},\eta)$ be a compact Sasakian manifold. If $(M,T^{1,0}_{M},\eta)$ admits a uniformizing $G$-variations of Hodge structure $(G,\rho,h)$ and $c_{1,B}(T_{M})=-C[d\eta]$ for a positive constant $C$, then the universal covering $\widetilde{M}$ with the lifting of some $A^{1}_{B}$-deformation of $(T^{1,0}_{M},\eta)$ is almost isomorphic to $\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D})$ with the lifting of the homogeneous Sasakian structure $(T^{1,0}_{D},\eta_{D})$ where $\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D})$ is the universal covering of $S^{1}(\bigwedge^{m}T^{1,0}_{D})$. $c_{1,B}(T_{M})$ is the first basic Chern class of $(M,T^{1,0}_{M},\eta)$ which is also very important for studying Sasaki-Einstein geometry (see [BG]). Note that the condition $c_{1,B}(T_{M})=C[d\eta]$ for a positive constant $C$ is a criterion for the existence of a Sasaki-Einstein metric on a compact Sasakian manifold in contrast to our result. An $A^{1}_{B}$-deformation is a slight modification of a second type deformation in the sense of Belgun [Bel2] (see also a deformation of type II in [BG, Definition 7.5.9]). An almost isomorphism is a diffeomorphism commuting with almost contact structures associated with Sasakian manifolds equivalently Sasakian isomorphism up to rescalings. Combining this criterion and the Sasakian version of Simpson’s correspondence between flat bundles and Higgs bundles ([Si1, Si]) proved in [BK, BK2], we obtain a criterion for uniformizing compact Sasakian manifolds in terms of stablities of Higgs bundles. For a Sasakian manifold $(M,T^{1,0}_{M},\eta)$, we have the canonical transverse holomorphic foliation ${\mathcal{F}}_{\xi}$ and hence we can define holomorphic structures and Higgs bundle structures on “basic” vector bundles over $(M,{\mathcal{F}}_{\xi})$. We have a canonical Basic Higgs bundle structure $(E_{M},\,\theta_{M})$ on the vector bundle $E_{M}=T^{1,0}_{M}\oplus{\mathbb{C}}_{M}$ where $T^{1,0}_{M}$ is the holomorphic tangent bundle associated with a CR structure and ${\mathbb{C}}_{M}$ is the trivial line bundle. ###### Theorem 1.2. We assume: 1. (1) The basic Higgs bundle $(E_{M},\,\theta_{M})$ is stable. 2. (2) The equality $\int_{M}\left(2c_{2,B}(T^{1,0}_{M})-\frac{n}{n+1}c_{1,B}(T^{1,0}_{M})^{2}\right)\wedge(d\eta)^{n-2}\wedge\eta\,=\,0$ holds. 3. (3) $c_{1,B}(T^{1,0}_{M})=-C[d\eta]$ for some positive constant $C$. Then the universal covering $\widetilde{M}$ with the lifting of some $A^{1}_{B}$-deformation of $(T^{1,0}_{M},\eta)$ is almost isomorphic to the $\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D})$ with the lifting of the homogeneous Sasakian structure $(T^{1,0}_{D},\eta_{D})$ for $D=SU(n,1)/S(U(n)\times U(1))$. This result is based on the existence of Hermitian-Einstein metrics on stable basic Higgs bundles over compact Sasakian manifolds. In [Zh], by using the existence of Sasaki-Einstein metrics, the result of uniformization by the odd- dimensional sphere is proved. As a corollary of Theorem 1.2, we obtain an analogue of the existence of a metric of constant negative curvature on a Riemannian surface of genus $\geq 2$. ###### Corollary 1.3. Let $(M,T^{1,0}_{M},\eta)$ be a compact $3$-dimensional Sasakian manifold with $c_{1,B}(T^{1,0})=-C[d\eta]$. Then the universal covering $\widetilde{M}$ with the lifting of some $A^{1}_{B}$-deformation of $(T^{1,0}_{M},\eta)$ is almost isomorphic to $\widetilde{PU}(1,1)$ with a left-invariant Sasakian structure. We have $\widetilde{PU}(1,1)=\widetilde{SL}_{2}({\mathbb{R}})$. This result is a part of Belgun’s classification in [Bel1, Bel2]. Belgun uses the non-Kähler complex structure on $M\times S^{1}$ for a compact Sasakian manifold $M$. On the other hand, we do not use such a complex structure. Acknowledgement. We are grateful to Professor Takuro Mochizuki for his valuable comments on the earlier works [BK, BK2] which led to the beginning of this project. ## 2\. Sasakian Geometry ### 2.1. Sasakian structures on CR manifolds Let $M$ be a $(2n+1)$-dimensional real smooth manifold. A CR-structure on $M$ is an $n$-dimensional complex involutive sub-bundle $T^{1,0}_{M}$ of the complexified tangent bundle $TM_{{\mathbb{C}}}\,=\,TM\otimes_{\mathbb{R}}{{\mathbb{C}}}$ such that $T^{1,0}_{M}\cap T^{0,1}_{M}=\\{0\\}$ where $T^{0,1}_{M}=\overline{T^{1,0}_{M}}$. Define the real $2n$-dimensional sub- bundle $S=TM\cap(T^{1,0}_{M}\oplus T^{0,1}_{M})\subset TM$. We have the almost complex structure $I:S\to S$ associated with the decomposition $S_{{\mathbb{C}}}=T^{1,0}_{M}\oplus T^{0,1}_{M}$. A strongly pseudo-convex CR structure on $M$ is a pair $(T^{1,0}_{M},\eta)$ of a CR structure $T^{1,0}_{M}$ and a contact $1$-form $\eta$ such that $\ker\eta=S$ and the bilinear form defined by $L_{\eta}(X,Y)=d\eta(X,IY)$ on $S$ is a Hermitian metric on $(S,I)$. Take the Reeb vector field $\xi$ associated with the contact $1$-form $\eta$. $(T^{1,0}_{M},\eta)$ is Sasakian if for any smooth section $X$ of $T^{1,0}_{M}$ the Lie bracket $[\xi,X]$ is also a smooth section of $T^{1,0}_{M}$. We consider the $1$-dimensional foliation ${\mathcal{F}}_{\xi}$ on $M$ generated by $\xi$. If $(T^{1,0}_{M},\eta)$ is Sasakian, then $T^{1,0}_{M}$ defines a transverse holomorphic structure on ${\mathcal{F}}_{\xi}$ and $d\eta$ is a transverse Kähler form on ${\mathcal{F}}_{\xi}$. We say that a Sasakian structure is quasi-regular if every leaf of ${\mathcal{F}}_{\xi}$ is closed. If $M$ is compact and a Sasakian structure on $M$ is quasi-regular, then the flow of $\xi$ induces an action $S^{1}\times M\to M$ (see [Wa]). Let $(M,T^{1,0}_{M},\eta)$ be a Sasakian manifold. A differential form $\omega$ on $M$ is called basic if the equations $i_{\xi}\omega=0=i_{\xi}d\omega$ hold. We denote by $A^{\ast}_{B}(M)$ the subspace of basic forms in the de Rham complex $A^{\ast}(M)$. Then $A^{\ast}_{B}(M)$ is a sub-complex of the de Rham complex $A^{\ast}(M)$. Denote by $H_{B}^{\ast}(M)$ the cohomology of the basic de Rham complex $A^{\ast}_{B}(M)$. We note that $d\eta\in A^{2}_{B}(M)$ and $[d\eta]\not=0\in H_{B}^{2}(M)$ if $M$ is compact. We have the bigrading $A^{r}_{B}(M)_{{\mathbb{C}}}=\bigoplus_{p+q=r}A^{p,q}_{B}(M)$ as well as the decomposition of the exterior differential $d_{|A^{r}_{B}(M)_{{\mathbb{C}}}}=\partial_{B}+\overline{\partial}_{B}$ on $A^{r}_{B}(M)_{{\mathbb{C}}}$, so that $\partial_{B}:A^{p,q}_{B}(M)\to A^{p+1,q}_{B}(M)\ \text{ and }\ \overline{\partial}_{B}:A^{p,q}_{B}(M)\to A^{p,q+1}_{B}(M)\,.$ We note that $d\eta\in A^{1,1}_{B}(M)$. For a strongly pseudo-convex CR- manifold $(M,T^{1,0}_{M},\eta)$ there exists a unique affine connection $\nabla^{TW}$ on $TM$ such that the following conditions hold ([Ta, We]): 1. (1) $S$ is parallel with respect to $\nabla^{TW}$. 2. (2) $\nabla^{TW}I\,=\nabla^{TW}d\eta=\nabla^{TW}\eta\,=\nabla^{TW}\xi=\,0$. 3. (3) The torsion $T^{TW}$ of the affine connection $\nabla^{TW}$ satisfies the equation $T^{TW}(X,\,Y)\,=\,-d\eta(X,\,Y)\xi$ for all $X,\,Y\,\in\,S_{x}$ and $x\,\in\,M$. This affine connection $\nabla^{TW}$ is called the Tanaka–Webster connection. It is known that $(T^{1,0},\eta)$ is a Sasakian manifold if and only if $T^{TW}(\xi,\,v)\,=\,0$ for all $v\,\in\,TM$. Define the homomorphism $\Phi_{\xi}:TM\to TM$ by the extension of $I:S\to S$ satisfying $\Phi_{\xi}(\xi)=0$. Then, $\Phi^{2}_{\xi}=-{\rm Id}+\xi\otimes\eta$. We call $\Phi_{\xi}$ the almost contact structure associated with a Sasakian structure $(T^{1,0}_{M},\eta)$. We define the Riemannian metric $g_{\eta}(X,Y)=\eta(X)\eta(Y)+d\eta(X,\Phi_{\xi}Y)$. We call $g_{\eta}$ the Sasakian metric associated with a Sasakian structure $(T^{1,0}_{M},\eta)$. Let $(N,J)$ be a complex manifold. A smooth map $f:M\to N$ is a holomorphic if $df\circ\Phi_{\xi}=J\circ df$. A holomorphic map $f:M\to N$ satisfies $df(\xi)=0$. We note that the holomorphic condition of a smooth map $f:M\to N$ depends only on the transverse holomorphic foliation ${\mathcal{F}}_{\xi}$. A smooth function $f:M\to{\mathbb{C}}$ is holomorphic if and only if $f\in A_{B}^{0}(M)$ and $\bar{\partial}_{B}f=0$. ### 2.2. Morphisms, equivalences and deformations Let $(M_{1},T^{1,0}_{M_{1}},\eta_{1})$ and $(M_{2},T^{1,0}_{M_{2}},\eta_{2})$ be two Sasakian manifolds. A smooth map $f:M_{1}\to M_{2}$ is a CR map if $df(T^{1,0}_{M_{1}})\subset T^{1,0}_{M_{2}}$. a CR map $f:M_{1}\to M_{2}$ satisfies $f^{\ast}\eta_{2}=g(x)\eta_{1}$ for some function $g(x)$ on $M_{1}$. By the standard arguments (see [BaK, Lemma 2.3]), we have: ###### Lemma 2.1. Let $f:M_{1}\to M_{2}$ be a diffeomorphism. Then the following two conditions are equivalent: * • $f$ is a CR map and $f^{\ast}\eta_{2}=\eta_{1}$. * • $f$ is isometry between $(M_{1},g_{\eta_{1}})$ and $(M_{2},g_{\eta_{2}})$ and $df(\xi_{1})=\xi_{2}$. A diffeomorphism satisfying the equivalent condition in this lemma is a Sasakian isomorphism. A smooth map $f:M_{1}\to M_{2}$ is a almost contact map if $df\circ\Phi_{\xi_{1}}=\Phi_{\xi_{2}}\circ df$. A almost contact map is a CR- map but the converse is not true. We can easily check that an almost contact map $f:M_{1}\to M_{2}$ satisfies $f^{\ast}\eta_{2}=c\eta_{1}$ and $df(\xi_{1x})=c\xi_{2f(x)}$ for some constant $c$. It is sufficient to check the above function $g(x)$ is constant. By the basicness ${\mathcal{L}}_{\xi_{1}}d\eta_{1}={\mathcal{L}}_{\xi_{2}}d\eta_{2}=0$ and $df(\xi_{1x})=c(x)\xi_{2f(x)}$, we have ${\mathcal{L}}_{\xi_{1}}(g)\eta_{1}={\mathcal{L}}_{\xi_{1}}(f^{\ast}\eta_{2})=(i_{\xi_{1}}d+di_{\xi_{1}})(f^{\ast}\eta_{2})=di_{\xi_{1}}f^{\ast}\eta_{2}=dg$ and so $dg=0$ on the sub-bundle $S={\rm ker}\eta_{1}\subset TM$. It is sufficient to prove $dg(\xi_{1x})=0$ for any $x\in M$. Since $d\eta$ is non- degenerate on $S$, we can take local sections $X,Y$ of $S$ such that $d\eta_{1}(X_{x},Y_{x})\not=0$. By the property of the Tanaka-Webster connection $\nabla^{TW}$, we have $\nabla^{TW}_{X}Y-\nabla^{TW}_{Y}X-[X,Y]=-d\eta(X,Y)\xi.$ Since $\nabla^{TW}_{X}Y-\nabla^{TW}_{Y}X$ is a local section of $S$, $-d\eta_{1}(X,Y)dg(\xi_{1})=dg(\nabla^{TW}_{X}Y-\nabla^{TW}_{Y}X)-X(Y(g))+Y(X(g))=0.$ This implies $dg(\xi_{1x})=0$. Let $(M,T^{1,0}_{M},\eta)$ be a Sasakian manifold. A rescaling of $(T^{1,0}_{M},\eta)$ is a Sasakian structure $(T^{1,0}_{M\tau},R\eta)$ for a real number $R>0$. The Reeb vector field of $R\eta$ is $\frac{1}{R}\xi$. A $A^{1}_{B}$-deformation of $(T^{1,0}_{M},\eta)$ is a Sasakian structure $(T^{1,0}_{M\tau},\eta^{\tau})$ such that $\eta^{\tau}=\eta+\tau$ for $\tau\in A^{1}_{B}$ and $T^{1,0}_{M\tau}=\\{X-\tau(X)\xi|X\in T^{1,0}_{M}\\}$. The Reeb vector field of $\eta^{\tau}$ is $\xi$ and the transverse holomorphic structure on ${\mathcal{F}}_{\xi}$ induced by $T^{1,0}_{M\tau}$ is same as the one induced by $T^{1,0}_{M}$. By the above argument on almost contact maps and the standard arguments (see [BG, Proposition 8.1.1]), we have: ###### Lemma 2.2. Let $f:M_{1}\to M_{2}$ be a diffeomorphism. Then the following two conditions are equivalent: * • $f$ is a Sasakian isomorphism up to rescalings. * • $f$ is an almost contact map. A diffeomorphism satisfying the equivalent condition in this lemma is an almost isomorphism. Obviously, an almost isomorphism is a CR-diffeomorphism. ## 3\. Basic vector bundles over Sasakian manifolds ### 3.1. Basic vector bundles We define structures on smooth vector bundles over Sasakian manifolds $(M,T_{M}^{1,0},\eta)$ related to ${\mathcal{F}}_{\xi}$ in terms of Rawnsley’s partial flat connections in [Ra]. A structure of basic vector bundle on a $C^{\infty}$ vector bundle $E$ is a linear operator $\nabla_{\xi}\,:\,{\mathcal{C}}^{\infty}(E)\,\longrightarrow\,{\mathcal{C}}^{\infty}(E)$ such that $\nabla_{\xi}(fs)\,=\,f\nabla_{\xi}s+\xi(f)s.$ Consider the flow ${\mathbb{R}}\times M\to M$ generated by $\xi$. Then if we have a lifting action ${\mathbb{R}}\times E\ni(t,v)\mapsto\phi_{t}(v)\in E$ of the flow ${\mathbb{R}}\times M\to M$ on a smooth vector bundle $E$, then we define the basic bundle structure on $E$ by $\nabla_{\xi}(s)_{x}=\frac{d}{dt}_{|t=0}\phi_{t}(v)_{x}.$ A differential form $\omega\,\in\,A^{\ast}(M,\,E)$ with values in $E$ is called basic if the equations $i_{\xi}\omega=0=\nabla_{\xi}\omega$ hold where we extend $\nabla_{\xi}:A^{\ast}(M,\,E)\to A^{\ast}(M,\,E)$. Let $A^{\ast}_{B}(M,\,E)\,\subset\,A^{\ast}(M,\,E)$ denote the subspace of basic forms in the space $A^{\ast}(M,\,E)$ of differential forms with values in $E$. A Hermitian metric on $E$ is called basic if it is $\nabla_{\xi}$-invariant. Unlike the usual Hermitian metric, a basic vector bundle may not admits a basic Hermitian metric in general. For a connection operator $\nabla:A^{\ast}(M,\,E)\to A^{\ast+1}(M,\,E)$ such that the covariant derivative of $\nabla$ along $\xi$ is $\nabla_{\xi}$, the curvature $R^{\nabla}\in A^{2}(M,\,{\rm End}(E))$ is basic i.e. $R^{\nabla}\in A^{2}_{B}(M,\,{\rm End}(E))$. For the Chern forms $c_{i}(E,\nabla)\in A^{2i}(M)$ associated with $\nabla$, we have $c_{i}(E,\nabla)\in A_{B}^{2i}(M)$ and we define the basic Chern classes $c_{i,B}(E)\in H^{2i}_{B}(M)$ by the cohomology classes of $c_{i}(E,\nabla)$. A structure of basic holomorphic bundle on a $C^{\infty}$ vector bundle $E$ is a linear differential operator $\nabla^{\prime\prime}\,:\,{\mathcal{C}}^{\infty}(E)\,\longrightarrow\,{\mathcal{C}}^{\infty}(E\otimes(\langle\xi\rangle\oplus T^{0,1})^{\ast})$ such that * • for any $X\in\langle\xi\rangle\oplus T^{0,1}$, and any smooth function $f$ on $M$, the equation $\nabla^{\prime\prime}_{X}(fs)\,=\,f\nabla^{\prime\prime}_{X}s+X(f)s$ holds for all smooth sections $s$ of $E$, and * • if we extend $\nabla^{\prime\prime}$ to $\nabla^{\prime\prime}\,:\,{\mathcal{C}}^{\infty}(E\otimes\bigwedge^{k}(\langle\xi\rangle\oplus T^{0,1})^{\ast})\,\longrightarrow\,{\mathcal{C}}^{\infty}(E\otimes\bigwedge^{k+1}(\langle\xi\rangle\oplus T^{0,1})^{\ast})$, then $\nabla^{\prime\prime}\circ\nabla^{\prime\prime}\,=\,0$. A basic holomorphic vector bundle $E$ has a canonical basic bundle structure corresponding to the derivative $\nabla^{\prime\prime}_{\xi}$. $\nabla^{\prime\prime}$ defines the linear operator $\bar{\partial}_{E}:A^{p,q}_{B}(M,\,E)\to A^{p,q+1}_{B}(M,\,E)$ so that $\bar{\partial}_{E}(f\omega)=\bar{\partial}_{B}f\wedge\omega+f\bar{\partial}_{E}(\omega)$ for $f\in A^{0}_{B}(M),\omega\in A^{p,q}_{B}(M,\,E)$. If a basic holomorphic vector bundle $E$ admits a basic Hermitian metric $h$, as complex case, we have a unique unitary connection $\nabla^{h}$ such that for any $X\in\langle\xi\rangle\oplus T^{0,1}$, $\nabla^{h}_{X}=\nabla^{\prime\prime}_{X}$. ###### Example 3.1. Consider the ${\mathcal{C}}^{\infty}$-trivial complex line bundle $E\,=\,M\times{\mathbb{C}}\,\longrightarrow\,M$. For any $C\in{\mathbb{R}}$, we define the connection $\nabla^{C}\,=\,d-2\pi\sqrt{-1}C\eta$ on $E$. Then, the curvature of $\nabla^{C}$ is $Cd\eta$. Since $d\eta\,\in\,A^{1,1}_{B}(M)$, this $\nabla^{C}$ induces a structure of a basic holomorphic bundle. Consequently, we have a non-trivial holomorphic vector bundle structure $E_{C}$ on $E$ that depends on $C\,\in\,{\mathbb{C}}$. The basic cohomology class of $Cd\eta$ is the basic first Chern class of the basic vector bundle $E_{C}$. Thus $\\{E_{C}\\}_{C\in{\mathbb{R}}}$ is a family of basic vector bundles such that $E_{C}\,\not\cong\,E_{C^{\prime}}$ for every $C\,\not=\,C^{\prime}$. The standard constant Hermitian metric $h$ on the ${\mathcal{C}}^{\infty}$ trivial line bundle $E\,=\,M\times{\mathbb{C}}$ is basic on $E_{C}$. The connection $\nabla^{C}$ is unitary for $h$, and hence $\nabla^{C}$ is the canonical connection for the basic Hermitian metric $h$. Consider the flow ${\mathbb{R}}\times M\to M$ generated by $\xi$. Define the lifting action ${\mathbb{R}}\times M\times{\mathbb{C}}\ni(t,x,v)\mapsto(x_{t},e^{-2\pi C\sqrt{-1}t}v)\in E$ of the flow ${\mathbb{R}}\times M\to M$. Then this induces the basic bundle structure $\nabla^{C}_{\xi}$ of $E_{C}$. ###### Example 3.2. The Tanaka-Webster connection $\nabla^{TW}$ defines a structure of a basic holomorphic bundle on $T^{1,0}$. Since we have $\nabla^{TW}_{\xi}(X)=[\xi,X]$, the basic bundle structure $\nabla^{TW}_{\xi}$ is given by the action ${\mathbb{R}}\times T^{1,0}X\to T^{1,0}X$ which is the differential of the action ${\mathbb{R}}\times M\to M$ generated by the flow of $\xi$. ### 3.2. Equivariant cohomology Assume that $M$ is compact and a Sasakian structure on $M$ is quasi-regular. Consider the equivariant cohomology $H^{\ast}_{S^{1}}(M,R)=H^{\ast}(M\times_{S^{1}}ES^{1},R)$ for the action $S^{1}\times M\to M$ with coefficients in a commutative ring $R$. $H^{\ast}_{S^{1}}(M,{\mathbb{R}})$ is the cohomology of the Cartan model $A^{\ast}(M)^{S^{1}}\otimes{\mathbb{R}}[u]$ with the differential $d_{S^{1}}$ so that $d_{S^{1}}u=0$, $d_{S^{1}}\omega=d\omega-i_{\xi}\omega u$ where $u$ is of degree $2$. The natural inclusion $A^{\ast}_{B}(M)\to A^{\ast}(M)^{S^{1}}\otimes{\mathbb{R}}[u]$ is a cochain complex homomorphism. Since the action $S^{1}\times M\to M$ is locally free, the induced map $H_{B}^{\ast}(M)\to H^{\ast}_{S^{1}}(M,{\mathbb{R}})$ is an isomorphism ([GS, Section 5]). We use the result in [MuR] which says that equivariant first Chern classes in $H^{\ast}_{S^{1}}(M,{\mathbb{Z}})$ classify lifts of the action $S^{1}\times M\to M$ on line bundles. For our case, basic bundle structures $\nabla_{\xi}$ on line bundles can be considered as infinitesimal lifts of the action $S^{1}\times M\to M$. Let $L_{1}\to M,L_{2}\to M$ be line bundles admitting lifts of the action $S^{1}\times M\to M$ and consider them as basic line bundles. Assume that we have basic Hermitian metrics $h_{1},h_{2}$ and basic unitary connections $\nabla_{1},\nabla_{2}$ on $L_{1}$ and $L_{2}$ respectively. We note that the equivariant first Chern classes in $H^{\ast}_{S^{1}}(M,{\mathbb{R}})$ are images of the basic first Chern classes via the canonical map $H_{B}^{\ast}(M)\to H^{\ast}_{S^{1}}(M,{\mathbb{R}})$. Since $H_{B}^{\ast}(M)\to H^{\ast}_{S^{1}}(M,{\mathbb{R}})$ is an isomorphism, we can say that if $c_{1,B}(L_{1})=c_{1,B}(L_{2})$, then $L_{1}^{l}$ and $L_{2}^{l}$ are equivariantly isomorphic for some positive integer $l$. ###### Proposition 3.3. If $c_{1,B}(T^{1,0}_{M})=-C[d\eta]$ for some positive constant $C$, then for some positive integer $l$, the $l$-th power of the anti-canonical bundle $\bigwedge^{n}T^{1,0}_{M}$ with the $S^{1}$-action given by the differential of the action $S^{1}\times M\to M$ is equivariantly isomorphic to $E_{-lC}$. ###### Remark 3.4. If $c_{1,B}(T^{1,0}_{M})=-C[d\eta]$ for a positive constant $C$, then the Sasakian structure $(T^{1,0}_{M},\eta)$ is quasi-regular ([BG, Theorem 8.1.14]). ## 4\. Variations of Hodge structure and uniformizations ### 4.1. Variations of Hodge structure over Sasakian manifolds Let $G$ be a connected semi-simple Lie group. Assume that $G$ has the finite center and no compact factor. A Hodge structure on $G$ is a real Hodge structure $\mathfrak{g}_{{\mathbb{C}}}=\bigoplus_{p\in{\mathbb{Z}}}\mathfrak{g}^{p,-p}$ of weight $0$ on the Lie algebra $\mathfrak{g}$ of $G$ such that the Lie bracket on $\mathfrak{g}$ is a morphism of real Hodge structures and $-B_{\mathfrak{g}}$ is a polarization where $B_{\mathfrak{g}}$ is the Killing form on $\mathfrak{g}$. Define the subalgebra $\mathfrak{v}=\mathfrak{g}^{0,0}\cup\mathfrak{g}\subset\mathfrak{g}$ and the subgroup $V=\exp(\mathfrak{v})\subset G$. Then $V$ is compact. The subalgebra $\bigoplus_{p>0}\mathfrak{g}^{p.-p}\subset\mathfrak{g}_{{\mathbb{C}}}$ defines a complex structure on $D=G/V$. The complex manifold $D$ is called the period domain. Define the subalgebra $\mathfrak{p}=\bigoplus_{p\geq 0}\mathfrak{g}^{p.-p}$ and the subgroup $P=\exp(\mathfrak{p})\subset G_{{\mathbb{C}}}$ in the complexification $G_{{\mathbb{C}}}$ of $G$. Then $\mathfrak{p}$ is a parabolic subalgebra in $\mathfrak{g}_{{\mathbb{C}}}$ and $G/V$ is a dual manifold of $G/P$ in the sense of [GrSc]. The holomorphic tangent bundle $T^{1,0}_{D}$ is a homogeneous vector bundle given by $(G\times\bigoplus_{p>0}\mathfrak{g}^{p.-p})/V$ and hence we have the decomposition $T^{1,0}_{D}=\bigoplus_{p>0}E^{p,-p}$. Consider the connection form $\lambda$ on the anti-canonical bundle $\bigwedge^{m}T^{1,0}_{D}$ given by the Maurer-Cartan form where $m=\dim_{{\mathbb{C}}}D$. Let $\eta_{D}=-\lambda$. Then $d\eta$ is a non-degenerate $2$-form on $D$ which is positive on $\bigoplus_{p\,odd}E^{p,-p}$ and negative on $\bigoplus_{p\,even}E^{p,-p}$ see [GrSc, (4.23)]. Let $(M,T^{1,0}_{M},\eta)$ be a Sasakian manifold. Denote by $\widetilde{M}$ the universal cover of $M$ at $x\in M$ and take the lifting Sasakian structure $(T^{1,0}_{\tilde{M}},\tilde{\eta})$. A $G$-variations of Hodge structure is $(G,\rho,h)$ such that: 1. (1) $G$ is a connected semi-simple Lie group equipped with a Hodge structure. 2. (2) $\rho:\pi_{1}(M,x)\to G$ is a representation. 3. (3) $h:\widetilde{M}\to D$ is a $\rho$-equivariant holomorphic map satisfying $dh(T^{1,0}_{\widetilde{M}})\subset E^{1,-1}$. ### 4.2. Uniformizing Variations of Hodge structure A Hodge structure on $G$ is Hermitian if $\mathfrak{g}_{{\mathbb{C}}}=\mathfrak{g}^{1,-1}\oplus\mathfrak{g}^{0,0}\oplus\mathfrak{g}^{-1,1}$. In this case, $D$ is a Hermitian symmetric manifold of non-compact type and $d\eta_{D}$ is an invariant Kähler form on $D$. Regarding $\eta$ as a contact $1$-form, $(T^{1,0}_{D},\eta_{D})$ is a $G$-invariant Sasakian structure on the circle bundle $S^{1}(\bigwedge^{m}T^{1,0}_{D})$. We have $S^{1}(\bigwedge^{m}T^{1,0}_{D})=(G\times S^{1})/V=G/V^{\prime}$ where $V^{\prime}$ is the kernel of the non-trivial representation $V\to GL(\bigwedge^{m}\mathfrak{g}^{1,-1})$. $S^{1}(\bigwedge^{m}T^{1,0}_{D})$ is also $G$-homogeneous and the Sasakian structure $(T^{1,0}_{D},\eta_{D})$ is homogeneous We can take the Reeb vector field $\xi_{D}$ as a fundamental vector field of the action $S^{1}\times(G\times S^{1})/V\ni(a,[g,t])\mapsto[g,a^{-1}t]\in(G\times S^{1})/V$. Since $D$ is contractible, by trivializing $S^{1}(\bigwedge^{m}T^{1,0}_{D})\cong S^{1}\times D$, we have the universal covering $\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D})\to S^{1}(\bigwedge^{m}T^{1,0}_{D})$ with the infinite cyclic covering group. Let $\Gamma\subset G$ be a discrete subgroup which acts freely and cocompactly on $S^{1}(\bigwedge^{m}T^{1,0}_{D})=G/V^{\prime}$. We consider the compact Sasakian manifold $(\Gamma\backslash S^{1}(\bigwedge^{m}T^{1,0}_{D}),T^{1,0}_{D}\eta_{D})$. We observe that: * • By the construction, we have $c_{1,B}(M)=-[d\eta_{D}]$. * • Defining $h:\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D})\to D$ by the composition of the covering $\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D})\to S^{1}(\bigwedge^{m}T^{1,0}_{D})$ and the projection $S^{1}(\bigwedge^{m}T^{1,0}_{D})\to D$ and $\rho:\pi_{1}(\Gamma\backslash S^{1}(\bigwedge^{m}T^{1,0}_{D}))\to G$ by the composition the quotient $\pi_{1}(\Gamma\backslash S^{1}(\bigwedge^{m}T^{1,0}_{D}))\to\Gamma$ by the covering group and the inclusion $\Gamma\to G$, we have the $G$-variation of Hodge structure $(G,\rho,h)$. The main purpose of this section is to prove that these properties uniformize compact Sasakian manifolds by the homogeneous Sasakian structure Let $(M,T^{1,0}_{M},\eta)$ be a Sasakian manifold. A $G$-variation of Hodge structure $(G,\rho,h)$ is uniformizing if a Hodge structure on $G$ is Hermitian and the differential $dh:T^{1,0}_{\widetilde{M}}\to T^{1,0}_{D}$ is an isomorphism at every fiber. ###### Theorem 4.1. Let $(M,T^{1,0}_{M},\eta)$ be a compact Sasakian manifold. If $(M,T^{1,0}_{M},\eta)$ admits a uniformizing $G$-variations of Hodge structure $(G,\rho,h)$ and $c_{1,B}(T_{M}^{1,0})=-C[d\eta]$ for a positive constant $C$, then the universal covering $\widetilde{M}$ with the lifting of some $A^{1}_{B}$-deformation of $(T^{1,0}_{M},\eta)$ is almost isomorphic to $\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D})$ with the lifting of the homogeneous Sasakian structure $(T^{1,0}_{D},\eta_{D})$. ###### Proof. By Proposition 3.3 and Remark 3.4, for some positive integer $l$, the $l$-th power of $\bigwedge^{n}T^{1,0}_{M}$ is equivariantly isomorphic to $E_{-lC}$. Consider the basic holomorphic line bundle $L$ over $M$ defined by the pull- back $h^{\ast}\bigwedge^{m}T^{1,0}_{D}$. Then, Since $f$ is $\rho$-equivariant holomorphic and $dh:T^{1,0}_{\tilde{M}}\to T^{1,0}_{D}$ is an isomorphism at every fiber, $L$ is identified with $\bigwedge^{m}T^{1,0}_{M}$ via the differential $dh$. We regard the $\rho$-equivariant map $h:\widetilde{M}\to D$ as a $V$-reduction $P$ of the flat bundle $(\widetilde{M}\times G)/\pi_{1}(M,x)$. Then $L$ is induced by $P$ associated with the representation $V\to GL(\bigwedge\mathfrak{g}^{1,-1})$. Take an equivariant isomorphism $L^{l}\cong E_{-lC}$. By $E_{-lC}=M\times{\mathbb{C}}$, corresponding to $M\ni x\mapsto(x,1)\in M\times{\mathbb{C}}$, we have a nowhere vanishing section of $L^{l}$ and it induces an $\rho$-equivariant map $f:\widetilde{M}\to(G\times S^{1})/V=(G\times S^{1})/V=G/V^{\prime\prime}$ where $V^{\prime\prime}$ is the kernel of the representation $\Lambda^{l}$. $f$ is a lift of $h:\widetilde{M}\to D$. $G/V^{\prime}$ is a $l$-sheeted covering of $G/V^{\prime\prime}$. $(T^{1,0}_{D},\eta_{D})$ defines a Sasakian structure on $G/V^{\prime\prime}$. By the equivariance of $L^{l}\cong E_{-lC}$, we have $df(\tilde{\xi})=lC\xi_{D}$. Since $M$ is compact and $f$ is $\rho$-equivariant, the Riemann metric $f^{\ast}g_{\eta_{D}}$ is complete on $\widetilde{M}$. $f:\widetilde{M}\to G/V^{\prime\prime}$ is a local isometry and hence covering map. Let $\tau=-\tilde{\eta}+\frac{1}{C}f^{\ast}\eta_{D}$. Then, by $\tau(\tilde{\xi})=0$ and $d\tau\in A^{2}_{B}(\widetilde{M})$, we have $\tau\in A^{1}_{B}(\widetilde{M})$. The $A^{1}_{B}$-deformation $(T^{1,0}_{\widetilde{M}\tau},\tilde{\eta}+\tau)$ of $(T^{1,0}_{\widetilde{M}},\tilde{\eta})$ is almost isomorphic to $(\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D}),T^{1,0}_{D},\eta_{D})$. Since $f$ is $\rho$-equivariant, $(T^{1,0}_{\widetilde{M}\tau},\tilde{\eta}+\tau)$ is defined on $M$. ∎ ###### Remark 4.2. In this proof, $f:\widetilde{M}\to G/V^{\prime}$ satisfies $df(T^{1,0}_{\widetilde{M}})=T^{1,0}_{D}$ modulo $\langle\xi\rangle$ but may not be a CR-map. Hence, the $A_{B}^{1}$-deformation $(T^{1,0}_{\widetilde{M}\tau},\tilde{\eta}+\tau)$ is essential. ## 5\. Higgs bundles and uniformizations ### 5.1. Higgs bundles over Sasakian manifolds Let $(M,T^{1,0}_{M},\eta)$ be a compact Sasakian manifold. A basic Higgs bundle over $M$ is a pair $(E,\,\theta)$ consisting of a basic holomorphic vector bundle $E$ and $\theta\,\in\,A^{1,0}_{B}(M,\,{\rm End}(E))$ satisfying the following two conditions: $\overline{\partial}_{{\rm End}(E)}\theta\,=\,0\ \ \text{ and }\ \ \theta\wedge\theta\,=\,0\,.$ We define the degree of a basic holomorphic vector bundle $E$ by ${\rm deg}(E)\,:=\,\int_{M}c_{1,B}(E)\wedge(d\eta)^{n-1}\wedge\eta\,.$ Denote by ${\mathcal{O}}_{B}$ the sheaf of holomorphic functions on $M$, and for a holomorphic vector bundle $E$ on $M$, denote by ${\mathcal{O}}_{B}(E)$ the sheaf of holomorphic sections of $E$. Consider ${\mathcal{O}}_{B}(E)$ as a coherent ${\mathcal{O}}_{B}$-sheaf. For a basic Higgs bundle $(E,\,\theta)$, a sub-Higgs sheaf of $(E,\,\theta)$ is a coherent ${\mathcal{O}}_{B}$-subsheaf $\mathcal{V}$ of ${\mathcal{O}}_{B}(E)$ such that $\theta({\mathcal{V}})\,\subset\,{\mathcal{V}}\otimes\Omega_{B}$, where $\Omega_{B}$ is the sheaf of basic holomorphic $1$-forms on $M$. By [BH, Proposition 3.21], if ${\rm rk}(\mathcal{V})\,<\,{\rm rk}(E)$ and ${\mathcal{O}}_{B}(E)/\mathcal{V}$ is torsion-free, then there is a transversely analytic sub-variety $S\,\subset\,M$ of complex co-dimension at least 2 such that ${\mathcal{V}}\big{|}_{M\setminus S}$ is given by a basic holomorphic sub-bundle $V\,\subset\,E\big{|}_{M\setminus S}$. The degree ${\rm deg}(\mathcal{V})$ can be defined by integrating $c_{1,B}(V)\wedge(d\eta)^{n-1}\wedge\eta$ on this complement $M\setminus S$. ###### Definition 5.1. We say that a basic Higgs bundle $(E,\,\theta)$ is stable if $E$ admits a basic Hermitian metric and for every sub-Higgs sheaf ${\mathcal{V}}$ of $(E,\,\theta)$ such that ${\rm rk}(\mathcal{V})\,<\,{\rm rk}(E)$ and ${\mathcal{O}}_{B}(E)/\mathcal{V}$ is torsion-free, the inequality $\frac{{\rm deg}(\mathcal{V})}{{\rm rk}(\mathcal{V})}\,<\,\frac{{\rm deg}(E)}{{\rm rk}(E)}$ holds. A basic Higgs bundle $(E,\,\theta)$ is called polystable if $(E,\,\theta)\,=\,\bigoplus_{i=1}^{k}(E_{i},\,\theta_{i})\,,$ where each $(E_{i},\,\theta_{i})$ is a stable Higgs bundle with $\frac{{\rm deg}(E_{i})}{{\rm rk}(E_{i})}\,=\,\frac{{\rm deg}(E)}{{\rm rk}(E)}\,.$ Let $(E,\,\theta)$ be a basic Higgs bundle over a compact Sasakian manifold $M$. Assume that $E$ admits a basic Hermitian metric $h$. Define $\overline{\theta}_{h}\,\in\,A^{0,1}_{B}(M,\,{\rm End}(E))$ by $h(\theta(e_{1}),\,e_{2})\,=\,h(e_{1},\,\overline{\theta}_{h}(e_{2}))$ for all $e_{1},\,e_{2}\,\in\,E_{x}$ and all $x\,\in\,M$. Define the canonical connection $D^{h}\,=\,\nabla^{h}+\theta+\overline{\theta}_{h}$ on $E$. ###### Theorem 5.2 ([BK, BK2]). If a basic Higgs bundle $(E,\,\theta)$ is stable and satisfies $c_{1,B}(E)\,=\,0\qquad{\rm and}\qquad\int_{M}c_{2,B}(E)\wedge(d\eta)^{n-2}\wedge\eta\,=\,0,$ then there exists a basic Hermitian metric $h$ so that the canonical connection $D^{h}$ is flat and such Hermitian metric is unique up to a positive constant. Moreover, such metric $h$ is a Harmonic metric on the flat bundle $(E,D^{h})$ with respect to the Sasakian metric $g_{\eta}$ and hence the flat bundle $(E,D^{h})$ is semi-simple ([Co]). This hermitian metric is said to be a harmonic metric on $(E,\theta)$. A polarized complex basic variation of Hodge structure of weight $w$ over $M$ is $(E=\bigoplus_{p+q=w}E^{p,q},D,H)$ so that 1. (1) $(E,D)$ is a complex flat vector bundle. 2. (2) $\bigoplus_{p+q=w}E^{p,q}$ is a direct sum of $C^{\infty}$-subbundles such that $D_{\xi}{\mathcal{C}}^{\infty}(E^{p,q})\subset{\mathcal{C}}^{\infty}(E^{p,q})$. 3. (3) (The Griffiths transversality conditions) For any $X\in T^{1,0}_{M}$ (resp. $X\in T^{0,1}_{M}$), $D_{X}{\mathcal{C}}^{\infty}(E^{p,q})\subset{\mathcal{C}}^{\infty}(E^{p,q})\oplus{\mathcal{C}}^{\infty}(E^{p-1,q+1})$ (resp. $D_{X}{\mathcal{C}}^{\infty}(E^{p,q})\subset{\mathcal{C}}^{\infty}(E^{p,q})\oplus{\mathcal{C}}^{\infty}(E^{p+1,q-1})$) 4. (4) $H$ is a parallel Hermitian form so that decomposition $\bigoplus_{p+q=w}E^{p,q}$ is orthogonal and $h$ is positive on $E^{p,q}$ for even $p$ and negative for odd $p$. Define the basic bundle structure $\nabla_{\xi}$ on $E$ and each $E^{p,q}$. By the flatness $D^{2}=0$ and the Griffiths transversality conditions, we have: * • We have $D=\nabla+\theta+\bar{\theta}$ such that $\nabla$ is a connection on each $E^{p,q}$ and $\theta\in A^{1}(M,{\rm Hom}(E^{p,q},E^{p-1,q+1}))$, $\bar{\theta}\in A^{1}(M,{\rm Hom}(E^{p,q},E^{p+1,q-1}))$. * • $\nabla$ gives a basic holomorphic bundle structure $\nabla^{\prime\prime}$, $\theta\in A^{1,0}(M,{\rm Hom}(E^{p,q},E^{p-1,q+1}))$, $\bar{\partial}_{{\rm End}(E)}\theta=0$ and $\theta\wedge\theta=0$. Thus $(E,\theta)$ is a basic Higgs bundle. * • Define the basic Hermitian metric $h$ on $E$ by $(-1)^{p}H$ on each $E^{p,q}$. Then $\nabla$ is the canonical connection associated with $h$ and $\bar{\theta}=\overline{\theta}_{h}$ as the above sense. Thus $h$ is a Harmonic metric on a basic Higgs bundle $(E,\theta)$. Let $G=SU(E_{x},H_{x})$ and $V=S(\Pi_{p+q=w}U(E^{p,q}_{x},H_{x}))$. Take the monodromy $\rho:\pi_{1}(M,x)\to GL(E_{x})$ of the flat bundle $(E,D)$. Since $h$ is parallel, we can assume $\rho(\pi_{1}(M,x))\subset G$. Let $\mathfrak{g}$ be the Lie algebra of $G$. The decomposition $E=\bigoplus_{p+q=w}E^{p,q}$ gives the decomposition $\mathfrak{g}_{{\mathbb{C}}}=\bigoplus_{r}\mathfrak{g}^{r,-r}$. This decomposition is a Hodge structure on $G$ and the period domain is $G/V$ (cf. [Sch]). Considering $h$ as a $V$-reduction of the flat bundle $(\tilde{M}\times G)/\pi_{1}(M,x)$, we have a $\rho$-equivariant map $h:\tilde{M}\to G/V$ such that $dh=\theta$ on $T^{1,0}_{\tilde{M}}$. By the Griffiths transversality conditions, $(G,\rho,h)$ is a $G$-variation of Hodge structure. Let $(E,\,\theta)$ be a polystable basic Higgs bundle satisfying $c_{1,B}(E)\,=\,0\qquad{\rm and}\qquad\int_{M}c_{2,B}(E)\wedge(d\eta)^{n-2}\wedge\eta\,=\,0.$ By Theorem 5.2, we have a harmonic metric $h$ on $(E,\theta)$. For any $t\in U(1)$, we have the new polystable basic Higgs bundle $(E,\,t\theta)$ and we can easily check that $h$ is also a harmonic metric on $(E,\,t\theta)$. Assume that $U(1)$ acts on $E$ by holomorphic automorphisms via $\mu:U(1)\to{\rm Aut}(E)$ so that $\mu(t)\theta\mu(t)^{-1}=t\theta$. Then, $\mu(t)$ is an isomorphism between basic Higgs bundles $(E,\,\theta)$ and $(E,\,t\theta)$. By the harmonicity of $h$ on both $(E,\,\theta)$ and $(E,\,t\theta)$ and the uniqueness of harmonic metrics, we can say that the action $\mu:U(1)\to{\rm Aut}(E)$ preserves a harmonic metric $h$. Fix $w\in{\mathbb{Z}}$, define the sub-bundle $E^{p,w-p}$ consisting of $e\in E$ such that for any $t\in U(1)$, $\mu(t)e=t^{p}e$. The polarized complex basic variation of Hodge structure of weight $w$ $(E=\bigoplus_{p+q=w}E^{p,q},D^{h},H)$ is defined by $H=(-1)^{p}h$ on each $E^{p,q}$. ### 5.2. Uniformizations associated with the canonical Higgs bundles Define the Higgs bundle $(E_{M},\,\theta_{M})$ by the following way: * • $E_{M}\,=\,{\mathbb{C}}_{M}\oplus T^{1,0}_{M}$ where ${\mathbb{C}}_{M}$ is the trivial basic holomorphic line bundle on $M$, and * • $\theta_{M}\,=\,\left(\begin{array}[]{cc}0&0\\\ 1&0\end{array}\right)$ where $1$ is the identity element in ${\rm End}(T^{1,0}_{M})=(T^{1,0}_{M})^{\ast}\otimes T^{1,0}_{M}$ regarded as an element in $A^{1,0}_{B}(M,{\rm Hom}({\mathbb{C}}_{M},T^{1,0}_{M}))$. ###### Theorem 5.3. We assume: 1. (1) The basic Higgs bundle $(E_{M},\,\theta_{M})$ is stable. 2. (2) The equality $\int_{M}\left(2c_{2,B}(T^{1,0}_{M})-\frac{n}{n+1}c_{1,B}(T^{1,0}_{M})^{2}\right)\wedge(d\eta)^{n-2}\wedge\eta\,=\,0$ holds. 3. (3) $c_{1,B}(T^{1,0}_{M})=-C[d\eta]$ for some positive constant $C$. Then the universal covering $\widetilde{M}$ with the lifting of some $A^{1}_{B}$-deformation of $(T^{1,0}_{M},\eta)$ is almost isomorphic to $\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D})$ with the lifting of the homogeneous Sasakian structure $(T^{1,0}_{D},\eta_{D})$ for $D=SU(n,1)/S(U(n)\times U(1))$. ###### Proof. By $c_{1,B}(T^{1,0}_{M})=-C[d\eta]$, $c_{1,B}(E\otimes E_{C^{\prime}})=0$ for some $C^{\prime}\in{\mathbb{R}}$. By $\int_{M}\left(2c_{2,B}(T^{1,0}_{M})-\frac{n}{n+1}c_{1,B}(T^{1,0}_{M})^{2}\right)\wedge(d\eta)^{n-2}\wedge\eta\,=\,0,$ we have $\int_{M}c_{2,B}(E_{M}\otimes E_{C^{\prime}})\wedge(d\eta)^{n-2}\wedge\eta\,=\,0.$ $\nabla^{\prime\prime}=\nabla_{|\langle\xi\rangle\oplus T^{0,1}}$ defines a holomorphic bundle structure on $L$. Thus, $(E_{M}\otimes E_{C^{\prime}},\,\theta\otimes{\rm Id}_{L})$ is a basic Higgs bundle. Since $(E_{M},\,\theta_{M})$ is stable, $(E_{M}\otimes E_{C^{\prime}}\,\theta_{M}\otimes{\rm Id}_{E_{C^{\prime}}})$ is also stable. Let $h$ be a harmonic metric on $(E\otimes E_{C^{\prime}},\,\theta\otimes{\rm Id}_{E_{C^{\prime}}})$. Define $E^{0,1}=T^{1,0}\otimes E_{C^{\prime}}$ and $E^{1,0}=E_{C^{\prime}}$. We have the polarized complex basic variation of Hodge structure of weight $1$ $(E_{M}\otimes E_{C^{\prime}}=E^{1,0}\oplus E^{0,1},D^{h},H)$. $SU(E_{x},H_{x})\cong SU(n,1)$ and the Hodge structure on $SU(n,1)$ is Hermitian and the period domain $D$ is the unit ball $SU(n,1)/S(U(n)\times U(1))$. We obtain a $SU(n,1)$-variation of Hodge structure $(SU(n,1),\rho,h)$. By the definition, for the map $h:\tilde{M}\to SU(n,1)/S(U(n)\times U(1))$, $dh=\theta$ is an isomorphism on each fiber of $T^{1,0}_{\widetilde{M}}$. Hence, $(SU(n,1),\rho,h)$ is uniformizing. ∎ ###### Remark 5.4. If the basic Higgs bundle $(E_{M},\theta_{M})$ is stable, then the Miyaoka-Yau type inequality $\int_{M}\left(2c_{2,B}(T^{1,0}_{M})-\frac{n}{n+1}c_{1,B}(T^{1,0}_{M})^{2}\right)\wedge(d\eta)^{n-2}\wedge\eta\,\geq\,0\,.$ holds. Thus the second condition means that the equality of the Miyaoka-Yau type inequality holds. In [Zh], Zhang proves that if $c_{1,B}(T^{1,0}_{M})=C[d\eta]$ for some positive constant $C$ and the certain condition in terms of Sasaki-Einstein geometry holds, then the Miyaoka-Yau type inequality holds and the equality is a criterion for uniformizing a higher dimensional compact Sasakian manifold by the odd-dimensional sphere. ###### Remark 5.5. In case $D=SU(n,1)/S(U(n)\times U(1))$, the line bundle $\bigwedge^{n}T^{1,0}_{D}$ is a homogeneous vector bundle corresponding to the representation $S(U(n)\times U(1))\ni\left(\begin{array}[]{cc}U&0\\\ 0&{\rm det}(U)^{-1}\end{array}\right)\mapsto{\rm det}(U)^{n+1}$. Define the subgroup $U^{\prime}(n)=\left\\{\left(\begin{array}[]{cc}U&0\\\ 0&{\rm det}(U)^{-1}\end{array}\right):{\rm det}(U)^{n+1}=1\right\\}$ in $SU(n,1)$. We have $S^{1}(\bigwedge^{n}T^{1,0}_{D})=SU(n,1)/U^{\prime}(n)=PU(n,1)/SU(n)$. Taking the universal covering $\widetilde{PU}(n,1)$ of $PU(n,1)$, we have $\widetilde{S^{1}}(\bigwedge^{m}T^{1,0}_{D})=\widetilde{PU}(n,1)/SU(n)$. The lifting of the Sasakian structure $(T_{D}^{1,0},\eta_{D})$ is $\widetilde{PU}(n,1)$-invariant. Assume $n=1$. Then $H^{2}_{B}(M)=\langle[d\eta]\rangle$. In this case, if $c_{1,B}(T^{1,0})=-C[d\eta]$ for some positive constant $C$, then the Higgs bundle $(E_{M},\,\theta_{M})$ is stable. By $H^{4}_{B}(M)=0$, the equality $\int_{M}\left(2c_{2,B}(T^{1,0})-\frac{n}{n+1}c_{1,B}(T^{1,0})^{2}\right)\wedge(d\eta)^{n-2}\wedge\eta\,=\,0$ is trivial. ###### Corollary 5.6. Let $(M,T^{1,0}_{M},\eta)$ be a compact $3$-dimensional Sasakian manifold with $c_{1,B}(T^{1,0})=-C[d\eta]$. Then the universal covering $\widetilde{M}$ with the lifting of some $A^{1}_{B}$-deformation of $(T^{1,0}_{M},\eta)$ is almost isomorphic to $\widetilde{PU}(1,1)$ with a left-invariant Sasakian structure. ## References * [BH] D. Baraglia and P. Hekmati, A foliated Hitchin-Kobayashi correspondence, arXiv:1802.09699 (2018). * [Bel1] F. A. Belgun, On the metric structure of non-Kähler complex surfaces. Math. Ann. 317 (2000), no. 1, 1–40. * [Bel2] F. A. Belgun, Normal CR structures on compact 3-manifolds. Math. Z. 238 (2001), no. 3, 441–460. * [BaK] O. Baues and Y. Kamishima Locally homogeneous aspherical Sasaki manifolds. Differential Geom. Appl. 70 (2020) 41 pp. * [BK] I. Biswas and H. Kasuya, Higgs bundles and flat connections over compact Sasakian manifolds, Comm. Math. Phys. 385 (2021), 267–290. * [BK2] I. Biswas and H. Kasuya, Higgs bundles and flat connections over compact Sasakian manifolds, II: quasi-regular bundles, arXiv:2110.10644. * [BSW] D. Burns, S. Shnider, and R. O. Wells, Deformations of strictly pseudoconvex domains, Invent. Math. 46 (1978),237–253. * [BG] C. P. Boyer and K. Galicki, Sasakian geometry, Oxford Mathematical Monographs. Oxford University Press, Oxford, 2008. * [Co] K. Corlette, Flat G-bundles with canonical metrics, J. Differential Geom. 28 (1988), 361–382. * [Hi] N. J. Hitchin, The self-duality equations on a Riemann surface, Proc. London Math. Soc. 55 (1987), 59–126. * [Ge] H.Geiges, Normal contact structures on 3-manifolds. Tohoku Math. J. (2) 49 (1997), no. 3, 415–422. * [GrSc] P. Griffiths, W. Schmid, Locally homogeneous complex manifolds, Acta Math., 123 (1969), 253–302. * [GS] V. Guillemin , S. Sternberg, Supersymmetry and Equivariant de Rham Theory, Springer, Berlin, 1999. * [MuR] I. Mundet i Riera, Lifts of smooth group actions to line bundles. Bull. Lond. Math. Soc. 33(2001), no 3 351–361. * [Ra] J. H. Rawnsley, Flat partial connections and holomorphic structures in $\mathcal{C}^{\infty}$ vector bundles, Proc. Amer. Math. Soc. 73 (1979), 391–397. * [Sch] W. Schmid, Variation of Hodge structure: the singularities of the period mapping. Invent. Math. 22 (1973), 211–319. * [Si1] C. T. Simpson, Constructing variations of Hodge structure using Yang-Mills theory and applications to uniformization. Jour. Amer. Math. Soc. 1 (1988), no. 4, 867–918. * [Si] C. T. Simpson, Higgs bundles and local systems. Inst. Hautes Études Sci. Publ. Math. No. 75 (1992), 5–95. * [Ta] N. Tanaka, A Differential Geometric Study on strongly pseudoconvex CR manifolds, Lecture Notes in Math., 9, Kyoto University, 1975. * [Wa] A. W. Wadsley, Geodesic foliations by circles, J. Differential Geometry 10 (1975), 541–549. * [We] S. Webster, Pseudo-Hermitian structures on a real hypersurface, J. Differential Geom. 13 (1978), 25–41. * [Zh] X. Zhang, Energy properness and Sasakian-Einstein metrics, Comm. Math. Phys. 306 (2011), 229–260.
# Dynamically Finding Optimal Observer States to Minimize Localization Error with Complex State-Dependent Noise Troi Williams1, Po-Lun Chen2, Sparsh Bhogavilli2, Vaibhav Sanjay1, and Pratap Tokekar1 *This work was supported by the Computing Innovation Fellows Project 2021 (NSF Award #2127309).1These authors are with the Department of Computer Science, University of Maryland, College Park, MD 20742, USA. <EMAIL_ADDRESS>authors are with the A. James Clark School of Engineering, University of Maryland, College Park, MD 20742, USA. <EMAIL_ADDRESS> ###### Abstract We present DyFOS, an active perception method that Dynamically Finds Optimal States to minimize localization error while avoiding obstacles and occlusions. We consider the scenario where a ground target without any exteroceptive sensors must rely on an aerial observer for pose and uncertainty estimates to localize itself along an obstacle-filled path. The observer uses a downward- facing camera to estimate the target’s pose and uncertainty. However, the pose uncertainty is a function of the states of the observer, target, and surrounding environment. To find an optimal state that minimizes the target’s localization uncertainty, DyFOS uses a localization error prediction pipeline in an optimization search. Given the states mentioned above, the pipeline predicts the target’s localization uncertainty with the help of a trained, complex state-dependent sensor measurement model (which is a probabilistic neural network in our case). Our pipeline also predicts target occlusion and obstacle collision to remove undesirable observer states. The output of the optimization search is an optimal observer state that minimizes target localization uncertainty while avoiding occlusion and collision. We evaluate the proposed method using numerical and simulated (Gazebo) experiments. Our results show that DyFOS is almost 100x faster than yet as good as brute force. Furthermore, DyFOS yielded lower localization errors than random and heuristic searches. Figure 1: This figure illustrates a high-level representation of the observer and target state machines (right) and one iteration of our motivating example (left). The target (grey rover) with no exteroceptive sensors moves along a path (solid purple line) with obstacles (green trees). The target’s localization uncertainty is drawn as an orange dashed ellipse. At time $i$, the target sends a rendezvous state to the observer (blue drone), predicting that it (the target) will arrive at time $j$. The observer finds and moves to an optimal state that will minimize the corrected localization uncertainty of the target (see Figure 2 for a detailed view of the optimization module). When the target arrives (at approximately time $j$), the observer estimates the target’s pose and uncertainty and sends them to the target. The target updates its pose and uncertainty and sends another rendezvous state to the observer. This process repeats until the target arrives at the goal. This paper focuses on the modules in rounded rectangles with solid lines. ## I Introduction Navigating through known or unknown environments is challenging when an autonomous robot is ill-equipped to perceive its surrounding environment. For example, many works explore how adverse weather affects sensors typically used on autonomous cars, such as cameras, LiDAR, radar, and GPS [1, 2, 3, 4, 5]. We note that sensor performance degradation or malfunction also occurs in other situations outside of adverse weather. Small robots may not even have exteroceptive sensing capable of supporting robust localization. We are interested in the situation where a robot can no longer access its exteroceptive sensors for an extended period. The ground robot (termed _target_) wants to navigate to a goal along a path that contains obstacles (Figure 1). We aim to minimize the target’s localization uncertainty by leveraging cooperation from an aerial robot (termed _observer_). The target receives pose and uncertainty measurements from the observer to update its pose estimate. The localization uncertainty is a function of the measurement uncertainty (from the observer). We consider the case where the measurement uncertainty is a function of the states of the observer, target, and surrounding environment (e.g., the ambient lighting conditions). Although we traditionally assume measurement uncertainty is constant, other works have shown that the measurement uncertainty can be state-dependent (some include [6, 7, 8, 9, 10]). Therefore, we aim to find an optimal observer state that minimizes the target’s localization uncertainty. To localize the target accurately, we propose DyFOS, a novel active perception method that Dynamically Finds Optimal States to minimize localization error (uncertainty) (Figure 2). The main contribution of DyFOS is a localization error prediction pipeline that is used in an optimization search. Given the state of the observer, target, and surrounding environment, the pipeline predicts the target’s corrected localization error with the help of a complex State-Dependent Sensor Measurement Model (SDSMM). The pipeline also allows constraints to avoid (predicted) obstacle collision or occlusion. Together, the optimization method and the pipeline select optimal observer poses that would minimize localization error and allow us to forego selecting optimal states in a heuristic manner (for example, requiring the target to be in the center of the sensor’s field of view [11]). Our paper is organized as follows. Section II places our work in the context of the literature. We formulate our problem in Section III. Section IV describes the DyFOS algorithm. We discuss our experiments and their results in Section V. Finally, Section VII concludes the paper. ## II Related Work Active perception is a process in which an agent performs a set of strategic actions that allow it to gather more insightful information about some phenomenon [12, 13, 14]. Active perception has been applied to various cases including autonomous scientific information gathering [15], foreground segmentation [16], object detection and target tracking [17, 9, 18, 19, 11], ocean flow and vehicle states [20], and searching for individuals [21, 22]. Of these related work, several are similar to our proposed method and motivating example [23, 11, 19]. Gürcüoglu et al. [23] proposed an active target tracking approach for a team of quadrotors that were equipped with 3D range sensors. Their approach contributes (1) a hierarchical controller that tracks desired optimal trajectories for each quadcopter and (2) a method that uses a Kalman Filter to fuse the position estimates of a target from each drone. The optimal trajectories attempt to minimize the fused position error of the target. Similar to [23], we compute optimal drone states that minimize the perceived error of a target. However, unlike [23], we focused on developing a complex, non-analytical measurement noise model that influenced the observer’s 2D pose. Therefore, our method must dynamically search for optimal solutions. We can also use any controller with our observer. Falanga et al. [11] proposed the first perception-aware, model predictive control framework for quadrotors. Their method used numerical optimization to compute trajectories that simultaneously optimized action and camera-specific, human-specified perception objectives, both of which can conflict. The human- specified perception objectives required the point of interest to be near the center of the image. Like Falanga et al., we optimize for perceptual objectives (namely, minimizing the target’s pose error). However, our localization error prediction pipeline influences the (output) observer pose rather than human-specified objectives. Furthermore, our method handles collision and occlusion constraints. Tallamraju et al. [19] proposed an active approach to a cooperative detection and tracking algorithm (in [24]) that tracked a person performing activities. Their method aimed to minimize the 3D position error of the person by ensuring optimal viewpoint configurations of the drones. Although we both perform target tracking, our methods differ. First, We employ a learned noise model to compute an optimal pose and, by extension, a viewpoint (via heading) instead of using heuristics and an analytical noise model. Second, we output an optimal pose (instead of a motion plan), allowing a flexible choice of motion planners. ## III Problem Formulation ### III-A Preliminaries Consider a world with a planar surface, an observer, a target, and a collection of static obstacles. The target is an autonomous ground vehicle (AGV) that wants to navigate along a path. Let ${\boldsymbol{\bar{x}}^{t}_{i}\in\text{SE}(2)}$, ${\boldsymbol{P}^{t}_{i}\in\mathbb{R}^{3\times 3}}$ denote an estimate of the target’s 2D pose and its covariance (uncertainty) of the state error at timestep $i$, respectively. The target has a map $\boldsymbol{M}$ that contains the locations and dimensions of obstacles. Due to damage or failure of the target’s exteroceptive sensors, the target cannot localize or avoid obstacles. Therefore, it relies on an observer to estimate its state and uncertainty. The observer is a rotorcraft that can localize itself accurately. Let ${\boldsymbol{\bar{x}}^{o}_{i}\in\text{SE}(3)}$ represent the 3D pose of the observer. The observer uses a downward, front-facing camera (with pitch $\beta$) to estimate the target’s state and uncertainty. We define the measurement model as $\boldsymbol{z}_{i}=\boldsymbol{x}^{t}_{i}+\boldsymbol{\sigma}_{i}$. Here, $\boldsymbol{x}^{t}_{i}$ is the target’s true 2D pose and $\boldsymbol{\sigma}_{i}\in\mathbb{R}^{3}$ is the state-dependent measurement noise [6] at time $i$. We assume each state-dependent noise $\boldsymbol{\sigma}_{i}\sim\mathcal{N}({0,\boldsymbol{\Sigma}_{i}})$ is normally distributed with covariance $\boldsymbol{\Sigma}_{i}$, and each covariance is a function of the states of the observer, target, and surrounding environment. Given the observer, target, and environment states, we also assume each covariance can be predicted or estimated dynamically by a state-dependent sensor measurement model (SDSMM). Finally, we assume the (predicted or estimated) states are available offline and online. The observer requires time to compute an optimal state, move to such a state, and observe the target. To account for this time, the observer rendezvouses with the target at rendezvous states $\boldsymbol{x}^{t}_{j}$ that are dynamically selected along the target’s path. Each rendezvous state is at a timestep $j$ in the future (where $i<j$) and is selected based on the approximate compute and travel times of the observer. ### III-B Problem Statement Our goal is two-fold. First, we find a future, optimal observer state $\boldsymbol{x}^{o\ast}_{j}$ that minimizes the predicted localization uncertainty of the target as it navigates along a path. Second, we seek observer states that avoid collisions and occlusions due to obstacles. To accomplish these goals, we propose an algorithm called DyFOS, which dynamically finds an optimal observer state that will 1) avoid obstacles, 2) avoid occlusions, and 3) minimize the target’s predicted localization uncertainty after a measurement update at timestep $j$. We search an optimal state $\boldsymbol{x}^{o\ast}_{j}$ for time $j$ using the following optimization problem: $\begin{split}\boldsymbol{x}^{o\ast}_{j}=&\operatorname*{arg\,max}_{\boldsymbol{x}^{o}_{j}\in\boldsymbol{\mathcal{X}}^{o}_{j}}\mathcal{O}\big{(}\boldsymbol{P}^{t}_{j}\big{)}\\\ \text{subject to}&~{}c_{l}\\!\left(\boldsymbol{x}^{o}_{j},\boldsymbol{\hat{x}}^{t}_{j},\boldsymbol{M}\right)=0,\forall l\in[1,L].\end{split}$ (1) Here, $(\boldsymbol{\hat{x}}^{t}_{j},\boldsymbol{\hat{P}}^{t}_{j})$ is the target’s rendezvous state and predicted state uncertainty at timestep $j$. $\boldsymbol{x}^{o}_{j}$ is a sampled observer state from $\boldsymbol{\mathcal{X}}^{o}$, the set of all observer states. ${\mathcal{C}=\\{c_{l}\\}_{l=1}^{L}}$ are a set of constraints that predict obstacle collision or occlusion. $\boldsymbol{P}^{t}_{j}$ is the predicted, corrected localization uncertainty after a measurement update at time $j$. Moreover, $\boldsymbol{M}$ is our environment model. Inspired by [25, 26], we choose $\boldsymbol{\hat{x}}^{t}_{j}$ as a waypoint along the target’s path. Furthermore, we compute $\boldsymbol{\hat{P}}^{t}_{j}$ by propagating the target’s current localization uncertainty $\boldsymbol{\hat{P}}^{t}_{i}$ from timestep $i$ to timestep $j$ using the Kalman Filter equations. Section IV discusses this optimization problem. ## IV Dynamically Finding Optimal States to Minimize Localization Error Figure 2: DyFOS is composed of an optimization method (dashed green) and a localization error prediction pipeline (LEPP) (solid orange). Together, the optimizer and pipeline find an optimal observer state $\boldsymbol{x}^{o\ast}_{j}$ that predicts the lowest target localization error $\hat{E}_{\textbf{P}}$ after a measurement update at time $j$. Then the observer flies to the optimal state to estimate the target’s pose and uncertainty. See Figure 1 for the entire state machine. We propose a method that searches for an optimal observer state $\boldsymbol{x}^{o\ast}_{j}$ that minimizes the predicted, corrected localization uncertainty $\boldsymbol{P}^{t}_{j}$ of a target at time $j$. Without the loss of generality, we describe our proposed method in the context of a multi-robot application, where a target (with no exteroceptive sensors) receives pose measurements $\boldsymbol{z}_{j}$ and their estimated uncertainties $\boldsymbol{\hat{\Sigma}}_{j}$ from an aerial observer. Figure 2 illustrates the proposed algorithm. The algorithm is composed of 1) a localization error prediction pipeline (LEPP) and 2) an optimization method. The pipeline contains four main steps. First, we sample an observer state $\boldsymbol{x}^{o}_{j}$ from a time-varying state space $\boldsymbol{\mathcal{X}}^{o}_{j}$ (Section IV-A). Next, we construct a feature vector $\boldsymbol{\lambda}_{j}$ using the sampled state $\boldsymbol{x}^{o}_{j}$, the rendezvous state $\boldsymbol{\hat{x}}^{t}_{j}$ at time $j$, and our environment model $\boldsymbol{M}$. Then an SDSMM predicts a covariance of the measurement noise $\boldsymbol{\hat{\Sigma}}_{j}$ given the feature vector $\boldsymbol{\lambda}_{j}$ (Section IV-B). The pipeline also evaluates the constraints given states of the observer, target, and environment (Section IV-C). Finally, the pipeline predicts a localization error $\hat{E}_{\textbf{P}}$ given the predicted measurement uncertainty and the output of our constraints (Section IV-D). The optimizer executes the pipeline many times to find one optimal observer state $\boldsymbol{x}^{o\ast}_{j}$ for a future time $j$. Discussions on selecting an optimizer and finding the appropriate parameters are beyond the scope of this paper. However, Section V mentions the optimizer used in our experiments. ### IV-A The Spatio-Temporal, Observer State Space The pipeline begins with sampling observer states from an observer state space that varies with space and time. Let $\boldsymbol{\mathcal{X}}^{o}$ define the set of all observer states in a map. Searching for an optimal pose $\boldsymbol{x}^{o\ast}_{j}$ within an arbitrarily large map with an arbitrarily long state vector $\boldsymbol{x}^{o}$ requires a significant amount of computation. Therefore, we define a new search space $\boldsymbol{\mathcal{X}}^{o}_{j}$ that reduces the computational complexity. First, we seek a 2D pose for the observer. We assume the observer’s altitude is near constant and its velocities, accelerations, pitch, and roll are near zero at time $j$. Second, we seek observer poses where the target (at the rendezvous state $\boldsymbol{\hat{x}}^{t}_{j}$) is in range and field-of-view of the observer. Using both statements, we define the set $\boldsymbol{\mathcal{X}}^{o}_{j}$ as: $\begin{split}\boldsymbol{\mathcal{X}}^{o}_{j}=&\\{\boldsymbol{x}^{o}_{j}\in\boldsymbol{\mathcal{X}}^{o}_{j},\boldsymbol{x}^{o}_{j}\subset\text{SE}(2)~{}|~{}\rho_{\text{min}}\leq\rho_{j}\leq\rho_{\text{max}},\\\ &|\psi_{j}|\leq\frac{\text{FoV}_{\psi}}{2},~{}|\phi_{j}|\leq\frac{\text{FoV}_{\phi}}{2}\\}.\end{split}$ (2) Here, $\rho_{\text{min}}$ and $\rho_{\text{max}}$ denote the minimum and maximum sensing range of the camera. $\text{FoV}_{\psi}$ and $\text{FoV}_{\phi}$ represent the camera’s horizontal and vertical field of views (FoVs). Finally, $\rho_{j}$ is the distance between the camera and the target, while $\psi_{j}$ and $\phi_{j}$ are the horizontal and vertical angles of the target relative to the camera (respectively). We note that $\boldsymbol{\mathcal{X}}^{o}_{j}$ also imposes an “in sensing range” constraint that varies with time and space due to the motion of the target and the positions of obstacles in the observer’s workspace. Figure 4 illustrates sample states from $\boldsymbol{\mathcal{X}}^{o}_{j}$, where each sampled state is within range and view of the target. Collision and occlusion predicting is handled in Section IV-C. Selecting poses from $\boldsymbol{\mathcal{X}}^{o}_{j}$ is user-specific. For example, one can sample uniformly and use gradient-based methods to select subsequent poses. ### IV-B State-Dependent Sensor Measurement Model Figure 3: We implement a state-dependent sensor measurement model (SDSMM) using a multi-layer perception (MLP) and a non-trainable Gaussian layer. We use the states of the observer, target, and surrounding environment to construct a feature vector $\boldsymbol{\lambda}_{k}$. Then, given $\boldsymbol{\lambda}_{k}$, the MLP outputs a free parameter vector $\boldsymbol{\upsilon}_{k}$. Afterward, the Gaussian layer computes the covariance of the measurement noise $\boldsymbol{\hat{\Sigma}}{k}$ using the parameter vector $\boldsymbol{\upsilon}_{k}$. A state-dependent sensor measurement model (SDSMM) is a trainable model that estimates the state-dependent covariance of the measurement noise [6, 27]. Following Williams and Sun [6, 7], we implement an SDSMM using a probabilistic neural network (Figure 3). The input to the SDSMM is a set of features $\boldsymbol{\lambda}\in\mathbb{R}^{d_{\lambda}{}}$ that correlate with the measurement noise111Williams and Sun [6, 27] refer to such features as a combined state. However, we avoid using the term to alleviate confusion between a combined state and states of robots.. The output of our SDSMM is an estimate of the covariance of the measurement noise $\boldsymbol{\hat{\Sigma}}{}$ of the target’s 2D pose. We train an SDSMM using a dataset ${\mathcal{D}=\\{(\boldsymbol{e}_{k},\boldsymbol{\lambda}_{k})\\}}_{k=1}^{K}$ that contains a collection of $K$ measurement errors $\boldsymbol{e}_{k}$ and their corresponding features $\boldsymbol{\lambda}_{k}$. The measurement error is defined as ${\boldsymbol{e}_{k}=(\boldsymbol{z}_{k}-\boldsymbol{x}^{t}_{k})}$, where $\boldsymbol{z}_{k}$ is a measurement of the target’s pose using a pose estimation algorithm and $\boldsymbol{x}^{t}_{k}$ is the ground truth pose of the target. Generally speaking, the feature vector $\boldsymbol{\lambda}_{k}$ for a given sensor and estimation problem varies. For instance, a feature vector for a camera and target localization problem may contain the ambient lighting of the area, the velocity of the camera, and the distance and relative velocity between the camera and the target. We note that these features can be computed given the ground truth, estimates, or predictions of the states of the observer, target, and surrounding environment. For example, we compute the features using ground truth states during training. However, we compute the feature vector $\boldsymbol{\lambda}_{k}$ using estimated or predicted states during runtime. Given the dataset $\mathcal{D}$, we use the negative log-likelihood loss function $\mathcal{L}$ to train our model. Since our measurement noise is normally-distributed, our loss function simplifies to $\mathcal{L}=-\frac{d_{\sigma}}{2}\ln{(2\pi)}-\frac{1}{2}\sum_{k=1}^{K}\left(\ln{(|\boldsymbol{\hat{\Sigma}}_{k}|)}+\boldsymbol{e}_{k}^{\top}\boldsymbol{\hat{\Sigma}}_{k}^{-1}\boldsymbol{e}_{k}\right),$ (3) where $\boldsymbol{\hat{\Sigma}}_{k}$ is the output of the SDSMM and $d_{\sigma}$ is the dimension of $\boldsymbol{\hat{\Sigma}}_{k}$. ### IV-C Observer State Constraints Figure 4: We show an example of predicting feasible and non-feasible observer states via our constraints (Section IV-C). For example, a sampled, future observer state $\boldsymbol{x}^{o}_{j}$ is not feasible if we predict the observer will collide with an obstacle (state 2) or an obstacle will occlude the target at the rendezvous state $\boldsymbol{\hat{x}}^{t}_{j}$ (state 1 with a partial FoV). Otherwise, we predict collision and occlusion will not occur (state 3 with the solid green FoV). We use constraints to predict which states are feasible during our online search (Figure 4). Let ${\mathcal{C}=\\{c_{{}_{l}}\\}_{l=1}^{L}}$ contain a set of constraints that determines if an observer state ${\boldsymbol{x}^{o}_{j}\in\boldsymbol{\mathcal{X}}^{o}_{j}}$ is a feasible solution. Collectively, the set of constraints $\mathcal{C}$ defines a feasible region that varies with time and space, and this feasible region contains the optimal observer state we seek. We use two constraints: collision avoidance and line-of-sight. Each constraint is defined as a function. The input to each function is a set of variables that describe a sampled observer, the target, or the surrounding environment. Each function returns $1$ if the prediction is true and $0$ otherwise. For example, we predict a sampled observer will collide with an obstacle if the two objects intersect. Likewise, we predict an obstacle will occlude the target if an obstacle is hit by rays emitting from the observer to the target. Our numerical and simulated experiments employed these constraints using simple geometric shapes [28] and an Octomap [29], respectively. ### IV-D Localization Error Prediction The final step in our pipeline predicts a localization error for timestep $j$. We remind the reader that we seek to minimize the target’s corrected localization uncertainty $\boldsymbol{P}^{t}_{j}$ using the measurement $\boldsymbol{z}_{j}$ and its estimated uncertainty $\boldsymbol{\hat{\Sigma}}_{j}$. Therefore, we define the localization error $\hat{E}_{\textbf{P}}$ as $\hat{E}_{\textbf{P}}=\mathcal{O}\big{(}\boldsymbol{P}^{t}_{j}\big{)}=\ln\det\left(\boldsymbol{P}^{t}_{j}\right)+\eta\sum_{l=1}^{L}\mathbf{1}(c_{l}),$ (4) where $\boldsymbol{P}^{t}_{j}$ is the uncertainty after a measurement update at time $j$, $\eta$ is a user-defined constraint penalty, and $c_{l}$ is the output of a constraint (Section IV-C). Suppose we use a Gaussian distribution to estimate the target’s pose and uncertainty. Then we can predict the corrected localization uncertainty $\boldsymbol{P}^{t}_{j}$ using the Riccati equations of the Kalman Filter [23, 9]. In the Riccati equations, we compute the target’s predicted localization uncertainty $\boldsymbol{\hat{P}}^{t}_{j}$ (prior to the measurement update) by propagating the target’s localization uncertainty from time $i$ up to time $j$. Furthermore, the predicted measurement uncertainty $\boldsymbol{\hat{\Sigma}}_{j}$ comes from the SDSMM (Section IV-B). In addition to varying with time, we note that $\boldsymbol{P}^{t}_{j}$ and, ultimately, $\hat{E}_{\textbf{P}}$ also vary with each sampled state because the predicted measurement uncertainty $\boldsymbol{\hat{\Sigma}}_{j}$ is a function of the sampled observer states (Section IV-A). ## V Experiments We performed numerical experiments and a simulated experiment in Gazebo. Each experiment used one of three algorithms to compute an observer pose. Our numerical experiments compared the pose to a brute-force solution. The algorithms used were random placement, heuristic placement, and DyFOS. The random algorithm arbitrarily placed the observer within view of the target. The heuristic algorithm arbitrarily placed the observer within $0.5$ to $3.1$ m of the target, which embodied a “sense close” rule of thumb. Finally, DyFOS used the method described in Section IV and the differential evolution global optimizer [30] in SciPy [31]. All algorithms employed the constraints (Section IV-C). We also denoted obstacles using 2D shapes and an Octomap in the numerical and simulated experiments, respectively. DyFOS used an SDSMM to predict the measurement noise. To train the SDSMM, we defined our camera’s ground truth uncertainty model as $\boldsymbol{\Sigma}=\text{diag}(\sigma_{x},\sigma_{y},\sigma_{\theta})$, where $\sigma_{x}=0.05(x+1)^{2}+0.05$, $\sigma_{y}=0.05(y-1)^{2}+0.05$, and $\sigma_{\theta}=\frac{4\pi}{180}$. The minimum uncertainty was at $(-1,1)$ (in the observer’s body frame). Here, $x$, $y$ were the target’s estimated position relative to the observer. These equations were chosen arbitrarily. During training and online, the input to the SDSMM was a noisy 2D position of the target relative to the observer. Each predicted covariance $\boldsymbol{\hat{\Sigma}}_{k}$ represented the uncertainty in the target’s 2D pose. Our training set was defined as $\mathcal{D}=\\{(\boldsymbol{\lambda}_{k},\boldsymbol{e}_{k})\\}_{k=1}^{K}$, where $K=10,000$. Here, $\boldsymbol{\lambda}_{k}$ is the position of the target relative to the observer, ${\boldsymbol{e}_{k}\sim\mathcal{N}({0,\boldsymbol{\Sigma}_{k}})}$ is the corresponding measurement error, and $\boldsymbol{\Sigma}_{k}$ is the measurement noise from the ground truth model. We used a brute force method to approximate ground truth poses that 1) satisfied all constraints and 2) had the lowest navigation loss using the ground truth noise model. The predicted localization error was computed using (4). ### V-A Numerical Placement in Random 2D Environments The placement experiment compared the output observer pose from each algorithm with the brute force solution. We generated 1,000 arbitrary 2D maps, which were comprised of one target and 25 obstacles. We selected arbitrary poses for the target and obstacles. The observer and target were 0.75 meters and 1.5 meters long, respectively. We sampled the length (in meters) of each obstacle using the uniform distribution $\mathcal{U}(0.5,2.5)$. The goal of each algorithm was to minimize the localization uncertainty $\boldsymbol{P}^{t}_{j}=\boldsymbol{I}\in\mathbb{R}^{3x3}$. ### V-B Numerical Target Localization The mobile target localization experiment featured a ground target that navigated along a path while receiving pose estimates and their uncertainty from the observer. The observer used the three algorithms to determine an optimal pose at each time step. We used three paths: a line, a circle, and a figure-$\infty$. The length or radius of each path was five meters. We ran each algorithm on a path 25 times to compute the mean performance. A Kalman Filter was used to estimate the target’s pose and incorporate pose estimates and their uncertainties from the observer. Finally, an LQG controller was used to compute control commands for the target. Figure 5: This figure shows a drone (the observer) localizing the Turtlebot (the target). The drone detects the Turtlebot via an ArUco marker. ### V-C Static Target Localization in Gazebo The target localization experiment in Gazebo featured a stationary Turtlebot (target) and an Iris drone (observer) with a 45-degree downward-facing camera. We performed nine trials where the drone used one of the three algorithms to find an observer pose, flew to that pose, and estimated the localization uncertainty of the Turtlebot (Figure 5). We performed the experiments in a world with a $30$ m2 planar surface. The world contained 25 static obstacles that were 10 m tall, had a diameter of one meter, and were in the workspace of both robots. ## VI Results and Discussion We used four metrics to evaluate each algorithm: 1) the predicted localization error, 2) each observer’s distance from the optimal solution, 3) each algorithm’s run times, and 4) the Kalman filter convergence. We computed the localization error using (4). The distance from the optimal solution was the 2D distance between the target and the observer. Figure 6: This figure shows the distributions for the localization errors (top) and distances from the optimal solution (bottom) for the placement experiment with $1,000$ randomly-generated worlds. DyFOS performed on par with the brute force method. Lower is better. Localization Error. No observer collided with any obstacle, and no obstacle occluded the target for any algorithm or world in our numerical placement experiment. We also observed that DyFOS had a mean error of ${-8.997\pm 0.046}$, performing on par with the brute force algorithm (with a mean error of $-8.815$) (Figure 6 top). By contrast, the localization error for the heuristic and random placement algorithms varied significantly, yielding mean errors of ${-8.101\pm 0.525}$ and ${-7.459\pm 0.738}$ (respectively). DyFOS also outperformed the random and heuristic algorithms in both localization experiments. Among the three algorithms, our method achieved the lowest loss the quickest and had the least variation for all three paths in the numerical experiment. Likewise, DyFOS also had the lowest localization errors in the simulated experiment. The mean localization errors for DyFOS, heuristic, and random were $-8.768\pm 0.002$, $-8.385\pm 0.299$, and $-7.699\pm 0.675$, respectively. We believe that the low loss and small variation are because DyFOS actively reduces the localization error. Finally, although the random and heuristic algorithms produced low errors, their respective errors plateaued higher than DyFOS (see Figure 7 for numerical results). Distance from the Optimal Solution. We observed similar performance trends as the above analyses for the numerical experiment (Figure 6). The mean distance from the optimal pose for the brute force algorithm and DyFOS were ${2.6\pm 12.7}$ cm and ${9.8\pm 17.6}$ cm, respectively. For $94.1\%$ of the (${1,000}$) worlds, DyFOS placed the observer within $25$ centimeters of the optimal pose. (Recall that the ground truth measurement uncertainty is least when the target is at $(1,-1)$ meters relative to the observer). By contrast, the heuristic and random algorithms placed the observer within $488$ and $768$ cm of the optimal pose (respectively) and yielded significant variation for the same number ($941$) of worlds. For perspective, a DJI Mini 3 PRO has a width of 24.5 cm unfolded and without propellers [32]. For the other $5.9\%$ of the worlds, DyFOS produced poses that were $25$ cm to $1.26$ m away from the optimal pose. These higher distances were due to fixing the observer’s heading, avoiding obstacles, and preventing occlusions. DyFOS and the brute force algorithm generally output similar poses in such cases. DyFOS also performed better than the other two searches in the simulated experiment. The minimum of the noise model was $(3,0.5)$ for this setup. The mean difference between the optimal solution and DyFOS was $2.6\pm 0.9$ cm. Meanwhile, the mean differences between the optimal solution and the heuristic and random searches were $80.4\pm 40.8$ cm and $165.3\pm 81.8$ cm, respectively. Finally, we note that DyFOS performed better in simulated experiments because we did not fix the observer’s heading in the Gazebo. Run Times. The run times vary due to the complexity of each algorithm. The heuristic and random placement algorithms took less than $0.01$ s to find an observer pose since they did not use any sensor model. The run time for DyFOS was ${0.90\pm 0.24}$ s with a maximum of $2.46$ s using one core, which is reasonable for real-time usage. Finally, the brute force algorithm took ${75.52\pm 2.34}$ s with a maximum of $83.82$ s using $32$ cores to find a solution, which is not reasonable for real-time usage. | Line | Circle | Figure-8 ---|---|---|--- Random | $11.1\pm 5.4$ | $16.4\pm 8.2$ | $24.9\pm 11.0$ Heuristic | $8.6\pm 2.8$ | $12.8\pm 6.1$ | $13.9\pm 6.0$ DyFOS | $\boldsymbol{5.1\pm 1.7}$ | $\boldsymbol{9.9\pm 3.7}$ | $\boldsymbol{12.6\pm 5.2}$ TABLE I: RMSE (cm) for each algorithm. Bolded means best. Figure 7: This figure shows the mean and standard deviation statistics for a Kalman Filter on the circle path over time. With DyFOS, the target’s estimated position converged the quickest (top). Furthermore, DyFOS had the lowest navigation loss and variation (bottom). Kalman Filter Convergence. We examined the time each algorithm took for the filter to converge with the target’s true pose and the error over time (Figure 7 top). For the Figure-$\infty$ and line paths, the three algorithms converged within a similar time and had a consistent error for the remainder of the trajectory. However, DyFOS converged significantly quicker for the circle path than the random and heuristic algorithms (within $15$ s for DyFOS versus at least $80$ s for the other algorithms). The results also showed that DyFOS had the lowest RMSEs for all three paths (Table I). ## VII Conclusion We proposed DyFOS, a novel active perception method that Dynamically Finds Optimal States to minimize localization error while avoiding obstacles and occlusions. DyFOS relies on user-defined constraints, a learned state- dependent sensor measurement model, and a loss function to evaluate the quality of each candidate state with respect to the localization task. DyFOS exploits a learned SDSMM to minimize the target’s localization uncertainty in a hazard-filled world. Our experiments showed that DyFOS performs on par with brute force methods but is almost 100x faster. The average runtime was less than one second using one CPU core, which may be reasonable for some real-time applications. DyFOS also achieves lower localization errors than random and heuristic search algorithms. ## References * [1] Shizhe Zang, Ming Ding, David Smith, Paul Tyler, Thierry Rakotoarivelo, and Mohamed Ali Kaafar. The Impact of Adverse Weather Conditions on Autonomous Vehicles: How Rain, Snow, Fog, and Hail Affect the Performance of a Self-Driving Car. IEEE Vehicular Technology Magazine, 14(2):103–111, 2019. * [2] Jorge Vargas, Suleiman Alsweiss, Onur Toker, Rahul Razdan, and Joshua Santos. An Overview of Autonomous Vehicles Sensors and Their Vulnerability to Weather Conditions. Sensors, 21(16), 2021. * [3] Robin Heinzler, Philipp Schindler, Jürgen Seekircher, Werner Ritter, and Wilhelm Stork. Weather Influence and Classification with Automotive Lidar Sensors. In 2019 IEEE Intelligent Vehicles Symposium (IV), pages 1527–1534, 2019. * [4] Marcel Sheeny, Emanuele De Pellegrin, Saptarshi Mukherjee, Alireza Ahrabian, Sen Wang, and Andrew Wallace. RADIATE: A Radar Dataset for Automotive Perception in Bad Weather. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 1–7, 2021. * [5] Li Tang, Yunpeng Shi, Qing He, Adel W. Sadek, and Chunming Qiao. Performance Test of Autonomous Vehicle Lidar Sensors Under Different Weather Conditions. Transportation Research Record, 2674:319 – 329, 2020. * [6] Troi Williams and Yu Sun. Learning State-Dependent, Sensor Measurement Models for Localization. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 3090–3097, 2019. * [7] Troi Williams and Yu Sun. Learning State-Dependent Sensor Measurement Models with Limited Sensor Measurements. In 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 86–93, 2021. * [8] G. Benet, F. Blanes, J.E. Simó, and P. Pérez. Using infrared sensors for distance measurement in mobile robots. Robotics and Autonomous Systems, 40(4):255–266, 2002. * [9] Zhongshun Zhang and Pratap Tokekar. Tree Search Techniques for Adversarial Target Tracking With Distance-Dependent Measurement Noise. IEEE Transactions on Control Systems Technology, 30(2):712–727, 2022. * [10] Baoqi Huang, Lihua Xie, and Zai Yang. tdoa-based source localization with distance-dependent noises. IEEE Transactions on Wireless Communications. * [11] Davide Falanga, Philipp Foehn, Peng Lu, and Davide Scaramuzza. PAMPC: Perception-Aware Model Predictive Control for Quadrotors. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 1–8, 2018. * [12] Ruzena Bajcsy, Yiannis Aloimonos, and John K Tsotsos. Revisiting Active Perception. Autonomous Robots, 42(2):177–196, 2018. * [13] Yiannis Aloimonos. Active Perception. Psychology Press, 2013. * [14] R. Bajcsy. Active Perception. Proceedings of the IEEE, 76(8):966–1005, 1988. * [15] Akash Arora, P. Michael Furlong, Robert Fitch, Salah Sukkarieh, and Terrence Fong. Multi-Modal Active Perception for Information Gathering in Science Missions. Auton. Robots, 43(7):1827–1853, oct 2019. * [16] Yuxiang Sun, Ming Liu, and Max Q.-H. Meng. Active Perception for Foreground Segmentation: An RGB-D Data-Based Background Modeling Method. IEEE Transactions on Automation Science and Engineering, 16(4):1596–1609, 2019. * [17] Pratap Tokekar, Joshua Vander Hook, and Volkan Isler. Active Target Localization for Bearing Based Robotic Telemetry. In 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 488–493, 2011. * [18] Alois Unterholzner, Michael Himmelsbach, and Hans-Joachim Wuensche. Active Perception for Autonomous Vehicles. In 2012 IEEE International Conference on Robotics and Automation, pages 1620–1627, 2012. * [19] Rahul Tallamraju, Eric Price, Roman Ludwig, Kamalakar Karlapalem, Heinrich H. Bülthoff, Michael J. Black, and Aamir Ahmad. Active Perception Based Formation Control for Multiple Aerial Vehicles. IEEE Robotics and Automation Letters, 4(4):4491–4498, 2019. * [20] Dongsik Chang, Matthew Johnson-Roberson, and Jing Sun. An Active Perception Framework for Autonomous Underwater Vehicle Navigation Under Sensor Constraints. IEEE Transactions on Control Systems Technology, pages 1–16, 2022\. * [21] José J. Acevedo, João Messias, Jesús Capitán, Rodrigo Ventura, Luis Merino, and Pedro U. Lima. A Dynamic Weighted Area Assignment Based on a Particle Filter for Active Cooperative Perception. IEEE Robotics and Automation Letters, 5(2):736–743, 2020. * [22] Juan Sandino, Fernando Vanegas, Felipe Gonzalez, and Frederic Maire. Autonomous UAV Navigation for Active Perception of Targets in Uncertain and Cluttered Environments. In 2020 IEEE Aerospace Conference, pages 1–12, 2020. * [23] Utku Gürcüoglu, Gustavo A. Puerto-Souza, Fabio Morbidi, and Gian Luca Mariottini. Hierarchical Control of a Team of Quadrotors for Cooperative Active Target Tracking. In 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 5730–5735, 2013. * [24] Eric Price, Guilherme Lawless, Roman Ludwig, Igor Martinovic, Heinrich H. Bülthoff, Michael J. Black, and Aamir Ahmad. Deep Neural Network-Based Cooperative Visual Tracking Through Multiple Micro Aerial Vehicles. IEEE Robotics and Automation Letters, 3(4):3193–3200, 2018. * [25] Ali-akbar Agha-mohammadi, Suman Chakravorty, and Nancy M. Amato. Firm: Feedback controller-based information-state roadmap - a framework for motion planning under uncertainty. In 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 4284–4291, 2011. * [26] Ali-Akbar Agha-Mohammadi, Suman Chakravorty, and Nancy M Amato. Firm: Sampling-based feedback motion-planning under motion uncertainty and imperfect measurements. The International Journal of Robotics Research, 33(2):268–304, 2014\. * [27] Yu Sun and Troi Williams. Learning State-Dependent Sensor Measurement Models for Localization, 2 2020. * [28] Sean Gillies et al. Shapely: Manipulation and Analysis of Geometric Objects, 2007–. * [29] Armin Hornung, Kai M. Wurm, Maren Bennewitz, Cyrill Stachniss, and Wolfram Burgard. OctoMap: An Efficient Probabilistic 3D Mapping Framework Based on Octrees. Autonomous Robots, 2013. Software available at https://octomap.github.io. * [30] Rainer Storn and Kenneth Price. Differential Evolution–A Simple and Efficient Heuristic for Global Optimization over Continuous Spaces. Journal of global optimization, 11(4):341–359, 1997. * [31] Pauli Virtanen, Ralf Gommers, Travis E. Oliphant, Matt Haberland, Tyler Reddy, David Cournapeau, Evgeni Burovski, Pearu Peterson, Warren Weckesser, Jonathan Bright, Stéfan J. van der Walt, Matthew Brett, Joshua Wilson, K. Jarrod Millman, Nikolay Mayorov, Andrew R. J. Nelson, Eric Jones, Robert Kern, Eric Larson, C J Carey, İlhan Polat, Yu Feng, Eric W. Moore, Jake VanderPlas, Denis Laxalde, Josef Perktold, Robert Cimrman, Ian Henriksen, E. A. Quintero, Charles R. Harris, Anne M. Archibald, Antônio H. Ribeiro, Fabian Pedregosa, Paul van Mulbregt, and SciPy 1.0 Contributors. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 17:261–272, 2020. * [32] DJI. DJI Mini 3 Pro - Specs - DJI, 2022.
# JWST NIRCam Defocused Imaging: Photometric Stability Performance and How it Can Sense Mirror Tilts Everett Schlawin Steward Observatory 933 North Cherry Avenue Tucson, AZ 85721, USA Thomas Beatty Department of Astronomy University of Wisconsin Madison, Madison, WI 53706 Brian Brooks Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 Nikolay K. Nikolov Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 Thomas P. Greene NASA Ames Research Center, Space Science and Astrobiology Division, MS 245-6, Moffett Field, CA, 94035, USA Néstor Espinoza Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 Department of Physics & Astronomy, Johns Hopkins University, Baltimore, MD 2121818, USA Kayli Glidic Steward Observatory 933 North Cherry Avenue Tucson, AZ 85721, USA Keith Baka Anton Pannekoek Institute, University of Amsterdam, Science Park 904, 1098 XH Amsterdam, The Netherlands Eiichi Egami Steward Observatory 933 North Cherry Avenue Tucson, AZ 85721, USA John Stansberry Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 Martha Boyer Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 Mario Gennaro Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 Department of Physics and Astronomy Johns Hopkins University 3400 North Charles Street, Baltimore, MD 21218, USA Jarron Leisenring Steward Observatory 933 North Cherry Avenue Tucson, AZ 85721, USA Bryan Hilbert Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 Karl Misselt Steward Observatory 933 North Cherry Avenue Tucson, AZ 85721, USA Doug Kelly Steward Observatory 933 North Cherry Avenue Tucson, AZ 85721, USA Alicia Canipe Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 Charles Beichman NASA Exoplanet Science Institute/IPAC Jet Propulsion Laboratory, California Institute of Technology 1200 E California Blvd Pasadena, CA 91125 Matteo Correnti Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 J. Scott Knight Ball Aerospace & technologies, Corp, 1600 Commerce Street, Boulder, CO 80301 Alden Jurling NASA Goddard Space Flight Center, 8800 Greenbelt Rd, Greenbelt, MD 20771, USA Marshall D. Perrin Space Telescope Science Institute 3700 San Martin Drive Baltimore, MD 21218 Lee D. Feinberg NASA Goddard Space Flight Center, 8800 Greenbelt Rd, Greenbelt, MD 20771, USA Michael W. McElwain NASA Goddard Space Flight Center, 8800 Greenbelt Rd, Greenbelt, MD 20771, USA Nicholas Bond ADNET Systems, Inc./Code 550, NASA’s Goddard Space Flight Center, Greenbelt, MD 20771 David Ciardi NASA Exoplanet Science Institute/IPAC Jet Propulsion Laboratory, California Institute of Technology 1200 E California Blvd Pasadena, CA 91125 Sarah Kendrew European Space Agency, Space Telescope Science Institute, 3700 San Martin Drive, Baltimore MD 21218 Marcia Rieke Steward Observatory 933 North Cherry Avenue Tucson, AZ 85721, USA ###### Abstract We use JWST NIRCam short wavelength photometry to capture a transit lightcurve of the exoplanet HAT-P-14 b to assess performance as part of instrument commissioning. The short wavelength precision is 152 ppm per 27 second integration as measured over the full time series compared to a theoretical limit of 107 ppm, after corrections to spatially correlated 1/f noise. Persistence effects from charge trapping are well fit by an exponential function with short characteristic timescales, settling on the order of 5-15 minutes. The short wavelength defocused photometry is also uniquely well suited to measure the realtime wavefront error of JWST. Analysis of the images and reconstructed wavefront maps indicate that two different hexagonal primary mirror segments exhibited “tilt events” where they changed orientation rapidly in less than $\sim$1.4 seconds. In some cases, the magnitude and timing of the flux jumps caused by tilt events can be accurately predicted with a telescope model. These tilt events can be sensed by simultaneous longer-wavelength NIRCam grism spectral images alone in the form of changes to the point spread function, diagnosed from the FWHM. They can also be sensed with the FGS instrument from difference images. Tilt events possibly from sudden releases of stress in the backplane structure behind the mirrors were expected during the commissioning period because they were found in ground-based testing. Tilt events have shown signs of decreasing in frequency but have not disappeared completely. The detectors exhibit some minor (less than 1%) deviations from linear behavior in the first few groups of each integration, potentially impacting absolute fluxes and transit depths on bright targets where only a handful of groups are possible. Overall, the noise is within 50% of the theoretical photon noise and read noise. This bodes well for high precision measurements of transiting exoplanets and other time variable targets. stars: atmospheres — stars: individual (HAT-P-14 (catalog )) — stars: variables: general ††facilities: JWST(NIRCam), JWST(FGS), Palomar††software: astropy (Astropy Collaboration et al., 2013), photutils v0.3 (Bradley et al., 2016), matplotlib (Hunter, 2007), numpy (van der Walt et al., 2011), scipy (Virtanen et al., 2020), starry (Luger et al., 2019), pynrc (Leisenring, 2019), webbpsf (Perrin et al., 2014), pysiaf (Sahlmann et al., 2019) jwst (Bushouse et al., 2022) ## 1 Introduction JWST will provide transformational new studies of exoplanet atmospheres with its unprecedented view in the infrared (Ahrer et al., 2022; Greene et al., 2016; Morley et al., 2017; Schlawin et al., 2018; Bean et al., 2018). The NIRCam instrument’s imaging, coronagraphic and spectrographic modes are providing powerful new insights into a range of astrophysical environments from the Solar System to the most distant galaxies from 0.6 µm to 5.0 µm (Rieke et al., 2005). The NIRCam instrument has two wavelength channels that simultaneously observe the same field. NIRCam’s grism time series mode produces simultaneous spectroscopy on the long wavelength channel and photometry on the short wavelength channel. The long wavelength spectroscopy employs a grism that enables spectroscopy from 2.4 to 5.0 µm at R=1100 to 1700, with the wavelength coverage determined by the filter selected on the filter wheel (Greene et al., 2017). This wavelength range includes many prominent features in exoplanet atmospheres including H2O, CH4, CO, CO2 and NH3 (Greene et al., 2016). The molecular, atomic and ionic spectral features, when detected in exoplanet atmospheres, can give clues about the formation of planets, composition, dynamics, structure and cloud and haze composition (e.g. Bitsch et al., 2021; Mordasini et al., 2016; Powell et al., 2019; Kempton et al., 2017). Given the small scale heights of planetary atmospheres ($\lesssim 10^{-3}R_{\odot}$) compared to their host star radii ($\sim 1R_{\odot}$) and their low temperatures compared to their host star temperatures ($T_{p}^{4}/T_{*}^{4}\lesssim 10^{-2}$, where $T_{p}$ and $T_{*}$ are the temperatures of the planet and star respectively), exoplanet transit spectroscopy requires extreme precision. The largest signals from even the most favorable planets are still very small. The ‘prominent’ 1.4 $\mu$m water vapor feature strength varies from 100 ppm in the prototypical hot Jupiter HD 209458 b (Deming et al., 2013) down to less than 30 ppm for the super-Earth GJ 1214 b (Kreidberg et al., 2018). The most favorable CO2 signatures are predicted to be 50 ppm and below in the temperate Earth-sized TRAPPIST-1 d and e (Barstow & Irwin, 2016; Lustig-Yaeger et al., 2019). Lessons from previous space telescopes such as Spitzer, Kepler and the Hubble Space Telescope (HST) have taught us that many instrument systematics can be present including intrapixel sensitivity, thermal breathing, charge trapping, and pointing errors (Beichman et al., 2014). Here we present the performance of the NIRCam short-wavelength time-series photometry based on commissioning observations of the HAT-P-14 system. A companion paper, Beatty et al. (2022) describes the simultaneous long- wavelength grism spectroscopy. The light that enters NIRCam is split by a dichroic beamsplitter so that the short wavelengths (0.6 to 2.3 µm) and long wavelengths (2.4 to 5.0 µm) are directed to different channels’ optics and detectors. The pupil wheel and filter wheel selections determine the wavelength extents of the observations ( a short wavelength F210M filter paired with the long wavelength filter F322W2 will cover approximately 1.99 to 2.23 µm on the short wavelength channel and 2.43 to 4.01 µm on the long wavelength channel). The flight software is configured to run the detectors with the same size subarrays and readout patterns so the two exposures are commanded to be simultaneous to within one frame time. For the grism time series mode, two short wavelength detectors (NRCA1 and NRCA3) centered on nearly the same piece of sky as the long wavelength (NRCALONG) are read out by detector electronics and downloaded from the observatory. Appendix A shows the relative layouts of the detectors.111Also see https://jwst- docs.stsci.edu/jwst-near-infrared-camera/nircam-instrumentation/nircam- detector-overview for the detector layout (STScI, 2016). When performing grism spectroscopy, the short wavelength channel can be (currently) configured to use the Weak Lens +4 or Weak Lens +8 lens optics, which are designed for mirror wavefront sensing (Greene et al., 2010). The Weak Lens +4 and Weak Lens +8 lens optics are a subset of all defocusing lenses, which also include Weak Lens -4 and Weak Lens -8. The Weak Lens +4 (on the filter wheel) and Weak Lens +/- 8 (on the pupil wheel) can be used in series simultaneously so that focuses of -8, -4, 0, +4 and +12 Peak-to-Valley (P-V) waves can be configured (Greene et al., 2010). The lenses were tuned to be sensitive to various spatial frequencies and levels of aberration in the wavefront at 2.12 µm (Dean & Bowers, 2003; Perrin et al., 2016). The Weak Lens +4 defocuses the light by 4.0 P-V units at 2.12 µm and results in a core PSF of approximately 31 pixels and the Weak Lens +8 defocuses the light by 8.0 P-V units at 2.12 µm and results in a core PSF of approximately 65 pixels. Physically, the Weak Lens +8 defocuses the light by modifying the wavefront by +8.0 waves P-V at 2.12 um at the exit pupil, where the positive convention is that the phase at the edge of the wavefront leads the phase at the center of the pupil. We assume that the equation for the phase in waves for the Weak Lens +8 is $8*(r^{2}-0.5)$ where r is the exit pupil radius normalized to unity at the edge and 0 in the center. This wavefront error displaces the best focus along the chief ray. The defocused PSF from a weak lens makes the short wavelength imaging very sensitive to changes in the wavefront and it also spreads the light over more pixels to decrease the rate of saturation for a given star brightness. While not yet supported, there is also an active proposal to allow the use of the Dispersed Hartmann Sensor (DHS) in the short wavelength channel instead of the weak lenses. Utilizing the DHS in the short wavelength channel would allow simultaneous spectroscopy across a large fraction of the full NIRCam wavelength regime (Schlawin et al., 2017). The target HAT-P-14 b (Torres et al., 2010) was selected for calibration purposes because it has no strongly detectable atmospheric features and a flat spectrum due to a high gravity (42 m s-2, calculated from Stassun et al., 2017) and small atmospheric scale height (2$\times 10^{-4}$ R⊙). The 3.4 MJ 1.4 RJ 1600 K planet (Stassun et al., 2017; Southworth, 2012) orbits a $K=8.85$, 6600 K F-type star with low chromospheric activity (Torres et al., 2010). It was observed by JWST NIRSpec (Espinoza et al., 2022), NIRISS and NIRCam to enable cross-instrument performance checks and to determine the transit depths are consistent. The expected variations in the transmission spectrum are 16 ppm, assuming a typical atmosphere that exhibits atmospheric features that are equivalent to 0.9 atmospheric scale heights (Wakeford et al., 2019), so the differences across wavelengths are expected to be about as large as the theoretical ideal photon noise $\sim$9 ppm for these observations. One option to characterize stability is to observe a known quiescent star and verify that the lightcurves are flat with time. However, in the event that the lightcurve is not perfectly flat, disentangling the measured precision from the baseline trend curves is challenging. Any de- trending done to the data, such as a Gaussian process regression, can remove trends but could also remove astrophysical signal on a time variable source. Therefore, we observed a known transiting planet with a known ephemeris to determine the ability of JWST to measure the transit depth. While the target was chosen to have no significant atmospheric signal (no significant change in transit depth with wavelength), it still has a transit depth of 6480 $\pm$ 240 ppm (Torres et al., 2010) or about 0.6% which is readily detectable. Figure 1: The Weak Lens +8 PSF spreads the light over a large region, which reduces the chance of saturation and reduces intra-pixel sensitivity. The Weak Lens +8 was designed for retrieving the primary mirror state and thus phase retrieval with the Weak Lens enables real-time monitoring of the mirror wavefront error. The above image is near the middle of a 2048$\times$128 SUBGRISM128 subarray with the F210M filter, a 27.0 second long integration time on HAT-P-14 with a plate scale of about 31 mas /px. The background subtraction annulus as shown as concentric circles and is the inner radius is the same as the source aperture used for photometry. The subarray position was adjusted downward after HAT-P-14 b commissioning to better center the hexagonal PSF on the subarray. Section 2 describes the observations during commissioning and how they were analyzed. Section 3 describes systematic effects in the data including mirror tilt events, tilt timescales, the charge trapping ramp and observed non- linearity. Section 4 describes the noise performance, pointing performance and time accuracy as measured from in-flight data and the transit depth precision when accounting for contamination from a background source. We present our conclusions in Section 5. Table 1: Summary of Key Times During Observation UT start time | Integration Number | Description ---|---|--- (YYYY-mm-ddThh:mm:ss) | | 2022-05-02T06:43:04 | 1 | Exposure Start 2022-05-02T09:21:34 | 344 | Transit Start (Contact 1) 2022-05-02T10:10:01 | 449 | HGA Move 2022-05-02T10:30:20 | 492 | Transit Midpoint 2022-05-02T10:30:41 | 493 | Tilt Event 1 2022-05-02T11:23:19 | 607 | Tilt Event 2 2022-05-02T11:39:05 | 641 | Transit End (Contact 4) 2022-05-02T12:43:25 | 780 | Exposure End Note. — Times are in Coordinated Universal Time. The barycentric dynamical times are approximately 4.63 minutes later. Integration numbers are given for 1-based counting ## 2 Observations and Image Processing ### 2.1 Observation Description We observed the hot Jupiter HAT-P-14 b to test the instrument performance and verify that the NIRCam Grism Time Series mode was suitable for science. The planet was selected to have a short duration transit (2.3 hr), high brightness that enables high precision, low stellar activity and high planet surface gravity to have a small atmospheric signal. The short duration requirement was selected to minimize the needed observing time but has the flip side of preferring targets with near-grazing transits (high impact parameters). HAT-P-14 b’s impact parameter is 0.91 (Fukui et al., 2016), so fitting the limb darkening without the planet crossing through the stellar mid-point can complicate the transit depth measurement and stellar models may be preferred for the limb darkening law, as described in Section 4.1. The observation exposure started at 2022-05-02T06:43 UTC, which was the earliest start time allowed by the Astronomy Proposal Tool (APT) special requirements to ensure a long baseline before transit ingress. The lightcurve was observation 1 of Commissioning Program ID 1442, which is available on the Barbara A. Mikulski Archive for Space Telescopes (MAST). Table 1 lists key times during the observations. The exposure duration was 6 hours, which enabled 2.6 hours of stabilization time before planet ingress. The SW channel was configured with the Weak Lens+8 pupil, F210M filter which covers approximately 1.9 to 2.2 $\mu$m. This was paired with the GRISMR pupil and F322W2 filter on the long wavelength channel, discussed in Beatty et al. (2022). The exposure setup was a single exposure with the BRIGHT2 readout pattern, 20 groups integrated up the ramp and 780 integrations for a 2048$\times$128 subarray (SUBGRISM128) with 4 output amplifier channels. NIRCam readout electronics sample non-destructive reads (ie. frames) to create a ramp and then reset the detector at the end of an integration. Detector groups are a set of 1 or more detector image frames where the frames have been either averaged or skipped by onboard electronics and the average frame (ie. a group) is downloaded from the spacecraft. The averaging or else skipping of frames between groups is done to reduce the data volume on some observations. Observations when more than one frame is averaged or skipped per group reduce the data volume as compared to saving every frame per integration. In BRIGHT2, there are 2 frames (ie. detector samples) averaged per group and zero frames skipped. This resulted in a frame time of 0.67597 seconds, a group time of 1.35194 seconds and a cadence of 27.72001 seconds per integration for 780 integrations. The 6 hour observation was continuous with no interruptions. Pointing performance was extremely stable as discussed in Section 4.3 below. There was one high gain antenna (HGA) move during the exposure but no interruptions or pauses to the integrations in the 6 hour exposure. This allows detector charge trapping effects to approach a steady state, unlike with HST where each satellite orbit around Earth results in a new charge trapping ramp effect approximately every 90 minutes (e.g. Berta et al., 2012). The point spread function (PSF) with the Weak Lens+8 pupil is shown in Figure 1. The Weak Lens +8 (along with the -8 wave and +4 wave lenses) is designed to measure the optical path differences of the primary mirror surface. It was enabled as an option in the grism time series and imaging time series modes to spread out the light over more pixels and also reduce the saturation on bright targets. Having the light spread out over more pixels reduces the intrapixel sensitivity, which strongly affected Spitzer Space Telescope lightcurves because the PSF was highly undersampled (Ingalls et al., 2016). It also allows realtime monitoring of the mirror surface along with the long wavelength grism spectroscopy, discussed in Section 3.1. We note that the full extent of the PSF is asymmetrically truncated by the subarray. This will be mitigated in Cycle 1 science observations because an update was made to better center the subarray position. Even with better centering, the SUBGRISM128 subarray truncates some of the wings of the PSF. In other words, the commissioning observations (Figure 1) had the PSF is asymmetrically truncated and in science observations it will be still truncated in the wings but more symmetrically. Fortunately, JWST pointing is so stable (to be discussed in Section 4.3) that there were no impacts from the PSF truncation on the lightcurve and zero- padding the image for wavefront analysis resulted in high quality wavefront phase retrievals (to be discussed in Section 3.1). Figure 2: The default rateints product for integration number 83 with no extra corrections (top) has large pre-amplifier offsets (rectangular-shaped offsets) and 1/f noise (horizontal banding). A row-by-row, odd/even by amplifier (ROEBA) correction, described in Section 2.2 uses background pixels to reduce the pre-amplifier offset and 1/f noise. The background pixels are the ones outside of the dashed black circle that has a radius of 201 pixels. ### 2.2 Data Analysis from Uncalibrated Data to Rates Per integration We begin with the uncal data (also known as “MAST level 1 uncalibrated”, “User Data Product Stage 0 Fully-Populated FITS file” or Data Processing Level 1b). We then run Stage 1 of the Calwebb STScI pipeline (Bushouse et al., 2022) with some modifications described below, to produce a rateints file which contains the counts in DN/s for each integration. We started with the STScI JWST Detector 1 Stage 1 pipeline version 1.6.0 and the Calibration Reference Data System (CRDS) context jwst_0822.pmap to convert the raw data values to signal slopes (DN/s) per integration. The Detector Stage 1 pipeline begins by scaling the groups (which doesn’t affect NIRCam readout modes), initializing the data quality flags, flagging saturated pixels and then subtracting a superbias. When examining the superbias subtraction, we note that every 4 integrations has a bias offset structure due to the fast frame resets that reset pixels across the whole detector to ensure that charge doesn’t migrate into the subarray. For this SUBGRISM128 subarray, the fast frame reset occurs every 4 integrations. However, accounting for the different superbias every 4 integrations resulted in very little change to the lightcurve so we used a single superbias subtraction for all integrations. Normally, detector stage 1 processing then does a reference pixel correction step. The reference pixels are light-insensitive pixels that share the same electronic noise sources as the regular light-sensitive pixels and thus allow efficient subtraction of common-mode noise sources. Normally, the bottom and side reference pixels are used to mitigate odd/even column effects, pre- amplifier resets and 1/f noise. However, the SUBGRISM128 subarray for the short-wavelength time series mode has no reference pixels along the top or bottom detector rows.222The SW subarray is designed to be centered on the defocused PSF when the same source is centered near the Y=34 long wavelength pixel on the LW subarray. See section A for a diagram. The relative positions of the short and long wavelength detectors when projected on the sky result in the weak lens +8 image being centered at Y= 167.5 in 1-based counting absolute coordinates on the short wavelength detector. The subarray position was adjusted to better center the PSF after HAT-P-14 was observed. Thus, the reference pixel step does not correct for large pre-amplifier offsets and smaller odd/even offsets. We instead replace the reference pixel step with a similar step that uses background instead of reference pixels to remove many of the same effects. We use a shorthand for this modified reference pixel step ROEBA, which stands for row-by-row, odd/even by amplifier correction, which lists the steps in reverse order. First, we select one of the four amplifiers for correction at a time. In this case a block pixels from 1 to 512, 513 to 1024, 1025 to 1536 or 1537 to 2048 pixels in the X direction and the entire 128 pixel tall subarray in the Y direction. For this correction, it is necessary to define a background region to use as a proxy for reference pixels that has minimal contamination from bright sources. We select all pixels that are more than 201 pixels away from the central PSF as shown in Figure 2. We then apply a slow- read correction, which subtracts the median odd count level in the background region from all odd numbered columns and the median even count level from in the background region from all even numbered columns. This removes the pre- amplifier offsets. We then calculate the row-by-row median count level from the background region and subtract this from all pixels in that row. This largely removes the longer time scale (low frequency) components of the 1/f noise (e.g. Schlawin et al., 2020). The resulting correction from ROEBA shown in Figure 2 (bottom) has substantially smaller pre-amplifier reset offsets in each amplifier as well as less 1/f noise banding along the horizontal direction. We note that the ROEBA method uses many more pixels across the horizontal fast-read direction than the side reference pixels on the detector so it can improve results even when reference pixels are available in a subarray. Another advantage of the ROEBA correction is that many fewer pixels are erroneously marked as jumps at the jump detection step (described below). One assumption in the ROEBA step is that it relies on a clean background region and could potentially over-subtract when there is faint scattered light or background sources so inspection for neighboring sources and faint emission is necessary before applying it to all NIRCam data. We next apply the non-linearity correction with default parameters. The result of the bias subtraction and ROEBA step is that the signal level is close to 0 DN at the background before applying the linearity correction step. The stage 1 pipeline normally has a dark current subtraction step after non-linearity correction, but dark current on the NIRCam short wavelength detectors is less than 0.05 e-/sec and there is currently too much signal from cosmic rays and persistence to measure well and subtract on these detectors. We next run the jump step correction with a threshold of 6 sigma. Without the ROEBA correction, many pixels are erroneously flagged as “jumps” (e.g. cosmic rays) because the pre-amplifier offsets and 1/f noise can appear as jumps in signal from one detector group to another. With the ROEBA correction, we note that these erroneous jumps are much less frequent and instead the jump step flags true outliers such as cosmic rays. One exception where the jump step does not flag a cosmic ray perfectly is the outer radii of ‘snowball“ events (Rigby et al., 2022), as shown in Figure 2 near X=737,Y=101. Here, only the brighter portions of the snowball are flagged with the jump detection algorithm used in the analysis for this paper. As the final step of stage 1, we run the ramp fit step with default parameters to calculate a count rate per pixel in Data Numbers (DN) per second. There is another gain scaling step for NIRSpec data but it does not affect the NIRCam data presented in this paper. ### 2.3 Photometric Extraction We next use aperture photometry with photutils (Bradley et al., 2016) to measure the flux for each integration to derive a flux as a function of time. We fit the central peak of the PSF with a Gaussian and used this to determine the aperture center for each integration. We use a circular source aperture with a radius of 81 pixels, and the mean background rate from a background annulus with an inner radius of 81 and an outer radius of 101 pixels. These pixels were found to produce the minimum lightcurve scatter from an aperture size grid search. Although the ROEBA step described above largely removes the background, we found that the additional background annulus subtraction improved the precision, likely because of residual 1/f not removed by ROEBA. We found the same level of noise whether we did aperture centering or kept the apertures fixed, with the exception of better photometry during the HGA move affecting one integration to compensate for the small amount of image motion within that one integration. We normalize the flux time series to a value near 1.0 (1000 parts per thousand, ppt) for lightcurve fitting and noise analysis, which we describe in Section 4.1. ## 3 Lightcurve Systematics The lightcurves overall exhibited very low level systematic errors, which we describe in this section. There were changes to the PSF due to “tilt” events as discussed in Section 3.1 through 3.3, an exponential ramp discussed in Section 3.4 and some evidence for non-linearity discussed in Section 3.5. ### 3.1 Tilt Events We originally noticed a large jump in the long wavelength time series (Beatty et al., 2022) but also found one at the same time at some aperture sizes in the short wavelength and a smaller jump later in the time series. The jumps in flux correspond to changes in the mirror surface, known as “tilt events” (Rigby et al., 2022). These were first noticed by choosing an arbitrary reference image (in this case the rate from the 6th integration) and dividing the rate from all other integrations by it. Figure 3 shows three example ratio images from 1) before the tilt events, 2) after the first tilt event and 3) after the second tilt event. The tilt events show up as clear wave patterns in these differential images due to changes in the mirror surface. The short wavelength Weak Lens +8 defocused PSF is part of the wavefront sensing tools aboard JWST and is thus highly sensitive to changes in the mirror surface. Figure 3: Weak Lens Ratio Images before and after both Tilt Events (top) and Phase retrieval, oriented as if looking down at the mirror surface from the secondary mirror (bottom plots). Figure 4: The jump in flux caused by the C6 tilt event is wavelength dependent. The black point and blue points with error bars show the measured flux jump size for the short wavelength (SW) aperture of 2.4″ and LW aperture of 0.630″, respectively. Our current model (green circle) correctly predicts the SW channel’s jump with a 2.4″ aperture but under-predicts the LW 0.63″ jump size (orange curve) for the grism spectrum, especially at 2.5 to 3.0 µm. We used a non-linear optimization based phase retrieval algorithm (Fienup, 1982, 1993) based on the algorithmic differentiation (Jurling & Fienup, 2014). The coherent propagation model used a flexibly sampled discrete Fourier transform (Jurling et al., 2018). We used a bias and gain invariant error metric (Thurman & Fienup, 2009) as the cost function. The phase retrieval model was initialized using the wavefront and pupil data calculated from the JWST optical model for the nearest available field point, the NRCA3-FP2MIMF multi-instrument multi-field point. Because PSF data were only available with the +8 wave weak lens, this was a single image phase retrieval problem. We configured the model with 128 pixels across the array storing the pupil and 256 pixels across the cropped and padded image. Because of the band-pass of the F210M filter used for imaging, we used a broadband point-spread function model (Fienup, 1999; Jurling et al., 2018). We ran the phase retrieval optimization in two steps. In the first step we used the first image in the time series to establish a starting point. We sequentially estimated image position (using a sub-pixel cross correlation method (Guizar-Sicairos et al., 2008)), global low order Zernike aberrations, segment level piston, and tip/tilt errors (1st order segment Zernike aberrations). Finally, we performed joint estimation over a smoothly interpolated low resolution grid in wavefront and amplitude. With a single image the ability to distinguish between phase and amplitude errors in this last step is limited. In the second step, we begin from the reference point established by the first phase retrieval. In this case the optimization was done jointly over global and segment Zernike aberrations without sequential bootstrapping. This facilitates change detection by requiring the pupil amplitude and higher order wavefront variation to be the same throughout the time series, but still allowing low order and segment wavefront changes through time. The phase retrieval results are shown in Figure 3 and it is clear that the C6 mirror segment changed orientation (tilted) at event 1 and that the C1 mirror segment tilted at event 2. In event 1, the wavefront over C6 changed by 18 nm RMS (evaluated over the segment) compared to the initial wavefront, and in event 2, C1 changed by 26 nm (again evaluated over the segment). These changes are much smaller than the wavelength of the light, so both the individual segments and the overall telescope remain diffraction limited. As shown in Figure 4, the C6 event produced a larger change with a magnitude that varies from 200 ppm to 1000 ppm across the 2.4 µm to 4.0 µm wavelength range for the long wavelength and 254 $\pm$ 22 ppm in the short wavelength photometry using an aperture radius of 79 pixels or 2.45 arcsec. Using the wavefront phase retrieval described in this section above (Jurling & Fienup, 2014) for the optical path differences, we make a prediction for the magnitude of the flux jump by comparing the enboxed or encircled energy before and after the jumps. We use enboxed energy for the long wavelength spectroscopic extraction box and encircled energy for the short wavelength photometry. The short wavelength (SW) model prediction of 260 ppm is very close to the measured 254 $\pm$ 22 ppm for the large aperture used in the short wavelength time series. The long wavelength (LW) jump model prediction under-predicts the jump size from 2.5 to 3.5 $\mu$m but agrees from 2.7 to 4.1 $\mu$m. Some of the difference between the model and measurement could be the difference between an imaging and dispersed PSF and how they are extracted. Preliminary models with a dispersed LW grism image can reproduce the 1000 ppm change at 2.5 $\mu$m. ### 3.2 Ways to Detect Tilt Events Figure 5: The lightcurves exhibited jumps most noticeably in the long wavelength spectroscopy, shown here as a broadband time series (plot A). These jumps (marked as dash-dot vertical lines) can be detected in a variety of ways. Apertures placed on the short wavelength weak lens (WL) PSF are highly sensitive to mirror tilts on mirrors C6 and C1 (plot B). The time series are shown on a group-by-group difference image (points in blue and orange, plot B) as well as on the full jump-corrected integration (solid blue and orange lines, plot B). The FGS instrument can also sense mirror tilts by a dot product of a difference images, as defined in Equation 1 (plot C). The FWHM of the long wavelength spectroscopy can also be used to sense mirror tilts (plot D). The HGA move (vertical dotted line) also can create spikes in some tilt statistics. The apertures used to sense the C6 and C1 mirror1 tilts (right plot) are the ones used to generate the time series in plot B. It is possible to quantitatively measure tilt events in addition to visually inspecting the differential images or doing the phase retrieval shown in Figure 3. The mirror changes can be evaluated with photometric apertures on key parts of mirror surfaces, such as in Figure 5 (right plot). The time series of these apertures can cleanly sense the two tilt events like step functions as shown in Figure 5 (left second from the top). We also calculate this photometry on the difference between pairs of adjacent detector groups to assess the mirror changes at a higher time cadence (colored points in Figure 5). The NIRCam grism time series and time series imaging modes are the only modes currently supported that allows simultaneous high cadence wavefront sensing along with the time series. In some time series modes such as the NIRSpec Bright Object Time Series (Birkmann et al., 2022) or MIRI low resolution time series (Kendrew et al., 2015), defocused imaging is not currently available for real-time wavefront sensing of the primary mirrors, so other measures are desirable to sense tilt events. NIRISS Single Object Slitless mode (SOSS) (Doyon et al., 2012) has a de-focusing lens so this mode has some analogous capabilities as the short wavelength weak lens. We explored some ways to sense tilt events with the NIRCam grism time series spectroscopy and the fine guidance sensor (FGS) fine guide data streams (Doyon et al., 2012). We experimented with a few different statistics for the FGS images. We found that the flux from the telemetry stream333See the SA_ZFGINSTCT keyword https://jwst-docs.stsci.edu/methods-and-roadmaps/jwst- time-series-observations/jwst-time-series-observations-noise-sources, (STScI, 2016) can sense the first tilt event (C6) but could not easily reveal the second tilt event (C1). In addition, the first tilt event (C6) could be detected with a dot product with the average PSF or the $3\times 3$ pixel flux of the time series but the second tilt event from the C1 mirror was harder to pull out of the noise. The FWHM of the FGS 8$\times$8 images also shows step- like changes with the two tilt events but at low signal to noise. We found that principal component analysis (PCA) could reveal evidence of both tilt events but is also not easily distinguishable above the noise. The noise-weighted “differential dot product” was most effective in detecting tilt events with FGS. $D_{i}\equiv\left((\vec{A_{i}}-\vec{R})/\vec{V}\right)\cdot\left(\vec{M}-\vec{R}\right),$ (1) where $D_{i}$ is the differential dot product of 8$\times$8 images reshaped to 64-pixel linear vectors. In the expression, $\vec{A_{i}}$ is a single FGS cal data image, $\vec{R}$ is a reference image after the tilt event, $\vec{V}$ is the variance image, and $\vec{M}$ is a reference image before the tilt. This method did use prior information about the timing of tilt events from NIRCam data to find the average image from before the tilt $\vec{M}$ and the average image after the tilt $\vec{R}$, but it may be possible to iteravely scan the time series to determine $\vec{M}$ and $\vec{R}$ from FGS alone. This statistic essentially magnifies the differences between the shape of the PSF before and after the tilt event and weights by the noise. Figure 5 shows that both tilt events can be sensed with FGS using a dot product, where the $\vec{M}$ and $\vec{R}$ are chosen for the first tilt event (C6) and second tilt event (C1). Additionally, the full width at half maximum (FWHM, bottom left plot in Figure 5) of the long wavelength grism PSF is sensitive to tilt events because the PSF changes with the mirror tilt. Thus, even when weak lens imaging is not available, the FGS images and science instrument FWHM can be used in all time series modes to give clues about tilt events. However, Equation 1 requires some knowledge of the timing to find $\vec{M}$ and $\vec{R}$ unless they can be found iteratively. Therefore, a more general-purpose indicator of tilt events may be the FWHM of science data, which can be tracked in all time series observations. Tilt events are hypothesized to be structural microdynamics in the telescope that may occur when stresses in the backplane structure behind the mirrors are suddenly relaxed. As these stresses are released, the frequency of tilt events is expected to decrease in time. Regular wavefront measurements indicate that they are less frequent in the transition from commissioning to science observations in July 2022 than earlier in commissioning. However, tilt events were common even 4.5 months between launch and these observations, and similar tilt events may occur during Cycle 1 science observations. ### 3.3 Tilt Timescale The mirror changes appear to be step functions in Figure 5, so the change happens on a timescale faster than the 27 second cadence of integrations. We inspect the mirror changes at a higher cadence (but noisier) time series by calculating the mirror tilt specific photometry on the group-by-group difference images, which have a cadence of 1.351 seconds. In this analysis, adjacent points are thus anti-correlated because they share a detector group so if one sample is anomalously high, the next one will be anomalously low. As shown in Figure 6, there is a rapid change that takes 2 sampling durations of 1.351 seconds each to go from low to high or high to low, or in other words one sample between the low and high values. There are two causes for the intermediate sample: 1) the tilt timescale is between 1.351 and 2.702 seconds and 2) the timescale is shorter than 1.351 seconds and occurs midway through the group. We consider the first scenario less likely as it would require some fine-tuning to cause a tophat-like disturbance between these two time intervals that shows little sign of change in the PSF beyond 2.702 seconds. We therefore expect that the timescale is faster than the shortest measurable time with the NIRCam photometry which is the 1.351 second group time. We also note that these observations used the BRIGHT2 mode, which includes two frames within a group so the tilt could happen between these two frames within a group or within one of the frames. We do not see any discontinuities within the image that would indicate a change within the frame, but the PSF only covers a fraction of the field of view so it is not a strong constraint. Figure 6: Mirror photometry on the pairwise-subtracted images, zoomed in on events. We use the same apertures as shown in Figure 5 to sense the tilts on the C1 and C6 primary mirror segments. These two happen on a timescale as fast or faster than the 1.351 second time between BRIGHT2 detector groups, which contain 2 frames each. For the C1 tilt event, the jump likely happened in the middle of a detector group so it has a sample between the low and high flux values. ### 3.4 Charge Trapping Ramp The HST WFC3 IR detector, which is an earlier generation HgCdTe detector, exhibits ramps with an exponential settling behavior (e.g. Berta et al., 2012; Zhou et al., 2017). Ground-based laboratory tests showed likely negligible effects for typical JWST detectors (e.g. Schlawin et al., 2021). However, the NRCA3 short wavelength detector exhibits the most persistent charge and likely charge trap density of the 10 NIRCam detectors that are used in flight (Leisenring et al., 2016). NRCA3 is the same detector used for the short wavelength component of the grism time-series observations when the long wavelength filter is selected to be F277W, F322W2, F356W.444https://jwst- docs.stsci.edu/jwst-near-infrared-camera/nircam-operations/nircam-target- acquisition/nircam-grism-time-series-target-acquisition The NRCA1 detector that is paired with the F444W filter, had about half the accumulated counts from charge trap release as the NRCA3 detector in ground based tests (Leisenring et al., 2016), so it is expected to have a smaller amplitude exponential ramp. We fit an exponential model to the charge trapping behavior, as is common with HST charge trapping ramps (e.g. Berta et al., 2012) and show the results in Figure 7. We first correct the small 260 ppm jump in the lightcurve using a wavefront model described in Section 3.1. We fit the lightcurve with the following planet and systematics model: $f(t)=\left(A+Bx^{\prime}+Cx^{\prime 2}\right)\left(1-R\exp{(x-x_{0})/\tau}\right)f_{a}(t),$ (2) where A, B and C are coefficients in the quadratic baseline trend and normalization, $x$ is barycentric time, $R$ is the exponential amplitude, $x_{0}$ is the exposure start time, $f_{a}(x)$ is the astrophysical limb darkened lightcurve (starry, Luger et al., 2019), with a 4 parameter polynomial limb darkening law from ExoCTK and $x^{\prime}$ is the normalized time, $x^{\prime}=\frac{x-x_{med}}{x_{max}-x_{min}},$ (3) where $x_{med}$ is the median time, $x_{max}$ is the maximum time and $x_{min}$ is the minimum time. We fit the lightcurves by fixing $C=0$ (ie. a linear fit) and also letting it be a free parameter (ie a quadratic fit). As will be shown in Section 4.2, significant correlated errors are seen in the residuals if a linear baseline trend is adopted instead of a quadratic trend. We find an exponential amplitude $R$ of 731 ppm and an exponential time $\tau$ of 5.1 minutes for a linear trend and $R$ of 656 ppm and $\tau=$15 of minutes for a quadratic trend. The normalization and polynomial trend constants for the quadratic trend are A=${1000.24\pm 0.01}$ parts per thousand (ppt), B=${-0.911\pm 0.03}$ ppt and C=${0.80\pm 0.14}$ ppt. In absolute units, this corresponds to a slope of -0.15 ppt/hr and derivate of the slope of 0.022 (ppt/hr2). We note that initially, we fit the baseline and exponential start to just the initial part of the lightcurve and found an order-of-magnitude exponential settling timescale of 11 minutes (Rigby et al., 2022). The 5.1 and 15 minutes minutes came from a MCMC Bayesian fit to the full lightcurve with a linear and quadratic baseline respectively, whereas the 11 minutes comes from least squared minimization of the first 300 points (before planet ingress) and has a steeper linear slope. We have not determined the cause of the long timescale trend, but note a that different linear trends were seen on the NRS1 and NRS2 detectors (Espinoza et al., 2022) so it may be detector-related. The quadratic slope and exponential ramp terms in Equation 3 are correlated and thus the change in slope is fit with a different exponential settling time. We also looked through the previous JWST activities and found that NIRCam’s previous use was 37 hours before for wavefront sensing. NIRCam detectors remained in idle reset mode for those 37 hours so it is unlikely any long timescale traps were filled and are being released to create the downward slope seen in the data. Figure 7: The lightcurve settling behavior is very fast with a fitted settling timescale of 5-15 minutes but is somewhat degenerate with the time series baseline. ### 3.5 Non-Linearity Effects The H2RG detectors used in the near infrared instruments on JWST are never strictly linear at any well filling fraction, ranging from sub-percent non- linearity to tens of percent at 98% the hard saturation value (e.g. Canipe et al., 2017). However, they become increasingly non-linear as the detector approaches full well capacity and saturation (e.g. Plazas et al., 2017). Correction polynomials are applied by the CalWebb JWST pipeline (Bushouse et al., 2022) to linearize the counts as function of counts (DN). A different but analogous method has been shown that linearity corrections are possible even up to high well filling fractions ($\sim$97% Canipe et al., 2017). We assessed the difference between pairs of groups up the ramp to look for evidence of non-linearity. The stellar flux is highly stable within a 27 second long integration so it is expected that a linearity-corrected ramp should have a constant difference between each successive group. We find however that the first 4 group differences and the last 9 have a higher rate of flux than the middle 6 group differences, as shown in Figure 8. The early higher fluxes may be related to the release of trapped charge (e.g. Smith et al., 2008; Leisenring et al., 2016) or a type of reset anomaly (Rauscher et al., 2007). We do not perform a dark current subtraction, as discussed in Section 2.2. However, the dark current is less than 0.05 e-/sec on the short wavelength NIRCam detectors, and the 1% change in the differential samplings (shown in Figure 8) on top of a representative rate of 400 e-/s would require 4 $e^{-}$/s of dark current. So the reset anomaly seen here may be different from the anomaly seen on other H2RG devices (Rauscher et al., 2007). Exoplanet time series observations with a small number of groups ($\lesssim 5$) could have an absolute photometric flux offsets but also systematic differences in the transit depth as compared to observations with many groups ($\gtrsim 5$). This is because the non-linearity effects can change the response of the detector to a differential signal (ie. a non-constant derivative of count rate as a function of signal in e-/s/Mjy). With the observed 1% change in signal from group 2-1 to group 4-3 in Figure 8, a 2% deep transit of a hot Jupiter could have a measured transit depth of up to $\sim$ 1% $\times$ 2% $\times$ 1/4 = 50 ppm deeper if just using group 2 and 1. This difference is below the measurement noise for a single transit of HAT-P-14 b using only group 2 - 1 but could be assessed on a brighter target. There is also a possibility of reciprocity failure with HgCdTe detectors that can cause systematic changes in transit depth (Biesiadzinski et al., 2011; Schlawin et al., 2021). Figure 8: The difference in counts between groups up the ramp is not constant across an integration, as would be expected after linearity correction curves are applied. The violin-shaped points are the distributions of fluxes as calculated by the pairwise difference image of two detector groups, using all of the out-of-transit integrations . The median flux from all out-of-transit integrations is shown as a short horizontal line. Future observations with just 2 or 3 total groups may exhibit absolute flux differences and transit depth differences due to this non-linearity soon after detector reset. ## 4 Lightcurve Performance Figure 9: Lightcurve and residuals before and after corrections for an exponential ramp and jump from the first tilt event. The blue symbols with error bars are the fluxes for each integration at a cadence of 27.7 seconds, while the orange symbols are 8 minute long time-integrated bins that aid in viewing smaller-level changes. Red symbols are outlier points not used in the fitting. ### 4.1 Lightcurve Fitting We fit the lightcurve with a transit model to determine the lightcurve performance, discussed in Section 4.2 and the timing accuracy, discussed in Section 4.5. We remove a 260 ppm jump in the F210M (2.1 µm) time series due to a tilt event using a wavefront model of the optics discussed in Section 3.1, allow a quadratic trend with time and fit an exponential settling ramp discussed in Section 3.4. We use the integration mid-times in the barycentric reference frame and barycentric dynamical time standard int_mid_BJD_TDB included in the INT_TIMES extension of the JWST data products. We fit the corrected lightcurve with a starry transit lightcurve model (Luger et al., 2019), an exponential ramp, a quadratic trend in time and the 260 ppm jump correction. The planet’s near-grazing impact parameter (0.91 Fukui et al., 2016)) means that the planet does not traverse near the stellar midpoint and fitting the limb darkening parameters as free parameters can lead to large uncertainties in the planet’s radius. We fix the limb darkening law to a 6600 K, [Fe/H]=0.11, $\log{g}$ = 4.25 (Stassun et al., 2017) ATLAS9 model (Kurucz, 2017) calculated with ExoCTK for the F210M filter for $\mu>0.05$ where $\mu$ is the cosine of the angle between the line of signt and the emergent intensity (e.g. Kipping, 2013) for a 4 parameter non-linear law. We use the starry lightcurve model, which uses a polynomial limb darkening law (Agol et al., 2020), so we fit the intensity function from a 4 parameter nonlinear law with a 6 parameter polynomial limb darkening law, which has been shown to be accurate to $\lesssim 0.5$ ppm differences in flux (Agol et al., 2020). We start with priors centered on the values from Bonomo et al. (2017) for the period, inclination, eccentricity and argument of pericenter, but widened these in case of systematic errors to 83.5 $\pm 0.3^{\circ}$, $e$=0.1071 $\pm$ 0.01 and $\omega$=106.1 $\pm 5^{\circ}$. We use the ephemeris from publicly vailable TESS data and a wide a/R∗ of 8.9 $\pm$ 1.0 centered on the value from Stassun et al. (2017). The resulting lightcurve fits and resiuals are shown in Figure 9. We show a model fit that does not account for two of the systematics to better illustrate them as well as another that accounts for them (the charge trap ramp and tilt event jump). ### 4.2 Photometric Performance After fitting the lightcurve with a transit model that includes a jump correction from a phase retrieval, an exponential charge trapping ramp and quadratic baseline described in Section 4.1, we analyze the statistical properties of the residuals. We include all points that are not marked as outliers (all points within 5 $\sigma$ of the best fit model) which includes 775 out of 780 total integrations. The standard deviation (scatter) in all of the non-outlier residuals is 152 ppm, compared to a theoretical limit of 107 ppm from photon and read noise. Thus, the measured noise is 42% larger than the theoretical limit, some of which could be due to 1/f noise (Schlawin et al., 2020). Even after ROEBA subtraction, there remains higher frequency (shorter timescale than the 5.24 ms row read time) noise. We note that the background annulus subtraction reduced the standard deviation of out-of- transit flux in the lightcurve by 37% over skipping the background annulus subtraction so there is likely residual 1/f noise. We next assess time- correlations in the data by binning the residuals. A key metric in the performance of the NIRCam photometry is how well the noise scales as a function of bin size. This commonly is presented in the form of an Allan Variance plot (e.g. Pont et al., 2006; Croll et al., 2011). In this plot, the photometric scatter is computed as a standard deviation as a function of bin size. In the case where each time sample is independent, the noise drops as $1/\sqrt{N}$ and this is plotted as “white noise scaling”. When the measured photometric scatter begins to diverge from the $1/\sqrt{N}$ power law, this is a sign that the time samples are correlated and time-dependent systematics are present. Figure 10 (left) shows that the noise falls with $1/\sqrt{N}$. Thus, no noise floor is measurable down to the 20 ppm level, but 1/f noise limits the precision in our current analysis to 42% above the theoretical limit. It may be possible to optimize the aperture to a hexagonal aperture, but previous tests with a circular annulus had little difference in noise performance on simulated data. For this Allan variance analysis, we have removed the first tilt event from the lightcurve using a model discussed in Section 3.1, a quadratic trend with time and an exponential settling ramp discussed in Section 3.4. For comparison, we also calculated an Allan variance curve when the baseline trend was fit with a linear polynomial instead of a quadratic polynomial in time. In this curve, the noise begins to grow near 5 minutes where the noise is about 60 ppm. Thus, there is some curvature to the lightcurve either related to instrument or astrophysical trends. Figure 10: The lightcurve errors per bin drop nearly as $1/\sqrt{N}$ for $N$ time points, after correcting for the exponential startup, quadratic trend and transit (left plot curve). Thus, the noise is largely independent for each integration. If we only include a linear baseline trend to fit the lightcurves (right), there is excess noise beginning for bin sizes around 4 minutes. The precision before time binning is about 42% worse than the ideal limit of photon and read noise, likely because of 1/f noise correlations between pixels within each frame. The measured 20 ppm lightcurve scatter at a time bin size of 30 minutes can be compared to the state-of-the-art best precision lightcurves from space-based photometers. For the brightest targets observed by Kepler, the precision was measured to be $\sim$15 ppm (Jenkins et al., 2010). For the bright exoplanet system 55 Cnc, the precision with TESS was 10 ppm for 30 minutes (Meier Valdés et al., 2022) and for CHEOPS the precision achievable on a 1.6 hour eclipse observed 41 times using only out-of-eclipse data was 3 ppm (Demory et al., 2022), which would scale to 3 ppm * sqrt(1.6 hours/0.5 hours * 41 eclipses /2)=20 ppm for 30 minutes. The scaling uses the relative time of the eclipse and the half hour benchmark, 41 separate eclipse events that were averaged and finally a factor of 2 accounting for the fact that the out-of-transit to in- transit ratio adds another factor of sqrt(2) in noise. Elsewhere, the instrumental noise on hour-long timescales for CHEOPS is reported to be 15 to 80 ppm (Maxted et al., 2022). In the JWST photometry presented in this work, precision is limited by photon counting statistics, so brighter sources with more photons per minute will be needed to assess what is the highest possible precision as compared to the state-of-the-art performance. ### 4.3 Pointing Performance Target acquisition was successful and placed the target at X=1060.7 px, Y=167.5 px on the NRCA3 detector in full frame 1-based Data Management System coordinates. Pointing was stable to 0.01 pixels (0.3 mas) in the X direction and 0.009 px (0.3 mas) in the Y direction. Some of the higher frequency jitter is thus averaged over the 27 second long integrations. Thus, JWST attitude control provides incredible pointing stability at the 10-2 pixel level and this produces no noticeable changes in flux with position and jitter, likely constituting a negligible part of the error budget. The guide star has an FGS magnitude of 15.3 and we expect similarly high precision pointing performance for guide stars from magnitude 12.5 to 15.5. Under the HgCdTe crosshatching structures on the NIRCam ALONG/A5 detector, 2 mas pointing was expected to produce 6 ppm changes of flux, so 0.3 mas jitter is likely to matter at the single digit ppm level and flux changes at this level cannot be measured for the HAT-P-14 target. ### 4.4 High Gain Antenna Move JWST high gain antenna moves maintain pointing at the designated ground stations on Earth’s Deep Space Network. While high gain antenna moves should happen during slews or between exposures, the moves are permitted to occur after 10,000 s during long time series observations with JWST. We had a high gain antenna adjustment at 2022-05-02T10:10:01 UTC, which was measurable with the Fine Guidance Sensor and NIRCam short wavelength time series centroids. As shown in Figure 11, the pointing change from the HGA move settles very quickly in less than 0.5 minutes. Furthermore, the position was returned back to the original pointing to within 1 mas. The data around an HGA move can be discarded and in the case of weak lens photometry that is spread over many pixels, only produced a transient 500 ppm change in flux. Thus, HGA moves are not expected to cause significant issues to most time series observations as long as the Fine Guide mode does not lose a guide star from its subarray. Figure 11: The High Gain Antenna moves are unlikely to have a big impact on time series observations. The HGA move only affects about 0.5 minutes of data, which is mostly corrected by shifting the aperture but could be excluded from the lightcurve if it is an outlier. The normalized flux changes by a maximum of about 0.5 parts per thousand (ppt). The plate scale of NIRCam’s Short Wavelength detector is 31 mas/px . ### 4.5 Timing Accuracy We find a transit center time of 2022-05-02T10:34:55 +/- 7 seconds BJDTDB, which is consistent (within 1.7 $\sigma$) with a prediction from the TESS ephemeris of 2022-05-02T10:36:35.1 +/- 60 seconds BJDTDB. Other planets with higher precision ephemerides will test JWST timing to higher accuracy. There is also a dedicated GO program program identification number 1666 (PI Poshak Gandhi) that will use a double white dwarf binary and is expected to calibrate the absolute timing of an exposure to the 100 ms level. ### 4.6 HAT-P-14 B Contamination The HAT-P-14 system was observed from the ground with Palomar adaptive optics (AO) on 2021-08-08 UTC to assess the contamination and dilution of the transit depth and find any unknown companions. No companions were detected with a contrast less than $\Delta$mag = 7.0 within 0.5″. The nearby stellar companion HAT-P-14 B (Ngo et al., 2015) was also imaged with Palomar AO to better constrain the infrared colors and contamination on the transit depth. The separation is 0.85″ at a position angle of 264 degrees, which is well within the 2.5″ source aperture used in the lightcurves of the central (hexagonal) portion of the PSF shown in Figure 1. With the Palomar AO imaging, we find a delta K magnitude between the HAT-P-14 A and B of $\Delta$K=4.99, which is a similar wavelength to this JWST F210M photometry. The JWST target acquisition image shown in Figure 12 shows both HAT-P-14 A and HAT-P-14 B. Using the target acquisition image, we find that the contrast is $\Delta$ MagF335M=5.03 $\pm$ 0.05 for the F335M filter (3.35 µm) and the separation is 0.83″. This contrast is a similar to $K$ band, as expected for the Raleigh Jeans limit for long wavelengths. Using webbpsf, we simulate the overlap of two Weak Lens +8 PSFs for a 81 pixel aperture and estimate the transit depth dilution. We use the flight wavefront optical path difference evaluated at a time of 2022-05-02T10:30 UTC using webbpsf and perform aperture photometry on the simulated PSFs of HAT-P-14 A and HAT-P-14 B separately. For the F210M filter at 2.1 µm, we find that HAT-P-14 B should dilute the fitted 6540 $\pm$ 22 ppm transit depth by 59 ppm for an undiluted transit depth 6599 $\pm$ 22 ppm. This is within 3.2 $\sigma$ of the long wavelength F322W2 broadband result of 6670 ppm, but with independent fitting methods and priors on orbital parameters and limb darkening. This is also consistent with the NIRSpec time series broadband fit with 6627 $\pm$ 8 ppm with an independent analysis (Espinoza et al., 2022), with different priors on orbital parameters and limb darkening. We do not expect the atmosphere to contribute much more than 15 ppm as described in Section 1 for 0.9 atmospheric scale heights, but atmospheric gases could contribute at 1.5$\sigma$ level for 2 atmospheric scale heights of absorption. Figure 12: Long Wavelength Target Acquisition centered near HAT-P-14 A with a faint ($\Delta$mag=5) stellar companion HAT-P-14 B (upper left) at 860 mas separation that dilutes the transit depth. ## 5 Conclusions Transiting exoplanet science and other time-variability studies will benefit from JWST high precision time series photometry using the defocused photometry that is collected simultaneously with the grism time series mode. Here, we present performance results from the hot Jupiter HAT-P-14 b that was observed to commission the instruments with a known flat transmission spectrum signal. We detect the planet transit at high precision and find that the transit depth is consistent with the NIRSpec result to within 28 ppm $\pm$ 20 ppm as expected for this relatively high gravity planet. The weak lens time series is particularly informative about small changes in the PSF as a result of wavefront variations, and we found two clear tilt events where two different hexagonal primary mirror segments quickly changed orientation with a timescale $\lesssim 1.4$ seconds. Difference images clearly show the tilt events associated with specific mirror segments. These tilt events can cause jumps in time series’ signal levels but can be predicted effectively for short wavelengths using a wavefront model. We also show how tilt events can be sensed with other instrument modes that do not use defocusing optics. FGS differential images and the FWHM of the PSF can both reveal tilt events as step functions and will be good metrics to assess mirror stability. The NRCA3 detector exhibited a charge trapping persistence ramp that is similar in shape to the ones seen on HST, but it settles out quickly on a 5-15 minutes decay timescale with an amplitude of 660 to 700 ppm and does not have a big impact on the time series. The overall precision is very high and has a low scatter over the residuals of 152 ppm compared to the photon and read noise value of 107 ppm per 27 second integration, so it is only 41% above the theoretical limit for this quiescent F-type $K$=8.85 star. Some of this excess noise may be due to residual 1/f noise not corrected by the row-by-row subtraction. The noise bins down approximately with the square root of the number of points in a time bin, indicating that systematic noise is minimal after the exponential and jump corrections and a quadratic polynomial as a function of time. Overall, the prospects are good for high precision time series measurements of exoplanets and other astrophysical phenomena. Defoucsed photometric monitoring will be valuable for measuring mirror tilt events and mitigating their impact on high precision measurements. ## acknowledgements Funding for E Schlawin is provided by NASA’s Goddard Space Flight Center. T Greene acknowledges funding from the JWST project via NASA WBS 411672.05.05.02.02. This research has made use of the SIMBAD database, operated at CDS, Strasbourg, France; and NASA’s Astrophysics Data System Bibliographic Services. Part of the research was carried out at the Jet Propulsion Laboratory, California Institute of Technology, under a contract with the National Aeronautics and Space Administration (80NM0018D0004). We respectfully acknowledge the University of Arizona is on the land and territories of Indigenous peoples. Today, Arizona is home to 22 federally recognized tribes, with Tucson being home to the O’odham and the Yaqui. Committed to diversity and inclusion, the University strives to build sustainable relationships with sovereign Native Nations and Indigenous communities through education offerings, partnerships, and community service. Thank you to Eddie Bergeron for sharing analysis on detecting tilt events with FGS. ## Appendix A Subarray Positions To help orient JWST users on the subarrays and relative positions of the short wavelength and long wavelength detector images, we provide diagrams of the locations as measured by in-flight data. Figure 13 displays the relative locations, in telescope coordinates, of the NIRCam LW grism trace (dashed lines) and the NIRCam SW hexagonal PSF (green hexagon) from the +8 wave defocused pupil element (Weak Lens +8), which are observed simultaneously. There is a small amount of refraction with the grism so there is a vertical offset between the two positions as placed in the same coordinate system. The edges of the full frame images for the A1 (short wavelength), A3 (short wavelength) and A5 (long wavelength) detectors are all shown as dotted lines while the edges of the SUBGRISM128 subarrays for the same 3 detectors are shown as solid lines. The positions here are shown for Cycle 1 after commissioning where small (8-10 SW px) tweaks were made to better-center the A1 and A3 subarrays on the hexagonal PSF. The reference pixels on JWST NIRCam detectors are at the full frame boundaries, so the relative positions of the both SUBGRISM64 and SUBGRISM128 subarrays necessitate excluding the bottom 4 reference pixel rows on the short wavelength NRCA1 and NRCA3 detectors. Significant improvements in performance can be made if background pixels are used in the same manner as reference pixels to correct for amplifier offsets and 1/f noise, as discussed in Section 2.2. Figure 13: The NIRCam SW and LW channels observe approximately the same region of the sky simultaneously, however the overlap of detectors is not exact. The relative position of the SW defocused weak lens image (green hexagon) is shown compared to the LW grism F322W2 (top figure) and F444W (bottom figure) central traces (brown dashed lines). The SW SUBGRISM128 subarrays on the A1 and A3 detectors (blue solid lines) are aligned to capture light from the hexagonal +8 wave PSF while the the LW SUBGRISM128 subarray (solid red line) is aligned to capture light from the LW grism trace and surrounding background area. The full frame detector boundaries (dotted lines) are where the 4 reference pixels are located, so the bottom references are included with the LW SUBGRISM128 subarray (solid red line) but no bottom reference pixels are available in the A1 and A3 SUBGRISM128 subarrays (solid blue lines). All positions are shown in V2 and V3 Telescope coordinate system using pysiaf (Sahlmann et al., 2019) after 8-10 SW pixel adjustments were made to the A1 and A3 subarrays following HAT-P-14 commissioning observations to better center the A1 and A3 subarrays on the hexagonal PSF. ## References * Agol et al. (2020) Agol, E., Luger, R., & Foreman-Mackey, D. 2020, AJ, 159, 123 [ADS] * Ahrer et al. (2022) Ahrer, E.-M., Stevenson, K. B., Mansfield, M., et al. 2022, arXiv e-prints, arXiv:2211.10489 [ADS] * Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33 [ADS] * Barstow & Irwin (2016) Barstow, J. K., & Irwin, P. G. J. 2016, MNRAS, 461, L92 [ADS] * Bean et al. (2018) Bean, J. L., Stevenson, K. B., Batalha, N. M., et al. 2018, Publications of the Astronomical Society of the Pacific, 130, 114402 [ADS] * Beatty et al. (2022) Beatty, T., Schlawin, E., Greene, T., et al. 2022, in prep * Beichman et al. (2014) Beichman, C., Benneke, B., Knutson, H., et al. 2014, PASP, 126, 1134 [ADS] * Berta et al. (2012) Berta, Z. K., Charbonneau, D., Désert, J.-M., et al. 2012, ApJ, 747, 35 [ADS] * Biesiadzinski et al. (2011) Biesiadzinski, T., Lorenzon, W., Newman, R., et al. 2011, PASP, 123, 958 [ADS] * Birkmann et al. (2022) Birkmann, S. M., Ferruit, P., Giardino, G., et al. 2022, A&A, 661, A83 [ADS] * Bitsch et al. (2021) Bitsch, B., Raymond, S. N., Buchhave, L. A., et al. 2021, A&A, 649, L5 [ADS] * Bonomo et al. (2017) Bonomo, A. S., Desidera, S., Benatti, S., et al. 2017, A&A, 602, A107 [ADS] * Bradley et al. (2016) Bradley, L., Sipocz, B., Robitaille, T., et al. 2016, astropy/photutils: v0.3, doi:10.5281/zenodo.164986 [LINK] * Bushouse et al. (2022) Bushouse, H., Eisenhamer, J., Dencheva, N., et al. 2022, Github, doi:10.5281/zenodo.7038885 [LINK] * Canipe et al. (2017) Canipe, A., Robberto, M., & Hilbert, B. 2017, JWST-STScI-005167, SM-12 [LINK] * Croll et al. (2011) Croll, B., Lafreniere, D., Albert, L., et al. 2011, AJ, 141, 30 [ADS] * Dean & Bowers (2003) Dean, B. H., & Bowers, C. W. 2003, Journal of the Optical Society of America A, 20, 1490 [ADS] * Deming et al. (2013) Deming, D., Wilkins, A., McCullough, P., et al. 2013, ApJ, 774, 95 [ADS] * Demory et al. (2022) Demory, B. O., Sulis, S., Meier Valdes, E., et al. 2022, arXiv e-prints, arXiv:2211.03582 [ADS] * Doyon et al. (2012) Doyon, R., Hutchings, J. B., Beaulieu, M., et al. 2012, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 8442, Space Telescopes and Instrumentation 2012: Optical, Infrared, and Millimeter Wave, 84422R [ADS] * Espinoza et al. (2022) Espinoza, N., Úbeda, L., Birkmann, S. M., et al. 2022, arXiv e-prints, arXiv:2211.01459 [ADS] * Fienup (1982) Fienup, J. R. 1982, Appl. Opt., 21, 2758 [ADS] * Fienup (1993) —. 1993, Appl. Opt., 32, 1737 [ADS] * Fienup (1999) —. 1999, Journal of the Optical Society of America A, 16, 1831 [ADS] * Fukui et al. (2016) Fukui, A., Narita, N., Kawashima, Y., et al. 2016, ApJ, 819, 27 [ADS] * Greene et al. (2010) Greene, T., Beichman, C., Gully-Santiago, M., et al. 2010, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 7731, Space Telescopes and Instrumentation 2010: Optical, Infrared, and Millimeter Wave, ed. J. Oschmann, Jacobus M., M. C. Clampin, & H. A. MacEwen, 77310C [ADS] * Greene et al. (2016) Greene, T. P., Line, M. R., Montero, C., et al. 2016, ApJ, 817, 17 [ADS] * Greene et al. (2017) Greene, T. P., Kelly, D. M., Stansberry, J., et al. 2017, Journal of Astronomical Telescopes, Instruments, and Systems, 3, 035001 [ADS] * Guizar-Sicairos et al. (2008) Guizar-Sicairos, M., Thurman, S. T., & Fienup, J. R. 2008, Optics Letters, 33, 156 [ADS] * Hunter (2007) Hunter, J. D. 2007, Computing In Science & Engineering, 9, 90 * Ingalls et al. (2016) Ingalls, J. G., Krick, J. E., Carey, S. J., et al. 2016, AJ, 152, 44 [ADS] * Jenkins et al. (2010) Jenkins, J. M., Caldwell, D. A., Chandrasekaran, H., et al. 2010, ApJ, 713, L120 [ADS] * Jurling et al. (2018) Jurling, A. S., Bergkoetter, M. D., & Fienup, J. R. 2018, Journal of the Optical Society of America A, 35, 1784 [ADS] * Jurling & Fienup (2014) Jurling, A. S., & Fienup, J. R. 2014, Journal of the Optical Society of America A, 31, 1348 [ADS] * Kempton et al. (2017) Kempton, E. M.-R., Bean, J. L., & Parmentier, V. 2017, ApJ, 845, L20 [ADS] * Kendrew et al. (2015) Kendrew, S., Scheithauer, S., Bouchet, P., et al. 2015, PASP, 127, 623 [ADS] * Kipping (2013) Kipping, D. M. 2013, MNRAS, 435, 2152 [ADS] * Kreidberg et al. (2018) Kreidberg, L., Line, M. R., Parmentier, V., et al. 2018, ArXiv e-prints, arXiv:1805.00029 [ADS] * Kurucz (2017) Kurucz, R. L. 2017, ATLAS9: Model atmosphere program with opacity distribution functions, Astrophysics Source Code Library, record ascl:1710.017, ascl:1710.017 [ADS] * Leisenring (2019) Leisenring, J. 2019, pynrc [LINK] * Leisenring et al. (2016) Leisenring, J. M., Rieke, M., Misselt, K., & Robberto, M. 2016, in Proc. SPIE, Vol. 9915, Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, 99152N [ADS] * Luger et al. (2019) Luger, R., Agol, E., Foreman-Mackey, D., et al. 2019, AJ, 157, 64 [ADS] * Lustig-Yaeger et al. (2019) Lustig-Yaeger, J., Meadows, V. S., & Lincowski, A. P. 2019, AJ, 158, 27 [ADS] * Maxted et al. (2022) Maxted, P. F. L., Ehrenreich, D., Wilson, T. G., et al. 2022, MNRAS, 514, 77 [ADS] * Meier Valdés et al. (2022) Meier Valdés, E. A., Morris, B. M., Wells, R. D., Schanche, N., & Demory, B. O. 2022, A&A, 663, A95 [ADS] * Mordasini et al. (2016) Mordasini, C., van Boekel, R., Mollière, P., Henning, T., & Benneke, B. 2016, ApJ, 832, 41 [ADS] * Morley et al. (2017) Morley, C. V., Kreidberg, L., Rustamkulov, Z., Robinson, T., & Fortney, J. J. 2017, ApJ, 850, 121 [ADS] * Ngo et al. (2015) Ngo, H., Knutson, H. A., Hinkley, S., et al. 2015, ApJ, 800, 138 [ADS] * Perrin et al. (2014) Perrin, M. D., Sivaramakrishnan, A., Lajoie, C.-P., et al. 2014, in Proc. SPIE, Vol. 9143, Space Telescopes and Instrumentation 2014: Optical, Infrared, and Millimeter Wave, 91433X [ADS] * Perrin et al. (2016) Perrin, M. D., Acton, D. S., Lajoie, C.-P., et al. 2016, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 9904, Space Telescopes and Instrumentation 2016: Optical, Infrared, and Millimeter Wave, ed. H. A. MacEwen, G. G. Fazio, M. Lystrup, N. Batalha, N. Siegler, & E. C. Tong, 99040F [ADS] * Plazas et al. (2017) Plazas, A. A., Shapiro, C., Smith, R., Rhodes, J., & Huff, E. 2017, Journal of Instrumentation, 12, C04009 [ADS] * Pont et al. (2006) Pont, F., Zucker, S., & Queloz, D. 2006, MNRAS, 373, 231 [ADS] * Powell et al. (2019) Powell, D., Louden, T., Kreidberg, L., et al. 2019, ApJ, 887, 170 [ADS] * Rauscher et al. (2007) Rauscher, B. J., Fox, O., Ferruit, P., et al. 2007, PASP, 119, 768 [ADS] * Rieke et al. (2005) Rieke, M. J., Kelly, D., & Horner, S. 2005, in SPIE Conf Series, Vol. 5904, Cryogenic Optical Systems and Instruments XI, ed. J. B. Heaney & L. G. Burriesci, 1–8 [ADS] * Rigby et al. (2022) Rigby, J., Perrin, M., McElwain, M., et al. 2022, arXiv e-prints, arXiv:2207.05632 [ADS] * Sahlmann et al. (2019) Sahlmann, J., Osborne, S., Cox, C., et al. 2019, spacetelescope/pysiaf: Introduce check against online version, doi:10.5281/zenodo.3516964 * Schlawin et al. (2018) Schlawin, E., Greene, T. P., Line, M., Fortney, J. J., & Rieke, M. 2018, AJ, 156, 40 [ADS] * Schlawin et al. (2020) Schlawin, E., Leisenring, J., Misselt, K., et al. 2020, AJ, 160, 231 [ADS] * Schlawin et al. (2017) Schlawin, E., Rieke, M., Leisenring, J., et al. 2017, PASP, 129, 015001 [ADS] * Schlawin et al. (2021) Schlawin, E., Leisenring, J., McElwain, M. W., et al. 2021, AJ, 161, 115 [ADS] * Smith et al. (2008) Smith, R. M., Zavodny, M., Rahmer, G., & Bonati, M. 2008, in High Energy, Optical, and Infrared Detectors for Astronomy III, Vol. 7021, 70210J [ADS] * Southworth (2012) Southworth, J. 2012, MNRAS, 426, 1291 [ADS] * Stassun et al. (2017) Stassun, K. G., Collins, K. A., & Gaudi, B. S. 2017, AJ, 153, 136 [ADS] * STScI (2016) STScI. 2016, JWST User Documentation (JDox), JWST User Documentation Website [ADS] * Thurman & Fienup (2009) Thurman, S. T., & Fienup, J. R. 2009, Journal of the Optical Society of America A, 26, 1008 [ADS] * Torres et al. (2010) Torres, G., Bakos, G. Á., Hartman, J., et al. 2010, ApJ, 715, 458 [ADS] * van der Walt et al. (2011) van der Walt, S., Colbert, S. C., & Varoquaux, G. 2011, Computing in Science and Engineering, 13, 22 [ADS] * Virtanen et al. (2020) Virtanen, P., Gommers, R., Oliphant, T. E., et al. 2020, Nature Methods, 17, 261 [LINK] * Wakeford et al. (2019) Wakeford, H. R., Wilson, T. J., Stevenson, K. B., & Lewis, N. K. 2019, Research Notes of the American Astronomical Society, 3, 7 [ADS] * Zhou et al. (2017) Zhou, Y., Apai, D., Lew, B. W. P., & Schneider, G. 2017, AJ, 153, 243 [ADS]
# Exact Controllability for a Refined Stochastic Plate Equation111This work is supported by the NSF of China under grants 12025105, 11971334 and 11931011, by the Chang Jiang Scholars Program from the Chinese Education Ministry, and by the Science Development Project of Sichuan University under grants 2020SCUNL101 and 2020SCUNL201. Qi Lü 222School of Mathematics, Sichuan University, Chengdu, P. R. China. Email<EMAIL_ADDRESS> and Yu Wang 333School of Mathematics, Sichuan University, Chengdu, P. R. China. Email<EMAIL_ADDRESS> ###### Abstract A widely used stochastic plate equation is the classical plate equation perturbed by a term of Itô’s integral. However, it is known that this equation is not exactly controllable even if the controls are effective everywhere in both the drift and the diffusion terms and also on the boundary. In some sense, this means that some key feature has been ignored in this model. Then, a one-dimensional refined stochastic plate equation is proposed and its exact controllability is established in [28]. In this paper, by means of a new global Carleman estimate, we establish the exact controllability of the multidimensional refined stochastic plate equation with two interior controls and two boundary controls. Moreover, we give a result about the lack of exact controllability, which shows that the action of two interior controls and at least one boundary control is necessary. AMS Mathematics Subject Classification. 93B05, 93B07. Keywords. Stochastic plate equation, exact controllability, observability estimate, Carleman estimate. ## 1 Introduction Let $T>0$ and $(\Omega,\mathcal{F},\mathbf{F},\mathbb{P})$ with $\mathbf{F}=\\{\mathcal{F}_{t}\\}_{t\geq 0}$ be a complete filtered probability space on which a one-dimensional standard Brownian motion $\\{W(t)\\}_{t\geq 0}$ is defined and $\mathbf{F}$ is the natural filtration generated by $W(\cdot)$, augmented by all the $\mathbb{P}$ null sets in $\mathcal{F}$. Write $\mathbb{F}$ for the progressive $\sigma$-field with respect to $\mathbf{F}$. Let $H$ be a Banach space. Denote by $L^{2}_{\mathcal{F}_{t}}(\Omega;H)$ the space of all $\mathcal{F}_{t}$-measurable random variables $\xi$ such that $\mathbb{E}|\xi|_{H}^{2}<\infty$; by $L_{\mathbb{F}}^{2}(0,T;H)$ the space consisting of all $H$-valued $\mathbf{F}$-adapted processes $X(\cdot)$ such that $\mathbb{E}\bigl{(}|X(\cdot)|_{L^{2}(0,T;H)}^{2}\bigr{)}<\infty$; by $L_{\mathbb{F}}^{\infty}(0,T;H)$ the space consisting of all $H$-valued $\mathbf{F}$-adapted bounded processes; and by $C_{\mathbb{F}}([0,T];L^{2}(\Omega;H))$ the space consisting of all $H$-valued $\mathbf{F}$-adapted processes $X(\cdot)$ such that $X(\cdot):[0,T]\rightarrow L^{2}_{\mathcal{F}_{\cdot}}(\Omega;H)$ is continuous. All these spaces are Banach spaces with the canonical norms (e.g., [24, Section 2.6]). Let $G\subset\mathbb{R}^{n}$ ($n\in\mathbb{N}$) be a bounded domain with a $C^{4}$ boundary $\Gamma$. Set $Q=(0,T)\times G$ and $\Sigma=(0,T)\times\Gamma$. Denote by $\nu(x)=(\nu^{1}(x),\cdots,\nu^{n}(x))$ the unit outward normal vector of $\Gamma$ at point $x$. Consider the following refined stochastic plate equation: $\left\\{\begin{aligned} &dy=\hat{y}dt+(a_{3}y+f)dW(t)&&\quad\text{ in }Q,\\\ &d\hat{y}+\Delta^{2}ydt=(a_{1}y+a_{2}\cdot\nabla y+a_{5}g)dt+(a_{4}y+g)dW(t)&&\quad\text{ in }Q,\\\ &y=h_{1},\frac{\partial y}{\partial\nu}=h_{2}&&\quad\text{ on }\Sigma,\\\ &(y(0),\hat{y}(0))=(y_{0},\hat{y}_{0})&&\quad\text{ in }G.\end{aligned}\right.$ (1.1) Here, $(y_{0},\hat{y}_{0})\in H^{-1}(G)\times(H^{3}(G)\cap H^{2}_{0}(G))^{*}$ (where $(H^{3}(G)\cap H^{2}_{0}(G))^{*}$ is the dual space of $H^{3}(G)\cap H^{2}_{0}(G)$ with respect to the pivot space $L^{2}(G)$), the coefficients $\displaystyle a_{1},a_{3},a_{4}\in L^{\infty}_{\mathbb{F}}(0,T;W^{1,\infty}(G)),\quad a_{2}\in L^{\infty}_{\mathbb{F}}(0,T;W^{2,\infty}(G;\mathbb{R}^{n})),\quad a_{5}\in L^{\infty}_{\mathbb{F}}(0,T;W^{3,\infty}(G)).$ and the controls $\displaystyle(f,g,h_{1},h_{2})$ $\displaystyle\in L^{2}_{\mathbb{F}}(0,T;H^{-1}(G))\times L^{2}_{\mathbb{F}}(0,T;(H^{3}(G)\cap H^{2}_{0}(G))^{*})\times L^{2}_{\mathbb{F}}(0,T;L^{2}(\Gamma))$ $\displaystyle\quad\times L^{2}_{\mathbb{F}}(0,T;H^{-1}(\Gamma)).$ ###### Remark 1.1. The term $a_{5}g$ reflects the influence of the control $g$ in the diffusion term on the drift term, i.e., if one puts a control $g$ in the diffusion term and $a_{5}g$ will appear as a side effect. This leads to some technical difficulties in the study of the exact controllability of Eq. 1.1. The control system Eq. 1.1 is a nonhomegeneous boundary value problem. Its solution is understood in the sense of transposition. For the readers’ convenience, we recall it briefly below. A systematic introduction to that can be found in [24, Section 7.2]. First, we introduce the following reference equation: $\displaystyle\left\\{\begin{aligned} &dz=\hat{z}dt+(Z-a_{5}z)dW(t)&&\quad\text{ in }Q_{\tau},\\\ &d\hat{z}+\Delta^{2}zdt=[(a_{1}-\operatorname{div}a_{2}-a_{4}a_{5})z-a_{2}\cdot\nabla z-a_{3}\hat{Z}+a_{4}Z]dt+\hat{Z}dW(t)&&\quad\text{ in }Q_{\tau},\\\ &z=\frac{\partial z}{\partial\nu}=0&&\quad\text{ on }\Sigma_{\tau},\\\ &(z(\tau),\hat{z}(\tau))=(z^{\tau},\hat{z}^{\tau})&&\quad\text{ in }G,\end{aligned}\right.$ (1.2) where $\tau\in(0,T]$, $Q_{\tau}\mathop{\buildrel\Delta\over{=}}(0,\tau)\times G$, $\Sigma_{\tau}\mathop{\buildrel\Delta\over{=}}(0,\tau)\times\Gamma$, and $(z^{\tau},\hat{z}^{\tau})\in L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{3}(G)\cap H^{2}_{0}(G))\times L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{1}_{0}(G))$. By the classical well-posedness result for backward stochastic evolution equations (e.g.,[24, Section 4.2]), we know that (1.2) admits a unique weak solution $\displaystyle(z,Z,\hat{z},\hat{Z})$ $\displaystyle\in L^{2}_{\mathbb{F}}(\Omega;C([0,\tau];(H^{3}(G)\cap H^{2}_{0}(G))))\times L^{2}_{\mathbb{F}}(0,\tau;(H^{3}(G)\cap H^{2}_{0}(G)))$ $\displaystyle\quad\times L^{2}_{\mathbb{F}}(\Omega;C([0,\tau];H_{0}^{1}(G)))\times L^{2}_{\mathbb{F}}(0,\tau;H_{0}^{1}(G)).$ Furthermore, for $0\leq s,t\leq\tau$, it holds that $\displaystyle|(z(t),\hat{z}(t))|_{L^{2}_{\mathcal{F}_{t}}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L^{2}_{\mathcal{F}_{t}}(\Omega;H_{0}^{1}(G))}$ (1.3) $\displaystyle\leq C(|(z(s),\hat{z}(s))|_{L^{2}_{\mathcal{F}_{s}}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L^{2}_{\mathcal{F}_{s}}(\Omega;H_{0}^{1}(G))}$ $\displaystyle\qquad+|(Z,\hat{Z})|_{L^{2}_{\mathbb{F}}(0,\tau;H^{3}(G)\cap H_{0}^{2}(G))\times L^{2}_{\mathbb{F}}(0,\tau;H_{0}^{1}(G))}).$ Here and in what follows, we denote by $C$ a generic positive constant depending on $G$, $T$, $\tau$ and $a_{i}$, $i=1,\cdots,5$, whose value may vary from line to line. Next, we give the following hidden regularity for solutions to Eq. 1.2. ###### Proposition 1.1. Let $(z^{\tau},\hat{z}^{\tau})\in L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{3}(G)\cap H^{2}_{0}(G))\times L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{1}_{0}(G))$. Then the solution $(z,Z,\hat{z},\hat{Z})$ of Eq. 1.2 satisfies $|\nabla\Delta z||_{\Gamma}\in L^{2}_{\mathbb{F}}(0,\tau;L^{2}(\Gamma))$. Furthermore, $\displaystyle|\nabla\Delta z|_{L^{2}_{\mathbb{F}}(0,\tau;L^{2}(\Gamma))}$ $\displaystyle\leq C|(z^{\tau},\hat{z}^{\tau})|_{L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{3}(G)\cap H^{2}_{0}(G))\times L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{1}_{0}(G))}.$ Proof of Proposition 1.1 is put in Section 2. Now we are in a position to give the definition of the transposition solution to Eq. 1.1. ###### Definition 1.1. A pair of stochastic processes $(y,\hat{y})\in C_{\mathbb{F}}([0,T];L^{2}(\Omega;H^{-1}(G)))\times$ $C_{\mathbb{F}}([0,T];$ $L^{2}(\Omega;(H^{3}(G)\cap H^{2}_{0}(G))^{*}))$ is a transposition solution to Eq. 1.1 if for any $\tau\in(0,T]$ and $(z^{\tau},\hat{z}^{\tau})\in L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{3}(G)\cap H^{2}_{0}(G))\times$ $L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{1}_{0}(G))$, we have $\begin{array}[]{ll}\displaystyle\mathbb{E}\langle\hat{y}(\tau),z^{\tau}\rangle_{(H^{3}(G)\cap H^{2}_{0}(G))^{*},H^{3}(G)\cap H^{2}_{0}(G)}-\langle\hat{y}_{0},z(0)\rangle_{(H^{3}(G)\cap H^{2}_{0}(G))^{*},H^{3}(G)\cap H^{2}_{0}(G)}\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle-\mathbb{E}\langle y(\tau),\hat{z}^{\tau}\rangle_{H^{-1}(G),H_{0}^{1}(G)}+\langle y_{0},\hat{z}(0)\rangle_{H^{-1}(G),H_{0}^{1}(G)}\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle=-\mathbb{E}\int_{0}^{\tau}\langle f,\widehat{Z}\rangle_{H^{-1}(G),H_{0}^{1}(G)}dt+\mathbb{E}\int_{0}^{\tau}\langle g,Z\rangle_{(H^{3}(G)\cap H^{2}_{0}(G))^{*},H^{3}(G)\cap H^{2}_{0}(G)}dt\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\quad+\mathbb{E}\int_{0}^{\tau}\int_{\Gamma}\frac{\partial\Delta z}{\partial\nu}h_{1}d\Gamma dt-\mathbb{E}\int_{0}^{\tau}\langle h_{2},\Delta z\rangle_{H^{-1}(\Gamma),H^{1}(\Gamma)}dt.\end{array}$ Here, $(z,Z,\hat{z},\hat{Z})$ solves Eq. 1.2. Combining Proposition 1.1 and the well-posedness for stochastic evolution equation with unbounded control operator in the sense of transposition solution (e.g., [24, Theorem 7.12]), we immediately get the following well- posedness result for Eq. 1.1. ###### Proposition 1.2. For each $(y_{0},\hat{y}_{0})\in H^{-1}(G)\times(H^{3}(G)\cap H^{2}_{0}(G))^{*}$, the system Eq. 1.1 admits a unique transposition solution $(y,\hat{y})$. Moreover, $\displaystyle|(y,\hat{y})|_{C_{\mathbb{F}}([0,T];L^{2}(\Omega;H^{-1}(G)))\times C_{\mathbb{F}}([0,T];L^{2}(\Omega;(H^{3}(G)\cap H^{2}_{0}(G))^{*}))}$ $\displaystyle\leq C\big{(}|y_{0}|_{H^{-1}(G)}+|\hat{y}_{0}|_{(H^{3}(G)\cap H^{2}_{0}(G))^{*}}+|f|_{L^{2}_{\mathbb{F}}(0,T;H^{-1}(G))}+|g|_{L^{2}_{\mathbb{F}}(0,T;(H^{3}(G)\cap H^{2}_{0}(G))^{*})}$ $\displaystyle\quad\quad+|h_{1}|_{L^{2}_{\mathbb{F}}(0,T;L^{2}(\Gamma))}+|h_{2}|_{L^{2}_{\mathbb{F}}(0,T;H^{-1}(\Gamma))}\big{)}.$ Now we give the definition of the exact controllability for Eq. 1.1. ###### Definition 1.2. The system Eq. 1.1 is called exactly controllable at time $T$ if for any $(y_{0},\hat{y}_{0})\in H^{-1}(G)\times(H^{3}(G)\cap H^{2}_{0}(G))^{*}$ and $(y_{1},\hat{y}_{1})\in L^{2}_{\mathcal{F}_{T}}(\Omega;H^{-1}(G))\times L^{2}_{\mathcal{F}_{T}}(\Omega;(H^{3}(G)\cap H^{2}_{0}(G))^{*})$, there exist controls $\displaystyle(f,g,h_{1},h_{2})$ $\displaystyle\in L^{2}_{\mathbb{F}}(0,T;H^{-1}(G))\times L^{2}_{\mathbb{F}}(0,T;(H^{3}(G)\cap H^{2}_{0}(G))^{*})\times L^{2}_{\mathbb{F}}(0,T;L^{2}(\Gamma))$ $\displaystyle\quad\times L^{2}_{\mathbb{F}}(0,T;H^{-1}(\Gamma))$ such that the solution $(y,\hat{y})$ to Eq. 1.1 satisfies that $(y(T,\cdot),\hat{y}(T,\cdot))=(y_{1},\hat{y}_{1})$, ${\mathbb{P}}$-a.s. ###### Remark 1.2. In the definition of the exact controllability for Eq. 1.1, we put the state space to be $L^{2}_{\mathcal{F}_{T}}(\Omega;H^{-1}(G))\times L^{2}_{\mathcal{F}_{T}}(\Omega;(H^{3}(G)\cap H^{2}_{0}(G))^{*})$. It is natural to choose the state space as $L^{2}_{\mathcal{F}_{T}}(\Omega;L^{2}(G))\times L^{2}_{\mathcal{F}_{T}}(\Omega;H^{-2}(G))$. Further, the controls $g\in L^{2}_{\mathbb{F}}(0,T;(H^{3}(G)\cap H^{2}_{0}(G))^{*})$ and $h_{2}\in L^{2}_{\mathbb{F}}(0,T;H^{-1}(\Gamma))$ are very irregular. We expect to use more regular controls to achieve the desired goal. However, we do not know how to do that now. Indeed, even for the deterministic plate equation, to the best of our knowledge, the existing results (e.g., [15]) can only prove the exact controllability in the space $H^{-1}(G)\times(H^{3}(G)\cap H^{2}_{0}(G))^{*}$ with controls in the Dirichlet and Neumann boundary conditions. The main result of this paper is the following. ###### Theorem 1.3. The system Eq. 1.1 is exactly controllable at any time $T>0$. ###### Remark 1.3. Similar to the derivation process in [28, 23], the refined stochastic plate equation Eq. 1.1 can be obtained from the classical stochastic plate equation: $\left\\{\begin{aligned} &dy_{t}+\Delta^{2}ydt=(a_{1}y+a_{2}\cdot\nabla y+f)dt+(a_{4}y+g)dW(t)&&\quad\text{ in }Q,\\\ &y=h_{1},\frac{\partial y}{\partial\nu}=h_{2}&&\quad\text{ on }\Sigma,\\\ &(y(0),y_{t}(0))=(y_{0},y_{1})&&\quad\text{ in }G.\end{aligned}\right.$ (1.4) Here, $(y_{0},y_{1})$ are the initial data, and $f,g,h_{1},h_{2}$ are controls. The system Eq. 1.4 is widely used in structural engineering, and can be applied to beams, bridges and other structures, see [3, 12, 4, 2]. In particular, Eq. 1.4 can be used to characterize fluttering or large-amplitude vibration of an elastic panel excited by aerodynamic forces which are perturbed by random fluctuations (e.g., [3]). However, similar to Theorem 4.1 in [28] and Theorem 2.1 in [23], one can show that the system Eq. 1.4 is not exactly controllable for any $T>0$. Inspired by the negative controllability result, and similar to [28, 23], we study a refined stochastic plate equation Eq. 1.1. We put four controls in the system Eq. 1.1. Similarly to Theorem 2.3 in [23], one can find that boundary controls $h_{1}$ and $h_{2}$ in Eq. 1.1 can not be dropped simultaneously, and internal controls $f$ and $g$ must be acted on the whole domain $G$. More precisely, we have the following result. ###### Theorem 1.4. The system Eq. 1.1 is not exactly controllable at any time $T>0$ provided that one of the following three conditions is satisfied: 1. (1). $a_{3}\in C_{\mathbb{F}}([0,T];L^{\infty}(G))$, $G\backslash\overline{G_{0}}\neq\emptyset$ and $\operatorname{supp}f\subset G_{0}$; 2. (2). $a_{4}\in C_{\mathbb{F}}([0,T];L^{\infty}(G))$, $G\backslash\overline{G_{0}}\neq\emptyset$ and $\operatorname{supp}g\subset G_{0}$; 3. (3). $h_{1}=h_{2}=0$. ###### Remark 1.4. It is worth studying whether one the of boundary controls can be removed. This can be done for deterministic plate equation (e.g., [15, 18]). However, we have no idea for how to do that. ###### Remark 1.5. By letting $G=(0,1)$, we can deduce from Theorem 3.2 that the system Eq. 1.1 is exactly controllable with controls in any nonempty subset of the boundary $\Gamma$, which has recently been proved in [28]. In fact, thanks to Remark 1.7, for any $(y_{0},\hat{y}_{0})$ and $(y_{1},\hat{y}_{1})$ satisfing Definition 1.2, one can find $(f,g,h_{1},h_{2})\in L^{2}_{\mathbb{F}}(0,T;H^{-1}(G))\times L^{2}_{\mathbb{F}}(0,T;(H^{3}(G)\cap H^{2}_{0}(G))^{*})\times(L^{2}_{\mathbb{F}}(0,T))^{2}$ such that the solution $(y,\hat{y})$ to Eq. 1.1, where the boundary conditions are $\displaystyle y(\cdot,0)=h_{1},\quad y_{x}(\cdot,0)=h_{2},\quad y(\cdot,1)=0,\quad y_{x}(\cdot,1)=0,\quad\text{~{} on }(0,T)$ satisfies that $(y(T,\cdot),\hat{y}(T,\cdot))=(y_{1},\hat{y}_{1})$. In the multidimensional case, it is worth to studying whether Eq. 1.1 is still exactly controllable under the assumption that $(h_{1},h_{2})\in(L^{2}_{\mathbb{F}}(0,T;L^{2}(\Gamma_{0})))^{2}$, where $\Gamma_{0}$ is a nonempty subset of $\Gamma$. By a standard duality argument, Theorem 1.3 is equivalent to the following observability estimate (e.g., [24, Theorem 7.17]). ###### Theorem 1.5. There exists a constant $C>0$ such that for every $(z^{T},\hat{z}^{T})\in L_{\mathcal{F}_{T}}^{2}(\Omega;H^{3}(G)\cap H^{2}_{0}(G))\times L_{\mathcal{F}_{T}}^{2}(\Omega;H^{1}_{0}(G))$, it holds that $\displaystyle|(z^{T},\hat{z}^{T})|^{2}_{L_{\mathcal{F}_{T}}^{2}(\Omega;H^{3}(G)\cap H^{2}_{0}(G))\times L_{\mathcal{F}_{T}}^{2}(\Omega;H^{1}_{0}(G))}$ $\displaystyle\leq C\mathbb{E}\int_{\Sigma}(|\nabla\Delta z|^{2}+|\Delta z|^{2})d\Gamma dt+C|(Z,\hat{Z})|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{3}(G)\cap H^{2}_{0}(G))\times L^{2}_{\mathbb{F}}(0,T;H^{1}_{0}(G))},$ where $(z,Z,\hat{z},\hat{Z})$ is the solution to the equation Eq. 1.2 with $\tau=T$, $z(T)=z^{T}$, and $\hat{z}(T)=\hat{z}^{T}$. ###### Remark 1.6. A sharp trace estimate for the deterministic plate equation is established in [17]. It suggests that we may need better regularity than $H_{0}^{2}(G)\times L^{2}(G)$ for the initial datum of dual system Eq. 1.2 to estimate $|\nabla\Delta z|$ on the boundary. This is the reason that we choose the final data of Eq. 1.2 in $L_{\mathcal{F}_{T}}^{2}(\Omega;H^{3}(G)\cap H^{2}_{0}(G))\times L_{\mathcal{F}_{T}}^{2}(\Omega;H^{1}_{0}(G))$. ###### Remark 1.7. Let $G=(0,1)$. By choosing $x_{0}>1$ in Eq. 3.4, from the proof of Theorems 3.2 and 1.5, we can deduce that $\displaystyle|(z^{T},\hat{z}^{T})|^{2}_{L_{\mathcal{F}_{T}}^{2}(\Omega;H^{3}(G)\cap H^{2}_{0}(G))\times L_{\mathcal{F}_{T}}^{2}(\Omega;H^{1}_{0}(G))}$ $\displaystyle\leq C\mathbb{E}\int_{0}^{T}\big{(}|z_{xxx}(0)|^{2}+|z_{xx}(0)|^{2}\big{)}dt+C|(Z,\hat{Z})|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{3}(G)\cap H^{2}_{0}(G))\times L^{2}_{\mathbb{F}}(0,T;H^{1}_{0}(G))}.$ This is the main result in [28]. ###### Remark 1.8. One can also consider the exact controllability of a refined stochastic palte equation when the boundary controls are as follows: $\displaystyle y=h_{1}\quad\mbox{ and }\quad\Delta y=h_{2}\text{~{} on ~{}}\Sigma.$ By the standard duality argument and the technique of transforming the controllability of the forward stochastic equation into the controllability of a backward equation (e.g., [24, Section 7.5]), we only need to prove that for any $(z_{0},\hat{z}_{0})\in\\{\eta\in H^{3}(G)|\Delta\eta\in H^{1}_{0}(G)\\}\times H^{1}_{0}(G)$, it holds that $\displaystyle|(\Delta z_{0},\hat{z}_{0})|^{2}_{H_{0}^{1}(G)\times H_{0}^{1}(G)}\leq C\mathbb{E}\int_{0}^{T}\int_{\Gamma}\bigg{(}\bigg{|}\frac{\partial\Delta z}{\partial\nu}\bigg{|}^{2}+\bigg{|}\frac{\partial z}{\partial\nu}\bigg{|}^{2}\bigg{)}d\Gamma dt,$ (1.5) where $(z,\hat{z})$ is the solution to $\left\\{\begin{aligned} &dz=\hat{z}dt-a_{5}zdW(t)&&\quad\text{ in }Q,\\\ &d\hat{z}+\Delta^{2}zdt=[(a_{1}-\operatorname{div}a_{2})z-a_{2}\cdot\nabla z-a_{4}a_{5}z]dt&&\quad\text{ in }Q,\\\ &z=\Delta z=0&&\quad\text{ on }\Sigma,\\\ &(z(0),\hat{z}(0))=(z_{0},\hat{z}_{0})&&\quad\text{ in }G.\end{aligned}\right.$ (1.6) Following the idea in [29], we can rewrite the equation (1.6) as two coupled stochastic Schrödinger equations, and the desired observability estimate can be obtained from the Carleman estimate for the latter. In fact, letting $u=i\hat{z}+\Delta z$, we have $\displaystyle idz+\Delta zdt=udt-ia_{5}zdW(t),$ and $\displaystyle-idu+\Delta udt$ $\displaystyle=[(a_{1}-\operatorname{div}a_{2})z-a_{2}\cdot\nabla z-a_{4}a_{5}z]dt+i\Delta(a_{5}z)dW(t).$ Clearly, it holds that $z=u=0$ on $\Sigma$. Thanks to Theorem 1.2 in [20], we can obtain that $\displaystyle|(\Delta z_{0},\hat{z}_{0})|^{2}_{H_{0}^{1}(G)\times H_{0}^{1}(G)}\leq C\mathbb{E}\int_{0}^{T}\int_{\Gamma}\bigg{(}\bigg{|}\frac{\partial\Delta z}{\partial\nu}\bigg{|}^{2}+\bigg{|}\frac{\partial z}{\partial\nu}\bigg{|}^{2}+\bigg{|}\frac{\partial\hat{z}}{\partial\nu}\bigg{|}^{2}\bigg{)}d\Gamma dt.$ (1.7) Then one may follow the rechnique in [16] to derive (1.5) from (1.7). On the other hand, one can also follow the technique in this paper to prove the desired observability (1.5). There are a large number of published works ([1, 9, 10, 11, 15, 18, 19, 26, 29, 5] and the references therein) studying the exact controllability for deterministic plate equations. However, as far as we know, [28] is the only published work investigating the exact controllability of stochastic beam equations, in which the authors show that Eq. 1.1 is exactly controllable when $G$ is an interval. In this paper, we shall use a stochastic Carleman estimate to prove Theorem 1.5. Such kind of estimate is one of the most useful tool in studying controllability for stochastic partial differential equations (see [6, 8, 21, 20, 24, 22, 27] and the references given there). Nevertheless, [28] is the only published work using this method to study the exact controllability of stochastic beam equations. The rest of this paper is organized as follows. In Section 2, we provide some preliminaries. Section 3 is devoted to establishing a Carleman estimate for the adjoint equation Eq. 1.2. By means of that Carleman estimate, we prove Theorem 1.5 in Section 4. At last, Section 5 is addressed to the proof of Theorem 1.4. ## 2 Some preliminary results This section provides some preliminary results. In the rest of this paper, the notation $y_{x_{i}}\equiv y_{x_{i}}(x)=\partial y(x)/\partial x_{i}$ will be used for simplicity, where $x_{i}$ is the $i$-th coordinate of a generic point $x=(x_{1},\cdots,x_{n})$ in $\mathbb{R}^{n}$. In a similar manner, we use notations $z_{x_{j}}$, $v_{x_{j}}$, etc. for the partial derivatives of $z$ and $v$ with respect to $x_{j}$. We first prove the hidden regularity for solutions to Eq. 1.2. ###### Proof of Proposition 1.1. For any $\rho\mathop{\buildrel\Delta\over{=}}(\rho^{1},\cdots,\rho^{n})\in C^{2}(\mathbb{R}^{n+1};\mathbb{R}^{n})$, by Itô’s formula and Eq. 1.2, we have $\begin{array}[]{ll}\displaystyle\sum_{j=1}^{n}(2\rho\cdot\nabla\Delta z\Delta z_{x_{j}}-\rho^{j}|\nabla\Delta z|^{2})_{x_{j}}dt\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle=-\operatorname{div}\rho|\nabla\Delta z|^{2}dt+2\rho\cdot\nabla\Delta z\big{[}(a_{1}-\operatorname{div}a_{2}-a_{4}a_{5})z-a_{2}\cdot\nabla z+a_{4}Z-a_{3}\hat{Z}\big{]}dt\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\quad+2\rho\cdot\nabla\Delta z\hat{Z}dW(t)+2\nabla\Delta zD\rho\cdot\nabla\Delta zdt-2d(\rho\cdot\nabla\Delta z\hat{z})+2\rho_{t}\cdot\nabla\Delta z\hat{z}dt\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\quad+2\rho\cdot\nabla\Delta(Z-a_{5}z)\hat{Z}dt+2\rho\cdot\nabla\Delta(Z-a_{5}z)\hat{z}dW(t)+2\operatorname{div}(\hat{z}\rho\nabla^{2}\hat{z})dt\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\quad-2\operatorname{div}(\hat{z}\nabla\hat{z}D\rho)dt+2\Delta\rho\cdot\nabla\hat{z}\hat{z}dt+2\nabla\hat{z}D\rho\cdot\nabla\hat{z}dt-\operatorname{div}(\rho|\nabla\hat{z}|^{2})dt+\operatorname{div}\rho|\nabla\hat{z}|^{2}dt.\end{array}$ (2.1) Since $\Gamma\in C^{3}$, there exists a vector field $\zeta\in C^{2}(\mathbb{R}^{n};\mathbb{R}^{n})$ such that $\zeta=\nu$ on $\Gamma$ (e.g., [13, Lemma 2.1]). Setting $\rho=\zeta$, integrating Eq. 2.1 in $Q_{\tau}$, and taking expectation on $\Omega$, we have $\displaystyle\mathbb{E}\int_{\Sigma_{\tau}}\biggl{(}2\bigg{|}\frac{\partial\Delta z}{\partial\nu}\bigg{|}^{2}-|\nabla\Delta z|^{2}\biggr{)}d\Gamma dt$ $\displaystyle=\mathbb{E}\int_{\Sigma_{\tau}}\sum_{j=1}^{n}(2\rho\cdot\nabla\Delta z\Delta z_{x_{j}}-\rho^{j}|\nabla\Delta z|^{2})\nu^{j}d\Gamma dt$ $\displaystyle=-2\mathbb{E}\int_{G}\rho\cdot\nabla\Delta z^{\tau}\hat{z}^{\tau}dx+2\mathbb{E}\int_{G}\rho\cdot\nabla\Delta z(0)\hat{z}(0)dx$ $\displaystyle\quad+\mathbb{E}\int_{Q_{\tau}}\bigl{\\{}-\operatorname{div}\rho|\nabla\Delta z|^{2}+2\rho\cdot\nabla\Delta z\big{[}(a_{1}-\operatorname{div}a_{2}-a_{4}a_{5})z-a_{2}\cdot\nabla z+a_{4}Z-a_{3}\hat{Z}\big{]}$ $\displaystyle\qquad\qquad\quad+2\nabla\Delta zD\rho\cdot\nabla\Delta z+2\rho_{t}\cdot\nabla\Delta z\hat{z}+2\rho\cdot\nabla\Delta(Z-a_{5}z)\hat{Z}+2\Delta\rho\cdot\nabla\hat{z}\hat{z}$ $\displaystyle\qquad\qquad\quad+2\nabla\hat{z}D\rho\cdot\nabla\hat{z}+\operatorname{div}\rho|\nabla\hat{z}|^{2}\bigr{\\}}dxdt.$ This implies $\displaystyle 2\bigg{|}\frac{\partial\Delta z}{\partial\nu}\bigg{|}^{2}_{L^{2}_{\mathbb{F}}(0,\tau;L^{2}(\Gamma))}-|\nabla\Delta z|^{2}_{L^{2}_{\mathbb{F}}(0,\tau;L^{2}(\Gamma))}\leq C|(z^{\tau},\hat{z}^{\tau})|^{2}_{L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L_{\mathcal{F}_{\tau}}^{2}(\Omega;H_{0}^{1}(G))}.$ (2.2) Denote by $\nabla_{\sigma}$ the tangential gradient on $\Gamma$. We have $\displaystyle|\nabla\Delta z|^{2}=\bigg{|}\frac{\partial\Delta z}{\partial\nu}\bigg{|}^{2}+|\nabla_{\sigma}\Delta z|^{2},$ which, together with Eq. 2.2, implies that $\displaystyle\bigg{|}\frac{\partial\Delta z}{\partial\nu}\bigg{|}^{2}_{L^{2}_{\mathbb{F}}(0,\tau;L^{2}(\Gamma))}\leq|\nabla_{\sigma}\Delta z|^{2}_{L^{2}_{\mathbb{F}}(0,\tau;L^{2}(\Gamma))}+C|(z^{\tau},\hat{z}^{\tau})|^{2}_{L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L_{\mathcal{F}_{\tau}}^{2}(\Omega;H_{0}^{1}(G))}.$ Now we are going to prove $\displaystyle|\nabla_{\sigma}\Delta z|^{2}_{L^{2}_{\mathbb{F}}(0,\tau;L^{2}(\Gamma))}\leq C|(z^{\tau},\hat{z}^{\tau})|^{2}_{L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L_{\mathcal{F}_{\tau}}^{2}(\Omega;H_{0}^{1}(G))}.$ (2.3) As the proof of Theorem 2.2 in [14], we introduce the following operator $\left\\{\begin{aligned} \mathscr{B}&=\sum_{i=1}^{n}b_{i}(x)\frac{\partial}{\partial x_{i}}=\text{ a first-order differential operator (time independent)}\\\ &\text{with coefficients $b_{i}\in C^{4}(\overline{G})$ and such that $\mathscr{B}$ is tangential to $\Gamma$, i.e.,}\\\ &\sum_{i=1}^{n}b_{i}(x)\nu^{i}=0\text{ on }\Gamma.\end{aligned}\right.$ (2.4) The operator $\mathscr{B}$ can be thought of as the pre-image, under the diffeomorphism via partitions of unity from $G$ onto half-space $\\{(x,y)\in\mathbb{R}^{n}\mid x>0,y\in\mathbb{R}^{n-1}\\}$ of the tangential derivative on the boundary $x=0$. Define $\displaystyle p\mathop{\buildrel\Delta\over{=}}\mathscr{B}z\in C_{\mathbb{F}}([0,\tau];L^{2}(\Omega;H^{2}(G))),\qquad\hat{p}\mathop{\buildrel\Delta\over{=}}\mathscr{B}\hat{z}\in C_{\mathbb{F}}([0,\tau];L^{2}(\Omega;L^{2}(G))).$ From Eq. 1.2, we have $\left\\{\begin{aligned} &dp=\hat{p}dt+g_{1}dW(t)&&\quad\text{ in }Q_{\tau},\\\ &d\hat{p}+\Delta^{2}pdt=f_{2}dt+g_{2}dW(t)&&\quad\text{ in }Q_{\tau},\\\ &p=0,\frac{\partial p}{\partial\nu}=\bigg{[}\frac{\partial}{\partial\nu},\mathscr{B}\bigg{]}z&&\quad\text{ on }\Sigma_{\tau},\\\ &(p(\tau),\hat{p}(\tau))=(\mathscr{B}z^{\tau},\mathscr{B}\hat{z}^{\tau})&&\quad\text{ in }G,\end{aligned}\right.$ (2.5) where $g_{1}=\mathscr{B}(Z-a_{5}z),\quad g_{2}=\mathscr{B}\hat{Z},$ and $f_{2}=\mathscr{B}\big{[}(a_{1}-\operatorname{div}a_{2}-a_{4}a_{5})z-a_{2}\cdot\nabla z-a_{3}\hat{Z}+a_{4}Z\big{]}+[\Delta^{2},\mathscr{B}]z,$ and $[\cdot,\cdot]$ denotes the commutator of two operators. From (2.4), we get that $\displaystyle|[\Delta^{2},\mathscr{B}]z|_{L^{2}_{\mathbb{F}}(0,\tau;H^{-1}(G))}\leq C|(z^{\tau},\hat{z}^{\tau})|_{L_{\mathcal{F}_{\tau}}^{2}(\Omega;(H^{3}(G)\cap H_{0}^{2}(G)))\times L_{\mathcal{F}_{\tau}}^{2}(\Omega;H_{0}^{1}(G))}$ (2.6) and that $\begin{array}[]{ll}\displaystyle\mathbb{E}\int_{\Sigma_{\tau}}|\nabla_{\sigma}\Delta z|^{2}d\Gamma dt=\mathbb{E}\int_{\Sigma_{\tau}}|\mathscr{B}\Delta z|^{2}d\Gamma dt=\mathbb{E}\int_{\Sigma_{\tau}}|\Delta p|^{2}d\Gamma dt+\mathbb{E}\int_{\Sigma_{\tau}}|[\mathscr{B},\Delta]z|^{2}d\Gamma dt.\end{array}$ (2.7) By (2.4) again, we find that $\begin{array}[]{ll}\displaystyle\mathbb{E}\int_{\Sigma_{\tau}}|[\mathscr{B},\Delta]z|^{2}d\Gamma dt&\displaystyle\leq C\mathbb{E}\int_{0}^{\tau}|z|_{H^{3}(G)}^{2}dt\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr&\displaystyle\leq C|(z^{\tau},\hat{z}^{\tau})|^{2}_{L_{\mathcal{F}_{\tau}}^{2}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L_{\mathcal{F}_{\tau}}^{2}(\Omega;H_{0}^{1}(G))}.\end{array}$ (2.8) Combining Eqs. 2.7 and 2.8, to show Eq. 2.3, we only need to prove $\displaystyle\mathbb{E}\int_{\Sigma_{\tau}}|\Delta p|^{2}d\Gamma dt\leq C|(z^{\tau},\hat{z}^{\tau})|^{2}_{L_{\mathcal{F}_{\tau}}^{2}(\Omega;(H^{3}(G)\cap H_{0}^{2}(G)))\times L_{\mathcal{F}_{\tau}}^{2}(\Omega;H_{0}^{1}(G))}.$ (2.9) For any $\rho\mathop{\buildrel\Delta\over{=}}(\rho^{1},\cdots,\rho^{n})\in C^{2}(\mathbb{R}^{n+1};\mathbb{R}^{n})$, by Itô’s formula and Eq. 2.5, we have $\begin{array}[]{ll}\displaystyle\sum_{i,k,l=1}^{n}(\rho^{i}p_{x_{k}x_{k}}p_{x_{l}x_{l}})_{x_{i}}dt\\\\[-2.84526pt] \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle=\sum_{i,k,l=1}^{n}\big{[}\rho^{i}_{x_{i}}p_{x_{k}x_{k}}p_{x_{l}x_{l}}+(2\rho^{i}p_{x_{i}x_{k}x_{k}}p_{x_{l}})_{x_{l}}-(2\rho^{i}_{x_{l}}p_{x_{i}x_{k}}p_{x_{l}})_{x_{k}}+2\rho^{i}_{x_{l}x_{k}}p_{x_{i}x_{k}}p_{x_{l}}\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\qquad\quad~{}~{}+2\rho^{i}_{x_{l}}p_{x_{i}x_{k}}p_{x_{l}x_{k}}-(2\rho^{i}p_{x_{l}x_{k}x_{k}}p_{x_{l}})_{x_{i}}+(2\rho^{i}_{x_{i}}p_{x_{l}x_{k}}p_{x_{l}})_{x_{k}}-2\rho^{i}_{x_{i}x_{k}}p_{x_{l}x_{k}}p_{x_{l}}\\\\[8.53581pt] \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\qquad\quad~{}~{}-2\rho^{i}_{x_{i}}p_{x_{l}x_{k}}p_{x_{l}x_{k}}+(2\rho^{i}p_{x_{l}x_{k}x_{k}}p_{x_{i}})_{x_{l}}-(2\rho^{i}_{x_{l}}p_{x_{l}x_{k}}p_{x_{i}})_{x_{k}}+2\rho^{i}_{x_{l}x_{k}}p_{x_{l}x_{k}}p_{x_{i}}\\\\[8.53581pt] \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\qquad\quad~{}~{}+2\rho^{i}_{x_{l}}p_{x_{l}x_{k}}p_{x_{i}x_{k}}\big{]}dt-2\rho\cdot\nabla p(f_{2}dt+g_{2}dW(t))+d(2\rho\cdot\nabla p\hat{p})-2\rho_{t}\cdot\nabla p\hat{p}dt\\\\[8.53581pt] \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\quad+\operatorname{div}\rho\hat{p}^{2}dt-2\rho\cdot d\nabla pd\hat{p}-\operatorname{div}(\rho\hat{p}^{2})dt-2\rho\cdot\nabla g_{1}\hat{p}dW(t).\end{array}$ (2.10) Setting $\rho=\zeta$, integrating Eq. 2.10 in $Q_{\tau}$, and taking expectation on $\Omega$, we have $\displaystyle\mathbb{E}\int_{\Sigma_{\tau}}|\Delta p|^{2}d\Gamma dt$ (2.11) $\displaystyle=\mathbb{E}\int_{\Sigma_{\tau}}\sum_{i,k,l=1}^{n}\rho^{i}\nu^{i}p_{x_{k}x_{k}}p_{x_{l}x_{l}}d\Gamma dt$ $\displaystyle=2\mathbb{E}\int_{G}\rho\cdot\nabla p^{\tau}\hat{p}^{\tau}dx-2\mathbb{E}\int_{G}\rho\cdot\nabla p(0)\hat{p}(0)dx$ $\displaystyle\quad+\mathbb{E}\int_{Q_{\tau}}\big{(}\operatorname{div}\rho\hat{p}^{2}-2\rho\cdot\nabla pf_{2}-2\rho_{t}\cdot\nabla p\hat{p}-2\rho\cdot\nabla g_{1}g_{2}\big{)}dxdt$ $\displaystyle\quad+\mathbb{E}\int_{Q_{\tau}}\sum_{i,k,l=1}^{n}\big{(}\rho^{i}_{x_{i}}p_{x_{k}x_{k}}p_{x_{l}x_{l}}+2\rho^{i}_{x_{l}x_{k}}p_{x_{i}x_{k}}p_{x_{l}}+2\rho^{i}_{x_{l}}p_{x_{i}x_{k}}p_{x_{l}x_{k}}-2\rho^{i}_{x_{i}x_{k}}p_{x_{l}x_{k}}p_{x_{l}}$ $\displaystyle\qquad\qquad\qquad\quad~{}~{}-2\rho^{i}_{x_{i}}p_{x_{l}x_{k}}p_{x_{l}x_{k}}+2\rho^{i}_{x_{l}x_{k}}p_{x_{l}x_{k}}p_{x_{i}}+2\rho^{i}_{x_{l}}p_{x_{l}x_{k}}p_{x_{i}x_{k}}\big{)}dxdt$ $\displaystyle\quad+2\mathbb{E}\int_{\Sigma_{\tau}}\sum_{i,k,l=1}^{n}\big{(}\rho^{i}_{x_{i}}p_{x_{l}x_{k}}p_{x_{l}}\nu^{k}-\rho^{i}_{x_{l}}p_{x_{i}x_{k}}p_{x_{l}}\nu^{k}+\rho^{i}p_{x_{l}x_{k}x_{k}}p_{x_{i}}\nu^{l}-\rho^{i}_{x_{l}}p_{x_{l}x_{k}}p_{x_{i}}\nu^{k}\big{)}d\Gamma dt,$ which, together with Eqs. 2.6, 2.4 and 2.5, implies Eq. 2.9. Then we complete the proof. ∎ Next, we give a pointwise weighted identity, which will play an important role in the proof of Theorem 3.2. We have the following fundamental identity. ###### Theorem 2.1. Let $v$ be an $H^{4}(G)$-valued Itô process and $\hat{v}$ be an $H^{2}(G)$-valued Itô process such that $dv=(\hat{v}+f_{1})dt+g_{1}dW(t)$ for some $f_{1}\in L^{2}_{\mathbb{F}}(0,T;H^{2}(G))$ and $g_{1}\in L^{2}_{\mathbb{F}}(0,T;H^{4}(G)\cap H_{0}^{2}(G)).$ Let $\eta\in C^{2}({\mathbb{R}}\times{\mathbb{R}}^{n})$. Set $\theta=e^{\ell}$, $\ell=s\xi$, $\xi=e^{\lambda\eta}$, $w=\theta v$, and $\hat{w}=\theta\hat{v}+\ell_{t}w$. Then, for any $t\in[0,T]$ and a.e. $(x,\omega)\in G\times\Omega$, $\displaystyle 2\theta I_{2}(d\hat{v}+\Delta^{2}vdt)-2\operatorname{div}(V_{1}+V_{2})dt$ $\displaystyle=2I_{2}^{2}dt+2I_{2}I_{3}+2(M_{1}+M_{2})dt+2\sum_{i,j,k,l=1}^{n}\Lambda^{ijkl}_{1}w_{x_{i}x_{j}}w_{x_{k}x_{l}}dt+2\sum_{i,j=1}^{n}\Lambda_{2}^{ij}w_{x_{i}}w_{x_{j}}dt$ (2.12) $\displaystyle\quad+2\Lambda_{3}w^{2}dt+2\Lambda_{4}+2d(I_{2}\hat{w})-\sum_{i=1}^{n}(\Phi_{1}^{i}\hat{w}d\Delta w)_{x_{i}}.$ Here $I_{1}=\Delta^{2}wdt+\Psi_{2}\Delta wdt+\sum_{i,j=1}^{n}\Psi_{3}^{ij}w_{x_{i}x_{j}}dt+\sum_{i=1}^{n}\Psi_{4}^{i}w_{x_{i}}dt+\sum_{i=1}^{n}\Psi_{5}^{i}w_{x_{i}}dt+\Psi_{6}wdt+d\hat{w},$ (2.13) with $\begin{cases}\Psi_{2}=2s^{2}\lambda^{2}\xi^{2}|\nabla\eta|^{2},\quad\Psi_{3}^{ij}=4s^{2}\lambda^{2}\xi^{2}\eta_{x_{i}}\eta_{x_{j}},\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\Psi_{4}^{i}=12s^{2}\lambda^{3}\xi^{2}|\nabla\eta|^{2}\eta_{x_{i}}+4s^{2}\lambda^{2}\xi^{2}\Delta\eta\eta_{x_{i}},\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\Psi_{5}^{i}=\sum_{j=1}^{n}8s^{2}\lambda^{2}\xi^{2}\eta_{x_{i}x_{j}}\eta_{x_{j}},\quad\Psi_{6}=s^{4}\lambda^{4}\xi^{4}|\nabla\eta|^{4},\end{cases}\quad i,j=1,\cdots,n$ (2.14) $I_{2}=\sum_{i=1}^{n}\Phi_{1}^{i}\Delta w_{x_{i}}+\Phi_{2}\Delta w+\sum_{i,j=1}^{n}\Phi_{3}^{ij}w_{x_{i}x_{j}}+\sum_{i=1}^{n}\Phi_{4}^{i}w_{x_{i}}+\Phi_{5}w,$ (2.15) with $\begin{cases}\displaystyle\Phi_{1}^{i}=-4s\lambda\xi\eta_{x_{i}},\quad\Phi_{2}=-2s\lambda^{2}\xi|\nabla\eta|^{2}-2s\lambda\xi\Delta\eta-\lambda,\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\Phi_{3}^{ij}=4s\lambda\xi\eta_{x_{i}x_{j}}-4s\lambda^{2}\xi\eta_{x_{i}}\eta_{x_{j}},\quad\Phi_{4}^{i}=-4s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\eta_{x_{i}},\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\Phi_{5}=-6s^{3}\lambda^{4}\xi^{3}|\nabla\eta|^{4}-12s^{3}\lambda^{3}\xi^{3}(\nabla^{2}\eta\nabla\eta\nabla\eta)\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\quad\qquad-2s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\Delta\eta-s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\nabla\eta|^{4},\end{cases}i,j=1,\cdots,n$ (2.16) $\displaystyle I_{3}$ $\displaystyle=-8s\lambda\xi\sum\limits_{i,j=1}^{n}\eta_{x_{i}x_{j}}w_{x_{i}x_{j}}dt-4\nabla\Delta\ell\cdot\nabla wdt+4(\nabla\ell\cdot\nabla\Delta\ell)wdt+2|\nabla^{2}\ell|^{2}wdt$ $\displaystyle\quad-\Delta^{2}\ell wdt+|\Delta\ell|^{2}wdt+8s^{3}\lambda^{3}\xi^{3}(\nabla^{2}\eta\nabla\eta\nabla\eta)wdt+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\nabla\eta|^{4}wdt+\lambda\Delta wdt$ (2.17) $\displaystyle\quad-\ell_{t}\theta f_{1}dt-\ell_{t}\theta g_{1}dW(t)+(\ell_{t}^{2}-\ell_{tt})wdt-2\ell_{t}\hat{w}dt,$ $\displaystyle V_{1}$ $\displaystyle=[V_{1}^{1},V_{1}^{2},\cdots,V_{1}^{n}],\quad V_{2}=[V_{2}^{1},V_{2}^{2},\cdots,V_{2}^{n}],$ $\displaystyle V_{1}^{j}$ $\displaystyle=\sum_{i,k=1}^{n}\Big{[}\sum_{l=1}^{n}\Phi_{1}^{l}w_{x_{k}x_{k}x_{l}}w_{x_{i}x_{i}x_{j}}-\frac{1}{2}\sum_{l=1}^{n}\Phi_{1}^{j}w_{x_{k}x_{k}x_{l}}w_{x_{i}x_{i}x_{l}}+\frac{1}{2}\Psi_{2}\Phi_{1}^{j}w_{x_{i}x_{i}}w_{x_{k}x_{k}}$ $\displaystyle\quad\quad\quad\ \ +\sum_{l=1}^{n}\Psi_{3}^{ik}\Phi_{1}^{l}w_{x_{i}x_{k}}w_{x_{l}x_{j}}-\frac{1}{2}\sum_{l=1}^{n}\Psi_{3}^{ij}\Phi_{1}^{l}w_{x_{i}x_{k}}w_{x_{k}x_{l}}+\Phi_{4}^{k}w_{x_{i}x_{i}x_{j}}w_{x_{k}}-\Phi_{4}^{k}w_{x_{i}x_{j}}w_{x_{i}x_{k}}$ $\displaystyle\quad\quad\quad\ \ +\frac{1}{2}\Phi_{4}^{j}w_{x_{i}x_{k}}^{2}+\Big{(}\Psi_{2}\Phi_{4}^{k}\delta_{ij}-\frac{1}{2}\Psi_{2}\Phi_{4}^{j}\delta_{ik}-\frac{1}{2}\Psi_{6}\Phi_{1}^{j}\delta_{ik}+\Psi_{3}^{ij}\Phi_{4}^{k}\Big{)}w_{x_{i}}w_{x_{k}}\Big{]},$ $\displaystyle V_{2}^{j}$ $\displaystyle=\sum_{i,l,r,m=1}^{n}\Theta_{1}^{ijlrm}w_{x_{i}x_{i}x_{l}}w_{x_{r}x_{m}}+\sum_{i,k,l,r=1}^{n}\Theta_{2}^{ijklr}w_{x_{i}x_{k}}w_{x_{l}x_{r}}+\sum_{i,k,l=1}^{n}\Theta_{3}^{ijkl}w_{x_{i}x_{k}}w_{x_{l}}$ $\displaystyle\quad+\sum_{i,k=1}^{n}\Theta_{4}^{ijk}w_{x_{i}}w_{x_{k}}+\sum_{i=1}^{n}\Theta_{5}w_{x_{i}x_{i}x_{j}}w+\sum_{i,k=1}^{n}\Theta_{6}^{ijk}w_{x_{i}x_{k}}w+\sum_{i=1}^{n}\Theta_{7}^{ij}w_{x_{i}}w+\Theta_{8}^{j}w^{2}+\Theta_{9}^{j},$ $\displaystyle M_{1}$ $\displaystyle=8s\lambda^{2}\xi|\nabla\Delta w\cdot\nabla\eta|^{2}+32s^{3}\lambda^{4}\xi^{3}|\nabla^{2}w\nabla\eta\nabla\eta|^{2}+48s^{3}\lambda^{3}\xi^{3}\nabla^{2}\eta(\nabla^{2}w\nabla\eta)(\nabla^{2}w\nabla\eta)$ $\displaystyle\quad+16s^{3}\lambda^{3}\xi^{3}(\nabla^{2}w\nabla\eta\nabla\eta)\sum\limits_{i,j=1}^{n}\eta_{x_{i}x_{j}}w_{x_{i}x_{j}}-16s^{3}\lambda^{4}\xi^{3}|\nabla\eta|^{2}|\nabla^{2}w\nabla\eta|^{2}$ $\displaystyle\quad+6s^{3}\lambda^{4}\xi^{3}|\nabla\eta|^{4}|\nabla^{2}w|^{2}+4s^{3}\lambda^{3}\xi^{3}(\nabla^{2}\eta\nabla\eta\nabla\eta)|\nabla^{2}w|^{2}+2s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\Delta\eta|\nabla^{2}w|^{2}$ $\displaystyle\quad+2s^{3}\lambda^{4}\xi^{3}|\nabla\eta|^{4}|\Delta w|^{2}-4s^{3}\lambda^{3}\xi^{3}(\nabla^{2}\eta\nabla\eta\nabla\eta)|\Delta w|^{2}-2s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\Delta\eta|\Delta w|^{2}$ $\displaystyle\quad+40s^{5}\lambda^{6}\xi^{5}|\nabla\eta|^{4}|\nabla w\cdot\nabla\eta|^{2}+64s^{5}\lambda^{5}\xi^{5}(\nabla^{2}\eta\nabla\eta\nabla\eta)|\nabla w\cdot\nabla\eta|^{2}$ $\displaystyle\quad-16s^{5}\lambda^{6}\xi^{5}|\nabla\eta|^{6}|\nabla w|^{2}+8s^{7}\lambda^{8}\xi^{7}|\nabla\eta|^{8}w^{2}+\lambda|\nabla\Delta w|^{2}-s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\nabla\eta|^{4}|\Delta w|^{2}$ $\displaystyle\quad+4s^{5}\lambda^{\frac{11}{2}}\xi^{5}|\nabla\eta|^{4}|\nabla w\cdot\nabla\eta|^{2}+2s^{5}\lambda^{\frac{11}{2}}\xi^{5}|\nabla\eta|^{6}|\nabla w|^{2}-s^{7}\lambda^{\frac{15}{2}}\xi^{7}|\nabla\eta|^{8}w^{2},$ $\displaystyle M_{2}$ $\displaystyle=\biggl{(}-\frac{1}{2}\sum_{i=1}^{n}\Phi_{1x_{i}}^{i}+\Phi_{2}\biggr{)}|\nabla\hat{w}|^{2}+\sum_{i,j=1}^{n}\big{(}-\Phi_{1x_{j}}^{i}+\Phi_{3}^{ij}\big{)}\hat{w}_{x_{i}}\hat{w}_{x_{j}}$ $\displaystyle\quad+\frac{1}{2}\biggl{[}\sum_{i,j=1}^{n}\big{(}\Phi_{1x_{i}x_{j}x_{j}}^{i}-\Phi_{2x_{i}x_{i}}\delta_{ij}-\Phi_{3x_{i}x_{j}}^{ij}+\Phi_{4x_{i}}^{i}\delta_{ij}\big{)}-2\Phi_{5}\biggr{]}\hat{w}^{2}$ $\displaystyle\quad-\sum_{i,k,j=1}^{n}\big{(}\Psi_{5}^{i}\Phi_{1}^{k}\big{)}_{x_{j}}w_{x_{i}}w_{x_{k}x_{j}}-4s^{2}\lambda^{3}\xi^{2}|\nabla^{2}w\nabla\eta|-2s^{2}\lambda^{3}\xi^{2}|\nabla\eta|^{2}|\Delta w|^{2}$ $\displaystyle\quad+s^{4}\lambda^{5}\xi^{4}|\nabla\eta|^{4}|\nabla w|^{2},$ $\displaystyle\Lambda_{1}^{ijkl}=\Phi_{3x_{i}x_{j}}^{kl}+\Phi_{3x_{i}x_{l}}^{kj}+\sum_{r=1}^{n}\Big{(}\frac{1}{2}\Phi_{2x_{r}x_{r}}\delta_{ij}\delta_{kl}-\Phi_{3x_{j}x_{r}}^{kr}\delta_{il}-\Phi_{3x_{i}x_{r}}^{kr}\delta_{lj}+\frac{1}{2}\sum_{m=1}^{n}\Phi_{3x_{r}x_{m}}^{rm}\delta_{ik}\delta_{lj}\Big{)},$ $\displaystyle\Lambda_{2}^{ij}=\sum_{k=1}^{n}\Big{[}-\Phi_{4x_{i}x_{k}x_{k}}^{j}+\frac{1}{2}\Phi_{4x_{i}x_{j}x_{k}}^{k}-2\Phi_{5x_{i}x_{j}}+(\Psi_{2}\Phi_{3}^{jk})_{x_{i}x_{k}}-\frac{1}{2}(\Psi_{2}\Phi_{3}^{ij})_{x_{k}x_{k}}$ $\displaystyle\qquad\qquad\quad-\sum_{l=1}^{n}\frac{1}{2}(\Psi_{2}\Phi_{3}^{kl}\delta_{ij})_{x_{k}x_{l}}+(\Psi_{3}^{ik}\Phi_{2})_{x_{k}x_{j}}-\sum_{l=1}^{n}\frac{1}{2}(\Psi_{3}^{kl}\Phi_{2}\delta_{ij})_{x_{k}x_{l}}-\frac{1}{2}(\Psi_{3}^{ij}\Phi_{2})_{x_{k}x_{k}}$ $\displaystyle\qquad\qquad\quad+\sum_{l=1}^{n}(\Psi_{3}^{ik}\Phi_{3}^{jl})_{x_{k}x_{l}}-\sum_{l=1}^{n}\frac{1}{2}(\Psi_{3}^{ij}\Phi_{3}^{kl})_{x_{k}x_{l}}-\frac{1}{2}(\Psi_{3}^{kl}\Phi_{3}^{ij})_{x_{k}x_{l}}+\frac{1}{2}(\Psi_{4}^{i}\Phi_{1}^{j})_{x_{k}x_{k}}$ $\displaystyle\qquad\qquad\quad+(\Psi_{4}^{i}\Phi_{2})_{x_{j}}+\frac{1}{2}(\Psi_{4}^{k}\Phi_{2}\delta_{ij})_{x_{k}}-(\Psi_{4}^{i}\Phi_{3}^{jk})_{x_{k}}+\frac{1}{2}(\Psi_{4}^{k}\Phi_{3}^{ij})_{x_{k}}+(\Psi_{5}^{i}\Phi_{2})_{x_{j}}$ $\displaystyle\qquad\qquad\quad+\frac{1}{2}(\Psi_{5}^{k}\Phi_{2}\delta_{ij})_{x_{k}}-(\Psi_{5}^{i}\Phi_{3}^{jk})_{x_{k}}+\frac{1}{2}(\Psi_{5}^{k}\Phi_{3}^{ij})_{x_{k}}\Big{]},$ $\displaystyle\Lambda_{3}=\sum_{i=1}^{n}\Big{[}\sum_{j=1}^{n}\frac{1}{2}\Phi_{5x_{i}x_{i}x_{j}x_{j}}+\sum_{j=1}^{n}\frac{1}{2}(\Psi_{2}\Phi_{5})_{x_{j}x_{j}}+\sum_{j=1}^{n}\frac{1}{2}(\Psi_{3}^{ij}\Phi_{5})_{x_{i}x_{j}}-\frac{1}{2}(\Psi_{4}^{i}\Phi_{5})_{x_{i}}-\frac{1}{2}(\Psi_{5}^{i}\Phi_{5})_{x_{i}}$ $\displaystyle\quad\quad\quad\quad-\sum_{j=1}^{n}\frac{1}{2}(\Psi_{6}\Phi_{1}^{i})_{x_{i}x_{j}x_{j}}+\frac{1}{2}(\Psi_{6}\Phi_{2})_{x_{i}x_{i}}\Big{]},$ $\displaystyle\Lambda_{4}=\sum_{i,j=1}^{n}\big{[}-\Phi_{1t}^{i}w_{x_{i}x_{j}x_{j}}+\big{(}-\Phi_{2t}\delta_{ij}-\Phi_{3t}^{ij}\big{)}w_{x_{i}x_{j}}\big{]}\hat{w}dt-\sum_{i=1}^{n}\Phi_{4t}^{i}w_{x_{i}}\hat{w}dt-\Phi_{5t}w\hat{w}dt$ $\displaystyle\qquad+\sum_{i,j=1}^{n}\big{(}\Phi^{i}_{1x_{i}}\hat{w}+\Phi^{i}_{1}\hat{w}_{x_{i}}-\Phi_{2}\hat{w}\delta_{ij}\big{)}\big{[}(\theta f_{1})_{x_{j}x_{j}}dt+(\theta g_{1})_{x_{j}x_{j}}dW(t)\big{]}-\Phi^{i}_{5}\hat{w}\big{(}\theta f_{1}dt+\theta g_{1}dW(t)\big{)}$ $\displaystyle\qquad-\sum_{i,j=1}^{n}\Phi^{ij}_{3}\hat{w}\big{[}(\theta f_{1})_{x_{i}x_{j}}dt+(\theta g_{1})_{x_{i}x_{j}}dW(t)\big{]}-\sum_{i=1}^{n}\Phi^{i}_{4}\hat{w}\big{[}(\theta f_{1})_{x_{i}}dt+(\theta g_{1})_{x_{i}}dW(t)\big{]}$ $\displaystyle\qquad-\sum_{i,j=1}^{n}\big{(}\Phi^{i}_{1}dw_{x_{i}x_{i}x_{j}}+\Phi_{2}\delta_{ij}dw_{x_{i}x_{i}}+\Phi_{3}^{ij}dw_{x_{i}x_{j}}+\Phi^{i}_{4}\delta_{ij}dw_{x_{i}}\big{)}d\hat{w}+\Phi_{5}dwd\hat{w},$ and $\displaystyle\Theta_{1}^{ijklrm}=\Phi_{2}\delta_{lj}\delta_{rm}+\Phi_{3}^{rm}\delta_{lj}-\Phi_{3}^{mj}\delta_{lr}+\Phi_{3}^{lr}\delta_{mj},$ $\displaystyle\Theta_{2}^{ijklr}=-\frac{1}{2}\Phi_{2x_{j}}\delta_{ik}\delta_{lr}-\Phi_{3x_{i}}^{lr}\delta_{kj}+\Phi_{3x_{k}}^{rj}\delta_{il}+\sum_{m=1}^{n}\Phi_{3x_{m}}^{rm}\delta_{kj}\delta_{il}-\frac{1}{2}\sum_{m=1}^{n}\Phi_{3x_{m}}^{jm}\delta_{il}\delta_{kr}-\Phi_{3x_{i}}^{kl}\delta_{rj},$ $\displaystyle\Theta_{3}^{ijkl}=-\Phi_{4x_{i}}^{l}\delta_{kj}-\Phi_{5}\delta_{ik}\delta_{lj}+\Psi_{2}\Phi_{3}^{ik}\delta_{lj}-\Psi_{2}\Phi_{3}^{ij}\delta_{kl}+\Psi_{3}\Phi_{2}\delta_{ik}-\Psi_{3}\Phi_{2}\delta_{ij}+\Psi_{3}^{ik}\Phi_{3}^{lj}$ $\displaystyle\qquad\quad\;-\Psi_{3}^{ij}\Phi_{3}^{kl}+\Psi_{4}^{l}\Phi_{1}^{k}\delta_{ij}+\Psi_{5}^{l}\Phi_{1}^{k}\delta_{ij},$ $\displaystyle\Theta_{4}^{ijk}=\Phi_{4x_{i}x_{j}}^{k}-\frac{1}{2}\Phi_{4x_{i}x_{k}}^{j}+2\Phi_{5x_{i}}\delta_{kj}-\Phi_{5x_{j}}\delta_{ik}-(\Psi_{2}\Phi_{3}^{jk})_{x_{i}}+\frac{1}{2}(\Psi_{2}\Phi_{3}^{ik})_{x_{j}}$ $\displaystyle\qquad\quad+\sum_{l=1}^{n}\frac{1}{2}(\Psi_{2}\Phi_{3}^{jl})_{x_{l}}\delta_{ik}-\sum_{l=1}^{n}(\Psi_{3}^{il}\Phi_{2})_{x_{l}}\delta_{kj}+\frac{1}{2}\sum_{l=1}^{n}(\Psi_{3}^{lj}\Phi_{2})_{x_{l}}\delta_{ik}+\frac{1}{2}(\Psi_{3}^{ik}\Phi_{2})_{x_{j}}$ $\displaystyle\qquad\quad-\sum_{l=1}^{n}(\Psi_{3}^{ij}\Phi_{3}^{kl})_{x_{l}}+\frac{1}{2}\sum_{l=1}^{n}(\Psi_{3}^{ik}\Phi_{3}^{jl})_{x_{l}}+\frac{1}{2}\sum_{l=1}^{n}(\Psi_{3}^{lj}\Phi_{3}^{ik})_{x_{l}}-\frac{1}{2}(\Psi_{4}^{i}\Phi_{1}^{k})_{x_{j}}+\Psi_{4}^{i}\Phi_{2}\delta_{kj}$ $\displaystyle\qquad\quad-\frac{1}{2}\Psi_{4}^{j}\Phi_{2}\delta_{ik}+\Psi_{4}^{i}\Phi_{3}^{kj}-\frac{1}{2}\Psi_{4}^{j}\Phi_{3}^{ik}+\Psi_{5}^{i}\Phi_{2}\delta_{kj}-\frac{1}{2}\Psi_{5}^{j}\Phi_{2}\delta_{ik}+\Psi_{5}^{i}\Phi_{3}^{kj}-\frac{1}{2}\Psi_{5}^{j}\Phi_{3}^{ik},$ $\displaystyle\Theta_{5}=\Phi_{5},$ $\displaystyle\Theta_{6}^{ijk}=-\Phi_{5x_{j}}\delta_{ik}+\Psi_{6}\Phi_{1}^{k}\delta_{ij},$ $\displaystyle\Theta_{7}^{ij}=\sum_{k=1}^{n}\Phi_{5x_{k}x_{k}}\delta_{ij}+\Psi_{2}\Phi_{5}\delta_{ij}+\Psi_{3}^{ij}\Phi_{5}-(\Psi_{6}\Phi_{1}^{i})_{x_{j}}+\Psi_{6}\Phi_{2}\delta_{ij}+\Psi_{6}\Phi_{3}^{ij},$ $\displaystyle\Theta_{8}^{j}=-\frac{1}{2}\sum_{i=1}^{n}\Phi_{5x_{i}x_{i}x_{j}}-\frac{1}{2}(\Psi_{2}\Phi_{5})_{x_{j}}-\frac{1}{2}\sum_{i=1}^{n}(\Psi_{3}^{ij}\Phi_{5})_{x_{i}}+\frac{1}{2}\Psi_{4}^{j}\Phi_{5}+\frac{1}{2}\Psi_{5}^{j}\Phi_{5}$ $\displaystyle\quad\quad\ +\frac{1}{2}\sum_{i=1}^{n}(\Psi_{6}\Phi_{1}^{j})_{x_{i}x_{i}}-\frac{1}{2}(\Psi_{6}\Phi_{2})_{x_{j}}-\frac{1}{2}\sum_{i=1}^{n}(\Psi_{6}\Phi_{3}^{ij})_{x_{i}}+\frac{1}{2}\Psi_{6}\Phi_{4}^{j}.$ $\displaystyle\Theta_{9}^{j}=\sum_{i=1}^{n}\biggl{(}\Phi^{i}_{1x_{i}}\hat{w}_{x_{j}}\hat{w}+\Phi_{1}^{i}\hat{w}_{x_{i}}\hat{w}_{x_{j}}-\frac{1}{2}\Phi_{1x_{i}x_{j}}^{i}\hat{w}^{2}-\frac{1}{2}\Phi_{1}^{j}\hat{w}_{x_{i}}^{2}-\Phi_{3}^{ij}\hat{w}\hat{w}_{x_{i}}+\frac{1}{2}\Phi_{3x_{i}}^{ij}\hat{w}^{2}\biggr{)}$ $\displaystyle\qquad~{}-\Phi_{2}\hat{w}\hat{w}_{x_{j}}+\frac{1}{2}\Phi_{2x_{j}}\hat{w}^{2}-\frac{1}{2}\Phi_{4}^{j}\hat{w}^{2}.$ ###### Remark 2.1. Since we do not put any further assumptions on $v$ and $\hat{v}$, the identity (2.12) seems to be very complicated. For solutions to (1.2) or (1.6), many terms, such as $V_{1}$ and $V_{2}$, will merge or vanish by means of the boundary conditions. Furthermore, compared with energy terms, such as $8s\lambda^{2}\xi|\nabla\Delta w\cdot\nabla\eta|^{2}$, $2s^{3}\lambda^{4}\xi^{3}|\nabla\eta|^{4}|\Delta w|^{2}$, $40s^{5}\lambda^{6}\xi^{5}|\nabla\eta|^{4}|\nabla w\cdot\nabla\eta|^{2}$, $\lambda|\nabla\Delta w|^{2}$, many terms in (2.12) is of no great importance. We only need to estimate their orders with respect to $s$ and $\lambda$. Hence, an effective way to simplify (2.12) is that we do not write these terms explicitly and just claim that the order of $s$ and $\lambda$ for them are lower than the terms with a “good sign”. However, we do not do this since we want to provide the full details for readers, and particularly for beginners. Although the form of the identity (2.12) is very complex, its proof follows from some basic computations. To avoid defocusing the main theme of this paper, we put the proof of Theorem 2.1 in the appendix. ## 3 Carleman estimate for the adjoint equation This section is devoted to establishing a Carleman estimate for a backward stochastic plate equation. To this end, we first introduce the weight function $\eta$. For any $\delta>0,T>0$ and $0<\varepsilon_{1}<\frac{1}{2}$, we choose $x_{0}\in\mathbb{R}^{n}\backslash\overline{G}$ such that $R_{0}\mathop{\buildrel\Delta\over{=}}\min_{x\in\overline{G}}|x-x_{0}|^{2}>2\delta,$ (3.1) and choose sufficiently large $\beta$ satisfing $R_{1}\mathop{\buildrel\Delta\over{=}}\max_{x\in\overline{G}}|x-x_{0}|^{2}\leq\beta\varepsilon_{1}^{2}T^{2}-\delta.$ (3.2) We also choose sufficiently small $\varepsilon_{0}$ with $0<\varepsilon_{0}<\varepsilon_{1}$ such that $R_{0}-\beta\varepsilon_{0}^{2}T^{2}\geq\delta.$ (3.3) Let $\eta(t,x)=|x-x_{0}|^{2}-\beta\bigg{(}t-\frac{T}{2}\bigg{)}^{2}.$ (3.4) From Eqs. 3.4, 3.1, 3.2 and 3.3, it is easy to see that $\eta$ satisfies the following conditions. ###### Condition 3.1. 1. (1). $|\eta(t,x)|_{C^{2}(\overline{Q})}\leq C_{1}$. 2. (2). $|\nabla\eta(t,x)|\geq C_{2}>0,\quad\forall~{}(t,x)\in\overline{Q}$. 3. (3). For all $(t,x)$ in $J_{1}\mathop{\buildrel\Delta\over{=}}[(0,T/2-\varepsilon_{1}T)\cup(T/2+\varepsilon_{1}T,T)]\times G,$ it holds that $\eta(t,x)\leq-\delta$. 4. (4). For all $(t,x)$ in $J_{2}\mathop{\buildrel\Delta\over{=}}(T/2-\varepsilon_{0}T,T/2+\varepsilon_{0}T)\times G,$ it holds that $\eta(t,x)\geq\delta$. Recall that $\theta=e^{\ell}$, $\ell=s\xi$ and $\xi=e^{\lambda\eta}$. With $\eta$ given by (3.4), the functions $\ell$ and $\theta$ are also defined. We also need the following known result. ###### Lemma 3.1. [7, Theorem 2.1] Let $q\in H^{2}_{0}(G)$. Then there exists a constant $C>0$ independent of $s$ and $\lambda$, and parameter $\widehat{\lambda}>1$ and $\widehat{s}>1$ such that, for all $\lambda\geq\widehat{\lambda}$ and $s\geq\widehat{s}$, $\displaystyle s^{4}\lambda^{6}\int_{Q}\xi^{4}\theta^{2}\big{(}s^{2}\lambda^{2}\xi^{2}|q|^{2}+|\nabla q|^{2}\big{)}dxdt\leq C\int_{Q}s^{3}\lambda^{4}\xi^{3}\theta^{2}|\Delta q|^{2}dxdt.$ We have the following Carleman estimate. ###### Theorem 3.2. There exist constants $C>0$ and $\lambda_{0}>0$ such that for all $\lambda\geq\lambda_{0}$, one can find $s_{0}=s_{0}(\lambda)>0$ so that for any $s\geq s_{0}$, $(v,\hat{v})\in L^{2}_{\mathbb{F}}(\Omega;C([0,T];H^{4}(G)\cap H_{0}^{2}(G)))\times L^{2}_{\mathbb{F}}(\Omega;C([0,T];H_{0}^{2}(G))),$ and $f_{1},f_{2},g_{2}\in L^{2}_{\mathbb{F}}(0,T;H_{0}^{2}(G)),\quad g_{1}\in L^{2}_{\mathbb{F}}(0,T;H^{4}(G)\cap H_{0}^{2}(G))$ satisfying $\left\\{\begin{array}[]{ll}\displaystyle dv=(\hat{v}+f_{1})dt+g_{1}dW(t)&\text{ in }Q,\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle d\hat{v}+\Delta^{2}vdt=f_{2}dt+g_{2}dW(t)&\text{ in }Q,\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle v=\frac{\partial v}{\partial\nu}=0&\text{ on }\Sigma,\end{array}\right.$ (3.5) and $v(0,\cdot)=v(T,\cdot)=\hat{v}(0,\cdot)=\hat{v}(T,\cdot)=0\quad\mbox{ in }G,\quad\text{$\mathbb{P}$-\rm a.s.},$ (3.6) it holds that $\displaystyle\mathbb{E}\int_{Q}\theta^{2}\bigl{(}s\lambda\xi|\nabla\hat{v}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\hat{v}|^{2}+\lambda|\nabla\Delta v|^{2}+s^{2}\lambda^{4}\xi^{2}|\nabla^{2}v|^{2}+s^{3}\lambda^{4}\xi^{3}|\Delta v|^{2}$ $\displaystyle\qquad\quad+s^{4}\lambda^{6}\xi^{4}|\nabla v|^{2}+s^{6}\lambda^{8}\xi^{6}|v|^{2}\bigr{)}dxdt$ $\displaystyle\leq C\mathbb{E}\int_{Q}\theta^{2}\big{(}s^{6}\lambda^{6}\xi^{6}f_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla f_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}f_{1}|^{2}+f_{2}^{2}+s^{6}\lambda^{6}\xi^{6}g_{1}^{2}$ (3.7) $\displaystyle\qquad\qquad\quad~{}+s^{4}\lambda^{4}\xi^{4}|\nabla g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}g_{1}|^{2}+|\nabla\Delta g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|g_{2}|^{2}\big{)}dxdt$ $\displaystyle\quad+C\mathbb{E}\int_{\Sigma}\theta^{2}\bigr{(}s\lambda\xi|\nabla\Delta v|^{2}+s^{3}\lambda^{3}\xi^{3}|\Delta v|^{2}\bigl{)}d\Gamma dt.$ ###### Proof of Theorem 3.2. In what follows, for a positive integer $r$, we denote by $O(\lambda^{r})$ a function of order $\lambda^{r}$ for large $\lambda$. Similarly, we use the notation $O(e^{C\lambda})$. In order to shorten the formulae, we define $\displaystyle\mathcal{A}_{1}\mathop{\buildrel\Delta\over{=}}\mathbb{E}\int_{Q}\big{(}s^{5}O(e^{C\lambda})+s^{6}\xi^{6}O(\lambda^{7})\big{)}|w|^{2}dxdt,$ $\displaystyle\mathcal{A}_{2}\mathop{\buildrel\Delta\over{=}}\mathbb{E}\int_{Q}\big{(}s^{3}O(e^{C\lambda})+s^{4}\xi^{4}O(\lambda^{5})\big{)}|\nabla w|^{2}dxdt,$ $\displaystyle\mathcal{A}_{3}\mathop{\buildrel\Delta\over{=}}\mathbb{E}\int_{Q}\big{(}sO(e^{C\lambda})+s^{2}\xi^{2}O(\lambda^{3})\big{)}|\nabla^{2}w|^{2}dxdt,$ $\displaystyle\mathcal{A}_{4}\mathop{\buildrel\Delta\over{=}}\mathbb{E}\int_{Q}O(1)|\nabla\Delta w|^{2}dxdt,$ (3.8) $\displaystyle\mathcal{B}_{2}\mathop{\buildrel\Delta\over{=}}\mathbb{E}\int_{\Sigma}\big{(}s^{2}O(e^{C\lambda})+s^{3}\xi^{3}O(\lambda^{2})\big{)}|\nabla^{2}w|^{2}dxdt,$ $\displaystyle\hat{\mathcal{A}}_{1}\mathop{\buildrel\Delta\over{=}}\mathbb{E}\int_{Q}\big{(}s^{2}O(e^{C\lambda})+s^{3}\xi^{3}O(\lambda^{3})\big{)}|\hat{w}|^{2}dxdt,$ $\displaystyle\hat{\mathcal{A}}_{2}\mathop{\buildrel\Delta\over{=}}\mathbb{E}\int_{Q}\big{(}O(\lambda)+s\xi O(1)\big{)}|\nabla\hat{w}|^{2}dxdt,$ and $\mathcal{A}\mathop{\buildrel\Delta\over{=}}\mathcal{A}_{1}+\mathcal{A}_{2}+\mathcal{A}_{3}+\mathcal{A}_{4}+\hat{\mathcal{A}}_{1}+\hat{\mathcal{A}}_{2}.$ Integrating the equality Eq. 2.12 on $Q$, taking mathematical expectation in both sides, and noting Eq. 3.6, we obtain $\displaystyle 2\mathbb{E}\int_{Q}\theta I_{2}(d\hat{v}+\Delta^{2}vdt)dx-2\mathbb{E}\int_{Q}\operatorname{div}(V_{1}+V_{2})dxdt$ $\displaystyle=2\mathbb{E}\int_{Q}I_{2}^{2}dxdt+2\mathbb{E}\int_{Q}I_{2}I_{3}dx+2\mathbb{E}\int_{Q}(M_{1}+M_{2})dxdt$ (3.9) $\displaystyle\quad+2\sum_{i,j,k,l=1}^{n}\mathbb{E}\int_{Q}\Lambda^{ijkl}_{1}w_{x_{i}x_{j}}w_{x_{k}x_{l}}dxdt+2\sum_{i,j=1}^{n}\mathbb{E}\int_{Q}\Lambda_{2}^{ij}w_{x_{i}}w_{x_{j}}dxdt$ $\displaystyle\quad+2\mathbb{E}\int_{Q}\Lambda_{3}w^{2}dxdt+2\mathbb{E}\int_{Q}\Lambda_{4}dx.$ By Condition 3.1, Eqs. 2.14, 2.16 and 3, we have $\sum_{i,j,k,l=1}^{n}\mathbb{E}\int_{Q}\Lambda^{ijkl}_{1}w_{x_{i}x_{j}}w_{x_{k}x_{l}}dxdt\geq-C\mathbb{E}\int_{Q}s\lambda^{4}\xi|\nabla^{2}w|^{2}dxdt\geq\mathcal{A}_{3},$ (3.10) $\sum_{i,j=1}^{n}\mathbb{E}\int_{Q}\Lambda_{2}^{ij}w_{x_{i}}w_{x_{j}}dxdt\geq-C\mathbb{E}\int_{Q}\big{(}s^{3}\lambda^{6}\xi^{3}+s^{2}\lambda^{6}\xi^{2}\big{)}|\nabla^{2}w|^{2}dxdt\geq\mathcal{A}_{2},$ (3.11) $\mathbb{E}\int_{Q}\Lambda_{3}w^{2}dxdt\geq-C\mathbb{E}\int_{Q}s^{3}\lambda^{8}\xi^{3}\big{(}1+s\xi+s^{2}\xi^{2}\big{)}|\nabla w|^{2}dxdt\geq\mathcal{A}_{1}$ (3.12) and $\displaystyle\mathbb{E}\int_{Q}\Lambda_{4}dx$ $\displaystyle\geq-C\mathbb{E}\int_{Q}\big{[}|\nabla\Delta w|^{2}+sO(e^{C\lambda})|\nabla^{2}w|^{2}+s^{3}O(e^{C\lambda})|\nabla w|^{2}+s^{5}O(e^{C\lambda})|w|^{2}$ $\displaystyle\qquad\qquad\quad~{}+s^{2}O(e^{C\lambda})|\hat{w}|^{2}+s^{3}\lambda^{3}\xi^{3}|\hat{w}|^{2}+(s\xi+\lambda)|\nabla\hat{w}|^{2}\big{]}dxdt$ $\displaystyle\quad-C\mathbb{E}\int_{Q}\theta^{2}\big{(}s^{6}\lambda^{6}\xi^{6}f_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla f_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}f_{1}|^{2}+s^{6}\lambda^{6}\xi^{6}g_{1}^{2}$ (3.13) $\displaystyle\qquad\qquad\quad~{}+s^{4}\lambda^{4}\xi^{4}|\nabla g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}g_{1}|^{2}+|\nabla\Delta g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|g_{2}|^{2}\big{)}dxdt$ $\displaystyle\geq-C\mathbb{E}\int_{Q}\theta^{2}\big{(}s^{6}\lambda^{6}\xi^{6}f_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla f_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}f_{1}|^{2}+s^{6}\lambda^{6}\xi^{6}g_{1}^{2}$ $\displaystyle\qquad\qquad\quad~{}+s^{4}\lambda^{4}\xi^{4}|\nabla g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}g_{1}|^{2}+|\nabla\Delta g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|g_{2}|^{2}\big{)}dxdt+\mathcal{A}.$ It follows from (3.9)–(3.13) that $\displaystyle 2\mathbb{E}\int_{Q}\theta I_{2}(d\hat{v}+\Delta^{2}vdt)dx-2\mathbb{E}\int_{Q}\operatorname{div}(V_{1}+V_{2})dxdt$ $\displaystyle\geq 2\mathbb{E}\int_{Q}I_{2}^{2}dxdt+2\mathbb{E}\int_{Q}I_{2}I_{3}dx+2\mathbb{E}\int_{Q}(M_{1}+M_{2})dxdt$ (3.14) $\displaystyle\qquad-C\mathbb{E}\int_{Q}\theta^{2}\big{(}s^{6}\lambda^{6}\xi^{6}f_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla f_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}f_{1}|^{2}+s^{6}\lambda^{6}\xi^{6}g_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla g_{1}|^{2}$ $\displaystyle\qquad\qquad\qquad\quad~{}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}g_{1}|^{2}+|\nabla\Delta g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|g_{2}|^{2}\big{)}dxdt+\mathcal{A}.$ We will estimate the terms in Eq. 3.14 one by one. The procedure is divided into three steps. _Step 1._ In this step, we consider the divergence terms. Thanks to the boundary conditions satisfied by $v$, it is easy to check that $\displaystyle w$ $\displaystyle=0,\quad\nabla w=0,\quad\nabla^{2}w=\theta\nabla^{2}v\quad\mbox{ on }\Sigma.$ (3.15) We also have $\displaystyle\frac{\partial w_{x_{i}}}{\partial\nu}\nu^{j}=w_{x_{i}x_{j}}=w_{x_{j}x_{i}}=\frac{\partial w_{x_{j}}}{\partial\nu}\nu^{i},$ (3.16) which implies $\displaystyle|\Delta w|^{2}$ $\displaystyle=\sum_{i,j=1}^{n}w_{x_{i}x_{i}}w_{x_{j}x_{j}}=\sum_{i,j=1}^{n}\frac{\partial w_{x_{i}}}{\partial\nu}\nu^{i}\frac{\partial w_{x_{j}}}{\partial\nu}\nu^{j}$ $\displaystyle=\sum_{i,j=1}^{n}\frac{\partial w_{x_{i}}}{\partial\nu}\nu^{j}\frac{\partial w_{x_{j}}}{\partial\nu}\nu^{i}=\sum_{i,j=1}^{n}w_{x_{i}x_{j}}^{2}$ (3.17) $\displaystyle=|\nabla^{2}w|^{2}=\sum_{i=1}^{n}\biggl{(}\frac{\partial w_{i}}{\partial\nu}\biggr{)}^{2}\quad\mbox{ on }\Sigma.$ Thanks to Eq. 3.15, we obtain that $\displaystyle V_{1}\cdot\nu$ $\displaystyle=\sum_{i,j,k,l=1}^{n}\Phi_{1}^{l}w_{x_{k}x_{k}x_{l}}w_{x_{i}x_{i}x_{j}}\nu^{j}-\frac{1}{2}\sum_{i,j,k,l=1}^{n}\Phi_{1}^{j}w_{x_{k}x_{k}x_{l}}w_{x_{i}x_{i}x_{l}}\nu^{j}$ $\displaystyle\quad+\frac{1}{2}\sum_{i,j,k=1}^{n}\Psi_{2}\Phi_{1}^{j}w_{x_{i}x_{i}}w_{x_{k}x_{k}}\nu^{j}+\sum_{i,j,k,l=1}^{n}\Psi_{3}^{ik}\Phi_{1}^{l}w_{x_{i}x_{k}}w_{x_{l}x_{j}}\nu^{j}$ (3.18) $\displaystyle\quad-\frac{1}{2}\sum_{i,j,k,l=1}^{n}\Psi_{3}^{ij}\Phi_{1}^{l}w_{x_{i}x_{k}}w_{x_{k}x_{l}}\nu^{j}-\sum_{i,j,k=1}^{n}\Phi_{4}^{k}w_{x_{i}x_{j}}w_{x_{i}x_{k}}\nu^{j}$ $\displaystyle\quad+\frac{1}{2}\sum_{i,j,k=1}^{n}\Phi_{4}^{j}w_{x_{i}x_{k}}^{2}\nu^{j}\qquad\mbox{ on }\Sigma.$ By Conditions 3.1 and 2.16, we have that $\displaystyle\sum_{i,j,k,l=1}^{n}\Phi_{1}^{l}w_{x_{k}x_{k}x_{l}}w_{x_{i}x_{i}x_{j}}\nu^{j}-\frac{1}{2}\sum_{i,j,k,l=1}^{n}\Phi_{1}^{j}\nu^{j}w_{x_{k}x_{k}x_{l}}w_{x_{i}x_{i}x_{l}}\geq- Cs\lambda\xi|\nabla\Delta w|^{2}\quad\mbox{ on }\Sigma.$ (3.19) From Eqs. 2.14 and 2.16, we get that $\displaystyle\frac{1}{2}\sum_{i,j,k=1}^{n}\Psi_{2}\Phi_{1}^{j}\nu^{j}w_{x_{i}x_{i}}w_{x_{k}x_{k}}=-4s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\frac{\partial\eta}{\partial\nu}|\Delta w|^{2}\quad\mbox{ on }\Sigma.$ (3.20) Combining Eq. 2.14, Eq. 2.16, Eq. 3.16 and Eq. 3.17, we find that $\displaystyle\sum_{i,j,k,l=1}^{n}\Psi_{3}^{ik}\Phi_{1}^{l}w_{x_{i}x_{k}}w_{x_{l}x_{j}}\nu^{j}$ $\displaystyle=-16\sum_{i,j,k,l=1}^{n}s^{3}\lambda^{3}\xi^{3}\eta_{x_{i}}\eta_{x_{k}}\eta_{x_{l}}\nu^{j}w_{x_{i}x_{k}}w_{x_{l}x_{j}}$ $\displaystyle=-16\sum_{i,j,k,l=1}^{n}s^{3}\lambda^{3}\xi^{3}\eta_{x_{i}}\eta_{x_{k}}\nu^{j}w_{x_{i}x_{k}}\frac{\partial w_{j}}{\partial\nu}\frac{\partial\eta}{\partial\nu}$ (3.21) $\displaystyle=-16\sum_{i,j,k,l=1}^{n}s^{3}\lambda^{3}\xi^{3}\biggl{(}\frac{\partial\eta}{\partial\nu}\biggr{)}^{3}\biggl{(}\frac{\partial w_{j}}{\partial\nu}\biggr{)}^{2}$ $\displaystyle=-16\sum_{i,j,k,l=1}^{n}s^{3}\lambda^{3}\xi^{3}\biggl{(}\frac{\partial\eta}{\partial\nu}\biggr{)}^{3}|\Delta w|^{2}\quad\mbox{ on }\Sigma$ and $\displaystyle-\frac{1}{2}\sum_{i,j,k,l=1}^{n}\Psi_{3}^{ij}\Phi_{1}^{l}w_{x_{i}x_{k}}w_{x_{k}x_{l}}\nu^{j}-\sum_{i,j,k=1}^{n}\Phi_{4}^{k}w_{x_{i}x_{j}}w_{x_{i}x_{k}}\nu^{j}+\frac{1}{2}\sum_{i,j,k=1}^{n}\Phi_{4}^{j}w_{x_{i}x_{k}}^{2}\nu^{j}$ $\displaystyle=8\sum_{i,j,k,l=1}^{n}s^{3}\lambda^{3}\xi^{3}\biggl{(}\frac{\partial\eta}{\partial\nu}\biggr{)}^{3}|\Delta w|^{2}+2s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\frac{\partial\eta}{\partial\nu}|\Delta w|^{2}\quad\mbox{ on }\Sigma.$ (3.22) Hence, combining Conditions 3.1, 3.18, 3.19, 3.20, 3.21 and 3.22, we obtain that $\displaystyle V_{1}\cdot\nu$ $\displaystyle\geq-Cs\lambda\xi|\nabla\Delta w|^{2}-8\sum_{i,j,k,l=1}^{n}s^{3}\lambda^{3}\xi^{3}\biggl{(}\frac{\partial\eta}{\partial\nu}\biggr{)}^{3}|\Delta w|^{2}-2s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\frac{\partial\eta}{\partial\nu}|\Delta w|^{2}$ $\displaystyle\geq-Cs\lambda\xi|\nabla\Delta w|^{2}-Cs^{3}\lambda^{3}\xi^{3}|\Delta w|^{2}\quad\mbox{ on }\Sigma.$ This implies that $\displaystyle\mathbb{E}\int_{Q}\operatorname{div}V_{1}dxdt$ $\displaystyle\geq-C\mathbb{E}\int_{\Sigma}s^{3}\lambda^{3}\xi^{3}|\Delta w|^{2}d\Gamma dt-C\mathbb{E}\int_{\Sigma}s\lambda\xi|\nabla\Delta w|^{2}d\Gamma dt.$ (3.23) Thanks to Condition 3.1, Eqs. 3.15, 2.16 and 3, we have $\displaystyle\mathbb{E}\int_{Q}\operatorname{div}V_{2}dxdt=\mathbb{E}\int_{\Sigma}V_{2}\cdot\nu d\Gamma dt$ $\displaystyle=\mathbb{E}\int_{\Sigma}\sum_{j=1}^{n}\Big{(}\sum_{i,k,l,r,m=1}^{n}\Theta_{1}^{ijklrm}w_{x_{i}x_{k}x_{l}}w_{x_{r}x_{m}}+\sum_{i,k,l,r=1}\Theta_{2}^{ijklr}w_{x_{i}x_{k}}w_{x_{l}x_{r}}+\Theta_{9}^{j}\Big{)}\nu^{j}d\Gamma dt$ $\displaystyle\geq\mathbb{E}\int_{\Sigma}\Big{[}s\xi O(\lambda)|\nabla\Delta w|^{2}+sO(e^{C\lambda})|\nabla^{2}w|^{2}\Big{]}d\Gamma dt$ (3.24) $\displaystyle\geq-C\mathbb{E}\int_{\Sigma}s\lambda\xi|\nabla\Delta w|^{2}d\Gamma dt+\mathcal{B}_{2}.$ Combining Eqs. 3.14, 3.23 and 3.24, we obtain that $\displaystyle 2\mathbb{E}\int_{Q}\theta I_{2}\big{(}d\hat{v}+\Delta^{2}vdt\big{)}dx+C\mathbb{E}\int_{\Sigma}s^{3}\lambda^{3}\xi^{3}|\Delta w|^{2}d\Gamma dt$ $\displaystyle+C\mathbb{E}\int_{Q}\theta^{2}\big{(}s^{6}\lambda^{6}\xi^{6}f_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla f_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}f_{1}|^{2}+s^{6}\lambda^{6}\xi^{6}g_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla g_{1}|^{2}$ $\displaystyle\qquad\qquad\quad~{}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}g_{1}|^{2}+|\nabla\Delta g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|g_{2}|^{2}\big{)}dxdt$ (3.25) $\displaystyle+C\mathbb{E}\int_{\Sigma}s\lambda\xi|\nabla\Delta w|^{2}d\Gamma dt+\mathcal{A}+\mathcal{B}_{2}$ $\displaystyle\geq 2\mathbb{E}\int_{Q}I_{2}^{2}dxdt+2\mathbb{E}\int_{Q}I_{2}I_{3}dx+2\mathbb{E}\int_{Q}(M_{1}+M_{2})dxdt.$ _Step 2._ In this step, we study $\displaystyle 2\mathbb{E}\int_{Q}(M_{1}+M_{2})dxdt$ via integration by parts. From Condition 3.1, section 3 and Eq. 3.15, we get that $\displaystyle-\mathbb{E}\int_{Q}16s^{3}\lambda^{4}\xi^{3}|\nabla\eta|^{2}|\nabla^{2}w\nabla\eta|^{2}dxdt$ $\displaystyle=-\mathbb{E}\int_{Q}16s^{3}\lambda^{4}\xi^{3}\sum_{i,j,k,l=1}^{n}\eta_{x_{i}}^{2}\eta_{x_{k}}\eta_{x_{l}}w_{x_{k}x_{j}}w_{x_{l}x_{j}}dxdt$ $\displaystyle=\mathbb{E}\int_{\Sigma}8s^{3}\lambda^{4}\sum_{i,j,k,l=1}^{n}[-2\xi^{3}\eta_{x_{i}}^{2}\eta_{x_{k}}\eta_{x_{l}}w_{x_{k}x_{j}}w_{x_{l}}+(\xi^{3}\eta_{x_{i}}^{2}\eta_{x_{k}}\eta_{x_{l}})_{x_{j}}w_{x_{k}}w_{x_{l}}]\nu^{j}d\Gamma dt$ (3.26) $\displaystyle\quad+\mathbb{E}\int_{Q}8s^{3}\lambda^{4}\sum_{i,j,k,l=1}^{n}[2\xi^{3}\eta_{x_{i}}^{2}\eta_{x_{k}}\eta_{x_{l}}w_{x_{k}x_{j}x_{j}}w_{x_{l}}-(\xi^{3}\eta_{x_{i}}^{2}\eta_{x_{k}}\eta_{x_{l}})_{x_{j}x_{j}}w_{x_{k}}w_{x_{l}}]dxdt$ $\displaystyle\geq-4\mathbb{E}\int_{Q}s\lambda^{2}\xi|\nabla\Delta w\nabla\eta|^{2}dxdt-16\mathbb{E}\int_{Q}s^{5}\lambda^{6}\xi^{5}|\nabla\eta|^{4}|\nabla w\nabla\eta|^{2}dxdt-\mathcal{A}_{2}.$ Thanks to Eqs. 3.4 and 3.1, we know that there esists $\lambda_{1}>0$ such that for all $\lambda\geq\lambda_{1}$, it holds that $\displaystyle 64\mathbb{E}\int_{Q}s^{5}\lambda^{5}\xi^{5}(\nabla^{2}\eta\nabla\eta\nabla\eta)|\nabla w\cdot\nabla\eta|^{2}dxdt$ $\displaystyle=128\mathbb{E}\int_{Q}s^{5}\lambda^{5}\xi^{5}|\nabla\eta|^{2}|\nabla w\cdot\nabla\eta|^{2}dxdt$ (3.27) $\displaystyle\geq-\mathbb{E}\int_{Q}s^{5}\lambda^{6}\xi^{5}|\nabla\eta|^{4}|\nabla w\cdot\nabla\eta|^{2}dxdt.$ Combining Condition 3.1, Eq. 3.4, section 3 and Eq. 3.15, we get that $\displaystyle\mathbb{E}\int_{Q}\big{[}4s^{3}\lambda^{3}\xi^{3}(\nabla^{2}\eta\nabla\eta\nabla\eta)+2s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\Delta\eta\big{]}(|\nabla^{2}w|^{2}-|\Delta w|^{2})$ $\displaystyle=\frac{1}{2}\mathbb{E}\int_{Q}\Delta\big{[}4s^{3}\lambda^{3}\xi^{3}(\nabla^{2}\eta\nabla\eta\nabla\eta)+2s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\Delta\eta\big{]}|\nabla w|^{2}dxdt$ $\displaystyle\quad+\mathbb{E}\int_{Q}\nabla\big{[}4s^{3}\lambda^{3}\xi^{3}(\nabla^{2}\eta\nabla\eta\nabla\eta)+2s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\Delta\eta\big{]}\cdot\nabla w\Delta wdxdt$ (3.28) $\displaystyle\geq-\mathbb{E}\int_{Q}s^{2}\xi^{2}O(\lambda^{3})|\nabla^{2}w|^{2}dxdt-\mathbb{E}\int_{Q}\big{(}s^{3}O(e^{C\lambda})+s^{4}\xi^{4}O(\lambda^{5})\big{)}|\nabla w|^{2}dxdt$ $\displaystyle\geq-\mathcal{A}_{2}-\mathcal{A}_{3},$ and that $\displaystyle 32\mathbb{E}\int_{Q}s^{3}\lambda^{3}\xi^{3}(\nabla^{2}w\nabla\eta\nabla\eta)\sum\limits_{i,j=1}^{n}\eta_{x_{i}x_{j}}w_{x_{i}x_{j}}dxdt$ $\displaystyle=32\sum_{i,j,k,l=1}^{n}\mathbb{E}\int_{Q}s^{3}\lambda^{3}\xi^{3}\eta_{x_{k}x_{l}}\eta_{x_{i}}\eta_{x_{j}}w_{x_{i}x_{j}}w_{x_{k}x_{l}}dxdt$ $\displaystyle=32\sum_{i,j,k,l=1}^{n}\mathbb{E}\int_{\Sigma}s^{3}\lambda^{3}\xi^{3}\eta_{x_{k}x_{l}}\eta_{x_{i}}\eta_{x_{j}}(w_{x_{i}}w_{x_{k}x_{l}}\nu^{j}-w_{x_{i}}w_{x_{j}x_{l}}\nu^{k})d\Gamma dt$ $\displaystyle\quad-32\sum_{i,j,k,l=1}^{n}\mathbb{E}\int_{Q}s^{3}\lambda^{3}(\xi^{3}\eta_{x_{k}x_{l}}\eta_{x_{i}}\eta_{x_{j}})_{x_{j}}w_{x_{i}}w_{x_{k}x_{l}}dxdt$ (3.29) $\displaystyle\quad+32\sum_{i,j,k,l=1}^{n}\mathbb{E}\int_{Q}s^{3}\lambda^{3}(\xi^{3}\eta_{x_{k}x_{l}}\eta_{x_{i}}\eta_{x_{j}})_{x_{k}}w_{x_{i}}w_{x_{j}x_{l}}dxdt$ $\displaystyle\quad+32\mathbb{E}\int_{Q}s^{3}\lambda^{3}\xi^{3}\nabla^{2}\eta(\nabla^{2}w\nabla\eta)(\nabla^{2}w\nabla\eta)dxdt$ $\displaystyle\geq 32\mathbb{E}\int_{Q}s^{3}\lambda^{3}\xi^{3}\nabla^{2}\eta(\nabla^{2}w\nabla\eta)(\nabla^{2}w\nabla\eta)dxdt-\mathcal{A}_{2}-\mathcal{A}_{3}$ $\displaystyle=64\mathbb{E}\int_{Q}s^{3}\lambda^{3}\xi^{3}|\nabla^{2}w\nabla\eta|^{2}dxdt-\mathcal{A}_{2}-\mathcal{A}_{3}.$ From sections 3 and 3.15, we see that $\displaystyle\mathbb{E}\int_{Q}\bigl{[}\big{(}8s^{3}\lambda^{4}\xi^{3}-s^{3}\lambda^{\frac{7}{2}}\xi^{3}\big{)}|\nabla\eta|^{4}|\Delta w|^{2}-2\big{(}8s^{3}\lambda^{4}\xi^{3}-s^{3}\lambda^{\frac{7}{2}}\xi^{3}\big{)}|\nabla\eta|^{6}s^{2}\lambda^{2}\xi^{2}|\nabla w|^{2}$ $\displaystyle\qquad~{}+\big{(}8s^{3}\lambda^{4}\xi^{3}-s^{3}\lambda^{\frac{7}{2}}\xi^{3}\big{)}|\nabla\eta|^{8}s^{4}\lambda^{4}\xi^{4}|w|^{2}\bigr{]}dxdt$ $\displaystyle=\mathbb{E}\int_{Q}\bigl{[}\big{(}8s^{3}\lambda^{4}\xi^{3}-s^{3}\lambda^{\frac{7}{2}}\xi^{3})|\nabla\eta|^{4}(\Delta w+s^{2}\lambda^{2}\xi^{2}|\nabla\eta|^{2}w\big{)}^{2}$ (3.30) $\displaystyle\qquad\qquad-2\big{(}8s^{3}\lambda^{4}\xi^{3}-s^{3}\lambda^{\frac{7}{2}}\xi^{3}\big{)}|\nabla\eta|^{6}s^{2}\lambda^{2}\xi^{2}(|\nabla w|^{2}+w\Delta w)\bigr{]}dxdt$ $\displaystyle\geq\mathbb{E}\int_{Q}\big{(}8s^{3}\lambda^{4}\xi^{3}-s^{3}\lambda^{\frac{7}{2}}\xi^{3}\big{)}|\nabla\eta|^{4}\big{(}\Delta w+s^{2}\lambda^{2}\xi^{2}|\nabla\eta|^{2}w\big{)}^{2}dxdt-\mathcal{A}_{1}.$ Noting that $(a+b+c)^{2}\leq 3(a^{2}+b^{2}+c^{2})$ for $a,b,c\in\mathbb{R}$, we find that $\displaystyle\mathbb{E}\int_{Q}s^{3}\lambda^{4}\xi^{3}\theta^{2}|\nabla\eta|^{4}|\Delta v|^{2}dxdt$ $\displaystyle=\mathbb{E}\int_{Q}s^{3}\lambda^{4}\xi^{3}|\nabla\eta|^{4}\big{(}\Delta w-2s\lambda\xi\nabla\eta\nabla w+s^{2}\lambda^{2}\xi^{2}|\nabla\eta|^{2}w-s\lambda^{2}\xi|\nabla\eta|^{2}w-s\lambda\xi|\Delta\eta|w\big{)}^{2}dxdt$ (3.31) $\displaystyle\leq 3\mathbb{E}\int_{Q}s^{3}\lambda^{4}\xi^{3}|\nabla\eta|^{4}\big{(}\Delta w+s^{2}\lambda^{2}\xi^{2}|\nabla\eta|^{2}w\big{)}^{2}dxdt+12\mathbb{E}\int_{Q}s^{5}\lambda^{6}\xi^{5}|\nabla\eta|^{4}|\nabla w\cdot\nabla\eta|^{2}+\mathcal{A}_{1}.$ Combining sections 3, 3.27, 3, 3, 3.30 and 3.31, we know there exists $\lambda_{2}\geq\lambda_{1}$ such that for all $\lambda\geq\lambda_{2}$, it holds that $\displaystyle\mathbb{E}\int_{Q}M_{1}dxdt+\mathcal{A}\geq\mathbb{E}\int_{Q}\lambda|\nabla\Delta w|^{2}dxdt+\mathbb{E}\int_{Q}s^{3}\lambda^{4}\xi^{3}\theta^{2}|\nabla\eta|^{4}|\Delta v|^{2}dxdt.$ (3.32) It follows from Condition 3.1, Eqs. 2.14, 2.16, 3.4 and 3 that $\displaystyle\mathbb{E}\int_{Q}M_{2}dxdt+\mathcal{A}\geq\mathbb{E}\int_{Q}\big{(}16s\lambda\xi|\nabla\hat{w}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\nabla\eta|^{4}\hat{w}^{2}\big{)}dxdt.$ (3.33) Thanks to sections 3, 3.32 and 3.33, for $\lambda\geq\lambda_{2}$, we have that $\displaystyle 2\mathbb{E}\int_{Q}\theta I_{2}(d\hat{v}+\Delta^{2}vdt)dx+C\mathbb{E}\int_{\Sigma}s^{3}\lambda^{3}\xi^{3}|\Delta w|^{2}d\Gamma dt$ $\displaystyle\quad+C\mathbb{E}\int_{\Sigma}s\lambda\xi|\nabla\Delta w|^{2}d\Gamma dt+\mathcal{A}+\mathcal{B}_{2}$ $\displaystyle+C\mathbb{E}\int_{Q}\theta^{2}\big{(}s^{6}\lambda^{6}\xi^{6}f_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla f_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}f_{1}|^{2}+s^{6}\lambda^{6}\xi^{6}g_{1}^{2}$ $\displaystyle\qquad\qquad\quad~{}+s^{4}\lambda^{4}\xi^{4}|\nabla g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}g_{1}|^{2}+|\nabla\Delta g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|g_{2}|^{2}\big{)}dxdt$ (3.34) $\displaystyle\geq 2\mathbb{E}\int_{Q}I_{2}^{2}dxdt+2\mathbb{E}\int_{Q}I_{2}I_{3}dx$ $\displaystyle\quad+\mathbb{E}\int_{Q}\big{(}\lambda|\nabla\Delta w|^{2}+s^{3}\lambda^{4}\xi^{3}\theta^{2}|\Delta v|^{2}+s\lambda\xi|\nabla\hat{w}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}\hat{w}^{2}\big{)}dxdt.$ _Step 3._ In this step, we get the estimate of $v$. From (3.5), we have that $2\mathbb{E}\int_{Q}\theta I_{2}\big{(}d\hat{v}+\Delta^{2}vdt\big{)}dx\leq\mathbb{E}\int_{Q}I_{2}^{2}dxdt+\mathbb{E}\int_{Q}\theta^{2}f_{2}^{2}dxdt.$ (3.35) Thanks to Condition 3.1, Eqs. 2.17 and 3, we find that $\displaystyle 2\mathbb{E}\int_{Q}I_{2}I_{3}dx\geq-\mathbb{E}\int_{Q}I_{2}^{2}dxdt-\mathcal{A}-C\mathbb{E}\int_{Q}s^{2}\lambda^{2}\xi^{2}\theta^{2}f_{1}^{2}dxdt.$ (3.36) Noting that $\displaystyle|\nabla\Delta w|^{2}\leq C\theta^{2}\big{(}|\nabla\Delta v|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}v|^{2}\big{)}\quad\mbox{ on }\Sigma,$ by Eqs. 3.17 and 3.15, there exists $\lambda_{3}>0$ such that for all $\lambda\geq\lambda_{3}$, there is $s_{1}=s_{1}(\lambda)>0$, such that for all $s\geq s_{1}$, we have that $\begin{array}[]{ll}\displaystyle\mathcal{B}_{2}+\mathbb{E}\int_{\Sigma}s^{3}\lambda^{3}\xi^{3}|\Delta w|^{2}d\Gamma dt+\mathbb{E}\int_{\Sigma}s\lambda\xi|\nabla\Delta w|^{2}d\Gamma dt\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\leq C\mathbb{E}\int_{\Sigma}s\lambda\xi\theta^{2}\big{(}|\nabla\Delta v|^{2}+s^{2}\lambda^{2}\xi^{2}|\Delta v|^{2}\big{)}d\Gamma dt.\end{array}$ (3.37) Thanks to Eqs. 3.34, 3.35, 3.36 and 3.37, for $\lambda\geq\lambda_{3}$ and $s\geq s_{1}$, we get that $\displaystyle C\mathbb{E}\int_{\Sigma}s\lambda\xi\theta^{2}\big{(}|\nabla\Delta v|^{2}+s^{2}\lambda^{2}\xi^{2}|\Delta v|^{2}\big{)}d\Gamma dt+\mathcal{A}$ $\displaystyle+C\mathbb{E}\int_{Q}\theta^{2}\big{(}s^{6}\lambda^{6}\xi^{6}f_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla f_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}f_{1}|^{2}+f_{2}^{2}+s^{6}\lambda^{6}\xi^{6}g_{1}^{2}$ (3.38) $\displaystyle\qquad\qquad\quad~{}+s^{4}\lambda^{4}\xi^{4}|\nabla g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}g_{1}|^{2}+|\nabla\Delta g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|g_{2}|^{2}\big{)}dxdt$ $\displaystyle\geq\mathbb{E}\int_{Q}\big{(}\lambda|\nabla\Delta w|^{2}+s^{3}\lambda^{4}\xi^{3}\theta^{2}|\Delta v|^{2}+s\lambda\xi|\nabla\hat{w}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}\hat{w}^{2}\big{)}dxdt.$ By Lemma 3.1, for $\lambda\geq\max\\{\lambda_{3},\widehat{\lambda}\\}$ and $s\geq\max\\{s_{1},\widehat{s}\\}$, we obtain that $\displaystyle\mathbb{E}\int_{Q}(s^{6}\lambda^{8}\xi^{6}\theta^{2}|v|^{2}+s^{4}\lambda^{6}\xi^{4}\theta^{2}|\nabla v|^{2})dxdt\leq C\mathbb{E}\int_{Q}s^{3}\lambda^{4}\xi^{3}\theta^{2}|\Delta v|^{2}dxdt,$ which, together with section 3, implies that $\displaystyle C\mathbb{E}\int_{\Sigma}s\lambda\xi\theta^{2}\big{(}|\nabla\Delta v|^{2}+s^{2}\lambda^{2}\xi^{2}|\Delta v|^{2}\big{)}d\Gamma dt+\mathcal{A}$ $\displaystyle+C\mathbb{E}\int_{Q}\theta^{2}\big{(}s^{6}\lambda^{6}\xi^{6}f_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla f_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}f_{1}|^{2}+f_{2}^{2}+s^{6}\lambda^{6}\xi^{6}g_{1}^{2}$ (3.39) $\displaystyle\qquad\qquad\quad~{}+s^{4}\lambda^{4}\xi^{4}|\nabla g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}g_{1}|^{2}+|\nabla\Delta g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|g_{2}|^{2}\big{)}dxdt$ $\displaystyle\geq\mathbb{E}\int_{Q}\big{(}\lambda|\nabla\Delta w|^{2}+s^{3}\lambda^{4}\xi^{3}\theta^{2}|\Delta v|^{2}+s^{4}\lambda^{6}\xi^{4}\theta^{2}|\nabla v|^{2}+s^{6}\lambda^{8}\xi^{6}\theta^{2}|v|^{2}$ $\displaystyle\qquad\qquad+s\lambda\xi|\nabla\hat{w}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}\hat{w}^{2}\big{)}dxdt.$ Let $\tilde{v}=s\lambda^{2}\xi e^{s\xi}v$. Then, we have that $\displaystyle\mathbb{E}\int_{Q}s^{2}\lambda^{4}\xi^{2}\theta^{2}|\nabla^{2}v|^{2}dxdt$ $\displaystyle=\mathbb{E}\int_{Q}s^{2}\lambda^{4}\xi^{2}\theta^{2}|\nabla^{2}(s^{-1}\lambda^{-2}\xi^{-1}\theta^{-1}\tilde{v})|^{2}dxdt$ $\displaystyle\leq C\mathbb{E}\int_{Q}s^{2}\lambda^{4}\xi^{2}(s^{-2}\lambda^{-4}\xi^{-2}|\nabla^{2}\tilde{v}|^{2}+\lambda^{-2}|\nabla\tilde{v}|^{2}+s^{2}\xi^{2}|\tilde{v}|^{2})dxdt$ $\displaystyle\leq C|\tilde{v}|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{2}(G))}+C\mathbb{E}\int_{Q}s^{2}\lambda^{2}\xi^{2}|\nabla\tilde{v}|^{2}dxdt+C\mathbb{E}\int_{Q}s^{4}\lambda^{4}\xi^{4}|\tilde{v}|^{2}dxdt$ (3.40) $\displaystyle\leq C|\tilde{v}|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{2}(G))}+C\mathbb{E}\int_{Q}s^{2}\lambda^{2}\xi^{2}(s^{4}\lambda^{6}\xi^{4}\theta^{2}v^{2}+s^{2}\lambda^{4}\xi^{2}\theta^{2}|\nabla v|^{2})dxdt$ $\displaystyle\quad+C\mathbb{E}\int_{Q}s^{6}\lambda^{8}\xi^{6}\theta^{2}|v|^{2}dxdt$ $\displaystyle\leq C|\tilde{v}|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{2}(G))}+C\mathbb{E}\int_{Q}(s^{6}\lambda^{8}\xi^{6}\theta^{2}|v|^{2}+s^{4}\lambda^{6}\xi^{4}\theta^{2}|\nabla v|^{2})dxdt.$ It follows from $\tilde{v}=0$ on $\Sigma$ that $\begin{array}[]{ll}\displaystyle|\tilde{v}|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{2}(G))}\\\\[2.84526pt] \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\leq C|\Delta\tilde{v}|^{2}_{L^{2}_{\mathbb{F}}(0,T;L^{2}(G))}\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle\leq C\mathbb{E}\int_{Q}\big{(}s^{6}\lambda^{8}\xi^{6}\theta^{2}|v|^{2}+s^{4}\lambda^{6}\xi^{4}\theta^{2}|\nabla v|^{2}+s^{2}\lambda^{4}\xi^{2}\theta^{2}|\Delta v|^{2}\big{)}dxdt.\end{array}$ (3.41) Combining Eqs. 3.40 and 3.41, we obtain that $\displaystyle\mathbb{E}\int_{Q}s^{2}\lambda^{4}\xi^{2}\theta^{2}|\nabla^{2}v|^{2}dxdt\leq C\mathbb{E}\int_{Q}\big{(}s^{6}\lambda^{8}\xi^{6}\theta^{2}|v|^{2}+s^{4}\lambda^{6}\xi^{4}\theta^{2}|\nabla v|^{2}+s^{2}\lambda^{4}\xi^{2}\theta^{2}|\Delta v|^{2}\big{)}dxdt.$ (3.42) From sections 3 and 3.42, there exists $\lambda_{4}\geq\max\\{\lambda_{3},\widehat{\lambda}\\}$ such that for all $\lambda\geq\lambda_{4}$, there is an $s_{2}=s_{2}(\lambda)>\max\\{s_{1},\widehat{s}\\}$, such that for all $s\geq s_{2}$, we have that $\displaystyle C\mathbb{E}\int_{\Sigma}s\lambda\xi\theta^{2}\big{(}|\nabla\Delta v|^{2}+s^{2}\lambda^{2}\xi^{2}|\Delta v|^{2}\big{)}d\Gamma dt+\mathcal{A}$ $\displaystyle+C\mathbb{E}\int_{Q}\theta^{2}\big{(}s^{6}\lambda^{6}\xi^{6}f_{1}^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla f_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}f_{1}|^{2}+f_{2}^{2}+s^{6}\lambda^{6}\xi^{6}g_{1}^{2}$ $\displaystyle\qquad\qquad\quad~{}+s^{4}\lambda^{4}\xi^{4}|\nabla g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}g_{1}|^{2}+|\nabla\Delta g_{1}|^{2}+s^{2}\lambda^{2}\xi^{2}|g_{2}|^{2}\big{)}dxdt$ (3.43) $\displaystyle\geq\mathbb{E}\int_{Q}\big{(}\lambda|\nabla\Delta w|^{2}+s^{3}\lambda^{4}\xi^{3}\theta^{2}|\Delta v|^{2}+s^{2}\lambda^{4}\xi^{2}\theta^{2}|\nabla^{2}v|^{2}+s^{4}\lambda^{6}\xi^{4}\theta^{2}|\nabla v|^{2}+s^{6}\lambda^{8}\xi^{6}\theta^{2}|v|^{2}$ $\displaystyle\qquad\qquad+s\lambda\xi|\nabla\hat{w}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}\hat{w}^{2}\big{)}dxdt.$ Recalling $v=\theta^{-1}w$ and $\hat{v}=\theta^{-1}(\hat{w}-\ell_{t}w)$, we get that $\displaystyle\mathbb{E}\int_{Q}\theta^{2}\big{(}\lambda|\nabla\Delta v|^{2}+s\lambda\xi|\nabla\hat{v}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\hat{v}|^{2}\big{)}dxdt$ $\displaystyle=\mathbb{E}\int_{Q}\theta^{2}\big{(}\lambda|\nabla\Delta(\theta^{-1}w)|^{2}+s\lambda\xi|\nabla[\theta^{-1}(\hat{w}-\ell_{t}w)]|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}\theta^{-2}|\hat{w}-\ell_{t}w|^{2}\big{)}dxdt$ (3.44) $\displaystyle\leq C\mathbb{E}\int_{Q}\big{[}\lambda\big{(}|\nabla\Delta w|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}w|^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla w|^{2}+s^{6}\lambda^{6}\xi^{6}|w|^{2}\big{)}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}\big{(}|\hat{w}|^{2}+s^{2}\lambda^{2}\xi^{2}|w|^{2}\big{)}$ $\displaystyle\qquad\qquad~{}+s\lambda\xi\big{(}|\nabla\hat{w}|^{2}+s^{2}\lambda^{2}\xi^{2}|\hat{w}|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla w|^{2}+s^{4}\lambda^{4}\xi^{4}|w|^{2}\big{)}\big{]}dxdt$ $\displaystyle\leq C\mathbb{E}\int_{Q}\big{(}\lambda|\nabla\Delta w|^{2}+s\lambda\xi|\nabla\hat{w}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\hat{w}|^{2}\big{)}dxdt+\mathcal{A}.$ Thanks to sections 3 and 3.1, there exists $\lambda_{5}>0$ such that for all $\lambda\geq\lambda_{5}$, there is an $s_{3}=s_{3}(\lambda)>0$, such that for all $s\geq s_{3}$, we have $\displaystyle\mathcal{A}$ $\displaystyle\leq\frac{1}{C}\mathbb{E}\int_{Q}\theta^{2}\bigl{(}s\lambda\xi|\nabla\hat{v}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\hat{v}|^{2}+s^{2}\lambda^{4}\xi^{2}|\nabla^{2}v|^{2}+s^{3}\lambda^{4}\xi^{3}|\Delta v|^{2}$ $\displaystyle\qquad\qquad\quad\;+s^{4}\lambda^{6}\xi^{4}|\nabla v|^{2}+s^{6}\lambda^{8}\xi^{6}|v|^{2}\bigr{)}dxdt.$ (3.45) Let us choose $\lambda_{0}\geq\max\\{\lambda_{4},\lambda_{5}\\}$. Combining sections 3, 3.44 and 3.45, for all $\lambda\geq\lambda_{0}$, one can find $s_{0}=s_{0}(\lambda)\geq\max\\{s_{2},s_{3}\\}$ so that for any $s\geq s_{0}$, inequality Theorem 3.2 holds. ∎ ## 4 Proof of the observability estimate ###### Proof of Theorem 1.5. Let $\chi\in C_{0}^{\infty}([0,T])$ satisfy $\chi=1\text{ in }\bigg{(}\frac{T}{2}-\varepsilon_{1}T,\frac{T}{2}+\varepsilon_{1}T\bigg{)}.$ Put $v=\chi z$ and $\hat{v}=\chi\hat{z}+\chi_{t}z$ for $(z,\hat{z})$ satisfying Eq. 1.2, then $(v,\hat{v})$ fulfills $v(0,\cdot)=v(T,\cdot)=\hat{v}(0,\cdot)=\hat{v}(T,\cdot)=0$ in $G$, and solves $\left\\{\begin{array}[]{ll}\displaystyle dv=\hat{v}dt+\chi(Z-a_{5}z)dW(t)&\text{ in }Q,\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle d\hat{v}+\Delta^{2}vdt=\tilde{f}_{2}dt+\tilde{g}_{2}dW(t)&\text{ in }Q,\\\ \vskip 6.0pt plus 2.0pt minus 2.0pt\cr\displaystyle v=\frac{\partial v}{\partial\nu}=0&\text{ on }\Sigma,\end{array}\right.$ (4.1) where $\tilde{f}_{2}=\chi[(a_{1}-\operatorname{div}a_{2}-a_{4}a_{5})z-a_{2}\nabla z-a_{3}\hat{Z}+a_{4}Z]+2\chi_{t}\hat{z}+\chi_{tt}z$ and $\tilde{g}_{2}=\chi\hat{Z}+\chi_{t}(Z-a_{5}z).$ By Theorem 3.2, for $\lambda\geq\lambda_{0}$ and $s\geq s_{0}$, we have $\displaystyle\mathbb{E}\int_{Q}\theta^{2}\chi^{2}\bigl{(}\lambda|\nabla\Delta z|^{2}+s^{2}\lambda^{4}\xi^{2}|\nabla^{2}z|^{2}+s^{3}\lambda^{4}\xi^{3}|\Delta z|^{2}+s^{4}\lambda^{6}\xi^{4}|\nabla z|^{2}+s^{6}\lambda^{8}\xi^{6}|z|^{2}\bigr{)}dxdt$ $\displaystyle\leq\mathbb{E}\int_{Q}\theta^{2}\bigl{(}s\lambda\xi|\nabla\hat{v}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\hat{v}|^{2}+\lambda|\nabla\Delta v|^{2}+s^{2}\lambda^{4}\xi^{2}|\nabla^{2}v|^{2}+s^{3}\lambda^{4}\xi^{3}|\Delta v|^{2}$ $\displaystyle\qquad\qquad~{}+s^{4}\lambda^{6}\xi^{4}|\nabla v|^{2}+s^{6}\lambda^{8}\xi^{6}|v|^{2}\bigr{)}dxdt$ $\displaystyle\leq C\mathbb{E}\int_{Q}\theta^{2}\chi^{2}\big{(}s^{6}\lambda^{6}\xi^{6}z^{2}+s^{4}\lambda^{4}\xi^{4}|\nabla z|^{2}+s^{2}\lambda^{2}\xi^{2}|\nabla^{2}z|^{2}+|\nabla\Delta z|^{2}+|z|^{2}+|\nabla z|^{2}\big{)}dxdt$ $\displaystyle\quad+C\mathbb{E}\int_{J_{1}}\theta^{2}\bigr{(}|\hat{z}|^{2}+z^{2}+s^{2}\lambda^{2}\xi^{2}z^{2}\bigl{)}dxdt+C(s,\lambda)|(Z,\hat{Z})|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{3}(G))\times L^{2}_{\mathbb{F}}(0,T;H^{1}(G))}$ $\displaystyle\quad+C\mathbb{E}\int_{\Sigma}\theta^{2}\bigr{(}s\lambda\xi|\nabla\Delta z|^{2}+s^{3}\lambda^{3}\xi^{3}|\Delta z|^{2}\bigl{)}d\Gamma dt.$ This, together with Condition 3.1, implies that there exists $\tilde{\lambda}_{1}\geq\lambda_{0}$ such that for all $\lambda\geq\tilde{\lambda}_{1}$, there is $\tilde{s}_{1}=\tilde{s}_{1}(\lambda)\geq s_{0}$, so that for any $s\geq\tilde{s}_{1}$, it holds that $\displaystyle\mathbb{E}\int_{Q}\theta^{2}\bigl{(}s\lambda\xi|\nabla\hat{v}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\hat{v}|^{2}+\lambda|\nabla\Delta v|^{2}+s^{2}\lambda^{4}\xi^{2}|\nabla^{2}v|^{2}+s^{3}\lambda^{4}\xi^{3}|\Delta v|^{2}$ $\displaystyle\qquad\quad+s^{4}\lambda^{6}\xi^{4}|\nabla v|^{2}+s^{6}\lambda^{8}\xi^{6}|v|^{2}\bigr{)}dxdt$ (4.2) $\displaystyle\leq C\mathbb{E}\int_{J_{1}}\theta^{2}\bigr{(}|\hat{z}|^{2}+z^{2}+s^{2}\lambda^{2}\xi^{2}z^{2}\bigl{)}dxdt+C\mathbb{E}\int_{\Sigma}\theta^{2}\bigr{(}s\lambda\xi|\nabla\Delta z|^{2}+s^{3}\lambda^{3}\xi^{3}|\Delta z|^{2}\bigl{)}d\Gamma dt$ $\displaystyle\quad+C(s,\lambda)|(Z,\hat{Z})|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{3}(G))\times L^{2}_{\mathbb{F}}(0,T;H^{1}(G))}.$ Thanks to Condition 3.1, we obtain $\displaystyle e^{2se^{\lambda\delta}-C\lambda-6\ln s}\mathbb{E}\int_{J_{2}}\bigl{(}|\nabla\Delta z|^{2}+|\nabla^{2}z|^{2}+|\nabla z|^{2}+z^{2}+|\nabla\hat{z}|^{2}+|\hat{z}|^{2}\bigr{)}dxdt$ $\displaystyle\leq e^{2se^{\lambda\delta}}\mathbb{E}\int_{J_{2}}\bigl{(}s\lambda\xi|\nabla\hat{v}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\hat{v}|^{2}+\lambda|\nabla\Delta v|^{2}+s^{2}\lambda^{4}\xi^{2}|\nabla^{2}v|^{2}+s^{3}\lambda^{4}\xi^{3}|\Delta v|^{2}$ $\displaystyle\qquad\qquad\qquad+s^{4}\lambda^{6}\xi^{4}|\nabla v|^{2}+s^{6}\lambda^{8}\xi^{6}|v|^{2}\bigr{)}dxdt$ (4.3) $\displaystyle\leq\mathbb{E}\int_{Q}\theta^{2}\bigl{(}s\lambda\xi|\nabla\hat{v}|^{2}+s^{3}\lambda^{\frac{7}{2}}\xi^{3}|\hat{v}|^{2}+\lambda|\nabla\Delta v|^{2}+s^{2}\lambda^{4}\xi^{2}|\nabla^{2}v|^{2}+s^{3}\lambda^{4}\xi^{3}|\Delta v|^{2}$ $\displaystyle\qquad\qquad~{}+s^{4}\lambda^{6}\xi^{4}|\nabla v|^{2}+s^{6}\lambda^{8}\xi^{6}|v|^{2}\bigr{)}dxdt.$ From Eqs. 1.3 and 3.1, we see that $\displaystyle|(z^{T},\hat{z}^{T})|_{L^{2}_{\mathcal{F}_{T}}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L^{2}_{\mathcal{F}_{T}}(\Omega;H_{0}^{1}(G))}$ $\displaystyle\leq C\mathbb{E}\int_{J_{2}}\bigl{(}|\nabla\Delta z|^{2}+|\nabla^{2}z|^{2}+|\nabla z|^{2}+z^{2}+|\nabla\hat{z}|^{2}+|\hat{z}|^{2}\bigr{)}dxdt$ (4.4) $\displaystyle\quad+C(s,\lambda)|(Z,\hat{Z})|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{3}(G)\cap H_{0}^{2}(G))\times L^{2}_{\mathbb{F}}(0,T;H_{0}^{1}(G))},$ and that $\displaystyle\mathbb{E}\int_{J_{1}}\theta^{2}\big{(}|\hat{z}|^{2}+z^{2}+s^{2}\lambda^{2}\xi^{2}z^{2}\big{)}dxdt$ $\displaystyle\displaystyle\leq e^{2se^{-\lambda\delta}+C\lambda+2\ln s}\mathbb{E}\int_{J_{1}}\big{(}|\hat{z}|^{2}+z^{2}\big{)}dxdt$ (4.5) $\displaystyle\leq Ce^{2se^{-\lambda\delta}+C\lambda+2\ln s}|(z^{T},\hat{z}^{T})|_{L^{2}_{\mathcal{F}_{T}}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L^{2}_{\mathcal{F}_{T}}(\Omega;H_{0}^{1}(G))}$ $\displaystyle\quad+C(s,\lambda)|(Z,\hat{Z})|^{2}_{L^{2}_{\mathbb{F}}(0,T;H^{3}(G)\cap H_{0}^{2}(G))\times L^{2}_{\mathbb{F}}(0,T;H_{0}^{1}(G))}.$ Combining (4.2)–(4) choosing $\lambda\geq\tilde{\lambda}_{1}$ and $s\geq\tilde{s}_{1}$ such that $\displaystyle C\exp\big{(}2se^{-\lambda\delta}-2se^{\lambda\delta}+C\lambda+8\ln s\big{)}\leq\frac{1}{2},$ we get the desired observability estimate. ∎ ## 5 Proof of Theorem 1.4 The proof of Theorem 1.4 is similar to that for [23, Theorem 2.3]. We provide it here for the convenience of the readers. To begin with, we recall the following result. ###### Lemma 5.1. [25, Lemma 2.1] There exists a random variable $\zeta\in L^{2}_{\mathcal{F}_{T}}(\Omega)$ such that it is impossible to find $\varsigma_{1},\varsigma_{2}\in L^{2}_{\mathbb{F}}(0,T)\times C_{\mathbb{F}}([0,T];L^{2}(\Omega))$ and $\alpha\in\mathbb{R}$ satisfying $\displaystyle\zeta=\alpha+\int_{0}^{T}\varsigma_{1}(t)dt+\int_{0}^{T}\varsigma_{2}(t)dW(t).$ ###### Proof of Theorem 1.4. We employ a contradiction argument and divide the proof into three cases. _Case 1: $a_{3}\in C_{\mathbb{F}}([0,T];L^{\infty}(G))$, $G\backslash\overline{G_{0}}\neq\emptyset$ and $\operatorname{supp}f\subset G_{0}$_. Let $\rho\in C_{0}^{\infty}(G\backslash G_{0})$ satisfying $|\rho|_{L^{2}(G)}=1$. Suppose that Eq. 1.1 was exactly controllable. By Definition 1.2, for $(y_{0},\hat{y}_{0})=(0,0)$, there exist controls $(f,g,h_{1},h_{2})$ with $\operatorname{supp}f\subset G_{0}$ a.e. $(t,\omega)\in(0,T)\times\Omega$ such that the solution to Eq. 1.1 fulfills $(y(T),\hat{y}(T))=(\rho\zeta,0)$, where $\zeta$ is given in Lemma 5.1. Hence, $\displaystyle\rho\zeta=\int_{0}^{T}\hat{y}dt+\int_{0}^{T}(a_{3}y+f)dW(t).$ (5.1) Multiplying Eq. 5.1 by $\rho$ and integrating it in $G$, we arrive that $\displaystyle\zeta=\int_{0}^{T}\langle\hat{y},\rho\rangle_{(H^{3}(G)\cap H_{0}^{2}(G))^{*},H^{3}(G)\cap H_{0}^{2}(G)}dt+\int_{0}^{T}\langle a_{3}y,\rho\rangle_{H^{-1}(G),H_{0}^{1}(G)}dW(t),$ which contradicts Lemma 5.1. _Case 2: $a_{4}\in C_{\mathbb{F}}([0,T];L^{\infty}(G))$, $G\backslash\overline{G_{0}}\neq\emptyset$ and $\operatorname{supp}g\subset G_{0}$_. Choose $\rho$ as in Case 1. Assume that Eq. 1.1 was exactly controllable. Then, for $(y_{0},\hat{y}_{0})=(0,0)$, there exist controls $(f,g,h_{1},h_{2})$ with $\operatorname{supp}g\subset G_{0}$ a.e. $(t,\omega)\in(0,T)\times\Omega$ such that the solution to Eq. 1.1 fulfills $(y(T),\hat{y}(T))=(0,\zeta)$. Clearly, $(\phi,\hat{\phi})\mathop{\buildrel\Delta\over{=}}(\rho y,\rho\hat{y})$ satisfies $\displaystyle\left\\{\begin{aligned} &d\phi=\hat{\phi}dt+(a_{3}\phi+\rho f)dW(t)&&\quad\text{ in }Q,\\\ &d\hat{\phi}+\Delta^{2}\phi dt=\tilde{f}_{2}dt+a_{4}\phi dW(t)&&\quad\text{ in }Q,\\\ &\phi=\frac{\partial\phi}{\partial\nu}=0&&\quad\text{ on }\Sigma,\\\ &(\phi(0),\hat{\phi}(0))=(0,0)&&\quad\text{ in }G,\end{aligned}\right.$ where $\tilde{f}_{2}=[\Delta^{2},\rho]y+a_{1}\phi+\rho a_{2}\cdot\nabla\phi$. Furthermore, we have $(\phi(T),\hat{\phi}(T))=(0,\rho\zeta)$. Hence, we have $\displaystyle\zeta=-\int_{0}^{T}(\langle\Delta^{2}\phi,\rho\rangle_{H^{-5}(G),H_{0}^{5}(G)}+\langle\tilde{f}_{2},\rho\rangle_{H^{-4}(G),H_{0}^{4}(G)})dt+\int_{0}^{T}\langle a_{3}\phi,\rho\rangle_{H^{-1}(G),H_{0}^{1}(G)}dW(t),$ which contradicts Lemma 5.1. _Case 3: $h_{1}=h_{2}=0$._ Assume that Eq. 1.1 was exactly controllable. Then, from the equivalence between the exact controllability of Eq. 1.1 and the observability estimate of Eq. 1.2, we get that for any $(z^{T},\hat{z}^{T})\in L^{2}_{\mathcal{F}_{T}}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L^{2}_{\mathcal{F}_{T}}(\Omega;H_{0}^{1}(G))$, the solution $(z,Z,\hat{z},\hat{Z})$ to Eq. 1.2 ( with $\tau=T$ and $(z(T),\hat{z}(T))=(z^{T},\hat{z}^{T})$) satisfies $\displaystyle|(z^{T},\hat{z}^{T})|_{L^{2}_{\mathcal{F}_{T}}(\Omega;H^{3}(G)\cap H_{0}^{2}(G))\times L^{2}_{\mathcal{F}_{T}}(\Omega;H_{0}^{1}(G))}\leq C(|Z|_{L^{2}_{\mathbb{F}}(0,T;H^{3}(G)\cap H^{2}_{0}(G))}+|\hat{Z}|_{L^{2}_{\mathbb{F}}(0,T;H_{0}^{1}(G))}).$ (5.2) For any nonzero $(\Phi_{0},\Phi_{1})\in(H^{3}(G)\cap H_{0}^{2}(G))\times H_{0}^{1}(G)$, let $(\Phi,\hat{\Phi})$ solve the equation $\displaystyle\left\\{\begin{aligned} &d\Phi=\hat{\Phi}dt-a_{5}\Phi dW(t)&&\quad\text{ in }Q,\\\ &d\hat{\Phi}+\Delta^{2}\Phi dt=[(a_{1}-\operatorname{div}a_{2}-a_{4}a_{5})\Phi- a_{2}\cdot\nabla\Phi]dt&&\quad\text{ in }Q,\\\ &\Phi=\frac{\partial\Phi}{\partial\nu}=0&&\quad\text{ on }\Sigma,\\\ &(\Phi(0),\hat{\Phi}(0))=(\Phi_{0},\Phi_{1})&&\quad\text{ in }G.\end{aligned}\right.$ Clearly, $(\Phi,0,\hat{\Phi},0)$ solves Eq. 1.2 with the final datum $(z^{T},\hat{z}^{T})=(\Phi(T),\hat{\Phi}(T))$, a contradiction to the inequality Eq. 5.2. ∎ ## Appendix A Proof of the weighted identity ###### Proof of Theorem 2.1. It is clear that $\displaystyle dw=d(\theta v)=\theta dv+\ell_{t}\theta vdt=\hat{w}dt+\theta f_{1}dt+\theta g_{1}dW(t),$ and $\displaystyle\theta d\hat{v}$ $\displaystyle=\theta d[\theta^{-1}(\hat{w}-\ell_{t}w)]=d\hat{w}-\ell_{t}dw-\ell_{tt}wdt-\ell_{t}\hat{w}dt+\ell_{t}^{2}dt$ $\displaystyle=d\hat{w}-2\ell_{t}\hat{w}dt+(\ell_{t}^{2}-\ell_{tt})wdt-\ell_{t}\theta f_{1}dt-\ell_{t}\theta g_{t}dW(t).$ (A.1) We also have $\displaystyle\theta\Delta^{2}v$ $\displaystyle=\Delta^{2}w-4s\lambda\xi\nabla\eta\cdot\nabla\Delta w-4s\lambda^{2}\xi(\nabla^{2}w\nabla\eta\nabla\eta)-4s\lambda\xi\sum\limits_{i,j=1}^{n}\eta_{x_{i}x_{j}}w_{x_{i}x_{j}}$ $\displaystyle\quad+2s^{2}\lambda^{2}\xi^{2}|\nabla\eta|^{2}\Delta w-2s\lambda^{2}\xi|\nabla\eta|^{2}\Delta w-2s\lambda\xi\Delta\eta\Delta w+4s^{2}\lambda^{2}\xi^{2}(\nabla^{2}w\nabla\eta\nabla\eta)$ $\displaystyle\quad-4\nabla\Delta\ell\cdot\nabla w+12s^{2}\lambda^{3}\xi^{2}|\nabla\eta|^{2}\nabla\eta\nabla w+8s^{2}\lambda^{2}\xi^{2}(\nabla^{2}\eta\nabla\eta\nabla w)$ (A.2) $\displaystyle\quad-4s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\nabla\eta\nabla w+4s^{2}\lambda^{2}\xi^{2}\Delta\eta\nabla\eta\nabla w+4(\nabla\ell\cdot\nabla\Delta\ell)w+2|\nabla^{2}\ell|^{2}w-\Delta^{2}\ell w$ $\displaystyle\quad-6s^{3}\lambda^{4}\xi^{3}|\nabla\eta|^{4}w-4s^{3}\lambda^{3}\xi^{3}(\nabla^{2}\eta\nabla\eta\nabla\eta)w+s^{4}\lambda^{4}\xi^{4}|\nabla\eta|^{4}w-2s^{3}\lambda^{3}\xi^{3}|\nabla\eta|^{2}\Delta\eta w+|\Delta\ell|^{2}w.$ From Eqs. 2.13, 2.14, 2.15, 2.16, A, 2.17 and A.2, we have $2\theta I_{2}(d\hat{v}+\Delta^{2}vdt)=2I_{2}(I_{1}+I_{2}dt+I_{3}).$ We will compute $I_{1}I_{2}$ under the form $\sum\limits_{i=1}^{7}\sum\limits_{j=1}^{5}I_{ij}$, where $I_{ij}$ is the product of the $i$-th term of $I_{1}$ with the $j$-th term of $I_{2}$. Note that $I_{ij}$ are the same as Appendix A in [22] for $i=1,\cdots,6$ and $j=1,\cdots,5$, except for $I_{13}$. We have $\displaystyle I_{13}$ $\displaystyle=\sum_{i,j,k,l=1}^{n}\Phi_{3}^{kl}w_{x_{i}x_{i}x_{j}x_{j}}w_{x_{k}x_{l}}dt$ $\displaystyle=\sum_{i,j,k,l=1}^{n}\biggl{(}\Phi_{3}^{kl}w_{x_{i}x_{i}x_{j}}w_{x_{k}x_{l}}-\Phi_{3}^{kj}w_{x_{i}x_{i}x_{l}}w_{x_{k}x_{l}}+\Phi_{3}^{kl}w_{x_{i}x_{i}x_{l}}w_{x_{k}x_{j}}-\Phi_{3_{x_{i}}}^{kl}w_{x_{i}x_{j}}w_{x_{k}x_{l}}$ $\displaystyle\qquad\qquad~{}~{}+\Phi_{3_{x_{l}}}^{kj}w_{x_{i}x_{l}}w_{x_{i}x_{k}}+\Phi_{3_{x_{l}}}^{kl}w_{x_{i}x_{j}}w_{x_{i}x_{k}}-\frac{1}{2}\Phi_{3_{x_{l}}}^{jl}w_{x_{i}x_{k}}^{2}-\Phi_{3_{x_{i}}}^{kl}w_{x_{j}x_{l}}w_{x_{i}x_{k}}\biggr{)}_{x_{j}}dt$ $\displaystyle\quad-\sum_{i,j,k,l=1}^{n}\Phi_{3}^{kl}w_{x_{i}x_{i}x_{l}}w_{x_{k}x_{j}x_{j}}dt+\sum_{i,j,k,l=1}^{n}\Phi_{3x_{i}x_{j}}^{kl}w_{x_{i}x_{j}}w_{x_{k}x_{l}}dt-\sum_{i,j,k,l=1}^{n}\Phi_{3x_{l}x_{j}}^{kl}w_{x_{i}x_{j}}w_{x_{i}x_{k}}dt$ $\displaystyle\quad-\sum_{i,j,k,l=1}^{n}\Phi_{3x_{i}x_{l}}^{kl}w_{x_{i}x_{j}}w_{x_{k}x_{j}}dt+\sum_{i,j,k,l=1}^{n}\Phi_{3x_{i}x_{j}}^{kl}w_{x_{i}x_{l}}w_{x_{k}x_{j}}dt+\frac{1}{2}\sum_{i,j,k,l=1}^{n}\Phi_{3x_{k}x_{l}}^{kl}w_{x_{i}x_{j}}^{2}dt.$ We have $\displaystyle I_{71}$ $\displaystyle=\sum_{i,j=1}^{n}\Phi_{1}^{i}w_{x_{i}x_{j}x_{j}}d\hat{w}$ $\displaystyle=\sum_{i,j=1}^{n}\biggl{(}-\Phi_{1}^{j}\hat{w}dw_{x_{i}x_{i}}+\Phi_{1x_{i}}^{i}\hat{w}\hat{w}_{x_{j}}-\frac{1}{2}\Phi_{1x_{i}x_{j}}^{i}\hat{w}^{2}+\Phi_{1}^{i}\hat{w}_{x_{i}}\hat{w}_{x_{j}}-\frac{1}{2}\Phi_{1}^{j}\hat{w}_{x_{i}}^{2}\biggr{)}_{x_{j}}dt$ $\displaystyle\quad+\sum_{i,j=1}^{n}d(\Phi^{i}_{1}w_{x_{i}x_{j}x_{j}}\hat{w})+\frac{1}{2}\sum_{i,j=1}^{n}\Phi_{1x_{i}x_{j}x_{j}}^{i}\hat{w}^{2}dt-\frac{1}{2}\sum_{i,j=1}^{n}\Phi_{1x_{i}}^{i}\hat{w}_{x_{j}}^{2}dt-\sum_{i,j=1}^{n}\Phi_{1x_{j}}^{i}\hat{w}_{x_{i}}\hat{w}_{x_{j}}dt$ $\displaystyle\quad-\sum_{i,j=1}^{n}\Phi_{1t}^{i}w_{x_{i}x_{j}x_{j}}\hat{w}dt+\sum_{i,j=1}^{n}(\Phi_{1x_{i}}^{i}\hat{w}+\Phi_{1}^{i}\hat{w}_{x_{i}})[(\theta f_{1})_{x_{j}x_{j}}dt+(\theta g_{1})_{x_{j}x_{j}}dW(t)]$ $\displaystyle\quad-\sum_{i,j=1}^{n}\Phi_{1}^{i}dw_{x_{i}x_{j}x_{j}}d\hat{w},$ $\displaystyle I_{72}$ $\displaystyle=\sum_{i=1}^{n}\Phi_{2}w_{x_{i}x_{i}}d\hat{w}$ $\displaystyle=\sum_{j=1}^{n}\biggl{(}-\Phi_{2}\hat{w}\hat{w}_{x_{j}}+\frac{1}{2}\Phi_{2x_{j}}\hat{w}^{2}\biggr{)}_{x_{j}}dt+\sum_{i=1}^{n}d(\Phi_{2}w_{x_{i}x_{i}}\hat{w})-\sum_{i=1}^{n}\Phi_{2t}w_{x_{i}x_{i}}\hat{w}dt$ $\displaystyle\quad-\frac{1}{2}\sum_{i=1}^{n}\Phi_{2x_{i}x_{i}}\hat{w}^{2}dt+\sum_{i=1}^{n}\Phi_{2}\hat{w}_{x_{i}}^{2}dt-\sum_{i=1}^{n}\Phi_{2}dw_{x_{i}x_{i}}d\hat{w}$ $\displaystyle\quad-\sum_{i=1}^{n}\Phi_{2}\hat{w}[(\theta f_{1})_{x_{i}x_{i}}dt+(\theta g_{1})_{x_{i}x_{i}}dW(t)],$ $\displaystyle I_{73}$ $\displaystyle=\sum_{i,j=1}^{n}\Phi_{3}^{ij}w_{x_{i}x_{j}}d\hat{w}$ $\displaystyle=\sum_{i,j=1}^{n}\biggl{(}-\Phi_{3}^{ij}\hat{w}\hat{w}_{x_{i}}+\frac{1}{2}\Phi_{3x_{i}}^{ij}\hat{w}^{2}\biggr{)}_{x_{j}}dt+\sum_{i,j=1}^{n}d(\Phi_{3}^{ij}w_{x_{i}x_{j}}\hat{w})-\sum_{i,j=1}^{n}\Phi_{3t}^{ij}w_{x_{i}x_{j}}\hat{w}dt$ $\displaystyle\quad-\frac{1}{2}\sum_{i,j=1}^{n}\Phi_{3x_{i}x_{j}}^{ij}\hat{w}^{2}dt-\sum_{i,j=1}^{n}\Phi_{3}^{ij}dw_{x_{i}x_{j}}d\hat{w}+\sum_{i,j=1}^{n}\Phi_{3}^{ij}\hat{w}_{x_{i}}\hat{w}_{x_{j}}dt$ $\displaystyle\quad-\sum_{i,j=1}^{n}\Phi_{3}^{ij}\hat{w}[(\theta f_{1})_{x_{i}x_{j}}dt+(\theta g_{1})_{x_{i}x_{j}}dW(t)],$ $\displaystyle I_{74}$ $\displaystyle=\sum_{i=1}^{n}\Phi_{4}^{i}w_{x_{i}}d\hat{w}$ $\displaystyle=\sum_{j=1}^{n}\biggl{(}-\frac{1}{2}\Phi_{4}^{j}\hat{w}^{2}\biggr{)}_{x_{j}}dt+\sum_{i=1}^{n}d(\Phi_{4}^{i}w_{x_{i}}\hat{w})-\sum_{i=1}^{n}\Phi_{4t}^{i}w_{x_{i}}\hat{w}dt+\frac{1}{2}\sum_{i=1}^{n}\Phi_{4x_{i}}^{i}\hat{w}^{2}dt$ $\displaystyle\quad-\sum_{i=1}^{n}\Phi^{i}_{4}dw_{x_{i}}d\hat{w}-\sum_{i=1}^{n}\Phi_{4}^{i}\hat{w}[(\theta f_{1})_{x_{i}}dt+(\theta g_{1})_{x_{i}}dW(t)],$ $\displaystyle I_{75}$ $\displaystyle=\Phi_{5}wd\hat{w}=d(\Phi_{5}w\hat{w})-\Phi_{5t}w\hat{w}dt-\Phi_{5}\hat{w}^{2}dt-\Phi_{5}dwd\hat{w}-\Phi_{5}\hat{w}(\theta f_{1}dt+\theta g_{1}dW(t)).$ By summing all the $I_{ij}$, we get Eq. 2.12. ∎ ## References * [1] J. M. Ball, J. E. Marsden, and M. Slemrod, Controllability for distributed bilinear systems, SIAM J. Control Optim., 20 (1982), pp. 575–597. * [2] Z. a. Brzeźniak, B. Maslowski, and J. Seidler, Stochastic nonlinear beam equations, Probab. Theory Related Fields, 132 (2005), pp. 119–149. * [3] P. L. Chow and J. L. Menaldi, Stochastic PDE for nonlinear vibration of elastic panels, Differential Integral Equations, 12 (1999), pp. 419–434. * [4] G. Da Prato and J. Zabczyk, Stochastic equations in infinite dimensions, vol. 152 of Encyclopedia of Mathematics and its Applications, Cambridge University Press, Cambridge, second ed., 2014. * [5] M. Eller and D. Toundykov, Semiglobal exact controllability of nonlinear plates, SIAM J. Control Optim., 53 (2015), pp. 2480–2513. * [6] X. Fu and X. Liu, Controllability and observability of some stochastic complex Ginzburg-Landau equations, SIAM J. Control Optim., 55 (2017), pp. 1102–1127. * [7] X. Fu, Q. Lü, and X. Zhang, Carleman estimates for second order partial differential operators and applications, SpringerBriefs in Mathematics, Springer, Cham, 2019. * [8] P. Gao, M. Chen, and Y. Li, Observability estimates and null controllability for forward and backward linear stochastic Kuramoto-Sivashinsky equations, SIAM J. Control Optim., 53 (2015), pp. 475–500. * [9] S. W. Hansen and O. Imanuvilov, Exact controllability of a multilayer Rao-Nakra plate with clamped boundary conditions, ESAIM Control Optim. Calc. Var., 17 (2011), pp. 1101–1132. * [10] A. Haraux, Séries lacunaires et contrôle semi-interne des vibrations d’une plaque rectangulaire, J. Math. Pures Appl. (9), 68 (1989), pp. 457–465 (1990). * [11] S. Jaffard, Contrôle interne exact des vibrations d’une plaque rectangulaire, Portugal. Math., 47 (1990), pp. 423–429. * [12] J. U. Kim, On a stochastic plate equation, Appl. Math. Optim., 44 (2001), pp. 33–48. * [13] V. Komornik, Exact controllability and stabilization, RAM: Research in Applied Mathematics, Masson, Paris; John Wiley & Sons, Ltd., Chichester, 1994\. * [14] I. Lasiecka, J.-L. Lions, and R. Triggiani, Nonhomogeneous boundary value problems for second order hyperbolic operators, J. Math. Pures Appl., 65 (1986), pp. 149–192. * [15] I. Lasiecka and R. Triggiani, Exact controllability of the Euler-Bernoulli equation with controls in the Dirichlet and Neumann boundary conditions: a nonconservative case, SIAM J. Control Optim., 27 (1989), pp. 330–373. * [16] , Exact controllability of the Euler-Bernoulli equation with boundary controls for displacement and moment, J. Math. Anal. Appl., 146 (1990), pp. 1–33. * [17] , Sharp trace estimates of solutions of Kirchhoff and Euler-Bernoulli equations, Appl. Math. Optim., 28 (1993), pp. 277–306. * [18] J.-L. Lions, Exact controllability, stabilization and perturbations for distributed systems, SIAM Rev., 30 (1988), pp. 1–68. * [19] W. Liu, Local boundary controllability for the semilinear plate equation, Comm. Partial Differential Equations, 23 (1998), pp. 201–221. * [20] Q. Lü, Exact controllability for stochastic Schrödinger equations, J. Differential Equations, 255 (2013), pp. 2484–2504. * [21] , Exact controllability for stochastic transport equations, SIAM J. Control Optim., 52 (2014), pp. 397–419. * [22] Q. Lü and Y. Wang, Null controllability for fourth order stochastic parabolic equations, SIAM J. Control Optim., 60 (2022), pp. 1563–1590. * [23] Q. Lü and X. Zhang, Exact controllability for a refined stochastic wave equation. ArXiv, Jan. 2019. * [24] , Mathematical theory for stochastic distributed parameter control systems, vol. 101 of Probability Theory and Stochastic Modelling, Springer Nature Switzerland AG., 2021\. * [25] S.-G. Peng, Backward stochastic differential equation and exact controllability of stochastic control systems, Progr. Natur. Sci. (English Ed.), 4 (1994), pp. 274–284. * [26] J.-P. Puel and E. Zuazua, Controllability of a multidimensional system of Schrödinger equations: application to a system of plate and beam equations, in Analysis and optimization of systems: state and frequency domain approaches for infinite-dimensional systems (Sophia-Antipolis, 1992), vol. 185 of Lect. Notes Control Inf. Sci., Springer, Berlin, 1993, pp. 500–511. * [27] S. Tang and X. Zhang, Null controllability for forward and backward stochastic parabolic equations, SIAM J. Control Optim., 48 (2009), pp. 2191–2216. * [28] Y. Yu and J.-F. Zhang, Carleman estimates of refined stochastic beam equations and applications, SIAM J. Control Optim., 60 (2022), pp. 2947–2970. * [29] X. Zhang, Exact controllability of semilinear plate equations, Asymptot. Anal., 27 (2001), pp. 95–125.
[table]capposition=top # Towards Training GNNs using Explanation Directed Message Passing Valentina Giunchiglia Imperial College London <EMAIL_ADDRESS>Varun Shukla11footnotemark: 1 Ludwig-Maximilians Universität München <EMAIL_ADDRESS>Gonzalez Imperial College London <EMAIL_ADDRESS>Agarwal Adobe <EMAIL_ADDRESS>Equal contribution. ###### Abstract With the increasing use of Graph Neural Networks (GNNs) in critical real-world applications, several post hoc explanation methods have been proposed to understand their predictions. However, there has been no work in generating explanations on the fly during model training and utilizing them to improve the expressive power of the underlying GNN models. In this work, we introduce a novel explanation-directed neural message passing framework for GNNs, Expass (EXplainable message PASSing), which aggregates only embeddings from nodes and edges identified as important by a GNN explanation method. Expass can be used with any existing GNN architecture and subgraph-optimizing explainer to learn accurate graph embeddings. We theoretically show that Expass alleviates the oversmoothing problem in GNNs by slowing the layer-wise loss of Dirichlet energy and that the embedding difference between the vanilla message passing and Expass framework can be upper bounded by the difference of their respective model weights. Our empirical results show that graph embeddings learned using Expass improve the predictive performance and alleviate the oversmoothing problems of GNNs, opening up new frontiers in graph machine learning to develop explanation-based training frameworks. ## 1 Introduction Graph Neural Networks (GNNs) are increasingly used as powerful tools for representing graph-structured data, such as social, information, chemical, and biological networks [1, 2]. With the deployment of GNN models in critical applications (e.g., financial systems and crime forecasting [3, 4]), it becomes essential to ensure that the relevant stakeholders understand and trust their decisions. To this end, several approaches [5, 6, 7, 8, 9, 10, 11, 12, 13] have been proposed in recent literature to generate post hoc explanations for predictions of GNN models. In contrast to other modalities like images and texts, generating instance- level explanations for graphs is non-trivial. In particular, it is more challenging since individual node embeddings in GNNs aggregate information using the entire graph structure, and, therefore, explanations can be on different levels (i.e., node attributes, nodes, and edges). While several categories of GNN explanation methods have been proposed: gradient-based [14, 10, 5], perturbation-based [13, 9, 11, 8, 15], and surrogate-based [7, 12], their utility is limited to generating post hoc node- and edge-level explanations for a given pre-trained GNN model. Thus, the capability of GNN explainers to improve the predictive performance of a GNN model lacks understanding as there is very little work on systematically analyzing the reliability of state-of-the-art GNN explanation methods on model performance [16]. To address this, recent works have explored the joint optimization of machine learning models and explanation methods to improve the reliability of explanations [17, 18]. Zhou et al. [18] proposed DropEdge as a technique to drop random edges (similar to generating random edge explanations) during training to reduce overfitting in GNNs. More recently, Spinelli et al. [17] used meta-learning frameworks to generate GNN explanations and show an improvement in the performance of specific GNN explanation methods. While these works make an initial attempt at jointly optimizing explainers and predictive models, they are neither generalizable nor exhaustive. They fail to show improvement in the downstream GNN performance [17] and degree of explainability [18] across diverse GNN architectures and explainers. Further, there is little to no work done on either theoretically analyzing the effect of GNN explanations on the neural message framework in GNNs or on important GNN properties like oversmoothing [19]. Present work. In this work, we introduce a novel explanation-directed neural message passing framework, Expass, which can be used with any GNN model and subgraph-optimizing explainer to learn accurate graph representations. In particular, Expass utilizes GNN explanations to steer the underlying GNN model to learn graph embeddings using only important nodes and edges. Expass aims to define local neighborhoods for neural message passing, i.e., identify the most important edges and nodes, using explanation weights, in the $k$-hop local neighborhood of every node in the graph. Formally, we augment existing message passing architectures to allow information flow along important edges while blocking information along irrelevant edges. We present an extensive theoretical and empirical analysis to show the effectiveness of Expass on the predictive, explainability, and oversmoothing performance of GNNs. Our theoretical results show that the embedding difference between vanilla message passing and Expass frameworks is upper- bounded by the difference between their model weights. Further, we show that embeddings learned using Expass relieve the oversmoothing problem in GNNs as they reduce information propagation by slowing the layer-wise loss of Dirichlet energy (Section 4.2). For our empirical analysis, we integrate Expass into state-of-the-art GNN models and evaluate their predictive, oversmoothing, and explainability performance on real-world graph datasets (Section 5). Our results show that, on average, across five GNN models, Expass improves the degree of explainability of the underlying GNNs by 39.68%. Our ablation studies show that for an increasing number of GNN layers, Expass achieves 34.4% better oversmoothing performance than its vanilla counterpart. Finally, our results demonstrate the effectiveness of using explanations during training, paving the way for new frontiers in GraphXAI research to develop explanation-based training algorithms. ## 2 Related works Graph Neural Networks. Graph Neural Networks (GNNs) are complex non-linear functions that transform input graph structures into a lower dimensional embedding space. The main goal of GNNs is to learn embeddings that reflect the underlying input graph structure, i.e., neighboring nodes in the graph are mapped to neighboring points in the embedding space. Prior works have proposed several GNN models using spectral and non-spectral approaches. Spectral models [20, 21, 22, 23, 24] leverage Fourier transform and graph Laplacian to define convolution approaches for GNN models. However, non-spectral approaches [25, 26, 27, 28, 29] define the convolution operation by leveraging the local neighborhood of individual nodes in the graph. Most modern non-spectral models are message-passing frameworks [30, 31], where nodes update their embedding by aggregating information from $k$-hop neighboring nodes. Post hoc Explanations. With the increasing development of complex high- performing GNN models [25, 26, 27, 28, 29], it becomes critical to understand their decisions. Prior works have focused on developing several post hoc explanation methods to explain the decisions of GNN models [5, 13, 7, 12, 11, 9, 32]. More specifically, these explanation methods can be broadly categorized into i) gradient-based methods [5] that leverage the gradients of the GNN model to generate explanations; ii) perturbation-based methods [13, 9, 11] that aim to generate explanations by calculating the change in GNN predictions upon perturbations of the input graph structure (nodes, edges, or subgraphs); and iii) surrogate-based methods [7, 12] that fit a simple interpretable model to approximate the predictive behavior of the given GNN model. Finally, recent works have introduced frameworks to theoretically and empirically analyze the behavior of state-of-the-art GNN explanation methods with respect to several desirable properties [33, 16]. ## 3 Preliminaries Notations. Let $\mathcal{G}=(\mathcal{V},\mathcal{E},\mathbf{X})$ denote an undirected graph comprising of a set of nodes $\mathcal{V}$ and a set of edges $\mathcal{E}$. Let $\mathbf{X}{=}\\{\mathbf{x}_{1},\mathbf{x}_{2},\dots,\mathbf{x}_{N}\\}$ denote the set of node feature vectors for all nodes in $\mathcal{V}$, where $\mathbf{x}_{v}\in\mathbb{R}^{d}$ captures the attribute values of a node $v$ and $N{=}|\mathcal{V}|$ denotes the number of nodes in the graph. Let $\mathbf{A}\in\mathbb{R}^{N\times N}$ be the graph adjacency matrix, where element $\mathbf{A}_{uv}=1$ if there exists an edge $e\in\mathcal{E}$ between nodes $u$ and $v$ and $\mathbf{A}_{uv}=0$ otherwise. We use ${\mathcal{N}}_{u}$ to denote the set of immediate neighbors of node $u$, i.e., , ${\mathcal{N}}_{u}=\\{v\in\mathcal{V}|A_{uv}=1\\}$. Finally, the function $\text{deg}:\mathcal{V}\mapsto\mathbb{Z}_{>0}$ is defined as $\text{deg}(v)=|\mathcal{N}_{v}|$ and outputs the degree of a node $v\in\mathcal{V}$ Graph Neural Networks (GNNs). Formally, GNNs can be formulated as message passing networks [30] specified by three key operators Msg, Agg, and Upd. These operators are recursively applied on a given graph $\mathcal{G}$ for a $L$-layer GNN model defining how neural messages are shared, aggregated, and updated between nodes to learn the final node representations in the $L^{\text{th}}$ layer of the GNN. Commonly, a message between a pair of nodes $(u,v)$ in layer $l$ is characterized as a function of their hidden representations $\mathbf{h}_{u}^{(l-1)}$ and $\mathbf{h}_{v}^{(l-1)}$ from the previous layer: $\mathbf{m}_{uv}^{(l)}=\textsc{Msg}(\mathbf{h}_{u}^{(l-1)},\mathbf{h}_{v}^{(l-1)}).$ The Agg operator retrieves the messages from the neighborhood of node $u$ and aggregates them as: $\mathbf{m}_{u}^{(l)}=\textsc{Agg}(\mathbf{m}_{uv}^{(l)}\,|\,v\in\mathcal{N}_{u})$. Next, the Upd operator takes the aggregated message $\mathbf{m}_{u}^{(l)}$ at layer $l$ and combines it with $\mathbf{h}_{u}^{(l-1)}$ to produce node $u$’s representation for layer $l$ as $\mathbf{h}_{u}^{(l)}=\textsc{Upd}(\mathbf{m}_{u}^{(l)},\mathbf{h}_{u}^{(l-1)})$. Lastly, the final node representation for node $u$ is given as $\mathbf{z}_{u}=\mathbf{h}_{u}^{(L)}$. Graph Explanations. In contrast to other modalities like images and texts, an explanation method for graphs can formally generate multi-level explanations. For instance, in a graph classification task, the explanations for a given graph prediction can be with respect to the node attributes $\mathbf{M}_{\text{x}}\in\mathbb{R}^{d}$, nodes $\mathbf{M}_{n}\in\mathbb{R}^{N}$, or edges $\mathbf{M}_{e}\in\mathbb{R}^{N\times N}$. Note that these explanation masks are continuous but can be discretized using specific thresholding strategies [33]. Oversmoothing. Cai et al. [34] and Zhou et al. [35] defined bounds for analyzing oversmoothing for a GNN using Dirichlet Energy. For a graph $\mathcal{G}$ with adjacency matrix $\mathbf{A}$ and degree matrix $\mathbf{D}$, we define $\tilde{\mathbf{A}}{=}\mathbf{A}+\mathbf{I}_{N}$ and $\tilde{\mathbf{D}}{=}\mathbf{D}+\mathbf{I}_{N}$ as the adjacency and degree matrices respectively of the graph $\mathcal{G}$ with self-loops. We also define the augmented normalized Laplacian of $\mathcal{G}$ as $\tilde{\Delta}{=}\mathbf{I}_{N}-\tilde{\mathbf{D}}^{-\frac{1}{2}}\tilde{A}\tilde{D}^{-\frac{1}{2}}$, and $\mathbf{P}{=}\mathbf{I}_{N}-\tilde{\Delta}$. ## 4 Our Framework: Expass Here, we describe Expass, our proposed explainable message-passing framework that aims to learn accurate and interpretable graph embeddings. In particular, Expass incorporates explanations into the message-passing framework of GNN models by only aggregating embeddings from key nodes and edges as identified using an explanation method. Problem formulation (Explanation Directed Message Passing). Given a graph $\mathcal{G}=(\mathcal{V},\mathcal{E},\textbf{X})$, Expass aims to generate a $d$-dimensional embedding $\mathbf{z}_{u}\in\mathbb{R}^{d}$ for each node $u\in\mathcal{V}$ using an explanation-directed message passing framework that filters out the noise from unimportant edges and improves the expressive power of GNNs. ### 4.1 Explanation Directed Message Passing Figure 1: Overview of Expass: a) Expass investigates the problem of injecting explanations into the message-passing framework to increase the expressive power and performance of GNNs. b) Shown is the general message passing scheme where, for node $u$, messages are aggregated from nodes $v_{i}\in\mathcal{N}_{u}$ in the 1-hop neighborhood of $u$. c) Expass injects explanations into the message passing framework translates by masking out messages from neighboring nodes $v_{i}\in\mathcal{N}_{u}$ with explanation scores $s_{uv_{i}}\approx 0$ when $u$ is correctly classified. The central idea of Expass is to propose a novel method for improving the neural message passing scheme of GNN models by utilizing explanations during model training and aggregating important neural messages along edges in graph neighborhoods. Next, we describe the existing message-passing scheme in GNNs and our explainable counterpart. Message Passing. As described in Section 3, each GNN layer can be described using the $\textsc{Msg},\textsc{Agg},$ and Upd operators. For each node $u\in\mathcal{V}$, the $(l+1)^{th}$ layer embeddings $\mathbf{h}^{(l+1)}_{u}$ is computed using a GNN operating on the node’s neighboring attributes. Formally, the GNN layer can be formulated as: where $\mathbf{h}^{(l+1)}_{u}$ represents the updated embedding of node $u$, $\psi$ is the Msg operator, $\bigoplus$ is the Agg operator (e.g., summation), $\phi$ is an Upd function (e.g., any non-linear activation function), and $\mathbf{h}^{(l)}_{u}$ represents the embedding of node $u$ from the previous layer. We obtain an embedding $\mathbf{z}_{u}$ for node $u$ by stacking $L$ GNN layers. Finally, the node embeddings $\mathbf{Z}\in\mathbb{R}$ are then passed to a Readout function to obtain an embedding for the graph. Expass. Here, we describe our proposed explainable message-passing scheme that incorporates explanations into the message-passing step in individual GNN layers on the fly during the training process. Given an explanation method, which generates an importance score $s_{uv}\in\mathbf{M}^{e}_{u}$ for every edge $e_{uv}\in\mathcal{E}$, we can weight the edge contribution in the neighborhood $\mathcal{N}_{u}$ of node $u$ as: Note that Expass is agnostic to explanation types and can also incorporate explanations on node attributes and node level. For instance, the importance scores for individual nodes can be computed by averaging the outgoing scores $s_{uv}$ for all $v\in\mathcal{N}_{u}$. Subsequently, we can replace the $s_{uv}$ score by using the average score $s_{u}$ to weight edges in the Expass layers, and for node attributes, we can multiply the node attribute explanation $\mathbf{M}^{a}_{u}$ to the original node attribute vector. To enable explainable message passing and only retain the important embeddings for node $u$, Expass removes the knowledge of irrelevant nodes and edges from the local neighborhood $\mathcal{N}_{u}$ of node $u$ using its explanations. For instance, if node $v$ is considered important to node $u$, Expass transforms the aggregated messages of node $u$ using the node importance scores $s_{uv}$. Note that since the explanations of node $u$ include important nodes and edges in the $L$-hop neighborhood of node $u$, even though node $u$ is only locally modified, the change will spread through all the nodes in every GNN layer. Furthermore, to avoid spurious correlations, we ensure that explanations are only generated for correctly classified nodes and graphs. Explanation weights infuse information from higher-order neighborhoods into each layer of the GNN model, specifically, from as many $L$-hop neighbors because explanation weights within each layer are computed using the $L$-layer GNN model. To illustrate this, we next show the weight computations for a GNN explanation method. Without loss of generality, let us consider GNNExplainer as our explanation method whose mask for the selected graph is formulated as: $\mathcal{G}_{\text{mask}}=(\mathbf{X}^{\prime},\mathbf{A}^{\prime})=(\mathbf{X}\odot\sigma(\mathbf{M}^{\text{x}}),\mathbf{A}\odot\sigma(\mathbf{M}^{\text{e}}))$, where $W=[\mathbf{M}^{\text{x}},\mathbf{M}^{\text{e}}]$ are the explainers parameters, $\sigma$ is the sigmoid function, and $\odot$ denotes element-wise multiplication. Here, $s_{uv}$ represents the element in row $v$ and column $u$ of $\mathbf{M}^{\text{e}}$. Gradient descent-based optimization is used to find the optimal values for the masks minimizing the following objective: $L_{e}=-\sum_{c=1}^{C}1[y=c]\log f_{\theta}(Y=y|\mathcal{G}_{\text{mask}})$, where $f_{\theta}$ is the $L$-layer GNN model and $C$ is the total number of classes. This shows that a $L$-hop neighborhood is used to compute $s_{uv}$. Formally, it minimizes the uncertainty of the predictive model when the GNN computation is limited to the explanation subgraph. This uncertainty is minimized as a proxy of the maximization of the mutual information between the prediction with the unmasked graph and masked graph. ### 4.2 Theoretical Analysis Here, we provide a detailed theoretical analysis of our proposed Expass framework. In particular, we (i) provide a theoretical upper bound on the embedding difference obtained from a vanilla message passing and Expass framework and (ii) show that graph embeddings learned using Expass relieves the oversmoothing problem in GNNs by reducing information propagation. ###### Theorem 1 (Differences between Expass and Vanilla Message Passing). Given a non-linear activation function $\sigma$ that is Lipschitz continuous, the difference between the node embeddings between a vanilla message passing and Expass framework can be bounded by the difference in their individual weights, i.e., $\|\mathbf{h}_{u}^{(l)}-\mathbf{h^{\prime}}_{u}^{(l)}\|_{2}\leq\|\mathbf{W}_{a}^{(l)}{-}\mathbf{W^{\prime}}_{a}^{(l)}\|_{2}\|\mathbf{h}_{u}^{(l-1)}\|_{2}+\|\mathbf{W}_{n}^{(l)}{-}\mathbf{W^{\prime}}_{n}^{(l)}\|_{2}\sum_{\mathclap{v\in\mathcal{N}_{u}\cap s_{v}=1}}\|\mathbf{h}_{v}^{(l-1)}\|_{2},$ (1) where $\mathbf{W}_{a}^{(l)}$ and $\mathbf{W^{\prime}}_{a}^{(l)}$ are the weights for node $u$ in layer $l$ of the vanilla message passing and Expass framework and $\mathbf{W}_{n}^{(l)}$ and $\mathbf{W^{\prime}}_{n}^{(l)}$ are their respective weight matrix with the neighbors of node $u$ at layer $l$. ###### Proof Sketch. In Theorem 1, we prove that the $\ell_{2}$-norm of the differences between the embeddings of vanilla message passing and Expass framework at layer $l$ is upper bounded by the difference between their weights and the embeddings of node $u$ and its subgraph. See Appendix A for more details. ∎ ###### Definition 1 (Dirichlet Energy for a Node Embedding Matrix [35]). Given a node embedding matrix $\mathbf{H}^{(l)}=[\mathbf{h}_{1}^{(l)},\dots,\mathbf{h}_{n}^{(l)}]^{T}$ learned from the GNN model at the $l^{th}$ layer, the Dirichlet Energy $E(\mathbf{H}^{(l)})$ is defined as: $E(\mathbf{H}^{(l)})=tr(\mathbf{H}^{(l)^{T}}\tilde{\Delta}\mathbf{H}^{(l)})=\frac{1}{2}\sum_{i,j\in\mathcal{V}}a_{ij}||\frac{\mathbf{H}_{i}^{(l)}}{\sqrt{1+\text{deg}_{i}}}-\frac{\mathbf{H}_{j}^{(l)}}{\sqrt{1+\text{deg}_{j}}}||_{2}^{2}$ (2) where $a_{ij}$ are elements in the adjacency matrix $\tilde{\mathbf{A}}$ and $\text{deg}_{i},\text{deg}_{j}$ is the degree of node $i$ and $j$, respectively. Cai et al. [34] extensively show that higher Dirichlet energies correspond to lower oversmoothing. Furthermore, they show that the removal of edges or, similarly, the reduction of edge weights on graphs helps alleviate oversmoothing. ###### Proposition 1 (Expass relieves Oversmoothing). Expass alleviates oversmoothing by slowing the layer-wise loss of Dirichlet energy. The complete proof is provided in Appendix A. ## 5 Experiments Next, we present experimental results for our Expass framework. More specifically, we address the following questions: Q1) Does Expass enable GNNs to learn more accurate embeddings and improve their degree of explainability? Q2) How does Expass affect the oversmoothing and predictive performance of GNNs with an increasing number of layers? Q3) Does Expass depend on the quality of explanations for improving the predictive and oversmoothing performance of GNNs and are they better than attention weights? Q4) How does Expass helps in the evolution of explanation during the training of the GNN model? 111Code to reproduce the results is available here ### 5.1 Datasets and Experimental setup We first describe the datasets used to study the utility of our proposed Expass framework and then outline the experimental setup. Datasets. We use real-world molecular chemistry datasets to evaluate the effectiveness of Expass w.r.t. the performance of the underlying GNN model and understand the trade-off between explainability and accuracy for a graph classification task. We consider four benchmark datasets, which includes Mutag [36], Alkane-Carbonyl [37], DD [38], and Proteins [39]. See Appendix B.1 for a detailed overview of the datasets. GNN Architectures and Explainers. To investigate the flexibility of Expass, we incorporate it into five different GNN models: GCN [40], GraphConv [41], LEConv [42], GraphSAGE [28], GAT [43], and GIN [27]. We use GNNExplainer [13] as our baseline GNN explanation method to generate edge-level explanations for most of our experiments. In addition, we use Integrated Gradients [44] and PGMExplainer [12], a node-level explanation method, to demonstrate Expass’s sensitivity to the choice of explainers. Implementation details. We consider DropEdge [45] as our baseline method for comparing the oversmoothing performance of Expass as DropEdge randomly removes edges from the input graph at each training epoch, acting like a message passing reducer. Across all experiments, we use topK (k=40%) node features/edges, and use them to generate explanations for all explanation methods. All other hyperparameters of the explanation and baseline methods were set following the author’s guidelines. For all our experiments (unless mentioned otherwise), we use the baseline architectures with three GNN layers followed by ReLU layers and set the hidden dimensionality to 32. Finally, we use a single linear layer to transform the graph embeddings to their respective classes. See Appendix B.2 for more details. Performance metrics for GNN Explainers. To measure the reliability of GNN explanation methods, we use the graph explanation faithfulness metric [16]: $\text{GEF}(\hat{y}_{u},\hat{y}_{u^{\prime}})=1-\exp^{-\text{KL}(\hat{y}_{u}||\hat{y}_{u^{\prime}})}$, where $\hat{y}_{u}$ is predicted probability vector using the whole subgraph and $\hat{y}_{u^{\prime}}$ is the predicted probability vector using the masked subgraph, where we generate the masked subgraph by only using the topK features identified by an explanation and the Kullback-Leibler (KL) divergence score (denoted by “$||$” operator) quantifies the distance between two probability distributions. Note that GEF is a measure of the unfaithfulness of the explanation. So, higher values indicate a higher degree of unfaithfulness. Performance metrics for Oversmoothing. Zhou et al. [18] introduced the Group Distance Ratio (GDR) metric to quantify oversmoothing in GNNs. It measures the ratio between the average of pairwise representation distances between graphs belonging to different (inter) and same (intra) groups. Formally, one would prefer to reduce the intra-group class representations and increase the inter- group distance to relieve the over-smoothing issue. Hence, lower GDR values denote higher oversmoothing in GNNs. Burn-in period. We defined the burn-in period as a number n of epochs during training in which no explanations are used. The burn-in period is necessary to avoid feeding spurious explanations to the model. The length of the burn-in period (i.e., the number of epochs) was treated as a hyperparameter and fine- tuned using the validation set. At the end of the burn-in period, a predefined percentage of correctly predicted graphs per batch is randomly sampled and their explanations are used in the model training. The percentage of correctly predicted graphs sampled in each batch was set to 0.4 for all our experiments. See Appendix C.2 for ablation on burn-in periods. ### 5.2 Results Q1) Expass improves the predictive performance and explainability of GNNs. To measure the predictive performance and degree of explainability of GNNs trained using Expass, we compute their average predictive performance (using AUROC and F1-score) and fidelity (using Graph Explanation Faithfulness) using different GNN models and datasets. Across four datasets and five GNN architectures, we find that Expass-augmented GNNs learn graph embeddings that are more accurate (higher AUROC and F1-score) and result in more faithful explanations (lower Graph Explanation Faithfulness score) than their vanilla counterparts. On average, Expass improves the AUROC and F1-score by 1.51% and 1.05%, respectively. In particular, we observe that Expass improves the predictive behavior of high-performing models like GIN (+2.06% in AUROC and +2.50% in F1-score) but shows little to no improvement in the case of LeConv, which utilizes a node-scoring mechanism through the similarity between a node and its neighbors’ embeddings. Finally, we find that Expass-augmented GNNs significantly improve the explainability of a GNN and achieve a 39.68% better faithfulness score as compared to vanilla GNNs (Table 1). See Appendix C.1 for results on node classification graph downstream tasks. Table 1: Results of Expass for five GNNs and four graph datasets. Shown is average performance across three independent runs. Arrows ($\uparrow$, $\downarrow$) indicate the direction of better performance. Expass improves the predictive power (AUROC and F1-score) and degree of explainability (Graph Explanation Faithfulness) of original GNNs across multiple datasets (shaded area). Values corresponding to best performance are bolded. Dataset | Method | AUROC ($\uparrow$) | F1-score ($\uparrow$) | GEF ($\downarrow$) ---|---|---|---|--- Alkane-Carbonyl | | GCN --- Expass-GCN | 0.97$\pm$0.01 --- 0.98$\pm$0.00 | 0.95$\pm$0.01 --- 0.96$\pm$0.01 | 0.33$\pm$0.02 --- 0.23$\pm$0.02 | GraphConv --- Expass-GraphConv | 0.97$\pm$0.01 --- 0.98$\pm$0.00 | 0.94$\pm$0.00 --- 0.97$\pm$0.00 | 0.38$\pm$0.05 --- 0.22$\pm$0.03 | LeConv --- Expass-LeConv | 0.98$\pm$0.01 --- 0.98$\pm$0.00 | 0.96$\pm$0.00 --- 0.96$\pm$0.01 | 0.37$\pm$0.03 --- 0.24$\pm$0.03 | GraphSAGE --- Expass-GraphSAGE | 0.98$\pm$0.00 --- 0.99$\pm$0.00 | 0.96$\pm$0.00 --- 0.97$\pm$0.01 | 0.40$\pm$0.12 --- 0.18$\pm$0.06 | GIN --- Expass-GIN | 0.96$\pm$0.01 --- 0.98$\pm$0.01 | 0.94$\pm$0.02 --- 0.96$\pm$0.02 | 0.35$\pm$0.06 --- 0.11$\pm$0.04 DD | | GCN --- Expass-GCN | 0.73$\pm$0.02 --- 0.74$\pm$0.01 | 0.70$\pm$0.02 --- 0.70$\pm$0.02 | 0.49$\pm$0.04 --- 0.30$\pm$0.09 | GraphConv --- Expass-GraphConv | 0.75$\pm$0.03 --- 0.77$\pm$0.03 | 0.73$\pm$0.03 --- 0.73$\pm$0.03 | 0.25$\pm$0.10 --- 0.19$\pm$0.04 | LeConv --- Expass-LeConv | 0.76$\pm$0.03 --- 0.77$\pm$0.03 | 0.74$\pm$0.02 --- 0.73$\pm$0.04 | 0.17$\pm$0.03 --- 0.31$\pm$0.10 | GraphSAGE --- Expass-GraphSAGE | 0.74$\pm$0.02 --- 0.76$\pm$0.03 | 0.70$\pm$0.02 --- 0.71$\pm$0.02 | 0.21$\pm$0.04 --- 0.20$\pm$0.03 | GIN --- Expass-GIN | 0.74$\pm$0.01 --- 0.76$\pm$0.01 | 0.70$\pm$0.01 --- 0.74$\pm$0.01 | 0.37$\pm$0.03 --- 0.35$\pm$0.05 Mutag | | GCN --- Expass-GCN | 0.71$\pm$0.11 --- 0.77$\pm$0.02 | 0.87$\pm$0.01 --- 0.89$\pm$0.00 | 0.09$\pm$0.03 --- 0.04$\pm$0.01 | GraphConv --- Expass-GraphConv | 0.91$\pm$0.02 --- 0.93$\pm$0.01 | 0.94$\pm$0.02 --- 0.94$\pm$0.01 | 0.66$\pm$0.03 --- 0.24$\pm$0.03 | LeConv --- Expass-LeConv | 0.92$\pm$0.03 --- 0.92$\pm$0.03 | 0.94$\pm$0.02 --- 0.96$\pm$0.01 | 0.65$\pm$0.05 --- 0.30$\pm$0.06 | GraphSAGE --- Expass-GraphSAGE | 0.76$\pm$0.02 --- 0.76$\pm$0.02 | 0.86$\pm$0.03 --- 0.87$\pm$0.03 | 0.24$\pm$0.08 --- 0.11$\pm$0.03 | GIN --- Expass-GIN | 0.92$\pm$0.02 --- 0.94$\pm$0.02 | 0.93$\pm$0.01 --- 0.95$\pm$0.01 | 0.61$\pm$0.05 --- 0.32$\pm$0.04 Proteins | | GCN --- Expass-GCN | 0.73$\pm$0.05 --- 0.74$\pm$0.03 | 0.68$\pm$0.04 --- 0.69$\pm$0.03 | 0.19$\pm$0.02 --- 0.08$\pm$0.02 | GraphConv --- Expass-GraphConv | 0.75$\pm$0.03 --- 0.75$\pm$0.03 | 0.70$\pm$0.03 --- 0.70$\pm$0.04 | 0.49$\pm$0.06 --- 0.10$\pm$0.03 | LeConv --- Expass-LeConv | 0.77$\pm$0.03 --- 0.76$\pm$0.02 | 0.72$\pm$0.04 --- 0.71$\pm$0.03 | 0.51$\pm$0.01 --- 0.15$\pm$0.07 | GraphSAGE --- Expass-GraphSAGE | 0.73$\pm$0.04 --- 0.73$\pm$0.04 | 0.69$\pm$0.04 --- 0.69$\pm$0.04 | 0.17$\pm$0.07 --- 0.06$\pm$0.01 | GIN --- Expass-GIN | 0.77$\pm$0.04 --- 0.78$\pm$0.03 | 0.73$\pm$0.05 --- 0.73$\pm$0.04 | 0.20$\pm$0.07 --- 0.19$\pm$0.01 Q2) Expass relieves Oversmoothing in GNNs. We examine the oversmoothing (using the Group Distance Ratio metric [18]) and predictive performance of GNNs trained using Expass with their vanilla counterparts. The oversmoothing problem in GNNs shows that the representations of nodes converge to similar vectors as the number of layers increases. Therefore, we analyze the oversmoothing of the GNNs for an increasing number of layers and find that, on average, across two architectures, Expass improves the group distance ratio by 34.4% (Figure 2). Further, we also analyzed the oversmoothing behavior of Expass for node classification tasks (in Appendix C.1) and our results indicate an inherent trade-off between oversmoothing and predictive performance of GNNs (Figures 5-7). Figure 2: The effects of the number of GNN layers on the oversmoothing performance of Expass (orange) and Vanilla (green) GCN (left column) and GIN (right column) models trained on Alkane-Carbonyl dataset. Across models with increasing number of layers, Expass achieves higher GDR performance without sacrificing the predictive performance of the GCN model. See Figs. 5-7 for predictive performance results. Q3) Ablation studies. We conduct ablations on several components of Expass with respect to its oversmoothing and predictive performance. Expass for different TopK Explanations. We investigate the oversmoothing and predictive performance of GNNs for different topK explanations (i.e., topK edges identified by a GNN explanation) chosen in the message passing. Results show that Expass alleviates oversmoothing by using only the topK edges to learn graph embeddings and explicitly filter out the noise from unimportant edges. In particular, we observe that the GDR values decrease (denoting higher oversmoothing) with the increase in the use of topK edges (Figure 3). More specifically, we find that the GDR value at topK=0.1 is 11.92% higher than vanilla message passing (i.e., using all edges in the graph). Expass vs. DropEdge. We compare the predictive and oversmoothing and predictive performance of Expass and DropEdge. Here, we show that message passing using optimized explanation-directed information outperforms random edge removal. We find that Expass outperforms DropEdge across both oversmoothing and accuracy metrics. In particular, on average, across different topK values, Expass improves the oversmoothing, AUROC, and F1-score performance of vanilla message passing by 71.16%, 9.53%, and 12.63%, respectively (Figure 3). (a) Group Distance Ratio (b) AUROC (c) F1-Score Figure 3: The effects of choosing only the topK percent of important edges on the (a) oversmoothing, (b) AUROC, and (c) F1-score performance of GCN model trained on Alkane-Carbonyl dataset. Over a wide range of topK values ($0.1<\text{topK}<1.0$), Expass outperforms DropEdge [45] on all the three metrics. Note that their performance converges for $\text{topK}=1.0$ as that denotes using all the edges in the graph. Expass using Node Explanations. We investigate the effect of the choice of the baseline explanation method on the performance of Expass with respect to the vanilla message passing framework. More specifically, we evaluate the predictive and explainability performance of Expass-augmented GNNs when trained using node explanations generated using Integrated Gradients (IG) [44]. Similar to the results of Expass with GNNExplainer as the baseline explanation method (Table 1), we find that Expass trained using IG explanations also improves the AUROC (+2.80%), F1-score (+1.11%), and GEF (+23.67%) of the vanilla GNN model. Our results show that the choice of explainer can make a difference in the Expass performance, depending on the dataset. For instance, IG is a node-masking explainer that is not considered a strong explanation method and its effects are variable across datasets [33]. We recommend using graph-specific explainers that optimize for fidelity and sparsity on the edges of the input graph, which would be a best fit to increase the performance of the network. See Appendix C.3 for results using PGMExplainer. Further, our results show that Expass is a model- and explainer- agnostic framework that can improve the downstream task and explainability performance across different GNN architectures using diverse GNN explainers. Expass vs. Attention. We demonstrate the utility of using explanations vs. attention weights in the message passing step using GAT [43] model architecture. On average, across four datasets, we find that Expass achieves higher AUROC (+3.85%) and F1-score (+2.24%) than the attention-based GAT model (Table 3). In addition, GNNExplainer [13] demonstrated that post hoc GNN explainers generate better explanations than attention weights, which further highlights the benefits of Expass. In comparison to Expass, GAT can be considered as a special case of our framework, where attention weights replace explanations. On the other hand, Expass has larger benefits since it can be applied to any existing GNN architectures that lack explainability. Table 2: Results of Expass for GCN using the node explanations from Integrated Gradients [44] for message passing for various datasets. Shown is average performance across three independent runs. Arrows ($\uparrow$, $\downarrow$) indicate the direction of better performance. Expass improves the predictive power (AUROC and F1-score) and degree of explainability (Graph Explanation Faithfulness) of original GNNs across multiple datasets (shaded area). Dataset | Method | AUROC ($\uparrow$) | F1-score ($\uparrow$) | GEF ($\downarrow$) ---|---|---|---|--- | DD --- | GCN --- Expass-GCN | 0.73$\pm$0.02 --- 0.75$\pm$0.01 | 0.70$\pm$0.02 --- 0.71$\pm$0.03 | 0.25$\pm$0.03 --- 0.23$\pm$0.04 | Alkane --- | GCN --- Expass-GCN | 0.97$\pm$0.01 --- 0.97$\pm$0.01 | 0.95$\pm$0.01 --- 0.95$\pm$0.01 | 0.09$\pm$0.01 --- 0.1$\pm$0.01 | Mutag --- | GCN --- Expass-GCN | 0.71$\pm$0.11 --- 0.77$\pm$0.02 | 0.87 $\pm$0.01 --- 0.88$\pm$0.01 | 0.09$\pm$0.02 --- 0.04$\pm$0.02 | Proteins --- | GCN --- Expass-GCN | 0.73$\pm$0.04 --- 0.73$\pm$0.04 | 0.68$\pm$0.04 --- 0.67$\pm$0.05 | 0.05$\pm$0.01 --- 0.04$\pm$0.01 Table 3: Results of Expass and GAT for various datasets. Shown is the average performance across three independent runs. Arrows ($\uparrow$, $\downarrow$) indicate the direction of better performance. Expass improves the predictive power (AUROC and F1-score) and degree of explainability (Graph Explanation Faithfulness) of original GNNs across multiple datasets (shaded area). Dataset | Method | AUROC ($\uparrow$) | F1-score ($\uparrow$) ---|---|---|--- | DD --- | GAT --- Expass-GCN | 0.72$\pm$0.02 --- 0.74$\pm$0.01 | 0.68$\pm$0.03 --- 0.70$\pm$0.01 | Alkane --- | GAT --- Expass-GCN | 0.97$\pm$0.01 --- 0.98$\pm$0.00 | 0.95$\pm$0.01 --- 0.96$\pm$0.01 | Mutag --- | GAT --- Expass-GCN | 0.69$\pm$0.10 --- 0.77$\pm$0.02 | 0.86 $\pm$0.01 --- 0.89$\pm$0.00 | Proteins --- | GAT --- Expass-GCN | 0.74$\pm$0.04 --- 0.74$\pm$0.03 | 0.68$\pm$0.04 --- 0.69$\pm$0.03 Q4) Visualizing explanations. Here, we visualize how the explanation develops over the training process of the GNN model. In particular, we visualize the generated explanations from Expass-GCN with GNNExplainer trained on the MUTAG dataset at different epochs during the training process and find that the explanations converge to the ground-truth explanation of a non-mutagenic molecule (i.e., the absence of a carbon ring alongside the highlighted NO2 molecules) as the training progresses (Figure 8). Further, we compare the generated explanations for a vanilla GCN and its Expass counterpart and find that the explanation for vanilla GCN falsely identifies the carbon-carbon bonds as important (Figure 4). This qualitative analysis provides further evidence for the observed higher faithfulness results (Table 1) of explanations generated using our proposed Expass framework. VanillaGCN Expass-GCN Figure 4: Visualizing the explanation generated for a non-mutagenic molecule prediction using Vanilla GCN (left) and Expass-GCN (right) with GNNExplainer method. Note that the explanation from vanilla GCN falsely identifies the carbon-carbon bonds as important. This qualitative analysis provides further evidence for the observed higher faithfulness results of explanations generated using our proposed Expass framework. ## 6 Conclusion In this work, we propose the problem of learning graph embeddings using explanation-directed message passing in GNNs. To this end, we introduce Expass, a novel message-passing framework that can be used with any existing GNN model and subgraph-optimizing explainer to learn accurate embeddings by aggregating only embeddings from nodes and edges identified as important by a GNN explainer. We perform an extensive theoretical analysis to show that Expass relieves the oversmoothing problem in GNNs, and the embedding difference between the vanilla message passing framework and Expass can be upper bounded by the difference of their respective layer weights. Our empirical results on benchmark datasets show that Expass improves the explainability of the underlying GNN model without sacrificing its predictive performance. However, the training of Expass depends on the choice of explanation method, the number of data points to explain, and the dataset of choice, which is computationally more expensive than its vanilla counterparts. We find that the training time of Expass can be improved by using techniques like batch processing and efficient sampling of correctly-classified nodes and graphs. Further, adapting post-hoc explainers to generate subgraphs utilizing the embedding space would also improve the computation time of Expass. Our proposed method and findings open exciting new avenues to learn graph representations by jointly training models and explanation methods. We anticipate that Expass could open new frontiers in graph machine learning for developing explanation-based training frameworks. ## Acknowledgements The authors would like to thank the anonymous reviewers for their helpful feedback that helped improve the work. CA would like to thank Lasse Mohr and Samuele Firmani for the helpful discussions at the beginning of the project and LOGML Summer School for connecting with the students. The views expressed here are those of the authors and do not reflect the official policy or position of the affiliated company. ## References * Zitnik et al. [2018] Marinka Zitnik, Monica Agrawal, and Jure Leskovec. Modeling polypharmacy side effects with graph convolutional networks. In _Bioinformatics_ , 2018. * Huang et al. [2020a] Kexin Huang, Cao Xiao, Lucas M Glass, Marinka Zitnik, and Jimeng Sun. Skipgnn: predicting molecular interactions with skip-graph networks. In _Scientific Reports_ , 2020a. * Jin et al. [2020] Guangyin Jin, Qi Wang, Cunchao Zhu, Yanghe Feng, Jincai Huang, and Jiangping Zhou. Addressing crime situation forecasting task with temporal graph convolutional neural network approach. In _ICMTMA_ , 2020. * Agarwal et al. [2021] Chirag Agarwal, Himabindu Lakkaraju, and Marinka Zitnik. Towards a unified framework for fair and stable graph representation learning. In _UAI_. PMLR, 2021. * Baldassarre and Azizpour [2019] Federico Baldassarre and Hossein Azizpour. Explainability techniques for graph convolutional networks. In _ICML Workshop on Learning and Reasoning with Graph-Structured Representations_ , 2019. * Faber et al. [2020] Lukas Faber, Amin K Moghaddam, and Roger Wattenhofer. Contrastive graph neural network explanation. In _ICML Workshop on Graph Representation Learning and Beyond_ , 2020\. * Huang et al. [2020b] Qiang Huang, Makoto Yamada, Yuan Tian, Dinesh Singh, Dawei Yin, and Yi Chang. Graphlime: Local interpretable model explanations for graph neural networks. _arXiv_ , 2020b. * Lucic et al. [2021] Ana Lucic, Maartje ter Hoeve, Gabriele Tolomei, Maarten de Rijke, and Fabrizio Silvestri. Cf-gnnexplainer: Counterfactual explanations for graph neural networks. _arXiv_ , 2021. * Luo et al. [2020] Dongsheng Luo, Wei Cheng, Dongkuan Xu, Wenchao Yu, Bo Zong, Haifeng Chen, and Xiang Zhang. Parameterized explainer for graph neural network. In _NeurIPS_ , 2020. * Pope et al. [2019] Phillip E Pope, Soheil Kolouri, Mohammad Rostami, Charles E Martin, and Heiko Hoffmann. Explainability methods for graph convolutional neural networks. In _CVPR_ , 2019. * Schlichtkrull et al. [2021] Michael Sejr Schlichtkrull, Nicola De Cao, and Ivan Titov. Interpreting graph neural networks for nlp with differentiable edge masking. In _ICLR_ , 2021. * Vu and Thai [2020] Minh N Vu and My T Thai. Pgm-explainer: Probabilistic graphical model explanations for graph neural networks. In _NeurIPS_ , 2020. * Ying et al. [2019] Rex Ying, Dylan Bourgeois, Jiaxuan You, Marinka Zitnik, and Jure Leskovec. Gnnexplainer: Generating explanations for graph neural networks. In _NeurIPS_ , 2019. * Simonyan et al. [2014] Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. Deep inside convolutional networks: Visualising image classification models and saliency maps. In _ICLR_ , 2014. * Yuan et al. [2021] Hao Yuan, Haiyang Yu, Jie Wang, Kang Li, and Shuiwang Ji. On explainability of graph neural networks via subgraph explorations. In _ICML_ , 2021. * Agarwal et al. [2022a] Chirag Agarwal, Owen Queen, Himabindu Lakkaraju, and Marinka Zitnik. Evaluating explainability for graph neural networks. _arXiv_ , 2022a. * Spinelli et al. [2022] Indro Spinelli, Simone Scardapane, and Aurelio Uncini. A meta-learning approach for training explainable graph neural networks. _IEEE TNNLS_ , 2022. * Zhou et al. [2020] Kaixiong Zhou, Xiao Huang, Yuening Li, Daochen Zha, Rui Chen, and Xia Hu. Towards deeper graph neural networks with differentiable group normalization. _NeurIPS_ , 2020. * Oono and Suzuki [2019] Kenta Oono and Taiji Suzuki. Graph neural networks exponentially lose expressive power for node classification. _arXiv_ , 2019. * Bruna et al. [2013] Joan Bruna, Wojciech Zaremba, Arthur Szlam, and Yann LeCun. Spectral networks and locally connected networks on graphs. _arXiv_ , 2013. * Henaff et al. [2015] Mikael Henaff, Joan Bruna, and Yann LeCun. Deep convolutional networks on graph-structured data. _arXiv_ , 2015. * Bianchi et al. [2020] Filippo Maria Bianchi, Daniele Grattarola, and Cesare Alippi. Spectral clustering with graph neural networks for graph pooling. In _ICML_ , 2020. * Stachenfeld et al. [2020] Kimberly Stachenfeld, Jonathan Godwin, and Peter Battaglia. Graph networks with spectral message passing. _arXiv_ , 2020. * Balcilar et al. [2021] Muhammet Balcilar, Renton Guillaume, Pierre Héroux, Benoit Gaüzère, Sébastien Adam, and Paul Honeine. Analyzing the expressive power of graph neural networks in a spectral perspective. In _ICLR_ , 2021. * Berg et al. [2017] Rianne van den Berg, Thomas N Kipf, and Max Welling. Graph convolutional matrix completion. _arXiv_ , 2017. * Xu et al. [2018] Keyulu Xu, Chengtao Li, Yonglong Tian, Tomohiro Sonobe, Ken-ichi Kawarabayashi, and Stefanie Jegelka. Representation learning on graphs with jumping knowledge networks. In _ICML_ , 2018. * Xu et al. [2019] Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. How powerful are graph neural networks? In _ICLR_ , 2019. * Hamilton et al. [2017] Will Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. In _NeurIPS_ , 2017. * Hu et al. [2020] Ziniu Hu, Yuxiao Dong, Kuansan Wang, and Yizhou Sun. Heterogeneous graph transformer. In _WWW_ , 2020. * Wu et al. [2020] Zonghan Wu, Shirui Pan, Fengwen Chen, Guodong Long, Chengqi Zhang, and S Yu Philip. A comprehensive survey on graph neural networks. In _IEEE TNNLS_ , 2020. * Gilmer et al. [2017] Justin Gilmer, Samuel S Schoenholz, Patrick F Riley, Oriol Vinyals, and George E Dahl. Neural message passing for quantum chemistry. In _ICML_ , 2017. * Han et al. [2020] Zhen Han, Peng Chen, Yunpu Ma, and Volker Tresp. Explainable subgraph reasoning for forecasting on temporal knowledge graphs. In _ICLR_ , 2020. * Agarwal et al. [2022b] Chirag Agarwal, Marinka Zitnik, and Himabindu Lakkaraju. Probing gnn explainers: A rigorous theoretical and empirical analysis of gnn explanation methods. In _AISTATS_ , 2022b. * Cai and Wang [2020] Chen Cai and Yusu Wang. A note on over-smoothing for graph neural networks. _ArXiv_ , 2020. * Zhou et al. [2021] Kaixiong Zhou, Xiao Huang, Daochen Zha, Rui Chen, Li Li, Soo-Hyun Choi, and Xia Hu. Dirichlet energy constrained learning for deep graph neural networks. In _NeurIPS_ , 2021. * Kazius et al. [2005] Jeroen Kazius, Ross McGuire, and Roberta Bursi. Derivation and validation of toxicophores for mutagenicity prediction. In _Journal of medicinal chemistry_ , 2005. * Sanchez-Lengeling et al. [2020] Benjamin Sanchez-Lengeling, Jennifer Wei, Brian Lee, Emily Reif, Peter Wang, Wesley Qian, Kevin McCloskey, Lucy Colwell, and Alexander Wiltschko. Evaluating attribution for graph neural networks. In _NeurIPS_ , 2020. * Shervashidze et al. [2009] Nino Shervashidze, SVN Vishwanathan, Tobias Petri, Kurt Mehlhorn, and Karsten Borgwardt. Efficient graphlet kernels for large graph comparison. In _AISTATS_ , 2009. * Borgwardt et al. [2005] Karsten M. Borgwardt, Cheng Soon Ong, Stefan Schönauer, S. V. N. Vishwanathan, Alex J. Smola, and Hans-Peter Kriegel. Protein function prediction via graph kernels. _Bioinformatics_ , 2005. * Kipf and Welling [2017] Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. In _ICLR_ , 2017. * Morris et al. [2019] Christopher Morris, Martin Ritzert, Matthias Fey, William L Hamilton, Jan Eric Lenssen, Gaurav Rattan, and Martin Grohe. Weisfeiler and leman go neural: Higher-order graph neural networks. In _AAAI_ , 2019. * Ranjan et al. [2020] Ekagra Ranjan, Soumya Sanyal, and Partha Talukdar. Asap: Adaptive structure aware pooling for learning hierarchical graph representations. In _AAAI_ , 2020. * Veličković et al. [2018] Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. Graph attention networks. In _ICLR_ , 2018. * Sundararajan et al. [2017] Mukund Sundararajan, Ankur Taly, and Qiqi Yan. Axiomatic attribution for deep networks. In _ICML_ , 2017. * Rong et al. [2020] Yu Rong, Wenbing Huang, Tingyang Xu, and Junzhou Huang. Dropedge: Towards deep graph convolutional networks on node classification. In _ICLR_ , 2020. * Dobson and Doig [2003] Paul D. Dobson and Andrew J. Doig. Distinguishing enzyme structures from non-enzymes without alignments. _Journal of Molecular Biology_ , 330(4), 2003. * Yang et al. [2016] Zhilin Yang, William Cohen, and Ruslan Salakhudinov. Revisiting semi-supervised learning with graph embeddings. In _ICML_ , 2016. ## Appendix A Proofs for Theorems in Section 4 Theorem 1. Given a non-linear activation function $\sigma$ that is Lipschitz continuous, the difference between the node embeddings between a vanilla message passing and Expass framework can be bounded by the difference in their individual weights, i.e., $\|\mathbf{h}_{u}^{(l)}-\mathbf{h^{\prime}}_{u}^{(l)}\|_{2}\leq\|\mathbf{W}_{a}^{(l)}{-}\mathbf{W^{\prime}}_{a}^{(l)}\|_{2}\|\mathbf{h}_{u}^{(l-1)}\|_{2}+\|\mathbf{W}_{n}^{(l)}{-}\mathbf{W^{\prime}}_{n}^{(l)}\|_{2}\sum_{\mathclap{v\in\mathcal{N}_{u}\cap s_{v}=1}}\|\mathbf{h}_{v}^{(l-1)}\|_{2},$ (3) where $\mathbf{W}_{a}^{(l)}$ and $\mathbf{W^{\prime}}_{a}^{(l)}$ are the weights for node $u$ in layer $l$ of the vanilla message passing and Expass framework and $\mathbf{W}_{n}^{(l)}$ and $\mathbf{W^{\prime}}_{n}^{(l)}$ are their respective weight matrix with the neighbors of node $u$ at layer $l$. ###### Proof. For a given node $u$, the node representation output by layer $l$ of the GNN is given by: $\mathbf{h}_{u}^{(l)}=\sigma\Big{(}\mathbf{W}_{a}^{(l)}\mathbf{h}_{u}^{(l-1)}+\mathbf{W}_{n}^{(l)}\sum_{\mathclap{v\in\mathcal{N}_{u}}}\mathbf{h}_{v}^{(l-1)}\Big{)},$ (4) where we consider the Agg operator as a fully-connected layer, Upd to be a sigmoid activation function $\sigma(\cdot)$, $\mathbf{W}_{a}^{(l)}$ is the weights for node $u$ in layer $l$ and $\mathbf{W}_{n}^{(l)}$ is the weight matrix with the neighbors of node $u$ at layer $l$. Let us consider an edge in-hoc explanation that generates a binary mask highlighting the important edges for the prediction of node $u$. Note that using the edge mask, we can also get a node-level mask signifying the importance of neighboring nodes. Let us denote that node explanation mask as $s_{v}$ where $s_{v}=1$ if the node is important, otherwise $s_{v}=0$. Formally, the corresponding message passing equations for Expass can be written as: $\mathbf{h^{\prime}}_{u}^{(l)}=\sigma\Big{(}\mathbf{W^{\prime}}_{a}^{(l)}\mathbf{h^{\prime}}_{u}^{(l-1)}+\mathbf{W^{\prime}}_{n}^{(l)}\sum_{\mathclap{v\in\mathcal{N}_{u}}}s_{v}\mathbf{h^{\prime}}_{v}^{(l-1)}\Big{)},$ (5) where $\mathbf{h^{\prime}}_{u}^{(l)}$ and $\mathbf{h^{\prime}}_{v}^{(l)}$ represents the embeddings of node $u$ and $v$ using the feedback explanation, and $\mathbf{W^{\prime}}_{a}^{(l)}$ and $\mathbf{W^{\prime}}_{n}^{(l)}$ represents the corresponding weights at layer $l$ for GNN model trained using Expass. The difference between the node embeddings obtained after the message-passing in layer $l$ from Equations 4-5 is given as: $\mathbf{h}_{u}^{(l)}-\mathbf{h^{\prime}}_{u}^{(l)}=\sigma\Big{(}\mathbf{W}_{a}^{(l)}\mathbf{h}_{u}^{(l-1)}+\mathbf{W}_{n}^{(l)}\sum_{\mathclap{v\in\mathcal{N}_{u}}}\mathbf{h}_{v}^{(l-1)}\Big{)}-\sigma\Big{(}\mathbf{W^{\prime}}_{a}^{(l)}~{}\mathbf{h^{\prime}}_{u}^{(l-1)}+\mathbf{W^{\prime}}_{n}^{(l)}~{}\sum_{\mathclap{v\in\mathcal{N}_{u}}}s_{v}\mathbf{h^{\prime}}_{v}^{(l-1)}\Big{)},$ (6) Taking the $\ell_{2}$-norm on both sides and assuming a normalized Lipschitz non-linear sigmoid activation, i.e., $\|\sigma(b)-\sigma(a)\|_{2}\leq\|b-a\|_{2}$, we get: $\displaystyle\|\mathbf{h}_{u}^{(l)}-\mathbf{h^{\prime}}_{u}^{(l)}\|_{2}$ $\displaystyle=\|\sigma\Big{(}\mathbf{W}_{a}^{(l)}\mathbf{h}_{u}^{(l-1)}{+}\mathbf{W}_{n}^{(l)}\sum_{\mathclap{v\in\mathcal{N}_{u}}}\mathbf{h}_{v}^{(l-1)}\Big{)}{-}\sigma\Big{(}\mathbf{W^{\prime}}_{a}^{(l)}~{}\mathbf{h^{\prime}}_{u}^{(l-1)}{+}\mathbf{W^{\prime}}_{n}^{(l)}~{}\sum_{\mathclap{v\in\mathcal{N}_{u}}}s_{v}\mathbf{h^{\prime}}_{v}^{(l-1)}\Big{)}\|_{2}$ $\displaystyle\leq\|\mathbf{W}_{a}^{(l)}\mathbf{h}_{u}^{(l-1)}+\mathbf{W}_{n}^{(l)}\sum_{\mathclap{v\in\mathcal{N}_{u}}}\mathbf{h}_{v}^{(l-1)}-\mathbf{W^{\prime}}_{a}^{(l)}~{}\mathbf{h^{\prime}}_{u}^{(l-1)}-\mathbf{W^{\prime}}_{n}^{(l)}~{}\sum_{\mathclap{v\in\mathcal{N}_{u}}}s_{v}\mathbf{h^{\prime}}_{v}^{(l-1)}\|_{2}$ $\displaystyle\leq\|\mathbf{W}_{a}^{(l)}\mathbf{h}_{u}^{(l-1)}-\mathbf{W^{\prime}}_{a}^{(l)}~{}\mathbf{h^{\prime}}_{u}^{(l-1)}+\mathbf{W}_{n}^{(l)}\sum_{\mathclap{v\in\mathcal{N}_{u}}}\mathbf{h}_{v}^{(l-1)}-\mathbf{W^{\prime}}_{n}^{(l)}\sum_{\mathclap{v\in\mathcal{N}_{u}}}s_{v}\mathbf{h^{\prime}}_{v}^{(l-1)}\|_{2}$ $\displaystyle\leq\|\mathbf{W}_{a}^{(l)}\mathbf{h}_{u}^{(l-1)}{-}\mathbf{W^{\prime}}_{a}^{(l)}~{}\mathbf{h}_{u}^{(l-1)}\|_{2}{+}\|\mathbf{W}_{n}^{(l)}\sum_{\mathclap{v\in\mathcal{N}_{u}\cap s_{v}=0}}\mathbf{h}_{v}^{(l-1)}{+}(\mathbf{W}_{n}^{(l)}-\mathbf{W^{\prime}}_{n}^{(l)})\sum_{\mathclap{v\in\mathcal{N}_{u}\cap s_{v}=1}}\mathbf{h}_{v}^{(l-1)}\|_{2}$ (Using Triangle Inequality and Faithfulness property of explanations) Given a faithful explanation, the node embeddings for node $u$ using the vanilla message passing network are equivalent to that Expass since most explainers optimize the mask to approximate the input embedding. More specifically, for a given node embedding $\mathbf{h^{\prime}}_{u}^{(l-1)}=\mathbf{h}_{u}^{(l-1)}+\epsilon_{u}$, a faithful explanation bounds the $\epsilon_{u}$ to zero. In addition to faithfulness, a GNN using vanilla message passing and Expass can predict a node $u$ to the same class only if both frameworks generate similar node embeddings (Proposition 1 in Agarwal et al. [4]). Using Matrix-norm and Triangle Inequality for the sum in the neighborhood, we get: $\displaystyle\|\mathbf{h}_{u}^{(l)}-\mathbf{h^{\prime}}_{u}^{(l)}\|_{2}\leq\|(\mathbf{W}_{a}^{(l)}-\mathbf{W^{\prime}}_{a}^{(l)})~{}\mathbf{h}_{u}^{(l-1)}\|_{2}+\|\mathbf{W}_{n}^{(l)}\|_{2}\sum_{\mathclap{v\in\mathcal{N}_{u}\cap s_{v}=0}}\|\mathbf{h}_{v}^{(l-1)}\|_{2}+$ $\displaystyle\|\mathbf{W}_{n}^{(l)}-\mathbf{W^{\prime}}_{n}^{(l)}\|_{2}\sum_{\mathclap{v\in\mathcal{N}_{u}\cap s_{v}=1}}\|\mathbf{h}_{v}^{(l-1)}\|_{2}$ Again, using the faithfulness property of explanations, the contribution of node embeddings from node $v\in\mathcal{N}(u)\cap s_{v}=0$ is irrelevant to the final embedding and can be removed. Finally, using Matrix-norm inequality on the first term, we get: $\displaystyle\|\mathbf{h}_{u}^{(l)}-\mathbf{h^{\prime}}_{u}^{(l)}\|_{2}$ $\displaystyle\leq\|\mathbf{W}_{a}^{(l)}{-}\mathbf{W^{\prime}}_{a}^{(l)}\|_{2}\|\mathbf{h}_{u}^{(l-1)}\|_{2}+\|\mathbf{W}_{n}^{(l)}{-}\mathbf{W^{\prime}}_{n}^{(l)}\|_{2}\sum_{\mathclap{v\in\mathcal{N}(u)\cap s_{v}=1}}\|\mathbf{h}_{v}^{(l-1)}\|_{2}$ Thus, we observe that the embedding difference at layer $l$ between a vanilla message passing network and the Expass is purely based on the difference between their weights and the embeddings of node $u$ and its subgraph. ∎ ###### Definition 2 (Dirichlet Energy for a Node Embedding Matrix [35]). Given a node embedding matrix $\mathbf{h}^{(l)}=[\mathbf{h}_{1}^{(l)},\dots,\mathbf{h}_{n}^{(l)}]^{T}$ learned from the GNN model at the $l^{th}$ layer, the Dirichlet Energy $E(\mathbf{h}^{(l)})$ is defined as: $E(\mathbf{h}^{(l)})=tr(\mathbf{h}^{(l)^{T}}\tilde{\Delta}\mathbf{h}^{(l)})=\frac{1}{2}\sum_{i,j\in\mathcal{V}}a_{ij}||\frac{\mathbf{h}_{i}^{(l)}}{\sqrt{1+d_{i}}}-\frac{\mathbf{h}_{j}^{(l)}}{\sqrt{1+d_{j}}}||_{2}^{2}$ (7) where $a_{ij}$ are elements in the adjacency matrix $\tilde{\mathbf{A}}$ and $d_{i},d_{j}$ is the degree of node $i$ and $j$, respectively. Cai et al. [34] extensively show that higher Dirichlet energies correspond to lower oversmoothing. Furthermore, they show that the removal of edges or ,similarly, reduction of edge weights on graphs help alleviate oversmoothing. Proposition 1 (Expass relieves Oversmoothing). Expass alleviates oversmoothing by slowing the layer-wise loss of Dirichlet energy. ###### Proof Sketch. Here, we show the capabilities of Expass as a framework that alleviates the oversmoothing problem in GNNs. To this end, we utilize the bounds on the Dirichlet energy of the Expass embeddings at the $l^{th}$ layer of the GNN model by Zhou et al. [35]: $(1-\lambda_{1})^{2}s^{(l)}_{min}E(\mathbf{h}^{(l-1)})\leq E(\mathbf{h}^{(l)})\leq(1-\lambda_{0})^{2}s^{(l)}_{max}E(\mathbf{h}^{(l-1)}),$ (8) where $\lambda_{1},\lambda_{0}$ are the non-zero eigenvalues of the symmetric normalized Laplacian $\tilde{\Delta}$ that is closest to 1 and 0, respectively, and $s^{(l)}_{min},s^{(l)}_{max}$ are the squares of the minimum and maximum singular values of weight $\mathbf{W}^{(l)}$, respectively. Since Expass reduces the input graph to its specific explanation, we argue that it can alleviate oversmoothing by reducing the information propagation along irrelevant nodes and edges. From the perspective of Dirichlet energy, we know from [19] that, for Erdős-Rényi graphs, $\lambda_{0}$ converges to 1 as the graph becomes denser. Oono et al. [19] state that GNNs oversmooth on sufficiently large graphs (similar to Erdős-Rényi graphs). Under this assumption, Expass, by definition introduces sparsity inside the $\tilde{\Delta}$ of the input graph by using a smaller set of topK important edges for learning embeddings and, thus, reduces $\lambda_{0}$ to tighten the upper-bound in Equation 8. In practice, the choice of explainer used in Expass can reduce $\lambda_{0}$ to varying degrees. More specifically, explainers that promote sparsity would push $\lambda_{0}$ closer to zero and slow down the decrease of Dirichlet energy in subsequent GNN layers. Finally, we know from Cai et al. [34] that higher values of Dirichlet energy per layer correspond to lower oversmoothing, we assert that Expass alleviates oversmoothing. ∎ ## Appendix B Experiment ### B.1 Datasets Mutag. The MUTAG [36] dataset contains 188 graph molecules labeled into two different classes according to their mutagenic properties, i.e., effect on the Gram-negative bacterium S. Typhimuriuma. Kazius et al. [36] identifies several toxicophores - motifs in the molecular graph - that correlate with mutagenicity. Alkane-Carbonyl. The Alkane-Carbonyl [37] dataset contains 1,125 molecular graphs categorized into two classes where an instance in the positive group indicates a molecule that contains an unbranched alkane and a carbonyl (C=O) functional group. DD. The DD [38] dataset was derived from [46] and contains 1,178 protein graphs where nodes represent individual amino-acids and edges represent their spatial proximity. The task is to predict whether a given protein is an enzyme or not. Proteins. The Proteins [39] dataset was derived from [46] and contains 1,113 protein graphs where nodes represent secondary structure elements and edges indicate neighborhood in the amino-acid sequence or the 3D space. The task is to predict whether a given protein is an enzyme or not. PubMed. The PubMed dataset [47] is a citation network from the PubMed database, with over 4 million nodes and edges respectively. It contains a bag- of-words representation of documents and citation links between documents. The task is to predict a node’s class among 3 classes. ### B.2 Implementation details GNN libraries and models. All our models were implemented using PyTorch Geometric (2.1.0) and PyTorch (1.11.0). For our experiments, we used baseline GNN architectures with three layers followed by ReLU layers and set the hidden dimensionality to 32. Finally, we used a single linear layer to transform the graph embeddings to their respective classes. We selected Adam as our optimizer and a weighted Cross Entropy Loss to train both vanilla and Expass frameworks. All models were trained over three independent runs with a learning rate of 0.01 for 200 epochs for DD and Proteins datasets and 150 epochs for Alkane and MUTAG datasets. Expass. We define the burn-in period as a number of epochs during training in which no explanations are used. The burn-in period is necessary to avoid feeding spurious explanations to the model since an untrained model can lead to unfaithful explanations. The length of the burn-in period was treated as a hyperparameter and fine-tuned during the model fine-tuning phase. After fine- tuning, we found that a burn-in period in the range [5, 15] worked best, whereas most Expass models outperformed their vanilla counterparts using a burn-in period of 5 and 10 epochs in our experiments. We generated explanations for a specific percentage of correctly predicted graphs sampled in each batch and were set to 0.4 for all our experiments. The generated explanations are normalized to [0, 1] and hard-masked over the topK most relevant edges, where topK is a percentage of the total number of edges in the input graph and was set to $\text{topK}\in[0.3,0.4]$ for our experiments. GNN explanation methods. At each epoch, the model weights were frozen to generate explanations, which were calculated as the median over n independent runs of GNNExplainer, in order to obtain consistent explanations. Then, the model weights were trained using generated explanations. We chose the median instead of the mean to prevent the individual outliers from significantly changing the final explanations. The number of individual runs of an explainer was treated as a hyperparameter and was set to five for GNNExplainer and one for Integrated Gradients (as it generates consistent explanations over multiple runs). In each run, the GNNExplainer was trained for 200 epochs (150 in the case of Alkane) with a learning rate of 0.01. All other hyperparameters of the explanation methods were set using the author’s guidelines. Note that these multiple iterations of the explainers are not required for Expass to perform well when using other stable GNN explainers. To summarize, the following parameters were treated as hyperparameters: the learning rate of the model, the learning rate of the explanation method, the number of epochs the explanation method was trained for, the number of times the GNNExplainer was computed at each epoch for each sampled graph, the percentage of correctly classified graphs that were randomly sampled to compute the explanations and the percentage of top edges/nodes that were selected as the most relevant. On the other hand, in the case of vanilla models, the learning rate was fine- tuned during the tuning phase. Dataset. The train, validation, and test split was at 80%, 10%, and 10% for Alkane, Proteins, and DD following prior works. In the case of MUTAG, no validation set was used due to the smaller dataset size, and the train and test split was at 80% and 20%. GNN performance metric. The GEF scores were evaluated as the mean over the individual scores of all generated explanations on the test dataset, where the explanations were hard-masked with $\textup{topK}=0.1$ for GNNExplainer, and $\textup{topK}=0.25$ for Integrated Gradients/PGMExplainer. Note that, since Integrated Gradients/PGMExplainer generates a node mask instead of an edge mask, we required a higher topK value to generate a non-empty hard mask over the input graphs, since we retain the topK most relevant nodes in the explanation mask. (a) Group Distance Ratio (b) AUROC (c) F1-Score Figure 5: The effects of the number of GNN layers on the (a) oversmoothing, (b) AUROC, and (c) F1-score performance of Expass-GCN and Vanilla-GCN trained on Alkane-Carbonyl dataset. Across models with increasing number of layers, Expass achieves higher GDR performance without sacrificing the predictive performance of the GCN model. (a) Group Distance Ratio (b) AUROC (c) F1-Score Figure 6: The effects of the number of GNN layers on the (a) oversmoothing, (b) AUROC, and (c) F1-score performance of Expass-GIN and Vanilla-GIN trained on Alkane-Carbonyl dataset. We observe that, across models with an increasing number of layers, Expass achieves higher GDR performance and there exists an inherent trade-off between oversmoothing and predictive performance of GIN. ## Appendix C Additional results ### C.1 Node classification results We extend our proposed framework to GNN models trained on different graph downstream tasks. In particular, we conduct additional experiments to obtain the over-smoothing and predictive behavior of Expass for node-level tasks. We train five state-of-the-art GNN models and their Expass counterparts on the PubMed node classification dataset. Our results show that Expass alleviates the over-smoothing effect in GNNs for models with higher depths (Figure 7) and achieves on-par or higher predictive performance (Table 4). We find that, on average, Expass augmented GCN achieves 19.53% better over-smoothing performance for node-classification GNN models with higher depths. Table 4: Results of Expass for five GNNs using PubMed node classification dataset. Shown is the average performance across five independent runs. Arrows ($\uparrow$, $\downarrow$) indicate the direction of better performance. Expass improves the predictive power (testing accuracy) of original GNNs across multiple datasets (shaded area). Method | Testing Accuracy ($\uparrow$) ---|--- | GCN --- Expass-GCN | 0.7596$\pm$0.002 --- 0.7616$\pm$0.002 | GraphConv --- Expass-GraphConv | 0.7652$\pm$0.002 --- 0.7682$\pm$0.002 | LeConv --- Expass-LeConv | 0.7424$\pm$0.003 --- 0.7244$\pm$0.010 | GraphSAGE --- Expass-GraphSAGE | 0.7462$\pm$0.004 --- 0.7533$\pm$0.002 | GIN --- Expass-GIN | 0.7233$\pm$0.002 --- 0.7310$\pm$0.009 (a) Group Distance Ratio (b) Accuracy Figure 7: The effects of the number of GNN layers on the (a) oversmoothing, (b) testing accuracy performance of Expass-GCN and Vanilla-GCN trained on PubMed node classification dataset. We observe that, across models with an increasing number of layers, Expass achieves higher GDR performance and achieves on-par or better testing accuracy. ### C.2 Burn-in period The burn-in period was treated as a hyperparameter and fine-tuned for each dataset and architecture. An example of the effect of the burn-in period on the AUROC and F1-score is reported in Table 5, where the change in performance is evaluated for the Proteins dataset when using a lag of 5, 10, and 15. Table 5: Results of Expass for various burn-in periods. Shown is the average performance across three independent runs (and standard error). Arrows ($\uparrow$, $\downarrow$) indicate the direction of better performance. Method | Burn-in period | AUROC ($\uparrow$) | F1-score ($\uparrow$) ---|---|---|--- | Expass-GIN --- | 5 --- 10 15 | 0.76 $\pm$0.04 --- 0.78$\pm$0.03 0.78$\pm$0.03 | 0.72$\pm$0.05 --- 0.73$\pm$0.04 0.73$\pm$0.03 | Expass-GraphSAGE --- | 5 --- 10 15 | 0.73 $\pm$0.04 --- 0.73$\pm$0.04 0.73$\pm$0.04 | 0.68$\pm$0.04 --- 0.69$\pm$0.04 0.69$\pm$0.04 | Expass-LeConv --- | 5 --- 10 15 | 0.76$\pm$0.02 --- 0.74$\pm$0.04 0.75$\pm$0.03 | 0.71$\pm$0.03 --- 0.69$\pm$0.04 0.71$\pm$0.03 ### C.3 PGMExplainer results Here, we showcase the flexibility of the Expass with respect to different GNN explainers. In Table 6, we show the effectiveness of Expass with respect to different explainers by further utilizing PGMExplainer [12] as the explanation generator. We utilize the graph explanations of PGMExplainer as node masks on the input graph (as they cannot generate edge-level masks) and incorporate them using our explanation-aware message-passing scheme with GCN as our underlying architecture. On average, across three datasets, we find that Expass trained using PGM-Explainer achieves higher GEF (+36.56%) than their vanilla counterparts (Table 6). Also, it achieves a boost of 11.27% in AUROC for MUTAG and a 1% improvement in the F1-score for the Alkane dataset. We observe that PGMExplainer, similar to Integrated Gradients, produces node masks, which lack detail and do not provide finer changes to the underlying GNN model, like edge masks. We hypothesize that this contributes to the large variation in the predictive performance across datasets. Table 6: Results of Expass for GCN using the node explanations from PGMExplainer [12] for message passing for various datasets. Shown is the average performance across three independent runs. Arrows ($\uparrow$, $\downarrow$) indicate the direction of better performance. Expass improves the predictive power (AUROC and F1-score) and degree of explainability (Graph Explanation Faithfulness) of original GNNs across multiple datasets (shaded area). Dataset | Method | AUROC ($\uparrow$) | F1-score ($\uparrow$) | GEF ($\downarrow$) ---|---|---|---|--- | Alkane --- | GCN --- Expass-GCN | 0.97$\pm$0.01 --- 0.97$\pm$0.01 | 0.95$\pm$0.01 --- 0.96$\pm$0.01 | 0.31$\pm$0.02 --- 0.28$\pm$0.03 | Mutag --- | GCN --- Expass-GCN | 0.71$\pm$0.11 --- 0.79$\pm$0.03 | 0.87$\pm$0.01 --- 0.86$\pm$0.01 | 0.21$\pm$0.07 --- 0.07$\pm$0.01 | Proteins --- | GCN --- Expass-GCN | 0.73$\pm$0.04 --- 0.66$\pm$0.02 | 0.68$\pm$0.04 --- 0.67$\pm$0.05 | 0.03$\pm$0.00 --- 0.02$\pm$0.00 Epoch 25 Epoch 50 Epoch 75 Epoch 100 Epoch 125 Epoch 149 Figure 8: Generated explanations from Expass-GCN trained on the MUTAG dataset at different epochs during the training process and find that the explanations does converge to the ground-truth explanation of a mutagenic molecule (i.e., the absence of a carbon ring) as the training progresses. This qualitative analysis provides further evidence for the observed higher faithfulness results of explanations generated using our proposed Expass framework.
# Determining Dust Properties in Protoplanetary Disks: SED-derived Masses and Settling With ALMA Anneliese M. Rilinger Department of Astronomy and Institute for Astrophysical Research, Boston University, 725 Commonwealth Avenue, Boston, MA 02215 Catherine C. Espaillat Department of Astronomy and Institute for Astrophysical Research, Boston University, 725 Commonwealth Avenue, Boston, MA 02215 Zihua Xin Department of Astronomy and Institute for Astrophysical Research, Boston University, 725 Commonwealth Avenue, Boston, MA 02215 Álvaro Ribas Institute of Astronomy, University of Cambridge, Madingley Road, Cambridge CB3 0HA, UK European Southern Observatory (ESO), Alonso de Córdova 3107, Vitacura, Casilla 19001, Santiago de Chile, Chile Enrique Macías ESO Garching, Karl- Schwarzschild-Str. 2, 85748, Garching bei Munchen, Germany Sarah Luettgen Department of Astronomy and Institute for Astrophysical Research, Boston University, 725 Commonwealth Avenue, Boston, MA 02215 Ann and H.J. Smead Department of Aerospace Engineering Sciences, University of Colorado at Boulder, 3775 Discovery Dr, Boulder, CO 80303 Anneliese M. Rilinger <EMAIL_ADDRESS> ###### Abstract We present spectral energy distribution (SED) modeling of 338 disks around T Tauri stars from eleven star-forming regions, ranging from $\sim$0.5 to 10 Myr old. The disk masses we infer from our SED models are typically greater than those reported from (sub)mm surveys by a factor of 1.5-5, with the discrepancy being generally higher for the more massive disks. Masses derived from (sub)mm fluxes rely on the assumption that the disks are optically thin at all millimeter wavelengths, which may cause the disk masses to be underestimated since the observed flux is not sensitive to the whole mass in the disk; SED models do not make this assumption and thus yield higher masses. Disks with more absorbing material should be optically thicker at a given wavelength; which could lead to a larger discrepancy for disks around massive stars when the disk temperature is scaled by the stellar luminosity. We also compare the disk masses and degree of dust settling across the different star-forming regions and find that disks in younger regions have more massive disks than disks in older regions, but a similar degree of dust settling. Together, these results offer potential partial solutions to the “missing” mass problem: disks around T Tauri stars may indeed have enough material to form planetary systems, though previous studies have underestimated the mass by assuming the disks to be optically thin; these planetary systems may also form earlier than previously theorized since significant dust evolution (i.e., settling) is already apparent in young disks. protoplanetary disks - star formation ## 1 Introduction The population of confirmed exoplanets detected by Kepler, TESS, radial velocity surveys, and other methods, is diverse and growing (Dressing & Charbonneau, 2013; Winn & Fabrycky, 2015). A complete picture of the formation process that can explain the origin of known exoplanets is not yet completely clear (see for example the recent review by Drazkowska et al., 2022). As the location of and source of material for planet formation (Williams & Cieza, 2011), protoplanetary disks are a key piece of the puzzle. Studying the properties of these disks can help inform and constrain various theories of planet formation, including the timescales over which the formation can occur. The Atacama Large Millimeter/submillimeter Array (ALMA) has been invaluable for studying protoplanetary disks. In particular, large surveys of various star-forming regions performed with ALMA have provided information about disk masses (see Table 1, Eisner et al., 2018; van Terwisga et al., 2020, 2022; Anderson et al., 2022) and dust substructures (e.g., Andrews et al., 2018) for disks around T Tauri stars (TTS). The regions studied by these surveys span from approximately $\sim$0.5–10 Myr, allowing for comparisons of how disk properties evolve over time. Table 1: (Sub)millimeter Surveys of Star-forming Regions Studied in this Work Region | Age | Approx. Dist. | Ref. | NtotalaaObjects in binary or multiple systems were excluded from these counts. | NdetaaObjects in binary or multiple systems were excluded from these counts. | Nsample ---|---|---|---|---|---|--- | Myr | pc | | | | $\rho$ Ophiuchus | 0.5 - 2 | 140 | Cieza et al. (2019); Williams et al. (2019) | 165 | 112 | 77 Taurus | 1 - 2 | 140 | Andrews et al. (2013); Long et al. (2019) | 98 | 65 | 25 L1641 | 1.5 | 428 | Grant et al. (2021) | 101 | 89 | 56 Cha II | 1 - 2 | 198 | Villenave et al. (2021) | 27 | 20 | 14 Lupus | 1 - 3 | 160 | Ansdell et al. (2016, 2018) | 87 | 62 | 41 Cha I | 2 - 3 | 180 | Pascucci et al. (2016) | 55 | 41 | 23 IC348 | 2 - 3 | 310 | Ruíz-Rodríguez et al. (2018) | 131 | 36 | 24 Corona Australis | 3 | 160 | Cazzoletti et al. (2019) | 33 | 16 | 10 $\sigma$ Ori | 3 - 5 | 385 | Ansdell et al. (2017) | 88 | 37 | 31 $\lambda$ Ori | 5 | 400 | Ansdell et al. (2020) | 33 | 13 | 12 Upper Sco | 5 - 10 | 145 | Barenfeld et al. (2016) | 79 | 42 | 25 Two disk properties are especially informative for understanding the planet formation process: disk dust mass and dust settling. As the mass reservoirs for planet formation, the amount of dust and gas in a disk directly constrains the number and size of planets that could potentially form in that system. The review by Bergin & Williams (2018) summarizes the various methods used to calculate disk masses (see also Miotello et al., 2022). One of the most commonly applied methods calculates disk masses from millimeter-wavelength flux measurements. The disk masses deduced from the ALMA surveys listed in Table 1 are typically larger in younger regions (1–3 Myr) than in older regions (see e.g., Barenfeld et al., 2016; van der Plas et al., 2016; van Terwisga et al., 2022); these results can be used to constrain the evolution of protoplanetary disks and the timescale of planet formation. Moreover, the estimated masses of the young disks are often too small to account for observed planetary systems (Manara et al., 2018) or require nearly 100% efficiency in the planet formation process (Mulders et al., 2021). A few possible solutions to this “missing mass” discrepancy have been offered. First, the masses obtained from millimeter observations may underestimate the actual disk masses (Ballering & Eisner, 2019; Ribas et al., 2020). Second, planet formation may occur rapidly, within the first few Myr of a disk’s lifetime (e.g., Najita & Kenyon, 2014). Third, as possibly observed by Ginski et al. (2021), the disk may accrete additional material (in significant amounts) from its surroundings, replenishing the disk (e.g., Throop & Bally, 2008; Kuffmeier et al., 2017); the amount of material we see in the disk would thus be only a fraction of the total material available for planet formation, alleviating the missing mass problem (Manara et al., 2018). Dust settling is an important early step in the planet formation process. The dust grains in disks originate from the interstellar medium (ISM) as small ($\sim$0.25$\mu$m) amorphous silicates. As the disk evolves, these dust grains collide with each other; if the collisions are inelastic the grains can grow to $\sim$micron sizes. These larger grains feel a greater drag force as they decouple from the bulk gas motion and move through the gas in the disk. The grains therefore experience a net force towards the disk midplane and settle out of the disk atmosphere (Dullemond & Dominik, 2004). Since this process takes time, it may be expected that older disks would be more settled than younger disks. However, infrared (IR) observations have shown that this process is already occurring as early as $\sim$1 Myr (Furlan et al., 2009; Ribas et al., 2017; Grant et al., 2018). See the recent review by Miotello et al. (2022) for an overview of dust settling and disk vertical structure. To build a complete picture of how disk masses and dust settling change over time – and therefore how the planet formation process is unfolding – the ideal disk sample would include as many disks as possible from star-forming regions spanning a wide range of ages. Until recently, such a study has been difficult to implement, given the complexity and computational demands of the physical models used to determine disk properties. Fortunately, we now have access to machine-learning technology that can reproduce the results of a physical model in a fraction of the computational time: Ribas et al. (2020) trained an Artificial Neural Network (ANN) to replicate the output of the D’Alessio Irradiated Accretion Disk (DIAD) models (D’Alessio et al., 1998, 1999, 2001, 2005, 2006). The ANN thus allows us to model hundreds of disks and perform a robust statistical analysis on their parameters. In this work, we use this modeling framework to consistently model a large sample of protoplanetary disks in multiple star-forming regions in order to probe how disk mass and dust settling vary over time. In Section 2, we present our sample of protoplanetary disks. Our modeling process and results are presented in Section 3, and we discuss our findings in Section 4. Finally, Section 5 provides a summary. ## 2 Disk Sample We present here our sample of consistently-modeled T Tauri stars and their surrounding disks. These objects are located in eleven star-forming regions, all of which have been studied at millimeter wavelengths: Ophiuchus (Cieza et al., 2019; Williams et al., 2019), Taurus (Andrews et al., 2013; Long et al., 2019), L1641 (Grant et al., 2021), Cha II (Villenave et al., 2021), Lupus (Ansdell et al., 2016, 2018), Cha I (Pascucci et al., 2016), IC348 (Ruíz- Rodríguez et al., 2018), Corona Australis (Cazzoletti et al., 2019), $\sigma$ Ori (Ansdell et al., 2017), $\lambda$ Ori (Ansdell et al., 2020), and Upper Scorpius (Barenfeld et al., 2016). The results from the two recent ALMA surveys of Orion A (van Terwisga et al., 2022) and Serpens (Anderson et al., 2022) are not included in our sample but will be discussed in a future work. See Table 1 for ages and distances for each region. Most of these surveys report to be complete down to the substellar limit (i.e., for spectral types $<$ M6). However, the objects surveyed in L1641, Cha II, IC 348, and Upper Sco were selected based on 24$\mu$m or 70$\mu$m detections, meaning that these surveys are likely incomplete at low stellar masses (spectral types M4 and M5). Recent results using data from the Gaia surveys (Galli et al., 2020; Luhman & Esplin, 2020; Galli et al., 2021; Grasser et al., 2021; Krolikowski et al., 2021; Esplin & Luhman, 2022) confirm the incompleteness of these samples. The total number of single Class II TTS observed in each region, Ntotal, is also reported in Table 1. Objects in binary or multiple systems with separations $<$ 1000 au were not included because companion objects can alter the disk properties and influence disk evolution (Harris et al., 2012; Kounkel et al., 2016; Ruíz-Rodríguez et al., 2016; Cox et al., 2017; Akeson et al., 2019; Barenfeld et al., 2019; Cazzoletti et al., 2019; Villenave et al., 2021; Rota et al., 2022). See the recent review by Offner et al. (2022) for a detailed discussion on the impact of binarity on disk systems. In addition to being single objects, certain other criteria must be met for an object to be included in our sample. We require the objects to have millimeter-wavelength photometry detections in order to constrain our SED models at long wavelengths. Though this requirement introduces a bias towards more massive disks since those disks are easier to detect in the millimeter, millimeter detections are crucial for constraining the model fit at long wavelengths and therefore the disk mass. The number of single Class II objects detected in the millimeter in each region, $N_{det}$ is reported in Table 1. We also require objects to have significant photometric and/or spectroscopic coverage in the infrared to constrain the rest of the SED. This requirement excludes all disks in NGC2024 (0.5 Myr, 414 pc; van Terwisga et al., 2020) and Orion (1 Myr, 400 pc; Eisner et al., 2018), despite their ALMA coverage, since they lack sufficient photometry points. No transitional disks (TDs) are included in our sample, since their inner cavities can affect the shape of an object’s SED (see for example the review by Andrews, 2020); this effect on the SED is not accounted for in our modeling framework, so we exclude these objects to avoid that source of uncertainty. TDs were determined by the shape of their SED (Cieza et al., 2010; Espaillat et al., 2012; Luhman & Mamajek, 2012; Maucó et al., 2016; van der Marel et al., 2016; Grant et al., 2018) or through resolved imaging (Isella et al., 2010; Andrews et al., 2011; Canovas et al., 2015). As noted by Andrews (2020), the current catalog of very high-resolution observations in which substructure could be detected is biased towards larger, brighter disks around more massive hosts. Excluding TDs from our sample may preferentially exclude more massive disks and introduce a bias into our sample. After applying these selection criteria, our modeling sample consists of 383 disks around T Tauri stars. Of these objects, 338 are used for our analysis; Section 3.2 describes our reasons for excluding 45 objects from the analysis. Table 1 lists the number of objects in our sample used for analysis per region, Nsample. ## 3 Analysis and Results ### 3.1 SED Models In order to determine disk properties for each of the TTS objects in our sample, we fit models to each object’s SED. SEDs were constructed using photometry points spanning visual to millimeter wavelengths, obtained using the Vizier catalog access tool (Ochsenbein et al., 2000). We include photometry from Gaia DR3, the American Association of Variable Star Observers Photometric All-Sky Survey (AAVSO; Henden et al., 2015, 2016), the Guide Star Catalog (GSC2.3; Lasker et al., 2008), 2MASS, WISE, Spitzer (labeled as c2d in the figures), Herschel, and any available millimeter photometry (usually ALMA, but also including the Submillimetre Common-User Bolometer Array on the James Clerk Maxwell Telescope (SCUBA; Holland et al., 2013) and the Submillimeter Array (SMA; Ho et al., 2004)). After visual inspection of the SED, we removed a few photometry data which were in obvious disagreement with the rest of the SED and which had apparent contamination in their images. We also include low- resolution (R $\sim$ 60–130) spectra from the InfraRed Spectrograph (IRS) on the Spitzer Space Telescope (Houck et al., 2004) for the 187 objects in our sample that were observed with this instrument. In order to counteract uneven weighting in the fitting process due to the much larger number of points in a spectrum than in the photometry for a given object, we bin the spectrum into a smaller number of points. Spectrum points are binned according to the nearest wavelength at which our model fluxes are calculated, yielding approximately 10-20 binned points, depending on the wavelength span of the spectrum. #### 3.1.1 The Artificial Neural Network Previous studies (e.g., Rubinstein et al., 2018; Macías et al., 2018; Rilinger et al., 2019; Rilinger & Espaillat, 2021) used the DIAD code (D’Alessio et al., 1998, 1999, 2001, 2005, 2006) to successfully model disk SEDs. However, these models are computationally expensive, and the ability to perform robust statistical analysis is limited. In this work, we use the ANN created by Ribas et al. (2020) to model the disk SEDs. The ANN was trained on tens of thousands of DIAD models, and can mimic DIAD outputs in a fraction of the computational time. A second ANN, ANNdiskmass is employed to determine the disk mass for each object based on the best-fit model parameters. Neither the DIAD models nor the ANN include disk mass as a free parameter; rather, disk masses are inferred from the other best-fit parameters. See the description of $\alpha$ and $\dot{M}$ below as well as Ribas et al. (2020) for details. DIAD takes various disk and stellar parameters as inputs to calculate the emission from the system. The dust in the disk is assumed to be distributed in two populations: smaller dust grains in the atmosphere of the disk and larger dust grains in the disk midplane. These populations are parameterized by the maximum dust grain size in each population (i.e., $a_{\rm max,upper}$ and $a_{\rm max,midplane}$, respectively). The amount of dust settling that has occurred is represented by the dimensionless parameter $\epsilon$ as described by D’Alessio et al. (2006) and defined as follows. We assume the grain size distributions and the dust-to-gas mass ratio ($\zeta$) to be constant in radius and vary only with height above and below the disk midplane. In order to keep the total $\zeta$ constant while still allowing for settling of large grains, $\zeta_{small}$, the dust-to-gas ratio of small grains in the atmosphere (with sizes up to $a_{\rm max,upper}$) decreases while $\zeta_{big}$, the dust-to-gas ratio of larger grains in the midplane (with sizes up to $a_{\rm max,midplane}$) increases. The settling parameter $\epsilon$ is thus defined such that $\epsilon=\zeta_{small}/\zeta_{std}$ (1) where $\zeta_{std}$ is the standard assumed dust-to-gas mass ratio of 0.01. Therefore, lower $\epsilon$ values correspond to larger $\zeta_{big}$, implying more settled disks. Disk viscosity is characterized by $\alpha$, following Shakura & Sunyaev (1973). In the DIAD models, $\alpha$ and the mass accretion rate $\dot{M}$ are the two main parameters that set the disk mass. The surface mass density $\Sigma$ is proportional to $\dot{M}$$\alpha^{-1}$; integrating $\Sigma$ over the disk radius yields disk mass. For a fixed mass accretion rate, smaller values of alpha thus correspond to larger disk masses, and vice versa. As discussed by Ribas et al. (2020), $\alpha$ and $\dot{M}$ can therefore be correlated, but they also have other smaller effects on the SED that can help constrain them. We allow alpha to vary between $10^{-4}$ and $10^{-1}$. These values are consistent with those derived by Rafikov (2017) and Ansdell et al. (2018); these works calculated $\alpha$ using measured mass accretion rates and disk sizes. ALMA molecular line observations generally support low turbulence values (on the order of 10-4 – 10${-3}$ Flaherty et al., 2015; Pinte et al., 2016; Flaherty et al., 2017), though some objects show higher values ($\sim$ 0.01 – 0.1 Flaherty et al., 2018, 2020). We vary $\dot{M}$ between $10^{-10}\ \rm{\it M}_{\odot}\,yr^{-1}{}$ and $10^{-6.5}\ \rm{\it M}_{\odot}\,yr^{-1}{}$; this is consistent with typical accretion rates reported for Class II disks (e.g., Valenti et al., 1993; Hartigan et al., 1995; Gullbring et al., 1998; Ingleby et al., 2013; Manara et al., 2016a; Simon et al., 2016). We note that DIAD calculates the radial and vertical disk density and temperature structure self-consistently. In general, the disk emission at a given wavelength will depend on the optical depth along the line of sight and the disk temperature structure. When the disk emission is optically thick, the temperature at the surface where the line-of-sight optical depth is unity will be the dominant component. As such, even when a disk is optically thick from the optical to the mm wavelengths, DIAD can still provide strong constraints on various disk parameters. The geometry of the disk is also important for fitting the SED, so the disk radius $R_{\rm disk}$ and inclination $i$ are also included as input parameters. Furthermore, the inner edge of the disk is an important source of disk emission, since the inner disk wall is assumed to be directly irradiated by the central star. The location of the inner wall is defined by the dust sublimation temperature, $T_{wall}$, and the height of the wall is scaled by a factor $z_{\rm wall}$. In addition to the nine disk parameters described above, the stellar temperature $T_{*}$, radius $R_{*}$, and mass $M_{*}$ are also important input parameters for DIAD (and therefore ANN). We ensure that we only allow for combinations of stellar parameters that are consistent with models of stellar evolution, following Ribas et al. (2020). We incorporate an $Age_{*}$ parameter into our fitting process (see Section 3.1.2); combined with $M_{*}$, we use this parameter to calculate $T_{*}$ and $R_{*}$ based on the MESA Isochrones and Stellar Tracks (MIST; Paxton et al., 2011, 2013, 2015; Dotter, 2016; Choi et al., 2016). The output $T_{*}$ and $R_{*}$ are therefore consistent with the input $M_{*}$, and all three parameters form a consistent set that can be used as input for DIAD. We generally find strong agreement between the $T_{*}$ and $R_{*}$ calculated by our models and the values reported in the literature. See Appendix A for a comparison with literature values. We include these stellar parameters in our fitting procedure in order to allow for a complete Bayesian analysis of the rest of the parameters. As discussed in Ribas et al. (2020), the stellar parameters derived in this fitting process are not intended to be precisely accurate values, as would be expected from a complete treatment of photospheric spectra. Rather, our values of $M_{*}$, $Age_{*}$, $T_{*}$, and $R_{*}$ are used to ensure that the models are internally consistent and to account for the uncertainties they produce in other parameters. Finally, once an SED has been calculated using these input parameters, we redden the output SED to the appropriate extinction value $A_{v}$ using the McClure (2009) extinction law. We determine the $A_{v}$ values using the Markov Chain Monte Carlo (MCMC) fitting process described below in Section 3.1.2. The reddened SEDs are ultimately scaled to the object’s distance, which is also fit with the MCMC. Gaia parallaxes (Gaia Collaboration et al., 2016, 2018, 2021) are used as priors where possible; the approximate distances to each region given in Table 1 are used for the 42 objects without Gaia measurements. #### 3.1.2 MCMC SED Modeling Including all of the free parameters described in the previous section, DIAD takes 13 input parameters. In addition to these parameters, we include four additional parameters to account for photometric uncertainties and outliers. One of these is the free parameter $f$ by which we scale uncertainties in flux density to account for possible systematic uncertainties, as follows: $s_{n}^{2}=\sigma_{n,model}^{2}+\sigma_{n,obs}^{2}+f^{2}\,y_{n,obs}^{2}\quad$ (2) where index n corresponds to each measurement, $\sigma_{n,model}$ is the adopted 10$\%$ uncertainty for the ANN prediction at the same wavelength, $\sigma_{n,obs}$ is the uncertainty in the flux for the measurement, $y_{n,obs}$ represents the observed fluxes of the measurement. The other three are parameters in a mixture model used to account for possible photometric outliers. Following Hogg et al. (2010), $P_{\rm out}$ is the probability that any photometry point is an outlier; $y_{\rm out}$ and $V_{\rm out}$ are the mean and variance of the outliers, respectively. Our model thus has a total of 17 free parameters. In order to sufficiently probe the large parameter space and to estimate posterior distributions for the model parameters, we implement a Bayesian analysis framework in the form of a Markov Chain Monte Carlo (MCMC) fitting process. We adopt the likelihood functions used by Xin et al. (in press), and the ptemcee (Vousden et al., 2016) parallel-tempering version of the more commonly used emcee MCMC code (Foreman-Mackey et al., 2013). Following Xin et al. (in press), we separate the photometry points into two categories: “critical”, which includes the 2MASS and millimeter-wavelength points, and “general”, which includes all other photometry points. This categorization allows us to ascribe a higher weighting to the 2MASS and ALMA photometry; fitting these points is crucial for accurately scaling the stellar photosphere and calculating the disk mass, respectively. We use four standard Gaussian likelihood functions, one each for the critical photometry data, the spectral data, $T_{*}$, and $R_{*}$, given as: $\text{\mathcal{L}}_{n,data}=\frac{1}{\sqrt{2\pi(s_{n}^{2})}}\exp{\biggl{(}-\frac{(y_{n,data}-y_{n,{\rm model}})^{2}}{2(s_{n}^{2})}\biggr{)}}$ (3) where $data$ is the data set (critical photometry, IRS spectrum, $T_{*}$, or $R_{*}$), index n corresponds to each measurement, $y_{n,x}$ represents the observed values, $y_{n,{\rm model}}$ represents the values predicted by ANN, and $s_{n}$ is the uncertainty scaling described above. We use the mixture model outlined above for the general photometry data (all other photometry points): $\text{\mathcal{L}}_{n,gen}=\frac{1-P_{\text{out}}}{\sqrt{2\pi(s_{n}^{2})}}\exp{\biggl{(}-\frac{(y_{n,gen}-y_{n,{\text{model}}})^{2}}{2(s_{n}^{2})}\biggr{)}}+\frac{P_{\text{out}}}{\sqrt{2\pi[s_{n}^{2}+V_{\text{out}}^{2}]}}\exp{\biggl{(}-\frac{(y_{n,gen}-y_{\text{out}})^{2}}{2[s_{n}^{2}+V_{\text{out}}^{2}]}\biggr{)}}$ (4) where $y_{n,gen}$ represents the observed fluxes of the photometry points. The overall likelihood function is thus: $\ln\mathcal{L}=\sum_{n=1}^{N_{\rm gen}}\ln\mathcal{L}_{n,gen}+\sum_{n=1}^{N_{\rm crit}}\ln\mathcal{L}_{n,crit}+\sum_{n=1}^{N_{\rm spect}}\ln\mathcal{L}_{n,spect}+\ln\text{\mathcal{L}}_{T_{\rm eff}}+\ln\text{\mathcal{L}}_{R_{*}}$ (5) We adopt flat priors for most parameters, except those for which we use reported values in the literature. The dust sublimation temperature $T_{wall}$ has a Gaussian prior centered around the typically-assumed value of 1400 K with a standard deviation of 50 K. Gaussian priors are also used for parallax and $M_{*}$ for objects with Gaia parallaxes, and reported $M_{*}$ values, respectively, using their values as the centers and their uncertainties as the standard deviations. For objects with reported disk inclinations, we adopt a Gaussian prior, truncated at the limits of 0 and 70 degrees of inclination (the range of values on which the ANN was trained). The maximum dust grain sizes $a_{\rm max,\ upper}$ and $a_{\rm max,\ midplane}$ are both fit with Jeffreys priors (Jeffreys, 1946, 1961), within their limits of 0.25 - 10 $\mu$m and 100 - 104 $\mu$m, respectively. All other parameters are assumed to have flat, uniform priors, within their allowed ranges of values (see Table C.1 in Ribas et al. (2020) for the ranges). We run ptemcee with three temperatures and 102 walkers (the minimum number of required walkers for three temperatures and 17 parameters) for 2 $\times$ 105 steps. This corresponds to $\sim$2000 times the autocorrelation time. We visually inspected the chain to confirm that we had achieved convergence, and discard the first 105 steps based on this inspection. The remaining 105 steps are therefore converged and represent $\sim$1000 times the autocorrelation time. Since using all 105 steps would be computationally expensive in our analysis, we sample 104 steps from these converged steps, which is still sufficient to explore the parameter space. These 104 steps were used to construct posterior distributions for the properties of interest, and to obtain the input parameters needed for ANNdiskmass to calculate the corresponding distribution of disk masses. To determine the best-fit parameters, we take the median of the $10^{4}$ steps for each parameter, and report the $1\sigma$ value for the uncertainties. ### 3.2 Model Results Using the MCMC fitting process described above, we obtain SED models for each of the 383 objects in our sample. In order to fully visualize the resulting fits for each object, we randomly sample 1000 steps from the converged MCMC chain and plot the corresponding models along with the photometry and, where available, an IRS spectrum. Some example SED fits are shown in Figure 1; SED fits for the other objects are available in an online figure set. The corner plots for all objects are available on https://zenodo.org/record/7235076 (catalog Zenodo). Figure 1: The observed SED and models for one example disk from each region. Results of the modeling process are shown by randomly selecting 1000 models (blue lines) from the posterior distributions. Instrument abbreviations are defined in Section 3.1. SEDs and models for all objects are available in an online figure set. Some objects have upper limit photometry points; these are represented by downward-facing triangles in the online figure set. We note that the photometry has not been dereddened; we redden the models as described in Section 3.1.1. After visual inspection, we determined that we obtained successful SED models, in which the models reproduce the observed SED, for 338 of the 383 objects in our sample, approximately 88%. Three examples of unsuccessful fits are shown in Figure 2. The unsuccessful fits can be attributed to a few explanations. First, TTSs are well-known to be variable on timescales ranging from minutes to decades (Siwak et al., 2018). This can explain discrepancies between photometric observations that were taken at different times. Since our model does not account for variability, differences between our model and the photometry – particularly at shorter wavelengths, where variability is more pronounced – are to be expected in some cases. Discrepancies between our model and near-IR photometry can also occur for disks around cool, very low mass stars, since the Pecaut & Mamajek (2013) colors used by DIAD are more uncertain for late spectral types (see for example, the left panel in Figure 2). Figure 2: Three examples of unsuccessful SED fits. In the left panel, the models fail to fit the J, H, and K band photometry at the peak; in the middle panel, the models fail to fit the millimeter photometry; in the right panel, the shape of the SED indicates that this object might be highly inclined (i.e., viewed nearly edge-on), which would require a scattered-light model component to be fit well. Another potential source of disagreement is our assumption regarding dust composition. The DIAD models upon which the ANN is based assume a standard mixture of olivine silicates and graphite and adopt standard opacities calculated from Mie theory. The optical constants are taken from Dorschner et al. (1995) and Draine & Lee (1984) for the silicates and graphite, respectively. DIAD, and therefore the ANN, do not fit spectral features in detail, but generally replicate the shape and size of the spectrum. In cases where our model does not perfectly replicate the IRS spectrum, the disk may have a different dust composition than assumed in our model. Millimeter photometry points with large uncertainties can also cause issues with the SED fit. As mentioned in Section 2, millimeter photometry is crucial for fitting the SED at long wavelengths, especially since in many cases, the millimeter point is the only photometry point beyond 24 $\mu$m. Points with larger uncertainties are given lower weighting in the MCMC fitting process; if the millimeter uncertainties are too large, the model can miss that point (see for example the middle panel of Figure 2). Future millimeter observations of these objects at higher sensitivities may result in lower uncertainties and therefore better-constrained model fits. Finally, the geometry and configuration of the system may add extra challenges to fitting the SED. If a disk is highly inclined (i.e., viewed nearly edge- on), a scattered light model component would be necessary for an accurate fit (see Figure 2, right panel); though DIAD can include scattered light, this feature is not yet available with the ANN. Additionally, though we attempted to exclude all objects with known companions and/or disk substructures, it is likely that some objects in our sample have companions or substructures that are yet undetected. Such a system may affect the SED in ways that are not accounted for by the DIAD models. Thus, some amount of unsuccessful SED fits is expected. Despite these challenges, we obtain successful model fits for 338 of the 383 objects in our sample. Moving forward, we consider only the 338 objects with satisfactory model fits. ## 4 Discussion ### 4.1 The Role of Optical Depth on Derived Disk Masses In this section, we compare the disk masses inferred from our modeling process to the masses previously reported in the literature. Typically, the following equation is used to convert an observed millimeter-wavelength flux to a dust mass (Hildebrand, 1983): $M_{d}=\frac{F_{\nu}d^{2}}{\kappa_{\nu}B_{\nu}(T_{d})}$ (6) where $M_{d}$ is the mass of dust in the disk, $F_{\nu}$ is the millimeter flux density, $\kappa_{\nu}$ is the dust opacity coefficient at frequency $\nu$, and $B_{\nu}$($T_{d}$) is the Planck function evaluated at dust temperature Td. This equation relies on some assumptions, namely, that the disk is isothermal and optically thin at the reference frequency. Furthermore, assumptions must be made regarding the value of $T_{d}$ (whether it is constant for all disks, or if it scales with luminosity), and the value of $\kappa_{\nu}$ and how it scales with frequency. Finally, many of the previous millimeter surveys of disks were published prior to the Gaia data releases, and therefore had to rely on constant distance assumptions for all objects, as opposed to precise measurements. Different studies chose different assumptions for the terms in Equation 6, so to ensure that we made a fair comparison, we recalculated all of the disk masses using Equation 6 according to a consistent set of assumptions. All distances were calculated from Gaia parallaxes, where possible (the distances in Table 1 were used for objects without Gaia data). In previous studies, dust temperature was assumed either to be a constant value (typically 20 K; Ansdell et al., 2016, 2017; Ruíz-Rodríguez et al., 2018; Cazzoletti et al., 2019; Williams et al., 2019; Ansdell et al., 2020) or to vary with stellar luminosity: $T_{d}=25(L_{*}/L_{\odot})^{0.25}K$ (7) as in Andrews et al. (2013), Barenfeld et al. (2016), Pascucci et al. (2016), and Villenave et al. (2021). We set $\kappa_{\nu}$ to 2.3 cm2 g-1 at 230 GHz (Beckwith et al., 1990). A common assumption is that $\kappa_{\nu}$ scales with frequency as $\nu^{1.0}$ (i.e., the power law index $\beta$ = 1.0, Ansdell et al., 2016, 2017; Cieza et al., 2019; Ansdell et al., 2020; Grant et al., 2021). Other works (i.e., Pascucci et al., 2016; Barenfeld et al., 2016) use $\beta$ = 0.4. We set $\nu$ to the frequency of the ALMA continuum observations: 225 GHz (Band 6) for most regions, except Cha I, Lupus, and Upper Sco which were observed at 338 GHz (Band 7). The choice of $\beta$ has no effect on the value of $\kappa_{\nu}$ for the surveys done in Band 6. For the surveys done in Band 7, using a $\beta$ of 0.4 yields a value of $\kappa_{\nu}$ that is $\sim$ 0.8 times the value obtained using a $\beta$ of 1, so the effect is minimal. Following Manara et al. (2022), we adopt a $\beta$ of 1 for this analysis. We calculated two sets of disk masses, one using each dust temperature assumption. The left panel of Figure 3 shows the recalculated values using the $T_{d}$ scaling in Equation 7; the right panel of Figure 3 shows recalculated values using a constant $T_{d}$ of 20 K. Figure 3: Comparison between our disk masses from SED modeling and disk masses calculated from single millimeter-wavelength fluxes. In Panel (a), we scale the dust temperature according to the stellar luminosity (calculated from our modeled T∗ and R∗ values), as in Andrews et al. (2013). In Panel (b), we assume a constant dust temperature of 20 K. In both panels, the black solid line represents the best-fit linear relationship and the black dotted line represents a one-to-one relationship. The linear fits presented here can be used to correct disk masses calculated using Equation 6. Uncertainties in the fitting parameters are provided in the text. Generally, as shown in Figure 3, the masses inferred from SED modeling are greater than the masses obtained via Equation 6. A growing number of studies report that Equation 6 underestimates the dust disk mass. In their study of disks in Taurus using radiative transfer models, Ballering & Eisner (2019) found that Equation 6 underestimates disk masses by a factor of $\sim$1-5. Similarly, using ANN, Ribas et al. (2020) found disk masses greater than those reported in Andrews et al. (2013) (calculated from Equation 6) by a factor of $\sim$3\. Macías et al. (2021) also found 3-5 times higher masses in TW Hya when comparing previous mm-flux-based mass estimates with a dust mass obtained from a radially resolved modelling of multi-wavelength ALMA observations that accounted for the radial variations in the optical depth of the disk. Most recently, Liu et al. (2022) investigated the effects of disk structure and dust properties on disk mass estimates through a radiative transfer parameter study; they find that mm-flux-based mass estimates typically underestimate disk masses by a factor of a few. For the majority of disks in our study, our model-derived masses are 1.5-5 times larger than those obtained using Equation 6, consistent with Ballering & Eisner (2019), Ribas et al. (2020), Macías et al. (2021), and Liu et al. (2022). Our result confirms the findings of these works: the assumptions required by Equation 6 can lead to an underestimate of the disk mass. Since radiative transfer models, and therefore ANN, do not rely on these assumptions, our disk masses do not suffer the same underestimation. Given the growing list of ALMA surveys of T Tauri disks, a method to determine disk masses from millimeter photometry, such as Equation 6, is a useful tool. We present here a correction to Equation 6, based on the relationships in Figure 3. In each panel of Figure 3, we used the LinearRegression model in the scikit-learn Python package to fit a line to the relationship between disk masses from our SED models and disk masses calculated using Equation 6. We repeated the linear fitting 500 times; the best-fit lines and their uncertainties reported here represent the mean and standard deviation of the coefficients from the 500 linear fits. Given a disk mass $M$ in Earth masses calculated from millimeter photometry using Equation 6, the slopes and intercepts of the best-fit lines in Figure 3 can be used to obtain a new $M_{corrected}$ in Earth masses: $log_{10}(M_{corrected})=1.09(\pm 0.02)*log_{10}(M)+0.19(\pm 0.02)$ (8) if the disk temperature was scaled by the stellar luminosity, or $log_{10}(M_{corrected})=0.97(\pm 0.02)*log_{10}(M)+0.29(\pm 0.01)$ (9) if the disk temperature was assumed to be a constant 20 K. While nearly all of our disk masses from SED modeling are larger than the recalculated literature disk masses, the discrepancy is not uniform when disk temperatures are scaled by stellar luminosity. As shown by the equations presented above, we find that disks with greater dust masses tend to deviate more from the recalculated literature values than disks with lower masses in the case where disk temperatures are scaled by the stellar luminosity. More luminous stars can supply more heat to the dust in the disk, increasing $T_{d}$. Assuming a constant dust temperature may result in an underestimate of $T_{d}$ for more massive stars, which, following Equation 6, would in principle result in an overestimate of $M_{dust}$. This has the result of flattening the linear relationship shown in Figure 3(b), since the more massive disks tend to be around more massive stars (Andrews et al., 2013; Manara et al., 2022). If we account for this effect by scaling $T_{d}$ with stellar luminosity, we observe the trend of greater discrepancy at higher masses (see Figure 3(a)). While it might seem that using the apparently incorrect fixed temperature can provide better results for more massive disks, we note that the trend is reversed for low mass disks: using a $T_{d}$ that scales with stellar luminosity provides a much better agreement with our SED- based masses. Regardless of the relationship between stellar luminosity and disk temperature, we find that Equation 6 underestimates disk mass compared to our SED-derived masses, regardless of the relationship between stellar luminosity and disk temperature. Another potential source of the discrepancy is the value of the reference dust opacity, $\kappa_{230\ GHz}$. As mentioned above, masses are typically calculated using a reference $\kappa_{230\ GHz}$ equal to 2.3 cm2 g-1, regardless of the size of the dust grains in the disk. The DIAD models compute different dust opacities depending on the size of the dust grains in the disk midplane, $a_{max,\ midplane}$. These dust opacities have a value of $\sim$2.3 cm2 g-1 for 400 $\mu$m grains, $\sim$1.9 cm2 g-1 for a grain size of 1 mm, and a minimum of 0.85 cm2 g-1 for 1 cm grains. In most disks in our sample, $a_{max,\ midplane}$ is not well-constrained, and in fact the posteriors are usually flat from 100 microns to 1 cm. The difference in opacity hence enlarges (and mostly sets) our error bars in dust mass, making sure that the uncertainties in dust opacity are properly accounted for. Even in objects where $a_{max,\ midplane}$ might be relatively well constrained, the different opacities would only be able to explain mass differences a factor of 1.5-2. To probe the effect of the value of $\kappa_{230\ GHz}$ on dust mass, we again recalculated masses using Equation 6. This time, we used the appropriate $\kappa_{230\ GHz}$ for each object as calculated by DIAD for its median $a_{max,\ midplane}$. All other assumptions were kept the same as described above and in Manara et al. (2022). Figure 4 shows the ratio of our SED-derived dust masses to these new recalculated dust masses plotted versus the new recalculated dust masses. As can be seen in the Figure, our SED-derived dust masses are still a few times greater than those calculated from Equation 6, even when accounting for different $\kappa_{230\ GHz}$. We note that uncertainties in dust composition and structure could affect the value of both $\kappa_{230\ GHz}$ and $\beta$ by as much as an order of magnitude (Testi et al., 2014; Miotello et al., 2022), but these would have the same effect both on our dust masses and the millimeter-flux ones. While these uncertainties could alter the dust masses, the differences in $\kappa_{230\ GHz}$ alone cannot explain the observed discrepancy between our dust masses and those derived from millimeter fluxes. Differences in $\kappa_{230\ GHz}$ cannot alone explain the observed discrepancy between our dust masses and those derived from millimeter fluxes. Figure 4: Comparison between our disk masses from SED modeling and disk masses calculated from single millimeter-wavelength fluxes, varying $\kappa_{230\ GHz}$ according to the median $a_{max,\ midplane}$ value of each object. The y-axis shows the ratio of our SED-derived masses to the flux-derived masses. The dashed black line shows a one-to-one relationship between the two masses, and the solid black lines show the linear fits from Figure 3. As in Figure 3, dust temperatures are scaled according to the stellar luminosity in Panel (a) and held at a constant value in Panel (b). In both cases, even when correcting for a more appropriate $\kappa_{230\ GHz}$ value, the dust masses we derive from SED fitting are still generally larger than those calculated from Equation 6. An alternative explanation for the observed trend is related to the assumption made in Equation 6 that the disk is optically thin at the reference frequency. More massive disks are expected to be optically thicker at a given wavelength due to the greater amount of absorbing material present; if these disks are optically thick, Equation 6 will underestimate the disk mass since the observed flux is not sensitive to the whole mass in the disk. Modeling by Mohanty et al. (2013), for example, showed that the 850 micron flux density becomes independent of disk mass in the optically thick limit. Ballering & Eisner (2019), Ribas et al. (2020), and Liu et al. (2022) all cite optical depth as a way to explain the discrepancy between disk mass estimates from (sub)mm fluxes and those from disk SED modeling. As mentioned above in Section 3.1.1, the SED models calculate the full disk structure, which includes the optical depth. Therefore, even when the observed millimeter emission is partially optically thick, the simulated millimeter emission takes this into account, and thus the total dust mass can be constrained. To test this explanation, we ran DIAD models for a subset of our disk sample in order to calculate optical depths. Using the median values for each model parameter from our fits, we used DIAD to calculate the structure of the disk and optical depth along the line of sight as a function of radius at 225 GHz (1.3 mm). We then computed a flux-weighted mean optical depth using the flux at each radius as the weight for the optical depth at that radius. In this way, we obtained a single average optical depth of the disk that takes into account the regions of the disk where most of the emission comes from. These flux-weighted mean optical depths are plotted versus dust mass in Figure 5 for each disk in Taurus, Lupus, Cha I and Upper Sco. (We selected these regions since they are the four regions studied by Pascucci et al. (2016); we compare our $M_{dust}$ – $M_{*}$ relationships to theirs in the following section.) In each region, optical depth generally increases with disk mass. Many disks in each region have a flux-weighted mean optical depth close to or greater than one, indicating that the optically thin assumption is not valid for these disks. Thus, disk masses calculated from Equation 6 are systematically underestimated; disk masses obtained with DIAD and ANNdiskmass, which do not assume the disks are optically thin, are more reliable for determining disk masses. We also show the disk size for each disk in Figure 5 by coloring each point according to its median outer radius from the MCMC fitting. These outer radii are larger than the FWHMs reported for resolved disks in Ansdell et al. (2016), Pascucci et al. (2016), and Barenfeld et al. (2016), though larger FWHMs generally correspond to larger outer radii in our sample. Disks with lower masses but high optical depths have smaller radii, indicating that they are compact and dense, while disks with higher masses but lower optical depths have larger radii and are comparatively less dense. #### 4.1.1 Impact on Planet Formation Models The higher masses resulting from our SED models can help alleviate the tension between the observed masses of exoplanetary systems and protoplanetary disks. Previously, Manara et al. (2018) reported that the dust disks in Lupus and Cha I are a factor of 3–5 too small to account for the typical mass of planetary cores. A more recent study by Mulders et al. (2021) reports more comparable masses between disks and exoplanetary systems, but requires nearly 100% efficiency in the planet formation process to explain the observed masses. The masses we report here are typically a factor of 1.5–5 greater than the recalculated literature values. The assumption that disks are optically thin can therefore result in a significant underestimation of their masses. Our disk masses ease the requirement for extremely high planet formation efficiency. We note, however, that increasing the available disk mass by a factor of 1.5-5 would imply planet formation efficiencies between $\sim$20% and $\sim$60%, to achieve the same exoplanet disk masses as reported by Mulders et al. (2021). Given that planet formation models do not still have a robust estimate of the planet formation efficiency, it is not fully clear if the optical depth alone could solve the “missing” mass problem or if alternative scenarios are still required (e.g., planetesimals already being formed, late infall of material onto the disk). Figure 5: Flux-weighted mean optical depth of each disk plotted versus disk mass for four star-forming regions: (a) Taurus, (b) Lupus, (c) Cha I, and (d) Upper Sco. Optical depths are flux-weighted averages of the 1.3 mm optical depth at each radius in the disk as calculated by the DIAD models. Disk masses are the median values from our models for each object. The horizontal dashed line denotes an optical depth of 1. Colors represent the disk outer radii. Some disks in each region have optical depth greater than 1, especially higher mass disks. ### 4.2 Disk Mass Trends with Host Mass Figure 6: The relationship between disk dust masses and host star masses for four star-forming regions: (a) Taurus, (b) Lupus, (c), Cha I, and (d) Upper Sco. Red points are median $M_{dust}$ and $M_{*}$ values from the MCMC output and black points are $M_{dust}$ and $M_{*}$ values from (Pascucci et al., 2016). Linear fits for these two groups are shown in their matching colors, with a 1$\sigma$ uncertainty shown by the shaded region. The values listed in the upper left corner of each panel are the Spearman correlation coefficient and p value for our modeled $M_{dust}$ and $M_{*}$ values. Taurus has a lower fraction of modeled objects than other regions, so we also include a fit to a mixed sample (purple line): our modeled objects plus single objects and non- detections not included in our sample, with masses recalculated from their mm fluxes (blue symbols, see Section 4.1). Disk masses have been shown to scale with the masses of their host stars in individual star forming regions (e.g., Pascucci et al., 2016; Testi et al., 2022), though the trend is murkier when multiple regions are considered together (Manara et al., 2022). In Figure 6 we plot disk dust mass versus host mass for the four star-forming regions studied by Pascucci et al. (2016). Following Pascucci et al. (2016), we use the Bayesian linear fitting procedure developed by Kelly (2007), which allows for measurement uncertainties as well as upper limits, to fit the recalculated dust mass relationship. Our MCMC fitting process results in asymmetric uncertainties, which are not supported by the Kelly (2007) fitting procedure, so we instead use Python Abstract Interfaces for Data Analysis (PAIDA)111paida.sourceforge.net to obtain a linear fit to the modeled dust masses that incorporates the uncertainties. For Lupus, Cha I, and Upper Sco, we find $M_{dust}$-$M_{*}$ relationships that are consistent with those reported in Pascucci et al. (2016); i.e., the slopes and intercepts of the fits reported in Table 2 for each region are within 1-2$\sigma$ of the fits reported by Pascucci et al. (2016). We generally find slightly larger intercepts than Pascucci et al. (2016); this can be attributed to the result reported above, that we find larger disk masses than previously reported. The relationship we find for disks in Taurus is shallower than previously reported. This difference in slope may be due to the inherent bias in our sample introduced by the exclusion of upper limits. Since most the disks in Taurus were observed at lower sensitivity by the SMA, many of these objects only have upper limits for their disk masses; the linear fit for Taurus reported by Pascucci et al. (2016) incorporates these points, but our sample does not and is limited to higher mass objects. To mediate the effect of this difference in sensitivity, we also present an $M_{dust}$-$M_{*}$ relationship for a mixed sample of Taurus objects. This mixed sample includes our modeled objects, plus any detected objects not known to be in binary or multiple systems but excluded for other reasons, and single non-detections. We adopt host and disk masses from our SED models where possible; for the remaining objects, host masses were taken from Pascucci et al. (2016) and disk masses were recalculated from literature fluxes using our consistent method described in Section 4.1, with a constant dust temperature of 20 K. The $M_{dust}$-$M_{*}$ relationship of this mixed sample, shown in Figure 6(a) is more consistent with the relationship from Pascucci et al. (2016), as well as the other three regions presented in Figure 6. We note that the correlation of the $M_{dust}$-$M_{*}$ relationship in each region is only moderate at best. Spearman correlation coefficients ($r_{s}$) for our modeled $M_{dust}$ and $M_{*}$ values are given in the upper left corner of each panel in Figure 6 and range between $\sim$ 0.3 and $\sim$ 0.5. The $r_{s}$ value describes how well the relationship between $M_{dust}$ and $M_{*}$ is described by a monotonic function; $r_{s}$ values closer to 1 indicate stronger correlation. For each region, we also report the p value, which gives the probability that a random, uncorrelated sample would produce a similar correlation. For Taurus, the $r_{s}$ and p values given in Figure 6 are for our sample only; for the mixed sample, $r_{s}$ = 0.425 and p = 0.001. Though the $r_{s}$ values show only a moderate correlation, the low p values for each of these four regions indicates that the moderate correlation is statistically significant. Pascucci et al. (2016) also reported moderate correlations, with dispersions of $\sim$0.8 dex in the $M_{dust}$ – $M_{*}$ relationship for each region. The $M_{dust}$-$M_{*}$ relationships for the remaining seven regions in our sample are shown with their $r_{s}$ and p values in Figure 7; we report the slopes and intercepts of the relationships in Table 2. We note that for three regions (Cha II, Corona Australis, and $\lambda$ Ori) the sample sizes may be too small to accurately assess correlation. Of the four regions with substantial samples (Ophiuchus, L1641, IC 348, and $\sigma$ Ori), only one, $\sigma$ Ori, has a statistically significant trend (i.e., p $<$ 0.05). The $r_{s}$ value for this region is comparable to the $r_{s}$ values for the regions shown in Figure 6, showing a moderate correlation between $M_{dust}$ and $M_{*}$. Ophiuchus, L1641, and IC 348 do not show any significant correlation. All eleven regions in this sample show a large dispersion in $M_{dust}$ for a given $M_{*}$. As noted in the review by Manara et al. (2022), the fact that this dispersion is observed in all regions suggests it is due to inherent variation in disk properties, as opposed to environmental effects in a specific region. Figure 7: The relationship between disk dust masses and host star masses for the star-forming regions not studied by Pascucci et al. (2016): (a) Ophiuchus, (b) L1641, (c), Cha II, (d) IC348, (e) Corona Australis, (f) $\sigma$ Ori, and (g) $\lambda$ Ori. Red points are median $M_{dust}$ and $M_{*}$ values from the MCMC output. Linear fits for each region are shown in red, with a 1$\sigma$ uncertainty shown by the shaded region. The values listed in the upper left corner of each panel are the Spearman correlation coefficient and p value for our modeled $M_{dust}$ and $M_{*}$ values. Table 2: $M_{dust}$ – $M_{*}$ Relationships Region | This work | Pascucci et al. (2016) ---|---|--- | $m$ | $b$ | $m$ | $b$ Ophiuchus TTS | 0.6$\pm$0.1 | 1.1$\pm$0.1 | … | … Ophiuchus TTS and BDs | 0.9$\pm$0.1 | 1.2$\pm$0.1 | … | … Taurus TTS | 0.2$\pm$0.2 | 1.9$\pm$0.1 | 1.1$\pm$0.2 | 1.0$\pm$0.1 | 2.1$\pm$0.3aaThese values are the fit to a mixed sample: our modeled objects plus single objects and non-detections not included in our sample, with masses recalculated from their mm fluxes. | 1.4$\pm$0.2aaThese values are the fit to a mixed sample: our modeled objects plus single objects and non-detections not included in our sample, with masses recalculated from their mm fluxes. | | Taurus TTS and BDs | 1.6$\pm$0.1 | 2.1$\pm$0.1 | … | … L1641 | 0.2$\pm$0.1 | 1.43$\pm$0.1 | … | … Cha II | 1.9$\pm$0.2 | 1.7$\pm$0.1 | … | … Lupus TTS | 1.0$\pm$0.1 | 1.5$\pm$0.1 | 1.1$\pm$0.3 | 1.4$\pm$0.2 Lupus TTS and BDs | 1.1$\pm$0.1 | 1.3$\pm$0.1 | … | … Cha I | 1.5$\pm$0.1 | 1.7$\pm$0.1 | 1.3$\pm$0.2 | 1.1$\pm$0.1 IC 348 | 0.0$\pm$0.2 | 0.9$\pm$0.1 | … | … Corona Australis | 0.0$\pm$0.2 | 0.4$\pm$0.1 | … | … $\sigma$ Ori | 1.1$\pm$0.2 | 1.5$\pm$0.1 | … | … $\lambda$ Ori | 0.6$\pm$0.2 | 0.6$\pm$0.1 | … | … Upper Sco TTS | 1.8$\pm$0.2 | 0.9$\pm$0.1 | 1.9$\pm$0.4 | 0.8$\pm$0.2 Upper Sco TTS and BDs | 0.8$\pm$0.1 | 0.6$\pm$0.1 | … | … Note. — The slopes ($m$) and intercepts ($b$) reported here correspond to the following linear relationship: log$(M_{dust}/M_{\oplus})$ = $m$ x log$(M_{*}/M_{\odot})$ \+ $b$. #### 4.2.1 TTS Compared to Brown Dwarfs Figure 8: Disk dust masses plotted versus host masses for the four star- forming regions studied in Rilinger & Espaillat (2021). TTS objects are shown as red open circles and BD objects are shown as black filled circles. TTS disk and host masses are from the SED models presented in this work. BD disk masses were derived via SED modeling by Rilinger & Espaillat (2021) and BD host masses were taken from the literature (see citations in Rilinger & Espaillat (2021)). Linear fits for just the TTS are represented by red lines; linear fits to the TTS and BDs together are shown in blue. As in Figure 6(a), we show linear fits to both our sample and the mixed sample for Taurus. Given the scaling of disk mass with host mass for TTS (e.g., Pascucci et al., 2016; Ansdell et al., 2016, 2017, and this work), we explored whether this trend extends to disks around lower mass brown dwarf (BD, $M_{*}\lesssim$ 0.08 $M_{\odot}$) hosts. Four of the star-forming regions presented in this work were also studied by Rilinger & Espaillat (2021), who used DIAD to obtain SED models for 49 disks around BDs in Ophiuchus, Taurus, Lupus, and Upper Sco. Since the modeling process is consistent between that work and this, we can directly compare the disk masses we infer for TTS to the disk masses Rilinger & Espaillat (2021) inferred for BDs. In Figure 8, we show how the inclusion of BD disks affects the $M_{dust}$ vs $M_{*}$ relationships presented above. Using PAIDA, we obtained linear fits for the combined sample of TTS and BDs in each region, which are shown in Figure 8 by the blue lines. We also reproduce the linear fits to the TTS sample alone from Figure 6 for comparison. In the three younger regions, the the combined TTS and BD fit is generally in good agreement with the fit to the TTS (see Table 2). In Upper Sco, however, including the BD disks in the fit results in a significantly shallower slope. In other words, in this region, BD disk dust masses are higher than would be expected from the TTS disk mass – host mass trend. Instead of decreasing with time (see Section 4.3), BD disk masses appear to remain consistent over a span of $\sim$ 10 Myr. Studies of protoplanetary disk fractions in various star-forming regions indicate that the fraction of stars that host disks increases for later spectral types (Carpenter et al., 2006; Luhman & Mamajek, 2012; Ribas et al., 2015; Luhman, 2022). In other words, lower-mass stars appear to retain their disks longer than higher-mass stars. The results presented here add additional evidence to this theory. BD disks are typically less massive than their TTS counterparts in a given star-forming region, as expected based on the established trend of decreasing disk mass with host mass. However, the BD disk dust masses remain constant across regions spanning as much as 10 Myr in age while the corresponding TTS disks exhibit a substantial decrease in disk dust mass (Section 4.3). One way in which disks are expected to lose mass is through photoevaporation due to radiation from the central star (Ercolano et al., 2017; Weber et al., 2020). Picogna et al. (2021) recently showed that the rate of photoevaporation scales with stellar mass. Given the lower masses of BDs compared to pre-main- sequence stars, photoevaporation rates should be lower in BD disks than TTS disks. That BD disks appear to maintain their disk masses over millions of years, while their TTS counterparts show a decrease in mass, may be the result of slower disk dissipation occurring in these objects. The potential for planet formation in a given disk is thus both enhanced by and inhibited by the mass of its host: a more massive YSO is more likely to host a larger disk early in its lifetime, but it will also likely emit more UV and X-ray radiation, which will dissipate the disk more quickly. ### 4.3 Comparison Between Regions of Various Ages In this work we studied disks in eleven star-forming regions, which vary in age from $\sim$0.5 to $\sim$10 Myr. By consistently modeling the SEDs of the disks in each of these regions, we can directly compare their properties and probe how the population of disks changes with time. In particular, we consider the disk masses and the degree of dust settling; both of these properties have implications for the planet-forming potential of these disks. #### 4.3.1 Disk Mass with Age Figure 9: Probability distribution functions of dust mass for each region and for all regions together (lower right panel). The red dashed line in each panel represents the median dust mass for the overall sample ($\sim$ 8 $M_{\oplus}$) and the blue dotted lines represent the median dust mass in each region. Disk mass distributions for each object are constructed using 10000 random samples from the object’s MCMC chain. We use these masses to construct normalized histograms of disk mass for each region. These normalized histograms are plotted in Figure 9 for each region, as well as the normalized overall mass distribution including all regions. As shown in Figure 9, the peak of the mass distribution tends to shift to lower masses as the age of the region increases. Young regions such as Taurus have mass distributions that peak above the median dust mass of the entire sample (represented by the red dashed line in Figure 9), while older regions such as Upper Sco have mass distributions that peak below the median dust mass. The mass distributions of moderately-aged regions such as IC 348 peak near the overall median. Two regions appear to deviate from this trend: Ophiuchus and Corona Australis. Ophiuchus has a mass distribution that peaks near the median of the overall sample, despite being the youngest region in our sample. This result is consistent with the findings of Williams et al. (2019), who report Ophiuchus disk dust masses to be smaller than the masses of disks in other, slightly older regions. A possible explanation suggested by Williams et al. (2019) is that the local cloud environment may significantly affect disk masses. Kuffmeier et al. (2020), for example, report that star-forming regions with higher ionization rates may contain less massive disks, since ionization can decrease disk size. The mass distribution for disks in Corona Australis also peaks at lower disk masses than may be expected for a moderately young star- forming region. Another potential explanation for the low disk masses in these regions is descibed by Testi et al. (2022). Dynamical modeling by Turrini et al. (2012, 2019); Gerbig et al. (2019) indicates that formation of planets within a disk can trigger collisional fragmentation of planetesimals, increasing the total mass of grains detectable at millimeter wavelengths. If planets can form within the first 0.5 – 1 Myr of a disk’s lifetime, the formation of this secondary population of dust grains may not have started yet in the disks in Ophiuchus and Corona Australis. Furthermore, recent studies of Corona Australis (Cazzoletti et al., 2019; Esplin & Luhman, 2022) suggest that the region may be comprised of two populations: the young ($<$ 3 Myr) Coronet Cluster, and an older, more extended population. The lower disk masses we find in this region support the existence of this older population. Our finding that disk dust masses generally decrease as the age of the region increases is consistent with previous studies (e.g., Ansdell et al., 2016; Barenfeld et al., 2016; Pascucci et al., 2016; Cieza et al., 2019; van Terwisga et al., 2019; Villenave et al., 2021; van Terwisga et al., 2022), which have found that older regions typically contain less massive disks than their younger counterparts. However, an important caveat is that we have only included objects which have millimeter-wavelength detections. The different ALMA surveys that studied each region had a range of sensitivities and thus different detection limits; regions observed with lower sensitivities will thus be biased towards higher masses, since the lower mass disks are likely to be less luminous, smaller in radius, and harder to detect. In particular, the survey of $\sigma$ Ori performed by Ansdell et al. (2017) was only sensitive to dust masses down to $\sim$2 $M_{\oplus}$. This upper limit is much higher than that achieved by surveys of other regions: Barenfeld et al. (2016), Pascucci et al. (2016), Ansdell et al. (2016), and Ansdell et al. (2020) all detect dust masses on the order of a few tenths of an Earth mass for Upper Sco, Cha I, Lupus, and $\lambda$ Ori, respectively. The lower sensitivity of the $\sigma$ Ori survey compared to that of other regions means that this region’s mass distribution is biased towards higher dust masses. In order to account for the different sensitivities of the ALMA surveys, we show cumulative distributions of the dust mass in each region in Figure 10. These distributions were constructed using a Kaplan-Meier esitmator from the lifelines package (Davidson-Pilon et al., 2021)222https://lifelines.readthedocs.io/. In order to ensure a fair comparison between regions, we include masses for all single objects in each region. We do not include any binary objects or objects in multiple systems, since the presence of a companion can significantly affect the evolution of a disk (e.g., Akeson et al., 2019; Barenfeld et al., 2019; Kounkel et al., 2019; Zurlo et al., 2020, 2021; Offner et al., 2022). For objects in our sample, we use the masses derived from DIAD. For single objects that were detected in the millimeter but either unable to be fit or not included in our sample, we use the dust mass calculated from the flux measurement (as in Section 4.1). For non-detections, we compute the upper limit fluxes as three times the uncertainty plus any positive measured flux density, as in Barenfeld et al. (2016); mass upper limits were calculated from these flux upper limits. The L1641 survey targets were selected based on Herschel 70 $\mu$m detections, which biases the sample towards brighter, more massive disks as noted by Grant et al. (2021). To counteract this bias we include objects from the more complete Survey of Orion Disks with ALMA (SODA) project (van Terwisga et al., 2022). Even when accounting for the sample biases and non-detections, the results reported above still hold: Figure 10 shows a decrease of disk mass with age, and Ophiuchus and Corona Australis still appear to have lower disk masses than expected from their ages. To further explore the decrease in disk mass with time, we plot in Figure 11 the median disk mass for each region versus age of the region. Median values are determined from the cumulative distribution functions shown in Figure 10 by taking the value of $M_{dust}$ at which 50% of disks have masses $\geq M_{dust}$. We do not include Corona Australis, $\lambda$ Ori, or $\sigma$ Ori since these regions have $<$ 50% detections in the millimeter. Following Testi et al. (2022), we also show the trend $M_{dust}\propto t^{-1}$. Testi et al. (2022) normalized this trend to the median dust mass in Ophiuchus. As we have argued above, disks in Ophiuchus tend to have unusually low disk masses for their age; we opt instead to scale the $M_{dust}\propto t^{-1}$ trend to the weighted average median dust mass at 1.5 Myr (i.e., the weighted average of the median values for Taurus, Cha II, and L1641). This trend agrees reasonably well with our median disk masses and indicates that the disk dust mass will be reduced to $\sim$ one-third of its initial amount after 3 Myr, which is consistent with typical disk dissipation timescales (Mamajek, 2009; Ribas et al., 2014). Figure 10: Cumulative distribution functions of dust mass for each region. We include dust masses from our SED modeling for objects in our sample, as well as consistently-recalculated dust masses (see Section 4.1) for objects not included in our sample (including non-detections) in order to more fairly compare between regions. Figure 11: Median disk dust masses for regions with $>$50% mm-flux detections, plotted versus age of the region. The black dashed line shows the relationship $M_{dust}\propto t^{-1}$, scaled to the weighted mean mass at 1.5 Myr. #### 4.3.2 Dust Settling with Age Figure 12 shows the settling parameter $\epsilon$ probability distribution functions for each of the eleven regions as well as the overall distribution for all regions together. Larger values of $\epsilon$ correspond to lower degrees of settling. As for the disk dust mass distributions, these normalized histograms were constructed from 10000 random $\epsilon$ values drawn from the MCMC chain for each object. The $\epsilon$ distributions for all regions are remarkably similar. Nearly every region peaks at the lowest $\epsilon$ value allowed by our SED modeling, though some regions have a second, smaller peak at higher $\epsilon$ values. We note that $\epsilon$ is defined as the ratio between the dust-to-gas mass ratio in the disk atmosphere and the global dust-to-gas mass ratio of the disk (assumed to be 0.01). Lower values of $\epsilon$ thus correspond to dust- depleted disk atmospheres caused by a high degree of settling. Our observed trend indicates that significant amounts of dust settling has occurred in these disks, regardless of their age. One may expect the degree of settling in a disk to increase with the age of the disk. However, our result is in agreement with recent studies, which have shown that disks show evidence of dust settling even at young ages ($\sim$ 1Myr; Furlan et al., 2009; Lewis & Lada, 2016; Grant et al., 2018; Rilinger & Espaillat, 2021). Large dust grains are expected to settle to the disk midplane faster than small dust grains, since larger grains decouple more easily from the gas in the disk, and feel a drag force as they orbit the central star (Dullemond & Dominik, 2004). The high degrees of settling that we find in young disks may therefore indicate that the dust grains in these disks grew significantly in the first $\sim$1 Myr. Alternatively, high degrees of settling may be the result of low turbulence in these disks. Turbulence causes the dust and gas in a disk to mix, which counteracts the settling process. If the gas density is sufficiently low (Dullemond & Dominik, 2004), or the gas velocity in the disk atmosphere is low (e.g., Ciesla, 2007; Flaherty et al., 2015, 2017), turbulence is less efficient and thus dust settling can occur more easily. The high degrees of settling we find in the disks studied here may therefore be the result of rapid dust grain growth and/or lack of turbulence in the upper layers of the disks. Figure 12: Probability distribution functions of the dust settling parameter $\epsilon$ for each region, and for all regions together (lower right panel). The red dashed line in each panel represents the median $\epsilon$ value for the overall sample ($\sim 10^{-3}$) and the blue dotted lines represent the median $\epsilon$ values for each region. ## 5 Summary and Conclusion Using the physically motivated DIAD models in conjunction with an ANN, we have obtained SED fits for 338 disks around T Tauri stars. These disks are located in the following eleven star-forming regions, ranging in age from $\sim$ 0.5 – 10 Myr old: Ophiuchus, Taurus, L1641, Cha II, Lupus, Cha I, IC 348, Corona Australis, $\sigma$ Ori, $\lambda$ Ori, and Upper Scorpius. Our main results are as follows: 1. 1. We confirm that the previously-reported mm-flux-based masses are a factor of 1.5-5 lower than the masses derived from physical models, consistent with results in Ballering & Eisner (2019); Ribas et al. (2020). Masses derived from millimeter fluxes depend on the assumption that the disk is optically thin at that wavelength; if the disk is optically thick, this assumption results in an underestimate of the disk mass. Our SED models do not rely on this assumption, and thus yield more accurate masses. The discrepancy between the two methods is higher for the more massive disks when the disk temperature is scaled by the stellar luminosity. We present two equations (Equations 8 and 9) to be used for correcting millimeter flux-derived disk masses to account for this effect. 2. 2. We find $M_{dust}$ – $M_{*}$ relationships that are generally consistent with those reported in Pascucci et al. (2016). Five of the star-forming regions studied here (Taurus, Lupus, Cha I, $\sigma$ Ori, and Upper Sco) show statistically significant (p $<$ 0.05), moderate correlations. Ophiuchus, L1641, and IC348 do not have a statistically significant trend; Cha II, Corona Australis, and $\lambda$ Ori have too few points to accurately assess their $M_{dust}$ – $M_{*}$ relationships. We do not find any trend of this relationship with age. 3. 3. Brown dwarfs are generally consistent with the $M_{dust}$ – $M_{*}$ relationships observed for TTS in younger regions. In the 5-10 Myr-old Upper Sco region, however, brown dwarf disks are more massive than predicted from their host masses. This result may indicate that brown dwarf disks dissipate more slowly than their higher-mass companions; photoevaporation is more efficient in higher-mass stars, which may explain this trend. 4. 4. We find clear evolution in the disk masses with time, with older regions having lower disk masses than younger regions, in agreement with previous studies (e.g., Barenfeld et al., 2016; van der Plas et al., 2016; van Terwisga et al., 2022). As an important caveat, different sensitivities of ALMA surveys of various star-forming regions may have biased portions of our sample towards higher masses. Future, more sensitive surveys may reveal a larger population of low-mass disks, especially in older regions. 5. 5. The degree of dust settling appears to be consistent across age, with even the youngest regions showing appreciable settling. This result agrees with Ribas et al. (2017) and Grant et al. (2018), who also reported appreciable levels of dust settling in regions as young as 1 Myr. In conclusion, these results may help to ease the reported “missing” mass problem. By assuming the disks to be optically thin, previous studies may have underestimated the masses of disks around T Tauri stars; our physical models are not limited by this assumption and yield masses that are more consistent with observed planetary systems. Furthermore, the high degrees of dust settling we find in disks of all ages may indicate that dust processing and evolution happens quickly; thus planetary systems may also form earlier than previously theorized. We thank the referee for the careful reading of the manuscript and helpful feedback. AMR thanks A. Meredith Hughes and Philip Muirhead for their insightful comments and discussion. This work was funded by NASA ADAP 80NSSC20K0451. We acknowledge support from the NRAO Student Observing Support program through award SOSPA8-007. Á.R. has been supported by the UK Science and Technology research Council (STFC) via the consolidated grant ST/S000623/1 and by the European Union’s Horizon 2020 research and innovation programme under the Marie Sklodowska-Curie grant agreement No. 823823 (RISE DUSTBUSTERS project).This work has made use of data from the European Space Agency (ESA) mission Gaia (https://www.cosmos.esa.int/gaia), processed by the Gaia Data Processing and Analysis Consortium (DPAC, https://www.cosmos.esa.int/web/gaia/dpac/consortium). Funding for the DPAC has been provided by national institutions, in particular the institutions participating in the Gaia Multilateral Agreement. This paper utilizes the D’Alessio Irradiated Accretion Disk (DIAD) code. We wish to recognize the work of Paola D’Alessio, who passed away in 2013. Her legacy and pioneering work live on through her substantial contributions to the field. ## Appendix A Agreement with Reported Literature Values Here we present a comparison between the stellar values obtained from our SED modeling and values reported in the literature. As described in Section 3.1.1, we incorporate literature values as priors for stellar parameters when literature values exist, and we ensure our $T_{*}$ and $R_{*}$ values are consistent with stellar evolution models using MIST (Paxton et al., 2011, 2013, 2015; Dotter, 2016; Choi et al., 2016). For a given $M_{*}$ and $Age_{*}$, $T_{*}$ and $R_{*}$ are calculated with MIST; these four values are then used in our MCMC fitting process. Literature values for $M_{*}$, $T_{*}$ and $R_{*}$ are taken from: Walter et al. (1997); Luhman & Rieke (1999); López Martí et al. (2005); Wilking et al. (2005); Lada et al. (2006); Cieza et al. (2007); Forbrich & Preibisch (2007); Luhman et al. (2007); Muench et al. (2007); Luhman et al. (2008); Sicilia- Aguilar et al. (2008); Meyer & Wilking (2009); Cieza et al. (2010); Hernández et al. (2010); McClure et al. (2010); Bayo et al. (2011); Erickson et al. (2011); Rigliaco et al. (2012); Andrews et al. (2013); Spezzi et al. (2013); Alcalá et al. (2014); Herczeg & Hillenbrand (2014); Hernández et al. (2014); Manara et al. (2015); Barenfeld et al. (2016); Manara et al. (2016a, b); Pascucci et al. (2016); Alcalá et al. (2017); Ansdell et al. (2017); Grant et al. (2018); Ruíz-Rodríguez et al. (2018); Cazzoletti et al. (2019); Ansdell et al. (2020). Stellar temperatures are determined from spectral types; for objects with reported spectral types but no reported temperatures, we used the Pecaut & Mamajek (2013) scaling to convert the reported spectral types to $T_{*}$ values. Stellar radii were either obtained via comparison to a reference SED (Spezzi et al., 2008, 2013), or calculated from the stellar luminosity and $T_{*}$. Luminosities were obtained by applying a bolometric correction to the observed $J$ band magnitude. In Figure A.1, we compare the previously reported $M_{*}$, $T_{*}$ and $R_{*}$ values to those obtained via our SED modeling. For all three parameters, we find very strong agreement. We also present a comparison of the best-fit parallaxes we obtain for each object versus the parallaxes reported by Gaia in Figure A.2, which also shows a strong agreement. As mentioned in Section 3.1.1, Gaia parallaxes are included as a prior in our model, so a strong correlation is expected. Figure A.3 shows a comparison between our modeled mass accretion rates ($\dot{M}$) and $\dot{M}$ values reported in the literature (scaled to Gaia distances). Literature $\dot{M}$ values are taken from Natta et al. (2006) for Ophiuchus, Antoniucci et al. (2011) for Cha II, Alcalá et al. (2017) for Lupus, Manara et al. (2017) for Cha I, Natta et al. (2014) for Sigma Ori, and Manara et al. (2020) for Upper Sco. These works calculate the accretion luminosity, and hence $\dot{M}$, from Br$\gamma$, Pa$\beta$ and other emission line equivalent widths (Natta et al., 2006; Antoniucci et al., 2011; Natta et al., 2014) or UV excess (Alcalá et al., 2017; Manara et al., 2017, 2020). We note that the ANN was trained on $log_{10}$($\dot{M}$) values between -10 and -6.5. Some objects have reported $\dot{M}$ values outside this range, though these objects have $\alpha$ values that are well-constrained, so this does not introduce a bias in our disk masses. The $\dot{M}$ error bars from DIAD are generally much larger than those obtained by measuring $\dot{M}$ with emission lines or UV excess. In principle these two accretion rates do not necessarily need to match, since the mass accretion rate onto the star could be different than in the disk. The mass accretion rate onto the star could also be variable (e.g., Robinson & Espaillat, 2019), by as much as two orders of magnitude (Claes et al., 2022), while the mass accretion rate probed by DIAD would be an average that sets the surface density profile in an alpha-disk. Finally, we note that chromospheric emission can create noise in UV observations, making it difficult to derive accurate $\dot{M}$ values below the noise threshold (Ingleby et al., 2011). Manara et al. (2013) report a noise threshold of $\sim 10^{-10.5}\ \rm{\it M}_{\odot}\,yr^{-1}{}$ for objects with similar $M_{*}$ and ages as the objects presented here. Despite these challenges, we find a general agreement between $\dot{M}$ values from our modeling and $\dot{M}$ values from the literature above the noise threshold. Figure A.1: Comparison between reported values and values obtained via our SED modeling for $M_{*}$ (panel a), $T_{*}$ (panel b) and $R_{*}$ (panel c). The black dotted line in each panel represents a one-to-one correlation. Our modeled values for each of these parameters agree very well with previously reported values. Figure A.2: Comparison between parallax values obtained via our SED modeling and those measured by Gaia. The black dotted line represents a one-to-one correlation. Our parallax values agree very well with previously reported values. Figure A.3: Comparison between mass accretion rate values obtained via our SED modeling and those reported in the literature. The black dotted line represents a one-to-one correlation. The black dashed line denotes the approximate threshold below which chromospheric activity interferes with derivations of $\dot{M}$ in the literature (Manara et al., 2013). ## References * Akeson et al. (2019) Akeson, R. L., Jensen, E. L. N., Carpenter, J., et al. 2019, ApJ, 872, 158 * Alcalá et al. (2014) Alcalá, J. M., Natta, A., Manara, C. F., et al. 2014, A&A, 561, A2 * Alcalá et al. (2017) Alcalá, J. M., Manara, C. F., Natta, A., et al. 2017, A&A, 600, A20 * Anderson et al. (2022) Anderson, A. R., Williams, J. P., van der Marel, N., et al. 2022, arXiv e-prints, arXiv:2204.08731 * Andrews (2020) Andrews, S. M. 2020, ARA&A, 58, 483 * Andrews et al. (2013) Andrews, S. M., Rosenfeld, K. A., Kraus, A. L., & Wilner, D. J. 2013, ApJ, 771, 129 * Andrews et al. (2011) Andrews, S. M., Wilner, D. J., Espaillat, C., et al. 2011, ApJ, 732, 42 * Andrews et al. (2018) Andrews, S. M., Huang, J., Pérez, L. M., et al. 2018, ApJ, 869, L41 * Ansdell et al. (2017) Ansdell, M., Williams, J. P., Manara, C. F., et al. 2017, AJ, 153, 240 * Ansdell et al. (2016) Ansdell, M., Williams, J. P., van der Marel, N., et al. 2016, ApJ, 828, 46 * Ansdell et al. (2018) Ansdell, M., Williams, J. P., Trapman, L., et al. 2018, ApJ, 859, 21 * Ansdell et al. (2020) Ansdell, M., Haworth, T. J., Williams, J. P., et al. 2020, AJ, 160, 248 * Antoniucci et al. (2011) Antoniucci, S., García López, R., Nisini, B., et al. 2011, A&A, 534, A32 * Ballering & Eisner (2019) Ballering, N. P., & Eisner, J. A. 2019, AJ, 157, 144 * Barenfeld et al. (2016) Barenfeld, S. A., Carpenter, J. M., Ricci, L., & Isella, A. 2016, ApJ, 827, 142 * Barenfeld et al. (2019) Barenfeld, S. A., Carpenter, J. M., Sargent, A. I., et al. 2019, ApJ, 878, 45 * Bayo et al. (2011) Bayo, A., Barrado, D., Stauffer, J., et al. 2011, A&A, 536, A63 * Beckwith et al. (1990) Beckwith, S. V. W., Sargent, A. I., Chini, R. S., & Guesten, R. 1990, AJ, 99, 924 * Bergin & Williams (2018) Bergin, E. A., & Williams, J. P. 2018, arXiv e-prints, arXiv:1807.09631 * Canovas et al. (2015) Canovas, H., Schreiber, M. R., Cáceres, C., et al. 2015, ApJ, 805, 21 * Carpenter et al. (2006) Carpenter, J. M., Mamajek, E. E., Hillenbrand, L. A., & Meyer, M. R. 2006, ApJ, 651, L49 * Cazzoletti et al. (2019) Cazzoletti, P., Manara, C. F., Baobab Liu, H., et al. 2019, A&A, 626, A11 * Choi et al. (2016) Choi, J., Dotter, A., Conroy, C., et al. 2016, ApJ, 823, 102 * Ciesla (2007) Ciesla, F. J. 2007, ApJ, 654, L159 * Cieza et al. (2007) Cieza, L., Padgett, D. L., Stapelfeldt, K. R., et al. 2007, ApJ, 667, 308 * Cieza et al. (2010) Cieza, L. A., Schreiber, M. R., Romero, G. A., et al. 2010, ApJ, 712, 925 * Cieza et al. (2019) Cieza, L. A., Ruíz-Rodríguez, D., Hales, A., et al. 2019, MNRAS, 482, 698 * Claes et al. (2022) Claes, R. A. B., Manara, C. F., Garcia-Lopez, R., et al. 2022, A&A, 664, L7 * Cox et al. (2017) Cox, E. G., Harris, R. J., Looney, L. W., et al. 2017, ApJ, 851, 83 * D’Alessio et al. (2001) D’Alessio, P., Calvet, N., & Hartmann, L. 2001, ApJ, 553, 321 * D’Alessio et al. (2006) D’Alessio, P., Calvet, N., Hartmann, L., Franco-Hernández, R., & Servín, H. 2006, ApJ, 638, 314 * D’Alessio et al. (1999) D’Alessio, P., Calvet, N., Hartmann, L., Lizano, S., & Cantó, J. 1999, ApJ, 527, 893 * D’Alessio et al. (1998) D’Alessio, P., Cantö, J., Calvet, N., & Lizano, S. 1998, ApJ, 500, 411 * D’Alessio et al. (2005) D’Alessio, P., Hartmann, L., Calvet, N., et al. 2005, ApJ, 621, 461 * Davidson-Pilon et al. (2021) Davidson-Pilon, C., Kalderstam, J., Jacobson, N., et al. 2021, CamDavidsonPilon/lifelines: 0.26.0, doi:10.5281/zenodo.4816284 * Dorschner et al. (1995) Dorschner, J., Begemann, B., Henning, T., Jaeger, C., & Mutschke, H. 1995, A&A, 300, 503 * Dotter (2016) Dotter, A. 2016, ApJS, 222, 8 * Draine & Lee (1984) Draine, B. T., & Lee, H. M. 1984, ApJ, 285, 89 * Drazkowska et al. (2022) Drazkowska, J., Bitsch, B., Lambrechts, M., et al. 2022, arXiv e-prints, arXiv:2203.09759 * Dressing & Charbonneau (2013) Dressing, C. D., & Charbonneau, D. 2013, ApJ, 767, 95 * Dullemond & Dominik (2004) Dullemond, C. P., & Dominik, C. 2004, A&A, 421, 1075 * Eisner et al. (2018) Eisner, J. A., Arce, H. G., Ballering, N. P., et al. 2018, ApJ, 860, 77 * Ercolano et al. (2017) Ercolano, B., Rosotti, G. P., Picogna, G., & Testi, L. 2017, MNRAS, 464, L95 * Erickson et al. (2011) Erickson, K. L., Wilking, B. A., Meyer, M. R., Robinson, J. G., & Stephenson, L. N. 2011, AJ, 142, 140 * Espaillat et al. (2012) Espaillat, C., Ingleby, L., Hernández, J., et al. 2012, ApJ, 747, 103 * Esplin & Luhman (2022) Esplin, T. L., & Luhman, K. L. 2022, AJ, 163, 64 * Flaherty et al. (2020) Flaherty, K., Hughes, A. M., Simon, J. B., et al. 2020, ApJ, 895, 109 * Flaherty et al. (2015) Flaherty, K. M., Hughes, A. M., Rosenfeld, K. A., et al. 2015, ApJ, 813, 99 * Flaherty et al. (2018) Flaherty, K. M., Hughes, A. M., Teague, R., et al. 2018, ApJ, 856, 117 * Flaherty et al. (2017) Flaherty, K. M., Hughes, A. M., Rose, S. C., et al. 2017, ApJ, 843, 150 * Forbrich & Preibisch (2007) Forbrich, J., & Preibisch, T. 2007, A&A, 475, 959 * Foreman-Mackey et al. (2013) Foreman-Mackey, D., Hogg, D. W., Lang, D., & Goodman, J. 2013, PASP, 125, 306 * Furlan et al. (2009) Furlan, E., Watson, D. M., McClure, M. K., et al. 2009, ApJ, 703, 1964 * Gaia Collaboration et al. (2016) Gaia Collaboration, Prusti, T., de Bruijne, J. H. J., et al. 2016, A&A, 595, A1 * Gaia Collaboration et al. (2018) Gaia Collaboration, Brown, A. G. A., Vallenari, A., et al. 2018, A&A, 616, A1 * Gaia Collaboration et al. (2021) —. 2021, A&A, 649, A1 * Galli et al. (2020) Galli, P. A. B., Bouy, H., Olivares, J., et al. 2020, A&A, 643, A148 * Galli et al. (2021) —. 2021, A&A, 646, A46 * Gerbig et al. (2019) Gerbig, K., Lenz, C. T., & Klahr, H. 2019, A&A, 629, A116 * Ginski et al. (2021) Ginski, C., Facchini, S., Huang, J., et al. 2021, ApJ, 908, L25 * Grant et al. (2018) Grant, S. L., Espaillat, C. C., Megeath, S. T., et al. 2018, ApJ, 863, 13 * Grant et al. (2021) Grant, S. L., Espaillat, C. C., Wendeborn, J., et al. 2021, ApJ, 913, 123 * Grasser et al. (2021) Grasser, N., Ratzenböck, S., Alves, J., et al. 2021, A&A, 652, A2 * Gullbring et al. (1998) Gullbring, E., Hartmann, L., Briceño, C., & Calvet, N. 1998, ApJ, 492, 323 * Harris et al. (2012) Harris, R. J., Andrews, S. M., Wilner, D. J., & Kraus, A. L. 2012, ApJ, 751, 115 * Hartigan et al. (1995) Hartigan, P., Edwards, S., & Ghandour, L. 1995, ApJ, 452, 736 * Henden et al. (2015) Henden, A. A., Levine, S., Terrell, D., & Welch, D. L. 2015, in American Astronomical Society Meeting Abstracts, Vol. 225, American Astronomical Society Meeting Abstracts #225, 336.16 * Henden et al. (2016) Henden, A. A., Templeton, M., Terrell, D., et al. 2016, VizieR Online Data Catalog, II/336 * Herczeg & Hillenbrand (2014) Herczeg, G. J., & Hillenbrand, L. A. 2014, ApJ, 786, 97 * Hernández et al. (2010) Hernández, J., Morales-Calderon, M., Calvet, N., et al. 2010, ApJ, 722, 1226 * Hernández et al. (2014) Hernández, J., Calvet, N., Perez, A., et al. 2014, ApJ, 794, 36 * Hildebrand (1983) Hildebrand, R. H. 1983, QJRAS, 24, 267 * Ho et al. (2004) Ho, P. T. P., Moran, J. M., & Lo, K. Y. 2004, ApJ, 616, L1 * Hogg et al. (2010) Hogg, D. W., Bovy, J., & Lang, D. 2010, arXiv e-prints, arXiv:1008.4686 * Holland et al. (2013) Holland, W. S., Bintley, D., Chapin, E. L., et al. 2013, MNRAS, 430, 2513 * Houck et al. (2004) Houck, J. R., Roellig, T. L., Van Cleve, J., et al. 2004, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 5487, Proc. SPIE, ed. J. C. Mather, 62–76 * Ingleby et al. (2011) Ingleby, L., Calvet, N., Bergin, E., et al. 2011, ApJ, 743, 105 * Ingleby et al. (2013) Ingleby, L., Calvet, N., Herczeg, G., et al. 2013, ApJ, 767, 112 * Isella et al. (2010) Isella, A., Carpenter, J. M., & Sargent, A. I. 2010, ApJ, 714, 1746 * Jeffreys (1946) Jeffreys, H. 1946, Proceedings of the Royal Society of London Series A, 186, 453 * Jeffreys (1961) Jeffreys, H. 1961, Theory of probability, 3rd edn. (Clarendon Press, Oxford), ix+447 * Kelly (2007) Kelly, B. C. 2007, ApJ, 665, 1489 * Kounkel et al. (2016) Kounkel, M., Megeath, S. T., Poteet, C. A., Fischer, W. J., & Hartmann, L. 2016, ApJ, 821, 52 * Kounkel et al. (2019) Kounkel, M., Covey, K., Moe, M., et al. 2019, AJ, 157, 196 * Krolikowski et al. (2021) Krolikowski, D. M., Kraus, A. L., & Rizzuto, A. C. 2021, AJ, 162, 110 * Kuffmeier et al. (2017) Kuffmeier, M., Haugbølle, T., & Nordlund, Å. 2017, ApJ, 846, 7 * Kuffmeier et al. (2020) Kuffmeier, M., Zhao, B., & Caselli, P. 2020, A&A, 639, A86 * Lada et al. (2006) Lada, C. J., Muench, A. A., Luhman, K. L., et al. 2006, AJ, 131, 1574 * Lasker et al. (2008) Lasker, B. M., Lattanzi, M. G., McLean, B. J., et al. 2008, AJ, 136, 735 * Lewis & Lada (2016) Lewis, J. A., & Lada, C. J. 2016, ApJ, 825, 91 * Liu et al. (2022) Liu, Y., Linz, H., Fang, M., et al. 2022, arXiv e-prints, arXiv:2210.07478 * Long et al. (2019) Long, F., Herczeg, G. J., Harsono, D., et al. 2019, ApJ, 882, 49 * López Martí et al. (2005) López Martí, B., Eislöffel, J., & Mundt, R. 2005, A&A, 444, 175 * Luhman (2022) Luhman, K. L. 2022, AJ, 163, 25 * Luhman & Esplin (2020) Luhman, K. L., & Esplin, T. L. 2020, AJ, 160, 44 * Luhman & Mamajek (2012) Luhman, K. L., & Mamajek, E. E. 2012, ApJ, 758, 31 * Luhman & Rieke (1999) Luhman, K. L., & Rieke, G. H. 1999, ApJ, 525, 440 * Luhman et al. (2007) Luhman, K. L., Adame, L., D’Alessio, P., et al. 2007, ApJ, 666, 1219 * Luhman et al. (2008) Luhman, K. L., Allen, L. E., Allen, P. R., et al. 2008, ApJ, 675, 1375 * Macías et al. (2021) Macías, E., Guerra-Alvarado, O., Carrasco-González, C., et al. 2021, A&A, 648, A33 * Macías et al. (2018) Macías, E., Espaillat, C. C., Ribas, Á., et al. 2018, ApJ, 865, 37 * Mamajek (2009) Mamajek, E. E. 2009, in American Institute of Physics Conference Series, Vol. 1158, Exoplanets and Disks: Their Formation and Diversity, ed. T. Usuda, M. Tamura, & M. Ishii, 3–10 * Manara et al. (2022) Manara, C. F., Ansdell, M., Rosotti, G. P., et al. 2022, arXiv e-prints, arXiv:2203.09930 * Manara et al. (2016a) Manara, C. F., Fedele, D., Herczeg, G. J., & Teixeira, P. S. 2016a, A&A, 585, A136 * Manara et al. (2018) Manara, C. F., Morbidelli, A., & Guillot, T. 2018, A&A, 618, L3 * Manara et al. (2015) Manara, C. F., Testi, L., Natta, A., & Alcalá, J. M. 2015, A&A, 579, A66 * Manara et al. (2013) Manara, C. F., Testi, L., Rigliaco, E., et al. 2013, A&A, 551, A107 * Manara et al. (2016b) Manara, C. F., Rosotti, G., Testi, L., et al. 2016b, A&A, 591, L3 * Manara et al. (2017) Manara, C. F., Testi, L., Herczeg, G. J., et al. 2017, A&A, 604, A127 * Manara et al. (2020) Manara, C. F., Natta, A., Rosotti, G. P., et al. 2020, A&A, 639, A58 * Maucó et al. (2016) Maucó, K., Hernández, J., Calvet, N., et al. 2016, ApJ, 829, 38 * McClure (2009) McClure, M. 2009, ApJ, 693, L81 * McClure et al. (2010) McClure, M. K., Furlan, E., Manoj, P., et al. 2010, ApJS, 188, 75 * Meyer & Wilking (2009) Meyer, M. R., & Wilking, B. A. 2009, PASP, 121, 350 * Miotello et al. (2022) Miotello, A., Kamp, I., Birnstiel, T., Cleeves, L. I., & Kataoka, A. 2022, arXiv e-prints, arXiv:2203.09818 * Mohanty et al. (2013) Mohanty, S., Greaves, J., Mortlock, D., et al. 2013, ApJ, 773, 168 * Muench et al. (2007) Muench, A. A., Lada, C. J., Luhman, K. L., Muzerolle, J., & Young, E. 2007, AJ, 134, 411 * Mulders et al. (2021) Mulders, G. D., Pascucci, I., Ciesla, F. J., & Fernandes, R. B. 2021, ApJ, 920, 66 * Najita & Kenyon (2014) Najita, J. R., & Kenyon, S. J. 2014, MNRAS, 445, 3315 * Natta et al. (2014) Natta, A., Testi, L., Alcalá, J. M., et al. 2014, A&A, 569, A5 * Natta et al. (2006) Natta, A., Testi, L., & Randich, S. 2006, A&A, 452, 245 * Ochsenbein et al. (2000) Ochsenbein, F., Bauer, P., & Marcout, J. 2000, A&AS, 143, 23 * Offner et al. (2022) Offner, S. S. R., Moe, M., Kratter, K. M., et al. 2022, arXiv e-prints, arXiv:2203.10066 * Pascucci et al. (2016) Pascucci, I., Testi, L., Herczeg, G. J., et al. 2016, ApJ, 831, 125 * Paxton et al. (2011) Paxton, B., Bildsten, L., Dotter, A., et al. 2011, ApJS, 192, 3 * Paxton et al. (2013) Paxton, B., Cantiello, M., Arras, P., et al. 2013, ApJS, 208, 4 * Paxton et al. (2015) Paxton, B., Marchant, P., Schwab, J., et al. 2015, ApJS, 220, 15 * Pecaut & Mamajek (2013) Pecaut, M. J., & Mamajek, E. E. 2013, ApJS, 208, 9 * Picogna et al. (2021) Picogna, G., Ercolano, B., & Espaillat, C. C. 2021, MNRAS, 508, 3611 * Pinte et al. (2016) Pinte, C., Dent, W. R. F., Ménard, F., et al. 2016, ApJ, 816, 25 * Rafikov (2017) Rafikov, R. R. 2017, ApJ, 837, 163 * Ribas et al. (2015) Ribas, Á., Bouy, H., & Merín, B. 2015, A&A, 576, A52 * Ribas et al. (2020) Ribas, Á., Espaillat, C. C., Macías, E., & Sarro, L. M. 2020, A&A, 642, A171 * Ribas et al. (2014) Ribas, Á., Merín, B., Bouy, H., & Maud, L. T. 2014, A&A, 561, A54 * Ribas et al. (2017) Ribas, Á., Espaillat, C. C., Macías, E., et al. 2017, ApJ, 849, 63 * Rigliaco et al. (2012) Rigliaco, E., Natta, A., Testi, L., et al. 2012, A&A, 548, A56 * Rilinger & Espaillat (2021) Rilinger, A. M., & Espaillat, C. C. 2021, ApJ, 921, 182 * Rilinger et al. (2019) Rilinger, A. M., Espaillat, C. C., & Macías, E. 2019, ApJ, 878, 103 * Robinson & Espaillat (2019) Robinson, C. E., & Espaillat, C. C. 2019, ApJ, 874, 129 * Rota et al. (2022) Rota, A. A., Manara, C. F., Miotello, A., et al. 2022, A&A, 662, A121 * Rubinstein et al. (2018) Rubinstein, A. E., Macías, E., Espaillat, C. C., et al. 2018, ApJ, 860, 7 * Ruíz-Rodríguez et al. (2016) Ruíz-Rodríguez, D., Ireland, M., Cieza, L., & Kraus, A. 2016, MNRAS, 463, 3829 * Ruíz-Rodríguez et al. (2018) Ruíz-Rodríguez, D., Cieza, L. A., Williams, J. P., et al. 2018, MNRAS, 478, 3674 * Shakura & Sunyaev (1973) Shakura, N. I., & Sunyaev, R. A. 1973, A&A, 500, 33 * Sicilia-Aguilar et al. (2008) Sicilia-Aguilar, A., Henning, T., Juhász, A., et al. 2008, ApJ, 687, 1145 * Simon et al. (2016) Simon, M. N., Pascucci, I., Edwards, S., et al. 2016, ApJ, 831, 169 * Siwak et al. (2018) Siwak, M., Ogloza, W., Moffat, A. F. J., et al. 2018, MNRAS, 478, 758 * Spezzi et al. (2008) Spezzi, L., Alcalá, J. M., Covino, E., et al. 2008, ApJ, 680, 1295 * Spezzi et al. (2013) Spezzi, L., Cox, N. L. J., Prusti, T., et al. 2013, A&A, 555, A71 * Testi et al. (2014) Testi, L., Birnstiel, T., Ricci, L., et al. 2014, in Protostars and Planets VI, ed. H. Beuther, R. S. Klessen, C. P. Dullemond, & T. Henning, 339 * Testi et al. (2022) Testi, L., Natta, A., Manara, C. F., et al. 2022, A&A, 663, A98 * Throop & Bally (2008) Throop, H. B., & Bally, J. 2008, AJ, 135, 2380 * Turrini et al. (2012) Turrini, D., Coradini, A., & Magni, G. 2012, ApJ, 750, 8 * Turrini et al. (2019) Turrini, D., Marzari, F., Polychroni, D., & Testi, L. 2019, ApJ, 877, 50 * Valenti et al. (1993) Valenti, J. A., Basri, G., & Johns, C. M. 1993, AJ, 106, 2024 * van der Marel et al. (2016) van der Marel, N., Verhaar, B. W., van Terwisga, S., et al. 2016, A&A, 592, A126 * van der Plas et al. (2016) van der Plas, G., Ménard, F., Ward-Duong, K., et al. 2016, ApJ, 819, 102 * van Terwisga et al. (2019) van Terwisga, S. E., Hacar, A., & van Dishoeck, E. F. 2019, A&A, 628, A85 * van Terwisga et al. (2022) van Terwisga, S. E., Hacar, A., van Dishoeck, E. F., Oonk, R., & Portegies Zwart, S. 2022, arXiv e-prints, arXiv:2202.11057 * van Terwisga et al. (2020) van Terwisga, S. E., van Dishoeck, E. F., Mann, R. K., et al. 2020, A&A, 640, A27 * Villenave et al. (2021) Villenave, M., Ménard, F., Dent, W. R. F., et al. 2021, A&A, 653, A46 * Vousden et al. (2016) Vousden, W. D., Farr, W. M., & Mandel, I. 2016, MNRAS, 455, 1919 * Walter et al. (1997) Walter, F. M., Vrba, F. J., Wolk, S. J., Mathieu, R. D., & Neuhauser, R. 1997, AJ, 114, 1544 * Weber et al. (2020) Weber, M. L., Ercolano, B., Picogna, G., Hartmann, L., & Rodenkirch, P. J. 2020, MNRAS, 496, 223 * Wilking et al. (2005) Wilking, B. A., Meyer, M. R., Robinson, J. G., & Greene, T. P. 2005, AJ, 130, 1733 * Williams et al. (2019) Williams, J. P., Cieza, L., Hales, A., et al. 2019, ApJ, 875, L9 * Williams & Cieza (2011) Williams, J. P., & Cieza, L. A. 2011, ARA&A, 49, 67 * Winn & Fabrycky (2015) Winn, J. N., & Fabrycky, D. C. 2015, ARA&A, 53, 409 * Zurlo et al. (2020) Zurlo, A., Cieza, L. A., Pérez, S., et al. 2020, MNRAS, 496, 5089 * Zurlo et al. (2021) Zurlo, A., Cieza, L. A., Ansdell, M., et al. 2021, MNRAS, 501, 2305
# The Force Balance of Electrons During Kinetic Anti-parallel Magnetic Reconnection J. Egedal H. Gurram S. Greess Department of Physics, University of Wisconsin-Madison, Madison, Wisconsin 53706, USA W. Daughton A. Lê Los Alamos National Laboratory, Los Alamos, New Mexico 87545, USA ###### Abstract Fully kinetic simulations are applied to the study of 2D anti-parallel reconnection, elucidating the dynamics by which the electron fluid maintains force balance within both the electron diffusion region (EDR) and the ion diffusion region (IDR). Inside the IDR, magnetic field-aligned electron pressure anisotropy (${p_{e\parallel}}\gg{p_{e\perp}})$ develops upstream of the EDR. Compared to previous investigations, the use of modern computer facilities allows for simulations at the natural proton to electron mass ratio $m_{i}/m_{e}=1836$. In this high-$m_{i}/m_{e}$-limit the electron dynamics changes qualitatively, as the electron inflow to the EDR is enhanced and mainly driven by the anisotropic pressure. Using a coordinate system with the $x$-direction aligned with the reconnecting magnetic field and the $y$-direction aligned with the central current layer, it is well-known that for the much studied 2D laminar anti-parallel and symmetric scenario the reconnection electric field at the $X$-line must be balanced by the $\partial p_{exy}/\partial x$ and $\partial p_{eyz}/\partial z$ off-diagonal electron pressure stress components. We find that the electron anisotropy upstream of the EDR imposes large values of $\partial p_{exy}/\partial x$ within the EDR, and along the direction of the reconnection $X$-line this stress cancels with the stress of a previously determined theoretical form for $\partial p_{eyz}/\partial z$. The electron frozen-in law is instead broken by pressure tensor gradients related to the direct heating of the electrons by the reconnection electric field. The reconnection rate is free to adjust to the value imposed externally by the plasma dynamics at larger scales. ###### pacs: ## I Introduction A key issue in plasma physics with importance to space weather modeling is related to how electrons decouple from the motion of magnetic field lines in small scale electron diffusion regions (EDRs). The phenomenon is known as magnetic reconnection dungey:1953 , which permits the magnetic field topology to rearrange, often accompanied by large scale release of stored magnetic energy. Well-known examples include reconnection associated with solar flares masuda:1994 and magnetic storms in the Earth’s magnetosphere mcpherron:1973 . Despite substantial efforts including laboratory, spacecraft, numerical and theoretical investigations zweibel:2016 , a detailed understanding of the electron dynamics within the EDR is still not fully developed. In particular, a full mathematical description of how the electron frozen-in condition is violated (permitting ${\bf E}+{\bf u}_{e}\times{\bf B}$ to be finite within the EDR) has still not been reached. Reconnection is observed in many different environments and geometries. For the Earth’s magnetotail, symmetric and anti-parallel reconnection (where the opposing inflow regions have similar strength magnetic fields but are oppositely directed) is of particular importance torbert:2018 . Open questions are concerned with the extent to which 3D effects fundamentally alter the structure of the reconnection region. For example, there is no doubt that reconnection current layers in the Earth’s magnetotail only have finite lengths and therefore must be 3D in nature. However, observations from the Magnetospheric Multiscale (MMS) mission demonstrate how the widths of reconnection layers typically approach the small length scales associated with the local electron orbit width. The length scales perpendicular to a particular reconnection current layer are therefore usually much shorter than the length scale along the current layer, and a range of observations further suggest that the local reconnection dynamics are well captured by 2D laminar models egedal:2018 ; egedal:2019 ; greess:2021 ; schroeder:2022 . For laminar and near steady state reconnection the force balance of the electron fluid is described by the generalized Ohm’s law, which similar to the Navier-Stokes equation for a regular fluid includes pressure and inertial effects: ${\bf E}+{\bf u}_{e}\times{\bf B}+\frac{\displaystyle 1}{\displaystyle ne}\nabla\cdot{\bf p}_{e}+\frac{\displaystyle m_{e}}{\displaystyle e}{\bf u}_{e}\cdot\nabla{\bf u}_{e}=0\quad.$ (1) Using a range of kinetic simulations, in the present paper we will discuss how the various terms of Eq. (1) become important within separate spatial areas of the IDR and EDR. Compared to similar investigations a decade or more ago hesse:1999 ; daughton:2006 ; drake:2008 , a main difference is that modern computing facilities now permit routine fully kinetic simulations at the full proton to electron mass ratio $m_{i}/m_{e}=1836$. The present analysis reveals that for simulations where this ratio approaches its natural value, the main assumption of whistler reconnection drake:2008 , that ${\bf E}+{\bf u}_{e}\times{\bf B}\simeq 0$ within the IDR, is not accurate. Instead, it is found that $|E_{\mbox{\scriptsize rec}}|\ll|\nabla\cdot{\bf p}_{e}/(ne)|$ such that ${\bf u}_{e}\times{\bf B}\simeq-\nabla\cdot{\bf p}_{e}/(ne)$ becomes the main driver of the Hall magnetic field perturbation. In fact, the familiar assumption of $u_{\mbox{\scriptsize in}}=E_{\mbox{\scriptsize rec}}/B_{\mbox{\scriptsize rec,up}}$ significantly underestimates the inflow speed of the electron fluid into the EDR, with important implications for previous theoretical work as well as the analysis of spacecraft data. Here, $E_{\mbox{\scriptsize rec}}$ is the reconnection electric field and $B_{\mbox{\scriptsize rec,up}}$ is the in-plane magnetic field just upstream of the EDR. The main topic of the paper, however, is the development of a theory to explain the terms of Eq. (1) which break the electron frozen-in condition at the very center of the EDR. The study reveals the importance of the electron anisotropy upstream of the EDR in driving the current within the EDR and cancelling an off-diagonal stress term identified in previous work kuznetsova1998kinetic ; hesse:1999 . Ultimately, the off-diagonal stress of ${\bf p}_{e}$ that is responsible for breaking the frozen-in condition is generated by $E_{\mbox{\scriptsize rec}}$ itself. Thus, the theory suggests that the electron fluid does not represent an obstacle (or bottleneck) for reconnection, which may then proceed at the rate imposed by dynamics external to the EDR stanier:2015 ; liu:2017 . The paper is organized as follows: In Section II we discuss the formation and role of electron pressure anisotropy within the IDR inflow regions, whereas Section III examines the length scales characterizing the EDR. A detailed account of the electron momentum balance at the $X$-line region is provided in Section IV, and the paper is summarized and concluded in Section V. ## II Formation of electron pressure anisotropy within the reconnection region ### II.1 Setup of kinetic simulations The 2D kinetic simulations applied in our study were implemented for anti- parallel Harris sheet reconnection using the code VPIC bowers:2009 with open boundary conditions daughton:2006 . The initial magnetic field and the density are $B_{x}=B_{0}\tanh(z/\lambda)$ and $n=n_{0}\cosh^{2}(z/\lambda)+n_{\infty}$, respectively, where $\lambda=0.5d_{i0}=0.5\sqrt{\varepsilon_{0}m_{i}c^{2}/n_{0}e^{2}}$. Other parameters are initial uniform temperatures with $T_{i\infty}/T_{e\infty}=5$, $\omega_{pe0}/\omega_{ce0}=2$, and $\sim 400$ particles per species per cell. The study incorporates 35 separate simulation runs implementing a matrix of 5 values for $m_{i}/m_{e}$ and 7 values for the normalized upstream electron pressure, $\beta_{e\infty}$. To be more specific, we use mass ratios of $m_{i}/m_{e}\in\\{100,200,400,800,1836\\}$, and the background density $n_{\infty}$ is varied so the initial upstream electron beta is $\beta_{e\infty}=2\mu_{0}n_{\infty}T_{e\infty}/B_{0}^{2}=2^{-k}$ with $k\in\\{1,2,\dots,7\\}$. In VPIC “natural” units the systems are implemented using $m_{e}=1$, $d_{e0}=1$, $c=1$, $n_{0}=1$, $B_{0}=1/2$, $T_{e\infty}=1/48\,m_{e}c^{2}$, and $n_{\infty}/n_{0}=(1+T_{i\infty}/T_{e\infty})2^{-k}$. For runs with $m_{i}/m_{e}\leq 400$ the domains were $4096\times 4096$ cells, corresponding to $50d_{i0}\times 50d_{i0}$, while runs at $m_{i}/m_{e}\geq 400$ were carried out at system sizes of $800d_{e0}\times 800d_{e0}$. Here, again, $d_{i0}$ and $d_{e0}$ are the ion and electron skin-depths, respectively, based on $n_{0}=1$. Note that $d_{e0}$ is defined with respect to the central Harris sheet density, $n_{0}$. Meanwhile, the upstream electron skin-depth becomes central to our analysis below and can be expressed as $d_{e\infty}=d_{0}/\sqrt{n_{\infty}/n_{0}}$. More details about the similarities and differences between the individual simulation runs are provided in Section III, where the length of the EDR electron jets is examined. ### II.2 Formation of electron pressure anisotropy in reconnection inflows Figure 1: VPIC simulations profiles of $B/B_{\infty}$, $T_{e\|}/T_{e\perp}$, $|{\bf u}_{e}|/v_{te}$, and $B_{y}/B_{\infty}$, obtained with $m_{i}/m_{e}=1836$ and $\beta_{e\infty}=2^{-4}=0.0625$. The arrows in b) represent the enhanced electron flow provided by ${\bf J}_{\mbox{\scriptsize extra}}=[({p_{e\parallel}}-{p_{e\perp}})/B]{\bf b}\times\mathbf{\kappa}$. In the inflow regions the electrons follow the magnetic flux-tubes in their convection toward the EDR, while the ions are unmagnetized. The ions by their inertia decouple from the motion of the magnetic field, and they dictate a near uniform plasma density within the region. In turn, the electrons respond strongly to match this uniform density and maintain quasi-neutrality. As illustrated in Fig. 1(a), the declining magnetic field strength $|{\bf B}|$ causes the widths of the magnetic flux-tubes to expand as the EDR is approached. To avoid a reduction in the electron density, field-aligned electric fields $E_{\|}$ develop egedal:2009pop , compressing the range of the parallel motion for trapped electrons. This boosts the electron density such that quasi-neutrality (i.e. $n_{e}\simeq n_{i}$) is maintained. The profiles of $E_{\|}$ in many cases trap all thermal electrons, limiting thermal heat conduction and yielding a regime that differs significantly from the standard Boltzmann regime where $T_{e}$ is constant. Rather, the reduction in $B$ and perpendicular expansion of the flux-tubes results in $T_{e\perp}$ cooling, whereas the parallel compression by $E_{\|}$ to maintain quasi-neutrality yields $T_{e\|}$ heating le:2009 ; le:2010grl ; egedal:2013 ; wetherton:2019 . Generally it is found that runs with $\beta_{e\infty}^{2}m_{i}/m_{e}\gtrsim 1$ are characterized by a regime of double adiabatic electron dynamics, where at the edge of the EDR the main electron pressure components are well described by the CGL-limit chew:1956 where ${p_{e\parallel}}\propto n^{3}/B^{2}$ and ${p_{e\perp}}\propto nB$. For $\beta_{e\infty}^{2}m_{i}/m_{e}\lesssim 1$ electron holes and double layers yield even stronger ${p_{e\parallel}}$ heating egedal:2013 ; egedal:2015 . The effects of the anisotropic heating are evident in the profile in Fig. 1(b), where $T_{e\|}/T_{e\perp}\simeq 4$ is observed just upstream of the EDR, and the marginal electron firehose condition is approached, ${p_{e\parallel}}-{p_{e\perp}}\simeq B^{2}/\mu_{0}$. Within the EDR the adiabatic invariance of the electron magnetic moments, $\mu=m{v_{\perp}}^{2}/(2B)$, break, leading to isotropization and pitch angle mixing. As a result, the electron pressure is approximately isotropic in the reconnection exhaust le:2013 ; egedal:2016b . The physical mechanisms which underpin the CGL-scalings can be understood through relatively simple arguments applicable to the electrons just upstream of the EDR, where the described trapping and parallel compression dynamics are sufficiently strong that they dominate the properties of the electron fluid. The area of a given flux-tube scales as $1/B$, and the total number of particles in the flux-tube section (of length $l$) therefore scales as $N\propto nl/B$. Particle conservation then requires that $l\propto B/n$. Next, similar to Fermi heating, the conservation of the parallel action for each trapped electron requires that $lv_{\parallel}$ is constant such that $v_{\parallel}\propto n/B$, yielding $T_{e\|}\propto v_{\parallel}^{2}\propto n^{2}/B^{2}$. Furthermore, because $\mu=m{v_{\perp}}^{2}/(2B)$ is conserved it is clear that ${v_{\perp}}^{2}\propto B$, such that for this trapped electron population $T_{e\perp}\propto B$. It then follows that ${p_{e\parallel}}=nT_{e\|}\propto n^{3}/B^{2}$ and ${p_{e\perp}}=nT_{e\perp}\propto nB$ coinciding with the CGL-scaling laws chew:1956 . A more accurate and rigorous model is provided in le:2009 ; egedal:2013 taking into account that not all electrons become trapped. For the present paper, however, it is sufficient to assume that within the inflow regions the electron pressures approximately follow the aforementioned CGL- scaling laws. ### II.3 The reconnection rate from the perspective of the electrons In Section III we document how the size of the EDR normalized by $d_{e\infty}$ remains approximately constant for varying values of $m_{i}/m_{e}$. Meanwhile, the size of the IDR scales with $d_{i\infty}=d_{e\infty}\sqrt{m_{i}/m_{e}}$, and from the perspective of the electrons the size of the ion diffusion region thus increases by a factor of $\sqrt{m_{i}/m_{e}}$. Likewise, relative to the time scale of the electron motion, the increasing inertia of the ions will slow the rate of reconnection. As a dimensionless measure of the reconnection electric field relevant to the electron orbit dynamics we introduce $\hat{E}_{\mbox{\scriptsize rec}}=ed_{e\infty}E_{\mbox{\scriptsize rec}}/T_{e\infty}$. This quantity represents the temperature-normalized energy gain an electron will acquire when traveling $d_{e\infty}$ in the direction of $E_{\mbox{\scriptsize rec}}$. Consistent with external MHD scale constraints for system sizes larger than $10d_{i}$ liu:2017 , in the present and previous kinetic simulations the absolute reconnection rate obeys $E_{\mbox{\scriptsize rec}}\simeq 0.1v_{A}B$, where the relevant Alfvén speed and $B$ are evaluated $1d_{i}$ upstream of the EDR shay:2004 . Corresponding to the shaded region in Fig. 2(a), it then follows that $\hat{E}_{\mbox{\scriptsize rec}}\propto\sqrt{m_{e}/(m_{i}\beta_{e\infty}^{2})}$, where reduced values of $m_{i}/m_{e}$ and low $\beta_{e\infty}$ impose the largest values of $\hat{E}_{\mbox{\scriptsize rec}}$. In Fig. 2(b) we show the profiles of $T_{e\|}/T_{e\infty}$ and $T_{e\perp}/T_{e\infty}$ corresponding to values in the various runs recorded just upstream of the electron diffusion region. Consistent with the discussion in Section II.B, for simulations within the double adiabatic regime ($\beta_{e\infty}^{2}m_{i}/m_{e}\gtrsim 1$) marked by the full lines, we observe that both of these profiles are largely independent of $m_{i}/m_{e}$. The normalized electron pressure anisotropy $(p_{e\|}-p_{e\perp})/(n_{\infty}T_{e\infty})$ is then also independent of $m_{i}/m_{e}$, and given the scaling of $\hat{E}_{\mbox{\scriptsize rec}}\propto\sqrt{m_{e}/m_{i}}$ the forces associated with the pressure anisotropy become most significant when compared to $e\hat{E}_{\mbox{\scriptsize rec}}$ at large value of $m_{i}/m_{e}$. In fact, as will be demonstrated below, the thermal forces of the electron pressure anisotropy strongly dominate the force balance of the electrons for $m_{i}/m_{e}=1836$, rendering the electron dynamics of the IDR and EDR significantly different when compared to reduced fluid models invoking isotropic electron pressure. Figure 2: Obtained from a matrix of 35 kinetic simulations, profiles as functions of $\log_{10}(\beta_{e\infty})$ are shown for a range of parameters as marked in the panels. As indicated in b) the line-type denotes the value of $m_{i}/m_{e}$. In a) $\hat{E}_{\mbox{\scriptsize rec}}$ is normalized by $T_{e\infty}/(ed_{e\infty})$, and the magenta shaded region is the expected range based on $E_{\mbox{\scriptsize rec}}=\alpha v_{A}B_{\mbox{\scriptsize rec}}$, with $\alpha=0.1$ and $100\leq m_{i}/m_{e}\leq 1836$ (see text). The wider yellow shaded region correspond to $0.08\leq\alpha\leq 0.12$. In b) the regions shaded in magenta are obtained from the CGL-limit of the Lê-2009 Equations of State le:2009 . The dashed lines mark runs in the regime $\beta_{e\infty}<\sqrt{m_{i}/m_{e}}$ where additional $T_{e\|}$ heating typically is observed egedal:2015 ; le:2016twostage . ### II.4 Dominant role of $p_{e\|}\gg p_{e\perp}$ for large $m_{i}/m_{e}$ Figure 3: Considering the same simulation as in Fig. 1 ($m_{i}/m_{e}=1836$ and $\beta_{e\infty}=2^{-4}$), color contours are presented for the LHS and RHS of Eq. (3), respectively. The EDR is outlined by the cyan rectangle. For the extended regions where $({\bf J}\times{\bf B})_{y}/(n_{e0}E_{y0})\gtrsim 4$ the electron fluid is moving significantly faster than the magnetic field. This motion is driven by electron pressure anisotropy such that $|{\bf J}\times{\bf B}|\simeq|\nabla\cdot{\bf p}_{e}|\gg en_{0}E_{\mbox{\scriptsize rec}}$. In the region outside the EDR where the electrons are well magnetized, the electron pressure anisotropy drives additional currents, ${\bf J}_{\mbox{\scriptsize extra}}=[({p_{e\parallel}}-{p_{e\perp}})/B]{\bf b}\times\mathbf{\kappa}$, which are not observed for isotropic pressure. Here $\mathbf{\kappa}=({\bf b}\cdot\nabla){\bf b}$ is the curvature vector which becomes large where the magnetic field lines have strong curvature. In Fig. 1(b) the arrows are proportional to $-{\bf J}_{\mbox{\scriptsize extra}}$ corresponding to a flow boosting the perpendicular drift of electrons into the EDR. To quantify the relative importance of the flow by ${\bf J}_{\mbox{\scriptsize extra}}$, in Fig. 3 we provide additional profiles for the simulation of Fig. 1, here focusing on the force balance in the $y$-direction in the vicinity of the EDR. Within the IDR we generally have $({\bf u}_{e}\times{\bf B})_{y}\gg({\bf u}_{i}\times{\bf B})_{y}$ and we thus introduce the approximation that $(-en_{e}{\bf u}_{e}\times{\bf B})_{y}\simeq({\bf J}\times{\bf B})_{y}$. Furthermore, it turns out that generally $\left|({\bf J}\times{\bf B})_{y}\right|\gg\left|(enm_{e}{\bf u}_{e}\cdot\nabla{\bf u}_{e})_{y}\right|$, and the $y$-component of the electron inertia can for the present analysis be neglected. With these approximation the $y$-component of the generalized Ohm’s law in Eq. (1) then becomes $-en_{e}E_{y}+({\bf J}\times{\bf B})_{y}=(\nabla\cdot{\bf p}_{e})_{y}\quad.$ (2) For the time point considered in Fig. 3 the reconnection geometry and rate is approximately steady. Furthermore, the density profile in this inner region is approximately uniform at a density we denote as $n_{e0}$. It then follows that $n_{e}E_{y}\simeq n_{e0}E_{y0}$, where $E_{y0}$ is the value of $E_{y}$ observed at the $X$-line. After dividing Eq. (2) by $n_{e0}E_{y0}$ we obtain: $\frac{\displaystyle({\bf J}\times{\bf B})_{y}}{\displaystyle n_{e0}E_{y0}}\simeq 1+\frac{\displaystyle(\nabla\cdot{\bf p}_{e})_{y}}{\displaystyle n_{e0}E_{y0}}\quad.$ (3) The LHS (left hand side) of Eq. (3) is shown in Fig. 3(a), while the RHS is shown in Fig. 3(b). The fact that these profiles are near-identical validates the approximations introduced above. From Fig. 3(a) it is apparent that $({\bf J}\times{\bf B})_{y}$ is up to 5 times larger that $n_{e0}E_{y0}$ in a significant fraction of the cross- section. Thus, rather than the $en_{e}E_{y}\simeq({\bf J}\times{\bf B})_{y}$ relation applied in whistler-wave reconnection models, we see that $|en_{e}E_{y}|\ll|({\bf J}\times{\bf B})_{y}|\simeq|(\nabla\cdot{\bf p})_{y}|$ throughout the majority of the ion and electron diffusion regions. This strongly enhanced inflow is provided by $p_{e\|}\gg p_{e\perp}$ through the drifts associated with ${\bf J}_{\mbox{\scriptsize extra}}$. In other words, while $E_{y}$ by Faraday’s law is fundamental for moving the magnetic flux across the reconnection region, Fig. 3 reveals how the $(\nabla\cdot{\bf p}_{e})_{y}$-term is responsible for driving the electrons into the EDR and through the elongated jets of the EDR at high rates, which for the present run are up to 5 times the speed of the magnetic field line motion. The electric field in the local rest frame of the electron fluid is given by ${\bf E}^{\prime}={\bf E}+{\bf v}_{e}\times{\bf B}$. In resistive Hall-fluid models with isotropic pressure the energization measure $-en_{e}{\bf v}_{e}\cdot{\bf E}^{\prime}$ is generally positive, such that magnetic energy is dissipated and energizes the electron fluid. This observation is altered by the presence of the strong electron pressure anisotropy. For example, within the regions in Fig. 3(a) where $({\bf J}\times{\bf B})_{y}/(en_{e}E_{y0})\simeq 4$ it follows that $E_{y}^{\prime}=E_{y}+({\bf v}_{e}\times{\bf B})_{y}\simeq-3E_{y}$. Thus, in these regions where the electrons “run ahead” of the magnetic field lines $-en{\bf v}_{e}\cdot{\bf E}^{\prime}$ may become negative, such that energy from the electrons is given back to the magnetic field. Such regions, known as generator regions, have been directly observed by the MMS mission payne:2021 . Furthermore, given the strong electron flow generated by $\nabla\cdot{\bf p}$ the aspects ratio of the electron diffusion region (i.e. the aspect ratio of the cyan boxes in Fig. 3) is not necessarily an accurate measure of the normalized reconnection rate, and may be important to newly develop methods for analysis of spacecraft data burch:2020 . ### II.5 Force balance along the length of the EDR and associated scaling laws for the electron pressure anisotropy a-c) Figure 4: Profiles of $u_{ex}/v_{te}$ for the full simulation domains for three separate runs at $\beta_{e\infty}=2^{-4}$ and $m_{i}/m_{e}=800$. The run in a) has the “standard” size described in Section II.A, whereas the runs in b,c) are for systems enhanced by factors $2\times$ and $4\times$, respectively. Zoomed-in views of the EDRs are shown in d-f), whereas g-i) show cuts of $u_{ex}$ and $u_{ey}$ for $z=0$, from which the length scale $l_{u}$ of current layer rotation is determined. Our analysis relies strongly on previous results regarding the overall force balance of the EDR le:2010grl ; egedal:2013 ; montag:2020 . For the convenience of the reader we here provide a short summary of how the upstream electron pressure anisotropy impacts the structure of the EDR. Also, in Appendix A we provide additional mathematical considerations to demonstrate how the upstream electron pressure anisotropy drives the current jets within the EDR. Consistent with Fig. 3, in Ref. le:2010grl it was demonstrated that the two dominant terms of Eq. (1) along the EDR are ${\bf u}_{e}\times{\bf B}$ and $(\nabla\cdot{\bf p}_{e})/ne$ le:2010grl . Expressed in terms of the Maxwell stress tensor, ${\bf T}=({\bf B}{\bf B}-B^{2}{\bf I})/(2\mu_{0})$, we have $-en{\bf u}_{e}\times{\bf B}\simeq{\bf J}\times{\bf B}=\nabla\cdot{\bf T}$, and momentum balance then requires that $\nabla\cdot({\bf p}_{e}-{\bf T})\simeq 0$. Considering the area encircled by the red square in Fig. 1(c), it can then be shown by the divergence theorem le:2010grl ; montag:2020 that at the edge of the EDR the electron firehose condition must be approximately satisfied, ${p_{e\parallel}^{\mbox{\scriptsize edge}}}-{p_{e\perp}^{\mbox{\scriptsize edge}}}\simeq B_{H}^{2}/\mu_{0}\quad,$ (4) where $B_{H}$ represents the magnetic field strength just upstream of the EDR (this field rotates into the Hall magnetic field explaining the subscript $H$). For example, for the red square in Fig. 1(c) the magnetic tension acting on the sides of the fluid element (corresponding to the integral over the element of the $-en{\bf u}_{e}\times{\bf B}$ force) is offset by pressure anisotropy. Again, this is discussed in more detail in Appendix A. We can derive important scaling laws for the level of pressure anisotropy for locations just upstream of the EDR. Using that ${p_{e\parallel}^{\mbox{\scriptsize edge}}}\gg{p_{e\perp}^{\mbox{\scriptsize edge}}}$ with the CGL-limits the maximal pressure differential ${\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}\equiv\mbox{max}({p_{e\parallel}}-{p_{e\perp}})={p_{e\parallel}^{\mbox{\scriptsize edge}}}-{p_{e\perp}^{\mbox{\scriptsize edge}}}\quad,$ (5) adheres to the asymptomatic scaling ${\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}\simeq{p_{e\parallel}^{\mbox{\scriptsize edge}}}\simeq(\pi\tilde{n}^{3}/(6\tilde{B}_{H}^{2}))\beta_{e\infty}B_{\infty}^{2}/(2\mu_{0})$. The factor $\pi/6$ is part of the Lê 2009 Equations of State le:2009 . The firehose condition in Eq. (4), representing the dominant momentum balance condition of the EDR, may then be expressed as $\frac{\displaystyle B_{H}}{\displaystyle B_{\infty}}=\frac{\displaystyle\mu_{0}K}{\displaystyle 2B_{\infty}}=\left(\frac{\displaystyle\pi\tilde{n}^{3}\beta_{e\infty}}{\displaystyle 12}\right)^{1/4}\quad.$ (6) Here $K$ is the total current integrated across the layer. $\tilde{n}\simeq 1$ is the local electron density normalized by the upstream value $n_{\infty}$ (and similar for $\tilde{B}_{H}$ used above) le:2010grl . Eq. (6) is significant as it demonstrates how the current per unit length of the EDR, $K$, is mainly controlled by the external electron pressure anisotropy. The above result is in contrast to predictions of two-fluid models with isotropic pressure where ${\bf u}_{e}\times{\bf B}$ is balanced by resistive terms, ${\bf E}$, and/or $(m_{e}/e){\bf u}_{e}\cdot\nabla{\bf u}_{e}$. It should also be noted that the half-width $\delta\simeq 2d_{e\infty}$ of the EDR region current layer is set by the meandering orbit width roytershteyn:2013 . From Ampère’s law it then follows that $B_{H}=\mu_{0}en_{e}u_{e}\delta\simeq 2\mu_{0}en_{e}u_{e}d_{e\infty}$, such that $u_{e}=B_{H}/(2en_{e}u_{e}d_{e\infty})\simeq B_{H}/(2\sqrt{\mu_{0}n_{e}m_{e}})=v_{Ae}/2$. Thus, $u_{e}$ must approach the corresponding value of $\sim v_{Ae}/2$, simply because of the narrow width of the current layer $\delta\simeq 2d_{e\infty}$. This analysis is also applicable to the Harris sheet type current layer and the condition $u_{e}=v_{Ae}/2$ is therefore reflective of the underlying meandering orbit dynamics setting the narrow width of the electron flow region roytershteyn:2013 . The EDR current layer rotates in the $xy$-plane hesse:2008 ; le:2010pop , and yields the $B_{y}$ perturbations shown in Fig. 1(d) hesse:2008 ; le:2010grl . The above results have also been applied in deriving scaling laws for the absolute electron heating within the reconnection inflow and EDR le:2016twostage . ## III Length scales of the EDR In resistive fluid models the aspect ratio of the EDR controls the rate of magnetic reconnection parker:1957 . This follows from the condition of mass continuity and the assumption that the electron fluid is frozen-in to the magnetic field (i.e. ${\bf E}+{\bf v}_{e}\times{\bf B}=0$) outside the EDR. However, as discussed above and documented in Fig. 3, in our kinetic runs at large values of $m_{i}/m_{e}$ the electrons decouple from the magnetic field within the IDR and “run ahead” of the magnetic field into the EDR. This simple observation has important implications, as it follows that for the kinetic limit with $p_{e\|}\gg p_{e\perp}$ upstream of the EDR, the rate of reconnection is then no longer tied to the aspect ratio of the EDR. Below we introduce the length $l_{u}$, which is a measure of the spatial scale at which the out-of-plane EDR electron current at the $X$-line rotates into the exhaust direction. Our analysis show that this length scale is shortened by the enhanced inflow of ${\bf J}_{\mbox{\scriptsize extra}}$ and we observe that $l_{u}\simeq 20\pm 7d_{e\infty}$ for all the simulation parameters considered with $\beta_{e\infty}\leq 2^{-3}$ ### III.1 Numerical evidence that $l_{u}$ does not scale with simulation system size In Ref. daughton:2006 , which employed relatively low mass ratio $m_{i}/m_{e}=25$, it was observed that the length of the EDR increases with the size of the numerical simulation domain, and evidence was presented that the long electron current layers of the EDR could act as a bottleneck for reconnection. These results do not to apply to our present simulations where $m_{i}/m_{e}$ approaches its natural value of 1836. For example, in Fig. 4(a) profiles of $u_{ex}$ are shown for a simulation at $m_{i}/m_{e}=800$ and $\beta_{e\infty}=2^{-4}$ for the “standard” setup described in Section II.A, whereas Figs. 4(b,c) apply domains factors of $2\times$ and $4\times$ larger, respectively. Zoomed-in views of the EDRs for the three runs are shown in Figs. 4(d-f), where we notice how the $2\times$ and $4\times$ runs have similarly sized EDRs. This indicates that for simulations at large $m_{i}/m_{e}$ implemented in large domains, the length of the EDR becomes independent of the system size. For the analytical analysis of the EDR the following length scale $l_{u}$ becomes essential to the presented theory for the force balance of the electron fluid. Here $l_{u}\equiv\left|\frac{u_{ey}}{du_{ex}/dx}\right|_{\mbox{\scriptsize X-line}},$ (7) represents a measure of the spatial distance that characterizes the rotation of the EDR electron current in the $xy$-plane . Figs. 4(g-i) provide the geometric interpretation of this quantity, which is observed to be on the order of $20d_{e}$ for the three runs. Below we show how $l_{u}$ is significantly shorter than the full length $l_{\mbox{\scriptsize jet}}$ of the EDR for many runs in our study, consistent with previously observed two-scale EDR structures karimabadi:2007 ; shay:2007 . ### III.2 Numerical results for $l_{u}$ over the full $(\beta_{e\infty},m_{i}/m_{e})$–matrix of kinetic runs Figure 5: Color contours of $u_{ex}$ for the matrix of numerical simulations applied in the study, $m_{i}/m_{e}\in\\{100,200,400,800,1836\\}$ and $\beta_{e\infty}=2^{-k}$ with $k\in\\{1,2,\dots,7\\}$. When falling within the domains considered, the ends of the jets are marked by the red stars corresponding to $|u_{ex}|$ dropping to 60% of its peak value. The similarly defined width of the $u_{ey}$ out-of-plane electron drift is marked in each panel at $x=0$. Figure 6: Recorded values of $l_{u}$ and $l_{\mbox{\scriptsize jet}}$. In b) the dashed lines indicate runs in the non- adiabatic regime with enhanced ${p_{e\parallel}}$-heating observed for $\beta_{e\infty}^{2}m_{i}/m_{e}\lesssim 1$. Evidently, this regime is also characterized by an enhancement in the length of the EDRs (while $l_{u}$ remains near constant). Fig. 5 provides zoomed-in views of the EDRs recorded in the full set of runs applied in our study. As is typical for kinetic simulations of reconnection based on the Harris sheet geometry, after fast reconnection has commenced the reconnection rate initially increases and then reduces to a near steady state. Each of the profiles considered here correspond to times $t\simeq 70\Omega_{ci}^{-1}$ during these later intervals of mostly steady state reconnection geometries. For $\beta_{e\infty}\geq 2^{-2}$ elongated EDRs are mostly observed $l_{\mbox{\scriptsize jet}}>100d_{e\infty}$. Here the out-of- plane electron flows are largely provided by the diamagnetic drifts similar to those of the Harris sheet. Meanwhile, for $\beta_{e\infty}\leq 2^{-3}$, the force balance condition of the EDR as expressed in Eq. (6) requires current densities boosted beyond those provided by the diamagnetic effect, and the EDRs acquire a shorter length. The marked ends of the jets are defined by the locations where $|u_{ex}|$ has fallen to 60% of its respective peak value. For the adiabatic regime ($\beta_{e\infty}^{2}m_{i}/m_{e}\gg 1$), the lengths of the electron diffusion region are consistently observed to be $l_{\mbox{\scriptsize jet}}\simeq 20d_{e\infty}$. For the regime $\beta_{e\infty}^{2}m_{i}/m_{e}\ll 1$ of enhanced ${p_{e\parallel}}$ energization egedal:2015 , a sharp transition occurs where the EDRs are characterized by longer outflow jets. In Fig. 6 we show the values of $l_{u}$ and $l_{\mbox{\scriptsize jet}}$ for all the runs applied in the study. The runs with $\beta_{e\infty}\leq 2^{-3}$ (most relevant to the analysis in Section IV) are all characterized by $l_{u}\simeq 18d_{e\infty}$. In Fig. 6(b) the dashed lines represent runs in the aforementioned regime $\beta_{e\infty}^{2}m_{i}/m_{e}>1$ of enhanced ${p_{e\parallel}}$ heating, and it is evident that this regime is characterized by larger values of $l_{\mbox{\scriptsize jet}}$. ### III.3 How continuity of electrons in the EDR sets the $l_{u}$ length scale Directly related to ${\bf J}_{\mbox{\scriptsize extra}}$, we find that $l_{u}$ is shortened by the enhancement of the electron flow into the EDR. Let $u_{ey0}$ be the flow in the $y$-direction at the $X$-line; given the definition of $l_{u}$ the flow in the $x$-direction is then $u_{ex}=u_{ey0}x/l_{u}$. For a uniform inflow velocity $u_{ez}$ mass continuity requires that $nu_{ez}x=nu_{ex}\delta=nu_{ey0}x\delta/l_{u}$ such that $l_{u}=\frac{\displaystyle u_{ey0}\delta}{\displaystyle u_{ez}}=\frac{\displaystyle B_{H}}{\displaystyle en\mu_{0}u_{ez}}\quad.$ Meanwhile, the strong parallel streaming of electrons in toward the EDR causes the inflow speed of the electrons to increase. Empirically we find $u_{ez}\simeq 4E_{y}/B_{H}\quad.$ This enhancement of the inflow speed above $E_{y}/B_{xz}$ is caused by the inflow pressure anisotropy through the term ${{\bf J}_{\perp}}_{\mbox{\scriptsize extra}}$, and is consistent with the profiles in Figs. 10(a,b). The reconnection electric field, $E_{y}$ follows the scaling laws of Ref. shay:2004 where $E_{y}\simeq 0.1v_{A}B_{1di}=\frac{\displaystyle 0.1B_{1di}^{2}}{\displaystyle\sqrt{\mu_{0}nm_{i}}}\quad.$ It then follows that $l_{u}=\frac{\displaystyle B_{H}^{2}}{\displaystyle 0.4B_{1di}^{2}}\sqrt{\frac{\displaystyle m_{i}}{\displaystyle m_{e}}}d_{e}\quad.$ Empirically, the magnetic field 1$d_{i}$ upstream of the EDR is approximately $B_{1di}\simeq B_{\infty}/2$. Furthermore, due to non-adiabatic effects $B_{H}$ depends on $m_{i}/m_{e}$ (see Fig. 9(b)), and the scaling is consistent with $B_{H}^{2}\sqrt{m_{i}/m_{e}}/(0.4B_{1di}^{2})\simeq 15$ for all the runs. In turn, this is in reasonable agreement with the observation that $l_{u}\simeq 18d_{e}$ in Fig. 6(a). ## IV Breaking the frozen-in law at the reconnection $X$-line The major force terms of the EDR are ${\bf J}\times{\bf B}$ and $\nabla\cdot{\bf p}_{e}$ which largely balance. As discussed above this force balance requirement can ultimately be expressed through the conditions in Eqs. (4) or (6). However, as seen in Fig. 3(a), in a small region ($4d_{e}\times 2d_{e}$) centered on the $X$-line the $({\bf J}\times{\bf B})_{y}$ force vanishes, such that the more detailed force balance right at the $X$-line then requires that the RHS of Eq. (3) also vanishes (consistent with Fig. 3(b)). Thus, for the considered anti-parallel and symmetric configurations and within this limited region around the $X$-line, the force balance constraint of Eq. (1) only involves the off-diagonal stress in the electron pressure tensor $E_{\mbox{\scriptsize rec}}=-\frac{1}{en}\left(\frac{\displaystyle\partial p_{exy}}{\displaystyle\partial x}+\frac{\displaystyle\partial p_{eyz}}{\displaystyle\partial z}\right)\quad.$ (8) Again, here $E_{\mbox{\scriptsize rec}}$ is the $E_{y}$ electric field along the $X$-line (that is aligned with the $y$ axis and runs through $(x,z)=(0,0)$). By Faraday’s law $E_{\mbox{\scriptsize rec}}$ then represents the rate at which magnetic flux cross the $X$-line. Below we will derive a new model for the terms in Eq. (8), providing a theory consistent with fast reconnection controlled by larger scale dynamics external to the EDR. ### IV.1 Meandering orbit motion of electrons within the EDR, setting the striated structure of the electron distribution function To elucidate how the upstream electron pressure anisotropy drives the current of the EDR, in Fig. 7 we consider the trajectories of electrons with field- aligned velocities ${\bf v}=3.3v_{te}{\bf B}/B$ injected with $z=-4d_{e\infty}$ at various values of $x<0$ just upstream of the EDR. These trajectories are representative because the upstream distributions (like in Figs. 8(a,d)) with ${p_{e\parallel}^{\mbox{\scriptsize edge}}}\gg{p_{e\perp}^{\mbox{\scriptsize edge}}}$ “feed” electrons to the EDR with $|v_{\parallel}|\gg{v_{\perp}}$ ng:2011 . In Figs. 7(a,b) the parts of the trajectories with $z<0$ ($z>0$) are represented by the blue (red) full lines. For each transit across $z=0$ in Fig. 7(b) additional orbits are initialized and shown by the dashed lines. This visualizes how the average force of $-e{\bf v}\times B_{z}{\bf e}_{z}$ causes the meandering Speiser-type motion speiser:1965 to diverge away from the $X$-line. The EDR current is caused in part by the $E\times B$-drift of the $E_{z}$ field displayed in Fig. 7(a). The large values of $E_{z}$ are consistent with momentum balance with the strong gradients of $p_{\|}$ at the interface of the EDR. The direction of the parallel thermal streaming is sensitive to the $B_{y}$ Hall magnetic field, and from Fig. 7(b) it is evident how this sets the angle in the $xy$-plane at which the electrons are injected into the EDR region. In Figs. 7(c-e) the colored and encircled triangles represent the particle velocities corresponding to the similarly colored velocity vectors in Fig. 7(b). For example, the magenta vector at $x=0$ in Fig. 7(b) pointing mostly in the $x$-direction, yields the velocity of an electron reaching the $X$-line directly from the upstream region, and the corresponding encircled magenta triangle in Fig. 7(d) has $|v_{x}|\gg|v_{y}|$. The $X$-line may also be reached through multiple meandering motions within the layer, steadily directing the velocity into the $-y$-direction. As shown in Figs. 7(c-e), repeating this procedure for four more initial velocities ${\bf v}=v_{\parallel}{\bf B}/B$ within the interval $0.1\leq v_{\parallel}/v_{te}\leq 3.3$ yields the lines of colored triangles, corresponding to the centers of the striated structures of the EDR electron distributions first described in Ref. ng:2011 . In Figs. 7(c-e), the similar points marked by green circles are the result of injecting electrons along field lines from the opposing sides of the EDR. Also, going from $x/d_{e\infty}=-3$ to $x/d_{e\infty}=3$, we note how the displayed structures rotates in the $v_{x}v_{y}$-plane (and similarly in Fig. 8), corresponding to the rotation of the EDR current layer mentioned above ng:2011 ; shuster:2015 . Figure 7: a,b) Trajectories of electrons injected along field lines into the EDR with $v_{\parallel}=3.3v_{te}$. Three points, $(x,z/d_{e\infty})=(-3,0);(0,0);(3,0)$, are selected in a) indicating the spatial locations where the velocity arrows in b) are evaluated. As indicated, the panels c-e) also correspond to these locations. The angles indicated by arrows in b) at which the trajectories reach the selected points, determine the location of the encircled triangles in c), d), and e), respectively. Similar trajectories are obtained by using a range of initial $v_{\parallel}$ values and result in the striated structures marked by the symbols in c), characteristic of the EDR distributions ng:2011 and observed in Fig. 8. ### IV.2 Origin for the off-diagonal $\partial p_{exy}/\partial x$-stress at the reconnection $X$-line To understand how the upstream pressure anisotropy influences $p_{exy}$ we consider the electron distributions in Fig. 8. At the edge of the EDR we have $p_{eyy}\simeq p_{ezz}$, and given Eq. (5), $p_{eyy}-p_{exx}=-{\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}$. Comparing the distributions in Figs. 8(a,b)) it is clear that only a fraction of this upstream anisotropy is carried by the orbit motion to the $X$-line. Empirically at the $X$-line we find that $\left.\left(p_{eyy}-p_{exx}\right)\right|_{\mbox{\scriptsize X-line}}=-\dfrac{3}{4}{\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}+\Delta p_{eyy,E}\quad.$ (9) Here the additional term, $\Delta p_{eyy,E}$, corresponds to an increase in $p_{eyy}$ due to direct heating of electrons by the reconnection electric field. From Fig. 7 it is clear that the “tips” of the triangular shaped distributions are composed of electrons meandering multiple times while streaming mostly in the $-y$-direction, energized most significantly by $E_{\mbox{\scriptsize rec}}$. The more pronounced triangular shape in Fig. 8(e) compared to Fig. 8(b) is consistent with stronger $\Delta p_{eyy,E}$ heating, as the value of $E_{\mbox{\scriptsize rec}}$ for $m_{i}/m_{e}=400$ is about twice as large compared to that observed in the $m_{i}/m_{e}=1836$ run. As detailed with Figs. 7 and 8, within the EDR the $xy$-projections of the electron distributions rotate approximately as a “solid-body” such that for small $x/l_{u}$ $p_{exy}=\frac{x}{l_{u}}\left.\left(p_{eyy}-p_{exx}\right)\right|_{\mbox{\scriptsize X-line}}\quad.$ (10) Combining Eqs. (9) and (10) we then obtain $\frac{\displaystyle\partial p_{exy}}{\displaystyle\partial x}=\frac{1}{l_{u}}\left(-\dfrac{3}{4}{\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}+\Delta p_{eyy,E}\right)\quad.$ (11) This equation represents the major new result of the present paper. Below we will find that the term proportional to ${\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}$ cancels out the contributions to $E_{\mbox{\scriptsize rec}}$ provided by $\partial p_{eyz}/\partial y$, leaving only the second term in Eq. (11) for balancing $E_{\mbox{\scriptsize rec}}$. Figure 8: Projections, $f_{xy}(v_{x},v_{y})=\int f({\bf v})dv_{z}$ of EDR electron distributions, recorded in simulations for $m_{i}/m_{e}=1836$ in a-c) and $m_{i}/m_{e}=400$ in d-f). The differences between the profiles are consistent with enhanced $\Delta p_{eyy,E}$ for $m_{i}/m_{e}=400$, caused by $\hat{E}_{\mbox{\scriptsize rec,400}}\simeq 2\hat{E}_{\mbox{\scriptsize rec,1836}}$. ### IV.3 Cancellation of the Hesse-$\partial p_{eyz}/\partial z$-stress Refs. kuznetsova1998kinetic ; hesse:1999 provide approximations for the off- diagonal pressure tensor derivatives ($\partial p_{exy}^{\mbox{\scriptsize Hesse}}/\partial x$ and $\partial p_{eyz}^{\mbox{\scriptsize Hesse}}/\partial z$). $\frac{\displaystyle\partial p_{exy}^{\mbox{\scriptsize Hesse}}}{\displaystyle\partial x}=\frac{\displaystyle\partial p_{eyz}^{\mbox{\scriptsize Hesse}}}{\displaystyle\partial z}=n\sqrt{\frac{\displaystyle m_{e}T_{e}}{\displaystyle 2}}\,\frac{\displaystyle\partial u_{ex}}{\displaystyle\partial x}\quad.$ These were derived through an assumption of isotropic upstream electron pressure. As such, given that upstream of the EDR $p_{exx}-p_{eyy}=\Delta p_{\|,\perp}$ is large, the $\partial p_{exy}^{\mbox{\scriptsize Hesse}}/\partial x$-approximation becomes inaccurate and is not consistent with the simulation data presented below. In contrast, the conditions required for the accuracy of $\partial p_{eyz}^{\mbox{\scriptsize Hesse}}/\partial z$ are well satisfied, as the upstream pressure in the $yz$-plane is isotropic, $p_{eyy}\simeq p_{ezz}\simeq{p_{e\perp}^{\mbox{\scriptsize edge}}}$. With $\partial u_{ex}/\partial x=u_{e}/l_{u}$, and following Refs. kuznetsova1998kinetic ; hesse:1999 that the width of the EDR is $\rho_{e}=m{v_{\perp}}/(eB_{H})$ such that from Ampère’s law $u_{e}=B_{H}/(\mu_{0}n\rho_{e})=B_{H}^{2}/(\mu_{0}nm{v_{\perp}})$, it then follows that $\frac{\displaystyle\partial p_{eyz}^{\mbox{\scriptsize Hesse}}}{\displaystyle\partial z}=\sqrt{\frac{m_{e}T_{e\perp}}{2}}\frac{B_{H}^{2}}{\mu_{0}m_{e}{v_{\perp}}l_{u}}=\frac{B_{H}^{2}}{2\mu_{0}l_{u}}=\frac{5}{7}\frac{{\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}}{l_{u}}\,\,.$ (12) As illustrated by the black contours in Figs. 1(c), for the present configurations the momentum balance in Eq. (6) corresponds to the condition $0.7B_{H}^{2}/\mu_{0}\simeq{\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}$, explaining the last equality above. It is now interesting to observe that both in Eqs. (11) and (12) the term ${\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}/l_{u}$ occurs with similar prefactors ($5/7$ vs. $3/4$), but opposite signs. Thus, to within the accuracy of their derivations, when calculating the sum of $\partial p_{exy}/\partial x$ and $\partial p_{eyz}^{\mbox{\scriptsize Hesse}}/\partial z$ the stress imposed on the layer by the rotation of the upstream pressure anisotropy cancels the $\partial p_{eyz}^{\mbox{\scriptsize Hesse}}/\partial z$-stress term. ### IV.4 The term breaking the electron frozen in law Given the described cancellation of terms, the theory suggests that the electron dynamics of the EDR do not represent a bottleneck for reconnection. In fact, inserting Eqs. (11) and (12) into Eq. (8) only one term remains for balancing the reconnection electric field $E_{\mbox{\scriptsize rec}}\simeq-\frac{1}{enl_{u}}\Delta p_{eyy,E}\quad.$ (13) Separate from the force balance constraint expressed in Eq. (13) we can obtain a second independent estimate for $\Delta p_{eyy,E}$: In Fig. 7 we observe that electrons in the $X$-line region typically travel a distance between $0$ and 2$l_{u}$ in the $y$-direction such that $\Delta p_{eyy,E}\simeq- enl_{u}E_{\mbox{\scriptsize rec}}$. Applying this in Eq. (13) we obtain the seemingly mundane expression that $E_{\mbox{\scriptsize rec}}\simeq E_{\mbox{\scriptsize rec}}$. Nevertheless, this is an important result, as it shows that the present geometry can accommodate any reconnection rate imposed at larger scales onto the EDR region. Consistent with the behavior of $\hat{E}_{\mbox{\scriptsize rec}}$ in Fig. 9(c), it confirms previous numerical results and theoretical conjectures that, with the formation of electron scale current layers, the electron dynamics can readily accommodate the rates of reconnection imposed by the plasma behavior at larger scales stanier:2015 ; liu:2017 . We also note that because $\hat{E}_{\mbox{\scriptsize rec}}\propto\Delta p_{eyy,E}/l_{u}$ and $\Delta p_{eyy,E}\propto l_{u}$, the reconnection rate becomes independent of $l_{u}$. ### IV.5 Numerical validation of the model As described above, the theory is aided by fully kinetic VPIC simulations carried out for a matrix of $\beta_{e\infty}$ and $m_{i}/m_{e}$ values. We next validate each of the separate theoretical predictions against the numerical results. First, the described dynamics involving trapped electrons require that within the inflow regions $|u_{e\|}|\ll v_{te}$, which is valid for $\beta_{e\infty}\gtrsim\sqrt{m_{e}/m_{i}}$ egedal:2015 ; le:2015 , and the predicted profiles of $T_{e\|}/T_{e\perp}$ as well as $B_{H}$ are confirmed in Figs. 9(a,b). Figure 9: Obtained from a matrix of 35 kinetic simulations, profiles as functions of $\log_{10}(\beta_{e\infty})$ are shown for a range of parameters as marked in the panels. In a) and b) the regions shaded in magenta are obtained from the CGL-limit of the Lê-2009 Equations of State le:2009 . In c) $\hat{E}_{\mbox{\scriptsize rec}}$ is normalized by $T_{e\infty}/(ed_{e\infty})$, and the magenta shaded region is the expected range based on $E_{\mbox{\scriptsize rec}}=\alpha v_{A}B_{\mbox{\scriptsize rec}}$, with $\alpha=0.1$ and $100\leq m_{i}/m_{e}\leq 1836$ (see text). The wider yellow shaded region correspond to $0.08\leq\alpha\leq 0.12$. The profiles in d–i) are various electron pressure tensor derivatives, normalized by $-nT_{e\infty}/d_{e\infty}$, such that the resulting dimensionless values can be compared directly to those of $\hat{E}_{\mbox{\scriptsize rec}}$ in c). For all the numerical runs in our study, the individual contributions of $\partial p_{exy}/\partial x$ and $\partial p_{eyz}/\partial z$ are shown in Figs. 9(d,e), respectively. Similar to $\hat{E}_{\mbox{\scriptsize rec}}$, all terms in Fig. 9 involving pressure derivatives are normalized by $-nT_{e\infty}/d_{e\infty}$. With this normalization, confirming Eq. (8), the sum of $\partial p_{exy}/\partial x+\partial p_{eyz}/\partial z$ in Fig. 9(f) reproduces with good accuracy $\hat{E}_{\mbox{\scriptsize rec}}$ in Fig. 9(c). Eqs. (9)-(11) represent a key insight for the theory which is confirmed numerically, as Fig. 9(g) provides an accurate representation of Fig. 9(d). Consistent with recent spacecraft observations egedal:2019 , $\partial p_{exy}/\partial x$ is mostly negative for runs with $m_{i}/m_{e}\geq 400$, and is thus dominated by the external stress imposed by ${p_{e\parallel}^{\mbox{\scriptsize edge}}}\gg{p_{e\perp}^{\mbox{\scriptsize edge}}}$. Likewise, Eq. (12) is confirmed numerically, as the profiles of $\frac{3}{4}{\Delta p_{e\|\perp}^{\mbox{\scriptsize max}}}/l_{u}$ in Fig. 9(h), are observed to provide a match to $\partial p_{eyz}/\partial z$ in Fig. 9(e). This is especially the case for $m_{i}/m_{e}\geq 400$, corresponding to the adiabatic limit required for the validity of Eq. (6), and for which the sums shown in Fig. 9(i) of the profiles of Figs. 9(g,h) provide a good match to $\hat{E}_{\mbox{\scriptsize rec}}$ in Fig. 9(c), confirming Eq. (13). ## V Summary and discussion The use of the VPIC code implemented in a modern super-computing facility enabled a matrix of simulations to be carried for anti-parallel magnetic reconnection. This matrix spans a range of the normalized electron pressures, $\beta_{e\infty}$, as well as the ion to electron mass ratios, $m_{i}/m_{e}$. Our study reveals a range of results, which require new interpretations of the electron dynamics of the EDR and IDR that are significantly different from previous models including those in Refs. hesse:1999 ; drake:2008 ; daughton:2006 . A main difference from the numerical studies a decade or more ago is our recent ability to carry out routine kinetic simulations at the natural proton to electron mass ratio of $m_{i}/m_{e}=1836$. At full mass ratio the effect of the electron pressure anisotropy that develops in the reconnection inflow becomes the dominant force-term not only within the EDR but also for a significant fraction of the IDR. It has previously been determined le:2009 ; egedal:2013 that upstream of the EDR strong electron pressure anisotropy with $p_{e\|}\gg p_{e\perp}$ is driven by the convection of electrons into the reconnection region characterized by low values of $B$. This upstream pressure anisotropy is responsible for driving the strong electron currents within the EDR. In addition, with Fig. 3 we visualize (for the first time) how the electron inflow speed for locations inside the IDR (but outside the traditional EDR) exceeds the inflow speed of the magnetic field by a factor of 5. Thus, the main assumption of “whistler reconnection” drake:2008 that ${\bf E}+{\bf v}_{e}\times{\bf B}=0$ within the IDR is not supported by the simulations. Instead, the strong Hall field $B_{y}$ perturbations are mainly a consequence of dynamics related to the electron pressure anisotropy yielding a new and dominate force balance constraint, namely $({\bf J}\times{\bf B})_{y}\simeq(\nabla\cdot{\bf p}_{e})_{y}\gg en_{e}E_{\mbox{\scriptsize rec}}$. While we cannot rule out that the condition $({\bf J}\times{\bf B})_{y}\simeq(\nabla\cdot{\bf p}_{e})_{y}$ is compatible with dispersive waves rogers:2001 , we conclude that for $m_{i}/m_{e}=1836$ the wave dynamics associated with the standing $B_{y}$-Hall structure around the EDR is not governed by the common whistler wave. Rather, the inflow into the EDR is strongly enhanced directly by ${\bf J}_{\mbox{\scriptsize extra}}=[({p_{e\parallel}}-{p_{e\perp}})/B]{\bf b}\times\mathbf{\kappa}$. This enhanced perpendicular flow into the EDR must in turn be sourced by electrons flowing along the field lines on the inflow- side of the separatrix layers. This is consistent with spacecraft observations that the Hall currents extend many tens of $d_{i}$ away from the EDR manapat:2006 . Aside from the results summarized above, the main goal of this paper is to develop a theory that can account for the electron momentum balance directly at the $X$-line within the center of the EDR for anti-parallel reconnection. Previously, Refs. kuznetsova1998kinetic ; hesse:1999 provided approximate expressions for the off-diagonal stress terms $\partial p_{exy}/\partial x$ and $\partial p_{eyz}/\partial z$. We observe that their theoretical form for $\partial p_{eyz}/\partial z$ is in excellent agreement with our numerical results. Meanwhile, the electron pressure anisotropy that develops upstream of the EDR turns out to have a large impact on $\partial p_{exy}/\partial x$ not previously considered. Our updated theory for $\partial p_{exy}/\partial x$ is cast in two separate terms. The first term is caused by the upstream electron pressure anisotropy and cancels with high accuracy the stress of $\partial p_{eyz}/\partial z$ predicted in Refs. kuznetsova1998kinetic ; hesse:1999 . In hindsight, this cancellation is not surprising as the EDR current, mainly driven by the upstream $p_{e\|}\gg p_{e\perp}$, must be in force balance independent of the value of $E_{\mbox{\scriptsize rec}}$. The additional second term for $\partial p_{exy}/\partial x$ in Eq. (11) is the term that scales with $E_{\mbox{\scriptsize rec}}$ and as such, can be considered the term that actually breaks the electron frozen-in condition at the $X$-line. This term is related to the increase of the $p_{eyy}$ pressure tensor element caused directly through heating by $E_{\mbox{\scriptsize rec}}$. Because the electron flow in the $y$-direction is largely fixed by the upstream $p_{e\|}\gg p_{e\perp}$ (see Eq. (6)), the increase in $p_{eyy}$ becomes linear in $E_{\mbox{\scriptsize rec}}$, and the rotation with length scale $l_{u}$ of $p_{eyy}$ in the $xy$-plane then yields the stress required to balance $E_{\mbox{\scriptsize rec}}$. Because the term is proportional to $E_{\mbox{\scriptsize rec}}$, it facilitates reconnection at any external rate imposed onto the EDR. From Fig. 9(c) it is clear how numerical runs with $m_{i}/m_{e}\ll 1836$ impose (from the electrons’ perspective) rates of reconnection which will be larger than those observed in naturally occurring systems (with $m_{i}/m_{e}=1836)$. Even so, the identified reconnection mechanism can still accommodate the resulting unrealistically large values of $\hat{E}_{\mbox{\scriptsize rec}}$. In closing, we note that our new theory predicts that $\partial p_{exy}/\partial x$ and $\partial p_{eyz}/\partial z$ have opposite signs for $m_{i}/m_{e}=1836$. This result has been directly confirmed in spacecraft observations of an EDR encounter by MMS in the Earth’s magnetotail egedal:2019 . Other MMS observations also support the approximation that reconnection in the Earth’s magnetotail occurs in regimes consistent with laminar 2D kinetic models. One reason for the apparent success of such laminar 2D models is perhaps that for $m_{i}/m_{e}=1836$ the force by $E_{\mbox{\scriptsize rec}}$ is small compared to the forces associated with electron pressure anisotropy, such that $E_{\mbox{\scriptsize rec}}$ only slightly perturbs the electron orbit dynamics within the EDR. Again, from the perspective of the electrons the reconnection rates imposed by the inertia and dynamics of the much heavier ions are in fact very modest and the small modifications imposed by $E_{\mbox{\scriptsize rec}}$ on the electron motion is not sufficient to drive strong instabilities. In order for instabilities to alter the momentum equation in Eq. (1) they much have an inverse growth rate similar to the short electron transit time through the EDR. For example, the Lower-Hybrid-Drift- Instability (LHDI) may perturb the out-of-plane structure of the EDR, but this does not necessarily cause a fundamental change in its underlying 2D dynamics le:2017 ; greess:2021 ; schroeder:2022 . Given recent progress in the understanding of how the larger scale ion dynamics influence the reconnection rate stanier:2015 ; liu:2017 , a more detailed and complete picture for reconnection now emerges consistent with local observations of reconnection within the Earth’s magnetotail. ## DATA AVAILABILITY The data that support the findings of this study are available from the corresponding author upon reasonable request. ## Appendix A: Force balance along the EDR Figure 10: Considering the same simulation ($m_{i}/m_{e}=1836$ and $\beta_{e\infty}=2^{-4}$) as in Fig. 1, color contours are presented for terms related to Eq. (3). The EDR is outlined by the cyan rectangle, and a fluid- element is selected for analysis by the black/magenta square. In Fig. 10 we explore how the large values of $({\bf J}\times{\bf B})_{y}$ inside the EDR are related to the electron pressure anisotropy that forms upstream of the EDR. For this, consider the small fluid element which in each panel is marked by the black/magenta square. Naturally, we can integrate Eq. (3) over the surface of this fluid element. For the RHS of Eq. (3) the largest contribution is from $\int(\nabla\cdot{\bf p}_{e})_{y}dA$. In turn, by the divergence theorem $\int(\nabla\cdot{\bf p}_{e})_{y}dA=\Delta{\bf p}_{exy}\,\Delta z+\Delta{\bf p}_{eyz}\,\Delta x$. Here, $\Delta{\bf p}_{exy}$ is the change in ${\bf p}_{exy}$ between the two black sides of the element, while $\Delta{\bf p}_{eyz}$ is the change in ${\bf p}_{eyz}$ between the two magenta sides. Comparing Figs. 10(a,b) it is clear that $|\Delta{\bf p}_{exy}|\ll|\Delta{\bf p}_{eyz}|$, such that $\int(\nabla\cdot{\bf p}_{e})_{y}dA\simeq\Delta{\bf p}_{eyz}\,\Delta x$. We have now demonstrated that $\int({\bf J}\times{\bf B})_{y}dA\simeq\Delta{\bf p}_{eyz}\,\Delta x$. On the other hand, ${\bf J}\times{\bf B}=\nabla\cdot{\bf T}$, where ${\bf T}=(B^{2}/\mu_{0})({\bf bb}-{\bf I}/2)$ is the Maxwell stress tensor. By similar arguments to those applied when integrating $\int(\nabla\cdot{\bf p}_{e})_{y}dA$, we then find that $\int({\bf J}\times{\bf B})_{y}dA=\int(\nabla\cdot{\bf T})_{y}dA\simeq\Delta{\bf T}_{yz}\,\Delta x$. Neglecting $E_{y}$, the force balance constraint in Eq. (2) can therefore be expressed as $\Delta{\bf T}_{yz}\simeq\Delta{\bf p}_{eyz}$, and given these elements are asymmetric about $z=0$ this further reduces to ${\bf T}_{yz}\simeq{\bf p}_{eyz}$ along the edge of the EDR. An identical analysis can be carried out for the forces in the $x$-direction, which yields the similar result that ${\bf T}_{xz}\simeq{\bf p}_{exz}$ along the edge of the EDR. The integrals over the fluid element involved the full pressure tensor ${\bf p}_{e}$, but by the divergence theorem the results only depend on the values of ${\bf p}_{e}$ at the edge of the EDR; here ${\bf p}_{e}$ is well approximated by its CGL-form ${\bf p}_{e}\simeq{\bf p}_{eg}=({p_{e\parallel}}-p_{e\perp}){\bf bb}+p_{e\perp}{\bf I}$. To illustrate that ${\bf p}_{e}\simeq{\bf p}_{eg}$ outside the EDR, in Figs. 10(b,c) the profiles of ${\bf p}_{eyz}$ and ${\bf p}_{eg,yz}$ can be compared directly. Simple manipulations then show that the two force balance constraints that ${\bf T}_{yz}\simeq{\bf p}_{eyz}$ and ${\bf T}_{xz}\simeq{\bf p}_{exz}$, can then be expressed on a common form through the marginal firehose condition ${p_{e\parallel}}-p_{e\perp}=B^{2}/\mu_{0}$, again, applicable along the edge of the EDR. To summarize this analysis, we have shown that for the present simulation at $m_{i}/m_{e}=1836$ that $|en_{e}E_{y}|\ll|({\bf J}\times{\bf B})_{y}|\simeq|(\nabla\cdot{\bf p}_{e})_{y}|$. By integrating over a fluid element that spans the width of the EDR, it is then found that the strong ${\bf J}\times{\bf B}$-forces within the EDR are balanced by the force of the pressure anisotropy at the edge of the EDR. In turn, at the edge of the EDR the Lê-2009 equations of state le:2009 are applicable, and together with the force balance condition ${p_{e\parallel}}-p_{e\perp}=B^{2}/\mu_{0}$, this leads to the scaling law for the current across the EDR as expressed in Eq. (6). It is noteworthy here how the current across the EDR is expressed in terms of parameters evaluated far upstream of the reconnection region. ## References * (1) J. Dungey, “Conditions for the occurence of electrical discharges in astrophysical systems,” Philosophical Magazine, vol. 44, p. 725, 1953. * (2) S. Masuda, T. Kosugi, H. Hara, and Y. Ogawaray, “A loop top hard x-reay source i a compact solar-flare as evidence for magnetic reconnection,” Nature, vol. 371, pp. 495–497, OCT 6 1994. * (3) R. McPherron, C. Russell, and M. Aubry, “Satellite studies of magnetospheric substorms on august 15, 1968 .9. phenomenological model for substorms,” J. Geophys. Res., vol. 78, no. 16, pp. 3131–3149, 1973. * (4) E. G. Zweibel and M. Yamada, “Perspectives on magnetic reconnection,” Proc. R. Soc. A, vol. 472, 2016. * (5) R. B. Torbert, J. L. Burch, T. D. Phan, M. Hesse, M. R. Argall, J. Shuster, R. E. Ergun, L. Alm, R. Nakamura, K. J. Genestreti, D. J. Gershman, W. R. Paterson, D. L. Turner, I. Cohen, B. L. Giles, C. J. Pollock, S. Wang, L.-J. Chen, J. E. Stawarz, J. P. Eastwood, K. J. Hwang, C. Farrugia, I. Dors, H. Vaith, C. Mouikis, A. Ardakani, B. H. Mauk, S. A. Fuselier, C. T. Russell, R. J. Strangeway, T. E. Moore, J. F. Drake, M. A. Shay, Y. V. Khotyaintsev, P.-A. Lindqvist, W. Baumjohann, F. D. Wilder, N. Ahmadi, J. C. Dorelli, L. A. Avanov, M. Oka, D. N. Baker, J. F. Fennell, J. B. Blake, A. N. Jaynes, O. Le Contel, S. M. Petrinec, B. Lavraud, and Y. Saito, “Electron-scale dynamics of the diffusion region during symmetric magnetic reconnection in space,” Science, 2018. * (6) J. Egedal, A. Le, W. Daughton, B. Wetherton, P. A. Cassak, J. L. Burch, B. Lavraud, J. Dorelli, D. J. Gershman, and L. A. Avanov, “Spacecraft Observations of Oblique Electron Beams Breaking the Frozen-In Law During Asymmetric Reconnection,” Phys. Rev. Lett., vol. 120, JAN 29 2018. * (7) J. Egedal, J. Ng, A. Le, W. Daughton, B. Wetherton, J. Dorelli, D. Gershman, and A. Rager, “Pressure tensor elements breaking the frozen-in law during reconnection in earth’s magnetotail,” Phys. Rev. Lett., vol. 123, p. 225101, Nov 2019. * (8) S. Greess, J. Egedal, A. Stanier, W. Daughton, J. Olson, A. Le, R. Myers, A. Millet-Ayala, M. Clark, J. Wallace, D. Endrizzi, and C. Forest, “Laboratory verification of electron-scale reconnection regions modulated by a three-dimensional instability,” J. Geophys. Res., vol. 126, JUL 2021\. * (9) J. M. Schroeder, J. Egedal, G. Cozzani, Y. V. Khotyaintsev, W. Daughton, R. E. Denton, and J. L. Burch, “2D Reconstruction of Magnetotail Electron Diffusion Region Measured by MMS,” Geophy. Res. Lett., vol. 49, no. 19, p. e2022GL100384, 2022. e2022GL100384 2022GL100384. * (10) M. Hesse, K. Schindler, J. Birn, and M. Kuznetsova, “The diffusion region in collisionless magnetic reconnection,” Phys. Plasmas, vol. 6, pp. 1781–1795, MAY 1999. 40th Annual Meeting of the Division of Plasma Physics of the American-Physical-Society, New Orleans, Louisiana, Nov 16-20, 1998. * (11) W. Daughton, J. Scudder, and K. Homa, “Fully kinetic simulations of undriven magnetic reconnection with open boundary conditions,” Phys. Plasmas, vol. 13, p. 072101, JUL 2006. * (12) J. F. Drake, M. A. Shay, and M. Swisdak, “The Hall fields and fast magnetic reconnection,” Phys. Plasmas, vol. 15, APR 2008. * (13) M. M. Kuznetsova, M. Hesse, and D. Winske, “Kinetic quasi-viscous and bulk flow inertia effects in collisionless magnetotail reconnection,” J. Geophys. Res., vol. 103, no. A1, pp. 199–213, 1998. * (14) A. Stanier, W. Daughton, L. Chacon, H. Karimabadi, J. Ng, Y. M. Huang, A. Hakim, and A. Bhattacharjee, “Role of ion kinetic physics in the interaction of magnetic flux ropes,” Phys. Rev. Lett., vol. 115, OCT 21 2015. * (15) “Why does Steady-State Magnetic Reconnection have a Maximum Local Rate of Order 0.1?,” Phys. Rev. Lett., vol. 118, no. 8, 2017. * (16) K. Bowers, B. Albright, L. Yin, W. Daughton, V. Roytershteyn, B. Bergen, and T. Kwan, “Advances in petascale kinetic plasma simulation with VPIC and Roadrunner,” Journal of Physics: Conference Series, vol. 180, p. 012055 (10 pp.), 2009 2009. * (17) J. Egedal, W. Daughton, J. F. Drake, N. Katz, and A. Le, “Formation of a localized acceleration potential during magnetic reconnection with a guide field,” Phys. Plasmas, vol. 16, MAY 2009. * (18) A. Le, J. Egedal, W. Daughton, W. Fox, and N. Katz, “Equations of State for Collisionless Guide-Field Reconnection,” Phys. Rev. Lett., vol. 102, FEB 27 2009. * (19) A. Le, J. Egedal, W. Daughton, J. F. Drake, W. Fox, and N. Katz, “Magnitude of the Hall fields during magnetic reconnection,” Geophy. Res. Lett., vol. 37, FEB 11 2010. * (20) J. Egedal, A. Le, and W. Daughton, “A review of pressure anisotropy caused by electron trapping in collisionless plasma, and its implications for magnetic reconnection,” Phys. Plasmas, vol. 20, JUN 2013. * (21) B. A. Wetherton, J. Egedal, A. Le, and W. Daughton, “Validation of Anisotropic Electron Fluid Closure Through In Situ Spacecraft Observations of Magnetic Reconnection,” Geophy. Res. Lett., vol. 46, pp. 6223–6229, JUN 28 2019. * (22) G. F. Chew, M. L. Goldberger, and F. E. Low, “The boltzmann equation and the one-fluid hydromagnetic equations in the absence of particle collisions,” Proc. Royal Soc. A, vol. 112, p. 236, 1956. * (23) J. Egedal, W. Daughton, A. Le, and A. L. Borg, “Double layer electric fields aiding the production of energetic flat-top distributions and superthermal electrons within magnetic reconnection exhausts,” Phys. Plasmas, vol. 22, OCT 2015. * (24) A. Le, J. Egedal, O. Ohia, W. Daughton, H. Karimabadi, and V. S. Lukin, “Regimes of the Electron Diffusion Region in Magnetic Reconnection,” Phys. Rev. Lett., vol. 110, MAR 28 2013. * (25) J. Egedal, B. Wetherton, W. Daughton, and A. Le, “Processes setting the structure of the electron distribution function within the exhausts of anti-parallel reconnection,” Phys. Plasmas, vol. 23, DEC 2016. * (26) M. Shay, J. Drake, M. Swisdak, and B. Rogers, “The scaling of embedded collisionless reconnection,” Phys. Plasmas, vol. 11, pp. 2199–2213, MAY 2004. * (27) A. Le, J. Egedal, and W. Daughton, “Two-stage bulk electron heating in the diffusion region of anti-parallel symmetric reconnection,” Phys. Plasmas, vol. 23, OCT 2016. * (28) D. S. Payne, C. J. Farrugia, R. B. Torbert, K. Germaschewski, A. R. Rogers, and M. R. Argall, “Origin and structure of electromagnetic generator regions at the edge of the electron diffusion region,” Phys. Plasmas, vol. 28, NOV 2021. * (29) J. L. Burch, J. M. Webster, M. Hesse, K. J. Genestreti, R. E. Denton, T. D. Phan, H. Hasegawa, P. A. Cassak, R. B. Torbert, B. L. Giles, D. J. Gershman, R. E. Ergun, C. T. Russell, R. J. Strangeway, O. Le Contel, K. R. Pritchard, A. T. Marshall, K. J. Hwang, K. Dokgo, S. A. Fuselier, L. J. Chen, S. Wang, M. Swisdak, J. F. Drake, M. R. Argall, K. J. Trattner, M. Yamada, and G. Paschmann, “Electron inflow velocities and reconnection rates at earth’s magnetopause and magnetosheath,” Geophy. Res. Lett., vol. 47, SEP 16 2020\. * (30) P. Montag, J. Egedal, and W. Daughton, “Influence of Inflow Density and Temperature Asymmetry on the Formation of Electron Jets during Magnetic Reconnection,” Geophy. Res. Lett., vol. 47, OCT 28 2020. * (31) V. Roytershteyn, S. Dorfman, W. Daughton, H. Ji, M. Yamada, and H. Karimabadi, “Electromagnetic instability of thin reconnection layers: Comparison of three-dimensional simulations with MRX observations,” Physics of Plasmas, 2013. * (32) M. Hesse, S. Zenitani, and A. Klimas, “The structure of the electron outflow jet in collisionless magnetic reconnection,” Phys. Plasmas, vol. 15, NOV 2008. * (33) A. Le, J. Egedal, W. Fox, N. Katz, A. Vrublevskis, W. Daughton, and J. F. Drake, “Equations of state in collisionless magnetic reconnection,” Phys. Plasmas, vol. 17, p. 055703, MAY 2010. 51st Annual Meeting of the Division of Plasma Physics of the American Physical Society, Atlanta, GA, NOV 02-06, 2009. * (34) E. N. Parker, “Sweet’s mechanism for merging magnetic fields in conducting fluids,” J. Geophys. Res., vol. 62, p. 509, 1957. * (35) H. Karimabadi, W. Daughton, and J. Scudder, “Multi-scale structure of the electron diffusion region,” Geophy. Res. Lett., vol. 34, JUL 11 2007\. * (36) M. A. Shay, J. F. Drake, and M. Swisdak, “Two-scale structure of the electron dissipation region during collisionless magnetic reconnection,” Phys. Rev. Lett., vol. 99, OCT 12 2007. * (37) J. Ng, J. Egedal, A. Le, W. Daughton, and L. J. Chen, “Kinetic structure of the electron diffusion region in antiparallel magnetic reconnection,” Phys. Rev. Lett., vol. 106, FEB 10 2011. * (38) T. Speiser, “Particle trajectories in model current sheets .i. analytical solutions,” J. Geophys. Res., vol. 70, p. 4219, 1965. * (39) J. R. Shuster, L. J. Chen, M. Hesse, M. R. Argall, W. Daughton, R. B. Torbert, and N. Bessho, “Spatiotemporal evolution of electron characteristics in the electron diffusion region of magnetic reconnection: Implications for acceleration and heating,” Geophy. Res. Lett., vol. 42, pp. 2586–2593, APR 28 2015. * (40) A. Le, J. Egedal, W. Daughton, V. Roytershteyn, H. Karimabadi, and C. Forest, “Transition in electron physics of magnetic reconnection in weakly collisional plasma,” Journal of Plasma Physics, vol. 81, 1 2015. * (41) B. N. Rogers, R. E. Denton, J. F. Drake, and M. A. Shay, “Role of dispersive waves in collisionless magnetic reconnection,” Phys. Rev. Lett., vol. 8719, p. 195004, NOV 5 2001. * (42) M. Manapat, M. Øieroset, T. Phan, R. Lin, and M. Fujimoto, “Field-aligned electrons at the lobe/plasma sheet boundary in the mid-to-distant magnetotail and their association with reconnection,” Geophy. Res. Lett., vol. 33, MAR 3 2006. * (43) A. Le, W. Daughton, L. J. Chen, and J. Egedal, “Enhanced electron mixing and heating in 3-D asymmetric reconnection at the Earth’s magnetopause,” Geophy. Res. Lett., vol. 44, pp. 2096–2104, MAR 16 2017.
spacing=nonfrench # Unsafe at Any Copy: Name Collisions from Mixing Case Sensitivities Aditya Basu$*$ <EMAIL_ADDRESS>John Sampson$*$ <EMAIL_ADDRESS>Zhiyun Qian† <EMAIL_ADDRESS>Trent Jaeger$*$ <EMAIL_ADDRESS>$*$The Pennsylvania State University †University of California, Riverside ###### Abstract File name confusion attacks, such as malicious symlinks and file squatting, have long been studied as sources of security vulnerabilities. However, a recently emerged type, i.e., case-sensitivity-induced name collisions, has not been scrutinized. These collisions are introduced by differences in name resolution under case-sensitive and case-insensitive file systems or directories. A prominent example is the recent Git vulnerability (CVE-2021-21300) which can lead to code execution on a victim client when it clones a maliciously crafted repository onto a case-insensitive file system. With trends including ext4 adding support for per-directory case-insensitivity and the broad deployment of the Windows Subsystem for Linux, the prerequisites for such vulnerabilities are increasingly likely to exist even in a single system. In this paper, we make a first effort to investigate how and where the lack of any uniform approach to handling name collisions leads to a diffusion of responsibility and resultant vulnerabilities. Interestingly, we demonstrate the existence of a range of novel security challenges arising from name collisions and their inconsistent handling by low-level utilities and applications. Specifically, our experiments show that utilities handle many name collision scenarios unsafely, leaving the responsibility to applications whose developers are unfortunately not yet aware of the threats. We examine three case studies as a first step towards systematically understanding the emerging type of name collision vulnerability. ## 1 Introduction A fundamental file system design choice is whether it will allow file names to be case sensitive or not, and modern file systems are diverse in their selection. A case-sensitive file system is one that allows the definition of multiple files whose names differ only in their case, such as Foo.c and foo.c. In a case-insensitive file system, only one file can be defined whose names differ only in their case. Historically, UNIX file systems are case sensitive, whereas Windows file systems are case insensitive. Further, case-insensitive file systems may be either case preserving (e.g. Apple File System (APFS), NTFS, etc.) or not (FAT), where a case-preserving file system preserves the case chosen (i.e., either Foo.c or foo.c), rather than converting all names to one case choice (e.g., all lowercase). Importantly, while choices in case sensitivity for a single file system may appear to be arbitrary or aesthetically driven, the precise semantics of interactions between two file systems with different case sensitivities can range from subtle to ill- defined, with associated consequences. Practitioners have long had concerns about the implications of leaving case sensitivity as an open design choice [31] Historically, these concerns were not considered as pressing when file systems were associated with their respective operating systems and associated singular assumptions about case. However, individual systems now frequently support a mixture of case-sensitive and case-insensitive file systems, creating opportunities for files to be moved between file systems with different cases and file identifier encodings. More troublingly, several file systems now support allowing the choice of case for individual directories [12], complicating file operations by having multiple case and encoding semantics within the same file system. Security risks related to this design choice therefore appear to be increasing. First, the Windows Subsystem for Linux [58] (WSL) integrates Linux and Windows platforms leading to expectations that files may be routinely copied from Linux (i.e., case-sensitive) to Windows (i.e., case-insensitive) file systems. Second, Linux ext4 now supports case-sensitive and case- insensitive naming in the same partition, configurable per directory [12, 34]. Linus Torvalds expressed concerns about adding such support to ext4 [31], stating that such features often cause "actual and very subtle security issues." Indeed, security issues caused by moving files from case-sensitive to case- insensitive file systems are starting to appear. For example, the git distributed version control system has suffered from multiple vulnerabilities (e.g., CVE-2014-9390, CVE-2021-21300), caused by how git clones repositories from case-sensitive file systems to case-insensitive file systems. To exploit this, an adversary creates a repository in a case-sensitive file system with a directory whose name will collide (i.e., only differs in case) with a symbolic link (to another directory) added by git when the repository is cloned to a case-insensitive file system. The name collision between the directory and the symbolic link enables adversaries to overwrite the scripts that git executes. Such attacks can alter both the target resource’s content and/or its metadata, including its permission assignments. Researchers have long been aware of hazards that may occur during file system name resolution [3, 4], particularly that programmers must validate safe use of symbolic links and check for "squatted" files when creating a new files. Many defenses have been proposed [7, 8, 55, 9, 40, 41, 42, 30, 50, 51, 52]. However, to the best of our knowledge, ours is the first work studying how case interplays cause name collisions that lead to incorrect, and in some cases, vulnerable behaviors. We show that utilities and applications currently do not recognize unsafe use of case-insensitive file systems, leading to these problems. In particular, we study the behavior of six utilities used to copy files, which is how case-sensitive and case-insensitive file systems can interact. We find a wide variety of responses to name collisions, including many that overwrite existing data and change file permissions silently. In addition, we examine three case studies in using such utilities that result in unsafe and sometimes exploitable behaviors. This paper demonstrates the potential implications of the name collision problem, focusing on Linux and its supported file systems, thereby motivating both more and broader (e.g., other OS-FS combination) investigations. This paper makes the following contributions: * • We examine the security and correctness implications of name collisions, when two distinct file system resources with two distinct names map to to a single name, due to file system case sensitivity and/or encoding mismatches. * • We develop an automated method to test common Linux utilities for unsafe reactions to name collisions, finding a wide variety of responses, many of which are unsafe and possibly exploitable. * • We perform detailed case studies of the impact of name collisions on three programs dpkg, rsync, and Apache, showing how they operate incorrectly in the face of name collisions and how they would be exploited when deployed on case- insensitive directories. ## 2 Background: From Cases to Collisions Beyond traditional, i.e. operating-system-entailed, decisions made with respect to case sensitivity, even Linux files systems now represent a surprising diversity of case sensitivity decisions. In particular, the desire to support some non-native applications, such as WINE and Samba from Windows systems, has motivated Linux file systems to support the case-insensitive file naming used in these non-native file systems. The ability to create case-insensitive file systems has long been possible in some Linux file systems, such as ZFS, JFS, and ciopfs. However, these options are applied to the entire filesystem, rather than just the relevant directories for individual applications. In 2019, Linux kernel version 5.2 added support for per-directory case-insensitivity to ext4[12, 34]. Later in 2019, similar support was added to the Flash-Friendly File System (F2FS) in Linux kernel version 5.4 [13, 14]. For case-insensitive directories, these file systems are case-preserving in nature. ### 2.1 Motivations for Increasing Case Diversity Samba: Samba [45] implements the Common Internet File System (CIFS) protocol which allows for sharing file systems over a network. Its primary use is sharing files with Windows clients that expect a case-insensitive file system. Hence, Samba implements user-space case-insensitive lookups even if the underlying file system is case-sensitive. Furthermore, it allows turning on/off case-sensitivity and case-preservation on a per-mount basis [46]. Note that this feature only works for non-Windows clients, which means that the actual file system can contain files differing only in case. This can lead to unexpected behaviors where Samba will choose to show only a subset of files. Deleting files which have collisions will now show the alternate versions, thereby giving rise to inconsistent behavior from the end user’s perspective. Samba’s requirement of case-insensitive matching, which is done in user-space, incurs a huge performance overhead [37] thereby motivating the support for case-insensitivity in the ext4 file system [34, 35, 36]. Other programs/systems such as Wine [57], Network File System (NFS), SteamOS [49, 48] and Android [32, 59] would also benefit from in-kernel case-insensitivity support. ext4: For ext4, the idea is that the filesystem at large can be configured to be "casefolding," which permits the mixing of case-sensitive and case- insensitive directories in the same file system. When creating an ext4 file system, the _casefold_ option is applied, e.g., mkfs -t ext4 -O casefold /dev/sda. Setting the +F inode attribute on an empty directory makes it case- insensitive, e.g., mkdir foo; chattr +F foo. Note that case-insensitive directories can contain case-sensitive directories. This means that for a given path, /foo/bar/bin/baz, any of foo, bar and bin can either be case- sensitive or case-insensitive. tmpfs: tmpfs recently added case-insensitivity support [33]. The use cases are similar to that of ext4 with the addition of supporting sandboxing and container tools such as Flatpak. ### 2.2 Name Collisions A name collision occurs when a file system maps two distinct names of two distinct resources to the same name. Name collisions can cause problems to occur if the names of distinct resources _collide_ when those resources are replicated to a target directory that does not provide a 1:1 mapping for all replicated objects. Suppose one directory has two files with distinct names in that file system. Should those files be copied to a second directory in which the two file names collide (i.e., are mapped to the same name), then only one file will be created, which may be either of the original files or an unpredictable combination of the two files’ content and metadata. Variation in case sensitivity between two file systems is a common origin of collisions, but diversity in other encoding properties, such as character choice (e.g., FAT does not support $``$, $:$, $*$, etc. 111http://elm- chan.org/fsw/ff/doc/filename.html) and canonicalization processes, can lead to the same effect. For example, NTFS uses UTF-16 while APFS (macOS) and ext4 (Linux) use UTF-8 and older file systems can use other encoding schemes, such as iso8859-1. Modern file systems perform the canonicalization of names using a technique called _case folding_[6]. Unlike traditional techniques, case folding uses lookup tables to transform each character of the filename to a pre-determined case. Furthermore, in Unicode several different characters (or code points) can be used to represent the same abstract character. Hence, a normalization step is needed after performing case folding to ensure that the binary representations match. Unfortunately, these case folding rules can differ across file systems. Additionally, the _locale_ (or language) also influences the case folding rules. Due to such differences, ‘temp_200K’ (where K = Kelvin Sign, U+212A in UTF-8) and ‘temp_200k’ are considered identical on NTFS and APFS, but on ZFS these filenames are considered different when using case-insensitive lookups. As a result, when two files of these names are copied from a ZFS file system to an NTFS file system, they will collide and only one filename and only one file will be created. Similarly, consider the filenames floß, FLOSS and floss. All can coexist on a case-sensitive file system supporting reasonable character encodings, but, since case-folding for both floß and FLOSS is floss, attempting to move these files to a case-insensitive system may only preserve one of the original triple. Modern encoding schemes such as UTF-8 have support for non-English characters which require case folding to perform case-insensitive matching. This only increases the number of case-insensitive matches, making the problem of name collisions even worse. For clarity and conciseness, we will use examples of ASCII-based, case-insensitive matching throughout the rest of the paper. We propose a taxonomy for _name confusions_ , shown in Figure 1, that captures the types of incorrect program behaviors that may stem from the ambiguous uses of names for file system resources. Name collisions are a subset of this broader class. Name confusions may be caused by three reasons: (1) because multiple names may refer to the same resource (i.e., aliasing); (2) because an adversary may create a resource of that name before the victim (i.e., squat); and (3) because the multiple resources may be associated with the same name (i.e., collisions). Of these, however, name collisions are the least explored for their correctness and security implications. As Linux is adding more support for case-insensitivity, it is crucial to understand the pitfalls and problems such functionality may incur. This work aims to study these issues. [Name Confusion (NC) [Alias [Symlink] [Hardlink] [Bind mount] ] [Squat [File] [Other] ] [Collision [Case] [Encoding] ] ] Figure 1: Taxonomy of name confusion vulnerabilities divided into alias (i.e., multiple names for a resource), collision (i.e., multiple resources for a name), and squat (temporal ambiguities in names vs. resources) classes ## 3 From Collisions to Calamities Name collisions can impair system functionality by modifying the content and/or metadata of files and directories in unexpected ways. Some name collisions have already led to security vulnerabilities [24]. In this section, we define the conditions in which a name collision occurs, the conditions under which such a collision may be exploitable by an adversary, and describe a known vulnerability that is caused by a name collision. ### 3.1 Causes of Name Collisions A process may cause a name collision under the following conditions. * • There exists a _source resource_ (e.g., file or directory) in a case-sensitive file system, whose name is _source name_. * • The process uses a relocation operation to place the source resource in a target directory, where the target directory is a case-insensitive or case- preserving directory. Examples of relocation operations include copy (e.g., cp, rsync, or an archive operation, such as tar or unzip) or move (e.g., mv). * • The relocation operation produces a destination name from the source name for the name of the source resource when placed in the target directory. * • There is a target resource with a target name whose name differs from the source name, but maps to the same name as the destination name does in the target directory (e.g., due to differences in case-folding rules between the source and target directory). * • If the process is authorized to modify the target resource, the process’s relocation operation results in a name collision between the target and source resources. * • If the relocation operation proceeds despite the name collision, then the target resource’s content and/or its metadata may be modified using the source resource content and/or metadata. When these conditions are met, a name collision occurs such that the target resource in the target directory will be modified using the source resource. In most cases, modifying a target resource using a source resource of a different name is an unexpected result. We test how common Linux utilities react to name collisions and examine case studies where name collisions cause incorrect operation. Given the above conditions, there are several clear scenarios where the movement of files involving the following types of file systems (following the categorization in § 2.2) could result in name collisions: * • Case-sensitive and case-insensitive file systems. * • Two distinct case-insensitive file systems with different case folding rules, e.g. ZFS to NTFS, etc. 222ASCII encoding contains a strict subset of characters representable under the UTF-8 encoding scheme. Hence, without loss of generality, we only consider ASCII encoded names for the examples in this paper. * • Two file system whose locales are different but they still use the same file system format (such as ext4). * • A single file system that supports per-directory case-insensitivity, e.g. ext4. Clearly, name collisions may impact system functionality by causing collateral damage to resources supposedly unrelated to the operation, even removing the target resource entirely. In addition, name collisions may be used to exploit the process performing the relocation operation in a version of a confused deputy attack [25]. An adversary only requires write access to the source directory to produce source names that may lead to name collisions to perform an attack. We note that adversaries require fewer permissions to perform attacks using name collisions than other name confusion classes, which require write access to a directory used in name resolution of the target resource [54]. Thus, remote attacks using file system archives, such as tarballs and zip files, as well as file repositories, such as GitHub, can be the sources of attacks. In practice, to perform a successful attack using a name collision, the victim process has to help the adversary in two ways. First, the victim process has to use the source resource in a relocation operation planted by an adversary as described above. In addition to archives, other activities, such as backups, may provide opportunities for exploitation of name collisions. In addition, ad hoc user actions copying files, e.g., from Linux to Windows in the Windows Subsystem for Linux, may result in unexpected and exploitable collisions. Second, the target directory of the relocation operation has to be predictable by the adversary to enable them to produce a source name that leads to a colliding destination name. Archives make this task much easier because the archive itself may be crafted to provide the target resource that is exploited by creating a collision with another archive file. A recent vulnerability in the git distributed revision control system demonstrates exactly this, as described below. ### 3.2 An Example Collision Vulnerability Security vulnerabilities due to filename collisions across different file systems have been demonstrated in the wild. Consider a recent vulnerability in the git distributed version control system (CVE-2021-21300). This vulnerability results in remote code execution after cloning a maliciously crafted repository created on a case-sensitive file system to a case- insensitive file system. .1 repo/. .2 .git/(contents omitted). .2 A/. .3 file1. .3 file2. .3 post- checkout(executable script). .2 a(symlink to .git/hooks/). Figure 2: Example for Git CVE-2021-21300 Figure 2 depicts the maliciously crafted repository structure. Note that this directory structure works correctly on a case-sensitive file system. However, on case-insensitive file systems, the presence of the ‘a’ (small) and ‘A’ (capital) directories creates a collision that exposes a vulnerability. This collision results in a vulnerability when using git’s out-of-order checkout machinery. Git Large File Storage (LFS) uses out-of-order checkouts for downloading binaries in the background. Say the repository creator (adversary) marks ‘A/post-checkout’ for an out-of-order checkout. When a user clones this repository to a case-insensitive file system (e.g., NTFS), git performs a sequence of operations that: (1) replaces ‘A’ with the symbolic link ‘a’ and (2) writes the script file ‘A/post-checkout’ to ‘.git/hooks/post-checkout’ due to the symbolic link ‘a’. After the files are downloaded, git runs the script ‘.git/hooks/post-checkout’ that the adversary provided, which is obviously undesirable. In this case, a maliciously crafted git repository can be designed to provide a target resource of the symbolic link ‘a’, which when collided by ‘A’ in resolving the source resource ‘A/post-checkout’ redirects the operation to a directory chosen by the adversary using the symbolic link. ### 3.3 The State of Defenses for Name Confusions Currently, operating systems provide no innate defenses to prevent name collisions, leaving the challenge to programmers. However, researchers have studied problems due to other types of name confusions extensively, proposing a variety of defenses [7, 8, 9, 40, 41, 42, 30, 50, 51, 52, 44]. However, researchers have shown that comprehensive program defenses are expensive [55] and that system-only defenses will always be prone to some false positives [5]. Leveraging limited program information [53, 28] still results in some false positives. As a result, library commands for opening files have been extended in a variety of ways to prevent name confusions from occurring. The open command has been extended with flags to detect file squats (i.e., O_EXCL|O_CREAT to detect the presence of an existing file during file creation) and prevent unexpected use of aliases (i.e., O_NOFOLLOW to prevent following symbolic links). However, the use of squats and aliases is desirable in some applications, despite their risks. Thus, the openat command has been added to enable programmers to avoid TOCTTOU attacks, by opening a file from a validated directory (i.e., file descriptor to the directory of the desired file). The successful use of openat requires the programmer to check for unwanted squats or aliases themselves. An alternative is proposed by the openat2 command, which instead controls how files may be opened, such as requiring all file components accessed to be descendants of the directory from which the operation originates. openat and openat2 limit the attack surface of squat and alias attacks, but do not eliminate them entirely, depending on the programmer’s additional actions to check for TOCTTOU attacks and configure the commands. At present, the above commands make no effort to help programmers address name collisions. As a result, utilities to perform copy and move operations and applications that may utilize file systems with multiple or mixed (e.g., ext4 and F2FS) case sensitivities or encodings may not detect and resolve name collisions correctly. We will examine the possible defenses for name collisions in § 8. ## 4 Overview In this paper, we aim to explore the impact that name collisions may have on file system security. To do this, we propose to examine three research questions. RQ1: How do applications invoke utilities that may allow unsafe name collisions? In § 6, we examine Linux packages to determine the most common options that applications employ for the utilities used to perform copies. We examine how frequently application packages use utilities in copy operations by scanning their scripts for such operations, as shown in Table 1. RQ2: When do the utilities for performing copy operations allow unsafe name collisions? Recall that § 3.1 defines the conditions under which an unsafe name collision may occur. This research question asks whether the utilities that applications may use to perform copy operations (e.g., cp and tar) prevent unsafe effects when a name collisions occur. For these utilities and the common options found in RQ1, we examine a variety of name collision scenarios to determine whether the utilities allow name collisions and their unsafe effects to occur as shown in Table 2. RQ3: What correctness and security problems are caused by name collisions? In § 7, we examine three case studies where we show how name collisions cause programs to behave incorrectly. In particular, we show concretely how applications can be vulnerable to name collisions when target resources are deployed on case-insensitive or case-preserving file systems. Impacts: A preview of our result is that: (1) many applications rely on these utilities to copy file system resources and repositories/archives; (2) the utilities used to copy file system resources and repositories/archives often allow unsafe name collisions, although the specific responses vary in ad hoc ways; and (3) applications currently lack defenses against name collisions, which can lead to incorrect operation and exploitable vulnerabilities. ## 5 Testing for Name Collisions This section details an automated tool for testing the responses of common Linux utilities used for relocation operations to name collisions. As described in § 3.1, a name collision is caused by creating a source name that will be converted to a destination name by the relocation operation that is equal to a target name in the target directory of the operation. Thus, our aim is develop a method to automate the generation of source resources with names that will lead to name collisions when relocated to case-insensitive targets and identify when operations allow the name collision to occur, detecting the effects of those operations. ### 5.1 Test Case Generation The individual test cases are generated to test file system resources of various types, including regular files, directories, symbolic links (to files and directories), hard links, pipes, and devices. In addition, we have found that creating collisions in non-trivial directory structures may also lead to incorrect behaviors. Figure 3 shows an example test case where the directories as well as their contents result in a collision when transferred to a case- insensitive file system. As a result, we aim to generate test cases that result in name collisions at different depths of the directory being copied, as evidenced by the collision between directory names at depth 2 (i.e., "dir" and "DIR") and the impact on colliding resources of different types (i.e., a regular file "foo" and a pipe "foo"). Input .1 src/. .2 dir/. .3 foo*. .2 DIR/. .3 foo|. copy Effect .1 target/. .2 dir/. .3 foo*|. Here, ‘*’ means a regular file and ‘|’ means file type is a named pipe. Figure 3: An Example Test Case Since we are testing the behavior of various utilities that perform relocation operations, we can control the source and target names in creating test cases. As a result, the choice of names is trivial. We create source directories that contain both the target resource (i.e., a resource copied first from the source to the target) and the source resource (i.e., a resource copied later by the utility to collide with the target resource (i.e., now in the target directory). This is similar to the way name collisions would occur when copying an archive or repository that causes a collision, as the git vulnerability. Since different utilities may process resources in different orders, we generate test cases with both orderings of resources that may cause collisions. USE[msg=10960,‘cp’.openat] 00:39|2389| /mnt/folding/dst/ROOT $\hookleftarrow$ CREATE[msg=10957,‘cp’. openat] 00:39|2389| /mnt/folding/dst/root device | inodeprogramaccessed pathsyscallauditd idoperation Figure 4: Example violation reported by name collision testing The only decisions then are what are the resource types of the source and target resources and where to place them in the source directory hierarchy to cause the desired collision to be created. Symbolic links, pipes, and devices only create interesting behaviors when used as target resources. For symbolic links, the unsafe effect is to follow the link to another file system resources, which only happens with the symbolic link as the target resource. For pipes and devices, the unsafe effect is to send the source resource’s content to the pipe or device, which also is only possible if these are target resources. As a result, the automated test generation produces test cases consisting of source and target resources of all combinations of potentially unsafe resource types and places these test cases at depth one and/or two of the file system hierarchy. For rsync, we specifically found an issue caused by a collision at depth two, without any collision at depth one, as detailed in Section 7.2. ### 5.2 Detecting Collision Effects The key idea is to record the file system operations sufficiently to detect that an unsafe name collision has occurred. Since we design the test cases to create a name collision on a relocation operation, we want to detect when such an operation is a successful collision. Then, we need to determine the impact of the operation to classify the effect according to one of the ten effect options defined in § 6.1. We monitor file system operations using auditd to detect successful collisions. An example of a log indicating a collision is shown in Figure 4. In this example, a create operation creates a target resource named “root” using openat command, but a later use operation to the same resource (i.e., same device-inode pair, see below) is associated with a name “ROOT”, which differs from the name used when the resource was created. Note that although the target resource was created on a case-insensitive file system, multiple names may be used that are resolved to the same name. We say that a collision is successful when we detect a use of a target resource with a different name than that used to create the target resource. To detect such collisions, we first identify the file system operations that create a target resource, recording its combination of device and inode identifiers, which form a unique resource identifier and its pathname. In Figure 4, the name component “root” will be important to detecting the collision. We then capture all the file system operations that use the target resource. In general, if a file pathname in a successful use operation on a target resource does not match the file pathname of a create for the same resource, then we record a positive (i.e., a collision). In Figure 4, the pathname of the use operation differs between “root” and “ROOT”, indicating a name collision. We also record a positive when a use operation deletes and replaces a resource from a prior create operation, as some collisions may cause the target resource to be deleted and the source resource to replace it. We validate that there is a create operation for the colliding destination name to verify the cause of the deletion is a collision. To detect the effect of a name collision, we examine the resulting resource that now maps to the target name. We compare the source resource and target resource content and metadata to the resultant resource to determine whose content and/or metadata (i.e., source, target, or neither) the resource has. For tests on directories and hardlinks, we examine the directories and the resultant directory entries. ## 6 Name Collisions on Linux Copy Utilities Table 1: Prevalence of copy utilities tar | zip | cp | cp* | rsync ---|---|---|---|--- | 10 | mc ---|--- 8 | perl-modules 7 | libkf5libkleo-data 6 | pluma 6 | mc-data | … 107 | TOTAL | 21 | texlive-plain-generic ---|--- 15 | aspell 11 | libarchive-zip-perl 7 | texlive-latex-recommended 5 | texlive-pictures | … 69 | TOTAL | 78 | hplip-data ---|--- 32 | dkms 22 | libltdl-dev 20 | autoconf 18 | ucf | … 538 | TOTAL | 12 | dkms ---|--- 2 | udev 2 | debian-reference-it 2 | debian-reference-es 1 | zsh-common | … 25 | TOTAL | 28 | mariadb-server ---|--- 5 | duplicity 4 | texlive-pictures 2 | vim-runtime 1 | rsync | … 42 | TOTAL We calculate the number of times that each command (tar, zip, etc.) is used inside scripts from various packages. We investigate 4752 .deb packages from the installation disk (DVD #1) of _Debian 11.2.0_. Only the top-five packages are shown (entries are sorted in descending order for each command). In this section, we examine how common Linux utilities that applications use to copy files333We focus on copy operations. The impact on move operations is similar because (unless both target and source are on the same file system) it simply performs a copy first and then deletes the source. from one part of the file system to another react when the copy operation causes a name collision in a case-insensitive directory. To quantify the ubiquity of these utilities, we survey their use by packages on Debian 11.2.0. We retrieve all packages from the Debian installation DVD and count the number of times the copy utilities are used inside the packages’ scripts. The results are summarized in Table 1. Note that the listed uses of these utilities are lower bounds because we do not parse executable binaries. Hence, we miss uses where the utilities are invoked via system calls such as system(...), execve(...), etc. ### 6.1 Collecting Responses to Name Collisions Table 2: Name Collision Responses for Popular Linux Utilities Name Collision between | | | | | | ---|---|---|---|---|---|--- Target Type | Source Type | tar | zip | cp | cp* | rsync | Dropbox file | file | $\times$ | $A$ | $E$ | $+$$\neq$ | $+$$\neq$ | $R$ symlink (to file) | file | $\times$ | $A$ | $E$ | $+$$T$ | $+$$\neq$ | $R$ pipe/device | file | $\times$ | $-$ | $E$ | $+$ | $+$ | $-$ hardlink | file | $\times$ | $-$ | $E$ | $+$$\neq$ | $+$$\neq$ | $-$ hardlink | hardlink | $C$$\times$ | $-$ | $E$ | $C$$\times$ | $C$$+$$\neq$ | $-$ directory | directory | $+$$\neq$ | $+$$\neq$ | $E$ | $+$$\neq$ | $+$$\neq$ | $R$ symlink (to directory) | directory | $+$ | $\infty$ | $E$ | $E$ | $+$$T$ | $R$ This table shows results of copying files/directories from a case-sensitive to a case-insensitive file system. cp* refers to cp being used with shell completion. For e.g., ‘cp * /target’ which copies all items from the current directory to /target directory. $\times$ Delete existing file and create new file $+$ Overwrite existing file. For directories, merge their contents. $\neq$ Mismatch between content and metadata $A$ Ask user to resolve the collision $T$ Traverse symlink $C$ Corrupts non-colliding files $E$ Deny operation and report error $\infty$ Program crashes, or hangs $-$ Ignore unsupported file type (for hardlinks create regular file instead) $R$ Rename colliding file/directory The name collision test cases and the responses of copy utilities are shown in Table 2. The ‘Target Type’ column represents the resource type of the target resource that may be overwritten. The ‘Source Type’ represents the resource type of the source that collides with the target. The rest of the columns represent individual utilities and their responses to name collisions between a source resource of the source type and a target resource of the target type. Below is a comprehensive list of the types of responses observed. Only "Deny" and "Rename" prevent name collisions from causing unsafe and possibly exploitable behaviors, although both may block legitimate functionality in some cases. "Ask the User" may result in an unsafe response if the user allows the target resource to be overwritten. Note that more than one response is possible for each test case. Delete & Recreate ($\times$) _Delete_ the target resource and _create_ a new resource based on the source resource. The new resource’s type, as well as its data and metadata, is determined by the source resource. The target resource is lost without any notification. Overwrite ($+$) Overwrite the data and metadata of the target resource using the source resource. Unlike Delete & Recreate, the name of the target resource is preserved. If file foo is being overwritten with file FOO, then the final file will be named foo but will have the contents and metadata of file FOO. Corrupt ($C$) Contents of a resource that is not involved in name collision (i.e., not the target resource) is modified. For a more in-depth discussion, refer to § 6.2.5. Metadata Mismatch ($\neq$) After a successful copy of a given source resource, some metadata, such as its name, UNIX permissions, user or group ID, extended attributes, or timestamp, remain from the target resource, creating a resource with a mismatch between the data (from the source) and the metadata (from the target). Follow Symlink ($T$) Follow symbolic links, even when explicitly directed not to do so. Rename ($R$) The source name is renamed automatically to avoid creating a name collision, such as by appending a counter, resulting in a copy of the source resource in the target resource’s directory with a non-colliding name. Ask the User ($A$) To resolve a collision, ask the user to choose from a list of actions, such as to overwrite the target resource, skip copying the source resource, rename the target resource, abort, etc. Note that the user can still choose a response that results in adverse consequences. For instance, if the user chooses to overwrite the target, the target’s data and metadata are modified using the source. Deny ($E$) Deny the copy associated with a collision and report an error. Crashes ($\infty$) Collisions can result in the program hanging (e.g., going into an infinite loop) or crashing. Unsupported file type ($-$) Does not support copying a resource if the source resource is of this file type. Note that if hardlinks are not recognized by a utility, then it simply creates a fresh copy of the underlying file. The exact command-line flags used used to generate Table 2 are listed in Appendix A. To identify these flags, we analyzed 4,752 .deb packages on Debian 11.2.0’s installation DVD. We found that the most commonly used flags enabled the following functionality. * • Support recursively copying all directories. * • Support copying symbolic-links and hard-links as-is but do not follow them. * • Preserve metadata such as UNIX permissions, extended attributes (xattr), timestamps, and owner/group IDs (uid/gid). Before examining the responses in Table 2, we briefly note some additional context for two of the columns. ##### cp vs. cp* Both of these represent the same executable binary. The difference is in the way the command-line arguments are passed to the binary. Specifically, the format of specifying the source directories is different. Consider that the source directory (to be copied) is foo. For cp, we will pass it as foo/ while for cp* we will use foo. Note the trailing / is missing in the latter case. Just this difference significantly changes the behavior of cp as noted in Table 2. We use the cp* method of invocation coupled with shell completion, e.g., ‘cp src/* /target’ where the shell replaces src/* with each individual entry present inside src sans the trailing /. When testing the cp method, we change the command to ‘cp src/ /target’. ##### Dropbox Strictly speaking, _Dropbox_[11] is not a copy utility but a popular file synchronization utility. It is intended to replicate entire directories across multiple machines and file systems. We mention Dropbox to highlight its distinct response to handling potential name collisions. Even when the underlying file system is case-sensitive, Dropbox treats it as case-insensitive. It proactively renames the files and directories to avoid name collisions that could occur if they were transferred to a case-insensitive file system. Note, however, that its renaming strategy is not even uniform across platforms: For example, the Dropbox application appends “(Case Conflicts)”, “(Case Conflicts 1)”, etc. to the file/directory names in case of a potential collision, whereas, when using their web-based interface, they append “(1)”, “(2)”, etc. instead. ### 6.2 Unsafe Responses to Name Collisions Several responses shown in Table 2 demonstrate that utilities often allow unsafe responses to name collisions. In this section, we examine some of the more concerning responses to show how utilities delegate responsibility for security against name collisions to the applications that invoke them. For the examples in upcoming sections, src/ and target/ are on case-sensitive and case-insensitive file systems respectively. #### 6.2.1 Silent data loss with tar, cp* & rsync Name collisions involving files generally result in silent data loss. From Table 2, we can see that tar deletes and recreates ($\times$) files when collisions occur. Hence, when there is a name collision between foo and FOO, only one of these files will remain in the target directory. The other file is permanently lost without any notification. Similar to tar, cp* and rsync also lose files silently. However, their behavior of overwriting ($+$) files results in other problems that are discussed later in this section. Unlike tar, zip and cp will ask a user for next steps ($A$) or report an error ($E$) respectively. Hence, they are not prone to silently losing files. #### 6.2.2 Merge directories with tar, zip, rsync & cp* Name collisions involving two directories results in their contents (files, directories, etc. inside the directory) being merged. All of tar, zip, rsync, and cp* will silently merge directory contents without notifying the user. Figure 5 highlights this issue using a directory listing. .1 src/. .2 dir/. .3 subdir/. .3 file1. .3 file2. .2 DIR/. .3 file2. — copy $\rightarrow$ .1 target/. .2 dir/. .3 subdir/. .3 file1. .3 file2. Figure 5: Impact of merging directories In this example in Figure 5, the data of file file2 is overwritten by the content written last in the copy operation. For example, if src/DIR’s contents are written last, then its content for file2 is preserved and src/dir’s is lost. Furthermore, when the colliding directories have different UNIX permissions, a collision results in metadata mismatch ($\neq$). With respect to Table 2, the UNIX permissions of the target resource are overwritten with permissions of the source resource. In Figure 5, consider src/dir/ with perms=700 and an adversary who creates src/DIR/ with perms=777. After a copy (using any of the above utilities), target/dir/ will have perms=777 effectively giving the adversary permission to the contents of the original src/dir/. #### 6.2.3 Stale names Whenever utilities resort to overwriting ($+$), we end up with stale file/directory names. For example, consider a name collision between a target resource foo (file content: ‘bar’) and a source resource FOO (file content: ‘BAR’). After copying with rsync or cp*, we will end up with file foo whose contents are ‘BAR’. The problem with such name collisions is that to the end user (or other programs), it will appear that foo was successfully copied while in reality FOO was copied. Just using the filename is not enough to discern which files were successfully copied. This is especially true for case-preserving file systems where the user has the expectation of the filenames being preserved. Hence, it is not unreasonable for the user to expect foo should contain bar. #### 6.2.4 Symbolic link traversal at target Name collisions between symlink (to file) and a regular file results in cp* following the symlink ($T$) and overwriting ($+$) its target’s contents with that of the regular source file. With regards to Table 2, if the target resource is a symbolic link and the source resource is a file, then cp* ends up following the symlink and writing data to the resource referenced by the symlink. .1 src/. .2 dat(to /foo). .2 DAT= pawn. .1 /foo= bar. — cp* $\rightarrow$ .1 target/. .2 dat(to /foo). .1 /foo= pawn. Figure 6: Following symlink Figure 6 illustrates this case with an example. src/dat is a symbolic link to /foo and /foo contains ‘bar’. Mallory (our adversary) does not have write access to /foo but does have access to src/. She creates src/DAT which contains ‘pawn’. Then the administrator starts the copy using: cp -a src/* target/. At this point, cp first creates the symlink target/dat. Then it overwrites ($+$) this symlink with the contents of src/DAT, effectively updating the file /foo. After the copy has completed, /foo contains ‘pawn’. cp* has no command-line options to prevent traversal of symbolic links at the target. Only link traversal at the source can be turned off via command-line flags. #### 6.2.5 The case of hardlink – hardlink name collisions During a copy when hardlinks (whose targets are different) collide, it can corrupt ($C$) other non-colliding files and create spurious hardlinks. Table 2 shows that this behavior is exhibited by tar, cp*, and rsync. An interesting observation is that, regardless of whether the utility’s behavior is Delete & Recreate ($\times$) or Overwrite ($+$), this problem affects both. .1 src/. .2 hfoo=foo. .2 zzz=foo. .2 hbar=bar. .2 ZZZ=bar. — rsync $\rightarrow$ .1 target/. .2 hfoo=bar. .2 zzz=bar. .2 hbar=bar. Figure 7: hardlink – hardlink name collision To understand this scenario, consider Figure 7 that uses rsync to perform the copy. The same color coding represents files that are hard-linked to each other. So src/hfoo and src/zzz are hard-linked, representing the same file. These files contain ‘foo’. Similarly, src/hbar and src/ZZZ are hard-linked and they contain ‘bar’. After copying using rsync, target/ contains three files that are all hard- linked to each other. Unlike the src/ directory, target/hfoo, target/hbar, target/zzz are all hardlinks of each other and they contain ‘bar’. Additionally, note that although the name collision happened between zzz and ZZZ, the contents of hfoo were replaced. Even tar, which deletes the old file and recreates it, exhibits this behavior. The following order of operations undertaken by rsync result in this behavior. 1. 1. Copy src/hbar to target/hbar. Now target/hbar contains ‘bar’. 2. 2. Copy src/zzz to target/zzz. Now target/zzz contains ‘foo’. 3. 3. In target/, hardlink ZZZ to hbar. Due to NC, this effectively changes zzz to be hard-linked to hbar. Now target/zzz contains ‘bar’. 4. 4. In target/, hardlink hfoo to zzz. Now target/foo contains ‘bar’. Additionally, all three files inside target/ are hard-linked to each other. The above copy is semantically different from the src/. Specifically, name collision results in distinct sets of files getting hard-linked with each other at the target/. ## 7 Case Studies In this section, we examine case studies where name collisions cause unsafe behaviors, some of which are exploitable. ### 7.1 dpkg Package Manager dpkg is the package manager on Debian OS and its derivatives such as Ubuntu. dpkg packages are compressed tarballs with extension .deb. When dpkg processes a package, it tracks all files it creates during package installations in a database. Before installing a new package, dpkg leverages this database to ensure that any files of previously installed packages will not by overwritten by this new package thereby preventing potentially malformed packages from corrupting the system. On the other hand, we have observed that dpkg will allow a package installation to replace any file whose name is not in its database, even privileged user files. Thus, as long as a file in a package has a filename that does not match the filename of another package’s file, dpkg will install the file, silently replacing any existing file. However, regardless of the underlying file system, the above database is matched in a _case-sensitive_ manner. This allows new packages to _replace files_ of previously installed packages via name collisions effectively circumventing the safeguards in dpkg. In addition, and perhaps even more seriously, dpkg may allow an adversary to replace a package’s customized config file with the default, reverting important changes. deb packages can mark certain files as configuration (or config) files. During package upgrades, if dpkg spots modifications to these config files then it prompts the user to review the changes. However, the config files are also matched in a case-sensitive manner. Under name collisions, dpkg will just replace the original package’s config file with the config file of the new package. For services, such as sshd, httpd, etc., config files are critical to their security, so such overwrites can potentially make the system vulnerable.. ##### Reporting We have reported these issues to the maintainers of dpkg. The maintainers of dpkg have since updated their package documentation[10] to warn end user communities not to use dpkg where targets may be case-insensitive (i.e. specific directories, or entire file systems). During our discussions, we analyzed 74,688 packages and found 12,237 filenames from those packages would collide if a case-insensitive file system were used, breaking multiple packages that contain these files. The name collision problem is fundamentally entrenched into the way dpkg is implemented because it reasons about _names_ without involving the underlying file system(s). ### 7.2 Rsync rsync demonstrates vulnerable behavior when processing name collisions involving _directories_. During copy, the default behavior of rsync is to simply recreate the symbolic links present at source. However, when colliding directories contain sub-directories and symbolic links with the same name, the collision causes rsync to suffer from link traversal444In this case, the name collision makes the alias exploitable, again combining name confusions.. Consider the source directory listed in Figure 8. Here, the directories topdir/ and TOPDIR/ only differ in case. So when copying to a case-insensitive file system, rsync will encounter a name collision. .1 src/. .2 topdir/. .3 secret/symlink to /tmp. .2 TOPDIR/. .3 secret/. .4 confidentialregular file. Figure 8: Case-sensitive source that rsync is copying We use the following command to perform the copy: rsync -a src/ dst/ where, | ---|--- -a | recursively copy directories, preserve symlinks, timestamps, DAC src/ | is case-sensitive dst/ | is case-insensitive After the copy is completed, the newly created files are shown in Figure 9. Note that the file named confidential ends up in /tmp. .1 dst/. .2 TOPDIR/. .3 secret/symlink to /tmp. .1 /tmp/confidentiallink traversal. Figure 9: After copying to case-insensitive destination rsync has created the /tmp/confidential file by following the symbolic link dst/TOPDIR/secret. Below, we describe how this situation can be exploited. Consider an adversary who wants to access a confidential file in TOPDIR/ to which she lacks any access. However, she knows that TOPDIR/ is processed by a backup operation using rsync. If she can create a sibling directory topdir/, to which she will have read-write access, she can direct rsync to write the confidential file (inside TOPDIR/) to any directory of her choosing by creating a symbolic link inside topdir/ to that directory. ##### Reporting We reported this issue to the rsync maintainers, and they told us that user’s should not use rsync with non-case honoring file systems. However, we have concerns about the user community following such a recommendation in this case, since rsync is often used by individuals. In the course of these discussions, we learned the cause of the incorrect behavior. rsync assumes a one-to-one mapping of directories between source and target file systems. When a name collision results in two source directories being mapped to a single directory in the target, rsync can be tricked into incorrectly predicting the target file type. In the presented scenario, a symbolic link src/topdir/secret (to a directory) is incorrectly inferred to be a regular directory src/TOPDIR/secret. rsync uses the O_NOFOLLOW flag with open() to prevent link traversal and uses openat()/openat2() to contain link traversals within a directory hierarchy, but it fails correctly use them because it assumes the symbolic link topdir/secret is a directory. ### 7.3 Apache httpd Security of certain applications relies on the security parameters of the underlying file system. One such application is Apache’s httpd. It allows access to the underlying file system via the HTTP protocol, relying on the UNIX Discretionary Access Control (DAC) permissions555If the system supports Mandatory Access Control (MAC), then DAC is used in conjunction with MAC. to mediate the access. For example, files are accessible over HTTP only if: (i) its UNIX group is www-data and has read permission for the group, or (ii) has world-readable UNIX permissions. Using utilities for copying directories between systems can silently alter these DAC permissions in unintended ways, leading to serious security lapses. We illustrate this scenario using Apache httpd and migration of its data using tar. To study the impact of name collisions on the security parameters, we assume that the migration happens from a case-sensitive to a case-insensitive file system. The behavior of tar discussed below draws from the discussion of Table 2. To protect sensitive directories, httpd can be configured to only allow authenticated users to access specific directories. A commonly used approach is to configure authentication via the .htaccess file [1] which lists the valid users/groups allowed to access a specific directory over HTTP. All sub- directories inside the sensitive directory are also protected. We show that the use of additional security-oriented files can be exploited under the presence of name collisions. ##### Scenario httpd serves the contents of www/ (of Figure 10) over HTTP. Initially, www/ is stored on a case-senstive file system. The directory hidden/ is inaccessible over HTTP since the others permissions are cleared. Next, protected/ is configured to be accessible only to specific users using the .htaccess file. .1 www/. .2 hidden/perm=700. .3 secret.txt. .2 protected/group=www-data, perm=750. .3 .htaccess(only allow valid users). .3 user-file1.txt. .2 index.html. Figure 10: www/ on case-sensitive file system ##### Adversary A UNIX user called Mallory has read-write access to www/ directory. However, DAC permissions prevent her from accessing hidden/ directory because its owner is another user. Additionally, protected is inaccessible since Mallory does not belong to the group www-data. She modifies www/ as shown in Figure 11 and adds the HIDDEN/ and PROTECTED/ directories with the intent of gaining access to hidden/ and protected/ via a name collision. .1 www/. .2 hidden/perm=700. .3 secret.txt. .2 HIDDEN/perm=755. .2 protected/group=www-data, perm=750. .3 .htaccess(only allow valid users). .3 user-file1.txt. .2 PROTECTED/perm=755. .3 .htaccess(empty file). .2 index.html. Figure 11: Adversary modified www/ on the case-sensitive file system ##### Vulnerability tar is used to migrate the adversary-modified www/ directory to another system that uses a case-insensitive file system. Figure 12 shows the state of the file system once the tarball (archive format of tar) is extracted. Now, the previously inaccessible hidden/ directory is now accessible over HTTP. Additionally, since the .htaccess file is cleared, unauthenticated users will be allowed to view protected/ over HTTP. .1 www/. .2 hidden/perm=755. .3 secret.txt. .2 protected/perm=755. .3 .htaccess(empty file). .3 user-file1.txt. .2 index.html. Figure 12: www/ after migrating to case-insensitive file system ##### Reporting We have reported this scenario to the Apache maintainers, but have not yet reached a resolution. Using Table 2, we can reason about the above problems. Under a directory – directory collision, tar incorrectly modifies metadata. This happens for the hidden/ – HIDDEN/ collision. Here, DAC permissions of the latter are applied to the former resulting in the leakage of secret files. For directory – directory collisions, tar will also merge contents of both directories. For protected/ – PROTECTED/ collision, this merger results in the empty .htaccess file overwriting the original one that restricts access to authorized users. The end result is that all users are now allowed access to the new protected/ directory. ## 8 Potential Defenses As discussed in the context of name confusion attacks in general in § 3.3, it can be difficult to produce defenses to prevent name collisions as well. In this section, we discuss some options and their limitations. Name collisions are due to differences in case folding rules among file systems, e.g., case sensitivity and encodings, so it is difficult to ensure that name collisions cannot happen. Suppose a system has only one file system. Even then, an archive constructed on another file system using conflicting case folding rules may cause name collisions to occur when copying the archive. Since user-space programs cannot determine the case-folding rules that may be applied to a file, user-space solutions alone will be unreliable. In addition, they may be prone to TOCTTOU attacks [4, 3]. Thus, extending library calls like realpath to detect name collisions will not sufficiently solve the problem. In addition, system solutions lack knowledge of the programmer intent that caused the collision, so as for name confusion defenses, systems defenses will suffer from one or more of the follow limitations: it must have side information about the programs it protects, it must protect only a subset of all programs, it must be vulnerable to DoS attacks, it must have false-positives, or it must fail to prevent some exploits [5]. For example, one idea may be to write a wrapper to vet archives prior to expansion operations (e.g., tar and zip) to validate that each file in the archive will result in distinct file after expansion. One way to do this is to check for name collisions due to case among all the files in the archive. Although the notion that no files in an archive should collide with any other files within the archive seems intuitively reasonable, there are critical drawbacks to implementing such a defense. For example, there may be files already in the target directory that may result in collisions, limit its utility. More fundamentally, the case folding rules applied by such a wrapper are not guaranteed to be the same as those of the target directory. As a result, we envision that defenses for name collisions will evolve in a manner similar to defenses for name confusions that utilize the open commands. Consider that the open command has flags to check whether a file of a corresponding name exists a creation time, only opening that file when created anew (i.e., O_CREAT|O_EXCL). This call prevents a name collision from overwriting an existing file, but it may be too strong a defense. Suppose one really wants to overwrite files of the same name, but prevent name collisions from modifying files that actually have differing names (i.e., that only match due to case folding). In this case, a new flag is necessary, such as O_EXCL_NAME, which prevents opening a file when the names differ (e.g., based on strcmp), but not when such names match. Unfortunately, even with variants of the open command and other defenses, such as FileProvider classes in Android, programmers continue to make mistakes that lead to errors and vulnerabilities. The challenge is for programmers to determine the intent of their operation, understand the threats faced in such an operation, and configure these complex, low-level commands in such a way that they block the threats while satisfying the intent. Until file system APIs enable this combination of requirements, errors will remain common. ## 9 Related Work Researchers have proposed system and program defenses to thwart name confusion attacks for alias and squat cases. To date, no defenses that are specific to name collisions have been proposed. ##### System Defenses Researchers have long known about name confusion attacks [3, 4] and have proposed a variety of system defenses [7, 8, 9, 40, 41, 42, 30, 50, 51, 52, 44]. In a system defense, the operating system aims to enforce an invariant that prevents name confusion attacks from succeeding. A challenge is that whether an operation is a feature or a vulnerability depends on the programmer’s intent. As a result, system defenses cannot prevent attacks completely without introducing the possibility of false positives [5]. Hybrid defenses have also been proposed [55, 53] where the operating introspects into the process to leverage program state along with file system state in enforcement. While reducing false positives, these techniques still lack programmer intent explicitly, resulting in some false positives. ##### Program Defenses As a result, systems provide APIs for programmers to decide how to handle name confusion attacks. Several file system APIs include flags to avoid using symbolic links entirely (e.g., O_NOFOLLOW flag for the open system call), but in many cases programmers want to be able to use symbolic links. Researchers have proposed program-specific defenses to configure APIs or program frameworks for preventing name confusion attacks [27, 43, 47, 56]. More advanced commands for file allow programmers to manage how files are open, including the impact of symbolic links. For example, the openat system call enables the user to open a directory first to validate its legitimacy before opening the remaining path. openat2 explicitly constrains how name resolution is performed to reduce the potential for attacks. ## 10 Conclusion Interactions among file systems with differing encoding/case-sensitivity design decisions can lead to name collisions when performing maliciously crafted, or even ostensibly benign, copy operations. In this paper, we explored the impact that these name collisions can have on file system security. Current operating systems do not directly prevent name collision- based attacks, delegating that responsibility to programmers. In investigating the utilities used to copy file system resources and repositories/archives, we demonstrate that they often allow unsafe name collisions and lack the sort of uniformity in name-collision handling against which safer use policies could be easily crafted. Further, we show that many applications rely on potentially unsafe use of these utilities, opening themselves up to exploitable vulnerabilities. We examine three case studies in greater detail and demonstrate concrete vulnerabilities to name collisions. Finally, we suggest a series of directions for future research to systematically defend against name collision attacks. ## References * [1] Apache HTTP Server: Authentication and authorization. https://httpd.apache.org/docs/2.4/howto/auth.html#gettingitworking. * [2] Bazaar’s handling of case insentitive file systems. http://doc.bazaar.canonical.com/bzr.1.12/developers/case-insensitive-file-systems.html. * [3] Richard Bisbey, Gerald Popek, Jim Carlstedt, et al. Protection errors in operating systems: Inconsistency of a single data value over time. Technical report, University Of Southern California Marina Del Rey Information Sciences, 1975. * [4] Matt Bishop, Michael Dilger, et al. Checking for race conditions in file accesses. Computing systems, 2(2):131–152, 1996. * [5] Xiang Cai, Yuwei Gui, and Rob Johnson. Exploiting unix file-system races via algorithmic complexity attacks. In 2009 30th IEEE Symposium on Security and Privacy, pages 27–41. IEEE, 2009. * [6] Case mapping vs. case folding. https://www.w3.org/TR/charmod-norm/#definitionCaseFolding. * [7] Suresh Chari, Shai Halevi, and Wietse Z Venema. Where do you want to go today? escalating privileges by pathname manipulation. In NDSS. Citeseer, 2010. * [8] Crispin Cowan, Steve Beattie, Chris Wright, and Greg Kroah-Hartman. RaceGuard: Kernel protection from temporary file race vulnerabilities. In 10th USENIX Security Symposium (USENIX Security 01), 2001. * [9] Drew Dean and Alan J Hu. Fixing races for fun and profit: How to use access (2). In USENIX Security Symposium, pages 195–206, 2004. * [10] dpkg faq: diff of updated documentation. https://wiki.debian.org/Teams/Dpkg/FAQ?action=diff&rev2=78&rev1=77. * [11] Dropbox. https://www.dropbox.com/. * [12] Linux kernel documentation (v5.2): ext4 support. https://www.kernel.org/doc/html/v5.2/admin-guide/ext4.html. * [13] Linux kernel documentation: Flash-Friendly File System (F2FS). https://docs.kernel.org/filesystems/f2fs.html. * [14] f2fs: Support case-insensitive file name lookups (patch<EMAIL_ADDRESS> * [15] ciopfs: case insensitive on purpose filesystem. https://www.brain-dump.org/projects/ciopfs/. * [16] ext3ci – case insensitive ext3 filesystem for Linux 2.6.32. http://bill.herrin.us/freebies/. * [17] Linux kernel documentation: FUSE. https://www.kernel.org/doc/html/latest/filesystems/fuse.html. * [18] IOMap. https://www.mono-project.com/docs/advanced/iomap/. * [19] Jfs filesystem (man page). https://www.unix.com/man-page/redhat/8/mkfs.jfs/. * [20] Linux kernel documentation: NTFS3. https://docs.kernel.org/filesystems/ntfs3.html. * [21] NTFS-3G driver. https://github.com/tuxera/ntfs-3g. * [22] Xfs filesystem (man page). https://manpages.ubuntu.com/manpages/trusty/man8/mkfs.xfs.8.html. * [23] ZFS on Linux project. https://zfsonlinux.org/. * [24] Git’s patch for CVE-2021-21300. https://github.com/git/git/commit/684dd4c2b414bcf648505e74498a608f28de4592. * [25] Norman Hardy. The confused deputy (or why capabilities might have been invented). ACM SIGOPS Operating Systems Review, 22:36–38, 1994. * [26] Daniel Kachakil. Multiple vulnerabilities in android’s download provider (CVE-2018-9468, CVE-2018-9493, CVE-2018-9546). https://ioactive.com/multiple-vulnerabilities-in-androids-download-provider-cve-2018-9468-cve-2018-9493-cve-2018-9546/. * [27] Maxwell Krohn, Alexander Yip, Micah Brodsky, Natan Cliffer, M Frans Kaashoek, Eddie Kohler, and Robert Morris. Information flow control for standard os abstractions. ACM SIGOPS Operating Systems Review, 41(6):321–334, 2007. * [28] James A Kupsch and Barton P Miller. How to open a file and not get hacked. In 2008 Third International Conference on Availability, Reliability and Security, pages 1196–1203. IEEE, 2008. * [29] Yu-Tsung Lee, William Enck, Haining Chen, Hayawardh Vijayakumar, Ninghui Li, Zhiyun Qian, Daimeng Wang, Giuseppe Petracca, and Trent Jaeger. PolyScope: Multi-Policy access control analysis to compute authorized attack operations in android systems. In 30th USENIX Security Symposium (USENIX Security 21), pages 2579–2596, 2021. * [30] Kyung-Suk Lhee and Steve J Chapin. Detection of file-based race conditions. International Journal of Information Security, 4(1):105–119, 2005\. * [31] Linus torvalds’s comments on case-insensitive file systems<EMAIL_ADDRESS> * [32] Eliminating android wrapfs “hackery”. https://lwn.net/Articles/718640/. * [33] mm: shmem: Add case-insensitive support for tmpfs. https://lwn.net/Articles/850214/. * [34] Case-insensitive ext4. https://lwn.net/Articles/784041/. * [35] Filesystems and case-insensitivity. https://lwn.net/Articles/772960/. * [36] Case-insensitive filesystem lookups. https://lwn.net/Articles/754508/. * [37] Network filesystem topics. https://lwn.net/Articles/685431/. * [38] Slava Makkaveev. Man-in-the-Disk: Android apps exposed via external storage. https://research.checkpoint.com/2018/androids-man-in-the-disk/. * [39] William S. McPhee. Operating system integrity in os/vs2. IBM Systems Journal, 13(3):230–252, 1974. * [40] OpenWall Project - information security software for open environments. http://www.openwall.com/. * [41] Jongwoon Park, Gunhee Lee, Sangha Lee, and Dong-Kyoo Kim. RPS: An extension of reference monitor to prevent race-attacks. In Pacific-Rim Conference on Multimedia, pages 556–563. Springer, 2004. * [42] Mathias Payer and Thomas R Gross. Protecting applications against tocttou races by user-space caching of file metadata. In Proceedings of the 8th ACM SIGPLAN/SIGOPS conference on Virtual Execution Environments, pages 215–226, 2012. * [43] Donald E Porter, Owen S Hofmann, Christopher J Rossbach, Alexander Benn, and Emmett Witchel. Operating system transactions. In Proceedings of the ACM SIGOPS 22nd symposium on Operating systems principles, pages 161–176, 2009. * [44] Calton Pu and Jinpeng Wei. A methodical defense against tocttou attacks: The edgi approach. In Proceedings of the 2006 International Symposium on Secure Software Engineering, 2006. * [45] Samba: Implementation of smb/cifs protocol. https://www.samba.org/. * [46] smb.conf.5 (man). https://www.samba.org/samba/docs/4.15/man-html/smb.conf.5.html. * [47] Jonathan S Shapiro, Jonathan M Smith, and David J Farber. EROS: A fast capability system. In Proceedings of the seventeenth ACM symposium on Operating systems principles, pages 170–185, 1999. * [48] Case sensitive filesystems not supported on mac. https://help.steampowered.com/en/faqs/view/0395-A862-13F3-6E82. * [49] SteamOS. https://store.steampowered.com/steamos. * [50] Dan Tsafrir, Tomer Hertz, David A Wagner, and Dilma Da Silva. Portably solving file tocttou races with hardness amplification. In FAST, volume 8, pages 1–18, 2008. * [51] Eugene Tsyrklevich and Bennet Yee. Dynamic detection and prevention of race conditions in file accesses. In 12th USENIX Security Symposium (USENIX Security 03), 2003. * [52] Prem Uppuluri, Uday Joshi, and Arnab Ray. Preventing race condition attacks on file-systems. In Proceedings of the 2005 ACM symposium on Applied computing, pages 346–353, 2005. * [53] Hayawardh Vijayakumar, Xinyang Ge, Mathias Payer, and Trent Jaeger. JIGSAW: Protecting resource access by inferring programmer expectations. In 23rd USENIX Security Symposium (USENIX Security 14), pages 973–988, 2014. * [54] Hayawardh Vijayakumar, Joshua Schiffman, and Trent Jaeger. STING: Finding name resolution vulnerabilities in programs. In 21st USENIX Security Symposium (USENIX Security 12), pages 585–599, 2012. * [55] Hayawardh Vijayakumar, Joshua Schiffman, and Trent Jaeger. Process firewalls: Protecting processes during resource access. In Proceedings of the 8th ACM European Conference on Computer Systems, pages 57–70, 2013. * [56] Robert NM Watson, Jonathan Anderson, Ben Laurie, and Kris Kennaway. Capsicum: Practical capabilities for UNIX. In 19th USENIX Security Symposium (USENIX Security 10), 2010. * [57] Wine: Wine is not an emulator. https://www.winehq.org/. * [58] What is the windows subsystem for linux? https://docs.microsoft.com/en-us/windows/wsl/about. * [59] Diving into SDCardFS: How Google’s FUSE Replacement Will Reduce I/O Overhead. https://www.xda-developers.com/diving-into-sdcardfs-how-googles-fuse-replacement-will-reduce-io-overhead/. ## Appendix A Command-line Flags for Utilities ### A.1 tar tar can be used to copy files/directories across different systems. In the first step, tar reads files and directories hierarchies at a given source and creates an archive (also known as a tarball) that contains information about the source contents. Subsequently, tar can recreate the source at another target using the previously created archive. The archive is portable and can be moved across computer systems. To generate Table 2, we used GNU tar (v1.30) with the following command-line flags. -cf create new archive and specify its name -x extract contents from the archive -v verbose mode (does not impact behavior) ### A.2 zip zip is similar to tar. It also creates an intermediate archive to copy files across computer systems. The following flags were used with zip (v3.0) to produce Table 2. -r recursively traverse all directories (during archive creation) –symlinks store symbolic links in zip archive ### A.3 cp and cp* cp is a commonly used utility to copy files from a source to a target. Both the source and target need to be accessible on the same machine. Unlike tar and zip, cp does not create an intermediate archive. The files and directories are directly created at the target by reading the source. The following flags were used with cp (GNU Coreutils v8.30) to generate Table 2. -a do not deference symlinks, recursively copy directories, and preserve attributes (mode, ownership, timestamps, xattr, hardlinks and context) cp and cp* use the same executable binary. The difference is in the way the command-line arguments are passed to the binary. Specifically, the format of specifying the source directories is different. Consider that the source directory (to be copied) is foo. For cp, we will pass it as foo/ while for cp* we will use foo. Note the trailing / is missing in the latter case. Just this difference changes the behavior of cp as noted in Table 2. ### A.4 rsync rsync can be used to copy, transfer, or sync file/directories on the same machine, or across machines (over a custom rsync:// protocol). The following flags were used with rsync (v3.1.3) to generate Table 2. -a recursively copy directories, preserve symlinks, timestamps, DAC permissions, owners, groups, device files and special files -H preserve hardlinks
# $f\left(R\right)$ gravity in the Jordan Frame as a Paradigm for the Hubble Tension Tiziano Schiavone,1,2 Giovanni Montani,3,4 1Department of Physics “E. Fermi", University of Pisa, Polo Fibonacci, Largo B. Pontecorvo 3, I-56127 Pisa, Italy 2INFN, Istituto Nazionale di Fisica Nucleare, Sezione di Pisa, Polo Fibonacci, Largo B. Pontecorvo 3, I-56127 Pisa, Italy 3ENEA, Fusion and Nuclear Safety Department, C.R. Frascati, Via E. Fermi 45, Frascati, I-00044 Rome, Italy 4Physics Department, “Sapienza" University of Rome, P.le Aldo Moro 5, I-00185 Rome, Italy E-mail<EMAIL_ADDRESS><EMAIL_ADDRESS> (Accepted XXX. Received YYY; in original form ZZZ) ###### Abstract We analyze the $f(R)$ gravity in the so-called Jordan frame, as implemented to the isotropic Universe dynamics. The goal of the present study is to show that, according to recent data analyses of the supernovae Ia Pantheon sample, it is possible to account for an effective redshift-dependence of the Hubble constant via the dynamics of a non-minimally coupled scalar field, emerging in the $f(R)$ gravity. We face the question both from an analytical and purely numerical point of view, following the same technical paradigm. We arrive to establish that the expected decay of the Hubble constant with the redshift $z$ is ensured by a form of the scalar field potential, which remains essentially constant for $z\lesssim 0.3$, independently if this request is made a priori, as in the analytical approach, or obtained a posteriori, when the numerical procedure is addressed. Thus, we demonstrate that an $f(R)$ dark energy model is able to account for an apparent variation of the Hubble constant due to the rescaling of the Einstein constant by the $f(R)$ scalar mode. ###### keywords: cosmology: theory – dark energy – cosmological parameters – galaxies: distances and redshifts – supernovae: general ††pubyear: 2022††pagerange: $f\left(R\right)$ gravity in the Jordan Frame as a Paradigm for the Hubble Tension–$f\left(R\right)$ gravity in the Jordan Frame as a Paradigm for the Hubble Tension ## Introduction The measurement of the Hubble constant $H_{0}$ has been the most important and challenging effort of large-scale observations of the present Universe since the very beginning of cosmological studies. For many decades, the value of $H_{0}$ has been determined with a very low degree of precision. However, since the beginning of the new century, the emergence of the so-called ’precision cosmology’ has allowed the determination of $H_{0}$ with an appropriate degree of accuracy, and the possibility to test an increasingly large set of cosmological parameters became concrete. Nowadays, a large number of different and independent measurements of $H_{0}$ are available (Di Valentino et al., 2021) and also other crucial cosmological indicators can be found, like the position of peaks in the cosmic microwave background (CMB) data thanks to the Planck satellite (Aghanim et al., 2020). However, the end of the previous century has been characterized by a really non-trivial surprise, coming from low-redshift observations of Type Ia supernovae (SNe Ia) recovered as standard candles: the present Universe is accelerating (Riess et al., 1998; Perlmutter et al., 1999). This experimental issue opened a new era on the understanding of the present Universe dynamics and physics, since either a dark energy component living in the cosmological space (identified in a cosmological constant term in the so-called $\Lambda$CDM model (Weinberg, 2008)), or a modified gravity theory in the late expansion phase must be postulated to represent the emerging acceleration phenomenology. In this very puzzling panorama, in recent years, in particular after the observations of the Planck satellite, an additional non-trivial observational evidence came out, called the Hubble tension (Di Valentino et al., 2021), i.e. a discrepancy in 4.9 $\sigma$ between the determination of $H_{0}$ via the CMB data ($H_{0}=67.4\pm 0.5\,\textrm{km s}^{-1}\,\textrm{Mpc}^{-1}$) (Aghanim et al., 2020) and that one coming out by using low-redshift cosmological testers, like the Cepheid-SN Ia sample ($H_{0}=73.04\pm 1.04\,\textrm{km s}^{-1}\,\textrm{Mpc}^{-1}$) (Riess et al., 2022). This discrepancy appears hard to be straightforwardly interpreted. Indeed, the emergence of different values of $H_{0}$ from several cosmological probes can be attributed essentially to two questions only: either the astrophysical characterization of the tester is inadequate (for instance, because their calibration or redshift evolution are not properly fixed) or new physics, in addition to the dark energy Universe component, must be considered for describing the late phase of expansion. In both these cases, the resulting observational or physical picture seems to require a deed revision. The analysis pursued by the SNe Ia community, mainly represented by the data of the Pantheon (Scolnic et al., 2018) and Pantheon+ (Riess et al., 2022) samples, seem to exclude the existence of a redshift evolution of these objects and show a reliable control of all the main sources of errors. However, some recent studies may report a possible redshift dependence of the marginalized absolute magnitude or the Hubble constant itself (Krishnan et al., 2020, 2021; Kazantzidis & Perivolaropoulos, 2020; Dainotti et al., 2021, 2022; Krishnan & Mondol, 2022; Schiavone et al., 2022; Colgáin et al., 2022). In particular, in the two analyses by Dainotti et al. (2021, 2022) it is argued that a binned distribution of the SNe Ia with equipopulated bins outlines a variation of $H_{0}$ as $(1+z)^{-\alpha}$ within 2 $\sigma$ confidence level, where the power exponent $\alpha$ is of the order of few percents. Furthermore, in Dainotti et al. (2021), it has been stressed that the extrapolation of the behavior of $H_{0}(z)$ up to the CMB redshift seems to naturally account for the Hubble tension feature, since a larger value of the Hubble constant today would slowly decrease to higher redshifts. Several alternative proposals have been addressed to explain the $H_{0}$ tension. First of all, a local underdensity of the Universe could account for the discrepancy between low redshift observations and the Planck data, simply because the latter are referred to the (larger) average density of the Universe, according to the original Hubble-bubble model (Zehavi et al., 1998). However, these proposals are not quite working, since we would need an enormous local underdensity of the actual Universe, which is not currently observed (Kenworthy et al., 2019; Camarena et al., 2022) and would affect critically photon paths. We can note that the Hubble tension is a serious puzzle because the observed discrepancy is a significant effect when stated on dynamical level. Moreover, to account for it, we can not alter other well- established observations. Another possibility to explain the Hubble tension is provided by modified gravity theories, as applied to the late Universe. In particular, in Dainotti et al. (2021), it was argued that the observed dependence of $H_{0}(z)$, predicted by the binning analysis of the Pantheon sample, can be interpreted as a variation of the Einstein constant, which can be naturally achieved with the $f(R)$ gravity theory (Nojiri & Odintsov, 2007; Starobinsky, 2007; Tsujikawa, 2008; Sotiriou & Faraoni, 2010) in the Jordan frame (Faraoni & Capozziello, 2011) via the non-minimally coupled scalar field related to the $f$ functional form. However, in Dainotti et al. (2022) it was shown how one of the most reliable dark energy models, the Hu-Sawicki proposal (Hu & Sawicki, 2007), is inappropriate to reproduce the desired effect, i.e. the resulting luminosity distance does not correspond to a restated constant $H_{0}$ value across the whole Pantheon sample. This negative result suggested the necessity to construct ad hoc a dark energy model able to account for the Universe acceleration, and at the same time responsible for the apparent variation $H_{0}(z)$. The present letter is dedicated to constructing a model having the requirements stated above, constructed following the prescriptions emerging from the binning analysis in Dainotti et al. (2021, 2022). We start from the evidence of a decreasing trend of $H_{0}(z)$ associated with the above sketched power-law decay with $z$, and then we construct the profile of the potential term of the non-minimally coupled scalar field $\phi$, which fixes the morphology of the required $f(R)$ theory. The analysis is divided into two parts: the first one is characterized by an analytical approach, while the second one relies on a pure numerical study. The analytical formulation starts with the assumption that a flat region of the potential exists, and then the deviation from this constant value (i.e. the global form of the scalar field potential) is calculated as the dynamics is defrozen, i.e. in correspondence to increasingly large redshifts. Unlike the analysis in Dainotti et al. (2021, 2022), we do not use an ansatz for $\phi(z)$, but we obtain it from the cosmological dynamics in the Jordan frame. In the numerical analysis, we assume again the evolution of $H_{0}(z)$, but we relax the request of a constant potential term in a given region. It is remarkable that both the analytical and the numerical formulations are consistent and predict a flat potential profile in a region $0<z\lesssim 0.3$, associated with a slow-rolling-like dynamics and a dark energy candidate. An important consistency check for the model formulation is provided by the determination of the $f(R)$ form in the limit of the low-redshift approximation. We get three contributions: a cosmological constant, a linear contribution in the Ricci scalar $R$, and eventually a quadratic correction as in the $R^{2}$-gravity theory (Starobinsky, 1980). It is worth emphasizing that this modified theory reduces to the standard $\Lambda$CDM model if the function $H_{0}(z)$ is frozen to a constant value and $df/dR\equiv 1$ today. The present study shows that the emerging $f(R)$ gravity in the Jordan frame can account for the behavior $H_{0}(z)$ within the whole redshift range of the Pantheon sample and, as it must be, also reproduces a reliable explanation for the present Universe acceleration. This work is organized as follows: in Sect. 1 we briefly introduce the $f(R)$ modified gravity in the Jordan frame within the framework of a homogeneous and isotropic Universe; in Sect. 2 we derive the scalar field potential, inferred from a running Hubble constant with the redshift; in Sect. 3 we show our numerical solutions; in Sect. 4 we obtain the functional form of $f(R)$ in the low-redshift limit; in Sect. 5 we summarize our key findings. The metric signature adopted here is $\left(-,+,+,+\right)$, and the speed of light is $c=1$. The Newton constant is denoted with $G$, while the Einstein constant is defined as $\chi\equiv 8\,\pi\,G$. ## 1 $f\left(R\right)$ gravity in the Jordan frame for a homogeneous and isotropic Universe The $f(R)$ modified gravity theories provide a useful extension of General Relativity (GR) up to cosmological scales. An extra scalar degree of freedom with respect to GR occurs in this environment, since the gravitational Lagrangian density is given by a function $f(R)$ of the Ricci scalar $R$. The $f(R)$ metric formalism involves higher-order derivatives in field equations up to the fourth-order with respect to the metric, but the Einstein gravitational field equations in GR are recovered if $f\left(R\right)=R$. The $f(R)$ proposal can be reformulated using a dynamically equivalent action in the so-called Jordan frame: $S_{J}=\frac{1}{2\,\chi}\,\int d^{4}x\,\sqrt{-g}\,\left[\phi\,R-V\left(\phi\right)\right]+S_{M}\left(g_{\mu\nu},\psi\right)\,,$ (1) where $g$ is the determinant of the metric tensor, $S_{M}$ is the matter action with matter fields $\psi$. Note that in the Jordan frame the additional degree of freedom is provided by the scalar field $\phi$, which is related to $f(R)$ through the relation $\phi=f^{\prime}\left(R\right)=df/dR$, and it is controlled by the scalar field potential $V\left(\phi\right)=\phi\,R\left(\phi\right)-f\left(R\left(\phi\right)\right)$. Considering a flat Friedmann-Lemaitre-Robertson-Walker (FLRW) metric (Weinberg, 2008), the field equations in the Jordan frame are the following: $\displaystyle H^{2}=\frac{\chi\,\rho}{3\,\phi}+\frac{V\left(\phi\right)}{6\,\phi}-H\,\frac{\dot{\phi}}{\phi}$ (2a) $\displaystyle\frac{\ddot{a}}{a}=-\frac{\chi}{6\,\phi}\,\left(\rho+3P\right)+\frac{V\left(\phi\right)}{6\,\phi}-\frac{H}{2}\,\frac{\dot{\phi}}{\phi}-\frac{1}{2}\,\frac{\ddot{\phi}}{\phi}$ (2b) $\displaystyle 3\ddot{\phi}-2\,V\left(\phi\right)+\phi\,\frac{dV}{d\phi}+9\,H\,\dot{\phi}=\chi\,\left(\rho-3P\right),$ (2c) where $\dot{}=\frac{d}{dt}$, being $t$ the cosmic time in the synchronous gauge, $H(t)$ the Hubble parameter, $\rho\left(t\right)$ and $P\left(t\right)$ the energy density and pressure of a cosmological fluid, respectively. Eq. (2a) is the generalized Friedmann equation, Eq. (2b) gives us the modified cosmic acceleration, and Eq. (2c) is the scalar field equation. Moreover, the divergenceless of the stress-energy tensor for a perfect fluid gives $\dot{\rho}+3\,H\,\left(\rho+P\right)\,.$ (3) In particular, Eqs. (2a) and (2c) are the two independent equations regarding the dynamics in the Jordan frame. To solve them, one must also specify the equation of state that for a barotropic fluid is $P\left(\rho\right)=w\,\rho$, according to the equation of state parameter of a cosmological fluid (i.e. $w=0$ for the matter component). The advantage of the Jordan frame is that now field equations contain derivatives with respect to the metric or the scalar field up to second order only, however the scalar field is non-minimally coupled to the metric. As a consequence of the scalar field dynamics, an effective Einstein constant $\chi/\phi$ emerges in the Jordan frame, as it can be seen in Eq. (2a). Cosmological models based on $f(R)$ modified gravity theories have been employed to predict deviations from the $\Lambda$CDM model and provide the cosmic acceleration in late times without a true cosmological constant. ## 2 Approximated solution for the scalar field potential We consider the dynamics of the gravitational field equations in the Jordan frame of the $f(R)$ modified gravity scenario to build the profile of the scalar field potential $V\left(\phi\right)$, assuming the evolution of $H_{0}$ with the redshift $H_{0}\left(z\right)=\frac{\tilde{H}_{0}}{\left(1+z\right)^{\alpha}}\,,$ (4) as a phenomenological result of the redshift binned analysis performed in Dainotti et al. (2021, 2022). Hence, the respective Hubble parameter in a flat almost like $\Lambda$CDM model is given by $H\left(z\right)=\frac{\tilde{H}_{0}}{\left(1+z\right)^{\alpha}}\,\sqrt{\Omega_{m0}\left(1+z\right)^{3}+1-\Omega_{m0}}\,,$ (5) where $\Omega_{m0}$ is the cosmological density parameter for the matter component. Moreover, we focus on a cosmological dust (pressureless matter $P=0$) in the late Universe, hence we neglect relativistic components, and $\rho\propto\left(1+z\right)^{3}$ by solving the continuity equation (3). Without loss of generality, we now set $\phi=F\left(z\right)$, so that Eq. (2a) rewrites $H^{2}=\frac{1}{F-\left(1+z\right)\,F^{\prime}}\frac{\chi}{3}\left(\rho+\frac{V\left(F\right)}{2\,\chi}\right)\,,$ (6) where $F^{\prime}\equiv dF/dz$. We used the definition of redshift $a_{0}/a=1+z$ with the standard assumption that the scale factor today is $a_{0}=1$, and also the fact that $dz/dt=-\left(1+z\right)\,H\left(z\right)$. In order to rewrite the generalized Friedmann equation in a form similar to the respective one in GR and discuss the $\Lambda$CDM limit, we assume the existence of a region in which $V\left[F\left(z\right)\right]\approx 2\chi\rho_{\Lambda}=\textrm{const}$ for $0<z\lesssim z^{*}$, where $\rho_{\Lambda}$ is the present value of the Universe dark energy density, and $z^{*}\sim 0.3$ is the redshift of matter-dark energy equality. Hence, we rewrite Eq. (6) as $H^{2}=\frac{1}{F-\left(1+z\right)\,F^{\prime}}\frac{\chi}{3}\left(\rho+\rho_{\Lambda}\right)\,,$ (7) where you can recognize the usual terms in the Friedmann equation in the $\Lambda$CDM scenario, unless a factor related to the scalar field $\phi$. Then, since we look for a rescaling of the Hubble constant by a factor $\left(1+z\right)^{-\alpha}$, comparing Eqs. (5) and (7), it is natural to require the following condition $F-\left(1+z\right)\,F^{\prime}=\left(1+z\right)^{2\alpha}\,,$ (8) which admits the solution: $\phi\left(z\right)=F\left(z\right)=\frac{\left(1+z\right)^{2\alpha}}{1-2\,\alpha}+\left(K-\frac{1}{1-2\,\alpha}\right)\,\left(1+z\right)\,.$ (9) We fixed the initial condition $\phi\left(0\right)=K$ at $z=0$, where $K$ denotes the deviation from a pure GR scenario ($\phi=1$). Let us now take into account the scalar field equation (2c). Using the relation $\phi=F\left(z\right)$ and the approximation for the scalar field potential $V\left[F\left(z\right)\right]\approx 2\chi\rho_{\Lambda}$, we obtain: $\displaystyle 3\,H^{2}\,\left[\left(1+z\right)^{2}\,F^{\prime\prime}-\left(1+z\right)\,F^{\prime}\right]-3\,\left(1+z\right)\,\frac{\ddot{a}}{a}\,F^{\prime}+F\,\frac{dV}{dF}=$ $\displaystyle=\chi\,\left(\rho+4\rho_{\Lambda}\right)\,.$ (10) It should be emphasized that we do not neglect the term $dV/dF$, since we want to check a posteriori the viability of the approximation for the scalar field potential at low redshifts. Furthermore, by substituting the term $\ddot{\phi}$ from Eq. (2c) in Eq. (2b), we have $\frac{\ddot{a}}{a}=-\frac{\chi}{3\,F}\,\left(\rho+\rho_{\Lambda}\right)+\frac{1}{6}\frac{dV\left(F\right)}{dF}-H^{2}\,\left(1+z\right)\frac{F^{\prime}}{F}\,.$ (11) Then, we combine Eqs. (10) and (11), and we employ the solution for $\phi\left(z\right)$ given by Eq. (9). Finally, we obtain an expression for $\displaystyle\frac{d\tilde{V}}{dz}=\frac{6\left[1+z-2\alpha\left(1+z\right)^{2\alpha}-K\left(1+z\right)\left(1-2\alpha\right)\right]}{\Omega_{m0}\left(1+z\right)\left[1+z+2\left(1+z\right)^{2\alpha}\left(\alpha-1\right)+K\left(1+z\right)\left(2\alpha-1\right)\right]}$ $\displaystyle\times\left[2\left(2+\alpha\right)+\Omega_{m0}\left(z\left(1+2\alpha\right)\left(3+z\left(3+z\right)\right)-3\right)\right]\,,$ (12) where we rescaled the potential as a dimensionless quantity $\tilde{V}\equiv V/m^{2}$ with the constant $m^{2}\equiv\chi\rho_{0}/3=\tilde{H}_{0}^{2}/\Omega_{m0}$. We denote with $\rho_{0}$ the actual matter density, and we used the definitions of the critical energy density of the Universe today $\rho_{c0}=3\tilde{H}_{0}^{2}/\chi$ and also of the cosmological density parameter $\Omega_{m0}=\rho_{0}/\rho_{c0}$. Note that Eq. (12) cannot be solved analytically, but we need to proceed numerically to obtain $\tilde{V}(z)$. We evaluate the numerical integral of Eq. (12) for different redshifts within the range of the Pantheon sample, i.e. $0<z<2.3$, and we build a grid with a step $\Delta z=10^{-3}$. If $V\left[F\left(z\right)\right]=2\chi\rho_{\Lambda}$ at $z=0$, then it is easy to show we need to add the following integration constant $\tilde{V}(0)=6\left(1-\Omega_{m0}\right)/\Omega_{m0}$, when we integrate Eq. (12). We fixed the values for $\alpha=9\cdot 10^{-3}$ (Dainotti et al., 2021), $\Omega_{m0}=0.298$ (Scolnic et al., 2018), and $K=1-10^{-7}$ (Hu & Sawicki, 2007). After obtaining different points as a result of numerical integrals, we use a polynomial fit up to the eighth-order in the redshift to calculate $\tilde{V}(z)$. The evolution of $\phi(z)$, given by Eq. (9), is plotted in Fig. 1 with a blue line. As a final check, in Fig. 2 you can see the resulting profile of $\tilde{V}$, which appears flat for $0<z\lesssim z^{*}$ in the dark-energy- dominated era, according to our preliminary assumptions on the flatness of $V(z)$ for low redshifts. ## 3 Numerical analysis of the model We now relax the assumption on the existence of a flat region of the scalar field potential for $0<z\lesssim z^{*}$, but we continue to consider the evolution of $H_{0}(z)$ given by Eq. (4). Let us proceed with a complete numerical analysis to solve the cosmological dynamics in the Jordan frame (Sect. 1) in the presence again of a matter component, in which we have two independent equations (2a) and (2c), and two unknown quantities, i.e. $\phi(z)$ and $V\left[\phi(z)\right]$. Firstly, we rewrite the generalized Friedmann equation (2a) in the variable $z$, isolating the dimensionless scalar field potential $\displaystyle\tilde{V}\left[\phi\left(z\right)\right]$ $\displaystyle=6\,\left[\frac{\phi(z)}{\left(1+z\right)^{2\alpha}}-\left(1+z\right)^{1-2\alpha}\,\frac{d\phi}{dz}\right]\left[\left(1+z\right)^{3}+\frac{1-\Omega_{mo}}{\Omega_{m0}}\right]$ $\displaystyle-6\left(1+z\right)^{3}\,,$ (13) where we used Eq. (5) and the fact that $\rho\sim\left(1+z\right)^{3}$. Secondly, we rewrite the scalar field equation (2c) likewise: $\displaystyle\left[\left(1+z\right)^{2-2\alpha}\frac{d^{2}\phi}{dz^{2}}-\left(2+\alpha\right)\left(1+z\right)^{1-2\alpha}\frac{d\phi}{dz}\right]\left[\left(1+z\right)^{3}+\frac{1-\Omega_{m0}}{\Omega_{m0}}\right]$ $\displaystyle+\frac{3}{2}\frac{d\phi}{dz}\left(1+z\right)^{4-2\alpha}-\frac{2}{3}\tilde{V}\left[\phi(z)\right]+\frac{\phi(z)}{3}\left(\frac{d\phi}{dz}\right)^{-1}\frac{d\tilde{V}}{dz}=\left(1+z\right)^{3}\,.$ (14) Then, by substituting $\tilde{V}\left[\phi(z)\right]$ from Eq. (13) into Eq. (14), we obtain a second-order differential equation in $\phi(z)$. We solve numerically this equation with the following initial conditions for $z=0$: $\phi(0)=K$, and $d\phi/dz\,(0)=-10^{-3}$, where we consider a slowly varying scalar field. We fixed the same values for $\alpha$, $\Omega_{m0}$, and $K$ adopted in Sect. 2. In Fig. 1 we show the evolution of $\phi$ with $z$ using a red line, while in Fig. 2 we plot the profile of $\tilde{V}$ in terms of $z$ and $\phi$. In all these figures, we also compare our numerical results with the respective profiles obtained from the approximated solution based on the assumption of a flat potential at low redshifts in Sect. 2, noting that corresponding solutions are very close for $z\ll 1$. It should be noted that the potential $\tilde{V}$ exhibits a flat profile for $0<z\lesssim z^{*}$ also for the numerical solution. Figure 1: Behavior of the scalar field $\phi$ VS redshift $z$ in the Jordan frame, assuming the phenomenological decreasing trend of $H_{0}\left(z\right)$ with redshift in Eq. (4). The blue line is referred to the approximated solution developed in Sect. 2, where we started from the hypothesis of a flat scalar field potential for low redshifts and $\phi(z)$ is given by Eq. (9), while the red line is obtained from the numerical analysis discussed in Sect. 3, after solving Eqs. (13) and (14). Figure 2: Profile of the scalar field potential in terms of the redshift $z$ (top panel) and the scalar field $\phi$ (bottom panel) in the Jordan frame, inferred from the assumption of a running Hubble constant $H_{0}\left(z\right)$, according to Eq. (4). Note that $\tilde{V}=V(\phi)/m^{2}$ is a dimensionless potential. The blue line is referred to the approximated solution developed in Sect. 2, while the red line is obtained from the numerical analysis discussed in Sect. 3. The potential is flat for $0<z\lesssim z^{*}$ for both these solutions. ## 4 The low-redshift $f(R)$ profile We are interested in obtaining an analytical expression for the scalar field potential and the respective $f(R)$ function, which provides both the late- time cosmic acceleration and a running Hubble constant with the redshift, according to Eq. (4). To this end, we expand the solution for $\phi(z)$ and the term $d\tilde{V}/dz$ in the limit of low redshifts for $z\ll 1$ up to the linear order in $z$. More specifically, starting from Eq. (9) and Eq. (12), we get $\displaystyle\phi(z)\approx K+\left(K-1\right)\,z\,,$ (15) $\displaystyle\frac{d\tilde{V}}{dz}\approx A_{1}+2\,A_{2}\,z\,,$ (16) respectively, where the constants $A_{1}$ and $A_{2}$ are defined as $\displaystyle A_{1}=6\,\frac{K-1}{K+1}\,\frac{4+2\alpha-3\Omega_{m0}}{\Omega_{m0}}\,,$ (17) $\displaystyle A_{2}=-6\,\frac{\alpha+2}{\Omega_{m0}}\,\frac{\left(K-1\right)^{2}+4\alpha K}{\left(K+1\right)^{2}}\,,$ (18) Looking at Eq. (16), we can integrate it analytically to write $\tilde{V}(z)$ in the low-redshift limit: $\tilde{V}(z)\approx A_{0}+A_{1}\,z+A_{2}\,z^{2}\,,$ (19) where the integration constant $A_{0}$ is set to be $A_{0}=6\frac{1-\Omega_{m0}}{\Omega_{m0}}\,,$ (20) assuming $V\left[F\left(0\right)\right]=2\chi\rho_{\Lambda}$. Note that $A_{0}$, $A_{1}$, and $A_{2}$ depend on the values of $\alpha$, $K$, and $\Omega_{m0}$. By inverting the relation (15) and substituting $z\left(\phi\right)$ in Eq. (19), we can easily write the scalar field potential $\tilde{V}\left(\phi\right)\approx A_{0}+A_{1}\,\frac{\phi-K}{K-1}+A_{2}\,\left(\frac{\phi-K}{K-1}\right)^{2}\,.$ (21) It should be noted that $\tilde{V}\left(\phi\right)$ has constant, linear, and quadratic contributions in $\phi$. Once we have the expression for $\tilde{V}\left(\phi\right)$, we use the equation $R=dV/d\phi$, which was obtained by varying the action in Eq. (1) with respect to $\phi$. Then, we end up in a relation $R=R\left(\phi\right)$, which admits the existence of the inverse $\phi=\phi\left(R\right)$. Finally, using the relation in the Jordan frame $f\left(R\right)=R\,\phi\left(R\right)-V\left[\phi\left(R\right)\right]$, we obtain $f\left(R\right)\approx B_{0}+B_{1}\,\frac{R}{m^{2}}+B_{2}\,\frac{R^{2}}{m^{2}}\,,$ (22) where we have defined the constants $\displaystyle B_{0}=m^{2}\,\left(\frac{A_{1}^{2}}{4A_{2}}-A_{0}\right)\qquad\quad B_{1}=m^{2}\,\left[K-\frac{A_{1}\,\left(K-1\right)}{2\,A_{2}}\right]$ (23) $\displaystyle B_{2}=\frac{\left(K-1\right)^{2}}{4\,A_{2}}$ (24) It should be stressed that Eq. (22) provides an approximated solution of the $f\left(R\right)$ function for $z\ll 1$, which contains constant, linear, and quadratic terms in $R$. Note that if $K\rightarrow 1$ and $\alpha\rightarrow 0$, we recover the $\Lambda$CDM model. ## 5 Conclusions We started our analysis from the results obtained by Dainotti et al. (2021, 2022), which outlined a dependence of the value of $H_{0}$ with the redshift via a binned data analysis of the SNe Ia Pantheon sample within 2 $\sigma$. The specific form of the decaying $H_{0}(z)$ given in Eq. (4) was the phenomenological input of our theoretical study. The idea proposed above consists in setting up a dark energy model that is also able to account for a variation with $z$ of the $H_{0}$ value. More specifically, we adopted the theoretical paradigm of a $f(R)$ extended theory of gravity, as viewed in the Jordan frame (Sect. 1). Thus, the non-minimally coupled scalar field dynamics was implemented to account for an effective variation of the Einstein constant, just the required feature to account the phenomenology predicted by the binned analysis. Starting from the basic equations for an isotropic Universe in the proposed theoretical framework, we assumed, without any loss of generality, that the scalar field is a function of the redshift, and we determined such specific relation in Eq. (9) by imposing the desired decaying of $H_{0}(z)$. Then, the scalar field dynamics was able to provide us the corresponding form of the scalar field potential term, which fixes the adopted $f(R)$ model. This picture has been performed both analytically and numerically: in the former case, in Sect. 2 we assumed the existence of flat region of the scalar field potential, there approximated via constant value, and then we pursued the explicit determination of the potential term derivative in Eq. (12); in the latter case, the scheme was implemented directly on the two basic equations (2a) and (2c) without any assumption on the potential form. It was rather remarkable that, in both analyses, the potential term singled out a flat region for $z\lesssim 0.3$ (Fig. 2), which is exactly when the dark energy contribution of the Universe dominates on the matter content. It is worth noting how the low-redshift limit of our model in Sect. 4 allowed an analytical determination of the potential term, and hence of the underlying $f(R)$ model. The resulting expression (22) for the modified Lagrangian of the modified gravity contains a cosmological constant, and also linear and quadratic contributions in the Ricci scalar. In particular, this result in the low-redshift limit is consistent with other $f\left(R\right)$ proposals based on quadratic gravity or the $f\left(R\right)$ power series expansion to describe deviations from GR and the $\Lambda$CDM cosmological model in the late Universe without introducing dark energy (Starobinsky, 1980; Sotiriou & Faraoni, 2010; Cosmai et al., 2016; Fanizza et al., 2020). It is very remarkable for the robustness of our proposed model that this modified scheme approaches the $\Lambda$CDM scenario when $\alpha\rightarrow 0$ and $df/dR\rightarrow 1$. In other words, even if we reduce the function $H_{0}(z)$ to a fixed constant value, our model can still contain a small deviation from the $\Lambda CDM$ Universe, according to its nature of a new proposal for explaining the dark energy phenomenon. Thus, we can claim that our study is able to simultaneously address two key points: on one hand, we get a modified gravity model as a suitable dark energy candidate; on the other hand, we provided a natural interpretation for the profile of $H_{0}(z)$ obtained in [nostri articoli], by identifying in the proposed modified gravity approach the lacking physics in the emerging picture from the Pantheon sample. The present study calls attention to further investigations on the proposed model as the redshift increases towards the CMB observations, in order to understand if it can satisfactory solve the Hubble tension. ## References * Aghanim et al. (2020) Aghanim N., et al., 2020, Astron. Astrophys., 641, A6 * Camarena et al. (2022) Camarena D., Marra V., Sakr Z., Clarkson C., 2022, Classical and Quantum Gravity, 39, 184001 * Colgáin et al. (2022) Colgáin E. Ó., Sheikh-Jabbari M. M., Solomon R., Dainotti M. G., Stojkovic D., 2022, arXiv e-prints, p. arXiv:2206.11447 * Cosmai et al. (2016) Cosmai L., Fanizza G., Tedesco L., 2016, International Journal of Theoretical Physics, 55, 754 * Dainotti et al. (2021) Dainotti M. G., De Simone B., Schiavone T., Montani G., Rinaldi E., Lambiase G., 2021, The Astrophysical Journal, 912, 150 * Dainotti et al. (2022) Dainotti M. G., De Simone B., Schiavone T., Montani G., Rinaldi E., Lambiase G., Bogdan M., Ugale S., 2022, Galaxies, 10, 24 * Di Valentino et al. (2021) Di Valentino E., et al., 2021, Class. Quant. Grav., 38, 153001 * Fanizza et al. (2020) Fanizza G., Franchini G., Gasperini M., Tedesco L., 2020, General Relativity and Gravitation, 52, 111 * Faraoni & Capozziello (2011) Faraoni V., Capozziello S., 2011, Beyond Einstein Gravity: A Survey of Gravitational Theories for Cosmology and Astrophysics. Springer, Dordrecht, doi:10.1007/978-94-007-0165-6 * Hu & Sawicki (2007) Hu W., Sawicki I., 2007, Phys. Rev. D, 76, 064004 * Kazantzidis & Perivolaropoulos (2020) Kazantzidis L., Perivolaropoulos L., 2020, Phys. Rev. D, 102, 023520 * Kenworthy et al. (2019) Kenworthy W. D., Scolnic D., Riess A., 2019, The Astrophysical Journal, 875, 145 * Krishnan & Mondol (2022) Krishnan C., Mondol R., 2022, arXiv e-prints, p. arXiv:2201.13384 * Krishnan et al. (2020) Krishnan C., Colgáin E. O., Ruchika Sen A. A., Sheikh-Jabbari M. M., Yang T., 2020, Phys. Rev. D, 102, 103525 * Krishnan et al. (2021) Krishnan C., Colgáin E. O., Sheikh-Jabbari M. M., Yang T., 2021, Phys. Rev. D, 103, 103509 * Nojiri & Odintsov (2007) Nojiri S., Odintsov S. D., 2007, International Journal of Geometric Methods in Modern Physics, 04, 115 * Perlmutter et al. (1999) Perlmutter S., et al., 1999, Astrophys. J., 517, 565 * Riess et al. (1998) Riess A. G., et al., 1998, Astron. J., 116, 1009 * Riess et al. (2022) Riess A. G., et al., 2022, Astrophys. J. Lett., 934, L7 * Schiavone et al. (2022) Schiavone T., Montani G., Dainotti M. G., De Simone B., Rinaldi E., Lambiase G., 2022, in 17th Italian-Korean Symposium on Relativistic Astrophysics. (arXiv:2205.07033) * Scolnic et al. (2018) Scolnic D. M., et al., 2018, Astrophys. J., 859, 101 * Sotiriou & Faraoni (2010) Sotiriou T. P., Faraoni V., 2010, Reviews of Modern Physics, 82, 451 * Starobinsky (1980) Starobinsky A., 1980, Physics Letters B, 91, 99 * Starobinsky (2007) Starobinsky A. A., 2007, JETP Lett., 86, 157 * Tsujikawa (2008) Tsujikawa S., 2008, Phys. Rev. D, 77, 023507 * Weinberg (2008) Weinberg S., 2008, Cosmology. Oxford University Press * Zehavi et al. (1998) Zehavi I., Riess A. G., Kirshner R. P., Dekel A., 1998, The Astrophysical Journal, 503, 483
# The binary and the disk: the beauty is found within NGC3132 with JWST Raghvendra Sahai Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA 91109, USA Valentin Bujarrabal Observatorio Astroómico Nacional (OAN/IGN), Ap 112, 28803 Alcalá de Henares, Spain Guillermo Quintana-Lacaci Dept. of Molecular Astrophysics, IFF-CSIC. C/ Serrano 123, E-28006, Madrid, Spain Nicole Reindl Institut für Physik und Astronomie, Universität Potsdam, Karl-Liebknecht-Stra$\beta$e 24/25, D-14476, Potsdam, Germany Griet Van de Steene Royal Observatory of Belgium, Astronomy and Astrophysics, Ringlaan 3, 1180 Brussels, Belgium Carmen Sánchez Contreras Centro de Astrobiologia (CAB), CSIC-INTA. Postal address: ESAC, Camino Bajo del Castillo s/n, 28692, Villanueva de la Cañada, Madrid, Spain Michael E. Ressler Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA 91109, USA ###### Abstract The planetary nebula (PN) NGC 3132 is a striking example of the dramatic but poorly understood, mass-loss phenomena that $(1-8)$ $M_{\odot}$ stars undergo during their death throes as they evolve into white dwarfs (WDs). From an analysis of JWST multiwavelength ($0.9-18$ µm) imaging of NGC 3132, we report the discovery of an asymmetrical dust cloud around the WD central star (CS) of NGC 3132, seen most prominently in the 18 µm image, with a surface-brightness limited radial extent of $\gtrsim 2{{}^{\prime\prime}}$. We show that the A2V star located $1\farcs 7$ to CS’s North-East (and 0.75 kpc from Earth) is gravitationally-bound to the latter, by the detection of relative orbital angular motion of $0.24\arcdeg\pm 0.045\arcdeg$ between these stars over $\sim$20 yr. Using aperture photometry of the CS extracted from the JWST images, together with published optical photometry and an archival UV spectrum, we have constructed the spectral-energy distribution (SED) of the CS and its extended emission over the UV to mid-IR ($0.091-18$ µm) range. We find that fitting the SED of the CS and the radial intensity distributions at $7.7,12.8$ and $18$ µm with thermal emission from dust requires a cloud that extends to a radius of $\gtrsim$1785 au, with a dust mass of $\sim 1.3\times 10^{-2}$ M⊕, and grains that are 70% silicate and 30% amorphous carbon. We propose plausible origins of the dust cloud and an evolutionary scenario in which a system of three stars – the CS, a close low-mass companion, and a more distant A2V star – forms a stable hierarchical triple system on the main- sequence but becomes dynamically unstable later, resulting in the spectacular mass-ejections that form the current, multipolar PN. circumstellar matter – stars: AGB and post-AGB – stars: individual (NGC 3132) – stars: mass loss – stars: jets cont#1 (cont.)#2#3 ## 1 Introduction In our quest to identify the signatures of extraterrestrial life, the search for planets and planetary systems around stars other than our Sun has become one of the most exciting areas of current astrophysical research. Planets are found to be common around other solar-type stars, but the disks in which these are produced dissipate as the stars reach the main-sequence. A Spitzer 24 µm survey of main-sequence A-type stars shows that up to $\sim$50% of young ($\lesssim$30 Myr) stars have little or no 24 µm excess emission from debris disks; and large debris-disk excesses decrease significantly at ages of $\sim$150 Myr, and much of the dust detected in these is likely produced (episodically) by large planetesimal collisions (Rieke et al., 2005). The dust in these debris disks would have dissipated long before the stars evolve off the main sequence. Thus it is remarkable that once these stars reach the end of their life cycle, observational surveys reveal disks or equatorially- flattened disk-like or toroidal structures (e.g., Hillen et al. (2017); Sahai et al. (2007, 2011)) – i.e., both the birth and the death of stars represent similar highly aspherical states that sandwich the more spherical life of stars on the main sequence (MS). The demise of most stars in the Universe that evolve in a Hubble time (i.e., in the 1–8 $M_{\odot}$ range) is believed to occur as a result of heavy mass- loss (with rates up to $10^{-4}$ $M_{\odot}$ yr-1) on the Asymptotic Giant Branch (AGB), when the stars are very luminous ($L\sim 5000-10,000$ $L_{\odot}$) and cool ($T_{eff}<3000$ K) (see e.g., review by Decin (2021)). After mass-loss – which may be via a quasi-steady wind over $\sim 10^{5}$ yr or rather sudden (if driven by a strong binary interaction, as e.g., in the Boomerang Nebula (Sahai et al., 2017)) – has depleted most of the stellar envelope, the stars evolve to higher temperatures through the post-AGB phase at almost constant luminosity, fading and becoming white dwarfs (WDs) at the ends of their lives. It is during these post-AGB and WD phases, that the renewed presence of puzzling disk-like structures around the central stars becomes obvious. How are such disks formed? Are there multiple mechanisms that can make disks during this phase? Can some of these be sufficiently dense (and long-lived), leading to a second phase of planet-formation? Such disks can generally be divided into two broad observational classes by their sizes – (Type 1) very small disks which are contained within the Roche limits of the central WD stars ($\lesssim 0.01$ AU), and (Type 2) large disks extending to radii of $\sim 10-100$ AU. The majority of the Type 2 disks are found around the central stars of PNe (CSPNe) (e.g., Bilíková et al. (2012)) and around post-AGB stars (e.g., De Ruyter et al. (2006)), whereas most Type 1 disks have been identified around old, naked WDs (i.e. those for which the PN shell is no longer visible). Type 1 dust disks around WDs were first detected around G29-38 and GD 362 through their excess IR emission (confirmed spectroscopically to be continuum), and since their discovery, it has been commonly accepted that they originate from tidally disrupted planetesimals (Becklin et al. (2005); Kilic et al. (2005); Jura (2003)), especially since the chemical composition of these planetesimals is similar to rocky bodies in the inner regions of our Solar System. The required gravitational scattering of planetesimals towards the WD implies the presence of a planet, and recent discoveries provide indirect and direct evidence for these. The WD J091405.30+191412.25 is found to be accreting hydrogen, oxygen, and sulfur, material which likely comes from the deep atmospheric layers of an icy giant planet (Gänsicke et al., 2019). From TESS data, Vanderburg et al. (2020) find a giant planet transiting the WD 1856+354 every 1.4 days. However, tidal shredding of a companion during a Common Envelope phase can also make such a disk (Kuruwita et al., 2016). The origin(s) of Type 2 disks is(are) more difficult to understand. Using Spitzer/MIPS imaging and IRS spectroscopy, Su et al. (2007) find evidence for the presence of a compact dust cloud around the central star of the Helix Nebula (WD 2226-210: $T_{eff}=110,000$ K) with a temperature of 90–130 K, located 40–100 AU from the central star, and inferred it to be a dusty disk. A more extensive search for 24 µm excesses around a sample of CSPNe by (Bilíková et al. (2012); Chu et al. (2011)) revealed spectral-energy- distributions (SEDs) and spectra (for a few objects) with a variety of IR excess characteristics, implying the operation of a different mechanism than planetesimal destruction for producing the dusty disk responsible for the excess emission. This mechanism involves the presence of a binary companion – binaries provide a source of angular momentum and free energy to form accretion discs. Several mechanisms for making disks in binary systems with a mass-losing companion have been identified (including Bondi-Hoyle-Eddington accretion, Roche-lobe and wind-Roche-lobe overflow, common envelope ejection, grazing envelope ejection). Numerical simulations of binary systems where the companion is detached show that the latter can gravitationally capture a fraction of the mass outflow from an AGB star into a small disk (hereafter Type 1b, $\sim$1 AU, Mastrodemos & Morris (1998)) – too small to be representative of Type 2 disks. Thus, the large Type 2 disks, if resulting from binarity, likely require interaction with a close binary companion that results in the ejection of a substantial fraction of the AGB mass outflow being directed along directions near the orbital plane forming a disk-like structure. But it has been observationally difficult to confirm the association of dust disks with binary CSPNe, which requires the detection of companions and disks around the same CSPNe. The spectral-type distribution for companions of WDs (which should be representative of CSPNe), peaks at spectral types M 3-4 (Farihi et al. (2005)). Spectra or sensitive photometry in the $\sim 0.6-5$ µm range can reveal a companion – e.g., an M3V companion would emit $\sim 210(D/0.5\,kpc)^{2}$ $\mu$Jy at 5 µm – or provide upper limits. In this paper, we report the detection of a dust cloud around the central star (CS) of the planetary nebula NGC 3132 using data from the JWST’s Early Release Observations (Pontoppidan et al. (2022)). This PN, also known as the Southern Ring, has been well-studied, with high-resolution optical images obtained with HST (e.g., Monteiro et al. (2000)), 2D spectroscopic imaging (Monreal-Ibero & Walsh (2020)), and mapping of molecular-line emission – 2.1 µm H2 (Storey (1984)) and 1.3 mm CO J=2–1 (Sahai et al. (1990)). These data reveal that NGC 3132 belongs to the primary morphological class “multipolar” (Sahai et al. (2011)) – with a bright central elliptical shell structure surrounded by fainter structures aligned along different axes. Located at the center of the elliptical structure, the CS of NGC 3132 is a hot white dwarf (discovered by Kohoutek & Laustsen (1977)) with $T_{eff}\sim 105,000$ K: Monreal-Ibero & Walsh (2020)). North-East of the CS, at a separation111all separations are as measured on the sky-plane, unless noted otherwise of $1\farcs 696$, lies a companion (HD 87892) that is a slightly evolved main-sequence star of spectral type A2V (Méndez (1978)). We focus in this study on the CS and its immediate surroundings; description and analysis of the full PN morphology is not within the scope of this paper, but has been reported by De Marco et al. (2022). ## 2 Observations We retrieved the pipeline-calibrated Level3 data on NGC 3132 (Proposal 2733, PI=K.M. Pontoppidan) from the MAST archive. NGC 3132 was imaged using filters F090W, F187N, F212N, F356W, F405N, F470N with NIRCam (plate-scale $0\farcs 031$ for $\lambda\lesssim 2.1$ µm and $0\farcs 063$ for longer wavelengths) (Jakobsen et al. (2022)), and F770W, F1130W, F1280W, and F1800W with MIRI (plate-scale $0\farcs 11$) (Bouchet et al. (2015)). JWST is diffraction- limited at wavelengths $\gtrsim$2 µm with a PSF size (FWHM) of $0\farcs 075$ at $\gtrsim$2 µm (Rigby et al. (2022)). The images have been processed and calibrated using data processing/calibration software version nos. “2022_2a” and “1.5.3”. Each of the Level 3 images is a combination of 8 dithered images (hereafter “sub-images”). For NIRCam data, 6/8 sub-images in each of the filters F090W, F187N and F202N include the CS in the FOV; for the remaining filters, the CS is within the FOV for all 8. For MIRI data, 6/8 sub-images in each of the filters include the CS in the FOV. We have examined each of these sub-images to look for artifacts that may affect the CS, and found that only one of these, for F090W, has 4 bad pixels at the location of the CS. Hence, for this filter only, the photometry described in the next section (§ 3) was carried out on each of the 5 good sub-images, and the reported value is a median average of these. ## 3 Observational Results The JWST images of NGC 3132 reveal the well-known, extended multipolar morphology of this object, the hot central star that excites the nebular material, and the nearby A2V star. The images (in selected filters) of the CS and its immediate surroundings, are shown in Fig. 1. A small but clearly extended emission region is seen around the CS in the F1800W image. In addition, the emission is not circularly symmetric in its outer parts (i.e., at radial offsets $\gtrsim 0\farcs 65$) – the contours appear flattened on the side closest to the A2V star, thus giving the emission source an overall elongated shape (Fig. 2a,b). Furthermore, there is a small “tail” at the Northern side of the emission source, that bends towards the A2V star. Radial intensity cuts centered on the CS, averaged over 90° wedges pointing towards and away from the A2V star, clearly show a significant mismatch between the intensities at offsets $\gtrsim 0\farcs 7$ (Fig. 2c). Although there is considerable structure in the foreground/background of the CS due to the PN itself, it is very unlikely that the above asymmetries are the result of a chance projection of a large extended nebular feature coinciding with the central dust clump, distorting its shape and producing such a specific tail shape curved towards the A2V star. Also, no hint of such an extended feature close to the compact dust clump is seen in the large-scale 18 µm image of the nebula (Appendix). A comparison of the CS’s radial intensity with that of relatively-bright field stars in the field-of-view of these images, shows that at 18 µm, the width (FWHM) of the emitting region, $0\farcs 96$, is significantly larger than that of the PSF ($0\farcs 66$). The observed radial extent of the dust cloud is surface-brightness limited – e.g., on the side away from the A2V star, it falls to an intensity equal to the standard deviation of the sky background, $\sim$4 MJy sr-1, at $r\sim 2\farcs 5$. The 12.8, 11.3 & 7.7 µm images also appear mildly extended, but to a progressively lesser extent; for shorter wavelengths, the CS appears to be point-like. ### 3.1 PSF subtraction The CS images are affected by the presence of the bright A2V type star which spreads its flux through its PSF. We describe our 5-step (iterative) PSF subtraction process below: 1. 1. Generate a simulated PSF (“sPSF”) using the PSF simulator tool (webbpsf222https://www.stsci.edu/jwst/science-planning/proposal-planning- toolbox/psf-simulation-tool), taking into account the spectral type of the source (A2V) and its offset from the center of the FOV in the NIRCam and MIRI images. The optics of the system are affected by additional effects (jitter, deformations,…) which are equivalent to an additional gaussian smoothing but are not accounted for in the simulator. We therefore applied a gaussian smoothing function to “sPSF”, whose FWHM was set iteratively (see step 3 below). 2. 2. Remove a median averaged sky-background from the “i2d” image. The value for the box size used for estimating a median background was set iteratively for each filter/camera. 3. 3. Apply the gaussian smoothing function to the sPSF, and scale the sPSF peak (central) intensity to that of the A2V star when the latter was not saturated in the center (F1800W, F1280W, F1130W, F770W), or to the intensity in the wings when the A2V star is saturated in the center (F090W, F187N, F212N, F356W, F405N, F470N). The width of the gaussian was set iteratively for each filter/camera. 4. 4. Compare intensity slices along the vertical and horizontal axis crossing through the center of the A2V star and the sPSF, and check the result of the subtraction of the current sPSF with the background-subtracted image. This allowed us to estimate the background subtraction performed in step 3, as well as re-adjust the gaussian spread applied to the PSF. Also, the registration of the sPSF relative to that of A2V star was re-adjusted, especially important in the NIRCam images for which the central pixels covering the A2V star are saturated. 5. 5. Iterate steps 2–4 (i.e. gaussian smoothing, registration, flux scaling) to bring the average flux in the region close to the A2V star as close to the median error in the sky background in this region. The background estimate and removal was performed by using the photutils333https://photutils.readthedocs.io/en/stable/background.html package (Bradley et al. (2015)). We estimated a 2D median background, by dividing the field of view into boxes and substituting the pixel intensities in each box with the median background for that box. This generated a low-resolution background which was subtracted from the image. The box size was estimated to remove as much background as possible but avoiding any source removal, including the PSF sidelobes. The box size, the width of the Gausian smoothing function, and the error in the PSF scale-factor444The actual value of the scale factor is unimportant, since it depends on the normalisation of the PSF, which is arbitrary are given in Table 3. The images of the CS and its immediate surroundings with the A2V star subtracted, for the same set of selected filters as in Fig. 1, are shown in Fig. 3. Most of the negative residuals in these images lie relatively close to the sharp diffraction structures and most likely are a result of inherent limitations in the PSF simulator tool’s ability to reproduce the observed PSF. ### 3.2 Photometry We have carried out aperture photometry of the CS in the NIRCam and MIRI images as follows. Except for the F1800W image, we have used circular apertures with sizes adjusted to include as much of the light from the CS, and exclude as much light from the surrounding structured background, as possible. For F1800W, we have used multiple methods to extract the photometry. Using the pipeline Level 3 image, we use an elliptical aperture of size $1\farcs 1\times 1\farcs 3$ that roughly matches the observed elongated shape of the source. Second, we extracted the average radial intensity over an angular wedge covering the position angle (PA) range $100\arcdeg-315\arcdeg$ that best excludes the region contaminated by the PSF of the A2V star; the background intensity was estimated from the cut intensity at large radii and subtracted. The resulting radial intensity was then integrated over circular apertures with varying outer radii to estimate the flux density as a function of aperture radius (Table 1). Radial intensity cuts have also been extracted from the F770W, F1130W and F1280W filter images in a similar manner as for F1800W. The background appears to be dominated by nebular emission in the F1130W, F1280W and F1800W filters; for shorter wavelengths the PSF of the A2V star becomes increasingly dominant while the CS emission becomes progressively fainter. Aperture corrections have been determined using relatively bright field stars in the images, and applied to the aperture photometry of the CS (Table 1). In order to assess errors, for each filter, we used apertures of different diameters, each with its own aperture correction; the results shown in Table 1 are those obtained with largest aperture that could be feasibly used. We have assigned conservative errors of $\pm$15% to the NIRCam photometry, and $\pm 10$% to the MIRI photometry. Noting the 0.1 to 0.2 mag difference between pipeline and manually-determined photometry for the HST data at 0.44 and 0.56 µm (Table 1), we have assumed $\pm$15% errors for the published photometry at shorter wavelengths, i.e. $\lambda<0.9$ µm. The aperture photometry was also carried out for the PSF-subtracted images, and yielded consistent results. We have verified that our photometry methodology is correct by measuring the F356W and F770W flux densities for a few field stars, and comparing these to those measured with Spitzer, IRAC 1 and IRAC 4 (Spitzer Science Center, 2021) – we find that our photometry is in good agreement with the latter (Table 2), accounting for the possibility of temporal variability in these stars. ### 3.3 UV Spectrum We downloaded a FUSE UV spectrum of NGC 3132 that covers the $910-1190$ Å wavelength range with a spectral resolving power $R\sim 20,000$ (Moos et al. (2000); Sahnow et al. (2000)) from the Mikulski Archive for Space Telescopes, calibrated with the final version of the CalFUSE calibration pipeline software package, CalFUSE 3.2.3 (Dixon et al. (2007)). The spectrum was obtained as part of FUSE program ID P133 (PI: L. Bianchi), and an analysis of the same (reporting the discovery of Ge III $\lambda$ 1088.46 in NGC 3132 and other PNe) was first presented in Sterling et al. (2002). The spectra are of relatively modest S/N at the short wavelength end, and we have re-binned it to reduce the resolution by a factor 50. The spectrum was taken through the LWRS aperture, which has a size of $30{{}^{\prime\prime}}\times 30{{}^{\prime\prime}}$, and therefore covers not only the CS, but also the A2V companion and a significant fraction of the ionized region. However, the continuum flux in the $910-1190$ Å wavelength region is dominated by the CS, allowing us to use the FUSE spectrum to complete the SED of the CS to 0.091 µm at the short wavelength end (see § 4.2). ## 4 Analysis ### 4.1 A Central Binary and Orbital Motion We have investigated the relationship between the CS and the A2V star as follows. First, there is little doubt that these two stars are a physical pair (De Marco et al., 2022), given the close agreement in their $Gaia$-based proper motions (A2V star: $PMra=-7.747\pm 0.027$ mas yr-1, $PMdec=-0.125\pm 0.031$ mas yr-1: CS: $PMra=-7.677\pm 0.235$ mas yr-1, $PMdec=0.197\pm 0.28$ mas yr-1) and radial velocities (A2V star: $VLSR=-24.1\pm 1.6$ km s-1, CS: $VLSR=-25\pm 0.9$ km s-1). We have therefore inferred a distance to NGC 3132 of 0.75 kpc by inverting the parallax of the A2V star companion555no parallax is available for the CS ($1.3198\pm 0.0344$ mas) listed in the $Gaia$ DR3 database (Gaia Collaboration, 2022). Hence, the separation between the CS and its A2V companion and the CS is $1277\pm 34$ au. We have found direct evidence for the CS and the A2V star to be a gravitationally-bound binary system from the detection of orbital motion, over a time interval of $\sim 20$ yr between the epoch of the first high angular- resolution image of the CS-A2V pair with HST (Epoch 1), and the epoch for $Gaia$ measurements (Epoch 2) as follows. We used the Epoch 1 F555W image (from GO 06119, PI: H. Bond; image taken on 1995 Dec 04), obtained with the PC camera of WFPC2 (pixel-size of 0$\farcs$05), to determine the coordinates of the CS and the A2V star. Since the S/N is quite high ($>50$ for the CS, $>4000$ for the A2V), the uncertainty in the location of the CS is $\sim$1 mas, and that of the A2V star much smaller. The position angle of the separation vector of the A2V, CS pair666measured from N towards E, using the CS as the origin, in the HST image is, $PA(HST)=47.82\arcdeg\pm 0.045\arcdeg$ (Fig. 4). The position angle of this vector during the $Gaia$ measurement epoch (Epoch 2: $\sim$JD$\sim$2457410), estimated from the DR3 coordinates of the CS and A2V stars, is $PA(gaia)=47.43\arcdeg$. The JWST data could not be used to determine the current orientation of the CS-A2V separation vector accurately enough because the locations of these stars cannot be measured with the required accuracy – at the shortest wavelengths, where the telescope is diffraction limited and the PSF is small (e.g., $0\farcs 075$ at $\sim 2$ µm) the A2V star is badly saturated and the CS is very faint and its image is partly contaminated by the wings of the A2V star’s PSF. We have checked for any (minute) difference between the cardinal directions in the $Gaia$ DR3 reference frame ($Gaia$ CRF3: Gaia Collaboration et al. (2022)) and the HST image reference frame, $PA(gaia-HST)$, using the following methodology. We identified 4 field stars in the HST image (fs1, fs4, fs6, and fs8)777these are the 4/5 stars seen in the full FOV for which $Gaia$ DR3 data are available, and computed the $PA$s of their separation vectors (Table 4). Then, using the proper motions of these stars from $Gaia$ DR3, we have computed their expected locations in Epoch 1, and from these, the expected $PA$s of their separation vectors in Epoch 1 (Table 4). Because the separation vectors for these stars are quite large, the errors in the $PA$s are quite small ($\lesssim 0\farcs 01$). The average of the $PA(gaia-HST)$ values is $0.147\arcdeg\pm 0.016$°. Applying this correction to the value of $PA(HST)$ computed above, the corrected value of the $PA$ of the CS-A2V separation vector is, $PA(HST,c)=PA(HST)-0.147\arcdeg=47.67\arcdeg$. Thus, the separation vector has rotated by $0.24\arcdeg\pm 0.045\arcdeg$ (clockwise) in 20.1 yr (time interval between Epoch 1 and Epoch 2), in good agreement with the expected 0.20° cos$(i)$/cos 45° rotation for an orbital period of $P\sim 25,500$ yr, based on the measured separation of 1277 au, assuming an intermediate angle between the orbital and sky planes of $i=45\arcdeg$ and masses of 0.7 $M_{\odot}$ and 2.5 $M_{\odot}$ for the CS and A2V stars. The angular distance between the CS-A2V pair in Epoch 1 is $\epsilon(HST)=1\farcs 693\pm 0\farcs 001$, compared to $1\farcs 696$ in Epoch 2. We have checked the “plate-scale” of the HST image, using the same 4 fields stars and a similar methodology as above. We computed the angular distances between the star pairs used above for Epoch 1 using their current $Gaia$ DR3 coordinates and proper motions to compute their Epoch 1 coordinates, and compared these with the directly measured angular distances in the HST image (Table 4): the average fractional difference for the distances between star pairs between Epoch 1 and 2 (Table 4) is $(3.92\pm 1.35)\times 10^{-5}$. This difference is too small to affect the value of $\epsilon(HST)$ derived above. Therefore, there is weak evidence for an increase in the angular distance between the CS-A2V pair of $2.7\pm 1$ mas from Epoch 1 to Epoch 2. Although additional HST imaging is available at intermediate epochs, these images proved to be inadequate for determining the locations of the two stars with sufficient accuracy, because of one or more of the following factors: (i) the pixelsize was twice as large (e.g., $0\farcs 1$ using the WFPC2 Wide-Field cameras), (ii) only emission-line images were available providing inadequate S/N to centroid the CS accurately, (iii) the core of the PSF of the A2V star appeared non-circular and distorted. ### 4.2 Dust Radiative Transfer Modeling We have used the MIRI and NIRCam photometry, together with the FUSE UV spectrum and published optical photometry, to construct the observed SED of the CS and the dust cloud around it over the UV to mid-IR ($0.091-18$ µm) region (Fig. 5). The photometry is not co-eval; we assume that there are no significant variations in the CS’s spectrum over the $\sim$4 decades spanned by epoch of the U-band photometry and that of the JWST observations. Although the CS is included in the catalog of large amplitude variables from Gaia DR2 (Gaia DR2 5420219732233481472: Mowlavi et al. (2021)) is probably erroneous. Mowlavi et al. (2021) report a variability amplitude proxy of 0.3566 mag for the CS – this proxy was computed from the uncertainty (keyword “pho_g_mean_mag_error”) in the G-band magnitude (keyword “phot_g_mean_mag”) published in DR2. A comparison of the Gaia DR2 and DR3 G-band magnitudes – 15.216140 (mean magnitude based on 125 single epoch measurements) for DR2, and 16.105730 (mean magnitude based on 193 single epoch measurements888these include the 125 measurements used for the DR2 photometry) for DR3 – shows that the CS was almost one order of magnitude fainter in DR3 compared to DR2. However, since the DR3 data supersede the DR2 data, it is likely that the DR2 G-band magnitude and error are incorrect, and possibly result from incorrect removal of the contribution by the bright A2V star’s PSF to the measured photometry of the CS. The simplest explanation for the extended emission seen most clearly at 18 µm is that it arises from thermal emission from warm dust. Although there are strong nebular emission lines that have been detected in NGC 3132 in the wavelength bands covered by the filters listed above, it is very unlikely that such gas is present in the immediate vicinity of the CS. We have made a rough estimate of the width of the intrinsic radial intensity distribution at 18 µm (FWHMint) as follows. We approximated the observed radial intensity distribution as well as the PSF at 18 µm, using Gaussians of widths FWHM${}_{obs}\sim 1\farcs 0$ and FWHM${}_{psf}\sim 0\farcs 7$, and determined FWHM${}_{int}\sim 0\farcs 7$ using ${\rm FWHM}_{obs}^{2}={\rm FWHM}_{int}^{2}+{\rm FWHM}_{psf}^{2}$. This intrinsic width implies a radial extent of $\sim$260 au at NGC 3132’s distance of 0.75 kpc. Any gas this close to the star would be quite hot, $>10^{4}$ K, and at the sound speed for such gas, 10 km s-1, would expand to radii $>10,000-20,000$ au in (say) a typical post-AGB age of $>5,000-10,000$ yr. Furthermore, the SED of NGC 3132’s CS dust emission is similar to that of the Helix, for which a spectrum of the CS shows no obvious line emission. We have used the DUSTY dust radiative transfer code (Ivezić et al., 2012) to model the CS and its dust emission. Although it is plausible, depending on its origin, that the dust cloud is flattened or disk-like (see § 5), since we have no direct information about this aspect, we have chosen to use 1D modeling. This is not a limitation, because (as shown below), the optical depth of this cloud is $\ll 1$, even at relatively short wavelengths ($\lesssim 0.001$ at $\lambda\gtrsim 0.05$ µm), hence the results are not sensitive to the specific geometry (i.e., sphere or disk) of the cloud. Even if the dust cloud had a disk configuration, the radial optical depth near and in the equatorial plane would remain well below unity. The main input parameters are (i) the dust temperature at the inner shell boundary ($T_{\mathrm{d}}$), (ii) the total radial optical depth at 0.55$\micron$ ($\tau_{V}$), (iii) the shell density (iv) the grain-size distribution for a choice of grain composition, (v) the relative shell thickness ($Y$ = ratio of the shell’s outer radius, $R_{ou}$, to its inner radius, $R_{in}$), (vi) the spectrum of the central star. The shell density was assumed to be a single power-law exponent, $n$, of the power-law ($\rho_{d}(r)\propto r^{-n}$), or a combination of two power-laws (described below). We have computed the stellar spectrum using the Tübingen NLTE Model Atmosphere Package (TMAP) (Rauch & Deetjen (2003); Werner et al. (2003, 2012)) for $T_{eff}=105,000$ K – as derived by Monreal-Ibero & Walsh (2020) using photoionization modeling. We have accounted for nebular extinction by using a visual extinction of $A_{V}=0.31$ – derived by Kohoutek & Laustsen (1977)) – to attenuate and redden the model SED. We require the grains to be relatively cool ($\sim 100$ K at $r\sim 1{{}^{\prime\prime}}$) in order for the average slope of the model spectrum from 7.7 µm to 18 µm to match the observed one (Fig. 5). The shape of the model SED longwards of $\sim 5$ µm, where dust emission is the dominant component, is most sensitive to the dust grain properties (composition, size distribution). We have investigated models with “cold” (“Sil-Oc”), “warm” silicate grains (“Sil-Ow”), and amorphous C (amC) grains in DUSTY999DUSTY provides 6 in-built choices for grain composition. The SED has a very distinctive shape in the wavelength region dominated by dust emission, i.e., longwards of $\sim 5$ µm – there is a steep rise from $\sim 7$ µm to $\sim$11 µm, followed by a flat region between 11 and 13 µm, and then a rise towards longer wavelengths. While models with silicate grains produce the observed roughly flat shape of the SED in the 11-13 µm region, pure amC-grain models produce a smooth rise with increasing wavelength, discrepant from the observed SED. Warm silicates produce modestly smaller ratios flux F(18 µm)/F(12.8 µm), F(18 µm)/F(11.3 µm) and F(18 µm)/F(7.7 µm) compared to cold silicates. We used a modified version of the Mathis, Rumpl, Nordsieck (MRN) distribution function for grain radius $a$, $n(a)\propto$ a-q for $a_{\mathrm{min}}$ $\leq a\leq$ $a_{\mathrm{max}}$ (Mathis et al. (1977)), with q=1.101010The standard MRN parameters are: $q=3.5$, $a_{\mathrm{min}}$ = 0.005$\micron$ and $a_{\mathrm{max}}$ = 0.25$\micron$. Models with a substantial fraction of large grains e.g., $a_{\mathrm{m}ax}\gtrsim 1$µm produced radial intensity profiles that are significantly more compact than observed; in addition, the detailed shaped of the SED between 7.7 and 18 µm (described above) cannot be reproduced – large grains produce a dust emission spectrum in which there is a smooth rise with increasing wavelength. De Marco et al. (2022) present a simple large-grain (size $\sim$100 µm) dust shell model, which suffers from these deficiencies. Keeping the grain size distribution the same, but allowing the grains to be cooler by decreasing $T_{\mathrm{d}}$ results in a decrease in the 7.7 µm to 18 µm flux ratio, and simultaneously makes the model emission more extended due to an increase in $R_{in}$. Keeping the value of $T_{\mathrm{d}}$ fixed, an increase in the proportion of larger grains by increasing a(max) or decreasing $q$, makes the dust emission more compact. The model radial intensity distribution, especially at the longer wavelengths ($\sim 11-18$ µm) is sensitive to (i) $T_{d}$, and (ii) the radial density gradient of the dust shell. Lower values of $T_{d}$ lead to smaller values of $R_{in}$, and therefore a more compact radial intensity distribution of the dust emission. Higher values of $n$ also result in more compact dust emission. The above competing constraints allow us to provide reasonable constraints on the model parameters. The DUSTY code generates a model SED, normalized to the bolometric flux, $F_{bol}$. We find that $F_{bol}=(6.5\pm 0.5)\times 10^{-9}$ erg s-1 cm-2 (implying a CS luminosity of 114 $L_{\odot}$) by scaling the (reddened) model SED to match the observed SED (Fig. 5a). The value of $F_{bol}$ is well- constrained (for a given choice of $T_{eff}$), independently of the specific dust model, because the model SED in the optical to near-IR wavelength range (i.e., $\sim 0.1-3$ µm) (Fig. 5d) is only affected by nebular extinction (which is modest), and not by the dust cloud which has a very low optical depth. While determining the best-fit model, we have ignored (i) photometry from the narrow band filters F187N, F212N, F405N, and F470N, as the data from these are most affected by the presence of nebular atomic or molecular hydrogen lines, and (ii) $Gaia$ DR3 photometry due to potential imperfect removal of the PSF of the A2V star. We did not convolve the SED with the filter responses, but we have checked that the difference between convolved and non-convolved model photometry is relatively small and well below difference between the best-fit model and the data. In order to quantitatively distinguish between models, we have defined a ‘goodness-of-fit’ measure for the SED modeling, $G(sed)$, as follows: $G(sed)$ =$\sum\left(\frac{O_{\rm j}-M_{\rm j}}{\sigma_{\rm j}}\right)^{2}$, where, Oj ($\sigma_{\rm j}$) is the observed flux (error), Mj is the model flux, and the index j refers to different wavelengths. The summation was carried out only for $\lambda\geq 3.56$ µm because at shorter wavelengths, the SED is dominated by the central star, and including the data for these would reduce the sensitivity of $G(sed)$ to the contribution of the dust emission. The ‘goodness-of-fit’ for the radial intensity disributions, $G(Fx)$, where $Fx$ is F770W, F1280W, or F1800W, is: $G(Fx)$ =$\sum[r_{\rm j}\,\delta\,r_{\rm j}\,(Oint_{\rm j}-Mint_{\rm j})/\sigma_{\rm j}]^{2}/\sum[r_{\rm j}\,\delta\,r_{\rm j}]^{2}$, where $Oint_{\rm j}$ ($\sigma_{\rm j}$) is the observed normalized intensity (error), $Mint_{\rm j}$ is the model normalized intensity, $r_{\rm j}$ ($\delta\,r_{\rm j}$) is the radial offset (width of annulus) at grid-point j, and the index j refers to different radial offsets. The inclusion of $r_{\rm j}\,\delta\,r_{\rm j}$ in the formula for $G(Fx)$ means that we have used the difference of between (normalized) observed and model flux at each radial grid point for determining the goodness-of-fit. Better-fitting models have lower values of $G(sed)$ and $G(Fx)$. The above “1-shell” models provide a reasonably good fit to the SED (Fig. 5a) as well as the radial intensity in the F1800W, F1280W and F770W images (Fig. 6a,b,c). The grain composition needed to fit the specific shape of the SED, requires a mixture of 70% cold silicate (Sil-cW) and 30% amorphous carbon (amC) grains. However, there are small but noticeable discrepancies – (i) the F1800W model radial intensity lies a little below the observed one for radial offsets $r\lesssim 1{"}$, and (ii) the F1280W model radial intensity lies a little above the observed one for radial offsets $r\lesssim 0\farcs 65$. Since the radial dust emission intensity is sensitive to the dust radial density distribution $\rho_{d}(r)$, we explore the possibility that a more complex density structure than the ones used for the 1-shell models can reduce the above discrepancy. We have investigated models in which $\rho_{d}(r)$ is described by a broken power-law, such that $\rho_{d}(r)\propto r^{-n1}$ for $1\leq Y\leq Y1$, and $\rho_{d}(r)\propto r^{-n2}$ for $Y1\leq Y\leq Y$ (hereafter “2-shell” models). We find that a 2-shell model with $n1=0.2$, $n2=-0.4$, and $Y1=5$ provides the best fit (Fig. 7a,b,c). For this model, $G(sed)$ is a factor 0.75 times and $G(Fx)$ is a factor 1.2, 0.96, and 0.41 times (for $Fx$ equal to F770W, F1280W, and F1800W, respectively), of that for the 1-shell model. While comparing the $G(Fx)$ values for these models, we give the most weight to $G(F1800W)$ and the least wight to $G(F770W)$ since the dust distribution is most (least) resolved at $18\,\micron$ ($7.7\,\micron$). Thus the 2-shell model provides an overall better fit to both the SED and the radial instensity distribution. The density structure in the 2-shell model suggests that the dust cloud around the CS has two shells with a low density region in between. The presence of double shells is not uncommon for the central stars of PNe that show dust emission – Bilíková et al. (2012) find the presence of double shells in $\gtrsim$50% of the CSs with IR-excesses due to dust. We have therefore also investigated models with a different variant of the 2-shell density structure – in this class of models (hereafter “gap” models), we have a geometrically thin shell close to the star, separated by a density gap from the extended shell. In these models, we varied $T_{\mathrm{d}}$, the ratio of the density in the inner shell (assumed constant) to the density of the outer shell at the outer edge of the gap, the radius and width of the gap, and the power-law density exponent for the outer shell. We find that although the best “gap” model can fit the SED and the F1800W and F1280W radial intensities as well as the 1-shell and 2-shell models, the F770W radial intensity distribution is signifiantly narrower than observed. Hence, we do not discuss these models further. We have derived the mass of the dust shell as follows, since DUSTY does not provide a direct measure of the shell dust mass. For objects obeying a $r^{-n}$ density distribution, the dust mass in the circumstellar component is given by (for $n\neq 1,3$): $M_{d}=4\pi\,[(n-1)/(3-n)]\,y(Y)\,R^{2}_{in}(\tau_{18}/\kappa_{18})$ (1) where $y(Y)=(Y^{3-n}-1)/(1-Y^{1-n})$, and $\tau_{18}$ and $\kappa_{18}$ are, respectively, the radial optical depth of the shell and the dust mass absorption coefficient, at 18 $\micron$. We assume $\kappa_{18}=10^{3}$ cm2g-1, based on the dust properties for silicate dust tabulated by Ossenkopf et al. (1992). For the ‘2-shell” models, we apply Eqn. 1 piecewise to the inner and outer shell. The input parameters and output properties for the best-fit models are given in Table 5. The total dust mass is about $1.45\times 10^{-2}$ M⊕. In the 2-shell model, the mass in the inner shell is a small percentage (0.5%) of the total. For comparison, De Marco et al. (2022) derive a dust mass of $\sim 5\times 10^{-2}$ M⊕ from their simple model. For both the 1-shell and 2-shell model classes, models covering an appropriately large input parameter space were computed, allowing us to estimate rough uncertainties in the derived output parameters for a fixed value of $T_{eff}$. The most accurately determined parameter is $T_{d}$ (uncertainty $\lesssim 5$%), followed by $R_{in}$ (uncertainty $\lesssim 10$%). The uncertainty in the value of q is about 50%; the maximum grain radius may be as high as $\sim 0.5$ µm. The optical depth has an uncertainty of $\sim$20%. The dust masses have an uncertainty of $\sim$50%. The temperature of the CS also affects the model results. At the relatively high effective temperature estimated for the CS, the flux at $\lambda\gtrsim 0.4$ µm follows the Rayleigh-Jeans approximation, i.e., $F(\nu,T)\propto(2\,k\,T_{eff}/\lambda^{2})$, and $F_{bol}=F(\nu,T)\,\lambda^{2}\,(\sigma\,T_{eff}^{3})/(2\,\pi\,k)$ where $k$ and $\sigma$ are the Boltzmann and Stefan-Boltzmann constants. Hence the slope of the observed SED at short wavelengths (UV to near-IR), where the stellar emission dominates, does not provide a constraint on $T_{eff}$. Since $F_{bol}$ depends on $T_{eff}^{3}$ for a given value of the flux $F(\nu,T)$ and the latter is constrained by the observed photometry, variations in $T_{eff}$ lead to corresponding variations in the CS’s inferred luminosity, $L=114\,\mbox{$L_{\odot}$}\,(T_{eff}/105,000K)^{3}$. Since a larger value of $L$ increases $R_{in}$ for a given value of $T_{d}$, making the model dust emission more extended than observed, models with larger $L$ require higher values of $T_{d}$. For example, if $T_{eff}=110,000$ K, then good 1-shell model fits (i.e., similar to the one shown in Fig. 6) are obtained with $F_{bol}=(110/105)^{3}\,(6.5\times 10^{-9})=7.5\times 10^{-9}$ erg s-1 cm-2, and $T_{d}\sim 255-260$ K, $\tau_{V}=8.6\times 10^{-4}$, and mass $\sim(0.78-0.72)\times 10^{-2}$ M⊕. ## 5 Discussion The dust cloud around the central star of NGC 3132 is probably not the result of current mass loss, since such phenomena are not expected in a WD surrounded by a relatively old planetary nebula. It is more likely a stable disk in Keplerian or quasi-Keplerian rotation. As mentioned earlier (§ 1), disks have been found to be associated with quite evolved WDs, often too old to show detectable PNe around them (e.g.Tokunaga et al. (1990); Manser et al. (2020)). These disks are very small, occupying just a few au, and have been proposed to be the result of disruption of planets or planetoids from a former planetary system. Central stars of relatively young PNe tend to show disks with smaller radii, inferred to typically $\sim$ 50 au (Bilíková et al. (2012); Su et al. (2007)). Many post-AGB stars (thus less evolved than old WDs) also show disks (e.g. De Ruyter et al. (2006); Bujarrabal et al. (2013, 2016)), but in this case their extents are much larger, $\stackrel{{\scriptstyle>}}{{\scriptstyle\sim}}$ 1000 au. In this case, the rotating disks are systematically associated with the presence of tight binary systems and probably consist of material ejected during the previous AGB phase, which gains angular momentum from interaction with the system and forms rotating circumbinary disks. We now discuss possible origins of the dust cloud around NGC 3132’s CS. ### 5.1 An Oort Cloud The large radial extent ($>1500$ au) of the dust cloud in NGC 3132, makes it unlikely that it is formed from material extracted from a planetary system, or a Kuiper Belt/debris-disk analog – the typical radius of the latter is about $\sim 50-150$ au. However, an Oort cloud analog might explain the origin of this dust cloud. The Oort cloud in our Solar system is thought to occupy a radius between 2,000 and 5,000 au and may extend as far as 50,000 au from the Sun. The dust cloud around the CS of NGC 7293 may also be extended – Su et al. (2007) find that the 24 µm emission from this cloud, after background subtraction, appeared slightly resolved with a FWHM of $\sim 9{{}^{\prime\prime}}$, 1.5 times that of a true point source – although, as stated by Su et al. (2007), this could be due to an imperfect subtraction of the background, it may also be real (as in the case of NGC 3132.) We now estimate the survival probability of Oort’s rocky bodies after collisions with the gas and dust ejected by the CS while it was on the AGB or RGB. For a cometary nucleus with a typical radius of $\sim$1 km, located at a distance of $\gtrsim 2000$ au, only about $0.56\times 10^{10}$ g of circumstellar ejecta will hit the nucleus assuming a total $\sim$1 $M_{\odot}$ ejected via mass-loss, much smaller than the typical mass of the comet nucleus ($>10^{16}$ g). So the cometary nuclei in the Oort cloud should easily survive the impact of the CS ejecta. Assuming that the impacting gas is expanding at 15 km s-1, and all of its linear momentum is transmitted to the cometary nucleus, the latter will gain a velocity of 0.85 cm s-1, significantly smaller than the escape velocity at 2000 au ($\sim 1$ km s-1), implying that an Oort cloud will easily survive the AGB mass-loss phase. Thus, the Oort cloud, with an estimated mass of $\sim$1.9 M⊕(Weissman (1983)), can easily supply the mass of dust observed around the CS. However, in this scenario, we require a mechanism to transport the dust inwards, since the inner radius of the dust cloud in our model is $\sim 70$ au. We think it is possible that the A2V companion provides sufficient perturbation to the orbits of the cometary bodies in the Oort cloud, causing them to collide and/or increase their eccentricity sufficiently so as to reach the inner regions of the dust cloud. ### 5.2 Interaction with a Binary Companion As mentioned earlier, the central star of NGC 3132 has a detached A2V sp. type companion (§ 4.1). A closer companion has not been detected but cannot be discarded (we discuss this possibility in more detail below) – the highest mass unresolved main-sequence companion that can be present and remain undetected is of lower mass than that of a M-dwarf with spectral type M6V (with L$\sim 7\times 10^{-4}$ $L_{\odot}$ and $T_{eff}\sim 2500$ K, e.g., Cifuentes et al. (2020)), i.e., $\lesssim 0.1$ $M_{\odot}$. Including the theoretical spectrum of such a companion – BT-NextGen (AGSS2009) model with $T_{eff}=2500$ K, log $g$ (cm s-2)=5, extracted from http://svo2.cab.inta- csic.es/theory/newov2/index.php, Allard et al. (2011) – increases the model SED flux in the F356W filter by 30%, well above the observed value (see inset of Fig. 5d). De Marco et al. (2022) propose that the CS has at least 2, and maybe even 3 close companions, in order to explain the origin of the dust cloud and the morphological structure of the extended PN. From angular momentum balance considerations alone, the dust cloud may be a result of interaction with either such a close companion or the detached A2V companion. The angular momentum of the cloud, assuming it have a disk geometry and in Keplerian rotation around a $\sim 0.65$ $M_{\odot}$ CS, is low ($\lesssim 0.5\times 10^{4}$ $M_{\odot}$ km2 s-1), orders of magnitude less than that of a compact binary system even if the secondary is just a big planet – e.g., the angular momentum of a Jupiter-mass planet in a 1 au orbit around the CS is $3.54\times 10^{6}$ $M_{\odot}$ km2 s-1. The A2V star’s angular momentum, $\sim 1.5\times 10^{11}$ $M_{\odot}$ km2 s-1, is also much larger than that of the CS’s dust cloud, assuming the latter has a disk geometry. From the point of view of the momentum transfer mechanism, however, the situation is very different for the above two scenarios. The currently adopted wind Roche-Lobe Overflow (wRLOF) mechanism of hydrodynamical interaction between a companion and a stellar wind in the red giant phase predicts strong effects when the separation of both stars is small, typically under 50–100 au (Mohamed & Podsiadlowski (2012); Chen et al. (2017); Kim et al. (2019)). The circumstellar gas at such distances is still being accelerated and shows a relatively low velocity, which is a basic ingredient for allowing strong interaction effects, including the formation of rotating circumbinary disks and symbiotic phenomena (e.g., Sánchez Contreras et al. (2022)). Rotation in inner suborbital regions can also be induced in those cases of strong interaction (e.g. Bermúdez-Bustamante et al. (2020)). However, when the distance between the two stars is significantly larger, the circumstellar expansion velocity is expected to be large and models predict that the gas passes by the companion with a minor interaction; the main effect is then the formation of spiral arcs due to the oscillation of the mass-ejecting primary. Formation of rotating disks is therefore not expected in detached binary systems. Hence, if the dust cloud is a disk resulting from binary interaction, we require that the CS has (or had in the past) another companion that is (was) much closer than the A2V star, and that this close companion (hereafter “Compc”) underwent a strong gravitational interaction with the CS. The A2V star may have played an active role in inducing the strong gravitational interaction of Compc with the CS. In this scenario, three stars – the CS, Compc (say at a separation of $\lesssim 5$ au), and the A2V star (in a more distant orbit, say at $\sim$400 au) – formed a stable hierarchical triple system while these stars were on the main-sequence. Such systems can become dynamically active on much longer time-scales due to the “Eccentric Kozai-Lidov” (EKL) mechanism causing the inner binary to undergo large- amplitude eccentricity and inclination oscillations (Kozai (1962); Lidov (1962); Naoz (2016)). The oscillations tend to drive the inner binary to have very small pericenter distances and even to merge (e.g., Prodan et al. (2015); Stephan et al. (2018)). Salas et al. (2019) have simulated such a triple system with a 2.2$M_{\odot}$ primary, a close companion (with a range of masses $\lesssim$0.9 $M_{\odot}$) and a tertiary star with a range of masses M$\lesssim$0.9 $M_{\odot}$) – they find that in 37% of their simulations, the tight binary merges. Although in the case of NGC 3132, the tertiary star is more massive ($>$2.5 $M_{\odot}$), Salas et al. (2019)’s results indicate that there is a significant possibility of the inner binary to merge. Such a merger would lead to a common envelope ejection of most of the stellar envelope of the primary star – the multipolar morphology observed in NGC 3132 would then be a result of interaction of multipolar collimated outflows with the ejecta, as appears to be the case in the Boomerang Nebula (Sahai et al. (2017)). The Boomerang is a pre-planetary nebula which has most likely resulted from CEE while the central star was still on the RGB and therefore in a much earlier post-AGB phase than NGC 3132. Since the primary must be initially more massive than the A2V star, say $\sim 2.8$ $M_{\odot}$, simple conservation of angular momentum indicates that the tertiary’s orbit, if initially equal to $\sim$400 au, would expand to a semi-major axis of $\sim$1200 au after the merger, bringing it to its current observed location. ### 5.3 The A2V Companion’s Effects on the Dust Cloud around the CS The JWST images show that the radial extent of the dust cloud is almost identical to the orbital radius of the CS-A2V detached binary system. In addition, there is a flattening of the 18 µm intensity contours defining the shape of the dust cloud on the side facing the A2V star. These two features suggest a physical interaction between the dust cloud formation and the A2V companion; we now discuss several physical mechanisms for such an interaction. The first is that the A2V star preferentially illuminates the dust closest to it, making it relatively hotter, and therefore produce an extra brightening of the dust emission, relative to the diametrically-opposed side of the dust cloud, contrary to what is observed. Next, we consider the effect of radiation pressure and a possible wind from the A2V star on the dust cloud around NGC 3132’s CS. The ratio of the stellar wind force ($F_{sw}$) to the gravitational force of star ($F_{gr}$) on a dust grain, is given by: $\beta_{sw}=3\mbox{$\dot{M}$}_{sw}V_{sw}C_{D}/(32\,\pi GM_{*}\rho_{g}a_{g}),$ (2) where $M_{*}$, $\mbox{$\dot{M}$}_{sw}$, and $V_{sw}$, and are the mass, mass- loss rate and outflow velocity for the A2V star, $G$ is the gravitational constant, $\rho_{g}$ and $a_{g}$ are the dust grain material density and radius, $r$ is the radial distance from the A2V star, $C_{D}$ is a coefficient $\sim 2$ (e.g., Eqn. 28 in Augereau & Beust (2006)). Taking $\mbox{$\dot{M}$}_{sw}=10^{-10}$ $M_{\odot}$ yr-1 (e.g., Lanz & Catala (1992)), $V_{sw}=300$ km s-1, $\rho_{g}=3$ g cm-3, $a_{g}=0.25$ µm, $M_{*}=2.5$ $M_{\odot}$, we get $\beta_{sw}=0.45$. The ratio of the radiation pressure force ($F_{rd}$) to the gravitational force of star ($F_{gr}$) on a dust grain, $\beta_{rd}$, has been computed by Lamy & Perrin (1997) for stars of various spectral types and luminosity classes, two of which bracket the A2V star – these are $\alpha$Aql (A5 IV-V) and $\alpha$CMi (F6 IV-V). This study shows that, for silicate grains of radii in the range 0.1–1 µm, $\beta_{rd}\sim 1-3$, i.e., significantly greater than the value of $\beta_{sw}$ derived above. Thus, radiation pressure is much more effective in pushing the grains away from the A2V star than the A2V star’s wind. Adopting an intermediate value, $\beta_{rd}=2$, the acceleration due to radiation pressure at a radial offset that is (say) halfway between the A2V star and the CS, i.e. at $r=0\farcs 85$, is $\sim 7.2\times 10^{-6}$ cm s-2 (and varies as $r^{-2}$). Thus, within about 1000 yr, radiation pressure will move the grains that are located (say) halfway between the A2V star and the CS, roughly 100 au or $0\farcs 13$ towards the CS. This pressure would push the outer regions of the dust cloud on the side facing the A2V star towards the CS, providing a plausible explanation for the observed flattening of the 18 µm intensity contours defining the shape of the dust cloud on the side facing the A2V star. But this process would also lead to an increase in the column density near this edge, which would result in a brightening of the dust emission compared to the outer regions of the dust cloud that are located on the side facing away from the A2V star, contrary to what is observed. A plausible explanation of why the dust cloud closer to the A2V star does not show enhanced emission, is that the wind from the A2V star also destroys some fraction of dust grains via sputtering. For example, e.g., Gray & Edmunds (2004) have studied sputtering as a function of impact energy of hydrogen nuclei, and their Fig. 2 shows a substantial sputtering yield for impact energies of $\sim$1 keV (H and He nuclei moving at 300 km s-1 have energies corresponding to 0.47 and 1.9 keV). Given the clockwise rotation that we find for the A2V star around the CS, the grains in the dust cloud must also be rotating clockwise (set by the sign of the global angular momentum of the primordial dense core in which these stars were formed). Hence, the dust grains approach the A2 star from the south-east as a result of their orbital motion. When they are far from the separation vector, the dominating force is the gravitational attraction to the CS, but that force becomes progressively less as the grains get closer to the A2V star. The “tail” seen in the 18 µm image in the north/north-west direction is thus a signature of grains trapped by the A2V star (as they pass from a dynamical regime dominated by gravitation attraction toward the CS to one that is dominated by attraction toward the A2V star) or perhaps escaping the CS-A2V system (because they pass a region with a very weak net gravitational force). ### 5.4 The Origin of the Density Discontinuity in the Dust Cloud The discontinuity in the 2-shell dust model suggests the presence of a radial gap in the dust cloud density, at a radius of $\sim$350 au. A plausible origin of this gap is the presence of a giant planet or brown dwarf with an initial orbital radius of $\sim 10$ au; following the mass-loss from the central star as described above (§ 5.2), the orbital radius (as in the case of the A2V companion) would increase to $\sim$350 au. This giant planet/ brown dwarf could then open up a fairly wide gap in the disk – e.g., Crida et al. (2006) have computed gap density profiles using semi-analytic calculations and numerical simulations for a Jupiter mass planet in a disk, and find that the widths are comparable or larger than the orbital radius for low viscosities. The migration of the giant planet/ brown dwarf radially outwards could also contribute to further broadening the gap. Multiple planets would produce larger gaps but with shallower depths (e.g., see Fig. 1 of Duffell & Dong (2015)). ## 6 Conclusions We have analysed new imaging data of the central star (CS) of NGC 3132 obtained using its NIRCam and MIRI instruments onboard JWST, through a set of filters spanning the $0.9-18$ µm range. Our main findings are as follows: 1. 1. The CS is located at an angular distance of $1\farcs 696$ from a bright A2V star to its north-east. We find that these stars form a wide gravitationally- bound system, separated by 1277 au and located at a distance of 0.75 kpc from Earth. The proper motions and radial velocities of these stars are consistent within uncertainties. In addition, we detect relative orbital motion between the two stars over a $20$ yr period, which is consistent with the expected value from the 25,500 yr period of the binary system estimated from the current separation and assuming an intermediate inclination angle of 45° between the orbital and sky-planes. 2. 2. The A2V star outshines the CS at all but the longest wavelength of 18 µm. Using PSF subtraction of the A2V star, we find that CS is clearly seen in the JWST images even at the shortest wavelength 0.9 µm. The CS is surrounded by extended emission, seen directly at $18$ µm. Radial intensity cuts show that the emission is extended in the $7.7-12.8$ µm range as well. This emission, which is surface-brightness limited and is somewhat asymmetrical in its outer regions, extends to a radius $\gtrsim$1600 au and most likely results from thermal dust emission. 3. 3. We have carried out aperture photometry of the CS in the JWST images. Using these data, together with an archival UV spectrum and published optical photometry, we have constructed the spectral-energy distribution (SED) of the CS and its extended emission over the UV to mid-IR ($0.091-18$ µm) range. The SED has a very distinctive shape in the $\sim 5-18$ µm region. 4. 4. Using dust radiative modeling, we have fitted the SED of the CS and the radial intensity distributions at $7.7,12.8$ and $18$ µm with a dust cloud that extends to a radius of $\gtrsim$1785 au. Models with a dust composition of 70% silicate and 30% amorphous carbon, and a modified MRN grain-size distribution that increases the proportion of larger grains, i.e., one in which the number of grains with radius $a$, varies as $n(a)\propto$ a-q for $a=0.005-0.25$ µm, with q=1, provide the best fit to the specific shape of the SED in the $5-18$ µm range. The radial dust optical depth of this cloud is $\sim 10^{-4}$ at 0.55 µm; and the dust temperature decreases from about $232$ to $73$ K from the inner to the outer radius. Our best-fit models give a total dust mass of $(1.3\pm 0.15)\times 10^{-2}$ M⊕ within a radius of $1785$ au; the dust mass estimate has a conservative uncertainty of about $\pm 25$%. 5. 5. The material in the dust cloud may have come from a pre-existing Oort cloud analog, or it may lie in a disk-like structure produced as a result of binary interaction. 6. 6. The radial extent of the dust cloud is almost identical to the orbital radius of the CS-A2V binary stellar system; the cloud appears flattened on the side facing the A2V star. These features suggest a physical interaction between the disk and the A2V companion, due to a combination of radiation pressure (due to the A2V star’s radiation), together with partial destruction of the dust grains by sputtering (due to a tenuous wind from the A2V star). 7. 7. A plausible evolutionary scenario that explains the spectacular mass-ejection that has resulted in the current, multipolar planetary nebula, is one in which three stars – the CS, a close low-mass companion (with, say, $a\lesssim 5$ au), and a much more distant A2V star (with, say, $a\sim$400 au) – formed a stable hierarchical triple system on the main-sequence, but which then became dynamically active much later due to the Eccentric Kozai-Lidov mechanism causing a strong binary interaction between the inner pair and leading to a loss of most of the primary’s envelope. The resulting severe reduction in the CS mass caused the A2V star’s orbit to expand, resulting in its current separation from the CS. 8. 8. We set an upper limit of $\sim$0.1 $M_{\odot}$ on the mass of any main- sequence star, i.e., spectral-type M6V (L$\sim 7\times 10^{-4}$ $L_{\odot}$, $T_{eff}\sim 2500$ K) or later, that may be located close enough to the CS to be unresolved and faint enough to be undetectable. Such a star could be the close binary companion in the above evolutionary scenario, provided the binary interaction did not lead to its merger with the CS. ## 7 acknowledgements We thank an (anonymous) referee for his/her timely and thorough review which has helped us improve our paper. The Early Release Observations and associated materials were developed, executed, and compiled by the ERO production team: Hannah Braun, Claire Blome, Matthew Brown, Margaret Carruthers, Dan Coe, Joseph DePasquale, Nestor Espinoza, Macarena Garcia Marin, Karl Gordon, Alaina Henry, Leah Hustak, Andi James, Ann Jenkins, Anton Koekemoer, Stephanie LaMassa, David Law, Alexandra Lockwood, Amaya Moro-Martin, Susan Mullally, Alyssa Pagan, Dani Player, Klaus Pontoppidan, Charles Proffitt, Christine Pulliam, Leah Ramsay, Swara Ravindranath, Neill Reid, Massimo Robberto, Elena Sabbi, Leonardo Ubeda. The EROs were also made possible by the foundational efforts and support from the JWST instruments, STScI planning and scheduling, and Data Management teams. Most of the data presented in this paper were obtained from the Mikulski Archive for Space Telescopes (MAST) at the Space Telescope Science Institute. The specific observations analyzed can be accessed via http://dx.doi.org/10.17909/6wd1-0170 (catalog http://dx.doi.org/10.17909/6wd1-0170). This work is also based in part on observations made with the Spitzer Space Telescope, which is operated by the Jet Propulsion Laboratory, California Institute of Technology under a contract with NASA, that can be accessed via https://doi.org/10.26131/irsa3 (catalog https://doi.org/10.26131/irsa3). RS’s contribution to the research described here was carried out at the Jet Propulsion Laboratory, California Institute of Technology, under a contract with NASA. VB acknowledges support from the EVENTs/NEBULAE WEB research program, Spanish AEI grant PID2019-105203GB-C21. CSC’s work is part of the I+D+i project PID2019-105203GB-C22 funded by the Spanish MCIN/AEI/10.13039/501100011033. The TMAW tool (http://astro.uni-tuebingen.de/ TMAW) used for this paper was constructed as part of the activities of the German Astrophysical Virtual Observatory. ## References * Allard et al. (2011) Allard, F., Homeier, D., & Freytag, B. 2011, 16th Cambridge Workshop on Cool Stars, Stellar Systems, and the Sun, 448, 91 * Augereau & Beust (2006) Augereau, J.-C. & Beust, H. 2006, A&A, 455, 987. doi:10.1051/0004-6361:20054250 * Becklin et al. (2005) Becklin, E. E., Farihi, J., Jura, M., et al. 2005, ApJ, 632, L119 * Bermúdez-Bustamante et al. (2020) Bermúdez-Bustamante, L. C., García-Segura, G., Steffen, W., et al. 2020, MNRAS, 493, 2606 * Bilíková et al. (2012) Bilíková, J., Chu, Y.-H., Gruendl, R. A., et al. 2012, ApJS, 200, 3 * Bouchet et al. (2015) Bouchet, P., García-Marín, M., Lagage, P.-O., et al. 2015, PASP, 127, 612. doi:10.1086/682254 * Bradley et al. (2015) Bradley, L., Sipőcz, B., Robitaille, T., et al. 2022, Zenodo. doi:10.5281/zenodo.6825092 * Bujarrabal et al. (2013) Bujarrabal, V., Alcolea, J., Van Winckel, H., et al. 2013, A&A, 557, A104 * Bujarrabal et al. (2016) Bujarrabal, V., Castro-Carrizo, A., Alcolea, J., et al. 2016, A&A, 593, A92 * Chen et al. (2017) Chen, Z., Frank, A., Blackman, E. G., et al. 2017, MNRAS, 468, 4465 * Chu et al. (2011) Chu, Y.-H., Su, K. Y. L., Bilikova, J., et al. 2011, AJ, 142, 75 * Cifuentes et al. (2020) Cifuentes, C., Caballero, J. A., Cortés-Contreras, M., et al. 2020, A&A, 642, A115. doi:10.1051/0004-6361/202038295 * Crida et al. (2006) Crida, A., Morbidelli, A., & Masset, F. 2006, Icarus, 181, 587. doi:10.1016/j.icarus.2005.10.007 * De Marco et al. (2022) De Marco, O. et al. 2022, Nat. Astron. (in press). doi:10.1038/s41550-022-01845-2 * De Ruyter et al. (2006) de Ruyter, S., van Winckel, H., Maas, T., et al. 2006, A&A, 448, 641 * Decin (2021) Decin, L. 2021, ARA&A, 59, 337. doi:10.1146/annurev-astro-090120-033712 * Dixon et al. (2007) Dixon, W. V., Sahnow, D. J., Barrett, P. E., et al. 2007, PASP, 119, 527. doi:10.1086/518617 * Duffell & Dong (2015) Duffell, P. C. & Dong, R. 2015, ApJ, 802, 42. doi:10.1088/0004-637X/802/1/42 * Farihi et al. (2005) Farihi, J., Becklin, E. E., & Zuckerman, B. 2005, ApJS, 161, 394 * Fazio et al. (2004) Fazio, G. G., Hora, J. L., Allen, L. E., et al. 2004, ApJS, 154, 10. doi:10.1086/422843 * Gaia Collaboration (2022) Gaia Collaboration 2022, VizieR Online Data Catalog, I/355 * Gaia Collaboration et al. (2022) Gaia Collaboration, Klioner, S. A., Lindegren, L., et al. 2022, arXiv:2204.12574 * Gray & Edmunds (2004) Gray, M. D. & Edmunds, M. G. 2004, MNRAS, 349, 491. doi:10.1111/j.1365-2966.2004.07502.x * Gänsicke et al. (2019) Gänsicke, B. T., Schreiber, M. R., Toloza, O., et al. 2019, Nature, 576, 61 * Hillen et al. (2017) Hillen, M., Van Winckel, H., Menu, J., et al. 2017, A&A, 599, A41. doi:10.1051/0004-6361/201629161 * Ivezić et al. (2012) Ivezić, Z., Nenkova, M., Heymann, F. & Elitzur, M. 2012, User Manual for DUSTY (V4) * Jakobsen et al. (2022) Jakobsen, P., Ferruit, P., Alves de Oliveira, C., et al. 2022, A&A, 661, A80. doi:10.1051/0004-6361/202142663 * Jura (2003) Jura, M. 2003, ApJ, 584, L91 * Kilic et al. (2005) Kilic, M., von Hippel, T., Leggett, S. K., et al. 2005, ApJ, 632, L115 * Kim et al. (2019) Kim, H., Liu, S.-Y., & Taam, R. E. 2019, ApJS, 243, 35 * Kohoutek & Laustsen (1977) Kohoutek, L. & Laustsen, S. 1977, A&A, 61, 761 * Kozai (1962) Kozai, Y. 1962, AJ, 67, 591. doi:10.1086/108790 * Kuruwita et al. (2016) Kuruwita, R. L., Staff, J., & De Marco, O. 2016, MNRAS, 461, 486 * Lamy & Perrin (1997) Lamy, P. L. & Perrin, J.-M. 1997, A&A, 327, 1147 * Lanz & Catala (1992) Lanz, T. & Catala, C. 1992, A&A, 257, 663 * Lidov (1962) Lidov, M. L. 1962, Planet. Space Sci., 9, 719. doi:10.1016/0032-0633(62)90129-0 * Manser et al. (2020) Manser, C. J., Gänsicke, B. T., Gentile Fusillo, N. P., et al. 2020, MNRAS, 493, 2127 * Mastrodemos & Morris (1998) Mastrodemos, N. & Morris, M. 1998, ApJ, 497, 303 * Mathis et al. (1977) Mathis, J. S., Rumpl, W. & Nordsieck, K. H. 1977, ApJ, 217, 425 * Mohamed & Podsiadlowski (2012) Mohamed, S. & Podsiadlowski, P. 2012, Baltic Astronomy, 21, 88 * Miller Bertolami (2016) Miller Bertolami, M. M. 2016, A&A, 588, A25. doi:10.1051/0004-6361/201526577 * Mowlavi et al. (2021) Mowlavi, N., Rimoldini, L., Evans, D. W., et al. 2021, A&A, 648, A44. doi:10.1051/0004-6361/202039450 * Méndez (1978) Méndez, R. H. 1978, MNRAS, 185, 647. doi:10.1093/mnras/185.3.647 * Monreal-Ibero & Walsh (2020) Monreal-Ibero, A. & Walsh, J. R. 2020, A&A, 634, A47. doi:10.1051/0004-6361/201936845 * Monteiro et al. (2000) Monteiro, H., Morisset, C., Gruenwald, R., et al. 2000, ApJ, 537, 853. doi:10.1086/309058 * Moos et al. (2000) Moos, H. W., Cash, W. C., Cowie, L. L., et al. 2000, ApJ, 538, L1. doi:10.1086/312795 * Ossenkopf et al. (1992) Ossenkopf, V., Henning, Th. & Mathis, J. S. 1992, A&A, 261, 567 * Pontoppidan et al. (2022) Pontoppidan, K. M., Barrientes, J., Blome, C., et al. 2022, ApJ, 936, L14. doi:10.3847/2041-8213/ac8a4e * Prodan et al. (2015) Prodan, S., Antonini, F., & Perets, H. B. 2015, ApJ, 799, 118. doi:10.1088/0004-637X/799/2/118 * Naoz (2016) Naoz, S. 2016, ARA&A, 54, 441. doi:10.1146/annurev-astro-081915-023315 * Rauch & Deetjen (2003) Rauch, T. & Deetjen, J. L. 2003, Stellar Atmosphere Modeling, 288, 103 * Ricker & Taam (2012) Ricker, P. M. & Taam, R. E. 2012, ApJ, 746, 74 * Rieke et al. (2005) Rieke, G. H., Su, K. Y. L., Stansberry, J. A., et al. 2005, ApJ, 620, 1010 * Rigby et al. (2022) Rigby, J., Perrin, M., McElwain, M., et al. 2022, arXiv:2207.05632 * Sahai et al. (2007) Sahai, R., Morris, M., Sánchez Contreras, C., et al. 2007, AJ, 134, 2200. doi:10.1086/522944 * Sahai et al. (2011) Sahai, R., Morris, M. R., & Villar, G. G. 2011, AJ, 141, 134 * Sahai et al. (2017) Sahai, R., Vlemmings, W. H. T., Nyman, L.-Å. 2017, ApJ, 841, 110 * Sahai et al. (1990) Sahai, R., Wootten, A., & Clegg, R. E. S. 1990, A&A, 234, L1 * Sahnow et al. (2000) Sahnow, D. J., Moos, H. W., Ake, T. B., et al. 2000, ApJ, 538, L7. doi:10.1086/312794 * Spitzer Science Center (SSC) Spitzer Science Center (SSC) & Infrared Science Archive (IRSA) 2021, VizieR Online Data Catalog, II/368 * Salas et al. (2019) Salas, J. M., Naoz, S., Morris, M. R., et al. 2019, MNRAS, 487, 3029. doi:10.1093/mnras/stz1515 * Stephan et al. (2018) Stephan, A. P., Naoz, S., & Gaudi, B. S. 2018, AJ, 156, 128. doi:10.3847/1538-3881/aad6e5 * Sterling et al. (2002) Sterling, N. C., Dinerstein, H. L., & Bowers, C. W. 2002, ApJ, 578, L55. doi:10.1086/344473 * Storey (1984) Storey, J. W. V. 1984, MNRAS, 206, 521. doi:10.1093/mnras/206.3.521 * Su et al. (2007) Su, K. Y. L., Chu, Y.-H., Rieke, G. H., et al. 2007, ApJ, 657, L41 * Sánchez Contreras et al. (2022) Sánchez Contreras, C., Alcolea, J., Rodríguez Cardoso, R., et al. 2022, arXiv:2206.12185 * Tokunaga et al. (1990) Tokunaga, A. T., Becklin, E. E., & Zuckerman, B. 1990, ApJ, 358, L21 * Vagnozzi (2019) Vagnozzi, S. 2019, Atoms, 7, 41. doi:10.3390/atoms7020041 * Vanderburg et al. (2020) Vanderburg, A., Rappaport, S. A., Xu, S., et al. 2020, Nature, 585, 363 * Weissman (1983) Weissman, P. R. 1983, A&A, 118, 90 * Werner et al. (2012) Werner, K., Dreizler, S., & Rauch, T. 2012, Astrophysics Source Code Library. ascl:1212.015 * Werner et al. (2003) Werner, K., Deetjen, J. L., Dreizler, S., et al. 2003, Stellar Atmosphere Modeling, 288, 31 Table 1: Photometry of the Central Star of NGC 3132 Filter | Wavelength | Flux | ErroraaPercentage Error in Flux in previous column | Apert. | Apert. | Phot ---|---|---|---|---|---|--- | (µm) | (mJy) | (%) | Rad.(′′) | Corr. | Ref.bbReferences for photometry: (1) Kohoutek & Laustsen 1977, (2) photometry on UVIS/WFC3 images (HST proposal 11699) from Monreal-Ibero & Walsh (2020), (3) Hubble Source Catalog V.3 (Whitmore et al. 2016), (4) $Gaia$ DR3, (5) this work; when pohotometry reference provides magnitudes, these are listed here U | 0.36 | 2.15 | 15 | … | … | 14.8ccVega Magnitude (1) F438W | 0.438 | 1.80 | 15 | … | … | 15.76ccVega Magnitude (3) F438W | 0.438 | 1.62 | 15 | … | … | 15.876ddAB Magnitude (2) F555W | 0.555 | 1.45 | 15 | … | … | 16.0ccVega Magnitude (3) F555W | 0.555 | 1.19 | 15 | … | … | 16.212ddAB Magnitude (2) G-band | 0.639 | 1.18 | 15 | … | … | (4) F814W | 0.814 | 0.56 | 15 | … | … | 17.03ddAB Magnitude (3) F090W | 0.90 | 0.38 | 15 | $0\farcs 10$ | 0.76 | (5) F187N | 1.87 | 0.11 | 15 | $0\farcs 10$ | 0.75 | (5) F212N | 2.12 | 0.106 | 15 | $0\farcs 10$ | 0.74 | (5) F356W | 3.56 | 0.037 | 15 | $0\farcs 15$ | 0.71 | (5) F405N | 4.05 | 0.020 | 15 | $0\farcs 15$ | 0.73 | (5) F470N | 4.70 | 0.018 | 15 | $0\farcs 15$ | 0.72 | (5) F770W | 7.7 | 0.070 | 10 | $0\farcs 35$ | 0.76 | (5) F1130W | 11.3 | 1.16 | 10 | $0\farcs 75$ | 0.97 | (5) F1280W | 12.8 | 1.2 | 10 | $0\farcs 75$ | 0.93 | (5) F1800W | 18.0 | 9.9 | 10 | $1\farcs 1\times 1\farcs 3$ | … | (5) F1800W | 18.0 | 10.3eeFlux derived from integration of radial intensity to outer radius in Col. (3) | 10 | $1\farcs 6$ | … | (5) F1800W | 18.0 | 11.1eeFlux derived from integration of radial intensity to outer radius in Col. (3) | 10 | $1\farcs 8$ | … | (5) F1800W | 18.0 | 12.1eeFlux derived from integration of radial intensity to outer radius in Col. (3) | 10 | $2\farcs 0$ | … | (5) Table 2: Photometry of Field Stars Star | WavelengthaaWavelength of JWST filter-passband | FluxbbMeasured photometry from this study | ErrorccPercentage Error in Flux in previous column | WavelengthddWavelength of filter-passband for published IRAC photometry (Fazio et al. (2004)) | FluxeePublished photometry | ErrorccPercentage Error in Flux in previous column | Instr.ffInstrument/ Detector for published photometry | Publ.Phot. ---|---|---|---|---|---|---|---|--- | (µm) | (mJy) | (%) | (µm) | (mJy) | (%) | | Ref.ggReference for published photometry: (1) The Spitzer (SEIP) source list (SSTSL2) (Spitzer Science Center, SSC) fs3 | 3.56 | 0.38 | 5 | 3.56 | 0.359 | 0.4 | IRAC 1 | 1 fs3 | 7.7 | 0.085 | 10 | 7.91 | 0.075 | 7 | IRAC 4 | 1 fs4 | 7.7 | 0.22 | 5 | 7.91 | 0.27 | 2.5 | IRAC 4 | 1 fs5 | 7.7 | 3.9 | 5 | 7.91 | 4.6 | 0.2 | IRAC 4 | 1 Table 3: Instr. | Filter | BoxaaBox size used for estimating a median background | % ErrorbbPercentage error in the scale-factor applied to the PSF for subtraction from A2V star image | G. smoothccFWHM of gaussian smoothing function applied to PSF before PSF-subtraction ---|---|---|---|--- | | (px/′′) | in PSF scale-factor | (′′) NIRCam | F090W | 50/1.54 | 0.16$\times 10^{-2}$ | 0.05 NIRCam | F187N | 50/1.54 | 0.20$\times 10^{-2}$ | 0.03 NIRCam | F212N | 200/6.17 | 0.13$\times 10^{-2}$ | 0.02 NIRCam | F356W | 50/3.15 | 0.47$\times 10^{-2}$ | 0.05 NIRCam | F405N | 250/15.75 | 0.15$\times 10^{-1}$ | 0.05 NIRCam | F470N | 50/3.15 | 0.17$\times 10^{-1}$ | 0.045 MIRI | F770W | 200/22.18 | 0.26$\times 10^{-1}$ | 0.045 MIRI | F1130W | 200/22.18 | 0.97$\times 10^{-1}$ | 0.07 MIRI | F1280W | 169/18.74 | 0.15 | 0.1 MIRI | F1800W | 135/14.97 | 0.49 | 0.1 Table 4: Position Angles and Separations of Field Star Pairs StarPair | $PA(HST)$aaPA of separation vector between star pair in Epoch 1, using HST image. PA is measured from N towards E, with the second star in the pair as the origin | $PA($Gaia$,Epoch1)$bbPA of separation vector between star pair in Epoch 1, estimated from $Gaia$ DR3 data | Ang.Dist.(HST)ccAngular distance between star pair in Epoch 1, using HST image | Err(HST)ddError in angular distance between star pair in Epoch 1, using HST image | Frac.Diff.eeAngular distance between star pair in Epoch 1 (using HST image) minus the angular distance between star pair in Epoch 1 (estimated from $Gaia$ DR3 data), divided by the average angular distance between star pair ---|---|---|---|---|--- | ° | ° | ′′ | ′′ | $10^{-5}$ fs6, fs1 | 22.016 | 21.850 | 98.1236 | 0.00055 | -2.58 fs6, fs8 | 21.219 | 21.071 | 162.7896 | 0.0012 | -4.25 fs1, fs8 | 20.031 | 19.912 | 64.6894 | 0.0011 | -6.36 fs4, fs8 | 13.816 | 13.674 | 141.6898 | 0.0017 | -3.55 fs4, fs1 | 8.647 | 8.486 | 77.6963 | 0.0013 | -2.85 Table 5: Models of the Dust Emission towards the Central Star of NGC 3132 Td(in)aaThe (input) dust temperature at shell inner radius | $R_{in}$bbThe (inferred) inner radius of the dust shell | $R_{out}$ccThe (inferred) outer radius of the dust shell | nddThe (input) exponent of the density power law ($\rho_{d}(r)\propto r^{-n}$) in the dust shell | $\tau_{V}$eeThe (input) dust shell’s optical depth at $0.55\micron$ | $F_{\rm bol}$ffBolometric Flux | MdggThe (inferred) circumstellar dust mass ---|---|---|---|---|---|--- (K) | (arcsec, au) | (arcsec, au) | | | (erg s-1 cm-2) | (M⊕) 1-shell model 232 | $0\farcs 095$, 71 | $\gtrsim 2\farcs 37$, 1785 | -0.05 | $9.9\times 10^{-4}$ | $6.5\times 10^{-9}$ | $1.15\times 10^{-2}$ 2-shell model 232 | $0\farcs 095$, 71 | $0\farcs 47$, 357 | 0.2 | $1.4\times 10^{-4}$ | $6.5\times 10^{-9}$ | $7.4\times 10^{-5}$ 125 | $0\farcs 47$, 357 | $\gtrsim 2\farcs 37$, 1785 | -0.4 | $9.6\times 10^{-4}$ | … | $1.45\times 10^{-2}$ Figure 1: Images of the central region of NGC 3132, taken with MIRI (a) F1800W, (b) F1280W, (c), F1130W, (d) F770W), and NIRCam (e) F187N (f) F212N (g) F356W (h) F470N, shown using a logarithmic strech for the intensity (MJy/sr). Black (white) dashed circles of diameter $1\farcs 0$ ($0\farcs 5$) locate the central white-dwarf star (CS) in the MIRI (NIRC) images; the A2V companion is located 1$\farcs$7 to the NE of the CS. Figure 2: Images of the dust cloud around the CS, taken with F1800W (a) pipeline image (logarithmic stretch), (b) PSF-sbtracted image (square-root strech). Contour levels are at 0.88, 0.50, 0.33, 0.25, 0.20, 0.167, 0.125, 0.10, 0.067, 0.05, 0.04, 0.033, 0.025 times the background-subtracted peak intensity (308 My/sr) in panel $a$; the median background level is 135 MJy/sr. Dashed box in panel $b$ demarcates the region of the “tail” feature. (c) Radial intensity cuts centered on the CS, averaged over 90° wedges pointing away (solid red curve) and towards (dashed red curve) from the A2V star. Brown curve shows the PSF extracted using a field star. Figure 3: As in Fig. 1, but with the A2V star subtracted. Figure 4: HST 0.55 µm (F555W) image of the CS and the A2V stars at the center of NGC 3132. Dashed yellow line shows the vector separating the two stars, with length $1\farcs 693$ and $PA=47.82\arcdeg$. North is up and East is to the left. Figure 5: Observed photometry (black symbols) and model spectral- energy-distributions (green curves) for the CS – panels $a$ and $b$ respectively show the 1-shell and 2-shell models. Panels $c$ and $d$ show the model SED at short wavelengths where it is dominated by direct starlight (reddened and attenuated by nebular extinction) and is unaffected by the low- optical depth dust cloud, together with observed spectrum (panel $c$) or photometry (panel $d$). The inset in panel $d$ shows the model in panel $a$ or $b$ (green curve), an M6V dwarf companion (red curve), and the addition of the two (black curve). Error bars ($\pm 15$%) on the observed photometry are conservative estimates. Figure 6: Observed (azimuthally-averaged) and model radial intensity distributions of the CS (+dust) of NGC 3132 as seen in the (a) F1800W, (b) F1280 and (c) F770W filters, together with (d) the radial distribution of density, visual optical depth, and dust temperature for the best-fit 1-shell model. All distributions have been normalized to their peak values. This 1-shell model uses a grain mixture consisting of 70% warm- silicates and 30% amorphous Carbon (amC). The model intensities (blue) have been convolved with the PSFs at 18, 12.8 and 7.7 µm, to generate the convolved models (green), for direct comparion with the observations (red). The sharp intensity peaks in the intrinsic model correspond to the inner radius of the dust shell. The F1280W and F770W intensities are shown over a smaller range of offsets compared to F1800W, because the observed radial intensity distributions for the former are much more narrow than for the latter. Figure 7: As in Fig. 6, but for a 2-shell model. We show here the large-scale structure of the PN, NGC 3132, imaged through the F1800W (18 µm) filter (Fig. A.1A). The image does not show the presence of any extended nebular feature (background or foreground) close to the compact dust clump around the central star (CS). Figure A.1A: The PN NGC 3132 imaged with MIRI, through the F1800W (18 µm) filter, displayed using a square-root stretch. The green dashed circle (diameter $5{{}^{\prime\prime}}$) encloses the dust cloud around the CS, and the A2V star to its North-East. The panel size is $80{{}^{\prime\prime}}\times 80{{}^{\prime\prime}}$.
aainstitutetext: School of Science, Huzhou University, Huzhou, Zhejiang 313000, Chinabbinstitutetext: National Astronomical Observatories, Chinese Academy of Sciences, Beijing, 100012, Chinaccinstitutetext: School of Fundamental Physics and Mathematical Sciences, Hangzhou Institute for Advanced Study, UCAS, Hangzhou 310024, Chinaddinstitutetext: International Centre for Theoretical Physics Asia-Pacific, Beijing/Hangzhou, China # Unitarity bounds on extensions of Higgs sector Bo-Qiang Lu b,c,d,2 Da Huang***Corresponding Author<EMAIL_ADDRESS><EMAIL_ADDRESS> ###### Abstract It is widely believed that extensions of the minimal Higgs sector is one of the promising directions for resolving many puzzles beyond the Standard Model (SM). In this work, we study the unitarity bounds on the models by extending the two-Higgs-doublet model with an additional real or complex Higgs triplet scalar. By noting that the SM gauge symmetries $SU(2)_{L}\times U(1)_{Y}$ are recovered at high energies, we can classify the two-body scattering states by decomposing the direct product of two scalar multiplets into their direct sum of irreducible representations of electroweak gauge groups. In such state bases, the s-wave amplitudes of two-body scalar scatterings can be written in the form of block-diagonalized scattering matrices. Then the application of the perturbative unitarity conditions on the eigenvalues of scattering matrices leads to the analytic constraints on the model parameters. Finally, we numerically investigated several specific models, finding that the perturbative unitarity places stringent bounds on the extra scalar mass spectrum, especially when there exist significant mixings among scalars in the extended Higgs sectors. ## 1 Introduction Despite the overwhelming success of the Standard Model (SM) by discovering the 125 GeV Higgs boson at the Large Hadron Collider (LHC) in 2012 ATLAS:2012yve ; CMS:2012qbp , it has been widely believed that new physics is required to explain various phenomena beyond the SM, such as tiny neutrino masses Formaggio:2021nfz , the nature of dark matter Planck:2015fie , and the origin of matter-antimatter asymmetry in the Universe Kuzmin:1985mm . One of the promising directions for resolving these puzzles is to extend the minimal SM Higgs sector by including additional scalars. Note that the shape of the SM Higgs potential is fully determined by the vacuum expectation value (VEV), $v$, and the quartic self-coupling, $\lambda_{H}$. However, in the non-minimal extension of the Higgs section, there will be inevitable deviations of the Higgs self-couplings with respect to the SM predictions. Therefore, the precise measurement of the Higgs self-couplings can help us to probe the new physics and to understand the electroweak symmetry breaking mechanism. Until now, the determination of the trilinear Higgs coupling has been performed at the LHC Run 2 and will be further searched for at the Run 3, by directly detecting the single and double Higgs boson productions Degrassi:2021uik and other indirect probes McCullough:2013rea ; Cao:2015oxx ; Bizon:2016wgr ; deBlas:2016ojx . In addition to the above experimental endeavors, there has already been considerable theoretical explorations in order to constrain the Higgs sector, such as the perturbative unitarity Gell-Mann:1969cuq ; Weinberg:1971fb ; Lee:1977yc ; Lee:1977eg , vacuum stability and triviality Cabibbo:1979ay ; Lindner:1985uk . In this work, we shall focus on the systematic derivation of the perturbative unitarity bounds on the non-minimal Higgs sector with two or three Higgs multiplets. As early as 1977, Lee, Quigg, and Thacker Lee:1977yc ; Lee:1977eg made use of the perturbative unitarity and found the Higgs boson mass upper bound $m_{h}<870$ GeV in the minimal SM. The perturbative unitarity has recently been calculated in various extensions of the Higgs sector and been identified as a significant constraint on the new physics. One of the most popular extensions is the two-Higgs-doublet model (2HDM) (see Refs. Branco:2011iw ; Wang:2022yhm for recent reviews) whose perturbative unitarity was firstly calculated in Refs. Huffel:1980sk ; Maalampi:1991fb ; Akeroyd:2000wc ; Arhrib:2000is with the assumptions of a softly broken $Z_{2}$ symmetry and $CP$ conservation. The perturbative unitarity for the most general 2HDM was given in Ref. Ginzburg:2005dt and the associated numerical investigationwas carried out in detail in Ref. Kanemura:2015ska . For other Beyond-SM theories, the unitarity bounds have been explored for the Georgi-Machacek model Georgi:1985nv in Ref. Aoki:2007ah , for the Type-II seesaw model Konetschny:1977bn ; Cheng:1980qt ; Magg:1980ut ; Schechter:1980gr ; Lazarides:1980nt ; Mohapatra:1980yp in Ref. Arhrib:2011uy , for extended scalar sector with a real triplet scalar in Ref. Khan:2016sxm , and for a complex triplet extension of the 2HDM with $CP$ conservation and a softly broken $Z_{2}$ symmetry in Ref. Ouazghour:2018mld , respectively. In this paper, we systematically study the unitarity bounds in extensions of the 2HDM by including an additional real Higgs triplet $\Sigma$ with hypercharge $Y_{\Sigma}=0$ or a complex Higgs triplet scalar $\Delta$ with $Y_{\Delta}=1$ in the most general setup, in order to ensure the validity of perturbation theory. Here we only concentrate on the high-energy limit where the SM gauge symmetry broken effects can be ignored. Thus, we can classify the two-scalar-particle states according to their conserved isospin and hypercharge quantum numbers, and construct the associated 2-to-2 scattering amplitude matrices in terms of the bases of $SU(2)_{L}\times U(1)_{Y}$ irreducible representations. Then we will consider the unitarity bounds in a few special cases, including the extensions of the SM or 2HDM by one additional real or complex scalar triplet, with or without a softly broken $Z_{2}$ symmetry. Finally, we will numerically apply our derived unitarity bounds to various concrete models, and show the corresponding constraints on the extended scalar potentials. This work is organized as follows. In Sec. 2, we provide the two-particle state eigenbasis according to the irreducible representations of given hypercharges and isospins. In Secs. 3 and 4, we present the scattering amplitude matrices in extensions of the 2HDM by including an additional real triplet and complex triplet, respectively. In Sec. 5, we show the unitarity bounds on the model parameters in the 2HDM with $CP$ conservation and a softly broken $Z_{2}$ symmetry, the complex 2HDM, and a complex triplet scalar extension of 2HDM, respectively. The conclusions are summarized in Sec. 6. We also include several appendix. In Appendix A, we provide the analytical solutions for eigenvalues of a 5-dimensional scattering matrix appearing in the real or complex triplet augment of the 2HDM with a softly broken $Z_{2}$ symmetry. In Appendix B, we provide elements of the mass matrices in the complex triplet extension of 2HDM. The relations between parameters in the generic scalar basis and the Higgs basis in this model are provided in Appendix C. Finally, the trilinear couplings of a neutral scalar with two charged Higgs particles are summarized in Appendix D. ## 2 Two-particle eigenstates and unitarity bounds on scattering matrices The calculation of the unitarity bounds in the minimal SM was firstly investigated in Refs. Lee:1977yc ; Lee:1977eg and has been applied to various extensions of the SM. It requires that the eigenvalues of this scattering matrix should be less than the unitarity limit Gell-Mann:1969cuq ; Weinberg:1971fb ; Ginzburg:2005dt , otherwise the perturbative calculation of scattering amplitudes at tree level is no more reliable. From another perspective, one can make a partial wave expansion of the scattering amplitudes for the interaction channels and put the unitarity bounds on the partial wave amplitudes. Concretely, the cross section of scalar scattering processes $s_{1}s_{2}\to s_{3}s_{4}$ can be expressed in terms of the partial wave decomposition as $\sigma=\frac{16\pi}{s}\sum_{l=1}^{\infty}(2l+1)\left|a_{l}(s)\right|^{2},$ (1) where $s$ is the Mandelstam variable and $a_{l}$ is the partial wave coefficients with the specific angular momenta $l$. Together with the optical theorem one finds the following bound of unitarity: $\left|\text{Re}\left(a_{l}\right)\right|<\frac{1}{2}\,,\quad\text{for all}~{}l.$ (2) In the high energy limit, it is found that the $s$-wave amplitude $a_{0}(s)$ is dominated by the point vertex processes since the $s$-, $t$-, $u$-channel processes are suppressed by the scattering energy. Furthermore, the equivalence theorem Cornwall:1973tb ; Cornwall:1974km ; Yao:1988aj ; Veltman:1989ud ; He:1992nga declares that at very high energy, the amplitudes of scattering processes involving longitudinal gauge bosons in the initial and final states are equivalent to those in which gauge bosons are replaced by the corresponding Nambu-Goldstone bosons. Thus, in the high energy limit $a_{0}(s)$ is fully determined by the quartic couplings of the scalar potential. Using the equivalence theorem, we can write down the two-particle state bases in terms of the components of the Higgs multiplets. Once given the scalar potential, we can determine the amplitudes for the $2\to 2$ scattering processes with the bases. This largely simplifies the calculations for scattering amplitudes. In Refs. Arhrib:2000is ; Aoki:2007ah ; Kanemura:2015ska ; Ginzburg:2005dt , the bases are further classified according to their electroweak (EW) charges, i.e., total hypercharge $Y$ and total isospin $I$, since the EW $SU(2)_{L}\times U(1)_{Y}$ gauge symmetries are recovered at high energies so that their associated quantum numbers becomes conserved again. In this approach, we decompose the direct product of two Higgs multiplets into the direct sums of irreducible representations under EW gauge symmetries. Field | $\Phi$ | $\tilde{\Phi}$ | $\Sigma$ | $\Delta$ | $\tilde{\Delta}$ ---|---|---|---|---|--- $SU(2)_{L}$ isospin | 2 | 2 | 3 | 3 | 3 Hypercharge | 1 | $-1$ | 0 | 2 | $-2$ Table 1: A summary of the quantum numbers of the Higgs multiplets. $\Phi$, $\Sigma$, and $\Delta$ denotes the $SU(2)_{L}$ doublet, real triplet, and complex triplet, respectively. We define $\tilde{\Phi}=i\tau_{2}\Phi^{*}$ and $\tilde{\Delta}_{ab}=(i\tau_{2})_{ac}(i\tau_{2})_{bd}(\Delta^{\dagger})^{cd}$, which have negative hypercharge. In this work, we adopt an intermediate route for the classification of the bases. Firstly, we classify the direct product of the two Higgs multiplets according to their total hypercharge, where the isospins and hypercharges of the Higgs multiplets considered in the present work are summarized in Table 1. Then we decompose the direct product into direct sums of the irreducible representations of the EW $SU(2)_{L}$ symmetry. There are three types of direct products of Higgs multiplets we are concerned about, which are given as follows: $2\otimes 2=1\oplus 3,~{}2\otimes 3=2\oplus 4,~{}{\rm and}~{}3\otimes 3=1\oplus 3\oplus 5.$ (3) In this way, we classify the two-particle bases according to their total isospins and hypercharges of the two Higgs multiplets. Furthermore, we express the bases of the irreducible representation in terms of components in the multiplets. The results are summarized in Tables 2-7, in which the eigenstates are rescaled so that they are normalized. Moreover, due to the symmetry property when exchanging two identical bosons, some representations of the two-particle eigenstates vanish, e.g., the $(Y,I)=(2,0)$ state in the $\Phi_{i}\times\Phi_{j}$ when $i=j$ in Table 2, the $(Y,I)=(0,1)$ state from $\Sigma\times\Sigma$ in Table 5, and the $(Y,I)=(4,0)$ state from $\Delta\times\Delta$ in Table 6. | $I=0$ | $I=1$ ---|---|--- $Y=0$ | $\frac{1}{\sqrt{2}}\left(w_{i}^{+}w_{j}^{-}+H_{i}^{0}H_{j}^{0*}\right)$ | $w_{i}^{+}H_{j}^{0*}$ $\frac{1}{\sqrt{2}}\left(-w_{i}^{+}w_{j}^{-}+H_{i}^{0}H_{j}^{0*}\right)$ $-H_{i}^{0}w_{j}^{-}$ $Y=2$ | $\frac{1}{\sqrt{2}}\left(-w_{i}^{+}H_{j}^{0}+H_{i}^{0}w_{j}^{+}\right)$ | $w_{i}^{+}w_{j}^{+}~{}\left(\times\frac{1}{\sqrt{2}}~{}{\rm for}~{}i=j\right)$ $\frac{1}{\sqrt{2}}\left(w_{i}^{+}H_{j}^{0}+H_{i}^{0}w_{j}^{+}\right)$ $H_{i}^{0}H_{j}^{0}~{}\left(\times\frac{1}{\sqrt{2}}~{}{\rm for}~{}i=j\right)$ Table 2: The bases of the irreducible representation for the two Higgs doublets direct product. The bases in the first and second row are corresponding to the direct product $\Phi_{i}\times\tilde{\Phi}_{j}$ $(Y=0)$ and $\Phi_{i}\times\Phi_{j}$ $(Y=2)$, respectively. Note that $i$ and $j$ indicate the Higgs doublet. We observe that the bases with $(Y=2,~{}I=2)$ vanish when the two Higgs doublets are identical, i.e., $i=j$. | $I=\frac{1}{2}$ | $I=\frac{3}{2}$ ---|---|--- $Y=1$ | $\sqrt{\frac{2}{3}}\left(-\frac{i}{\sqrt{2}}w^{+}\sigma^{0}+H^{0}\sigma^{+}\right)$ $\sqrt{\frac{2}{3}}\left(-w^{+}\sigma^{-}+\frac{i}{\sqrt{2}}H^{0}\sigma^{0}\right)$ | $w^{+}\sigma^{+}$ $\frac{1}{\sqrt{3}}\left(i\sqrt{2}w^{+}\sigma^{0}+H^{0}\sigma^{+}\right)$ $\frac{1}{\sqrt{3}}\left(w^{+}\sigma^{-}+i\sqrt{2}H^{0}\sigma^{0}\right)$ $H^{0}\sigma^{-}$ Table 3: The bases of the irreducible representation for the direct product of a Higgs doublet and a real Higgs triplet scalar, $\Phi\times\Sigma$. | $I=\frac{1}{2}$ | $I=\frac{3}{2}$ ---|---|--- $Y=1$ | $\sqrt{\frac{2}{3}}\left(-\frac{1}{\sqrt{2}}H^{0*}\delta^{+}+w^{-}\delta^{++}\right)$ $\sqrt{\frac{2}{3}}\left(-H^{0*}\delta^{0}-\frac{1}{\sqrt{2}}w^{-}\delta^{+}\right)$ | $-H^{0*}\delta^{++}$ $\frac{1}{\sqrt{3}}\left(\sqrt{2}H^{0*}\delta^{+}+w^{-}\delta^{++}\right)$ $\frac{1}{\sqrt{3}}\left(H^{0*}\delta^{0}-\sqrt{2}w^{-}\delta^{+}\right)$ $-w^{-}\delta^{0}$ $Y=3$ | $\sqrt{\frac{2}{3}}\left(-\frac{1}{\sqrt{2}}w^{+}\delta^{+}-H^{0}\delta^{++}\right)$ $\sqrt{\frac{2}{3}}\left(-w^{+}\delta^{0}+\frac{1}{\sqrt{2}}H^{0}\delta^{+}\right)$ | $-w^{+}\delta^{++}$ $\frac{1}{\sqrt{3}}\left(\sqrt{2}w^{+}\delta^{+}-H^{0}\delta^{++}\right)$ $\frac{1}{\sqrt{3}}\left(w^{+}\delta^{0}+\sqrt{2}H^{0}\delta^{+}\right)$ $H^{0}\delta^{0}$ Table 4: The bases of the irreducible representation for the direct product of a Higgs doublet and a complex Higgs triplet scalar. The bases in the first and second row are corresponding to the direct product $\tilde{\Phi}\times\Delta$ $(Y=1)$ and $\Phi\times\Delta$ $(Y=3)$, respectively. | $I=0$ | $I=1$ | $I=2$ ---|---|---|--- $Y=0$ | $\sqrt{\frac{2}{3}}\left(\sigma^{+}\sigma^{-}+\frac{1}{2}\sigma^{0}\sigma^{0}\right)$ | 0 | $\frac{1}{\sqrt{2}}\sigma^{+}\sigma^{+}$ $i\sigma^{+}\sigma^{0}$ $\frac{1}{\sqrt{3}}\left(\sigma^{+}\sigma^{-}-\sigma^{0}\sigma^{0}\right)$ $i\sigma^{-}\sigma^{0}$ $\frac{1}{\sqrt{2}}\sigma^{-}\sigma^{-}$ Table 5: The bases of the irreducible representation for the two real Higgs triplets direct product, $\Sigma\times\Sigma$. We only consider the case of the direct product of two identical triplet scalar (i.e., $i=j$). We find the bases with $I=1$ vanish in this case. | $I=0$ | $I=1$ | $I=2$ ---|---|---|--- $Y=0$ | $\frac{1}{\sqrt{3}}\left(\delta^{++}\delta^{--}+\delta^{+}\delta^{-}+\delta^{0}\delta^{0*}\right)$ | $\frac{1}{\sqrt{2}}\left(-\delta^{++}\delta^{-}+\delta^{+}\delta^{0*}\right)$ $-\frac{1}{\sqrt{2}}\left(\delta^{++}\delta^{--}-\delta^{0}\delta^{0*}\right)$ $-\frac{1}{\sqrt{2}}\left(-\delta^{+}\delta^{--}+\delta^{-}\delta^{0}\right)$ | $-\delta^{++}\delta^{0*}$ $\frac{1}{\sqrt{2}}\left(\delta^{++}\delta^{-}+\delta^{+}\delta^{0*}\right)$ $\frac{1}{\sqrt{6}}\left(-2\delta^{+}\delta^{-}+\delta^{++}\delta^{--}+\delta^{0}\delta^{0*}\right)$ $\frac{1}{\sqrt{2}}\left(-\delta^{+}\delta^{--}-\delta^{0}\delta^{-}\right)$ $-\delta^{--}\delta^{0}$ $Y=4$ | $\sqrt{\frac{2}{3}}\left(-\delta^{++}\delta^{0}-\frac{1}{2}\delta^{+}\delta^{+}\right)$ | 0 | $\frac{1}{\sqrt{2}}\delta^{++}\delta^{++}$ $-\delta^{++}\delta^{+}$ $\frac{1}{\sqrt{3}}\left(\delta^{+}\delta^{+}-\delta^{++}\delta^{0}\right)$ $\delta^{0}\delta^{+}$ $\frac{1}{\sqrt{2}}\delta^{0}\delta^{0}$ Table 6: The bases of the irreducible representation for the direct product of two complex Higgs triplet scalars. The bases in the first and second row are corresponding to the direct product $\Delta\times\tilde{\Delta}$ $(Y=0)$ and $\Delta\times\Delta$ $(Y=4)$, respectively. We observe again that the bases with $(Y=4,~{}I=1)$ vanish because the two triplets are identical. | $I=0$ | $I=1$ | $I=2$ ---|---|---|--- $Y=2$ | $\frac{1}{\sqrt{3}}\left(\sigma^{+}\delta^{0}-i\sigma^{0}\delta^{+}-\sigma^{-}\delta^{++}\right)$ | $-\frac{1}{\sqrt{2}}\left(\sigma^{+}\delta^{+}+i\sigma^{0}\delta^{++}\right)$ $-\frac{1}{\sqrt{2}}\left(\sigma^{+}\delta^{0}+\sigma^{-}\delta^{++}\right)$ $-\frac{1}{\sqrt{2}}\left(i\sigma^{0}\delta^{0}-\sigma^{-}\delta^{+}\right)$ | $\sigma^{-}\delta^{++}$ $\frac{1}{\sqrt{2}}\left(\sigma^{+}\delta^{+}-i\sigma^{0}\delta^{++}\right)$ $\frac{1}{\sqrt{6}}\left(\sigma^{+}\delta^{0}+i2\sigma^{0}\delta^{+}-\sigma^{-}\delta^{++}\right)$ $\frac{1}{\sqrt{2}}\left(i\sigma^{0}\delta^{0}+\sigma^{-}\delta^{+}\right)$ $\sigma^{-}\sigma^{0}$ Table 7: The bases of the irreducible representation for the direct product of a real Higgs triplet and a complex Higgs triplet, $\Sigma\times\Delta$. Based on the above two-particle basis, we can determine the $2\to 2$ scattering amplitudes as follows Ginzburg:2005dt $S_{(Y,I)}=\left\langle(\phi\phi)_{Y,I}^{f}|\hat{S}|(\phi\phi)_{Y,I}^{i}\right\rangle,$ (4) in the sector with definite EW charges $(Y,I)$ with $Y$ and $I$ as the total hypercharge and isospin, respectively. We do not distinguish states by the third components of isospin $I_{3}$, since the states with same $(Y,I)$ but different $I_{3}$ would lead to exactly the same scattering matrix. In the tree-level approximation, the elements of the scattering matrix among scalars are determined by the quartic couplings in the scalar potential. Here we do not decompose a complex scalar field into its real and imaginary parts either in the external state basis or in the scalar potential due to the recovered $SU(2)_{L}\times U(1)_{Y}$ symmetry at high energies. As argued before, the unitarity of scattering amplitudes requires that the $s$-wave amplitude $a_{0}(s)$ in the partial-wave expansion should fulfill the bound in Eq. (2). Note that the amplitude of two scalar scatterings is dominated by the $s$-wave one at the tree level, so that the unitarity bounds can be transformed into the following condition on the eigenvalues $\Lambda_{(Y,I)}$ of the scattering matrices $16\pi S_{(Y,I)}$ as follows $|\Lambda_{(Y,I)}|\leq 8\pi\,.$ (5) ## 3 Two-Higgs-doublet model plus a real triplet In this section, we will focus on the model that contains two Higgs doublets and a real Higgs triplet scalar. The scattering matrix for the scalar potential is provided in sec. 3.2. Using these results, we consider the perturbative unitarity constraints on two simplified cases: the model with a softly broken $Z_{2}$ symmetry and the $\Sigma$SM model FileviezPerez:2008bj . ### 3.1 The scalar potential The scalar potential for the extension of the 2HDM with an additional real Higgs triplet field $\Sigma$ is given by $V_{r}=V\left(\Phi_{1},\Phi_{2}\right)+V(\Sigma)+V\left(\Phi_{1},\Phi_{2},\Sigma\right),$ (6) where the Higgs doublet and real triplet scalar are $\Phi_{i}=\begin{pmatrix}w_{i}^{+}\\\ H_{i}^{0}\end{pmatrix},~{}{\rm and}~{}\Sigma=\left(\begin{array}[]{cc}\sigma^{0}/\sqrt{2}&\sigma^{+}\\\ \sigma^{-}&-\sigma^{0}/\sqrt{2}\end{array}\right),$ (7) where we can further expand the $H_{i}^{0}=\frac{1}{\sqrt{2}}(\varphi_{i}+iz_{i})$. The most general renormalizable scalar potential for the 2HDM in the generic basis $\\{\Phi_{1},\Phi_{2}\\}$ is commonly written as Davidson:2005cw $\displaystyle V\left(\Phi_{1},\Phi_{2}\right)=$ $\displaystyle m_{1}^{2}\Phi_{1}^{\dagger}\Phi_{1}+m_{2}^{2}\Phi_{2}^{\dagger}\Phi_{2}-\left(m_{12}^{2}\Phi_{1}^{\dagger}\Phi_{2}+\text{H.c.}\right)+\frac{1}{2}\lambda_{1}\left(\Phi_{1}^{\dagger}\Phi_{1}\right)^{2}$ (8) $\displaystyle+\frac{1}{2}\lambda_{2}\left(\Phi_{2}^{\dagger}\Phi_{2}\right)^{2}+\lambda_{3}\Phi_{1}^{\dagger}\Phi_{1}\Phi_{2}^{\dagger}\Phi_{2}+\lambda_{4}\Phi_{1}^{\dagger}\Phi_{2}\Phi_{2}^{\dagger}\Phi_{1}+\left[\frac{1}{2}\lambda_{5}\left(\Phi_{1}^{\dagger}\Phi_{2}\right)^{2}\right.$ $\displaystyle\left.+\lambda_{6}\left(\Phi_{1}^{\dagger}\Phi_{1}\right)\left(\Phi_{2}^{\dagger}\Phi_{1}\right)+\lambda_{7}\left(\Phi_{2}^{\dagger}\Phi_{2}\right)\left(\Phi_{2}^{\dagger}\Phi_{1}\right)+\text{H.c.}\right].$ The parameters $m_{12}^{2}$, $\lambda_{5}$, $\lambda_{6}$, and $\lambda_{7}$ should be real if we impose the CP conservation on the potential. If the $Z_{2}$ symmetry with $\Phi_{1}\to\Phi_{1}$ and $\Phi_{2}\to-\Phi_{2}$ is only softly broken by the term proportional to $m_{12}^{2}$, we should require $\lambda_{6}=\lambda_{7}=0$. The potential for the real Higgs triplet scalar is given by $V(\Sigma)=\frac{1}{2}m_{\Sigma}^{2}\operatorname{Tr}\Sigma^{2}+\frac{1}{4}\lambda_{\Sigma}\operatorname{Tr}\Sigma^{4}\,.$ (9) while the interactions between the Higgs doublets and the real triplet read as follows $\displaystyle V(\Phi_{1},\Phi_{2},\Sigma)$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left[a_{1}\Phi_{1}^{\dagger}\Sigma\Phi_{1}+a_{2}\Phi_{2}^{\dagger}\Sigma\Phi_{2}+\left(a_{12}\Phi_{1}^{\dagger}\Sigma\Phi_{2}+\text{H.c.}\right)\right]$ (10) $\displaystyle+$ $\displaystyle\frac{1}{2}\operatorname{Tr}\Sigma^{2}\left[\lambda_{8}\Phi_{1}^{\dagger}\Phi_{1}+\lambda_{9}\Phi_{2}^{\dagger}\Phi_{2}+\left(\lambda_{10}\Phi_{1}^{\dagger}\Phi_{2}+\text{H.c.}\right)\right]\,.$ For a real triplet, the possible terms $\operatorname{Tr}\left(\Sigma^{4}\right)$ and $\Phi^{\dagger}\Sigma^{2}\Phi$ are not independent since they can be expressed as the combination of $\left[\operatorname{Tr}\left(\Sigma^{2}\right)\right]^{2}$ and $\operatorname{Tr}\left(\Sigma^{2}\right)\Phi^{\dagger}\Phi$. Also, the potential cubic terms in the first line of Eq. (10) break the $Z_{2}^{\Sigma}$ symmetry: $\Sigma\to-\Sigma$, and are negligible for the $2\to 2$ scalar scattering in the high energy limit. Therefore, these terms play no roles in deriving the perturbative unitarity bounds. Furthermore, $\lambda_{10}$ can be a complex parameter and should vanish when the $Z_{2}$ symmetry involving the two Higgs doublets, softly-broken or not, is imposed. ### 3.2 Scattering matrix Based on the two-particle bases given in Tables 2, 3, and 5 classified according to the conserved quantum numbers $(Y,I)$, we can expand the general potential in Eq. (6) with the scalar components defined in Eq. (7) and obtain the following scattering matrices of given $(Y,I)$: $16\pi S_{(0,0)}=\left(\begin{array}[]{ccccc}3\lambda_{1}&2\lambda_{3}+\lambda_{4}&3\lambda_{6}&3\lambda_{6}^{*}&\sqrt{3}\lambda_{8}\\\ 2\lambda_{3}+\lambda_{4}&3\lambda_{2}&3\lambda_{7}&3\lambda_{7}^{*}&\sqrt{3}\lambda_{9}\\\ 3\lambda_{6}^{*}&3\lambda_{7}^{*}&\lambda_{3}+2\lambda_{4}&3\lambda_{5}^{*}&\sqrt{3}\lambda_{10}^{*}\\\ 3\lambda_{6}&3\lambda_{7}&3\lambda_{5}&\lambda_{3}+2\lambda_{4}&\sqrt{3}\lambda_{10}\\\ \sqrt{3}\lambda_{8}&\sqrt{3}\lambda_{9}&\sqrt{3}\lambda_{10}&\sqrt{3}\lambda_{10}^{*}&5\lambda_{\Sigma}\end{array}\right)$ (11) $16\pi S_{(0,1)}=\left(\begin{array}[]{cccc}\lambda_{1}&\lambda_{4}&\lambda_{6}&\lambda_{6}^{*}\\\ \lambda_{4}&\lambda_{2}&\lambda_{7}&\lambda_{7}^{*}\\\ \lambda_{6}^{*}&\lambda_{7}^{*}&\lambda_{3}&\lambda_{5}^{*}\\\ \lambda_{6}&\lambda_{7}&\lambda_{5}&\lambda_{3}\end{array}\right)$ (12) $16\pi S_{(0,2)}=2\lambda_{\Sigma}$ (13) $16\pi S_{(1,\frac{1}{2})}=16\pi S_{(1,\frac{3}{2})}=\left(\begin{array}[]{cc}\lambda_{8}&\lambda_{10}^{*}\\\ \lambda_{10}&\lambda_{9}\end{array}\right)$ (14) $16\pi S_{(2,0)}=\lambda_{3}-\lambda_{4}$ (15) $16\pi S_{(2,1)}=\left(\begin{array}[]{ccc}\lambda_{1}&\lambda_{5}^{*}&\sqrt{2}\lambda_{6}^{*}\\\ \lambda_{5}&\lambda_{2}&\sqrt{2}\lambda_{7}\\\ \sqrt{2}\lambda_{6}&\sqrt{2}\lambda_{7}^{*}&\lambda_{3}+\lambda_{4}\end{array}\right)$ (16) Comparing with the 2HDM results in Ref. Ginzburg:2005dt , the scattering matrix $16\pi S_{(0,0)}$ now becomes 5-dimensional, since there is an additional state with $(Y,I)=(0,0)$ composed solely by components in the triplet $\Sigma$. Furthermore, the scattering processes in the sectors with $(Y,I)=(0,2),~{}(1,\frac{1}{2})$, and $(1,\frac{3}{2})$ take place only between two scalar triplets. ### 3.3 The case with a $Z_{2}$ symmetry Now we simplify our discussion by imposing the softly broken $Z_{2}$ symmetry with $\Phi_{1}\to\Phi_{1}$ and $\Phi_{2}\to-\Phi_{2}$ on the scalar potential (6), so that we have $\lambda_{6}=\lambda_{7}=\lambda_{10}=0$ but leaving a nonzero $m_{12}^{2}$. Such a model is phenomenologically important because it protects the theory from flavor changing neutral currents at tree level. Using the results given in Appendix A, the matrix (11) can be block diagonalized into a $2\times 2$ matrix $16\pi S_{(0,0)}^{(2)}$ and a $3\times 3$ one $16\pi S_{(0,0)}^{(2)}$ as follows $16\pi S_{(0,0)}^{(2)}=\left(\begin{array}[]{cc}\lambda_{3}+2\lambda_{4}&3\lambda_{5}^{*}\\\ 3\lambda_{5}&\lambda_{3}+2\lambda_{4}\end{array}\right)\,,\,\,16\pi S_{(0,0)}^{(3)}=\left(\begin{array}[]{ccc}3\lambda_{1}&2\lambda_{3}+\lambda_{4}&\sqrt{3}\lambda_{8}\\\ 2\lambda_{3}+\lambda_{4}&3\lambda_{2}&\sqrt{3}\lambda_{9}\\\ \sqrt{3}\lambda_{8}&\sqrt{3}\lambda_{9}&5\lambda_{\Sigma}\end{array}\right)\,.$ (17) The eigenvalues for $16\pi S_{(0,0)}^{(3)}$ can be found numerically or analytically by applying Eq. (76). Furthermore, the matrix $16\pi S_{(0,1)}$ in Eq. (12) can also be decomposed into the following two matrices, $\displaystyle 16\pi S_{(0,1)}^{\text{u}}=\left(\begin{array}[]{cc}\lambda_{1}&\lambda_{4}\\\ \lambda_{4}&\lambda_{1}\end{array}\right)\,,\quad 16\pi S_{(0,1)}^{\text{d}}=\left(\begin{array}[]{cc}\lambda_{3}&\lambda_{5}^{*}\\\ \lambda_{5}&\lambda_{3}\end{array}\right)\,.$ (22) Apart from $16\pi S_{(0,0)}^{(3)}$, the eigenvalues for the scattering matrices are summarized as follows: $\displaystyle\Lambda_{(0,0)}^{(2)\pm}=\lambda_{3}+2\lambda_{4}\pm 3\left|\lambda_{5}\right|,$ (23) $\displaystyle\Lambda_{(0,1)}^{\text{u}\pm}=\frac{1}{2}\left(\lambda_{1}+\lambda_{2}\pm\sqrt{\left(\lambda_{1}-\lambda_{2}\right)^{2}+4\lambda_{4}^{2}}\right),$ $\displaystyle\Lambda_{(0,1)}^{\text{d}\pm}=\lambda_{3}\pm\left|\lambda_{5}\right|,$ $\displaystyle\Lambda_{(2,0)}=\lambda_{3}-\lambda_{4},~{}~{}\Lambda_{(2,1)}=\lambda_{3}+\lambda_{4},$ $\displaystyle\Lambda_{(2,1)}^{\pm}=\frac{1}{2}\left(\lambda_{1}+\lambda_{2}\pm\sqrt{\left(\lambda_{1}-\lambda_{2}\right)^{2}+4\left|\lambda_{5}\right|^{2}}\right),$ $\displaystyle\Lambda_{(0,2)}=2\lambda_{\Sigma},~{}~{}\Lambda_{(1,\frac{1}{2})}^{1}=\Lambda_{(1,\frac{3}{2})}^{1}=\lambda_{8},~{}~{}\Lambda_{(1,\frac{1}{2})}^{2}=\Lambda_{(1,\frac{3}{2})}^{2}=\lambda_{9},$ where $\Lambda_{(0,0)}^{2\pm}$, $\Lambda_{(0,1)}^{\text{u}\pm}$, and $\Lambda_{(0,1)}^{\text{d}\pm}$ are the eigenvalues for $16\pi S_{(0,0)}^{(2)}$, $16\pi S_{(0,1)}^{\text{u}}$, and $16\pi S_{(0,1)}^{\text{d}}$. By further assuming $\lambda_{8}=\lambda_{9}=\lambda_{\Sigma}=0$ in the matrix $S_{(0,0)}^{(3)}$, we can reproduce the eigenvalues $\Lambda_{00\pm}^{\text{even }}$ in Eq. (10) of Ref. Ginzburg:2005dt . The last line of Eq. (LABEL:eq:2HDMEV) gives the eigenvalues for the scattering matrices involving only components of the real triplet. Together with numerical eigenvalues of $16\pi S_{(0,0)}^{(3)}$, we have provided all eigenvalues for the model (6) with the softly broken $Z_{2}$ symmetry. ### 3.4 The $\Sigma$SM model The $\Sigma$SM model is a simple extension of the SM by a real triplet scalar, many aspects of which has been extensively investigated in the literature, such as the dark matter phenomenology FileviezPerez:2008bj ; YaserAyazi:2014jby ; Chiang:2020rcv , the LHC searches Wang:2013jba ; Bandyopadhyay:2014vma , and the strongly first-order electroweak phase transition Niemi:2018asa ; Bell:2020gug . Now we only have one Higgs doublet $\Phi$ which should be identified as the $\Phi_{1}$ in the 2HDM case. Thus, $\Sigma$SM can be reproduced by setting all the couplings in Eq. (6) involving the second Higgs doublet $\Phi_{2}$ to vanish, with the potential given by $V_{r}=V(\Phi)+V(\Sigma)+V(\Phi,\Sigma)\,,$ (24) where $V(\Sigma)$ is already given in Eq. (9) and $\displaystyle V(\Phi)$ $\displaystyle=$ $\displaystyle\mu^{2}\Phi^{\dagger}\Phi+\lambda_{\Phi}\left(\Phi^{\dagger}\Phi\right)^{2}\,,$ $\displaystyle V(\Phi,\Sigma)$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}a_{1}\Phi^{\dagger}\Sigma\Phi+\frac{\lambda_{8}}{2}\left({\rm Tr}\Sigma^{2}\right)\Phi^{\dagger}\Phi\,.$ (25) By using Eqs. (17) and (LABEL:eq:2HDMEV), the unitarity bounds on the $\Sigma$SM are then found to be $\displaystyle|\lambda_{\Phi}|\leq 4\pi,~{}|\lambda_{\Sigma}|\leq 4\pi,~{}|\lambda_{8}|\leq 8\pi,$ (26) $\displaystyle|6\lambda_{\Phi}+5\lambda_{\Sigma}\pm\sqrt{\left(6\lambda_{\Phi}-5\lambda_{\Sigma}\right)^{2}+12\lambda_{8}}|\leq 16\pi\,,$ which confirm the unitarity bounds provided in Ref. Khan:2016sxm . ## 4 Two-Higgs-doublet model plus a complex triplet In this section, we will consider the extension of 2HDM by a complex Higgs triplet $\Delta$ with $Y=2$ Chen:2021jok . By using the state bases provided in Sec. 2, we shall calculate the scattering matrix for the most general case of the model. We shall then impose the perturbative unitarity constraints on the eigenvalues of the scattering matrix for several simplified models, such as the one with a softly broken $Z_{2}$ symmetry and the Type-II seesaw model Konetschny:1977bn ; Cheng:1980qt ; Magg:1980ut ; Schechter:1980gr ; Lazarides:1980nt ; Mohapatra:1980yp ; Mohapatra:1999zr . ### 4.1 The general scalar potential The general scalar potential for the model with two Higgs doublets and a complex Higgs triplet scalar $\Delta$ is given by $V_{c}=V\left(\Phi_{1},\Phi_{2}\right)+V(\Delta)+V\left(\Phi_{1},\Phi_{2},\Delta\right),$ (27) where the complex Higgs triplet is written as $\Delta=\left(\begin{array}[]{cc}\delta^{+}/\sqrt{2}&\delta^{++}\\\ \delta^{0}&-\delta^{+}/\sqrt{2}\end{array}\right)\,.$ (28) Note that the neutral component $\delta^{0}$ is a complex scalar. The 2HDM potential $V\left(\Phi_{1},\Phi_{2}\right)$ has been provided in Eq. (8), while the part related to the self-interactions of the complex Higgs triplet is given by $V(\Delta)=m_{\Delta}^{2}\operatorname{Tr}\Delta^{\dagger}\Delta+\lambda_{\Delta 1}\left(\operatorname{Tr}\Delta^{\dagger}\Delta\right)^{2}+\lambda_{\Delta 2}\operatorname{Tr}\left(\Delta^{\dagger}\Delta\right)^{2}.$ (29) The third part in Eq. (27) gives the interactions among the Higgs doublets and the triplet Chen:2021jok $\displaystyle V\left(\Phi_{1},\Phi_{2},\Delta\right)=$ $\displaystyle\left(\mu_{1}\Phi_{1}^{T}i\tau_{2}\Delta^{\dagger}\Phi_{1}+\mu_{2}\Phi_{2}^{T}i\tau_{2}\Delta^{\dagger}\Phi_{2}+\mu_{3}\Phi_{1}^{T}i\tau_{2}\Delta^{\dagger}\Phi_{2}+\text{H.c.}\right)$ (30) $\displaystyle+\left[\lambda_{8}\Phi_{1}^{\dagger}\Phi_{1}+\lambda_{9}\Phi_{2}^{\dagger}\Phi_{2}+\left(\lambda_{10}\Phi_{1}^{\dagger}\Phi_{2}+\text{H.c.}\right)\right]\operatorname{Tr}\Delta^{\dagger}\Delta$ $\displaystyle+\lambda_{11}\Phi_{1}^{\dagger}\Delta\Delta^{\dagger}\Phi_{1}+\lambda_{12}\Phi_{2}^{\dagger}\Delta\Delta^{\dagger}\Phi_{2}+\left(\lambda_{13}\Phi_{1}^{\dagger}\Delta\Delta^{\dagger}\Phi_{2}+\text{H.c.}\right).$ The $Z_{2}^{\Delta}$ symmetry of the transformation $\Delta\to-\Delta$ is only softly broken by the cubic terms in the first line of Eq. (30), which are negligible for the $2\to 2$ scalar scatterings in the high energy limit, so that they cannot be constrained by the unitarity bounds. Note that there are possibly additional cubic interactions like $\Delta^{a}_{b}\Delta^{b}_{c}\Delta^{c}_{a}$. But we ignore them since they do not contribute to the unitarity bounds. The parameters $\lambda_{10}$ and $\lambda_{13}$ can be complex, and the associated interactions explicitly break the $Z_{2}$ symmetry involved in the two Higgs doublets. ### 4.2 Scattering matrix We expand the scalar potential (27) in terms of components in the two doublets and the triplet. With the two-particle eigenstates given in Tables. 2, 4, and 6, we can determine the scattering matrices for different conserved quantum numbers $(Y,I)$, which are summarized as follows: $16\pi S_{(0,0)}=\left(\begin{array}[]{ccccc}3\lambda_{1}&2\lambda_{3}+\lambda_{4}&3\lambda_{6}&3\lambda_{6}^{*}&\lambda_{a}\\\ 2\lambda_{3}+\lambda_{4}&3\lambda_{2}&3\lambda_{7}&3\lambda_{7}^{*}&\lambda_{b}\\\ 3\lambda_{6}^{*}&3\lambda_{7}^{*}&\lambda_{3}+2\lambda_{4}&3\lambda_{5}^{*}&\lambda_{c}^{*}\\\ 3\lambda_{6}&3\lambda_{7}&3\lambda_{5}&\lambda_{3}+2\lambda_{4}&\lambda_{c}\\\ \lambda_{a}&\lambda_{b}&\lambda_{c}&\lambda_{c}^{*}&\lambda_{\Delta}\end{array}\right)~{}{\rm with}~{}\left\\{\begin{matrix}\lambda_{a}&=&\sqrt{\frac{3}{2}}(2\lambda_{8}+\lambda_{11})\\\ \lambda_{b}&=&\sqrt{\frac{3}{2}}(2\lambda_{9}+\lambda_{12})\\\ \lambda_{c}&=&\sqrt{\frac{3}{2}}(2\lambda_{10}+\lambda_{13})\\\ \lambda_{\Delta}&=&2(4\lambda_{\Delta 1}+3\lambda_{\Delta 2})\end{matrix}\right.$ (31) $16\pi S_{(0,1)}=\left(\begin{array}[]{ccccc}\lambda_{1}&\lambda_{4}&\lambda_{6}&\lambda_{6}^{*}&\lambda_{11}\\\ \lambda_{4}&\lambda_{2}&\lambda_{7}&\lambda_{7}^{*}&\lambda_{12}\\\ \lambda_{6}^{*}&\lambda_{7}^{*}&\lambda_{3}&\lambda_{5}^{*}&\lambda_{13}^{*}\\\ \lambda_{6}&\lambda_{7}&\lambda_{5}&\lambda_{3}&\lambda_{13}\\\ \lambda_{11}&\lambda_{12}&\lambda_{13}&\lambda_{13}^{*}&2\lambda_{\Delta 1}+4\lambda_{\Delta 2}\end{array}\right)$ (32) $16\pi S_{(0,2)}=2\lambda_{\Delta 1}$ (33) $16\pi S_{(1,\frac{1}{2})}=\left(\begin{array}[]{cc}\lambda_{8}+3\lambda_{11}/2&\lambda_{10}+3\lambda_{13}/2\\\ \lambda_{10}^{*}+3\lambda_{13}^{*}/2&\lambda_{9}+3\lambda_{12}/2\end{array}\right)$ (34) $16\pi S_{(1,\frac{3}{2})}=\left(\begin{array}[]{cc}\lambda_{8}&\lambda_{10}\\\ \lambda_{10}^{*}&\lambda_{9}\end{array}\right)$ (35) $16\pi S_{(2,0)}=\lambda_{3}-\lambda_{4}$ (36) $16\pi S_{2,1}=\left(\begin{array}[]{ccc}\lambda_{1}&\lambda_{5}^{*}&\sqrt{2}\lambda_{6}\\\ \lambda_{5}&\lambda_{2}&\sqrt{2}\lambda_{7}^{*}\\\ \sqrt{2}\lambda_{6}^{*}&\sqrt{2}\lambda_{7}&\lambda_{3}+\lambda_{4}\end{array}\right)$ (37) $16\pi S_{(3,\frac{1}{2})}=\left(\begin{array}[]{cc}\lambda_{8}-\lambda_{11}/2&\lambda_{10}^{*}-\lambda_{13}^{*}/2\\\ \lambda_{10}-\lambda_{13}/2&\lambda_{9}-\lambda_{12}/2\end{array}\right)$ (38) $16\pi S_{(3,\frac{3}{2})}=\left(\begin{array}[]{cc}\lambda_{8}+\lambda_{11}&\lambda_{10}^{*}+\lambda_{13}^{*}\\\ \lambda_{10}+\lambda_{13}&\lambda_{9}+\lambda_{12}\end{array}\right)$ (39) $16\pi S_{(4,0)}=2\lambda_{\Delta 1}-\lambda_{\Delta 2}$ (40) $16\pi S_{(4,2)}=2(\lambda_{\Delta 1}+\lambda_{\Delta 2})$ (41) We observe that the scattering matrices $16\pi S_{(0,0)}$ and $16\pi S_{(0,1)}$ are now five-dimensional, which is compared with the four- dimensional matrices in the 2HDM. This is caused by the fact that irreducible representations of the product of two Higgs triplets contain the states with $(Y,I)=(0,0)$ and $(0,1)$, which can scatter into two components of Higgs doublets with the same quantum numbers. On the other hand, the scattering processes with $(Y,I)=(0,2),~{}(1,\frac{1}{2}),~{}(1,\frac{3}{2})$, $(3,\frac{1}{2}),~{}(3,\frac{3}{2})$, $(4,0)$, and $(4,2)$ take place only among the complex Higgs triplets. ### 4.3 Special case with a softly broken $Z_{2}$ symmetry Now we consider some simplified models in the above complex triplet extension of the 2HDM, which may allow us to obtain the eigenvalues of scattering matrices analytically. The first example is to impose a softly broken $Z_{2}$ symmetry on the potential of the two Higgs doublets in Eq. (27), so that the $m_{12}^{2}$, we have the condition $\lambda_{6}=\lambda_{7}=\lambda_{10}=\lambda_{13}=0$ for the potential. In this case, the $5\times 5$ scattering matrix (31) and (32) can be decomposed into a 2-dimensional (which is correspondingly denoted as $16\pi S_{(0,0)}^{(2)}$ and $16\pi S_{(0,1)}^{(2)}$) and a 3-dimensional matrices. The corresponding 2-dimensional and 3-dimensional matrices are given by $16\pi S_{(0,0)}^{(2)}=\left(\begin{array}[]{cc}\lambda_{3}+2\lambda_{4}&3\lambda_{5}^{*}\\\ 3\lambda_{5}&\lambda_{3}+2\lambda_{4}\end{array}\right),~{}16\pi S_{(0,1)}^{(2)}=\left(\begin{array}[]{cc}\lambda_{3}&\lambda_{5}^{*}\\\ \lambda_{5}&\lambda_{3}\end{array}\right).$ (42) $16\pi S_{(0,0)}^{(3)}=\left(\begin{array}[]{ccc}3\lambda_{1}&2\lambda_{3}+\lambda_{4}&\lambda_{a}\\\ 2\lambda_{3}+\lambda_{4}&3\lambda_{2}&\lambda_{b}\\\ \lambda_{a}&\lambda_{b}&\lambda_{\Delta}\end{array}\right),~{}16\pi S_{(0,1)}^{(3)}=\left(\begin{array}[]{ccc}\lambda_{1}&\lambda_{4}&\lambda_{11}\\\ \lambda_{4}&\lambda_{2}&\lambda_{12}\\\ \lambda_{11}&\lambda_{12}&2\lambda_{\Delta 1}+4\lambda_{\Delta 2}\end{array}\right).$ (43) It is convenient to find the eigenvalues for $16\pi S_{(0,0)}^{(3)}$ and $16\pi S_{(0,1)}^{(3)}$ numerically, we also provide the analytical solutions in Eq. (76). We collect the eigenvalues for the remaining scattering matrices as follows: $\displaystyle\Lambda_{(0,0)}^{2\pm}=\lambda_{3}+2\lambda_{4}\pm 3\left|\lambda_{5}\right|,$ (44) $\displaystyle\Lambda_{(0,1)}^{2\pm}=\lambda_{3}\pm\left|\lambda_{5}\right|,$ $\displaystyle\Lambda_{(0,2)}=2\lambda_{\Delta 1},$ $\displaystyle\Lambda_{(1,\frac{1}{2})}^{1}=\lambda_{8}+3\lambda_{11}/2,~{}~{}\Lambda_{(1,\frac{1}{2})}^{2}=\lambda_{9}+3\lambda_{12}/2,$ $\displaystyle\Lambda_{(1,\frac{3}{2})}^{1}=\lambda_{8},~{}~{}\Lambda_{(1,\frac{3}{2})}^{2}=\lambda_{9},$ $\displaystyle\Lambda_{(2,0)}=\lambda_{3}-\lambda_{4},~{}~{}\Lambda_{(2,1)}^{1}=\lambda_{3}+\lambda_{4},$ $\displaystyle\Lambda_{(2,1)}^{2\pm}=\frac{1}{2}\left(\lambda_{1}+\lambda_{2}\pm\sqrt{\left(\lambda_{1}-\lambda_{2}\right)^{2}+4\left|\lambda_{5}\right|^{2}}\right),$ $\displaystyle\Lambda_{(3,\frac{1}{2})}^{1}=\lambda_{8}-\lambda_{11}/2,~{}~{}\Lambda_{(3,\frac{1}{2})}^{2}=\lambda_{9}-\lambda_{12}/2,$ $\displaystyle\Lambda_{(3,\frac{3}{2})}^{1}=\lambda_{8}+\lambda_{11},~{}~{}\Lambda_{(3,\frac{3}{2})}^{2}=\lambda_{9}+\lambda_{12},$ $\displaystyle\Lambda_{(4,0)}=2\lambda_{\Delta 1}-\lambda_{\Delta 2},~{}~{}\Lambda_{(4,2)}=2(\lambda_{\Delta 1}+\lambda_{\Delta 2}),$ where $\Lambda_{(0,0)}^{2\pm}$ and $\Lambda_{(0,1)}^{2\pm}$ are the eigenvalues for $16\pi S_{(0,0)}^{(2)}$ and $16\pi S_{(0,1)}^{(2)}$, respectively. The remaining results in Eq. (44) represent the eigenvalues for the matrices (33)-(41). Combining with numerical eigenvalues of the matrices $16\pi S_{(0,0)}^{(3)}$ and $16\pi S_{(0,1)}^{(3)}$, we have provided all eigenvalues for the scattering matrices in the model (27) with a softly broken $Z_{2}$ symmetry. ### 4.4 The $\Delta$SM Another simple example belonging the present class is to consider a model with only one Higgs doublet and one complex triplet, the so-called $\Delta$SM, which has been widely employed to explain the tiny neutrino masses by the type-II seesaw mechanism Konetschny:1977bn ; Cheng:1980qt ; Magg:1980ut ; Schechter:1980gr ; Lazarides:1980nt ; Mohapatra:1980yp ; Mohapatra:1999zr ; Gu:2006wj ; Chao:2007mz ; FileviezPerez:2008jbu . The investigations of this model are extended to the searches at colliders Melfo:2011nx ; Chen:2013dh ; Han:2015sca ; Dev:2018sel ; Du:2018eaw ; Cheng:2022jyi , dark matter Ding:2017jdr and electroweak phase transition (EWPT) phenomena Arhrib:2011uy ; Primulando:2019evb ; Zhou:2022mlz . From potential (27), we can reproduce the Higgs sector of $\Delta$SM by keeping the quartic terms with couplings $\lambda_{1}$, $\lambda_{8}$, $\lambda_{11}$, $\lambda_{\Delta 1}$, and $\lambda_{\Delta 2}$ while setting all other quartic couplings to zero. Again, we take $\lambda_{1}=2\lambda_{\Phi}$ to reproduce the commonly seen SM Higgs potential in Eq. (3.4). By using the matrices in Eq. (43) and the eigenvalues in Eq. (44), the unitarity bounds for the $\Delta$SM are given by $\displaystyle|\lambda_{\Phi}|\leq 4\pi,~{}|\lambda_{\Delta 1}|\leq 4\pi,~{}|\lambda_{8}|\leq 8\pi$ (45) $\displaystyle|2\lambda_{\Delta 1}-\lambda_{\Delta 2}|\leq 8\pi,~{}|\lambda_{\Delta 1}+\lambda_{\Delta 2}|\leq 4\pi,$ $\displaystyle|\lambda_{8}+3\lambda_{11}/2|\leq 8\pi,~{}|\lambda_{8}-\lambda_{11}/2|\leq 8\pi,~{}|\lambda_{8}+\lambda_{11}|\leq 8\pi,$ $\displaystyle|\lambda_{\Phi}+\lambda_{\Delta 1}+2\lambda_{\Delta 2}\pm\sqrt{(\lambda_{\Phi}-\lambda_{\Delta 1}-2\lambda_{\Delta 2})+\lambda_{11}^{2}}|\leq 8\pi,$ $\displaystyle|6\lambda_{\Phi}+8\lambda_{\Delta 1}+6\lambda_{\Delta 2}\pm\sqrt{(6\lambda_{\Phi}-8\lambda_{\Delta 1}-6\lambda_{\Delta 2})^{2}+6(2\lambda_{8}+\lambda_{11})^{2}}|\leq 16\pi\,,$ which are in agreement with the results given in Ref. Arhrib:2011uy . ## 5 Applications The unitarity constraint on the quartic couplings can be translated into the upper bounds on the Higgs boson masses if $\sqrt{\lambda_{i}}v$ dominates the contributions to the Higgs boson masses. In this section, we will apply the perturbative unitarity to three specific models and show the quantitative constraints on the model parameters. ### 5.1 $CP$-conserving 2HDM with a $Z_{2}$ symmetry Let us firstly focus on the $CP$-conserving 2HDM with a softly broken $Z_{2}$ symmetry, in which the potential is given by Eq. (8) . In this simplified case, we have $\lambda_{6}=\lambda_{7}=0$ and all parameters in the scalar potential are real. We can rotate the generic scalar basis $\\{\Phi_{1},\Phi_{2}\\}$ into the Higgs basis $\\{H_{1},H_{2}\\}$ via the transformation $\left(\begin{array}[]{c}H_{1}\\\ H_{2}\end{array}\right)=\left(\begin{array}[]{cc}\cos\beta&\sin\beta\\\ -\sin\beta&\cos\beta\end{array}\right)\left(\begin{array}[]{c}\Phi_{1}\\\ \Phi_{2}\end{array}\right)$ (46) so that only $H_{1}$ has a non-vanishing VEV $v=\sqrt{v_{1}^{2}+v_{2}^{2}}\simeq 246$ GeV. Here the quantity $\tan\beta$ is defined by the ratio of two Higgs field VEVs, i.e., $\tan\beta=v_{2}/v_{1}$, and therefore, depends on the choices of the doublet field basis. In the most general 2HDM, $\tan\beta$ is of no particular importance since there is no preferred basis choice in this model Davidson:2005cw ; Haber:2006ue . On the other hand, in the case with a discrete $Z_{2}$ symmetry: $\Phi_{1}\to\Phi_{1}$ and $\Phi_{2}\to-\Phi_{2}$, $\tan\beta$ is a meaningful parameter that connects the $Z_{2}$ charge basis with the Higgs basis. In the following, we will use shorthanded notations $t_{\theta}\equiv\tan\theta$, $s_{\theta}\equiv\sin\theta$, and $c_{\theta}\equiv\cos\theta$, in which $\theta$ is an arbitrary angle. In the Higgs basis, the two doublet scalars can be expressed in components as follows, $H_{1}=\left(\begin{array}[]{c}H_{1}^{+}\\\ \frac{1}{\sqrt{2}}\left(v+h_{1}^{0}+iA_{1}\right)\end{array}\right),~{}~{}{\rm and}~{}~{}H_{2}=\left(\begin{array}[]{c}H^{+}\\\ \frac{1}{\sqrt{2}}\left(h_{2}^{0}+iA_{2}\right),\end{array}\right)$ (47) where $h_{1}^{0}$ and $h_{2}^{0}$ are CP-even neutral Higgs bosons, $A_{2}$ and $H^{+}$ are the physical neutral pseudoscalar and the charged scalar, respectively, while $H_{1}^{\pm}$and $A_{1}$ are the Goldstone bosons associated with the $W^{\pm}$ and $Z$ gauge bosons. Since $A_{2}$ here is a mass eigenstate, we will denote it as $A\equiv A_{2}$ for latter convenience. The scalar potential (8) in the Higgs basis $\\{H_{1},H_{2}\\}$ takes the same form as that in the generic scalar basis $\\{\Phi_{1},\Phi_{2}\\}$, with the coefficients in Eq. (8) being replaced by $m_{11}^{2}\to\tilde{m}_{11}^{2},~{}m_{22}^{2}\to\tilde{m}_{22}^{2},~{}m_{12}^{2}\to\tilde{m}_{12}^{2},~{}{\rm and}~{}\lambda_{i}\to\tilde{\lambda}_{i}~{}(i=1,2,...,7).$ (48) The explicit relations between two sets of parameters can be found in Appendix C. With Eq. (47), the scalar potential can be expanded in terms of component fields. By using the global minimum conditions for the 2HDM potential in the Higgs basis, we obtain $\tilde{m}_{11}^{2}=-\frac{1}{2}\tilde{\lambda}_{1}v^{2}~{}~{}{\rm and}~{}~{}\tilde{m}_{12}^{2}=\frac{1}{2}\tilde{\lambda}_{6}v^{2}.$ (49) With the second derivatives of the potential with respect to the component fields as in Appendix B, the masses of the pseudoscalar and singly-charged Higgs boson can be obtained as follows $\displaystyle m_{A}^{2}$ $\displaystyle=$ $\displaystyle\tilde{m}_{22}^{2}+\frac{v^{2}}{2}\left(\tilde{\lambda}_{3}+\tilde{\lambda}_{4}-\tilde{\lambda}_{5}\right),$ (50) $\displaystyle m_{H^{\pm}}^{2}$ $\displaystyle=$ $\displaystyle\tilde{m}_{22}^{2}+\frac{v^{2}}{2}\tilde{\lambda}_{3},$ (51) and the mass matrix for the CP-even Higgs scalars $h_{1}^{0}$ and $h_{2}^{0}$ is given by $\mathcal{M}_{h_{1}^{0}h_{2}^{0}}^{2}=\left(\begin{array}[]{cc}\tilde{\lambda}_{1}v^{2}&\tilde{\lambda}_{6}v^{2}\\\ \tilde{\lambda}_{6}v^{2}&m_{A_{2}}^{2}+\tilde{\lambda}_{5}v^{2}\end{array}\right).$ (52) The mass eigenstates $h$ and $H$ of the CP-even Higgs scalars $h_{1}^{0}$ and $h_{2}^{0}$ can be obtained by the rotation Haber:2006ue $\left(\begin{array}[]{c}h\\\ H\end{array}\right)=\left(\begin{array}[]{cc}s_{\beta-\alpha}&c_{\beta-\alpha}\\\ c_{\beta-\alpha}&-s_{\beta-\alpha}\end{array}\right)\left(\begin{array}[]{l}h_{1}^{0}\\\ h_{2}^{0}\end{array}\right),$ (53) where the mixing angle $\beta-\alpha$ is determined by $s_{2(\beta-\alpha)}=-\frac{2\tilde{\lambda}_{6}v^{2}}{m_{H}^{2}-m_{h}^{2}}\,.$ (54) The masses of $h$ and $H$ are given by $m_{H,h}^{2}=\frac{1}{2}\left\\{m_{A}^{2}+v^{2}\left(\tilde{\lambda}_{1}+\tilde{\lambda}_{5}\right)\pm\sqrt{\left[m_{A}^{2}+v^{2}\left(\tilde{\lambda}_{5}-\tilde{\lambda}_{1}\right)\right]^{2}+4v^{4}\tilde{\lambda}_{6}^{2}}\right\\}\,,$ (55) where $m_{h}=125$ GeV is the mass of the SM-like Higgs boson $h$. In the present work, we assume that the mass of $H$ is larger than that of the SM- like Higgs $h$. Combining Eq. (54) and Eq. (55) gives us the following relation $\tilde{\lambda}_{6}=\frac{t_{2(\beta-\alpha)}}{2v^{2}}\left[m_{A}^{2}+v^{2}\left(\tilde{\lambda}_{5}-\tilde{\lambda}_{1}\right)\right].$ (56) It is seen from Eq. (53) that the angle $\beta-\alpha$ represents the mixing between the two $CP$-even neutral Higgs states. In the alignment or decoupling limit, $s_{\beta-\alpha}$ or $\tilde{\lambda}_{6}$ approaches zero so that $h$ becomes more SM-like. Figure 1: Unitarity bound on the scalar masses. Left: scalar mass distributions as a function of $c_{\beta-\alpha}$. Right: scalar mass distributions as a function of $t_{\beta}$. The red, blue, and green scatter points denote the mass distributions for $H$, $A$, and $H^{\pm}$, respectively. Figure 2: Unitarity bound on the scalar mass distributions. It is well-known Ginzburg:2005dt ; Kanemura:2015ska that The perturbative unitarity can set strong limits on scalar masses in new-physics models. In order to achieve this, we make a random scan over model parameter space to numerically study the unitarity bounds. There are two distinctive sets of parameters: one set is those directly appearing in the scalar potential, including quartic couplings $\lambda_{i}$ while the other set contains the scalar masses and mixing angles, which are derived but more directly related to observations. In Figs. 1 and 2, the scan is performed for quartic couplings in the following ranges: $0<\lambda_{1},\lambda_{2}<8\pi,~{}{\rm and}~{}-8\pi<\lambda_{3},\lambda_{4},\lambda_{5}<8\pi.$ (57) Note that the bound $\lambda_{1},\lambda_{2}>0$ is the requirement from the positivity of the scalar potential. In the left plots of Figs. 1 and 2, we take $t_{\beta}=5$ and scan over the mixing angle $c_{\beta-\alpha}$ in the range $10^{-3}-10^{-1}$. The observations from collider searches have strongly constrained the mixing between the SM Higgs $h$ and the heavy one $H$. The recent global fit Herrero-Garcia:2019mcy ; Eberhardt:2020dat ; Athron:2021auq to the Higgs signal strengths and flavor measurements has restricted $c_{\beta-\alpha}\lesssim 10^{-1}$. Therefore, we only consider the case with a small mixing. On the other hand, in the right plot of Fig. 1, we fix $c_{\beta-\alpha}=0.02$ and scan over $t_{\beta}$ in the range of $10^{-1}-10$. The masses of scalars can be obtained with Eqs. (50)- (55) once the values of quartic couplings and mixing angles are given. In the left plot of Fig. 1, we show the scanning results on the scalar masses and $c_{\beta-\alpha}$ plane. The green, blue, and red points represent the masses for $H^{\pm}$, $A$, and $H$, respectively. As clearly shown by this figure, the unitarity puts upper bounds on scalar masses for a given $c_{\beta-\alpha}$. Specifically, when the mixing angle $c_{\beta-\alpha}\simeq 0.1$, the scalar masses have been restricted to be $\lesssim 1.4$ TeV. The bounds become weak with the decrease of $c_{\beta-\alpha}$. When the mixing angle $c_{\beta-\alpha}\simeq 10^{-3}$, the unitarity allowed upper bound climbs to around $\sim 12$ TeV. This feature can be easily understood as folows. From Eqs. (50) and (51) we observe that the masses for $H^{\pm}$, $A$ and $H$ can all be expressed by the form $\tilde{m}_{22}^{2}+\lambda_{i}v^{2}$. For small values of $\tilde{m}_{22}^{2}$, $\lambda_{i}v^{2}$ can dominate the contribution to scalar masses with large values of $\lambda_{i}$’s and $c_{\beta-\alpha}$ which are, however, strongly constrained by unitarity. Thus, the mass contributions from the term $\sqrt{\lambda_{i}}v$ should be smaller than $\sim$ TeV. On the other hand, for $m_{H,A,H^{\pm}}^{2}\gg v^{2}$, the scalar masses mainly come from the parameter $\tilde{m}_{22}$, which is not constrained by the perturbative unitarity. In this case, the scalar masses can be arbitrarily large. From Fig. 2 we can directly observe that the ranges of $m_{H}$, $m_{A}$, and $m_{H^{\pm}}$ become overlapped with each other when their values are much above several TeV. In the right plot of Fig. 1, we show the scan results on the scalar masses as a function of $t_{\beta}$. The upper limit on the scalar masses has its maximum value at $t_{\beta}\simeq 0.5$ and 3. When $t_{\beta}$ deviates from these two values, the unitarity bounds get stronger, so that the allowed mass ranges shrink. ### 5.2 Complex 2HDM Next we consider the so-called complex 2HDM (C2HDM) model WahabElKaffas:2007xd , in which the model admits a softly-broken $Z_{2}$ symmetry and the $CP$-violation by the complex $m_{12}$ and $\lambda_{5}$. Therefore, there are two $CP$-violating phases, which are defined by $m_{12}^{2}=\left|m_{12}^{2}\right|e^{i\phi\left(m_{12}^{2}\right)},\quad\lambda_{5}=\left|\lambda_{5}\right|e^{i\phi\left(\lambda_{5}\right)}\,.$ (58) However, these two phases are not independent. They are related to each other by one of the minimum conditions $2\operatorname{Im}\left(m_{12}^{2}\right)=v_{1}v_{2}\operatorname{Im}\left(\lambda_{5}\right).$ (59) As in the CP-conserving case, we can first rotate the scalar basis into the Higgs basis by Eq. (46). In the presence of the explicit $CP$ violation, $A_{2}$ is not yet a mass eigenstate, but would further mix with the scalars $h_{1}^{0}$ and $h_{2}^{0}$. This can be observed directly from the following off-diagonal elements in the neutral scalar mass matrix Fontes:2017zfn $m_{h_{1}^{0}A_{2}}^{2}=-\frac{1}{2}{\rm Im}\lambda_{5}s_{\beta}\,,\quad m_{h_{2}^{0}A_{2}}^{2}=-\frac{1}{2}{\rm Im}\lambda_{5}c_{\beta},$ (60) where the minimum conditions for the potential have been taken into account. We define an orthogonal rotation matrix $R$ to transform from the basis of the neutral components $\\{h_{1}^{0},h_{2}^{0},A_{2}\\}$ into the mass basis $\\{h,H,A\\}$, where the associated mass matrix is diagonalized as $R\mathcal{M}^{2}R^{\mathrm{T}}=\operatorname{diag}\left(m_{h}^{2},m_{H}^{2},m_{A}^{2}\right)$ with the rotation matrix $R$ given by WahabElKaffas:2007xd ; Fontes:2017zfn $R=\left(\begin{array}[]{ccc}c_{1}c_{2}&s_{1}c_{2}&s_{2}\\\ -\left(c_{1}s_{2}s_{3}+s_{1}c_{3}\right)&c_{1}c_{3}-s_{1}s_{2}s_{3}&c_{2}s_{3}\\\ -c_{1}s_{2}c_{3}+s_{1}s_{3}&-\left(c_{1}s_{3}+s_{1}s_{2}c_{3}\right)&c_{2}c_{3}\end{array}\right),$ (61) where $s_{i}=\sin\alpha_{i},~{}c_{i}=\cos\alpha_{i}(i=1,2,3)$ with the rotation angles defined in the following range $-\pi/2<\alpha_{1\,,2\,,3}\leq\pi/2\,.$ (62) The angles $\alpha_{2}$ and $\alpha_{3}$ parametrize the mixing between the pseudoscalar $A_{2}$ with two scalars $h_{1,2}^{0}$, which signals the $CP$ violation Khater:2003wq ; Inoue:2014nva . The five quartic couplings in the potential as functions of scalar masses and mixing angles have been provided in previous literature ElKaffas:2007rq . The ranges of other input parameters are given as follows: $m_{H}\in[126~{}{\rm GeV},15~{}{\rm TeV}]$, $m_{H^{\pm}}\in[10~{}{\rm GeV},15~{}{\rm TeV}]$, $t_{\beta}\in[10^{-1},10]$, and $\text{Re}m_{12}^{2}/(v^{2}s_{2\beta})\in[-10,10]$, where $m_{H}$ and $m_{H^{\pm}}$ stand for the masses of bosons $H$ and $H^{\pm}$, and $t_{\beta}$ for the VEV ratio between two Higgs doublet as usual. Here we have assumed the neutral state $h$ to be the SM-like Higgs with its mass to be $m_{h}=125$ GeV, while the mass squared of the third component $A$ is given by Fontes:2017zfn $m_{A}^{2}=\frac{m_{h}^{2}R_{13}\left(R_{12}t_{\beta}-R_{11}\right)+m_{H}^{2}R_{23}\left(R_{22}t_{\beta}-R_{21}\right)}{R_{33}\left(R_{31}-R_{32}t_{\beta}\right)},$ (63) where $R_{ij}$ are elements of the rotation matrix in Eq. (61). Figure 3: The perturbative unitarity bounds on the scalar masses. The colorbars in the left and right two plots represent the values of mixing angles $\alpha_{1}$ and $\alpha_{3}$, respectively. In Fig. 3, we have shown the results by imposing the perturbative unitarity bounds on the scalar masses in the C2HDM.It is shown that the upper limits on the masses of the scalar $H$, $A$, and $H^{\pm}$ from the unitarity are estimated as $m_{H}\lesssim 800~{}{\rm GeV},~{}~{}m_{A}\lesssim 2~{}{\rm TeV},~{}~{}{\rm and}~{}~{}m_{H^{\pm}}\lesssim 1~{}{\rm TeV}\,.$ (64) Comparing with the $CP$-conserving case, the unitarity bounds become stronger in the C2HDM. This is because both the neutral component $h_{2}^{0}$ and $A_{2}$ mix with $h_{1}^{0}$, the SM-like component of the Higgs boson. In most of parameter space, terms proportional to $\lambda_{i}v^{2}$ dominate the contribution to neutral Higgs masses squared, so that the constraints on quratic couplings from unitarity can significantly restrict the masses of neutral scalars. Note that the mass bounds in Eq. (64) becomes ineffective in the decoupling and $CP$-conserving limits, where the mixings among scalars vanish. The colorbars in the two left plots of Fig. 3 denotes the distributions of the mixing angle $\alpha_{1}$, which show that the model with a large $m_{A}$ tends to have a big value of $\alpha_{1}$. In the right two plots of Fig. 3, the colorbars represent the size of the mixing angle $\alpha_{3}$. We find that large values of $|\alpha_{3}|$ mainly concentrate around the upper boundary of the distribution of $m_{A}$, while small values of $|\alpha_{3}|$ mainly distribute around $m_{A}\sim 100$ GeV. ### 5.3 Complex triplet extension of the 2HDM The muon anomalous magnetic dipole moment (denoted by $(g-2)_{\mu}$) is one of the long-standing anomalies in the particle physics. The recent muon $g-2$ measurement performed by the Muon experiment at Fermilab Muong-2:2021ojo has further confirmed this discrepancy. Possible solutions to the muon $g-2$ anomaly has been widely discussed in the 2HDM content in Refs. Cheung:2001hz ; Cheung:2003pw ; Zhou:2001ew ; Aoki:2009ha ; Han:2022juu . At one-loop level, both the charged and neutral Higgs bosons in the 2HDM contribute to the muon $g-2$, but it is found that these corrections are too small to explain the observed deviation. On the other hand, the two-loop Barr-Zee diagrams can give rise to the dominant contribution to the muon $g-2$ in some parameter space. However, it has been shown that the explanation of the muon $g-2$ anomaly with the Barr-Zee mechanism requires a light pseudo-scalar mass with $m_{A}\lesssim 100$ GeV and $t_{\beta}\sim 50$ when various constraints are imposed Cheung:2003pw ; Ferreira:2021gke ; Kim:2022xuo . Note that one class of the strictest constraints is provided by the unitarity bounds in the theory. In particular, Ref. Ferreira:2021gke has shown that most of the parameter space with $m_{A}\gtrsim 100$ GeV in the typical 2HDM is already excluded by the unitarity alone. Figure 4: The Barr-Zee type Feynman diagram for the muon $g-2$, with charged scalar $\delta^{\pm}$ and $\delta^{\pm\pm}$ running in the loop. Figure 5: Upper: unitarity bound on the mass difference of $m_{\delta^{\pm}}-m_{\delta^{\pm\pm}}$. Lower: unitarity bound on the trilinear couplings $\lambda_{H\delta^{\pm}\delta^{\pm}}$ and $\lambda_{H\delta^{\pm\pm}\delta^{\pm\pm}}$. The colorbar represents the values of $\Delta a_{\mu}(\times 10^{-9})$. The recent work in Ref. Chen:2021jok shows that if a Higgs triplet with hypercharge $Y=2$ is added to the 2HDM, the charged components of the Higgs triplet can induce new Barr-Zee-type contributions illustrated in Fig. 4, which may explain the muon $g-2$ while easily evading other experimental constraints. From these Feynman diagrams, it is clear that the new contribution to muon $g-2$ is proportional to the trilinear scalar couplings $\lambda_{i}v$ which might be well constrained by the perturbative unitarity. However, Ref. Chen:2021jok has not appropriately taken into account the unitarity issue, which is the main topic in this subsection. Following Ref. Chen:2021jok , we will consider the decoupling limit of the model. By using the minimization conditions for the scalar potential in Eq. (27) and the mass matrices provided in Appendix B, we find that the decoupling between components in Higgs doublets and those in the triplet can be achieved when $v_{\Delta},~{}\tilde{\mu}_{3}\ll 1$ GeV. In the two-Higgs-doublet sector, we consider the case with a softly broken $Z_{2}$ symmetry and $CP$ conservation, in which all parameters in the scalar potential are real Also, following Ref. Chen:2021jok , we shall consider the aligned limit of the two Higgs doublets, i.e., $c_{\beta-\alpha}\approx 0$. In this case, the mass eigenstates $h$ and $H$ are almost $h_{1}^{0}$ and $h_{2}^{0}$ in Eq. (47), so that the trilinear scalar couplings are given by Eq. (D) in Appendix D. To calculate the Barr-Zee diagram shown in Fig. 4, we need to know the coupling between muons and $H$. As in Ref. Chen:2021jok , we shall consider the aligned two-Higgs-doublet model (A2HDM) case Pich:2009sp (see Ref. Eberhardt:2020dat for the recent global fit of A2HDM), in which the lepton Yukawa couplings with $H$ are given by $-{\cal L}_{Y}=\sum_{f}y_{f}^{H}\frac{M_{f}}{v}\bar{f}_{L}f_{R}H+{\rm H.c.},$ (65) where $M_{f}$ is the mass of the lepton flavor $f$ and $y_{f}^{H}=\left(s_{\beta-\alpha}\zeta_{f}-c_{\beta-\alpha}\right)\,.$ (66) Here $\zeta_{f}$ is a parameter in the A2HDM, whose benchmark value is taken to be $\zeta_{f}=-100$ following Ref. Chen:2021jok . The contribution to the muon $(g-2)$ from the Barr-Zee diagrams is given by Ilisie:2015tra $\Delta a_{\mu}=\sum_{\phi_{i}}\frac{\alpha m_{\mu}^{2}}{8\pi^{3}m_{H}^{2}}\operatorname{Re}\left(y_{f}^{H}\right)\lambda_{H\phi_{i}\phi_{i}^{*}}\mathcal{F}\left(\frac{m_{\phi_{i}}^{2}}{m_{H}^{2}}\right),$ (67) where $\phi_{i}=\delta^{\pm},~{}\delta^{\pm\pm}$, the trilinear couplings $\lambda_{H\phi_{i}\phi_{i}^{*}}$ are given in Eq. (D), and the loop function is given by $\mathcal{F}(\omega)=\frac{1}{2}\int_{0}^{1}dx\frac{x(x-1)}{\omega-x(1-x)}\ln\left(\frac{\omega}{x(1-x)}\right).$ (68) Since the Barr-Zee diagrams in Fig. 4 dominate the anomalous muon $g-2$, we can ignore other one- or two-loop $(g-2)_{\mu}$ contributions in our following numerical calculations. In order to search for the parameter space allowed by the perturbative unitarity, we scan over the quartic couplings $\lambda_{8}$, $\lambda_{9}$, $\lambda_{11}$, and $\lambda_{12}$ in the range of $(-8\pi-8\pi)$ and the doubly-charged scalar mass $m_{\delta^{\pm\pm}}$ in the range of $(10-1000)$ GeV. The mass of $\delta^{\pm}$ is determined by equations given in Appendix B. For the 2HDM sector, we take $\lambda_{1,2,...,5}=0.2$, $t_{\beta}=5$, and $m_{H}=300$ GeV for conservative estimations. We show the scan results in Fig. 5. From the upper two plots, we observe that the mass squared difference between the singly-charged and doubly-charged scalars in the triplet should be $|m^{2}_{\delta^{\pm}}-m^{2}_{\delta^{\pm\pm}}|/v^{2}\lesssim 6$, which is restricted by unitarity bound on the quartic coupling $\tilde{\lambda}_{11}$ as seen in Eqs. (B) and (82). Furthermore, for $m_{\delta^{\pm\pm}}\lesssim 200$ GeV we have $m_{\delta^{\pm}}-m_{\delta^{\pm\pm}}>0$. The colorbar of this figure represents the value distribution of $\Delta a_{\mu}$. The lower two plots of Fig. 5 show that large values of $|\Delta a_{\mu}|$ prefer large values of trilinear scalar couplings, $|\lambda_{H\delta^{\pm}\delta^{\pm}}|$ and $|\lambda_{H\delta^{\pm\pm}\delta^{\pm\pm}}|$, as well as small values of $m_{\delta^{\pm\pm}}$. Note that if $\Delta a_{\mu}$ is positive as required by experiments, it picks the parameter space with negative values of $\lambda_{H\delta^{\pm}\delta^{\pm}}$ and/or $\lambda_{H\delta^{\pm\pm}\delta^{\pm\pm}}$, which are well constrained by the unitarity consideration with $|\lambda_{H\delta^{\pm}\delta^{\pm}}|\lesssim 2.6$ and $|\lambda_{H\delta^{\pm\pm}\delta^{\pm\pm}}|\lesssim 5.0$. From the Barr-Zee Feynman diagrams and their expressions in Eq. (67), the trilinear couplings are directly related to the dominant contribution to the muon $g-2$, and perturbative unitarity can thus put very useful constraints on this model. Finally, we note that the unitarity bounds given in this section are rather conservative since the doublet-triplet mixings are ignored due to the nearly vanishing triplet VEV. In the case with $v_{\Delta}\sim 1$ GeV, the mixings between Higgs doublet and triplet components can become significant, which would further enhance the unitarity bounds on the scalar masses. ## 6 Conclusions The perturbative unitarity is one of the most significant theoretical constraints on the Higgs sector, beyond which the perturbation calculation in the theory breaks down. It has proven to be successful in predicting the upper limit on the Higgs boson mass in the minimal SM, and in constraining many new physics models such the 2HDM. In this work, we focus on deriving the perturbative unitarity bounds on several interesting extensions of the 2HDM with an additional real or complex Higgs triplet scalar. Since the total hypercharge and isospin are conserved in the high-energy limit of scatterings, we explicitly give the two-particle state basis according to their $SU(2)_{L}\times U(1)_{Y}$ charges by decomposing the direct product of two Higgs multiplets into direct sums of irreducible representations under the electroweak gauge groups. The classification of the two-particle state basis is summarized in Tables 2-7, in which the states are expressed in terms of component fields. With these two-particle bases, the $2\to 2$ scattering amplitudes among scalars can be simplified into the block-diagonal forms, which are easily determined by expanding the quartic scalar terms in the potential. We then impose the unitarity bounds on the eigenvalues of the scattering matrices. The associated analytical results are summarized in Secs. 3.2 and 4.2. We then numerically apply our derived unitarity bounds to several models of phenomenological interest, including the softly broken $Z_{2}$ symmetric 2HDM with and without $CP$-conservation, as well as the extension of 2HDM with a complex Higgs triplet. We have shown that the perturbative unitarity can put useful and stringent constraints on the Higgs scalar masses and mixings. The bounds become stronger when the mixings among Higgs scalars are significant. It is found that the neutral scalar masses in the C2HDM have been restricted to be below about $1-2$ TeV. Also, in the complex triplet extension of the 2HDM, the unitarity bounds on the couplings can also constrain new solutions to the long-standing muon $g-2$ anomaly. In the near future, together with the experimental measurements of the Higgs trilinear coupling and the Higgs signal strengths of different channels at the LHC Run 3, we hope that the unitarity bounds would help us to understand the structure of Higgs sector more deeply. ## Acknowledgments BQL is supported in part by the Huzhou University under startup Grant No. RK21094 and National Natural Science Foundation of China (NSFC) under Grant No. 12147219, No. 12175066, and No. 11975009. DH is supported in part by the National Natural Science Foundation of China (NSFC) under Grant No. 12005254, the National Key Research and Development Program of China under Grant No. 2021YFC2203003, and the Key Research Program of Chinese Academy of Sciences under grant No. XDPB15 ## Appendix A Eigenvalues of the $(Y,I)=(0,0)$ scattering matrix in the triplet extension of the 2HDM with a $Z_{2}$ symmetry In this section we analytically solve the eigenvalues for the 5-dimensional scattering matrix, which appears in the $(Y,I)=(0,0)$ sector of the extension of 2HDM with a softly-broken $Z_{2}$ symmetry. By imposing the $Z_{2}$ symmetry, the 5-dimensional scattering matrix in Eq. (31) can be simplified into the following form $X=\left(\begin{array}[]{ccccc}a_{1}&a_{2}&0&0&c_{5}\\\ a_{3}&a_{4}&0&0&c_{6}\\\ 0&0&b_{1}&b_{2}&0\\\ 0&0&b_{3}&b_{4}&0\\\ c_{5}&c_{6}&0&0&c_{7}\end{array}\right)\,,$ (69) which can be further decomposed into a $2\times 2$ matrix and a $3\times 3$ one as follows $X^{(2)}=\left(\begin{array}[]{cc}b_{1}&b_{2}\\\ b_{3}&b_{4}\\\ \end{array}\right),~{}~{}X^{(3)}=\left(\begin{array}[]{ccc}a_{1}&a_{2}&c_{5}\\\ a_{3}&a_{4}&c_{6}\\\ c_{5}&c_{6}&c_{7}\end{array}\right).$ (70) The eigenvalues for $X^{(2)}$ and $X^{(3)}$ are the same as that directly obtained from the 5-dimensional matrix $X$. Note that, for a general matrix $A$, the eigenvalue $f$ can be obtained by solving the equation $|fI-A|=0\,.$ (71) For $X^{(2)}$ and $X^{(3)}$, the eigenvalue equation can be transformed into the following equations $(f-b_{1})(f-b_{4})-b_{2}b_{3}=0$ (72) $(f-a_{1})(f-a_{4})(f-c_{7})-(f-a_{1})c_{6}^{2}-a_{2}a_{3}(f-c_{7})-a_{2}c_{5}c_{6}-a_{3}c_{5}c_{6}-(f-a_{4})c_{5}^{2}=0\,,$ (73) respectively. The solutions for Eq. (72) can be easily solved by $f_{1,2}=\frac{1}{2}\left(b_{1}+b_{4}\pm\sqrt{(b_{1}-b_{4})^{2}+4b_{2}b_{3}}\right).$ (74) For the case with $b_{1}=b_{4}$ and $b_{2}=b_{3}^{*}$, we have $f_{1,2}=b_{1}\pm|b_{2}|.$ (75) Note that the scattering matrix should be Hermitian, which means $a_{2}=a_{3}^{*}$ and $b_{2}=b_{3}^{*}$. Furthermore, the eigenvalues for the Hermitian matrix are always real. Note that, for a general cubic equation $f^{3}+bf^{2}+cf+d=0$, one representations of the three roots is given by $\displaystyle f_{3}=-\frac{b}{3}+2\sqrt[3]{r}\cos\theta\,,$ (76) $\displaystyle f_{4}=-\frac{b}{3}+2\sqrt[3]{r}\cos\left(\theta+\frac{2}{3}\pi\right)\,,$ $\displaystyle f_{5}=-\frac{b}{3}+2\sqrt[3]{r}\cos\left(\theta+\frac{4}{3}\pi\right)\,,$ where $r=\sqrt{-\left(\frac{p}{3}\right)^{3}},~{}\theta=\frac{1}{3}\arccos\left(-\frac{q}{2r}\right),~{}{\rm with}~{}p=\frac{3c-b^{2}}{3},~{}q=\frac{27d-9bc+2b^{3}}{27}\,.$ (77) By comparing Eq. (73) with the general cubic equation, it is found that $\displaystyle b=-(a_{1}+a_{4}+c_{7})\,,$ (78) $\displaystyle c=a_{1}a_{4}+a_{1}c_{7}+a_{4}c_{7}-a_{2}a_{3}-c_{5}^{2}-c_{6}^{2}\,,$ $\displaystyle d=a_{1}c_{6}^{2}+a_{2}a_{3}c_{7}+a_{4}c_{5}^{2}-a_{2}c_{5}c_{6}-a_{3}c_{5}c_{6}\,.$ In this way, we give the analytic solutions to the eigenvalues for the three- rank scattering matrix $X^{(3)}$. ## Appendix B Mass matrices in the complex triplet extension of 2HDM Here we provide some of the mass matrices elements for the extension of 2HDM with an additional complex Higgs triplet. In this appendix we express the neutral scalar in the triplet (28) as $\delta^{0}=\frac{1}{\sqrt{2}}(d^{0}+i\eta^{0})$. The mass matrix elements for the $CP$-even components of the neutral scalars in the model are given by $\displaystyle m_{h_{1}^{0}h_{1}^{0}}^{2}$ $\displaystyle=$ $\displaystyle\frac{3\tilde{\lambda}_{1}v^{2}}{2}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{8}}{2}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{11}}{2}-\sqrt{2}v_{\Delta}\tilde{\mu}_{1}+\tilde{m}_{1}^{2},$ $\displaystyle m_{h_{2}^{0}h_{2}^{0}}^{2}$ $\displaystyle=$ $\displaystyle\frac{\tilde{\lambda}_{3}v^{2}}{2}+\frac{\tilde{\lambda}_{4}v^{2}}{2}+\frac{1}{4}v^{2}{\rm Re}\tilde{\lambda}_{5}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{9}}{2}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{12}}{2}-\sqrt{2}v_{\Delta}\tilde{\mu}_{2}+\tilde{m}_{2}^{2},$ $\displaystyle m_{d^{0}d^{0}}^{2}$ $\displaystyle=$ $\displaystyle 3v_{\Delta}^{2}\tilde{\lambda}_{\Delta 1}+3v_{\Delta}^{2}\tilde{\lambda}_{\Delta 2}+\tilde{m}_{\Delta}^{2}+\frac{\tilde{\lambda}_{8}v^{2}}{2}+\frac{\tilde{\lambda}_{11}v^{2}}{2},$ $\displaystyle m_{h_{1}^{0}h_{2}^{0}}^{2}$ $\displaystyle=$ $\displaystyle\frac{3}{4}v^{2}{\rm Re}\tilde{\lambda}_{6}+\frac{1}{4}v_{\Delta}^{2}{\rm Re}\tilde{\lambda}_{10}+\frac{1}{4}v_{\Delta}^{2}{\rm Re}\tilde{\lambda}_{13}-\frac{v_{\Delta}{\rm Re}\tilde{\mu}_{3}}{2\sqrt{2}}-\frac{{\rm Re}\tilde{m}_{12}^{2}}{2},$ $\displaystyle m_{h_{1}^{0}d^{0}}^{2}$ $\displaystyle=$ $\displaystyle v_{\Delta}\tilde{\lambda}_{8}v+v_{\Delta}\tilde{\lambda}_{11}v-\sqrt{2}\tilde{\mu}_{1}v,$ $\displaystyle m_{h_{2}^{0}d^{0}}^{2}$ $\displaystyle=$ $\displaystyle\frac{1}{2}v_{\Delta}v{\rm Re}\tilde{\lambda}_{10}+\frac{1}{2}v_{\Delta}v{\rm Re}\tilde{\lambda}_{13}-\frac{v{\rm Re}\tilde{\mu}_{3}}{2\sqrt{2}}\,,$ (79) while the mass matrix elements for the $CP$-odd components are shown as follows $\displaystyle m_{A_{1}^{0}A_{1}^{0}}^{2}$ $\displaystyle=$ $\displaystyle\frac{\tilde{\lambda}_{1}v^{2}}{2}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{8}}{2}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{11}}{2}+\sqrt{2}v_{\Delta}\tilde{\mu}_{1}+\tilde{m}_{1}^{2},$ $\displaystyle m_{A_{2}^{0}A_{2}^{0}}^{2}$ $\displaystyle=$ $\displaystyle\frac{\tilde{\lambda}_{3}v^{2}}{2}+\frac{\tilde{\lambda}_{4}v^{2}}{2}-\frac{1}{4}v^{2}{\rm Re}\tilde{\lambda}_{5}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{9}}{2}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{12}}{2}+\sqrt{2}v_{\Delta}\tilde{\mu}_{2}+\tilde{m}_{2}^{2},$ $\displaystyle m_{\eta^{0}\eta^{0}}^{2}$ $\displaystyle=$ $\displaystyle v_{\Delta}^{2}\tilde{\lambda}_{\Delta 1}+v_{\Delta}^{2}\tilde{\lambda}_{\Delta 2}+\tilde{m}_{\Delta}^{2}+\frac{\tilde{\lambda}_{8}v^{2}}{2}+\frac{\tilde{\lambda}_{11}v^{2}}{2},$ $\displaystyle m_{A_{1}^{0}A_{2}^{0}}^{2}$ $\displaystyle=$ $\displaystyle\frac{v_{\Delta}^{2}{\rm Re}\tilde{\lambda}_{10}}{4}+\frac{v_{\Delta}^{2}{\rm Re}\tilde{\lambda}_{13}}{4}+\frac{v_{\Delta}{\rm Re}\tilde{\mu}_{3}}{2\sqrt{2}}-\frac{{\rm Re}\tilde{m}_{12}^{2}}{2}+\frac{{\rm Re}\tilde{\lambda}_{6}v^{2}}{4},$ $\displaystyle m_{A_{1}^{0}\eta^{0}}^{2}$ $\displaystyle=$ $\displaystyle-\sqrt{2}\tilde{\mu}_{1}v,$ $\displaystyle m_{A_{2}^{0}\eta^{0}}^{2}$ $\displaystyle=$ $\displaystyle-\frac{{\rm Re}\tilde{\mu}_{3}v}{2\sqrt{2}}.$ (80) The mass matrix elements for the singly-charged scalars are $\displaystyle m_{H_{1}^{+}H_{1}^{-}}^{2}$ $\displaystyle=$ $\displaystyle\frac{\tilde{\lambda}_{1}v^{2}}{2}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{8}}{2}+\tilde{m}_{1}^{2},$ $\displaystyle m_{H_{2}^{+}H_{2}^{-}}^{2}$ $\displaystyle=$ $\displaystyle\frac{\tilde{\lambda}_{3}v^{2}}{2}+\frac{v_{\Delta}^{2}\tilde{\lambda}_{9}}{2}+\tilde{m}_{2}^{2},$ $\displaystyle m_{\delta^{+}\delta^{-}}^{2}$ $\displaystyle=$ $\displaystyle\frac{\tilde{\lambda}_{8}v^{2}}{2}+\frac{\tilde{\lambda}_{11}v^{2}}{4}+v_{\Delta}^{2}\tilde{\lambda}_{\Delta 1}+v_{\Delta}^{2}\tilde{\lambda}_{\Delta 2}+\tilde{m}_{\Delta}^{2},$ $\displaystyle m_{H_{1}^{+}H_{2}^{-}}^{2}$ $\displaystyle=$ $\displaystyle\frac{1}{2}v^{2}\tilde{\lambda}_{6}^{*}+\frac{1}{2}v_{\Delta}^{2}\tilde{\lambda}_{10}^{*}-\tilde{m}_{12}^{2*}=m_{H_{1}^{-}H_{2}^{+}}^{2*},$ $\displaystyle m_{H_{1}^{+}\delta^{-}}^{2}$ $\displaystyle=$ $\displaystyle\frac{v_{\Delta}\tilde{\lambda}_{11}v}{2\sqrt{2}}-\tilde{\mu}_{1}v=m_{H_{1}^{-}\delta^{+}}^{2*},$ $\displaystyle m_{H_{2}^{+}\delta^{-}}^{2}$ $\displaystyle=$ $\displaystyle\frac{v_{\Delta}\tilde{\lambda}_{13}v}{2\sqrt{2}}-\frac{\tilde{\mu}_{3}v}{2}=m_{H_{2}^{-}\delta^{+}}^{2*}.$ (81) Since there is only one doubly-charged scalar in the model, so there is not any mixing and its mass is simply given by $m_{\delta^{++}\delta^{--}}^{2}=v_{\Delta}^{2}\tilde{\lambda}_{\Delta 1}+\tilde{m}_{\Delta}^{2}+\frac{\tilde{\lambda}_{8}v^{2}}{2}$ (82) The parameters with tilde denote those in the Higgs basis, with the transformation relation to the parameters in the generic basis given in Appendix C. Due to $CP$-violating effects, there may exist mixings between the $CP$-even and $CP$-odd components in the neutral scalars. Since we do not use them in our work, we do not provide their explicit formulae here. ## Appendix C Parameters in the Higgs basis for the complex triplet extension of 2HDM The electroweak gauge symmetry is spontaneously broken when the neutral components of the Higgs multiplets obtain VEVs. The VEVs can be complex and there may be a relative phase between them. Here we use $\xi$ to denote the phase between the VEVs of doublets $\Phi_{1}$ and $\Phi_{2}$ in the triplet extension of the 2HDM. Concretely, one assumes real $v_{1}$ and complex $v_{2}e^{i\xi}$. Such a phase can be absorbed by the following phase redefinitions of the complex parameters: $\lambda_{5}\rightarrow e^{2i\xi}\lambda_{5}\text{ and }m_{12}^{2},\lambda_{6},\lambda_{7},\lambda_{10},\lambda_{13}\rightarrow e^{i\xi}\left\\{m_{12}^{2},\lambda_{6},\lambda_{7},\lambda_{10},\lambda_{13}\right\\}.$ (83) so that the form of the potential keeps unchanged. Thus, we can start with real VEVs for scalars. In the following, we summarize the parameters of the potential in the Higgs basis as functions of those defined in the generic basis. For the parameters with mass dimensions, we have $\displaystyle\tilde{m}_{11}^{2}$ $\displaystyle=$ $\displaystyle c_{\beta}^{2}m_{11}^{2}+s_{\beta}^{2}m_{2}^{2}-c_{\beta}s_{\beta}\left(m_{12}^{2}+m_{12}^{2*}\right),$ $\displaystyle\tilde{m}_{22}^{2}$ $\displaystyle=$ $\displaystyle s_{\beta}^{2}m_{11}^{2}+c_{\beta}^{2}m_{2}^{2}+c_{\beta}s_{\beta}\left(m_{12}^{2}+m_{12}^{2*}\right),$ $\displaystyle\tilde{m}_{12}^{2}$ $\displaystyle=$ $\displaystyle c_{\beta}s_{\beta}\left(m_{11}^{2}-m_{22}^{2}\right)+\left(c_{\beta}^{2}m_{12}^{2}-s_{\beta}^{2}m_{12}^{2*}\right),$ $\displaystyle\tilde{m}_{\Delta}^{2}$ $\displaystyle=$ $\displaystyle m_{\Delta}^{2},$ $\displaystyle\tilde{\mu}_{1}$ $\displaystyle=$ $\displaystyle\mu_{1}c_{\beta}^{2}+\mu_{3}c_{\beta}s_{\beta}+\mu_{2}s_{\beta}^{2},$ $\displaystyle\tilde{\mu}_{2}$ $\displaystyle=$ $\displaystyle\mu_{2}c_{\beta}^{2}-\mu_{3}c_{\beta}s_{\beta}+\mu_{1}s_{\beta}^{2},$ $\displaystyle\tilde{\mu}_{3}$ $\displaystyle=$ $\displaystyle\mu_{3}\left(c_{\beta}^{2}-s_{\beta}^{2}\right)+2\left(\mu_{2}-\mu_{1}\right)c_{\beta}s_{\beta}.$ (84) The quartic couplings that relate only to the two Higgs doublets are given by $\displaystyle\tilde{\lambda}_{1}$ $\displaystyle=$ $\displaystyle\lambda_{1}c_{\beta}^{4}+\lambda_{2}s_{\beta}^{4}+2\left(\lambda_{3}+\lambda_{4}+{\rm Re}\lambda_{5}\right)c_{\beta}^{2}s_{\beta}^{2}+4{\rm Re}\lambda_{6}c_{\beta}^{3}s_{\beta}+4{\rm Re}\lambda_{7}c_{\beta}s_{\beta}^{3},$ $\displaystyle\tilde{\lambda}_{2}$ $\displaystyle=$ $\displaystyle\lambda_{1}s_{\beta}^{4}+\lambda_{2}c_{\beta}^{4}+2\left(\lambda_{3}+\lambda_{4}+{\rm Re}\lambda_{5}\right)c_{\beta}^{2}s_{\beta}^{2}-4{\rm Re}\lambda_{6}c_{\beta}s_{\beta}^{3}-4{\rm Re}\lambda_{7}c_{\beta}^{3}s_{\beta},$ $\displaystyle\tilde{\lambda}_{3}$ $\displaystyle=$ $\displaystyle\frac{1}{4}s_{2\beta}^{2}\left[\lambda_{1}+\lambda_{2}-2\left(\lambda_{3}+\lambda_{4}+{\rm Re}\lambda_{5}\right)\right]+\lambda_{3}-\left({\rm Re}\lambda_{6}-{\rm Re}\lambda_{7}\right)c_{2\beta}s_{2\beta},$ $\displaystyle\tilde{\lambda}_{4}$ $\displaystyle=$ $\displaystyle\frac{1}{4}s_{2\beta}^{2}\left[\lambda_{1}+\lambda_{2}-2\left(\lambda_{3}+\lambda_{4}+{\rm Re}\lambda_{5}\right)\right]+\lambda_{4}-\left({\rm Re}\lambda_{6}-{\rm Re}\lambda_{7}\right)c_{2\beta}s_{2\beta},$ $\displaystyle\tilde{\lambda}_{5}$ $\displaystyle=$ $\displaystyle\left[\lambda_{1}+\lambda_{2}-2\left(\lambda_{3}+\lambda_{4}\right)\right]c_{\beta}^{2}s_{\beta}^{2}+\lambda_{5}c_{\beta}^{4}+\lambda_{5}^{*}s_{\beta}^{4}-2\left(\lambda_{6}-\lambda_{7}\right)c_{\beta}^{3}s_{\beta}$ $\displaystyle+2\left(\lambda_{6}^{*}-\lambda_{7}^{*}\right)c_{\beta}s_{\beta}^{3},$ $\displaystyle\tilde{\lambda}_{6}$ $\displaystyle=$ $\displaystyle\left(-\lambda_{1}+\lambda_{3}+\lambda_{4}+\lambda_{5}^{*}\right)c_{\beta}^{3}s_{\beta}+\left(\lambda_{2}-\lambda_{3}-\lambda_{4}-\lambda_{5}\right)c_{\beta}s_{\beta}^{3}+\lambda_{6}^{*}c_{\beta}^{4}-\lambda_{7}s_{\beta}^{4}$ $\displaystyle-\left(\lambda_{6}^{*}-2\lambda_{7}^{*}+2\lambda_{6}-\lambda_{7}\right)c_{\beta}^{2}s_{\beta}^{2},$ $\displaystyle\tilde{\lambda}_{7}$ $\displaystyle=$ $\displaystyle\left(-\lambda_{1}+\lambda_{3}+\lambda_{4}+\lambda_{5}\right)c_{\beta}s_{\beta}^{3}-\left(-\lambda_{2}+\lambda_{3}+\lambda_{4}+\lambda_{5}^{*}\right)c_{\beta}^{3}s_{\beta}-\lambda_{6}s_{\beta}^{4}+\lambda_{7}^{*}c_{\beta}^{4}$ (85) $\displaystyle+\left(2\lambda_{6}^{*}-\lambda_{7}^{*}+\lambda_{6}-2\lambda_{7}\right)c_{\beta}^{2}s_{\beta}^{2}.$ Finally, the quartic couplings involving the complex Higgs triplet are give by $\displaystyle\tilde{\lambda}_{8}$ $\displaystyle=$ $\displaystyle\lambda_{8}c_{\beta}^{2}+\lambda_{9}s_{\beta}^{2}+\lambda_{10}^{*}c_{\beta}s_{\beta},$ $\displaystyle\tilde{\lambda}_{9}$ $\displaystyle=$ $\displaystyle\lambda_{8}s_{\beta}^{2}+\lambda_{9}c_{\beta}^{2}-\lambda_{10}^{*}c_{\beta}s_{\beta},$ $\displaystyle\tilde{\lambda}_{10}$ $\displaystyle=$ $\displaystyle-\left(\lambda_{8}-\lambda_{9}\right)c_{\beta}s_{\beta}-\lambda_{10}^{*}s_{\beta}^{2},$ $\displaystyle\tilde{\lambda}_{11}$ $\displaystyle=$ $\displaystyle\lambda_{11}c_{\beta}^{2}+\lambda_{12}s_{\beta}^{2}+{\rm Re\lambda_{13}}s_{2\beta},$ $\displaystyle\tilde{\lambda}_{12}$ $\displaystyle=$ $\displaystyle\lambda_{11}s_{\beta}^{2}+\lambda_{12}c_{\beta}^{2}-{\rm Re\lambda_{13}}s_{2\beta},$ $\displaystyle\tilde{\lambda}_{13}$ $\displaystyle=$ $\displaystyle\left(-\lambda_{11}+\lambda_{12}\right)c_{\beta}s_{\beta}+\lambda_{13}c_{\beta}^{2}-\lambda_{13}^{*}s_{\beta}^{2},$ $\displaystyle\tilde{\lambda}_{\Delta 1}$ $\displaystyle=$ $\displaystyle\lambda_{\Delta 1},~{}~{}\tilde{\lambda}_{\Delta 2}=\lambda_{\Delta 2}.$ (86) Note that under the $U(1)$ transformation $H_{1}\rightarrow e^{i\chi}H_{1}$ and $H_{2}\rightarrow e^{-i\chi}H_{2}$, the scalar potential remains unchanged if the complex parameters of the scalar potential in the Higgs basis are transformed by the corresponding phase rotation Davidson:2005cw : $\tilde{\lambda}_{5}\rightarrow e^{4i\chi}\tilde{\lambda}_{5}\text{ and }\tilde{m}_{12}^{2},\tilde{\lambda}_{6},\tilde{\lambda}_{7},\tilde{\lambda}_{10},\tilde{\lambda}_{13}\rightarrow e^{2i\chi}\left\\{\tilde{m}_{12}^{2},\tilde{\lambda}_{6},\tilde{\lambda}_{7},\tilde{\lambda}_{10},\tilde{\lambda}_{13}\right\\}.$ (87) Therefore, beginning with the Eqs. (C)-(C) in the phase $\\{\xi=0,\chi=0\\}$, we can obtain the relations between the two sets of parameters with arbitrary phases $\\{\xi,\chi\\}$ by applying the replacements (83) and (87) to the Eqs. (C)-(C). Finally, The inversion of Eqs. (C)-(C) can be obtained by making the replacements $\tilde{m}_{12}^{2}\to m_{12}^{2}$, $\tilde{\lambda}_{i}\to\lambda_{i}$, and $\beta\to-\beta$. ## Appendix D Trilinear couplings in the complex triplet extension of 2HDM The trilinear couplings between a neutral scalar and a pair of charged scalars in the triplet extension of the 2HDM are summarized as follows: $\displaystyle\lambda_{h_{1}^{0}H_{2}^{+}H_{2}^{-}}=\tilde{\lambda}_{3}$ , $\displaystyle~{}~{}\lambda_{h_{2}^{0}H_{2}^{+}H_{2}^{-}}={\rm Re}\tilde{\lambda}_{7},$ $\displaystyle\lambda_{h_{1}^{0}\delta^{+}\delta^{-}}=\tilde{\lambda}_{8}+\frac{1}{2}\tilde{\lambda}_{11}$ , $\displaystyle~{}~{}\lambda_{h_{2}^{0}\delta^{+}\delta^{-}}={\rm Re}\tilde{\lambda}_{10}+\frac{1}{2}{\rm Re}\tilde{\lambda}_{13},$ $\displaystyle\lambda_{h_{1}^{0}\delta^{++}\delta^{--}}=\tilde{\lambda}_{8}$ , $\displaystyle~{}~{}\lambda_{h_{2}^{0}\delta^{++}\delta^{--}}={\rm Re}\tilde{\lambda}_{10}.$ (88) Note that in the aligned and decoupling limits, we have $h_{1}^{0}\equiv h$ and $h_{2}^{0}\equiv H$. ## References * (1) G. Aad et al. [ATLAS], “Observation of a new particle in the search for the Standard Model Higgs boson with the ATLAS detector at the LHC,” Phys. Lett. B 716, 1-29 (2012) doi:10.1016/j.physletb.2012.08.020 [arXiv:1207.7214 [hep-ex]]. * (2) S. Chatrchyan et al. [CMS], “Observation of a New Boson at a Mass of 125 GeV with the CMS Experiment at the LHC,” Phys. Lett. B 716, 30-61 (2012) doi:10.1016/j.physletb.2012.08.021 [arXiv:1207.7235 [hep-ex]]. * (3) J. A. Formaggio, A. L. C. de Gouvêa and R. G. H. Robertson, “Direct Measurements of Neutrino Mass,” Phys. Rept. 914, 1-54 (2021) doi:10.1016/j.physrep.2021.02.002 [arXiv:2102.00594 [nucl-ex]]. * (4) P. A. R. Ade et al. [Planck], “Planck 2015 results. XIII. Cosmological parameters,” Astron. Astrophys. 594, A13 (2016) doi:10.1051/0004-6361/201525830 [arXiv:1502.01589 [astro-ph.CO]]. * (5) V. A. Kuzmin, V. A. Rubakov and M. E. Shaposhnikov, “On the Anomalous Electroweak Baryon Number Nonconservation in the Early Universe,” Phys. Lett. B 155, 36 (1985) doi:10.1016/0370-2693(85)91028-7 * (6) G. Degrassi, B. Di Micco, P. P. Giardino and E. Rossi, “Higgs boson self-coupling constraints from single Higgs, double Higgs and Electroweak measurements,” Phys. Lett. B 817, 136307 (2021) doi:10.1016/j.physletb.2021.136307 [arXiv:2102.07651 [hep-ph]]. * (7) M. McCullough, “An Indirect Model-Dependent Probe of the Higgs Self-Coupling,” Phys. Rev. D 90, no.1, 015001 (2014) [erratum: Phys. Rev. D 92, no.3, 039903 (2015)] doi:10.1103/PhysRevD.90.015001 [arXiv:1312.3322 [hep-ph]]. * (8) Q. H. Cao, Y. Liu and B. Yan, “Measuring trilinear Higgs coupling in WHH and ZHH productions at the high-luminosity LHC,” Phys. Rev. D 95, no.7, 073006 (2017) doi:10.1103/PhysRevD.95.073006 [arXiv:1511.03311 [hep-ph]]. * (9) W. Bizon, M. Gorbahn, U. Haisch and G. Zanderighi, “Constraints on the trilinear Higgs coupling from vector boson fusion and associated Higgs production at the LHC,” JHEP 07, 083 (2017) doi:10.1007/JHEP07(2017)083 [arXiv:1610.05771 [hep-ph]]. * (10) J. de Blas, M. Ciuchini, E. Franco, S. Mishima, M. Pierini, L. Reina and L. Silvestrini, “Electroweak precision observables and Higgs-boson signal strengths in the Standard Model and beyond: present and future,” JHEP 12, 135 (2016) doi:10.1007/JHEP12(2016)135 [arXiv:1608.01509 [hep-ph]]. * (11) M. Gell-Mann, M. L. Goldberger, N. M. Kroll and F. E. Low, “Amelioration of divergence difficulties in the theory of weak interactions,” Phys. Rev. 179, 1518-1527 (1969) doi:10.1103/PhysRev.179.1518 * (12) S. Weinberg, “Physical Processes in a Convergent Theory of the Weak and Electromagnetic Interactions,” Phys. Rev. Lett. 27, 1688-1691 (1971) doi:10.1103/PhysRevLett.27.1688 * (13) B. W. Lee, C. Quigg and H. B. Thacker, “The Strength of Weak Interactions at Very High-Energies and the Higgs Boson Mass,” Phys. Rev. Lett. 38, 883-885 (1977) doi:10.1103/PhysRevLett.38.883 * (14) B. W. Lee, C. Quigg and H. B. Thacker, “Weak Interactions at Very High-Energies: The Role of the Higgs Boson Mass,” Phys. Rev. D 16, 1519 (1977) doi:10.1103/PhysRevD.16.1519 * (15) N. Cabibbo, L. Maiani, G. Parisi and R. Petronzio, “Bounds on the Fermions and Higgs Boson Masses in Grand Unified Theories,” Nucl. Phys. B 158, 295-305 (1979) doi:10.1016/0550-3213(79)90167-6 * (16) M. Lindner, “Implications of Triviality for the Standard Model,” Z. Phys. C 31, 295 (1986) doi:10.1007/BF01479540 * (17) G. C. Branco, P. M. Ferreira, L. Lavoura, M. N. Rebelo, M. Sher and J. P. Silva, “Theory and phenomenology of two-Higgs-doublet models,” Phys. Rept. 516, 1-102 (2012) doi:10.1016/j.physrep.2012.02.002 [arXiv:1106.0034 [hep-ph]]. * (18) L. Wang, J. M. Yang and Y. Zhang, “Two-Higgs-doublet models in light of current experiments: a brief review,” Commun. Theor. Phys. 74, no.9, 097202 (2022) doi:10.1088/1572-9494/ac7fe9 [arXiv:2203.07244 [hep-ph]]. * (19) H. Huffel and G. Pocsik, “Unitarity Bounds on Higgs Boson Masses in the Weinberg-Salam Model With Two Higgs Doublets,” Z. Phys. C 8, 13 (1981) doi:10.1007/BF01429824 * (20) J. Maalampi, J. Sirkka and I. Vilja, “Tree level unitarity and triviality bounds for two Higgs models,” Phys. Lett. B 265, 371-376 (1991) doi:10.1016/0370-2693(91)90068-2 * (21) A. G. Akeroyd, A. Arhrib and E. M. Naimi, “Note on tree level unitarity in the general two Higgs doublet model,” Phys. Lett. B 490, 119-124 (2000) doi:10.1016/S0370-2693(00)00962-X [arXiv:hep-ph/0006035 [hep-ph]]. * (22) A. Arhrib, “Unitarity constraints on scalar parameters of the standard and two Higgs doublets model,” [arXiv:hep-ph/0012353 [hep-ph]]. * (23) I. F. Ginzburg and I. P. Ivanov, “Tree-level unitarity constraints in the most general 2HDM,” Phys. Rev. D 72, 115010 (2005) doi:10.1103/PhysRevD.72.115010 [arXiv:hep-ph/0508020 [hep-ph]]. * (24) S. Kanemura and K. Yagyu, “Unitarity bound in the most general two Higgs doublet model,” Phys. Lett. B 751, 289-296 (2015) doi:10.1016/j.physletb.2015.10.047 [arXiv:1509.06060 [hep-ph]]. * (25) H. Georgi and M. Machacek, “DOUBLY CHARGED HIGGS BOSONS,” Nucl. Phys. B 262, 463-477 (1985) doi:10.1016/0550-3213(85)90325-6 * (26) M. Aoki and S. Kanemura, “Unitarity bounds in the Higgs model including triplet fields with custodial symmetry,” Phys. Rev. D 77, no.9, 095009 (2008) [erratum: Phys. Rev. D 89, no.5, 059902 (2014)] doi:10.1103/PhysRevD.77.095009 [arXiv:0712.4053 [hep-ph]]. * (27) W. Konetschny and W. Kummer, “Nonconservation of Total Lepton Number with Scalar Bosons,” Phys. Lett. B 70, 433-435 (1977) doi:10.1016/0370-2693(77)90407-5 * (28) T. P. Cheng and L. F. Li, “Neutrino Masses, Mixings and Oscillations in SU(2) x U(1) Models of Electroweak Interactions,” Phys. Rev. D 22, 2860 (1980) doi:10.1103/PhysRevD.22.2860 * (29) M. Magg and C. Wetterich, “Neutrino Mass Problem and Gauge Hierarchy,” Phys. Lett. B 94, 61-64 (1980) doi:10.1016/0370-2693(80)90825-4 * (30) J. Schechter and J. W. F. Valle, “Neutrino Masses in SU(2) x U(1) Theories,” Phys. Rev. D 22, 2227 (1980) doi:10.1103/PhysRevD.22.2227 * (31) G. Lazarides, Q. Shafi and C. Wetterich, Nucl. Phys. B 181, 287-300 (1981) doi:10.1016/0550-3213(81)90354-0 * (32) R. N. Mohapatra and G. Senjanovic, Phys. Rev. D 23, 165 (1981) doi:10.1103/PhysRevD.23.165 * (33) A. Arhrib, R. Benbrik, M. Chabab, G. Moultaka, M. C. Peyranere, L. Rahili and J. Ramadan, “The Higgs Potential in the Type II Seesaw Model,” Phys. Rev. D 84, 095005 (2011) doi:10.1103/PhysRevD.84.095005 [arXiv:1105.1925 [hep-ph]]. * (34) N. Khan, “Exploring the hyperchargeless Higgs triplet model up to the Planck scale,” Eur. Phys. J. C 78, no.4, 341 (2018) doi:10.1140/epjc/s10052-018-5766-4 [arXiv:1610.03178 [hep-ph]]. * (35) B. A. Ouazghour, A. Arhrib, R. Benbrik, M. Chabab and L. Rahili, “Theory and phenomenology of a two-Higgs-doublet type-II seesaw model at the LHC run 2,” Phys. Rev. D 100, no.3, 035031 (2019) doi:10.1103/PhysRevD.100.035031 [arXiv:1812.07719 [hep-ph]]. * (36) J. M. Cornwall, D. N. Levin and G. Tiktopoulos, “Uniqueness of spontaneously broken gauge theories,” Phys. Rev. Lett. 30, 1268-1270 (1973) [erratum: Phys. Rev. Lett. 31, 572 (1973)] doi:10.1103/PhysRevLett.30.1268 * (37) J. M. Cornwall, D. N. Levin and G. Tiktopoulos, “Derivation of Gauge Invariance from High-Energy Unitarity Bounds on the s Matrix,” Phys. Rev. D 10, 1145 (1974) [erratum: Phys. Rev. D 11, 972 (1975)] doi:10.1103/PhysRevD.10.1145 * (38) Y. P. Yao and C. P. Yuan, “Modification of the Equivalence Theorem Due to Loop Corrections,” Phys. Rev. D 38 (1988), 2237 doi:10.1103/PhysRevD.38.2237 * (39) H. G. J. Veltman, “The Equivalence Theorem,” Phys. Rev. D 41 (1990), 2294 doi:10.1103/PhysRevD.41.2294 * (40) H. J. He, Y. P. Kuang and X. y. Li, “On the precise formulation of equivalence theorem,” Phys. Rev. Lett. 69 (1992), 2619-2622 doi:10.1103/PhysRevLett.69.2619 * (41) P. Fileviez Perez, H. H. Patel, M. J. Ramsey-Musolf and K. Wang, “Triplet Scalars and Dark Matter at the LHC,” Phys. Rev. D 79, 055024 (2009) doi:10.1103/PhysRevD.79.055024 [arXiv:0811.3957 [hep-ph]]. * (42) S. Yaser Ayazi and S. M. Firouzabadi, “Constraining Inert Triplet Dark Matter by the LHC and FermiLAT,” JCAP 11, 005 (2014) doi:10.1088/1475-7516/2014/11/005 [arXiv:1408.0654 [hep-ph]]. * (43) C. W. Chiang, G. Cottin, Y. Du, K. Fuyuto and M. J. Ramsey-Musolf, “Collider Probes of Real Triplet Scalar Dark Matter,” JHEP 01, 198 (2021) doi:10.1007/JHEP01(2021)198 [arXiv:2003.07867 [hep-ph]]. * (44) L. Wang and X. F. Han, “LHC diphoton and Z+photon Higgs signals in the Higgs triplet model with Y = 0,” JHEP 03, 010 (2014) doi:10.1007/JHEP03(2014)010 [arXiv:1303.4490 [hep-ph]]. * (45) P. Bandyopadhyay, K. Huitu and A. Sabanci Keceli, “Multi-Lepton Signatures of the Triplet Like Charged Higgs at the LHC,” JHEP 05, 026 (2015) doi:10.1007/JHEP05(2015)026 [arXiv:1412.7359 [hep-ph]]. * (46) L. Niemi, H. H. Patel, M. J. Ramsey-Musolf, T. V. I. Tenkanen and D. J. Weir, “Electroweak phase transition in the real triplet extension of the SM: Dimensional reduction,” Phys. Rev. D 100, no.3, 035002 (2019) doi:10.1103/PhysRevD.100.035002 [arXiv:1802.10500 [hep-ph]]. * (47) N. F. Bell, M. J. Dolan, L. S. Friedrich, M. J. Ramsey-Musolf and R. R. Volkas, “Two-Step Electroweak Symmetry-Breaking: Theory Meets Experiment,” JHEP 05, 050 (2020) doi:10.1007/JHEP05(2020)050 [arXiv:2001.05335 [hep-ph]]. * (48) C. H. Chen, C. W. Chiang and T. Nomura, “Muon g-2 in a two-Higgs-doublet model with a type-II seesaw mechanism,” Phys. Rev. D 104, no.5, 055011 (2021) doi:10.1103/PhysRevD.104.055011 [arXiv:2104.03275 [hep-ph]]. * (49) R. N. Mohapatra, A. Perez-Lorenzana and C. A. de Sousa Pires, “Type II seesaw and a gauge model for the bimaximal mixing explanation of neutrino puzzles,” Phys. Lett. B 474, 355-360 (2000) doi:10.1016/S0370-2693(00)00026-5 [arXiv:hep-ph/9911395 [hep-ph]]. * (50) P. H. Gu, H. Zhang and S. Zhou, “A Minimal Type II Seesaw Model,” Phys. Rev. D 74, 076002 (2006) doi:10.1103/PhysRevD.74.076002 [arXiv:hep-ph/0606302 [hep-ph]]. * (51) W. Chao, S. Luo, Z. z. Xing and S. Zhou, “A Compromise between Neutrino Masses and Collider Signatures in the Type-II Seesaw Model,” Phys. Rev. D 77, 016001 (2008) doi:10.1103/PhysRevD.77.016001 [arXiv:0709.1069 [hep-ph]]. * (52) P. Fileviez Perez, T. Han, G. y. Huang, T. Li and K. Wang, “Neutrino Masses and the CERN LHC: Testing Type II Seesaw,” Phys. Rev. D 78, 015018 (2008) doi:10.1103/PhysRevD.78.015018 [arXiv:0805.3536 [hep-ph]]. * (53) A. Melfo, M. Nemevsek, F. Nesti, G. Senjanovic and Y. Zhang, “Type II Seesaw at LHC: The Roadmap,” Phys. Rev. D 85, 055018 (2012) doi:10.1103/PhysRevD.85.055018 [arXiv:1108.4416 [hep-ph]]. * (54) C. S. Chen, C. Q. Geng, D. Huang and L. H. Tsai, “$h\rightarrow Z\gamma$ in Type-II seesaw neutrino model,” Phys. Lett. B 723, 156-160 (2013) doi:10.1016/j.physletb.2013.05.007 [arXiv:1302.0502 [hep-ph]]. * (55) Z. L. Han, R. Ding and Y. Liao, “LHC phenomenology of the type II seesaw mechanism: Observability of neutral scalars in the nondegenerate case,” Phys. Rev. D 92, no.3, 033014 (2015) doi:10.1103/PhysRevD.92.033014 [arXiv:1506.08996 [hep-ph]]. * (56) P. S. B. Dev, M. J. Ramsey-Musolf and Y. Zhang, “Doubly-Charged Scalars in the Type-II Seesaw Mechanism: Fundamental Symmetry Tests and High-Energy Searches,” Phys. Rev. D 98, no.5, 055013 (2018) doi:10.1103/PhysRevD.98.055013 [arXiv:1806.08499 [hep-ph]]. * (57) Y. Du, A. Dunbrack, M. J. Ramsey-Musolf and J. H. Yu, “Type-II Seesaw Scalar Triplet Model at a 100 TeV $pp$ Collider: Discovery and Higgs Portal Coupling Determination,” JHEP 01, 101 (2019) doi:10.1007/JHEP01(2019)101 [arXiv:1810.09450 [hep-ph]]. * (58) Y. Cheng, X. G. He, Z. L. Huang and M. W. Li, “Type-II seesaw triplet scalar effects on neutrino trident scattering,” Phys. Lett. B 831, 137218 (2022) doi:10.1016/j.physletb.2022.137218 [arXiv:2204.05031 [hep-ph]]. * (59) R. Ding, Z. L. Han, L. Feng and B. Zhu, “Confronting the DAMPE Excess with the Scotogenic Type-II Seesaw Model,” Chin. Phys. C 42, no.8, 083104 (2018) doi:10.1088/1674-1137/42/8/083104 [arXiv:1712.02021 [hep-ph]]. * (60) R. Primulando, J. Julio and P. Uttayarat, “Scalar phenomenology in type-II seesaw model,” JHEP 08, 024 (2019) doi:10.1007/JHEP08(2019)024 [arXiv:1903.02493 [hep-ph]]. * (61) R. Zhou, L. Bian and Y. Du, “Electroweak phase transition and gravitational waves in the type-II seesaw model,” JHEP 08, 205 (2022) doi:10.1007/JHEP08(2022)205 [arXiv:2203.01561 [hep-ph]]. * (62) S. Davidson and H. E. Haber, “Basis-independent methods for the two-Higgs-doublet model,” Phys. Rev. D 72, 035004 (2005) [erratum: Phys. Rev. D 72, 099902 (2005)] doi:10.1103/PhysRevD.72.099902 [arXiv:hep-ph/0504050 [hep-ph]]. * (63) H. E. Haber and D. O’Neil, “Basis-independent methods for the two-Higgs-doublet model. II. The Significance of tan$\beta$,” Phys. Rev. D 74, 015018 (2006) [erratum: Phys. Rev. D 74, no.5, 059905 (2006)] doi:10.1103/PhysRevD.74.015018 [arXiv:hep-ph/0602242 [hep-ph]]. * (64) J. Herrero-Garcia, M. Nebot, F. Rajec, M. White and A. G. Williams, “Higgs Quark Flavor Violation: Simplified Models and Status of General Two-Higgs-Doublet Model,” JHEP 02, 147 (2020) doi:10.1007/JHEP02(2020)147 [arXiv:1907.05900 [hep-ph]]. * (65) O. Eberhardt, A. P. Martínez and A. Pich, “Global fits in the Aligned Two-Higgs-Doublet model,” JHEP 05, 005 (2021) doi:10.1007/JHEP05(2021)005 [arXiv:2012.09200 [hep-ph]]. * (66) P. Athron, C. Balazs, T. E. Gonzalo, D. Jacob, F. Mahmoudi and C. Sierra, “Likelihood analysis of the flavour anomalies and g – 2 in the general two Higgs doublet model,” JHEP 01, 037 (2022) doi:10.1007/JHEP01(2022)037 [arXiv:2111.10464 [hep-ph]]. * (67) A. Wahab El Kaffas, P. Osland and O. M. Ogreid, “Constraining the Two-Higgs-Doublet-Model parameter space,” Phys. Rev. D 76, 095001 (2007) doi:10.1103/PhysRevD.76.095001 [arXiv:0706.2997 [hep-ph]]. * (68) D. Fontes, M. Mühlleitner, J. C. Romão, R. Santos, J. P. Silva and J. Wittbrodt, “The C2HDM revisited,” JHEP 02, 073 (2018) doi:10.1007/JHEP02(2018)073 [arXiv:1711.09419 [hep-ph]]. * (69) W. Khater and P. Osland, “CP violation in top quark production at the LHC and two Higgs doublet models,” Nucl. Phys. B 661, 209-234 (2003) doi:10.1016/S0550-3213(03)00300-6 [arXiv:hep-ph/0302004 [hep-ph]]. * (70) S. Inoue, M. J. Ramsey-Musolf and Y. Zhang, “CP-violating phenomenology of flavor conserving two Higgs doublet models,” Phys. Rev. D 89, no.11, 115023 (2014) doi:10.1103/PhysRevD.89.115023 [arXiv:1403.4257 [hep-ph]]. * (71) A. W. El Kaffas, P. Osland and O. M. Ogreid, “CP violation, stability and unitarity of the two Higgs doublet model,” Nonlin. Phenom. Complex Syst. 10, 347-357 (2007) [arXiv:hep-ph/0702097 [hep-ph]]. * (72) B. Abi et al. [Muon g-2], “Measurement of the Positive Muon Anomalous Magnetic Moment to 0.46 ppm,” Phys. Rev. Lett. 126, no.14, 141801 (2021) doi:10.1103/PhysRevLett.126.141801 [arXiv:2104.03281 [hep-ex]]. * (73) K. m. Cheung, C. H. Chou and O. C. W. Kong, “Muon anomalous magnetic moment, two Higgs doublet model, and supersymmetry,” Phys. Rev. D 64, 111301 (2001) doi:10.1103/PhysRevD.64.111301 [arXiv:hep-ph/0103183 [hep-ph]]. * (74) K. Cheung and O. C. W. Kong, “Can the two Higgs doublet model survive the constraint from the muon anomalous magnetic moment as suggested?,” Phys. Rev. D 68, 053003 (2003) doi:10.1103/PhysRevD.68.053003 [arXiv:hep-ph/0302111 [hep-ph]]. * (75) Y. F. Zhou and Y. L. Wu, “Lepton flavor changing scalar interactions and muon g-2,” Eur. Phys. J. C 27, 577-585 (2003) doi:10.1140/epjc/s2003-01137-1 [arXiv:hep-ph/0110302 [hep-ph]]. * (76) M. Aoki, S. Kanemura, K. Tsumura and K. Yagyu, “Models of Yukawa interaction in the two Higgs doublet model, and their collider phenomenology,” Phys. Rev. D 80, 015017 (2009) doi:10.1103/PhysRevD.80.015017 [arXiv:0902.4665 [hep-ph]]. * (77) X. F. Han, F. Wang, L. Wang, J. M. Yang and Y. Zhang, “Joint explanation of W-mass and muon g–2 in the 2HDM*,” Chin. Phys. C 46, no.10, 103105 (2022) doi:10.1088/1674-1137/ac7c63 [arXiv:2204.06505 [hep-ph]]. * (78) P. M. Ferreira, B. L. Gonçalves, F. R. Joaquim and M. Sher, “(g-2)$\mu$ in the 2HDM and slightly beyond: An updated view,” Phys. Rev. D 104, no.5, 053008 (2021) doi:10.1103/PhysRevD.104.053008 [arXiv:2104.03367 [hep-ph]]. * (79) J. Kim, “Compatibility of muon g-2, W mass anomaly in type-X 2HDM,” Phys. Lett. B 832, 137220 (2022) doi:10.1016/j.physletb.2022.137220 [arXiv:2205.01437 [hep-ph]]. * (80) A. Pich and P. Tuzon, “Yukawa Alignment in the Two-Higgs-Doublet Model,” Phys. Rev. D 80, 091702 (2009) doi:10.1103/PhysRevD.80.091702 [arXiv:0908.1554 [hep-ph]]. * (81) V. Ilisie, “New Barr-Zee contributions to $\mathbf{(g-2)_{\mu}}$ in two-Higgs-doublet models,” JHEP 04, 077 (2015) doi:10.1007/JHEP04(2015)077 [arXiv:1502.04199 [hep-ph]].
# Spin-polarized antichiral exciton-polariton edge states Ruiqi Bao S. Mandal<EMAIL_ADDRESS>Huawen Xu Xingran Xu R. Banerjee Timothy C. H. Liew<EMAIL_ADDRESS>Division of Physics and Applied Physics, School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore 637371, Singapore ###### Abstract We consider theoretically a system of exciton-polariton micropillars arranged in a honeycomb lattice. The naturally present TE-TM splitting and an alternating Zeeman splitting, where the different sublattices experience opposite Zeeman splitting, shifts the Dirac points in energy, giving rise to antichiral behavior. In a strip geometry having zigzag edges, two pairs of edge states exist and propagate in the same direction (including the states at the opposite edges). The edge modes localized at the opposite edges have opposite spins (circular polarizations), which leads to co-propagating $``\pm"$ spin channels. The antichiral edge states are protected by non-zero winding numbers and can propagate around a 60 degree bend without being reflected. We further compare the transport properties of these edge states with chiral edge modes and propose a scheme to realize them experimentally. ## I Introduction The quantum Hall effect Klitzing _et al._ (1980), where free electron gas subjected to a perpendicular magnetic field shows quantized Hall conductance, has laid the platform for topological phases. A few years later, the Haldane model was proposed Haldane (1988), where the quantum Hall effect was shown without any net magnetic field. The Haldane model has gone onto become the backbone of many noble topological systems such as anomalous Chern insulators Jotzu _et al._ (2014); Zhao _et al._ (2020), quantum spin-Hall insulators Kane and Mele (2005), etc. Recently, a new type of topological phase, which shows antichiral edge states, has been introduced based on a modified-Haldane model Colomés and Franz (2018). In contrast to previous topological phases, systems with antichiral edge states do not host a bulk band gap. Instead, in a strip geometry both the edge states propagate in the same direction and counter-propagating states within the same energy window lie in the bulk. Although, the antichiral edge states reside along with the bulk modes, they were shown to be reasonably robust against disorder Colomés and Franz (2018). Because of their peculiar properties, the antichiral edge states have been under intense investigation in various system including in exciton-polaritonics Mandal _et al._ (2019a), photonics Chen _et al._ (2020); Zhou _et al._ (2020); Chen and Li (2022), acoustics Yu _et al._ (2021); Wu _et al._ (2020), electric circuits Yang _et al._ (2021) and others Bhowmick and Sengupta (2020); Mannaï and Haddad (2020); Breitkreiz (2020); Mizoguchi and Koma (2021); Cheng _et al._ (2021). Figure 1: Schematic of a honeycomb lattice consisting exciton-polariton micropillars. “A” and “B” correspond to the different sublattices of the system, which experience opposite Zeeman splittings. In presence of the sublattice dependent Zeeman splitting and the TE-TM splitting, spin-polarized antichiral edge states appear, where opposite edge attain opposite spins and propagate in the same direction. Although, the antichiral edge states have been explored over a variety of systems, the spin degree of freedom has not been associated with them till now. In this work, we propose theoretically spin-polarized antichiral edge states in an exciton-polariton system. Exciton-polaritons are quasi particles that arise due to the strong coupling between quantum well excitons and microcavity photons Deng _et al._ (2010); Carusotto and Ciuti (2013); Byrnes _et al._ (2014). Due to their excitonic component, polaritons exhibit strong nonlinearity, which has lead to a variety of optical devices Sanvitto and Kéna-Cohen (2016); Feng _et al._ (2021), optical computation Kyriienko _et al._ (2019); Xu _et al._ (2020); Banerjee and Liew (2020); Harrison _et al._ (2022), and others Xu _et al._ (2021a); Zvyagintseva _et al._ (2022). The finite lifetime of the photons gives rise to interesting non-Hermitian physics. Circularly polarized photons provide a spin degree of freedom to the polaritons, which show many rich effects. For example, the naturally present energy splitting between the transverse electric (TE) and transverse magnetic (TM) modes (also known as the TE-TM splitting) acts as an effective magnetic field Kavokin _et al._ (2005), which leads to the realization of the optical spin-Hall effect both in the linear Leyder _et al._ (2007) and nonlinear Kammann _et al._ (2012) regimes. The excitonic component along with the spin degree of freedom leads to the Zeeman splitting in presence of perpendicular magnetic field Larionov _et al._ (2010). These fascinating properties of polaritons make them ideal for studying topological phases in polariton lattices. Polariton Chern insulators Karzig _et al._ (2015); Bardyn _et al._ (2015); Nalitov _et al._ (2015); Banerjee _et al._ (2018); Li _et al._ (2018); Sun _et al._ (2019) and related phases Solnyshkov _et al._ (2018); Mandal _et al._ (2019b) have been explored theoretically and later realized experimentally Klembt _et al._ (2018). Strong nonlinear polariton-polariton interaction gives rise to different kinds of nonlinear topological behavior including topological solitons Kartashov and Skryabin (2016); Gulevich _et al._ (2017); Pernet _et al._ (2022); Solnyshkov _et al._ (2017), interaction induced chiral edge states Bleu _et al._ (2017), bistable topological insulators Kartashov and Skryabin (2017), nonlinear higher-order topological insulators Banerjee _et al._ (2020), etc. In some cases, nonlinearity can alone induce topological phase transition Bardyn _et al._ (2016); Sigurdsson _et al._ (2017); Banerjee _et al._ (2021). The inherent non-Hermiticity of the polaritons has also lead to a variety of non-Hermitian topological behavior such as exceptional points Gao _et al._ (2015a, 2018); Khurgin (2020); Comaron _et al._ (2020); Su _et al._ (2021); Liao _et al._ (2021); Krol _et al._ (2022), non-reciprocal transport Mandal _et al._ (2020); Xu _et al._ (2021b), skin effect Mandal _et al._ (2022), double-sided skin effect Xu _et al._ (2021c), non-Hermitian topological edge-modes Pickup _et al._ (2020); Comaron _et al._ (2020), etc. However, none of the above works can show topologically protected spin transport and the search for a system exhibiting robust spin transport is still an open question. Here, we consider exciton-polaritons in a honeycomb lattice in the presence of the TE-TM splitting and subjected to an alternating Zeeman splitting, where all the “A” (“B”) sublattices are considered to have positive (negative) Zeeman splitting. The same system with a uniform Zeeman splitting is known to give rise to a Chern insulating phase, where in a strip geometry, the bulk band gap hosts two pairs of counter-propagating edge states Nalitov _et al._ (2015); Bardyn _et al._ (2015); Klembt _et al._ (2018). Unlike the Chern insulator case, our present system does not show a bulk bandgap. Instead, the Dirac points split in energy. The non-trivial topology of the system connects the Dirac points with two pairs of zigzag edge modes with each pair residing on the opposite edges of a finite strip. Consequently, the edge modes including the opposite edges obtain the same sign of the group velocity. Unlike the Chern insulator case, here the system becomes antichiral, where all the edge modes propagate in the same direction. We further find that the edge states residing at the opposite edges of the strips have opposite circular polarization, which enables robust spin transport along the edges of the system. However, due to the fact that the edge modes reside with the counter- propagating bulk modes, the system is less robust compared to their Chern insulator counterpart. Our simulation shows that the system shows fair robustness against lattices defects and the efficiency of pulse propagation remains lower than but close to the Chern insulators. Figure 2: (a)Honeycomb lattice consisting of two types of sublattices, “A” and “B”. $a_{1}$ and $a_{2}$ are the primitive lattice vectors. (b) Lines connecting the high symmetry points of the 2D Brillouin zone, along which the band structures are calculated. (c) Band structure for a trivial Honeycomb lattice ($\Delta_{T}=\Delta_{A}=\Delta_{B}=0$). (d) Effect of Zeeman splitting on the honeycomb lattice band structure ($\Delta_{T}=0,~{}\Delta_{A}=\Delta_{B}=0.3J$). (e) Effect of the TE-TM splitting on the band structure ($\Delta_{T}=0.3J,~{}\Delta_{A}=\Delta_{B}=0$). (f) Chern insulating band structure with a non-trivial bulk bandgap ($\Delta_{T}=\Delta_{A}=\Delta_{B}=0.3J$). (g) Antichiral band structure with the two valleys shifted in energy ($\Delta_{T}=\Delta_{A}=-\Delta_{B}=0.3J$). (h) Winding numbers as a function of $k_{1}$ for the antichiral bands, which become nontrivial for $k_{1}>|k_{D}|$, predicting the existence of two pairs of antichiral edge states in that momentum window. ## II The model We start by considering two coupled micropillars “A” and “B” in presence of the Zeeman splitting and the TE-TM splitting. The corresponding Hamiltonian using the bispinor basis $\Psi=[\Psi_{A}^{+},\Psi_{A}^{-},\Psi_{B}^{+},\Psi_{B}^{-}]^{T}$ can be expressed in the tight-binding limit as $\displaystyle\mathcal{H}_{AB}=\begin{bmatrix}\Delta_{A}&0&J&\Delta_{T}e^{-2i\theta}\\\ 0&-\Delta_{A}&\Delta_{T}e^{2i\theta}&J\\\ J&\Delta_{T}e^{-2i\theta}&\Delta_{B}&0\\\ \Delta_{T}e^{2i\theta}&J&0&-\Delta_{B}\end{bmatrix}.$ (1) Here $J$ is the coupling strength between the pillars, $\Delta_{T}$ is the TE- TM splitting, $\theta$ is the orientation angle connecting the two pillars with respect to a reference axis, and $\Delta_{A(B)}$ is the Zeeman splitting corresponding to the pillar “A”(“B”). We stress that while staggered potentials were considered in previous work predicting a valley-Hall effect Bleu _et al._ (2018), here it is the Zeeman splitting that is staggered. We discuss explicit ways that it can be realized in section VI. The difference between staggered potentials and staggered Zeeman splitting is that in the case that a uniform magnetic field is applied together with a staggered potential, we would generally obtain the chiral edge states of a Chern insulator rather than antichiral ones. However, by properly tunning the parameters it is possible to completely gap out one of the valleys, leaving the other one gapped. Although, the system no longer remains a Chern insulator, it can act as a robust spin filter, where only one of the two spins are allowed to propagate along one of the edges. Such a regime is discussed in details in Ref. Mandal _et al._ (2019b). Next, we arrange such pillar pairs in a honeycomb lattice structure and write down its corresponding Hamiltonian in the reciprocal space by taking periodic boundaries in both the directions. This can be achieved by writing $\Psi^{\pm}_{A,B,n}=\Psi^{\pm}_{A,B}e^{i\mathbf{k}.{\mathbf{x}}_{n}}$ where ${\mathbf{x}}_{n}$ is the position of each site and reducing the basis back to that of $[\Psi_{A}^{+},\Psi_{A}^{-},\Psi_{B}^{+},\Psi_{B}^{-}]^{T}$. $\displaystyle\mathcal{H}_{\mathbf{k}}=\begin{bmatrix}\Delta_{A}&0&-g_{\mathbf{k}}J&-g^{+}_{\mathbf{k}}\Delta_{T}\\\ 0&-\Delta_{A}&-g^{-}_{\mathbf{k}}\Delta_{T}&-g_{\mathbf{k}}J\\\ -g^{*}_{\mathbf{k}}J&-{g^{-}_{\mathbf{k}}}^{*}\Delta_{T}&\Delta_{B}&0\\\ -{g^{+}_{\mathbf{k}}}^{*}\Delta_{T}&-g^{*}_{\mathbf{k}}J&0&-\Delta_{B}\end{bmatrix},$ (2) where $g_{\mathbf{k}}=\sum_{n=1}^{3}\exp\left({-i\mathbf{k}.{\mathbf{r}}_{n}}\right)$ and $g^{\pm}_{\mathbf{k}}=\sum_{n=1}^{3}\exp\left({-i\left[\mathbf{k}.{\mathbf{r}}_{n}\mp 2\theta_{n}\right]}\right).$ Here $\mathbf{r}_{n}$ represent the vectors connecting the three nearest “B” sites from a single “A” site (Fig. 2(a)) and $\theta_{n}=2\pi(n-1)/3$ are the angles of those vectors with respect to one of them (say $\mathbf{r}_{1}$). To investigate the effect of $\Delta_{T}$, $\Delta_{A}$, and $\Delta_{B}$, we plot the band structure of the system along a line connecting the high symmetry points in the 2D Brillouin zone. Such a line is shown in Fig. 2(b). Before proceeding, we define the degree of circular polarization corresponding to a state as $\displaystyle S_{z}=\frac{\left|\Psi^{+}_{A}\right|^{2}+\left|\Psi^{+}_{B}\right|^{2}-\left|\Psi^{-}_{A}\right|^{2}-\left|\Psi^{-}_{B}\right|^{2}}{\left|\Psi^{+}_{A}\right|^{2}+\left|\Psi^{+}_{B}\right|^{2}+\left|\Psi^{-}_{A}\right|^{2}+\left|\Psi^{-}_{B}\right|^{2}}.$ (3) The Hamiltonian of the system for $\Delta_{T}=\Delta_{A}=\Delta_{B}=0$ corresponds to trivial Graphene. Consequently, we obtain the Graphene band structure by diagonalizing Eq. (2), where the lower and upper band touch each other at the $K$ and $K^{\prime}$ valleys forming the Dirac cones (see Fig. 2(c)). In this case, the bands corresponding to the “$\pm$” spins are degenerate or equivalently they are linearly polarized with $S_{z}=0$. This degeneracy can be lifted by making $\Delta_{A}=\Delta_{B}=\Delta$, which breaks the time-reversal symmetry (TR) and shifts the bands corresponding to the “$\pm$” spins in energy by $2\Delta$ (see Fig. 2(d)). Next, we make $\Delta_{T}\neq 0$ keeping $\Delta_{A}=\Delta_{B}=0$. In this case, the TR symmetry is restored and the bands become linearly polarized. One significant difference from the previous two cases is that the band structure is no longer linear near the $K$ and $K^{\prime}$ valleys (see Fig. 2(e)). Making $\Delta_{T}\neq 0$ and $\Delta_{A}=\Delta_{B}\neq 0$ gives rise to the non- trivial topological band structure with the system transiting to a Chern insulating phase having Chern number 2 Nalitov _et al._ (2015); Bardyn _et al._ (2015); Klembt _et al._ (2018). A nontrivial bulk band gap opens up due to the band inversion, which can be identified by the opposite values of $S_{z}$ of the second and third bands near the valleys (see Fig. 2(f)). Finally, we introduce the antichiral band structure in Fig. 2(g) by choosing $\Delta_{T}\neq 0$ and $\Delta_{A}=-\Delta_{B}\neq 0$. Unlike the Chern insulator band structure, the bulk bands are not gapped here. Instead the states near the $K$ and $K^{\prime}$ prime valleys shift in energy. The introduction of $\Delta_{A(B)}$ breaks the TR symmetry. However, unlike Figs. 2(d,f) the bulk bands do not have any preferred spins and stay linearly polarized. The antichiral system does not have a bulk band gap. Consequently, it is not possible to assign a Chern number to characterize the topology of the system. It is possible to calculate a relevant winding number. However, unlike all the previous antichiral systems, instead of one, two bands are involved here. Consequently, one needs to define the non-abelian Berry connection in order to calculate the winding number (see appendix). This is defined as Qi _et al._ (2008) $\displaystyle\mathcal{F}^{m,n}_{k_{1}}=i\langle u^{m}_{k_{1}}(k_{2})\left|\frac{\partial}{\partial k_{2}}\right|u^{n}_{k_{1}}(k_{2})\rangle,$ (4) where $m,n=1,2$ corresponds to the lowest two bands as indicated in Fig. 2(g). $u^{n}_{k_{1}}(k_{2})$ is the Bloch wave function. $k_{1}$ and $k_{2}$ are the wave vectors along the $a_{1}$ and $a_{2}$ directions, respectively (see Fig. 2(a)). The winding number is given as $\displaystyle\nu_{k_{1}}=\nu^{+}_{k_{1}}+\nu^{-}_{k_{1}}=\frac{1}{\pi}\oint_{\text{BZ}}dk_{2}\text{Tr}[\mathcal{F}_{k_{1}}],$ (5) where $\nu_{k_{1}}$ is the total winding number of the system, $\nu^{\pm}_{k_{1}}$ correspond to the contribution from the two bands, and the integration is taken over the Brillouin zone along the ${k_{2}}$ direction for each value of $k_{1}$. In Fig. 2(h) the winding numbers are plotted as a function of $k_{1}$, which shows $\displaystyle\nu^{+}_{k_{1}}=-\nu^{-}_{k_{1}}\approx\left\\{\begin{matrix}0,~{}\text{for}\left|k_{1}\right|<k_{D}\\\ 1,~{}\text{for}\left|k_{1}\right|>k_{D}\end{matrix}\right.$ (6) where $k_{D}$ is the position of the Dirac point. The total winding number of the system is always 0. However, $\nu^{\pm}_{k_{1}}$ deviates from 0 and becomes close to $\pm 1$ as shown in the shaded region in Fig. 2(h). Non-zero winding corresponds to the non-trivial topology and predicts the existence of antichiral edge states Colomés and Franz (2018). Two non-trivial branches in the winding number predicts two pairs of antichiral edge modes. This is different from previously studied systems, where only a pair of antichiral edge modes are obtained. Figure 3: Topological band structures for antichiral (a-b) and Chern insulator (c-d) cases. The states are color coded with respect to the contribution from the edges in (a,c) and with respect to the contribution from the spins in (b,d). Antichiral edge states are co-propagating having almost pure spins in a certain energy window. The chiral edge states in the Chern insulator case are counter-propagating and have mixed spins. Here $a$ is the periodicity along the $x$ direction, $\langle y\rangle$ represents of the mean position of the states along the $y$ axis with $-L$ and $+L$ being the boundaries. $S_{z}$ is defined in Eq. (3). All the parameters are kept the same as those in Fig. 2. ## III Topological band structure In this section, we provide topological band structures by considering the system periodic along the $x$ direction and truncated along the $y$ direction with the zigzag boundaries. In Figs. 3(a-b) the numerically calculated band structure for the antichiral system $(\Delta_{T}=\Delta_{A}=-\Delta_{B})$ is shown. As predicted from the non-zero winding numbers, indeed two pairs of antichiral edge modes connecting the Dirac points appear and co-exist with the bulk modes. Generally, the edge modes localized at the opposite edges are degenerate. However, for better visualization, we have lifted the degeneracy of the edge modes by adding a inversion symmetry breaking term $m\sigma_{z}\otimes\sigma_{o}$ to the Hamiltonian in Eq. (2). Here $\sigma_{z}$ is the $z$ component of the Pauli matrix, $\sigma_{o}$ is a $2\times 2$ identity matrix, and $\otimes$ represents the tensor product. We have chosen a small $m=10^{-2}$ for Fig. 3(a-b), which does not alter any property of the system. This is necessary for visualizing the properties of the edge modes from the band structure, otherwise the opposite edge modes in each pair would overlap with each other making it difficult to distinguish them. To study the properties of the edge modes, we color code the band structure as a contribution of localization in Fig. 3(a) and as a contribution of $S_{z}$ in Fig. 3(b). All the edge modes have the same group velocity, which makes them propagate in the same direction. Such one directional propagation at the edges is balanced by the counter propagating bulk modes. Although, the bulk remains linearly polarized, the antichiral edge modes at the two opposite edges have opposite values of $S_{z}$ and in a particular energy window $S_{z}\rightarrow\pm 1$. As we show in the next section, this leads to different spin channels in the same sample, where opposite spins propagate along the different edges of a finite sample in the same direction. In Figs. 3(c-d), we provide the Chern insulating band structure $(\Delta_{T}=\Delta_{A}=\Delta_{B})$ for comparison. Unlike the antichiral system, here the bulk bands are gapped and two pairs of edge modes appear. Since the edge states at the different edges have opposite group velocities, they propagate in the opposite directions. Here the bulk bands are not linearly polarized and have some preferred value of $S_{z}$. Near the Dirac points, $S_{z}\rightarrow\pm 1$. However, the edge modes deviate from the pure spins and acquire some mixed spins. ## IV Spin propagation Figure 4: (a-c) Polariton propagation under the effect of two pulses. The arrows in (b) represents the positions of two defects deliberately introduced by removing a site at each edge. Polaritons can propagate around the defect without significant backscattering. (d) Polariton propagation under continuous coherent pumps. Polaritons with “$-$” spin propagate along the upper edge, while the polaritons with “$+$” spin propagate along the lower edge. The yellow stars correspond to the positions of linearly polarized coherent pulses (a-c) and pumps (d). Parameters: $F_{0}=J$, $\Gamma=0.002J$, $\sigma=3$, $\tau_{0}/\hbar=50/J$, $\hbar\omega_{p}=-0.207J$, $t_{u}=\hbar/J$ and $ak_{p}=3.42$. All other parameters are kept the same as those in Figs. 3(a-b). Figure 5: Schematic diagram of antichiral system arranged in a triangle (a) and a parallelogram shape (d). The triangle shaped system has only one type of edge. (b-c) Only spin “+” polaritons can propagate around the bending without being backscattered. (e-f) The parallelogram consists of two types of zigzag edges and “$\pm$” spins propagate along those edges, separately. (g) At the top left corner, where both edges meet, polaritons scatter into the bulk. (h) For the Chern insulator system, polaritons can propagate along the edges and do not scatter into the bulk. Each arm consists of 15 sites in (a-c) and 25 sites in (d-h). The yellow stars represent linearly polarized coherent pumps. Parameters: (a-g) All other parameters are kept the same as those in Fig. 4. (h) $\Delta_{T}=\Delta_{A}=\Delta_{B}=0.3J$, $\hbar\omega_{p}=0.07J$, and $ak_{x}=2.89$. In this section, we investigate the polariton propagation in the antichiral system and find out their robustness by removing one site from each edge. We consider a stripe lattice formed by 12 unit cells along the $y$ direction and 20 unit cells along the $x$ direction. The upper (lower) edge of the system is “A” (“B”) type zigzag edge (see Fig. 1). The evolution of the polaritons under the coherent excitation is governed by the following time dependent Schrödinger equation $i\hbar\frac{\partial\Psi_{\pm}}{\partial t}=H\Psi_{\pm}-i\Gamma\Psi_{\pm}+F_{\pm}(x,y,t)\exp{\left[i\left(k_{p}x-\omega_{p}t\right)\right]},$ (7) where $\displaystyle F_{\pm}=F_{0}\left\\{\begin{matrix}\exp\left[{-\frac{[(x-x_{0})^{2}+(y-y_{0})^{2}]}{2\sigma^{2}}-\frac{t^{2}}{2\tau_{0}^{2}}}\right],~{}\text{for pulse,}\\\ \exp\left[{-\frac{[(x-x_{0})^{2}+(y-y_{0})^{2}]}{2\sigma^{2}}}\right],~{}\text{for pump}\end{matrix}\right..$ (8) Here $H$ is the Hamiltonian representing the system which can be written by repeating the Hamiltonian $H_{AB}$, $F_{\pm}$ represents the circularly polarized components of Gaussian shaped coherent pulses or pumps having amplitude $F_{0}$, energy $\hbar\omega_{p}$, and momentum $k_{p}$; and $\Gamma$ is the decay rate of polaritons. The expression in Eq. (8) represents the profiles of $F_{\pm}$ in space and in time. $(x_{0},y_{0})$ represent the position of $F_{\pm}$ having widths $\sigma$ in space and $\tau_{0}$ in time. In Fig. 4 the polariton propagation under two linearly polarized ($F_{+}=F_{-}$) pulses positioned at the two edges (represented by the yellow stars) are shown. As it can be seen from the band structure in Fig. 3(a-b) the edge modes have negative group velocity. Consequently, polaritons propagate from right to left along both the edges. To check the robustness, we have removed one site at each edge as indicated by the arrows in 4(b). As it can be seen in Figs. 4(a-c) the polaritons can propagate around the defect without significant backscattering. However, the propagation around the defect is not perfect and in the next section, we have shown that the backscattering is around 10%. Because of the presence of the decay term $\Gamma$, the intensity created by the pulse decreases exponentially. Consequently, for visulization purpose the intensity is renormalized at each time step in Figs. 4(a-c). Next, to show that the edges are circularly polarized, we use continuous pump instead of pulses. It should be noted that we use linearly polarized pumps to inject both $\pm$ polaritons. However, the system only allows $``+"$ polaritons to propagate along the lower edge and $``-"$ polaritons to propagate along the upper edge, creating two spin channels. The dynamics of the polaritons become more interesting if instead of a strip geometry, triangular and parallelogram shaped geometries are considered. In Fig. 5(a), (d) schematics of triangular and parallelogram shaped geometries are shown, respectively. For triangular geometry, one of the edges of the system can be considered as periodic while the other edge does not appear in the structure. As a result, under a linearly polarized coherent pump, spin “+” polaritons can propagate along the edges without being backscattered into the bulk (see Figs. 5(b-c)). This scenario changes for parallelogram shaped geometry as it consists of both type of edges. Consequently, $``\pm"$ polaritons propagate along different edges (see Figs. 5(e-f)). Due to the presence of $\Gamma$ the intensity of the polaritons decreases as they propagate away from the excitation spot. It should be noted that in both the structures, polaritons can go around the 60 degree bends perfectly. At 120 degree bends the two types of edge states meet (see the top left corner in Fig. 5(d)). Since both the edge states propagate in the same direction, polaritons can not continue propagating along the edges. Instead they scatter into the bulk. Such a situation is shown in Fig. 5(g), where the pump is positioned at the top left corner, where the opposite edges meet. For comparison, in Fig. 5(h) we have considered the case for the Chern insulator, which understandably does not scatter into the bulk and continues propagating along the edges. ## V Efficiency comparison To estimate the backscattering while going around the defect, we consider the similar geometry as shown in Figs. 4(a-c), where the edge contains a defect. Next, we position a linearly polarized coherent pulse at the right end of the strip and define the efficiency of the system as $\eta=\frac{I_{l}-I_{r}}{I_{l}+I_{r}}$ (9) where $I_{l}$ is the total intensity of the sites situated at the left of the defect (from both edge and bulk) and $I_{r}$ is the total intensity of sites situated at the right of the defect (from both edge and bulk). The intensities are considered at longer times when the polaritons pass the defect. In Fig. 6 the efficiency of the antichiral system is plotted in blue as a function of different system size $N$ as defined in the inset. For larger $N$, the efficiency stays near the 90%, but never reaches 100%, which is an indication that there is always around 10% backscattering when the antichiral edge states go around a defect. This is understandable, as the antichiral edge modes reside together with the bulk modes in the same energy window, a defect can couple the edge and bulk modes. However, Since the edge modes are spatially separated from the bulk modes, the system still shows fair robustness and the efficiency remains around 90 %. The efficiency decreases rapidly as $N$ decreases. For comparison with the Chern insulator edge states, we also perform the same steps and obtain the intensity as shown in red. As expected, the Chiral edge modes of the Chern insulators show near perfect transmission around a defect and the efficiency remains around 100 %. Similar to antichiral system the efficiency decreases rapidly for lower $N$. The low $\eta$ for lower $N$ is understandable from the fact that, being topological in nature, the robustness of the edge modes in both the systems is associated with the bulk properties. However, if sufficient bulk is not provided to the system, the system can not show the expected robustness. We should notice that the nature of backscattering is different for the antichiral and Chern insulator case. For the antichiral system, the drop in efficiency arises due to the coupling of co-propagating antichiral edge states with the counter propagating bulk states. However, for the Chern insulator case, when $N$ is small, a defect couples the two counter-propagating edge states, reducing the efficiency. Figure 6: Transmission efficiency comparison between anti-chiral edge states (in blue) and the Chern insulator edge states (in red) with different number of layers $N$, as defined in the inset. The Chern insulator can reach almost perfect transmission without backscattering when number of layers is larger than 10. Understandably, the efficiency of antichiral edge states remains lower than that of the Chern insulator, but still reaches around $90\%$. Both types of system has a site missing at the edge. ## VI Experimental proposal Figure 7: (a) Spatial profile of $P_{+}$ with peaks occurring in only one type of sublattices. The green circles represent the micropillars. (b-e) Band structure of the system. All four of them corresponds to the same system. For presentation purpose, upper edge states are omitted in (b-c) and lower edge states are omitted in (d-e). The bulk is linearly polarized, while the opposite edge states have opposite spins. In this section, we discuss and propose a scheme to realize the spin-polarized antichiral edge states in the experiments. The key elements in our scheme are 1) honeycomb lattice potential for the polaritons, 2) TE-TM splitting, and 3) sublattice dependent Zeeman splitting. Arranging exciton-polariton micropillars in a variety of periodic lattices is a routine task in experiments. The TE-TM splitting is known to occur naturally in microcavities due to the polarization dependent reflection from the cavity mirrors Panzarini _et al._ (1999). A Chern insulator was realized based on these ingredients, when a honeycomb lattice was subjected to a strong perpendicular magnetic field, which induced a uniform Zeeman splitting Klembt _et al._ (2018). In order to realize the antichiral system, instead of a uniform Zeeman splitting, the sign of the Zeeman splitting needs to be opposite for different sublattices. For this reason, the set up with real magnetic field used for the realization of Chern insulators seems challenging and requires an alternative way. In Ref. Sun _et al._ (2019), an additional layer of ferromagnetic material was proposed to induced the Zeeman splitting. Instead of one, two ferromagnetic material layers with opposite predefined magnetic momentums may be used and etched to make the Zeeman splitting opposite in different sublattices. However, using the inherent nonlinearity of the polaritons seems to be the most straight-forward way. In a honeycomb lattice under the nonpolarized incoherent pump, it is possible to form polariton condensate in an antiferromagnetic configuration, where different types of sublattices attain opposite circular polarization Sigurdsson _et al._ (2019). In such a scenario, due to the nonlinear interactions, polaritons with opposite spins will experience blueshifts in different sublattices, giving rise to the necessary sublattice dependent Zeeman splitting in our scheme. Another approach could be using circularly polarized incoherent pumps. It was recently verified experimentally that in micropillar structures, a circularly polarized incoherent pump can give rise to optically induced Zeeman splitting up to 0.2 meV Real _et al._ (2021). In what follows, we implement this recent experimental finding in our scheme and realize the spin-polarized antichiral edge states. Polaritons under the effect of external incoherent pump can be described using the following Gross-Pitaevskii equation (GPE) Carusotto and Ciuti (2013) $\displaystyle i\hbar\frac{\partial\psi_{\pm}}{\partial t}=$ $\displaystyle\left[-\frac{\hbar^{2}\nabla^{2}}{2m_{o}}+V(x,y)+gP_{\pm}(x,y)\right]\psi_{\pm}$ $\displaystyle+\left(\alpha_{1}|\psi_{\pm}|^{2}+\alpha_{2}|\psi_{\mp}|^{2}-i\alpha_{NL}|\psi_{\pm}|^{2}\right)\psi_{\pm}$ $\displaystyle+i\left[P_{\pm}(x,y)-\Gamma\right]\psi_{\pm}+\Delta_{T}\left(i\frac{\partial}{\partial x}\pm\frac{\partial}{\partial y}\right)^{2}\psi_{\mp}.$ (10) Here $\psi_{\pm}$ is the polariton wave function corresponding to the “$\pm$” spins, $m_{o}$ is the polartion effective mass, and $V(x,y)$ is the underlying honeycomb lattice potential. $gP_{\pm}$ is the spin dependent potential experienced by the polaritons due to interaction with the excitonic reservoir created by the circularly polarized incoherent pumps $P_{\pm}$ with $g$ being a dimensionless parameter. $\Delta_{T}$ is the TE-TM splitting. $\alpha_{1(2)}$ is the polariton-polariton interaction coefficient between same (opposite) spins. $\alpha_{NL}$ is the nonlinear loss and $\Gamma$ is the linear decay. The incoherent pump can be patterned into a lattice shape. One way is to use a spatial light-modulator, which has been used to make lattice potentials in a variety of works Ohadi _et al._ (2017, 2018); Töpfer _et al._ (2021); Opala _et al._ (2022). An alternative technique is to interference plane waves with different in-plane wavevector components which has been demonstrated experimentally in Luo _et al._ (2020). Here in this work, we use superposition of four waves to prepare the incoherent pumps, such that the resultant wave has maxima at one type of sublattice, while minima at the other type of sublattices. Here we emphasize that the four waves we are using here come from the same laser by using additional external optics devices. In this case, $P_{+}$ can be expressed as $\displaystyle P_{+}(x,y)=P_{0}f(x,y),$ (11) where $\displaystyle f(x,y)=$ $\displaystyle\cos\left(K_{1}\cdot R\right)+\cos\left(K_{2}\cdot R\right)+\cos\left(K_{3}\cdot R\right)$ $\displaystyle+\cos\left(K_{4}\cdot R\right)$ (12) with $K_{1}=4\pi(0,1/\sqrt{3})/a$, $K_{2,3}=2\pi(1,\pm 1/\sqrt{3})/a$, $K_{4}=(0,0)$. In the above $a$ is the periodicity of the lattice along the $x$ direction, $P_{0}$ controls the amplitude of the pump, $f$ is the superposition of four waves that creates the maxima at a particular type of sublattices, and the last term is a wave which has zero in-plane wave vector that fixes the minima of $P_{+}$ to zeros. In Fig. 7(a) the spatial profile of $P_{+}$ is plotted, which shows that indeed the maxima of $P_{+}$ coincide with only one type of sublattices. $P_{-}$ also has a similar spatial form, however, one needs to shift $f$ spatially such that the maxima now coincide with the other type of sublattice. We choose the pillar diameter to be $2~{}\mu$m, the lattice periodicity around $a=2.95~{}\mu$m Whittaker _et al._ (2018); Su _et al._ (2020); Jacqmin _et al._ (2014), effective mass $m_{0}=5\times 10^{-5}m_{e}$, where $m_{e}$ is the free electron mass, and the potential depth of the micropillars to be around 4 meV with 0 meV being the minimum. We also fix $\Delta_{T}=0.12$ meV.$\mu$m2 Klembt _et al._ (2018); Dufferwiel _et al._ (2015); Sedov _et al._ (2019) and $\Gamma=5.9~{}\mu$eV Nelsen _et al._ (2013), which corresponds to a polariton lifetime around 55 ps. By fixing $P_{0}=2.3~{}\mu$eV we stay near the condensation threshold, such that $|\psi_{\pm}|^{2}\approx$0\. This allows us to be in the linear limit and calculate the band structure of the system using the Bloch theorem. We also set $g=2.3$, which induces optical Zeeman shift of the lower energy modes around 0.2 meV, which is consistent with experiment Real _et al._ (2021). Now that we have all the ingredients, by applying the Bloch theorem to the linear Hamiltonian corresponding to the Eq. (VI), we calculate the band structure of the system corresponding to a strip geometry with zigzag edges as shown in Figs. 7(b-e). Similar to the tight binding model, our continuous model also predicts the antichiral band structure, where the Dirac points shifts in energy and two pairs of edge states appear. As already discussed, the opposite edge states in our system are degenerate. Consequently, to clearly show that opposite edge states have opposite spins, we plot the same band structure four times. In Figs. 7(b-c) we have omitted the edge states located at the upper edge of the strip ($\langle y\rangle=+L$), while in Figs. 7(d-e) we have omitted the edge states located at the lower edge of the strip ($\langle y\rangle=-L$). Indeed, the opposite edge states attain opposite spins, while the bulk remains linearly polarized. Although, the topological properties of the continuous model band structure is the same with those from the tight-binding band structure, there are a few differences. For example, the upper of pair edge states does not have similar dispersion as the lower pair. This may arise due to the next nearest neighbour hopping, which is naturally present in the continuous model as well as due to the continuous nature of the effective Zeeman and TE-TM splittings. The parameters that we have used are flexible. However, the TE-TM splitting and effective sublattice dependent Zeeman splitting are the crucial parameters. For the present parameters, the energy window, where the antichiral edge states resides is around 0.12 meV. This is similar to the topological bandgap obtained experimentally for the polariton Chern insulator Klembt _et al._ (2018). Being topological in nature, a slight change in the parameters will not hamper the antichiral modes. However, the case where the Zeeman splitting decreases will result in a decrease in the energy window of interest. For simplicity, we have only considered only the necessary terms in the GPE in Eq. (VI) to obtain the desired effect. For example, we have ignored the spin relaxation of the excitonic reservoir and assumed it to preserve their circular polarization perfectly. Although this is not entirely true, experiment has shown that under circularly polarized incoherent pump the reservoir can have up to 17% preferred $S_{z}$ C. Zambon _et al._ (2019), which should be enough as long as the blueshift difference between opposite circularly polarized polariton modes in a pillar is around 0.2 meV. ## VII Discussion and conclusion The spin degree of freedom of exciton-polaritons makes them suitable for optical spintronic devices. Devices based on polariton spins have been realized experimentally, which includes spin switches Amo _et al._ (2010), logic gates Gao _et al._ (2015b), etc. However, for the realization of a complete optical network, it is important to be able to communicate among different elements without losing the spin information. Unfortunately, the Chiral polaritonic edge modes do not preserve spins while propagating. Our proposed antichiral edge states could be useful in this purpose. Besides all the edge modes in our system propagate in one direction, which may be helpful in transferring oneway information through both the edges, unlike in the Chern insulator where only one edge can be used. We note that antichiral edge states for exciton-polaritons were proposed in a previous work Mandal _et al._ (2019a). The present work is significantly different from the previous one in many ways. First, in the previous work the antichiral edge modes were not coherent polaritonic states. Instead, those were fluctuations on top of a steady state. Second, in Ref. Mandal _et al._ (2019a) the antichiral edge states located at opposite edges did not have opposite spins. Last, in the previous work, the antichiral edge states were shown to propagate only in a straight strip, while their propagation around different types of bendings was not discussed. To conclude, we have presented a theoretical scheme to obtain spin-polarized antichiral edge states for the first time. We use the naturally present TE-TM splitting and sublattice depend Zeeman splitting in a honeycomb lattice of exciton-polariton micropillar, where the Dirac points shift in energy and two pairs of antichiral edge states appear that propagate in the same directions. While the bulk of the system is linearly polarized, states located at the opposite edges have opposite spins. Although the antichiral edge states reside together with the bulk modes, we found that they show fair robustness and can go around 60 degree bend without being reflected. Our work may be useful in transferring spins and connecting different spin dependent polariton logic elements. ## VIII Acknowledgements The work was supported by the Ministry of Education, Singapore (Grant No. MOE2019-T2-1-004). ## IX Appendix Here in the appendix we discuss how to calculate the winding number in detail. Different from the model considered in Colomés and Franz (2018), which has only one band below the Dirac points, our present model involves two bands. This makes the Hamiltonian complex and restricts us from calculating the winding number analytically. Additionally, because of the presence of two bands the Berry connections becomes non-Abelian (a matrix instead of a scalar) and one must rely on numerical techniques to calculate the winding number. To make the integral Gauge invariance, we use the Wilson loop approach Benalcazar _et al._ (2017). We first transform the hexagonal BZ to a rhombic shape with its axes aligned with the $k_{1}$ and $k_{2}$ direction as shown in Fig. 8.We notice that such a choice of the BZ is not unique, and the directions $(k_{1},k_{2})$ used in Colomés and Franz (2018) is related to ours by a rotation of $\pi/12$. However, we choose such a BZ as the Dirac points are situated at the central of the BZ when projected on any of the two directions. The Wilson loop operator along $k_{2}$ for a fixed value of $k_{1}$ can be defined as: Figure 8: A rhombic shape Brillouin zone with its axis aligned with the $k_{1}$ and $k_{2}$ direction. $\displaystyle W(k_{1})=$ $\displaystyle F(k_{1},k_{2}+(N-1)\Delta k)F(k_{1},k_{2}+(N-2)\Delta k)...$ $\displaystyle F(k_{1},k_{2}+\Delta k)F(k_{1},k_{2}),$ (13) where $F$ is a matrix whose elements are given by $\left[F(k_{1},k_{2})\right]^{mn}=\left\langle u^{m}(k_{1},k_{2}+\Delta k)|u^{n}(k_{1},k_{2})\right\rangle.$ (14) Here $\Delta k$ is the grid spacing in the reciprocal space, the integral is over the unit cell, $u$ is the Bloch state and $m,n=1,2$ corresponds to the lowest two bands. Next we define a Wannier Hamiltonian $H_{W(k_{1})}$ as $W(k_{1})=e^{iH_{W(k_{1})}}.$ (15) The eigenvalues of $H_{W(k_{1})}$ gives the winding numbers for each $k_{1}$ as shown in Fig. 2(h) in the main text. ## References * Klitzing _et al._ (1980) K. V. Klitzing, G. Dorda, and M. Pepper, “New method for high-accuracy determination of the fine-structure constant based on quantized hall resistance,” Phys. Rev. Lett. 45, 494 (1980). * Haldane (1988) F. D. M. Haldane, “Model for a quantum hall effect without landau levels: Condensed-matter realization of the ”parity anomaly”,” Phys. Rev. Lett. 61, 2015 (1988). * Jotzu _et al._ (2014) G. Jotzu, M. Messer, R. Desbuquois, M. Lebrat, T. Uehlinger, D. Greif, and T. Esslinger, “Experimental realization of the topological Haldane model with ultracold fermions,” Nature 515, 237 (2014). * Zhao _et al._ (2020) Y. Zhao, R. Zhang, R. Mei, L. Zhou, H. Yi, Y. Zhang, J. Yu, R. Xiao, K. Wang, N. Samarth, M. H. W. Chan, C. Liu, and C. Chang, “Tuning the Chern number in quantum anomalous Hall insulators,” Nature 588, 419 (2020). * Kane and Mele (2005) C. L. Kane and E. J. Mele, “Quantum spin hall effect in graphene,” Phys. Rev. Lett. 95, 226801 (2005). * Colomés and Franz (2018) E. Colomés and M. Franz, “Antichiral edge states in a modified haldane nanoribbon,” Phys. Rev. Lett. 120, 086603 (2018). * Mandal _et al._ (2019a) S. Mandal, R. Ge, and T. C. H. Liew, “Antichiral edge states in an exciton polariton strip,” Phys. Rev. B 99, 115423 (2019a). * Chen _et al._ (2020) J. Chen, W. Liang, and Z. Li, “Antichiral one-way edge states in a gyromagnetic photonic crystal,” Phys. Rev. B 101, 214102 (2020). * Zhou _et al._ (2020) P. Zhou, G. Liu, Y. Yang, Y. Hu, S. Ma, H. Xue, Q. Wang, L. Deng, and B. Zhang, “Observation of photonic antichiral edge states,” Phys. Rev. Lett. 125, 263603 (2020). * Chen and Li (2022) J. Chen and Z. Li, “Configurable topological beam splitting via antichiral gyromagnetic photonic crystal,” Opto-Electronic Science 1, 220001 (2022). * Yu _et al._ (2021) L. Yu, H. Xue, and B. Zhang, “Antichiral edge states in an acoustic resonator lattice with staggered air flow,” Journal of Appl. Phys. 129, 235103 (2021). * Wu _et al._ (2020) X. Wu, X. Li, R. Zhang, X. Xiang, J. Tian, Y. Huang, S. Wang, B. Hou, C. T. Chan, and W. Wen, “Deterministic scheme for two-dimensional type-ii dirac points and experimental realization in acoustics,” Phys. Rev. Lett. 124, 075501 (2020). * Yang _et al._ (2021) Y. Yang, D. Zhu, Z. Hang, and Y. Chong, “Observation of antichiral edge states in a circuit lattice,” Science China Physics, Mechanics & Astronomy 64, 257011 (2021). * Bhowmick and Sengupta (2020) D. Bhowmick and P. Sengupta, “Antichiral edge states in heisenberg ferromagnet on a honeycomb lattice,” Phys. Rev. B 101, 195133 (2020). * Mannaï and Haddad (2020) M. Mannaï and S. Haddad, “Strain tuned topology in the haldane and the modified haldane models,” Journal of Physics: Condensed Matter 32, 225501 (2020). * Breitkreiz (2020) M. Breitkreiz, “Parabolic hall effect due to copropagating surface modes,” Phys. Rev. Research 2, 012071 (2020). * Mizoguchi and Koma (2021) T. Mizoguchi and T. Koma, “Bulk-edge correspondence in two-dimensional topological semimetals: A transfer matrix study of antichiral edge modes,” Phys. Rev. B 103, 195310 (2021). * Cheng _et al._ (2021) X. Cheng, J. Chen, L. Zhang, L. Xiao, and S. Jia, “Antichiral edge states and hinge states based on the haldane model,” Phys. Rev. B 104, L081401 (2021). * Deng _et al._ (2010) H. Deng, H. Haug, and Y. Yamamoto, “Exciton-polariton bose-einstein condensation,” Rev. Mod. Phys. 82, 1489 (2010). * Carusotto and Ciuti (2013) I. Carusotto and C. Ciuti, “Quantum fluids of light,” Rev. Mod. Phys. 85, 299 (2013). * Byrnes _et al._ (2014) T. Byrnes, N. Y. Kim, and Y. Yamamoto, “Exciton–polariton condensates,” Nature Phys. 10, 803 (2014). * Sanvitto and Kéna-Cohen (2016) D. Sanvitto and S. Kéna-Cohen, “The road towards polaritonic devices,” Nature Mater. 15, 1061 (2016). * Feng _et al._ (2021) J. Feng, J. Wang, A. Fieramosca, R. Bao, J. Zhao, R. Su, Y. Peng, T. C. H. Liew, D. Sanvitto, and Q. Xiong, “All-optical switching based on interacting exciton polaritons in self-assembled perovskite microwires,” Science Advances 7, eabj6627 (2021). * Kyriienko _et al._ (2019) O. Kyriienko, H. Sigurdsson, and T. C. H. Liew, “Probabilistic solving of $np$-hard problems with bistable nonlinear optical networks,” Phys. Rev. B 99, 195301 (2019). * Xu _et al._ (2020) H. Xu, S. Ghosh, M. Matuszewski, and T. C.H. Liew, “Universal self-correcting computing with disordered exciton-polariton neural networks,” Phys. Rev. Applied 13, 064074 (2020). * Banerjee and Liew (2020) R. Banerjee and T. C. H. Liew, “Artificial life in an exciton-polariton lattice,” New Journal of Phys. 22, 103062 (2020). * Harrison _et al._ (2022) S. L. Harrison, H. Sigurdsson, S. Alyatkin, J.D. Töpfer, and P. G. Lagoudakis, “Solving the max-3-cut problem with coherent networks,” Phys. Rev. Applied 17, 024063 (2022). * Xu _et al._ (2021a) H. Xu, T. Krisnanda, W. Verstraelen, T. C. H. Liew, and S. Ghosh, “Superpolynomial quantum enhancement in polaritonic neuromorphic computing,” Phys. Rev. B 103, 195302 (2021a). * Zvyagintseva _et al._ (2022) D. Zvyagintseva, H. Sigurdsson, V. K. Kozin, I. Iorsh, I. A. Shelykh, V. Ulyantsev, and O. Kyriienko, “Machine learning of phase transitions in nonlinear polariton lattices,” Commun. Phys. 5, 8 (2022). * Kavokin _et al._ (2005) A. Kavokin, G. Malpuech, and M. Glazov, “Optical spin hall effect,” Phys. Rev. Lett. 95, 136601 (2005). * Leyder _et al._ (2007) C. Leyder, M. Romanelli, J. Ph. Karr, E. Giacobino, T. C. H. Liew, M. M. Glazov, A. V. Kavokin, G. Malpuech, and A. Bramati, “Observation of the optical spin Hall effect,” Nature Phys. 3, 628 (2007). * Kammann _et al._ (2012) E. Kammann, T. C. H. Liew, H. Ohadi, P. Cilibrizzi, P. Tsotsis, Z. Hatzopoulos, P. G. Savvidis, A. V. Kavokin, and P. G. Lagoudakis, “Nonlinear optical spin hall effect and long-range spin transport in polariton lasers,” Phys. Rev. Lett. 109, 036404 (2012). * Larionov _et al._ (2010) A. V. Larionov, V. D. Kulakovskii, S. Höfling, C. Schneider, L. Worschech, and A. Forchel, “Polarized nonequilibrium bose-einstein condensates of spinor exciton polaritons in a magnetic field,” Phys. Rev. Lett. 105, 256401 (2010). * Karzig _et al._ (2015) T. Karzig, C. E. Bardyn, N. H. Lindner, and G. Refael, “Topological polaritons,” Phys. Rev. X 5, 031001 (2015). * Bardyn _et al._ (2015) C. E. Bardyn, T. Karzig, G. Refael, and T. C. H. Liew, “Topological polaritons and excitons in garden-variety systems,” Phys. Rev. B 91, 161413 (2015). * Nalitov _et al._ (2015) A. V. Nalitov, D. D. Solnyshkov, and G. Malpuech, “Polariton $\mathbb{Z}$ topological insulator,” Phys. Rev. Lett. 114, 116401 (2015). * Banerjee _et al._ (2018) R. Banerjee, T. C. H. Liew, and O. Kyriienko, “Realization of hofstadter’s butterfly and a one-way edge mode in a polaritonic system,” Phys. Rev. B 98, 075412 (2018). * Li _et al._ (2018) C. Li, F. Ye, X. Chen, Y. V. Kartashov, A. Ferrando, L. Torner, and D. V. Skryabin, “Lieb polariton topological insulators,” Phys. Rev. B 97, 081103 (2018). * Sun _et al._ (2019) M. Sun, D. Ko, D. Leykam, V. M. Kovalev, and I. G. Savenko, “Exciton-polariton topological insulator with an array of magnetic dots,” Phys. Rev. Applied 12, 064028 (2019). * Solnyshkov _et al._ (2018) D. D. Solnyshkov, O. Bleu, and G. Malpuech, “Topological optical isolator based on polariton graphene,” Appl. Phys. Lett. 112, 031106 (2018). * Mandal _et al._ (2019b) S. Mandal, R. Banerjee, and T. C. H. Liew, “One-way reflection-free exciton-polariton spin-filtering channel,” Phys. Rev. Applied 12, 054058 (2019b). * Klembt _et al._ (2018) S. Klembt, T. H. Harder, O. A. Egorov, K. Winkler, R. Ge, M. A. Bandres, M. Emmerling, L. Worschech, T. C. H. Liew, M. Segev, C. Schneider, and S. Höfling, “Exciton-polariton topological insulator,” Nature 562, 552 (2018). * Kartashov and Skryabin (2016) Y. V. Kartashov and D. V. Skryabin, “Modulational instability and solitary waves in polariton topological insulators,” Optica 3, 1228 (2016). * Gulevich _et al._ (2017) D. R. Gulevich, D. Yudin, D. V. Skryabin, I. V. Iorsh, and I. A. Shelykh, “Exploring nonlinear topological states of matter with exciton-polaritons: Edge solitons in kagome lattice,” Scientific Reports 7, 1780 (2017). * Pernet _et al._ (2022) N. Pernet, P. St-Jean, D. D. Solnyshkov, G. Malpuech, N. Carlon Zambon, Q. Fontaine, B. Real, O. Jamadi, A. Lemaître, M. Morassi, L. Le Gratiet, T. Baptiste, A. Harouri, I. Sagnes, A. Amo, S. Ravets, and J. Bloch, “Gap solitons in a one-dimensional driven-dissipative topological lattice,” Nature Physics (2022), 10.1038/s41567-022-01599-8. * Solnyshkov _et al._ (2017) D. D. Solnyshkov, O. Bleu, B. Teklu, and G. Malpuech, “Chirality of topological gap solitons in bosonic dimer chains,” Phys. Rev. Lett. 118, 023901 (2017). * Bleu _et al._ (2017) O. Bleu, D. D. Solnyshkov, and G. Malpuech, “Photonic versus electronic quantum anomalous hall effect,” Phys. Rev. B 95, 115415 (2017). * Kartashov and Skryabin (2017) Y. V. Kartashov and D. V. Skryabin, “Bistable topological insulator with exciton-polaritons,” Phys. Rev. Lett. 119, 253904 (2017). * Banerjee _et al._ (2020) R. Banerjee, S. Mandal, and T. C. H. Liew, “Coupling between exciton-polariton corner modes through edge states,” Phys. Rev. Lett. 124, 063901 (2020). * Bardyn _et al._ (2016) C. E. Bardyn, T. Karzig, G. Refael, and T. C. H. Liew, “Chiral bogoliubov excitations in nonlinear bosonic systems,” Phys. Rev. B 93, 020502 (2016). * Sigurdsson _et al._ (2017) H. Sigurdsson, G. Li, and T. C. H. Liew, “Spontaneous and superfluid chiral edge states in exciton-polariton condensates,” Phys. Rev. B 96, 115453 (2017). * Banerjee _et al._ (2021) R. Banerjee, S. Mandal, and T. C. H. Liew, “Optically induced topological spin-valley hall effect for exciton polaritons,” Phys. Rev. B 103, L201406 (2021). * Gao _et al._ (2015a) T. Gao, E. Estrecho, K. Y. Bliokh, T. C. H. Liew, M. D. Fraser, S. Brodbeck, M. Kamp, C. Schneider, S. Höfling, Y. Yamamoto, F. Nori, Y. S. Kivshar, A. G. Truscott, R. G. Dall, and E. A. Ostrovskaya, “Observation of non-hermitian degeneracies in a chaotic exciton-polariton billiard,” Nature 526, 554–558 (2015a). * Gao _et al._ (2018) W. Gao, X. Li, M. Bamba, and J. Kono, “Continuous transition between weak and ultrastrong coupling through exceptional points in carbon nanotube microcavity exciton–polaritons,” Nature Photonics 12, 362–367 (2018). * Khurgin (2020) J. B. Khurgin, “Exceptional points in polaritonic cavities and subthreshold fabry–perot lasers,” Optica 7, 1015–1023 (2020). * Comaron _et al._ (2020) P. Comaron, V. Shahnazaryan, W. Brzezicki, T. Hyart, and M. Matuszewski, “Non-hermitian topological end-mode lasing in polariton systems,” Phys. Rev. Research 2, 022051 (2020). * Su _et al._ (2021) R. Su, E. Estrecho, D. Bieganska, Y. Huang, M. Wurdack, M. Pieczarka, A. G. Truscott, T. C. H. Liew, E. A. Ostrovskaya, and Q. Xiong, “Direct measurement of a non-hermitian topological invariant in a hybrid light-matter system,” Science Advances 7, eabj8905 (2021). * Liao _et al._ (2021) Q. Liao, C. Leblanc, J. Ren, F. Li, Y. Li, D. Solnyshkov, G. Malpuech, J. Yao, and H. Fu, “Experimental measurement of the divergent quantum metric of an exceptional point,” Phys. Rev. Lett. 127, 107402 (2021). * Krol _et al._ (2022) M. Krol, I. Septembre, P. Oliwa, M. Kedziora, K. Lempicka-Mirek, M. Muszyński, R. Mazur, P. Morawiak, W. Piecek, P. Kula, W. Bardyszewski, P. G. Lagoudakis, D. D. Solnyshkov, G. Malpuech, B. Pietka, and J. Szczytko, “Annihilation of exceptional points from different dirac valleys in a 2d photonic system,” Nature Communications 13, 5340 (2022). * Mandal _et al._ (2020) S. Mandal, R. Banerjee, Elena A. Ostrovskaya, and T. C. H. Liew, “Nonreciprocal transport of exciton polaritons in a non-hermitian chain,” Phys. Rev. Lett. 125, 123902 (2020). * Xu _et al._ (2021b) H. Xu, K. Dini, X. Xu, R. Banerjee, S. Mandal, and T. C. H. Liew, “Nonreciprocal exciton-polariton ring lattices,” Phys. Rev. B 104, 195301 (2021b). * Mandal _et al._ (2022) S. Mandal, R. Banerjee, and T. C. H. Liew, “From the topological spin-hall effect to the non-hermitian skin effect in an elliptical micropillar chain,” ACS Photon. 9, 527 (2022). * Xu _et al._ (2021c) X. Xu, H. Xu, S. Mandal, R. Banerjee, S. Ghosh, and T. C. H. Liew, “Interaction-induced double-sided skin effect in an exciton-polariton system,” Phys. Rev. B 103, 235306 (2021c). * Pickup _et al._ (2020) L. Pickup, H. Sigurdsson, J. Ruostekoski, and P. G. Lagoudakis, “Synthetic band-structure engineering in polariton crystals with non-Hermitian topological phases,” Nature Commun. 11, 4431 (2020). * Bleu _et al._ (2018) O. Bleu, G. Malpuech, and D. D. Solnyshkov, “Robust quantum valley hall effect for vortices in an interacting bosonic quantum fluid,” Nature Communications 9, 3991 (2018). * Qi _et al._ (2008) X. L. Qi, T. L. Hughes, and S. C. Zhang, “Topological field theory of time-reversal invariant insulators,” Phys. Rev. B 78, 195424 (2008). * Panzarini _et al._ (1999) G. Panzarini, L. C. Andreani, A. Armitage, D. Baxter, M. S. Skolnick, V. N. Astratov, J. S. Roberts, A. V. Kavokin, M. R. Vladimirova, and M. A. Kaliteevski, “Exciton-light coupling in single and coupled semiconductor microcavities: Polariton dispersion and polarization splitting,” Phys. Rev. B 59, 5082 (1999). * Sigurdsson _et al._ (2019) H. Sigurdsson, Y. S. Krivosenko, I. V. Iorsh, I. A. Shelykh, and A. V. Nalitov, “Spontaneous topological transitions in a honeycomb lattice of exciton-polariton condensates due to spin bifurcations,” Phys. Rev. B 100, 235444 (2019). * Real _et al._ (2021) B. Real, N. Carlon Zambon, P. St-Jean, I. Sagnes, A. Lemaître, L. Le Gratiet, A. Harouri, S. Ravets, J. Bloch, and A. Amo, “Chiral emission induced by optical zeeman effect in polariton micropillars,” Phys. Rev. Research 3, 043161 (2021). * Ohadi _et al._ (2017) H. Ohadi, A. J. Ramsay, H. Sigurdsson, Y. del Valle-Inclan Redondo, S. I. Tsintzos, Z. Hatzopoulos, T. C. H. Liew, I. A. Shelykh, Y. G. Rubo, P. G. Savvidis, and J. J. Baumberg, “Spin order and phase transitions in chains of polariton condensates,” Phys. Rev. Lett. 119, 067401 (2017). * Ohadi _et al._ (2018) H. Ohadi, Y. del Valle-Inclan Redondo, A. J. Ramsay, Z. Hatzopoulos, T. C. H. Liew, P. R. Eastham, P. G. Savvidis, and J. J. Baumberg, “Synchronization crossover of polariton condensates in weakly disordered lattices,” Phys. Rev. B 97, 195109 (2018). * Töpfer _et al._ (2021) J. D. Töpfer, I. Chatzopoulos, H. Sigurdsson, T. Cookson, Y. G. Rubo, and P. G. Lagoudakis, “Engineering spatial coherence in lattices of polariton condensates,” Optica 8, 106–113 (2021). * Opala _et al._ (2022) A. Opala, R. Panico, V. Ardizzone, B. Pietka, J. Szczytko, D. Sanvitto, M. Matuszewski, and D. Ballarini, “Training a neural network with exciton-polariton optical nonlinearity,” Phys. Rev. Applied 18, 024028 (2022). * Luo _et al._ (2020) S. Luo, L. Liao, Z. Zhang, J. Wang, X. Shen, and Z. Chen, “Classical spin chains mimicked by room-temperature polariton condensates,” Phys. Rev. Applied 13, 044052 (2020). * Whittaker _et al._ (2018) C. E. Whittaker, E. Cancellieri, P. M. Walker, D. R. Gulevich, H. Schomerus, D. Vaitiekus, B. Royall, D. M. Whittaker, E. Clarke, I. V. Iorsh, I. A. Shelykh, M. S. Skolnick, and D. N. Krizhanovskii, “Exciton polaritons in a two-dimensional lieb lattice with spin-orbit coupling,” Phys. Rev. Lett. 120, 097401 (2018). * Su _et al._ (2020) R. Su, Sanjib G., J. Wang, S. Liu, C. Diederichs, T. C. H. Liew, and Q. Xiong, “Observation of exciton polariton condensation in a perovskite lattice at room temperature,” Nature Physics 16, 301–306 (2020). * Jacqmin _et al._ (2014) T. Jacqmin, I. Carusotto, I. Sagnes, M. Abbarchi, D. D. Solnyshkov, G. Malpuech, E. Galopin, A. Lemaître, J. Bloch, and A. Amo, “Direct observation of dirac cones and a flatband in a honeycomb lattice for polaritons,” Phys. Rev. Lett. 112, 116402 (2014). * Dufferwiel _et al._ (2015) S. Dufferwiel, Feng Li, E. Cancellieri, L. Giriunas, A. A. P. Trichet, D. M. Whittaker, P. M. Walker, F. Fras, E. Clarke, J. M. Smith, M. S. Skolnick, and D. N. Krizhanovskii, “Spin textures of exciton-polaritons in a tunable microcavity with large te-tm splitting,” Phys. Rev. Lett. 115, 246401 (2015). * Sedov _et al._ (2019) E. S. Sedov, Y. G. Rubo, and A. V. Kavokin, “Polariton polarization rectifier,” Light: Science & Applications 8, 79 (2019). * Nelsen _et al._ (2013) B. Nelsen, G. Liu, M. Steger, D. W. Snoke, R. Balili, K West, and L. Pfeiffer, “Dissipationless flow and sharp threshold of a polariton condensate with long lifetime,” Phys. Rev. X 3, 041015 (2013). * C. Zambon _et al._ (2019) N. C. Zambon, P. St-Jean, M. Milićević, A. Lemaître, A. Harouri, L. Le Gratiet, O. Bleu, D. D. Solnyshkov, G. Malpuech, I. Sagnes, S. Ravets, A. Amo, and J. Bloch, “Optically controlling the emission chirality of microlasers,” Nature Photon. 13, 283 (2019). * Amo _et al._ (2010) A. Amo, T. C. H. Liew, C. Adrados, R. Houdré, E. Giacobino, A. V. Kavokin, and A. Bramati, “Exciton–polariton spin switches,” Nature Photon. 4, 361 (2010). * Gao _et al._ (2015b) T. Gao, C. Antón, T. C. H. Liew, M. D. Martín, Z. Hatzopoulos, L. Viña, P. S. Eldridge, and P. G. Savvidis, “Spin selective filtering of polariton condensate flow,” Appl. Phys. Lett. 107, 011106 (2015b). * Benalcazar _et al._ (2017) W. A. Benalcazar, B. A. Bernevig, and T. L. Hughes, “Quantized electric multipole insulators,” Science 357, 61–66 (2017).
A Pseudo-Value Regression Approach for Differential Network Analysis of Co- Expression Data–A Pseudo-Value Regression Approach for Differential Network Analysis of Co-Expression Data October # A Pseudo-Value Regression Approach for Differential Network Analysis of Co- Expression Data Seungjun Ahn1 Tyler Grimes2 and Somnath Datta1 1Department of Biostatistics University of Florida Gainesville Florida U.S.A. 2Department of Mathematics and Statistics University of North Florida Jacksonville Florida U.S.A (2022) ###### Abstract The differential network (DN) analysis detects changes in measures of association among genes under two or more conditions. In this study, we introduce a Pseudo-value Regression Approach for Network Analysis (PRANA). This is a novel method of DN analysis that also adjusts for additional clinical covariates. We start from mutual information (MI) criteria, followed by pseudo-value calculations, which are then entered into a robust regression model. Performance in terms of precision, recall, and F1 score of differentially connected (DC) genes is assessed in both univariable and multivariable settings through variety of simulations. By and large, PRANA outperformed dnapath and DINGO, neither of which is equipped to adjust for available covariates such as patient-age. Lastly, we employ PRANA in a real data application from the Gene Expression Omnibus (GEO) database to identify DC genes that are associated with chronic obstructive pulmonary disease (COPD) to demonstrate its utility. This is the first attempt of utilizing a regression modeling for DN analysis by collective gene expression levels between two or more groups with the inclusion of additional covariates. ###### keywords: Differential network analysis; Gene regulatory network; Pseudo-value; Regression method; RNA-Seq data. ## 1 Introduction The rapid advancement of RNA-sequencing (RNA-Seq) data from high-throughput sequencing technologies has provided clear advantages in gene expression studies. It has broadened our understanding of genetics and pathogenesis of human diseases (Reuter et al., 2015; Gao et al., 2017). Compared to microarrays, RNA-Seq has a wider dynamic range, the ability to detect novel transcripts, and often results in higher sensitivity and specificity of detection of differential gene expression (Wang et al., 2009; Zhao et al., 2014). With the increase of gene expression studies, the statistical methods to analyze these gene expression data have also accordingly adapted and progressed. The regression modelling for RNA-Seq differential expression (DE) analysis has been established to compare the number of DE genes under different biological or clinical states, including linear model based limma (Ritchie et al., 2015), negative binomial model based edgeR (Robinson et al., 2010), or Poisson log-linear model approach (Li et al., 2012). The analysis of DE has a major limitation in that it looks at one gene at a time, even though a set of genes are often involved in the same biological process (Ideker and Krogan, 2012; Kim et al., 2018). In contrast, the differential network (DN) analysis complements the DE analysis (de la Fuente, 2010) by looking at genes collectively. The DN analysis identifies changes in measures of association (i.e. network properties or topologies) of the networks across biological conditions, which makes it distinct from a single network analysis. Several groups have proposed statistical methods for DN analysis (Ha et al., 2015; McKenzie et al., 2016; Gill et al., 2010; Grimes et al., 2019). In particular, DINGO (Ha et al., 2015) and dnapath (Grimes et al., 2019) have developed methods for RNA-Seq data, to find differentially connected (DC) genes in subnetworks corresponding to different pathways, between two groups of patients; e.g. ‘high-risk’ vs. ‘low-risk’ or ‘long-term survivors’ vs. ‘short-term survivors.’ While these methods are convenient to use and applicable, they do not consider other observed covariates that may be associated with gene expression. For instance, a previous study has shown that the expression levels of oxidative stress-associated genes were differentially expressed with smokers with chronic obstructive pulmonary disease (COPD) through gene set enrichment analysis using microarray data (Marballi et al., 2016). Let us suppose we want to carry out DN analysis on expression data that includes oxidative stress- associated genes and smoking status, which would be used as a grouping variable. In practice, clinicians would also want to include additional covariates such as patient history of cardiac arrhythmia (Bruha et al., 2012) and lung carcinoma (Herreros-Villanueva et al., 2013) to garner more information for better prognosis. However, there is no available direct regression modeling for DN analysis regressing gene expression level between the smoking statuses with the inclusion of additional clinical covariates described above. The pseudo-value approach was first developed from the leave-one-out jackknife subsampling procedure, applied to a marginal quantity representing some aspect of a marginal distribution of the response variable. It was originally introduced by Andersen and his colleagues (Andersen et al., 2003; Andersen and Klein, 2007) for multi-state survival models. Several studies (Dutta et al., 2017; Graw et al., 2009) purported that the pseudo-value regression has advantages that the pseudo-values derived from an asymptotically linear and unbiased estimator are approximately independent and identically distributed with the same conditional expectation. Ahn and Logan (Ahn and Logan, 2016) and Ahn and Mendolia (Ahn and Mendolia, 2014) showed that their pseudo-value approaches controlled the type I error while maintaining high power with clustered survival data. With these benefits, we propose a regression modeling method that regresses the jackknife pseudo-values (Efron and Tibshirani, 1993) derived from a measure of connectivity of genes in a network to estimate the effects of predictors. Note that the grouping variable itself could also be included in the regression model along with additional clinical covariates while regressing the pseudo-values. We loosely refer to this as a “multivariate setting”, whereas in “univariate settings” only the group variable is utilized in a DN analysis. Thus, in this paper, we introduce a Pseudo-value Regression Approach for Network Analysis (PRANA). This is a novel method of DN analysis that can adjust for additional covariates. We start from mutual information (MI) criteria, followed by pseudo-value calculations, which are then entered into a robust regression model. This article assesses the model performances of our pseudo-value approach in a multivariable setting, followed by a comparison to dnapath and DINGO in the univariable setting through simulations. Lastly, we employ our method in a real data application (Wang et al., 2021) from the Gene Expression Omnibus (GEO) database (Barrett et al., 2013) to identify DC genes that are associated with COPD. All statistical analyses are performed in R version 4.0.2 (R Foundation for Statistical Computing, Vienna, Austria). ## 2 Methods ### 2.1 Mutual Information and ARACNE Mutual information (MI) determines whether and how two genes interact. That is, it is a measure of their relatedness and calculated from their joint expression profiles. MI is zero if and only if the joint distribution between the expression level of gene $j$ and gene $k$ satisfies $P(g_{j},g_{k})=P(g_{j})P(g_{k})$ for $j\neq k$, or if $j=k$; in other words, they are statistically independent. Algorithm for the Reconstruction of Accurate Cellular Networks (ARACNE) which are proposed by Margolin et al. (2006) and Zoppoli et al. (2010) estimates MI using a computationally efficient Gaussian kernel estimator. The estimate of MI is used to quantify the connectivity of each pair of genes in a network. Given a set of two genes measurements, $\overrightarrow{u_{i}}\equiv(g_{ij},g_{ik})$, $i=1,\ldots,n$, the joint probability distribution is approximated as $f(\overrightarrow{u})=1/n\sum_{i}h^{-2}\Phi(h^{-1}(\overrightarrow{u}-\overrightarrow{u_{i}}))$, where $\Phi$ is the bivariate standard normal density and $h$ is the position- dependent kernel width. Then the MI can be expressed as (Margolin et al., 2006): $\hat{I}_{jk}=\frac{1}{n}\sum_{i}\log\frac{f(g_{ij},g_{ik})}{f(g_{ij})f(g_{ik})},$ where $f(g_{j})$ and $f(g_{k})$ are the marginals of $f(\overrightarrow{u})$. The matrix containing entries $\hat{I}_{jk}$ is defined as the association matrix. The ARACNE algorithm copula-transform the profiles for MI estimation because MI is reparameterization invariant; thus, the range of these transformed variables is between 0 and 1 (Basso et al., 2005). ### 2.2 Pseudo-Value Approach Let $\hat{I}_{jk}$ be the MI estimate for a pair of genes $j,k\in\\{1,\dots,p\\}$ of an estimated network from $n$ individuals. For each gene $k$, we sum the edges (MI estimates) around the gene by taking the column sum of the association matrix to obtain the total connectivity (which can be deemed as a continuous version of degree centrality) of gene $k$: $\hat{\theta}_{k}=\sum_{j=1}^{p}\hat{I}_{jk},$ where $k=1,\dots,p$. The jackknife pseudo-values (Efron and Tibshirani, 1993) for the $i^{\text{th}}$ individual and $k^{\text{th}}$ gene are defined by: $\tilde{\theta}_{ik}=n\hat{\theta}_{k}-(n-1)\hat{\theta}_{k(i)},$ (1) where $\hat{\theta}_{k(i)}$ is the column sum of a gene calculated from the re-estimated association matrix using the RNA-Seq data without the $i^{\text{th}}$ subject. For each gene $k$, the re-estimation process requires $n$ such calculations with the data size of $n-1$. Let $Z$ a binary group indicator. Let $\mathcal{G}_{1}=\\{i:Z_{i}=1\\}$, $\mathcal{G}_{2}=\\{i:Z_{i}=2\\}$, and $n_{z}=|\mathcal{G}_{z}|$ is the sample size for the two groups $z=1,2$ and $n=\sum n_{z}$. The jackknife pseudo- values are separately obtained within groups. Following the general formula above, for gene $k$ and group $z$, we similarly define $\hat{\theta}_{k}^{z}$ and $\hat{\theta}_{k(i)}^{z}$, where $i=1,\ldots,n_{z}$. Then for each $i\in\mathcal{G}_{z}$, the $k^{\text{th}}$ gene jackknife pseudo-values are calculated by $\tilde{\theta}_{ik}=n_{z}\hat{\theta}_{k}^{z}-(n_{z}-1)\hat{\theta}_{k(i)}^{z}$. Next, a robust regression is applied to regress the pseudo-values on a set of covariates, including $Z$ and X, where $Z$ is the group indicator and $\textbf{X}=(X_{1},\dots,X_{q}$) are the potential confounders, such as age and gender. For the $i^{\text{th}}$ individual and $k^{\text{th}}$ gene, we posit the model $E[\tilde{\theta}_{ik}|Z_{i},\textbf{X}_{i}]=\alpha_{k}+\beta_{k}Z_{i}+\gamma_{k1}X_{i1}+\dots+\gamma_{kq}X_{iq},$ (2) where $\alpha_{k}$ is the intercept, $\beta_{k}$ is the regression coefficient for $Z$, and $\gamma_{k1},\dots,\gamma_{kq}$ is the set of regression coefficients to be estimated for $X$. The main parameter of interest is $\beta_{k}$ to test for the change in total connectivity (or degree centrality) of the $k^{\text{th}}$ gene between groups. Least trimmed squares (LTS), also known as least trimmed sum of squares (Rousseeuw, 1984), is then implemented to perform a robust regression. The LTS estimator is defined by $\min_{\alpha_{k},\beta_{k},\gamma_{k1},\ldots,\gamma_{kq}}\sum_{i=1}^{h}r_{(i)}(\alpha_{k},\beta_{k},\gamma_{k1},\ldots,\gamma_{kq})^{2},$ where $r_{(i)}$ is the set of ordered absolute values of the residuals (in increasing order of absolute value) and $h$ may depend on some pre-defined trimming proportion $c$, for instance by means of $h=[n(1-c)]+1$. In general, $c$ is chosen between 0.5 and 1 (Pison et al., 2002). ### 2.3 Hypothesis Testing To test whether the true difference in total connectivity of $k^{\text{th}}$ gene differs between groups, we test the null hypothesis of $H_{0}:\beta_{k}=0$ against the research hypothesis $H_{1}:\beta_{k}\neq 0$. The t-statistic is computed by $\hat{\beta}_{k}/SE(\hat{\beta}_{k})$, where $SE(\hat{\beta}_{k})$ standard error of $\hat{\beta}_{k}$, obtained using large sample theory, and which is the least-squares estimator of $\beta_{k}$ for $k=1,\dots,p$ from the robust regression described in equation (2). P-values are calculated using a t-distribution as in robustbase R package (Todorov and Filzmoser, 2009). It is important to control the false discovery rate (FDR), since multiple hypothesis tests are conducted in the DN analysis. The FDR measures the proportion of false discoveries among a set of genes which are significantly DC between groups. The empirical Bayes screening (EBS) approach (Datta and Datta, 2005) has been applied to control the FDR, which is an extension of Westfall and Young step-down adjusted p-values (Westfall and Young, 1993). The EBS procedure is robust against model mis-specification, as it utilizes nonparametric function estimation techniques for the estimation of the marginal density of the transformed p-values. ## 3 Materials This section details step-by-step procedures how the simulation is performed. The performance of our proposed method is assessed by an extensive simulation study. Data are simulated with different number of genes $p$ and sample size $n$. In this simulation, the regression model includes two covariates $Z$ and $X$, where $Z$ is the group indicator and $X\sim N(55,10)$ is a continuous covariate such as the age of a patient. Three different simulation scenarios are considered. ### 3.1 Data Generation Simulate weighted networks and RNA-Seq data with a dependence structure that depends on $Z$ and/or $X$ using the SeqNet R package (Grimes and Datta, 2021). In this setting, there are total of six networks for the combination of two groups and three age categories (younger than 50, 50-60, and older than 60). We consider three different scenarios incorporating group information only (scenario I), age and group information (scenario II), and age and group information with unequal sampling proportions with different distributions of the age in the two groups (scenario III) (see Figures 1–3 for visual demonstrations). #### Scenarios I (a–b) and II (a–c) * a. Generate the first random network with $p$ nodes for $z=1$. The $p\times p$ adjacency matrix, where the diagonal elements are 0 and non-diagonal elements are in $\\{0,1\\}$, is extracted from this first graph. It is a symmetric matrix indicating whether a pair of nodes are connected by an edge. Take the column sum of the adjacency matrix to see the total number of connected edges to the node. Record the indices of this vector with column sum for the use of effect size adjustment in later step. * b. Perturb the first network to generate the second network for $z=2$ by removing the edges around nodes using the indices obtained in previous step. To assess the effect size of group, the top $5\%$, $10\%$, and $20\%$ of total nodes with the most number of edges in a network lose their edges (e.g. 2 nodes with the most number of edges for a network with $p=20$ for the effect size of $10\%$, Figure 1). This is the end of scenario I. * c. For scenario II, further perturb remaining networks by removing edges of one additional node with the next most number of edges, coming after Step (b) above. This is to simulate networks with a covariate dependence structure on both age and group (see Figure 2). #### Scenario III We created a scenario where age is acting like a confounder. In other words, for a given each category that two networks are the same, but the distributions of the age of the patients are different in the two groups. Therefore, there will be an observed difference in network connectivity, which is explained through age. * a. Generate the first random network with $p$ nodes for younger than 50 category. The $p\times p$ adjacency matrix, where the diagonal elements are 0 and non- diagonal elements are either $\\{0,1\\}$, is extracted from this first graph. Record the indices of connected edges for the perturbation of network in later steps below. * b. Perturb the first network to generate the second network for age 50-60 category by removing the edges of the two nodes with the most number of edges in a network lose all of their edges. In other words, we refer to the indices, recorded in the adjacency matrix from the earlier step, and remove all the connected edges around the two nodes. * c. Next, we repeat the same to perturb the second network to obtain the third network for older than 60 category (see Figure 3). Figure 1: Network plots visualizing the gene network (p = 20) without a covariate dependence structure that depends on binary group only (scenario I). The row represents group whereas the column represents age categories. The three networks in each row are identical, since there is no effect of age on the structure of network. The edges of the hub nodes are removed based on the effect size of the binary group. Figure 2: Network plots visualizing the gene network (p = 20) with a covariate dependence structure that depends on age and group information (scenario II). The row represents group whereas the column represents age categories. All six networks have unique structure of the network. The edges of the hub nodes are firstly removed based on the effect size of the group, as shown in Figure 1 above. For this scenario II, additional edges of nodes with greater number of connected edges are removed for each age category. Figure 3: Network plots visualizing the gene network (p = 20) with a covariate dependence structure that depends on age and group information with unequal sampling proportions with respect to different distribution of the age in the two groups (scenario III). The row represents group whereas the column represents age categories. All six networks have unique structure of the network. The edges of the two hub nodes are removed for each age category. To employ the effect of group, $10\%/10\%/80\%$ of the subjects in $z=1$ will have a network structure to each of the first, second, and third networks in the first row. In contrast, $80\%/10\%/10\%$ of the subjects in $z=2$ will have a network structure to each of the first, second, and third networks in the second row. For all three scenarios, we assign a partial correlation to edges to obtain weighted networks (Grimes and Datta, 2021). Note that adjacency matrices of these weighted networks are used for the the true connection per gene. Generate RNA-Seq samples based on weighted networks with equal sampling proportions for scenarios I and II. However, specifically for scenario III, a sampling proportion differs across age categories and groups. That is, $10\%/10\%/80\%$ for $z=1$ and $80\%/10\%/10\%$ for $z=2$. The data generation involves with two major steps. Firstly, we generate gene expressions (Gaussian values) from a group-specific weighted network for each gene, denoted as $\tilde{x}_{i}\sim N(0,1)$. These Gaussian values are then mapped into RNA-Seq data column-wise by using the inverse CDF of empirical distribution of the reference data using expression data with accession number GSE158699 (Wang et al., 2021) from the Gene Expression Omnibus (GEO) database (Barrett et al., 2013). We will have $n_{z}\times p$ matrices for each group $z=1,2$. ### 3.2 Algorithm [enumerate] Obtain an association matrix with ARACNE from the data generated in steps from the “Data Generation” section to fit an estimated network using minet (Margolin et al., 2006; Meyer et al., 2008) for each group. For each gene $k$, calculate the column sums of association matrix for each group $z$ separately, denoted by $\hat{\theta}_{k}^{z}$. For each gene $k$ and individual $i\in\mathcal{G}_{z}$, compute $\hat{\theta}_{k(i)}^{z}$ from the association matrix that is re-estimated based on RNA-Seq data without the $i^{\text{th}}$ subject of $n_{z}\times p$ data from the “Data Generation” section for each group $z$ separately, where $i=1,\dots,n_{z}$. Calculate $\tilde{\theta}_{ik}$ using equation (1) based on Step 2 and 3. For each gene $k$, fit a multivariable robust regression with binary group variable and continuous age variable to obtain the p-values of the group variable, computed from the t-test. These p-values are used to compute the performance measures of simulation study. More details on the performance measures are stated next. ### 3.3 Performance Measures To evaluate the performance of our proposed method, precision, recall, and the F1 score are calculated. Let $\Omega^{z}\in\mathbf{R}^{p\times p}$ be the adjacency matrix for group $z$, where $\Omega_{jk}^{z}=\begin{cases}1&\text{if }j^{\text{th}}\text{ gene and }k^{\text{th}}\text{ gene are connected}\\\ 0&\text{otherwise},\end{cases}$ for $z=1,2$. Then, for each gene $k$, we calculate $\eta_{k}=I\bigg{(}\sum_{j=1}^{p}|\Omega_{jk}^{1}-\Omega_{jk}^{2}|\geq 1\bigg{)},$ where $I(\cdot)$ is an indicator function to determine whether gene $k$ has differential connectivity. The gene $k$ is truly DC if $\eta_{k}=1$, and is not DC if $\eta_{k}=0$ for the true gene network. Similarly, for the covariate dependence structure, the following quantities are obtained $\eta_{k}=I\bigg{(}\frac{1}{c}\sum_{c}\sum_{j=1}^{p}|\Omega_{jk}^{1,c}-\Omega_{jk}^{2,c}|\geq 1\bigg{)},$ where $\Omega^{z,c}\in\mathbf{R}^{p\times p}$ be the adjacency matrix for group $z$ and age category $c=1,2,3$. Denote that $S$ is the total number of Monte Carlo simulation replicates. Let $q_{ks}$ be adjusted p-value as in following procedure (Datta and Datta, 2005) of $k^{\text{th}}$ gene at the $s^{\text{th}}$ simulation replicate. $\alpha$ represents the magnitude of error control, and 0.05 was used throughout the simulation. * - Precision is the proportion of genes that are inferred to be significantly DC from the test which have true connection between two comparing groups: $\text{Precision}=\frac{\sum_{k=1}^{p}\eta_{k}\,I(q_{ks}<\alpha)}{\sum_{k=1}^{p}I(q_{ks}<\alpha)}.$ * - Recall is the proportion of genes that have true connection which are correctly inferred to be significantly DC between two comparing groups from the test: $\text{Recall}=\frac{\sum_{k=1}^{p}\eta_{k}\,I(q_{ks}<\alpha)}{\sum_{k=1}^{p}\eta_{k}}.$ * - F1 is calculated based on the harmonic mean of precision and recall obtained from the simulation. A higher F1 score suggests lower false negative and false positive rate: $\text{F1}=2\cdot\frac{\text{Precision}\cdot\text{Recall}}{\text{Precision}+\text{Recall}}.$ ### 3.4 COPDGene Data A recent genome-wide association study (Sakornsakolpat et al., 2019) identified 35 new COPD-related genes from the UK Biobank and International COPD Genetics Consortium data. Among these 35 COPD-related genes, 28 genes are available in the data from the COPDGene study for the analysis using PRANA and other DN analysis methods including dnapath and DINGO. The 28 COPD-related genes are the following: CITED2, TESK2, COL15A1, AMZ1, RASEF, DDX1, DMWD, MED13L, ZBTB38, CCDC69, EML4, HSPA4, ITGB8, TEPP, TNPO1, ARNTL, DTWD1, ADAMTSL3, RREB1, THRA, SLMAP, DENND2D, STN1, SYN3, ASAP2, IER3, MFHAS1, and VGLL4. Among 2,561 samples from the initial phenotype data, we have used 406 samples that were provided as the validation set in the analysis of the original study. For the analysis with PRANA, binary current smoking status variable is used as the grouping variable, and smoking pack years, age, gender, race, and FEV1 are included as additional covariates in a multivariable model. The binary current smoking status variable is used as the grouping variable for dnapath and DINGO. ## 4 Results ### 4.1 Simulation Study More details of the simulation setup are available in the “Materials” section above. We select $p=20,50,100$ genes to test our pseudo-value approach in smaller to larger gene networks. For each gene network, five different sample sizes $n=40,100,200,500,1000$ are considered and in each setting, 1,000 Monte Carlo replicates. We draw 1,000 random samples first, then take the subsamples from this pool for a simulation with a smaller sample size to reduce computational burden. A random network is generated at each simulation replicate in which a layer of randomness is imposed to account for biological variability of the network structure. For additional details on the generation of simulated RNA-Seq data, see the Materials section. Simulations are repeated to show the performance of our method by altering the effect size from 5$\%$, 10$\%$, to 20$\%$ for simulation scenarios I and II. Results are compared with the true parent network in order to compute the performance measures described in the “Performance Measures” section. In the true network setting, a gene is considered truly DC between groups if it has at least one DC edge connected to other genes. Tables 1 and 2 summarize simulation results in the multivariable setting for scenarios I and II, respectively, when the continuous variable is added as a covariate with the binary group variable in the regression. Table 2 incorporates the effect of covariate when generating random networks, whereas Table 1 does not. For both scenarios, results show that the pseudo-value regression method generally yields a high precision and recall across all specifications of network size, sample size, and effect size. The pseudo-value regression method maintains a high precision while having an acceptable recall, especially, when a smaller sample size is considered. Table 1: Scenario I simulation results of binary group variable in the multivariable robust regression model (continuous age and binary group) using pseudo-value approach with 1,000 replicates. Random network is generated at each simulation replicate. | | Effect size ---|---|--- | | 5$\%$ | 10$\%$ | 20$\%$ $p$ | $n$ | Precision | Recall | F1 | Precision | Recall | F1 | Precision | Recall | F1 20 | 40 | 0.76 | 0.81 | 0.77 | 0.90 | 0.79 | 0.83 | 0.98 | 0.79 | 0.87 | 100 | 0.75 | 0.91 | 0.82 | 0.90 | 0.91 | 0.90 | 0.98 | 0.91 | 0.94 | 200 | 0.74 | 0.94 | 0.82 | 0.89 | 0.94 | 0.91 | 0.97 | 0.94 | 0.96 | 500 | 0.73 | 0.97 | 0.83 | 0.88 | 0.97 | 0.92 | 0.97 | 0.97 | 0.97 | 1,000 | 0.73 | 0.98 | 0.83 | 0.88 | 0.98 | 0.92 | 0.97 | 0.98 | 0.97 50 | 40 | 0.95 | 0.65 | 0.77 | 0.98 | 0.65 | 0.78 | 1.00 | 0.64 | 0.77 | 100 | 0.95 | 0.77 | 0.85 | 0.98 | 0.77 | 0.86 | 1.00 | 0.77 | 0.86 | 200 | 0.96 | 0.85 | 0.90 | 0.99 | 0.85 | 0.91 | 1.00 | 0.85 | 0.91 | 500 | 0.95 | 0.93 | 0.94 | 0.99 | 0.92 | 0.95 | 1.00 | 0.92 | 0.96 | 1,000 | 0.95 | 0.96 | 0.95 | 0.98 | 0.96 | 0.97 | 1.00 | 0.96 | 0.98 100 | 40 | 0.96 | 0.57 | 0.71 | 0.98 | 0.57 | 0.72 | 1.00 | 0.57 | 0.72 | 100 | 0.96 | 0.67 | 0.79 | 0.99 | 0.68 | 0.80 | 1.00 | 0.67 | 0.80 | 200 | 0.96 | 0.74 | 0.83 | 0.99 | 0.74 | 0.84 | 1.00 | 0.74 | 0.85 | 500 | 0.97 | 0.82 | 0.89 | 0.99 | 0.81 | 0.89 | 1.00 | 0.81 | 0.89 | 1,000 | 0.97 | 0.90 | 0.93 | 0.99 | 0.89 | 0.94 | 1.00 | 0.89 | 0.94 Table 2: Scenario II simulation results of binary group variable in the multivariable robust regression model (continuous age and binary group) using pseudo-value approach with 1,000 replicates. Random network is generated at each simulation replicate. | | Effect size ---|---|--- | | 5$\%$ | 10$\%$ | 20$\%$ $p$ | $n$ | Precision | Recall | F1 | Precision | Recall | F1 | Precision | Recall | F1 20 | 40 | 0.75 | 0.59 | 0.64 | 0.90 | 0.69 | 0.70 | 0.98 | 0.61 | 0.73 | 100 | 0.78 | 0.70 | 0.71 | 0.91 | 0.72 | 0.79 | 0.98 | 0.74 | 0.83 | 200 | 0.78 | 0.81 | 0.78 | 0.91 | 0.83 | 0.86 | 0.98 | 0.85 | 0.91 | 500 | 0.76 | 0.90 | 0.82 | 0.90 | 0.92 | 0.91 | 0.98 | 0.94 | 0.95 | 1,000 | 0.75 | 0.94 | 0.83 | 0.89 | 0.95 | 0.92 | 0.97 | 0.96 | 0.97 50 | 40 | 0.95 | 0.56 | 0.70 | 0.98 | 0.56 | 0.71 | 1.00 | 0.57 | 0.72 | 100 | 0.95 | 0.65 | 0.77 | 0.98 | 0.66 | 0.78 | 1.00 | 0.68 | 0.80 | 200 | 0.96 | 0.75 | 0.84 | 0.99 | 0.75 | 0.85 | 1.00 | 0.77 | 0.87 | 500 | 0.96 | 0.88 | 0.92 | 0.99 | 0.87 | 0.92 | 1.00 | 0.89 | 0.94 | 1,000 | 0.96 | 0.93 | 0.94 | 0.99 | 0.93 | 0.96 | 1.00 | 0.94 | 0.97 100 | 40 | 0.96 | 0.55 | 0.69 | 0.99 | 0.55 | 0.70 | 0.99 | 0.55 | 0.70 | 100 | 0.95 | 0.63 | 0.75 | 0.98 | 0.63 | 0.76 | 0.98 | 0.63 | 0.76 | 200 | 0.96 | 0.68 | 0.79 | 0.99 | 0.68 | 0.81 | 0.99 | 0.68 | 0.80 | 500 | 0.97 | 0.77 | 0.86 | 0.99 | 0.77 | 0.86 | 0.99 | 0.77 | 0.86 | 1,000 | 0.97 | 0.87 | 0.92 | 0.99 | 0.86 | 0.92 | 0.99 | 0.86 | 0.92 Tables 3 and 4 summarize simulation results for scenarios I and II, respectively, when only the binary group variable is included in the model for pseudo-value calculation. Thus, age dependent networks are simulated for Table 4 but not for Table 3. Two competing univariable methods, dnapath and DINGO are included for these simulations. Overall, a similar pattern is observed in the univariable setting; i.e., PRANA consistently reaches a high precision and recall. The performance improves as n increases, as to be expected. It is noteworthy that PRANA outperforms dnapath in simulation when the sample size is relatively small regardless of the network size. Our method also shows a better recall value and F1-score than dnapath with small sample sizes ($\textit{n}=40,100$). As DINGO requires substantially large computational time, it was considered for the gene network with smaller sample sizes only. To be more specific, simulations with larger sample sizes ($n=500,1000$) are stopped after 20 days for DINGO from the University of Florida Research Computing Linux server, HiPerGator 3.0 with 10CPU cores and 10GB of RAM per node. See Table S1 in Additional File 1 for more details. Table 3: Scenario I simulation results of binary group variable in the univariable robust regression model using pseudo-value approach with 1,000 replicates. The network structure does not depend on age covariate. Random network is generated at each simulation replicate. Sample size n = (500, 1000) for gene size p = 100 were not included for DINGO due to heavy computational time. The best results are highlighted in boldface. | | Precision | Recall | F1 ---|---|---|---|--- $p$ | $n$ | PRANA | dnapath | DINGO | PRANA | dnapath | DINGO | PRANA | dnapath | DINGO 20 | 40 | 0.90 | 0.95 | 0.87 | 0.81 | 0.64 | 0.78 | 0.84 | 0.76 | 0.82 | 100 | 0.90 | 0.93 | 0.87 | 0.90 | 0.88 | 0.79 | 0.84 | 0.90 | 0.82 | 200 | 0.89 | 0.91 | 0.87 | 0.94 | 0.94 | 0.76 | 0.91 | 0.92 | 0.81 | 500 | 0.88 | 0.89 | - | 0.97 | 0.98 | - | 0.92 | 0.93 | - | 1,000 | 0.88 | 0.89 | - | 0.98 | 0.99 | - | 0.92 | 0.93 | - 50 | 40 | 0.98 | 1.00 | 0.99 | 0.65 | 0.39 | 0.70 | 0.78 | 0.56 | 0.81 | 100 | 0.98 | 1.00 | 0.98 | 0.77 | 0.61 | 0.84 | 0.86 | 0.75 | 0.90 | 200 | 0.99 | 1.00 | 0.98 | 0.85 | 0.85 | 0.85 | 0.91 | 0.91 | 0.91 | 500 | 0.99 | 0.99 | - | 0.92 | 0.95 | - | 0.95 | 0.97 | - | 1,000 | 0.98 | 0.99 | - | 0.96 | 0.98 | - | 0.97 | 0.98 | - 100 | 40 | 0.98 | 1.00 | 0.98 | 0.57 | 0.27 | 0.69 | 0.72 | 0.42 | 0.85 | 100 | 0.98 | 1.00 | 0.99 | 0.68 | 0.32 | 0.75 | 0.80 | 0.48 | 0.85 | 200 | 0.99 | 1.00 | 0.98 | 0.74 | 0.62 | 0.82 | 0.84 | 0.77 | 0.89 | 500 | 0.99 | 1.00 | - | 0.81 | 0.88 | - | 0.89 | 0.93 | - | 1,000 | 0.99 | 0.99 | - | 0.89 | 0.94 | - | 0.94 | 0.97 | - Table 4: Scenario II simulation results of binary group variable in the univariable robust regression model using pseudo-value approach with 1,000 replicates. The network structure depends on age covariate. Random network is generated at each simulation replicate. Sample size n = (500, 1000) or gene size p = 100 were not included for DINGO due to heavy computational time. The best results are highlighted in boldface. | | Precision | Recall | F1 ---|---|---|---|--- $p$ | $n$ | PRANA | dnapath | DINGO | PRANA | dnapath | DINGO | PRANA | dnapath | DINGO 20 | 40 | 0.90 | 0.97 | 0.89 | 0.59 | 0.38 | 0.63 | 0.69 | 0.53 | 0.72 | 100 | 0.91 | 0.97 | 0.88 | 0.72 | 0.66 | 0.75 | 0.79 | 0.77 | 0.80 | 200 | 0.91 | 0.96 | 0.89 | 0.83 | 0.83 | 0.73 | 0.86 | 0.88 | 0.79 | 500 | 0.90 | 0.93 | - | 0.93 | 0.94 | - | 0.91 | 0.93 | - | 1,000 | 0.89 | 0.91 | - | 0.95 | 0.97 | - | 0.92 | 0.94 | - 50 | 40 | 0.98 | 1.00 | 0.99 | 0.56 | 0.28 | 0.67 | 0.71 | 0.43 | 0.80 | 100 | 0.98 | 1.00 | 0.98 | 0.65 | 0.45 | 0.71 | 0.78 | 0.61 | 0.82 | 200 | 0.99 | 1.00 | 0.98 | 0.75 | 0.68 | 0.80 | 0.85 | 0.80 | 0.88 | 500 | 0.99 | 0.99 | - | 0.87 | 0.91 | - | 0.93 | 0.95 | - | 1,000 | 0.99 | 0.99 | - | 0.94 | 0.97 | - | 0.96 | 0.98 | - 100 | 40 | 0.99 | 1.00 | 0.55 | 0.55 | 0.21 | 0.77 | 0.70 | 0.34 | 0.63 | 100 | 0.98 | 1.00 | 0.55 | 0.63 | 0.27 | 0.75 | 0.76 | 0.42 | 0.63 | 200 | 0.99 | 1.00 | 0.55 | 0.68 | 0.46 | 0.77 | 0.80 | 0.62 | 0.63 | 500 | 0.99 | 1.00 | - | 0.77 | 0.82 | - | 0.86 | 0.90 | - | 1,000 | 0.99 | 1.00 | - | 0.86 | 0.92 | - | 0.92 | 0.96 | - Table 5 presents results of scenario III, where age acts as a confounder. That is, an observed difference in connectivity may be due to a difference in the distribution of age in the two groups. Higher precision values from the multivariable pseudo-value regression indicate that PRANA correctly identifies the DC genes, compared with dnapath and DINGO, neither of which accounts for the effects of age. By and large, PRANA has higher precision than DINGO and higher recall than dnapath. Table 5: Scenario III simulation results of binary group variable in the multivariable and univariable robust regression model using pseudo-value approach with 1,000 replicates. The network structure depends on age covariate by unequal sampling proportion depending on age categories. Random network is generated at each simulation replicate. Sample size n = (500, 1000) or gene size p = 100 were not included for DINGO due to heavy computational costs. The best results are highlighted in boldface. | | Precision | Recall | F1 ---|---|---|---|--- $p$ | $n$ | PRANA (Mult) | PRANA (Univ) | dnapath | DINGO | PRANA (Mult) | PRANA (Univ) | dnapath | DINGO | PRANA (Mult) | PRANA (Univ) | dnapath | DINGO 20 | 40 | 0.67 | 0.60 | 0.64 | 0.58 | 0.57 | 0.74 | 0.50 | 0.75 | 0.59 | 0.65 | 0.55 | 0.64 | 100 | 0.67 | 0.58 | 0.61 | 0.58 | 0.65 | 0.85 | 0.73 | 0.79 | 0.64 | 0.68 | 0.65 | 0.66 | 200 | 0.66 | 0.58 | 0.59 | 0.58 | 0.76 | 0.91 | 0.85 | 0.80 | 0.69 | 0.70 | 0.69 | 0.67 | 500 | 0.64 | 0.57 | 0.58 | - | 0.87 | 0.95 | 0.95 | - | 0.73 | 0.71 | 0.71 | - | 1,000 | 0.64 | 0.57 | 0.58 | - | 0.92 | 0.97 | 0.97 | - | 0.75 | 0.71 | 0.72 | - 50 | 40 | 0.57 | 0.50 | 0.54 | 0.49 | 0.47 | 0.62 | 0.33 | 0.67 | 0.50 | 0.54 | 0.40 | 0.55 | 100 | 0.58 | 0.49 | 0.52 | 0.49 | 0.50 | 0.71 | 0.51 | 0.79 | 0.52 | 0.58 | 0.51 | 0.60 | 200 | 0.58 | 0.49 | 0.51 | 0.49 | 0.52 | 0.76 | 0.60 | 0.83 | 0.53 | 0.59 | 0.54 | 0.61 | 500 | 0.55 | 0.48 | 0.48 | - | 0.67 | 0.88 | 0.83 | - | 0.59 | 0.61 | 0.60 | - | 1,000 | 0.54 | 0.48 | 0.48 | - | 0.81 | 0.93 | 0.91 | - | 0.64 | 0.62 | 0.62 | - 100 | 40 | 0.57 | 0.49 | 0.53 | 0.48 | 0.44 | 0.57 | 0.22 | 0.75 | 0.49 | 0.52 | 0.30 | 0.58 | 100 | 0.58 | 0.49 | 0.53 | 0.48 | 0.47 | 0.64 | 0.35 | 0.76 | 0.51 | 0.55 | 0.41 | 0.58 | 200 | 0.58 | 0.49 | 0.53 | 0.48 | 0.46 | 0.65 | 0.40 | 0.80 | 0.50 | 0.55 | 0.45 | 0.60 | 500 | 0.56 | 0.47 | 0.49 | - | 0.44 | 0.74 | 0.60 | - | 0.48 | 0.57 | 0.53 | - | 1,000 | 0.53 | 0.47 | 0.47 | - | 0.63 | 0.85 | 0.82 | - | 0.57 | 0.60 | 0.59 | - Table S1: Comparison of computational time of PRANA with that of dnapath and DINGO. Scenario I is depicted for the illustrative purposes. Random network is generated at each simulation replicate. Large sample sizes n = (500, 1000) are stopped after 20 days for DINGO from the high-performance Linux cluster using machines with 10CPU cores and 10GB of RAM per node. | | Time (in hours) ---|---|--- $p$ | $n$ | PRANA | dnapath | DINGO 20 | 40 | 0.12 | 0.13 | 29.25 | 100 | 0.22 | 0.2 | 86.58 | 200 | 0.23 | 0.17 | 206.18 | 500 | 0.42 | 0.27 | $>$ 480 | 1,000 | 0.68 | 0.45 | $>$ 480 50 | 40 | 0.2 | 0.22 | 26.57 | 100 | 0.35 | 0.32 | 31.48 | 200 | 0.67 | 0.45 | 137.58 | 500 | 1.52 | 0.98 | $>$ 480 | 1,000 | 2.48 | 1.08 | $>$ 480 100 | 40 | 0.63 | 0.52 | 68.73 | 100 | 0.82 | 0.7 | 61.78 | 200 | 1.1 | 1.23 | 147.57 | 500 | 3.58 | 2.37 | $>$ 480 | 1,000 | 4.68 | 3.57 | $>$ 480 ### 4.2 Analysis of COPDGene Data 23 out of 28 COPD-related genes are predicted to be DC between current and non-current smokers with PRANA while accounting for smoking pack years, age, gender, race, and FEV1. A complete list of DC genes found from the pseudo- value approach are CITED2, TESK2, AMZ1, DDX1, DMWD, MED13L, ZBTB38, EML4, HSPA4, ITGB8, TEPP, TNPO1, ARNTL, DTWD1, ADAMTSL3, THRA, SLMAP, DENND2D, STN1, SYN3, ASAP2, IER3, and MFHAS1. We compared results of PRANA with dnapath (Grimes et al., 2019) and DINGO (Ha et al., 2015). With DINGO, a total of 19 out of 28 COPD-related genes were selected as DC genes between current and non-current smokers. A complete list of DC genes found in DINGO are the following: ARNTL, DDX1, HSPA4, ITGB8, SLMAP, SYN3, ASAP2, IER3, MFHAS1, VGLL4, CITED2, TESK2, CCDC69, EML4, ADAMTSL3, DENND2D, AMZ1, RASEF, and ZBTB38. Lastly, 3 genes were found DC between current smoking groups with dnapath, namely DTWD1, EML4, and TEPP. Of the 23 DC genes from PRANA, 5 are found exclusive to PRANA (DMWD, MED13L, TNPO1, THRA, and STN1). Notably, DMWD is linked to myotonic dystrophy, a rare genetic muscular disorder (Westerlaken et al., 2003). Thyroid hormone receptor alpha (THRA) is related to congenital hypothyroidism (Tylki-Szymańska et al., 2015). These findings about additional genes will facilitate harnessing of the possible mechanisms at work in COPD exacerbation. Heat shock protein family A (Hsp70) member 4 (HSPA4) is associated with gastric ulcer (Sakurai et al., 2015). Multifunctional ROCO family signaling regulator 1 (MFHAS1) is linked to soft tissue tumor and cell cycle (Stelzer et al., 2016). HSPA4 and MFHAS1 are DC genes identified in both PRANA and DINGO. Echinoderm microtubule-associated protein-like 4 (EML4) is found in all three methods. It has been studied for its association with lung cancer (Stelzer et al., 2016; Adib et al., 2019). A Venn diagram is provided to show the overlap between and among three methods (Figure 4). In addition, a diagram is included to summarize the findings of this application study (Figure 5). Figure 4: A Venn diagram displaying the number of overlapping DC genes between and among univariable analysis such as DINGO and dnapath versus multivariable robust regression with pseudo-value approach using COPDGene study data from GEO database. Figure 5: A diagram summarizing results using each methods analyzing the COPDGene study data from GEO database. A full list of DCGs (differentially connected genes) are provided in each box. ## 5 Discussion Simulations and real-data analysis have elucidated that PRANA is superior to existing alternatives and a practical tool, which includes covariates in the model. To the best of our knowledge, this is the first attempt to develop a regression modeling in DN analysis. Our working objective is to propose a statistical method that determines whether a gene is significantly DC between groups with the covariate included in the model. In this paper, we have shown through simulations that PRANA reaches a consistently high degree of precision and recall to identify DC genes with varying simulation parameters such as network size, sample size, and effect size. We also analyzed a COPD-related gene expression data from the GEO database. When comparing results from our method to dnapath and DINGO, five COPD-related genes are additionally found DC between current versus non-current smokers: DMWD, MED13L, TNPO1, THRA, and STN1. There are a number of limitations to be highlighted in this study. We have used the absolute value of the differences between the two adjacency matrices as a proxy to determine the true DC genes. Certainly, this is a practical way to detect differences in the number of edges for each genes in a network. The comparison of maximum values between adjacency matrices was also considered. However, we concluded that they are more useful describing the global characteristic of a network, which deviates from our objective, namely, a gene-specific characteristic of a network. Another limitation is the inability to perturb simulated networks in a continuous way. Right now, we have discretized the effect of a covariate into three groups. Perhaps, there are other models where a truly continuous covariate could be incorporated. Lastly, the Pearson correlation, partial correlation, and degree-weighted LASSO were also examined as alternatives to the ARACNE as a measure of association or connectedness, albeit not reported in the paper, due to relatively poor performance and heavy computational costs. It remains an interesting task for future studies to extend our work to other measures of association of a network which better assess different structural changes in the network. We conclude by foregrounding the future direction of the pseudo- value regression approach for the DN analysis, which are potentially extensible to other data types, such as the microbiome data. ## 6 Conclusion The adjustment of covariate is an important step in differential network analysis. In this paper, we presented PRANA, a novel pseudo-value regression approach for the DN analysis, which can incorporate additional clinical covariates in the model. This is a direct regression modeling, and it is therefore computationally amenable for the most users. ## References * Adib et al. (2019) Adib, R., Montgomery, J., Atherton, J., O’Regan, L., Richards, M., Straatman, K., et al. (2019). Mitotic phosphorylation by NEK6 and NEK7 reduces the microtubule affinity of EML4 to promote chromosome congression. Science Signaling 12, 594\. * Ahn and Logan (2016) Ahn, K. and Logan, B. (2016). Pseudo-value approach for conditional quantile residual lifetime analysis for clustered survival and competing risks data with applications to bone marrow transplant data. Annals of Applied Statistics 10, 618–637. * Ahn and Mendolia (2014) Ahn, K. and Mendolia, F. (2014). Pseudo-value approach for comparing survival medians for dependent data. Statistics in Medicine 33, 1531–1538. * Andersen and Klein (2007) Andersen, P. and Klein, J. (2007). Regression analysis for multistate models based on a pseudo-value approach, with applications to bone marrow transplantation studies. Scandinavian Journal of Statistics 34, 3–16. * Andersen et al. (2003) Andersen, P., Klein, J., and Rosthøj, S. (2003). Generalised linear models for correlated pseudo-observations, with applications to multi-state models. Biometrika 90, 15–27. * Barrett et al. (2013) Barrett, T., Wilhite, S., Ledoux, P., Evangelista, C., Kim, I., Tomashevsky, M., et al. (2013). NCBI GEO: archive for functional genomics data sets–update. Nucleic Acids Research 41, D991–D995. * Basso et al. (2005) Basso, K., Margolin, A., Stolovitzky, G., Klein, U., Dalla-Favera, R., and Califano, A. (2005). Reverse engineering of regulatory networks in human b cells. Nature Genetics 37, 382–390. * Bruha et al. (2012) Bruha, R., Dvorak, K., and Petrtyl, J. (2012). Alcoholic liver disease. World Journal of Hepatology 4, 81–90. * Datta and Datta (2005) Datta, S. and Datta, S. (2005). Empirical Bayes screening of many p-values with applications to microarray studies. Bioinformatics 21, 1987–1994. * de la Fuente (2010) de la Fuente, A. (2010). From ‘differential expression’ to ‘differential networking’ - identification of dysfunctional regulatory networks in diseases. Trends in Genetics 26, 326–333. * Dutta et al. (2017) Dutta, S., Datta, S., and Datta, S. (2017). Temporal prediction of future state occupation in a multistate model from high-dimensional baseline covariates via pseudo-value regression. Journal of Statistical Computation and Simulation 87, 1363–1378. * Efron and Tibshirani (1993) Efron, B. and Tibshirani, R. (1993). An Introduction to the Bootstrap. Chapman & Hall, New York. * Gao et al. (2017) Gao, M., Zhong, A., Patel, N., Alur, C., and Vyas, D. (2017). High throughput RNA sequencing utility for diagnosis and prognosis in colon diseases. World Journal of Gastroenterology 23, 2819–2825. * Gill et al. (2010) Gill, R., Datta, S., and Datta, S. (2010). A statistical framework for differential network analysis from microarray data. BMC Bioinformatics 11, 95\. * Graw et al. (2009) Graw, F., Gerds, T., and Schumacher, M. (2009). On pseudo-values for regression analysis in competing risks models. Lifetime Data Analysis 15, 241–255. * Grimes and Datta (2021) Grimes, T. and Datta, S. (2021). SeqNet: An R package for generating gene-gene networks and simulating RNA-seq data. Journal of Statistical Software 98, 21\. * Grimes et al. (2019) Grimes, T., Potter, S., and Datta, S. (2019). Integrating gene regulatory pathways into differential network analysis of gene expression data. Scientific Reports 9, 1–12. * Ha et al. (2015) Ha, M., Baladandayuthapani, V., and Do, K. (2015). DINGO: differential network analysis in genomics. Bioinformatics 31, 3413–3420. * Herreros-Villanueva et al. (2013) Herreros-Villanueva, M., Hijona, E., Bañales, J., Cosme, A., and Bujanda, L. (2013). Alcohol consumption on pancreatic diseases. World Journal of Gastroenterology 19, 638–647. * Ideker and Krogan (2012) Ideker, T. and Krogan, N. (2012). Differential network biology. Molecular Systems Biology 8, 565\. * Kim et al. (2018) Kim, Y., Hao, J., Gautam, Y., Mersha, T., and Kang, M. (2018). DiffGRN: differential gene regulatory network analysis. International Journal of Data Mining and Bioinformatics 20, 362–379. * Li et al. (2012) Li, J., Witten, D., Johnstone, I., and Tibshirani, R. (2012). Normalization, testing, and false discovery rate estimation for RNA-sequencing data. Biostatistics 13, 523–538. * Marballi et al. (2016) Marballi, K., Genabai, N., Blednov, Y., Harris, R., and Ponomarev, I. (2016). Alcohol consumption induces global gene expression changes in VTA dopaminergic neurons. Genes, Brain and Behavior 15, 318–326. * Margolin et al. (2006) Margolin, A., Nemenman, I., Basso, K., Wiggins, C., Stolovitzky, G., Dalla-Favera, R., et al. (2006). ARACNE: an algorithm for the reconstruction of gene regulatory networks in a mammalian cellular context. BMC Bioinformatics 7, S7. * McKenzie et al. (2016) McKenzie, A., Katsyv, I., Song, W., Wang, M., and Zhang, B. (2016). DGCA: a comprehensive R package for differential gene correlation analysis. BMC Systems Biology 10, 106\. * Meyer et al. (2008) Meyer, P., Lafitte, F., and Bontempi, G. (2008). minet: A R/Bioconductor package for inferring large transcriptional networks using mutual information. BMC Bioinformatics 9, 461\. * Pison et al. (2002) Pison, G., Van Aelst, S., and Willems, G. (2002). Small sample corrections for lts and mcd. Metrika 55, 111–123. * Reuter et al. (2015) Reuter, J., Spacek, D., and Snyder, M. (2015). High-throughput sequencing technologies. Molecular Cell 58, 586–597. * Ritchie et al. (2015) Ritchie, M., Phipson, B., Wu, D., Hu, Y., Law, C., Shi, W., et al. (2015). limma powers differential expression analyses for RNA-sequencing and microarray studies. Nucleic Acids Research 43, 7\. * Robinson et al. (2010) Robinson, M., McCarthy, D., and Smyth, G. (2010). edgeR: a Bioconductor package for differential expression analysis of digital gene expression data. Bioinformatics 26, 139–140. * Rousseeuw (1984) Rousseeuw, P. (1984). Least median of squares regression. Journal of the American Statistical Association 79, 871–880. * Sakornsakolpat et al. (2019) Sakornsakolpat, P., Prokopenko, D., Lamontagne, M., Reeve, N., Guyatt, A., Jackson, V., et al. (2019). Genetic landscape of chronic obstructive pulmonary disease identifies heterogeneous cell-type and phenotype associations. Nature Genetics 51, 494–505. * Sakurai et al. (2015) Sakurai, T., Kashida, H., Hagiwara, S., Nishida, N., Watanabe, T., Fujita, J., et al. (2015). Heat shock protein A4 controls cell migration and gastric ulcer healing. Digestive Diseases and Sciences 60, 850–857. * Stelzer et al. (2016) Stelzer, G., Rosen, N., Plaschkes, I., Zimmerman, S., Twik, M., Fishilevich, S., et al. (2016). The GeneCards suite: from gene data mining to disease genome sequence analyses. Digestive Diseases and Sciences 54, 1.30.1–1.30.33. * Todorov and Filzmoser (2009) Todorov, V. and Filzmoser, P. (2009). An object-oriented framework for robust multivariate analysis. Journal of Statistical Software 32, 1–47. * Tylki-Szymańska et al. (2015) Tylki-Szymańska, A., Acuna-Hidalgo, R., Krajewska-Walasek, M., Lecka-Ambroziak, A., Steehouwer, M., Gilissen, C., et al. (2015). Thyroid hormone resistance syndrome due to mutations in the thyroid hormone receptor $\alpha$ gene (THRA). Journal of Medical Genetics 52, 312–316. * Wang et al. (2009) Wang, Z., Gerstein, M., and Snyder, M. (2009). RNA-Seq: a revolutionary tool for transcriptomics. Nature Reviews Genetics 10, 57–63. * Wang et al. (2021) Wang, Z., Masoomi, A., Xu, Z., Boueiz, A., Lee, S., Zhao, T., et al. (2021). Improved prediction of smoking status via isoform-aware RNA-seq deep learning models. PLoS Computational Biology 17, 10\. * Westerlaken et al. (2003) Westerlaken, J., Van der Zee, C., Peters, W., and Wieringa, B. (2003). The DMWD protein from the myotonic dystrophy (DM1) gene region is developmentally regulated and is present most prominently in synapse-dense brain areas. Brain Research 971, 116–127. * Westfall and Young (1993) Westfall, P. and Young, S. (1993). Resampling-based Multiple Testing: Examples and Methods for p-value Adjustment. Wiley-Interscience, New York. * Zhao et al. (2014) Zhao, S., Fung-Leung, W., Bittner, A., Ngo, K., and Liu, X. (2014). Comparison of RNA-Seq and microarray in transcriptome profiling of activated t cells. PLoS One 9, e78644. * Zoppoli et al. (2010) Zoppoli, P., Morganella, S., and Ceccarelli, M. (2010). TimeDelay-ARACNE: Reverse engineering of gene networks from time-course data by an information theoretic approach. BMC Bioinformatics 11, 154\.
# Approximate minimum cuts and their enumeration111University of Illinois, Urbana-Champaign. Email: {calvinb2, karthe<EMAIL_ADDRESS>Supported in part by NSF grants CCF-1814613 and CCF-1907937. Calvin Beideman Karthekeyan Chandrasekaran Weihang Wang ###### Abstract We show that every $\alpha$-approximate minimum cut in a connected graph is the _unique_ minimum $(S,T)$-terminal cut for some subsets $S$ and $T$ of vertices each of size at most $\lfloor 2\alpha\rfloor+1$. This leads to an alternative proof that the number of $\alpha$-approximate minimum cuts in a $n$-vertex connected graph is $n^{O(\alpha)}$ and they can all be enumerated in deterministic polynomial time for constant $\alpha$. ## 1 Introduction Let $G=(V,E)$ be a graph with positive edge costs $c:E\rightarrow\mathbb{R}_{+}$. A partitioning of the vertex set into $2$ non- empty parts is known as a _cut_. For a non-empty proper subset $U$ of vertices, we will use $\overline{U}$ to denote $V\setminus U$, $(U,\overline{U})$ to denote the cut associated with $U$, $\delta(U)$ to denote the set of edges crossing the cut $(U,\overline{U})$, and $d(U):=\sum_{e\in\delta(U)}c(e)$ to denote the cut value of $U$. We will use $\lambda$ to denote the value of a minimum cut in $G$, i.e., $\lambda:=\min\\{d(U):\emptyset\neq U\subsetneq V\\}.$ For $\alpha\geq 1$, a cut $(U,\overline{U})$ is an _$\alpha$ -approximate minimum cut_ if $d(U)\leq\alpha\lambda$. A fundamental graph structural result states that the number of $\alpha$-approximate minimum cuts in a $n$-vertex connected graph is $O(n^{2\alpha})$. This structural result forms the backbone of several algorithmic and representation results in graphs—e.g., fast randomized construction of skeletons [12], existence of cut sparsifiers which in turn finds applications in streaming and sketching [1, 2, 3, 14], approximation algorithms for TSP [17, 4, 6, 5], computing reliability of probabilistic networks [11], and polynomial time algorithms for connectivity augmentation [16]. The structural result along with a randomized polynomial- time algorithm to enumerate all constant-approximate minimum cuts was first shown via Karger’s random contraction technique [13]. Subsequently, the splitting-off technique and the tree packing technique have been used to show the structural result along with a _deterministic_ polynomial-time algorithm to enumerate all constant-approximate minimum cuts [15, 8]. In this work, we give a fourth technique to bound the number of $\alpha$-approximate minimum cuts by $n^{O(\alpha)}$ along with a deterministic polynomial-time algorithm to enumerate all constant-approximate minimum cuts. Let $S$, $T$ be disjoint non-empty subsets of vertices. A $2$-partition $(U,\overline{U})$ is an $(S,T)$-terminal cut if $S\subseteq U\subseteq V\setminus T$. Here, the set $U$ is known as the _source set_ and the set $\overline{U}$ is known as the _sink set_. An $(S,T)$-terminal cut with minimum cut value will be denoted as a _minimum $(S,T)$-terminal cut_. The following is our main result. ###### Theorem 1.1. Let $G=(V,E)$ be a connected graph with positive edge costs and $(U,\overline{U})$ be an $\alpha$-approximate minimum cut for some $\alpha\geq 1$. Then, there exist subsets $S,T\subseteq V$ with $|S|,|T|\leq\lfloor 2\alpha\rfloor+1$ such that $(U,\overline{U})$ is the unique minimum $(S,T)$-terminal cut. In other words, every $\alpha$-approximate minimum cut $(U,\overline{U})$ in a connected graph can be recovered as the _unique_ minimum $(S,T)$-terminal cut for some subsets $S$ and $T$ of sizes at most $\lfloor 2\alpha\rfloor+1$. A few remarks are in order. ###### Remark 1.1. An immediate consequence of Theorem 1.1 is that the number of distinct $\alpha$-approximate minimum cuts in a $n$-vertex connected graph is at most $n^{4\alpha+2}$ and they can all be enumerated using $n^{4\alpha+2}$ minimum $(S,T)$-terminal cut computations. We recall that minimum $(S,T)$-terminal cut in a given graph with given subsets $S$ and $T$ of vertices can be computed in deterministic polynomial time. ###### Remark 1.2. The size bound for subsets $S$ and $T$ in Theorem 1.1 is tight up to an additive factor of one: Consider the $n$-vertex cycle graph $G$ with $\alpha$ being a positive integer where $n\geq 4\alpha$ and with all edge costs being $1$. Let $(U,\overline{U})$ be a cut in $G$ with $d(U)=2\alpha$ and moreover, each connected component in $G-\delta(U)$ has at least $2$ vertices (see Figure 1 for such an example where $\alpha=4$). For such a cut $(U,\overline{U})$, the only choice of $S$ and $T$ for which $(U,\overline{U})$ is the _unique_ minimum $(S,T)$-terminal cut is given by $\displaystyle S$ $\displaystyle:=\\{v\in U:v\text{ is an end vertex of an edge in }\delta(U)\\}\text{ and}$ $\displaystyle T$ $\displaystyle:=\\{v\in\overline{U}:v\text{ is an end vertex of an edge in }\delta(U)\\}.$ Hence, this example shows that both $S$ and $T$ need to have at least $2\alpha$ vertices in order to recover $(U,\overline{U})$ as the _unique_ minimum $(S,T)$-terminal cut. Figure 1: An example showing the tightness of Theorem 1.1: The set $U$ consists of vertices within all gray boxes and the set $\overline{U}$ consists of all vertices outside gray boxes; the dotted edges correspond to the edges in the cut set $\delta(U)$. The only choice of $S$ and $T$ for which $(U,\overline{U})$ is the unique minimum $(S,T)$-terminal cut is as follows: $S$ consists of all vertices that are shaded black and $T$ consists of all vertices that are shaded gray. Theorem 1.1 is inspired by a result of Goemans and Ramakrishnan [9]. They showed the following structural theorem for $(4/3-\epsilon)$-approximate minimum cuts: fix an arbitrary vertex $t$ in a connected graph $G=(V,E)$ and let $(U,\overline{U})$ be a $(4/3-\epsilon)$-approximate minimum cut with $t\in\overline{U}$ for some $\epsilon>0$. Then, there exists a subset $S$ of size at most $2$ such that $(U,\overline{U})$ is the unique minimum $(S,\\{t\\})$-terminal cut. Their result leads to the following natural question: for every $\alpha\geq 1$, can every $\alpha$-approximate minimum cut $(U,\overline{U})$ be obtained as a minimum $(S,T)$-terminal cut for some subsets $S$ and $T$ of vertices each of size at most some function of $\alpha$? Our Theorem 1.1 answers this question affirmatively. However, our proof of Theorem 1.1 deviates significantly from the proof approach of Goemans-Ramakrishnan’s structural result for $(4/3-\epsilon)$-approximate minimum cuts. Their proof proceeds via contradiction and relies on the _submodular triple inequality_ which holds for the graph cut function. The submodular triple inequality shows that $3$ non-empty sets satisfying certain crossing properties can be uncrossed to obtain $4$ disjoint non-empty sets without increasing the sum of their cut values; thus, if all $3$ sets have cut value less than $(4/3)\lambda$, then one of the $4$ sets should have cut value strictly smaller than $\lambda$, a contradiction to the definition of $\lambda$. However, there does not seem to be a generalization of the submodular triple inequality to larger number of sets. Our proof of Theorem 1.1 circumvents the submodular triple inequality but achieves its intended purpose in this context—we show that a sufficiently large number of $\alpha$-approximate minimum cuts with certain crossing properties can be uncrossed to obtain a cut of value cheaper than $\lambda$, thus leading to a contradiction again. ### 1.1 Preliminaries Let $G=(V,E)$ be a graph with positive edge costs $c:E\rightarrow\mathbb{R}_{+}$. For disjoint non-empty subsets $S,T\subseteq V$, there can be multiple minimum $(S,T)$-terminal cuts. We will be interested in _source minimal_ minimum $(S,T)$-terminal cuts. A minimum $(S,T)$-terminal cut $(U,\overline{U})$ is a source minimal minimum $(S,T)$-terminal cut if for all minimum $(S,T)$-terminal cut $(X,\overline{X})$, we have $U\subseteq X$. For every pair of disjoint non-empty subsets $S$ and $T$ of vertices, there exists a unique source minimal minimum $(S,T)$-terminal cut (this is due to submodularity—e.g., see [9]). We recall that the graph cut function $d:2^{V}\rightarrow\mathbb{R}$ is symmetric and submodular, i.e., $d(A)=d(V\setminus A)$ and $d(A)+d(B)\geq d(A\cap B)+d(A\cup B)$ for all $A,B\subseteq V$. We will need an uncrossing result that relies on more careful counting of edges than simply employing the submodularity inequality. We begin with some notation that will help in such careful counting. Let $(Y_{1},\ldots,Y_{p},W,Z)$ be a partition of $V$ into $p+2$ non-empty parts. We define $\displaystyle\sigma(Y_{1},\ldots,Y_{p},W,Z)$ $\displaystyle:=2\left(\sum_{uv\in E:\ u\in Y_{i},\ v\in Y_{j}\text{ for distinct }i,\ j\in[p]}c(uv)+\sum_{uv\in E:\ u\in W,\ v\in Z}c(uv)\right)$ $\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad+\sum_{uv\in E:\ u\in\cup_{i\in[p]}Y_{i},\ v\in W\cup Z}c(uv)$ We note that $\displaystyle\sum_{i\in[p]}d(Y_{i})$ $\displaystyle\leq\sigma(Y_{1},\ldots,Y_{p},W,Z).$ (1) The following lemma uncrosses a collection of $p$ sets to obtain a partition of the vertex set into $p+2$ non-empty parts with small $\sigma$-value. The lemma was shown by Kamidoi, Yoshida, and Nagamochi in the context of minimum $k$-cuts [10]. It was also extended to hypergraphs by Chandrasekaran and Chekuri [7]. See Figure 2 for an illustration of the sets that appear in the statement of the lemma. ###### Lemma 1.1. [10, 7] Let $G=(V,E)$ be a graph with positive edge costs and $\emptyset\neq R\subsetneq U\subsetneq V$. Let $S=\\{u_{1},\ldots,u_{p}\\}\subseteq U\setminus R$ for $p\geq 2$. For each $i\in[p]$, let $(\overline{A_{i}},A_{i})$ be a minimum $((S\cup R)\setminus\\{u_{i}\\},\overline{U})$-terminal cut. Suppose that $u_{i}\in A_{i}\setminus(\cup_{j\in[p]\setminus\\{i\\}}A_{j})$ for every $i\in[p]$. Let $Z:=\cap_{i=1}^{p}\overline{A_{i}},\ W:=\cup_{1\leq i<j\leq p}(A_{i}\cap A_{j}),\ \text{and}\ Y_{i}:=A_{i}-W\ \forall i\in[p].$ Then, $(Y_{1},\ldots,Y_{p},W,Z)$ is a partition of $V$ into $p+2$ non-empty parts with $\sigma(Y_{1},\ldots,Y_{p},W,Z)\leq\min\\{d(A_{i})+d(A_{j}):i,j\in[p],i\neq j\\}.$ Figure 2: Illustration of the sets that appear in the statement of Lemma 1.1. We briefly remark on the lemma: known proofs of the lemma are by induction on $p$ with an uncrossing argument that involves careful edge counting. The use of $\sigma(Y_{1},\ldots,Y_{p},W,Z)$ instead of $\sum_{i\in[p]}d(Y_{i})$ on the LHS is crucial for the induction argument—this is why we prefer to state the lemma as giving an upper bound on $\sigma(Y_{1},\ldots,Y_{p},W,Z)$ although we will use it only to conclude an upper bound on $\sum_{i\in[p]}d(Y_{i})$ via inequality (1). ## 2 Proof of Theorem 1.1 We will derive Theorem 1.1 from the following theorem. ###### Theorem 2.1. Let $G=(V,E)$ be a connected graph with positive edge costs and $(U,\overline{U})$ be an $\alpha$-approximate minimum cut for some $\alpha\geq 1$. Then, there exists a subset $S\subseteq U$ with $|S|\leq\lfloor 2\alpha\rfloor+1$ such that $(U,\overline{U})$ is the unique minimum $(S,\overline{U})$-terminal cut. ###### Proof. Consider the collection $\mathcal{C}:=\\{Q\subsetneq V:\overline{U}\subsetneq Q,d(Q)\leq d(U)\\}$. Suppose $\mathcal{C}$ is empty. Let $S=\\{x\\}$ for some arbitrary $x\in U$ and let $(X,\overline{X})$ be a minimum $(S,\overline{U})$-terminal cut. Then, $\overline{U}\subseteq\overline{X}\subsetneq V$ and $d(\overline{X})\leq d(U)$. Since $\overline{X}\not\in\mathcal{C}$ (because $\mathcal{C}$ is empty), we must have $\overline{X}=\overline{U}$. Therefore, $(U,\overline{U})$ is the unique minimum $(S,\overline{U})$-terminal cut. Next, suppose $\mathcal{C}\neq\emptyset$. Let $S\subseteq U$ be a minimal transversal of the collection $\mathcal{C}$, i.e., $S\subseteq U$ is a minimal set with $S\cap Q\neq\emptyset$ for all $Q\in\mathcal{C}$. Proposition 2.1 and Lemma 2.1 complete the proof of Theorem 2.1. ∎ ###### Proposition 2.1. $(U,\overline{U})$ is the unique minimum $(S,\overline{U})$-terminal cut. ###### Proof. Suppose $(X,\overline{X})$ be the source minimal minimum $(S,\overline{U})$-terminal cut with $X\neq U$. Since $(U,\overline{U})$ is a $(S,\overline{U})$-terminal cut, we have $d(\overline{X})=d(X)\leq d(U)$. Moreover, we have $\emptyset\neq X\subsetneq U$, which implies that $\overline{U}\subsetneq\overline{X}\subsetneq V$. Hence, we have $\overline{X}\in\mathcal{C}$, and thus $S\cap\overline{X}\neq\emptyset$. This contradicts with the assumption that $(X,\overline{X})$ is a $(S,\overline{U})$-terminal cut. Therefore, we must have $X=U$. ∎ ###### Lemma 2.1. $|S|\leq\lfloor 2\alpha\rfloor+1$. ###### Proof. For the sake of contradiction, suppose that $|S|\geq\lfloor 2\alpha\rfloor+2$. Let $\lambda$ denote the value of a minimum cut in $G$. Our proof strategy is to arrive at a cut with value cheaper than $\lambda$, thus contradicting the definition of $\lambda$. For convenience, we will denote $p:=|S|$ and write $S=\\{u_{1},\ldots,u_{p}\\}$. For each $i\in[p]$, let $(\overline{A_{i}},A_{i})$ be the source minimal minimum $(S-u_{i},\overline{U})$-terminal cut. ###### Claim 2.1. For every $i\in[p]$, we have $d(A_{i})\leq d(U)$ and $u_{i}\in A_{i}$. ###### Proof. Let $i\in[p]$. Since $S$ is a minimal transversal of the collection $\mathcal{C}$, there exists a set $B_{i}\in\mathcal{C}$ such that $B_{i}\cap S=\\{u_{i}\\}$. Hence, $(\overline{B_{i}},B_{i})$ is a $(S-u_{i},\overline{U})$-terminal cut. Therefore, $d(A_{i})\leq d(B_{i})\leq d(U).$ We will show that $A_{i}$ is in the collection $\mathcal{C}$. By definition, $A_{i}\subseteq V\setminus(S-u_{i})\subsetneq V$ and $\overline{U}\subseteq A_{i}$. If $A_{i}=\overline{U}$, then the above inequalities are equations (since $d(A_{i})=d(\overline{U})=d(U)$) implying that $(\overline{B_{i}},B_{i})$ is a minimum $(S-u_{i},\overline{U})$-terminal cut, and consequently, $(\overline{B_{i}},B_{i})$ contradicts source minimality of the minimum $(S-u_{i},\overline{U})$-terminal cut $(\overline{A_{i}},A_{i})$. Therefore, $\overline{U}\subsetneq A_{i}$. Hence, $A_{i}$ is in the collection $\mathcal{C}$. We recall that the set $S$ is a transversal for the collection $\mathcal{C}$ and moreover, none of the elements of $S-u_{i}$ are in $A_{i}$ by definition of $A_{i}$. Therefore, the vertex $u_{i}$ must be in $A_{i}$. ∎ We recall our assumption that $p\geq\lfloor 2\alpha\rfloor+2$. We note that $p-1>2\alpha\geq 2$. Using Claim 2.1, we observe that the sets $U$, $R:=\\{u_{p}\\}$, $S=\\{u_{1},\ldots,u_{p-1}\\}$, and partitions $(\overline{A_{i}},A_{i})$ for $i\in[p-1]$ satisfy the conditions of Lemma 1.1. Therefore, applying Lemma 1.1 shows the existence of a partition $(Y_{1},\ldots,Y_{p-1},W,Z)$ of $V$ into $p+1$ non-empty parts such that $\displaystyle\sigma(Y_{1},\ldots,Y_{p-1},W,Z)$ $\displaystyle\leq\min\\{d(A_{i})+d(A_{j}):i,j\in[p-1],i\neq j\\}\leq 2d(U)\leq 2\alpha\lambda.$ By inequality (1), this implies that $\displaystyle\sum_{i\in[p-1]}d(Y_{i})\leq 2\alpha\lambda.$ Therefore, there exists $i\in[p-1]$ such that $\displaystyle d(Y_{i})\leq\frac{2\alpha\lambda}{p-1}<\lambda.$ The last inequality is because $p-1>2\alpha$ and $\lambda>0$ since the graph is connected. Thus, we have a cut $(Y_{i},\overline{Y_{i}})$ with cut value smaller than $\lambda$, a contradiction to the definition of $\lambda$. ∎ Applying Theorem 2.1 to $(\overline{U},U)$ yields the following corollary. ###### Corollary 2.1. Let $G=(V,E)$ be a connected graph with positive edge costs and $(U,\overline{U})$ be an $\alpha$-approximate minimum cut for some $\alpha\geq 1$. Then, there exists a subset $T\subseteq\overline{U}$ with $|T|\leq\lfloor 2\alpha\rfloor+1$ such that $(U,\overline{U})$ is the unique minimum $(U,T)$-terminal cut. We now restate Theorem 1.1 and prove it using Theorem 2.1 and Corollary 2.1. See 1.1 ###### Proof. By Theorem 2.1, there exists a subset $S\subseteq U$ with $|S|\leq\lfloor 2\alpha\rfloor+1$ such that $(U,\overline{U})$ is the unique minimum $(S,\overline{U})$-terminal cut. By Corollary 2.1, there exists a subset $T\subseteq\overline{U}$ with $|T|\leq\lfloor 2\alpha\rfloor+1$ such that $(U,\overline{U})$ is the unique minimum $(U,T)$-terminal cut. For these choices of subsets $S$ and $T$, we now show that $(U,\overline{U})$ is the unique minimum $(S,T)$-terminal cut. Let $(Y,\overline{Y})$ be an arbitrary minimum $(S,T)$-terminal cut. It suffices to show that $Y=U$. Since $(U,\overline{U})$ is a $(S,T)$-terminal cut, we have that $d(Y)\leq d(U).$ Since $(Y\cap U,\overline{Y\cap U})$ is a $(S,\overline{U})$-terminal cut and $(U,\overline{U})$ is a minimum $(S,\overline{U})$-terminal cut, we have that $d(U)\leq d(Y\cap U).$ Since $(Y\cup U,\overline{Y\cup U})$ is a $(U,T)$-terminal cut and $(U,\overline{U})$ is a minimum $(U,T)$-terminal cut, we have that $d(U)\leq d(Y\cup U).$ Using the above inequalities in conjunction with the submodularity of the cut function, we obtain that $2d(U)\leq d(U\cap Y)+d(U\cup Y)\leq d(U)+d(Y)\leq 2d(U).$ Hence, all of the above inequalities should be equations. Consequently, $(Y\cap U,\overline{Y\cap U})$ is a minimum $(S,\overline{U})$-terminal cut, and $(Y\cup U,\overline{Y\cup U})$ is a minimum $(U,T)$-terminal cut. This implies that $Y\cap U=U$ since $(U,\overline{U})$ is the unique minimum $(S,\overline{U})$-terminal cut and $Y\cup U=U$ since $(U,\overline{U})$ is the unique minimum $(U,T)$-terminal cut. Thus, we have that $Y=U$. ∎ ## 3 An Open Problem The fundamental structural result that the number of $\alpha$-approximate minimum cuts in a $n$-vertex connected graph is at most $n^{O(\alpha)}$ has been extended to $r$-rank hypergraphs (the _rank_ of a hypergraph is the size of the largest hyperedge): Kogan and Krauthgamer [14] used the random contraction technique to show that the number of $\alpha$-approximate minimum cuts in a $r$-rank $n$-vertex connected hypergraph is $O(2^{\alpha r}n^{2\alpha})$ and they can all be enumerated in randomized polynomial time for constant $\alpha$ and $r$. It would be interesting to give a _deterministic_ proof of this bound. ## References * [1] K. Ahn and S. Guha, _Graph sparsification in the semi-streaming model_ , Proceedings of the 36th International Colloquium on Automata, Languages and Programming: Part II, ICALP, 2009, pp. 328–338. * [2] K. Ahn, S. Guha, and A. McGregor, _Analyzing graph structure via linear measurements_ , Proceedings of the 23rd Annual ACM-SIAM Symposium on Discrete Algorithms, SODA, 2012, pp. 459–467. * [3] , _Graph sketches: Sparsification, spanners, and subgraphs_ , Proceedings of the 31st Symposium on Principles of Database Systems, PODS, 2012, pp. 5–14. * [4] A. Benczur, _A representation of cuts within 6/5 times the edge connectivity with applications_ , Proceedings of the 36th Annual IEEE Foundations of Computer Science, FOCS, 1995, pp. 92–102. * [5] , _Cut structures and randomized algorithms in edge-connectivity problems_ , Ph.D. thesis, MIT, 1997. * [6] A. Benczur and M. Goemans, _Deformable polygon representation and near-mincuts_ , Building Bridges: Between Mathematics and Computer Science, M. Groetschel and G.O.H. Katona, Eds., Bolyai Society Mathematical Studies 19 (2008), 103–135. * [7] K. Chandrasekaran and C. Chekuri, _Hypergraph $k$-cut for fixed $k$ in deterministic polynomial time_, Mathematics of Operations Research (2022), Prelim. version in Proceedings of the 61st Annual Symposium on Foundations of Computer Science (FOCS 2020) pages 810–821. * [8] C. Chekuri, K. Quanrud, and C. Xu, _LP relaxation and tree packing for minimum $k$-cut_, SIAM Journal on Discrete Mathematics 34 (2020), no. 2, 1334–1353. * [9] M. X. Goemans and V. S. Ramakrishnan, _Minimizing submodular functions over families of sets_ , Combinatorica 15 (1995), 499–513. * [10] Y. Kamidoi, N. Yoshida, and H. Nagamochi, _A Deterministic Algorithm for Finding All Minimum $k$-Way Cuts_, SIAM Journal on Computing 36 (2007), no. 5, 1329–1341. * [11] D. Karger, _A randomized fully polynomial time approximation scheme for the all-terminal network reliability problem_ , SIAM Journal on Computing 29 (1999), no. 2, 492–514. * [12] , _Minimum cuts in near-linear time_ , Journal of the ACM 47 (2000), no. 1, 46–76. * [13] D. Karger and C. Stein, _A new approach to the minimum cut problem_ , Journal of the ACM 43 (1996), no. 4, 601–640. * [14] D. Kogan and R. Krauthgamer, _Sketching cuts in graphs and hypergraphs_ , Proceedings of the 2015 Conference on Innovations in Theoretical Computer Science, ITCS, 2015, pp. 367–376. * [15] H. Nagamochi, K. Nishimura, and T. Ibaraki, _Computing all small cuts in an undirected network_ , SIAM Journal on Discrete Mathematics 10 (1997), no. 3, 469–481. * [16] D. Naor, D. Gusfield, and C. Martel, _A fast algorithm for optimally increasing the edge connectivity_ , SIAM Journal on Computing 26 (1997), no. 4, 1139–1165. * [17] R. Zenklusen, _A $1.5$-Approximation for Path TSP_, Proceedings of the 30th Annual ACM-SIAM Symposium on Discrete Algorithms, SODA, 2019, pp. 1539–1549.
# Extreme nature of four blue-excess dust-obscured galaxies revealed by optical spectroscopy Akatoki Noboriguchi School of General Education, Shinshu University, 3-1-1 Asahi, Matsumoto, Nagano 390-8621, Japan Frontier Research Institute for Interdisciplinary Sciences, Tohoku University, 6-3 Aramaki, Aoba-ku, Sendai, Miyagi 980-8578, Japan Graduate School of Science and Engineering, Ehime University, 2-5 Bunkyo-cho, Matsuyama, Ehime 790-8577, Japan Tohru Nagao Research Center for Space and Cosmic Evolution, Ehime University, 2-5 Bunkyo- cho, Matsuyama, Ehime 790-8577, Japan Yoshiki Toba National Astronomical Observatory of Japan, 2-21-1 Osawa, Mitaka, Tokyo 181-8588, Japan Department of Astronomy, Kyoto University, Kitashirakawa-Oiwake-cho, Kyoto 606-8502, Japan Academia Sinica Institute of Astronomy and Astrophysics, 11F of Astronomy-Mathematics Building, AS/NTU, No.1, Section 4, Roosevelt Road, Taipei 10617, Taiwan Research Center for Space and Cosmic Evolution, Ehime University, 2-5 Bunkyo-cho, Matsuyama, Ehime 790-8577, Japan Kohei Ichikawa Frontier Research Institute for Interdisciplinary Sciences, Tohoku University, 6-3 Aramaki, Aoba-ku, Sendai, Miyagi 980-8578, Japan Masaru Kajisawa Graduate School of Science and Engineering, Ehime University, 2-5 Bunkyo-cho, Matsuyama, Ehime 790-8577, Japan Nanako Kato Graduate School of Science and Engineering, Ehime University, 2-5 Bunkyo-cho, Matsuyama, Ehime 790-8577, Japan Toshihiro Kawaguchi Department of Economics, Management and Information Science Onomichi City University, Hisayamada 1600-2, Onomichi, Hiroshima 722-8506, Japan Hideo Matsuhara Institute of Space and Astronautical Science, Japan Aerospace Exploration Agency, 3-1-1 Yoshinodai, Chuo-ku, Sagamihara, Kanagawa 252-5210, Japan Department of Space and Astronautical Science, The Graduate University for Advanced Studies, SOKENDAI, 3-1-1 Yoshinodai, Chuo-ku, Sagamihara, Kanagawa 252-5210, Japan Yoshiki Matsuoka Research Center for Space and Cosmic Evolution, Ehime University, 2-5 Bunkyo- cho, Matsuyama, Ehime 790-8577, Japan Kyoko Onishi Department of Space, Earth and Environment, Chalmers University of Technology, Onsala Observatory, SE-439 92 Onsala, Sweden Masafusa Onoue Kavli Institute for the Physics and Mathematics of the Universe (Kavli IPMU, WPI), The University of Tokyo, 5-1-5 Kashiwanoha, Kashiwa, Chiba 277-8583, Japan Kavli Institute for Astronomy and Astrophysics, Peking University, Beijing 100871, China Max-Planck-Institut für Astronomie, Königstuhl 17, D-69117, Heidelberg, Germany Nozomu Tamada Graduate School of Science and Engineering, Ehime University, 2-5 Bunkyo-cho, Matsuyama, Ehime 790-8577, Japan Koki Terao Subaru Telescope, National Astronomical Observatory of Japan, 650 North A’ohoku Place, Hilo, HI 96720, USA Astronomical Institute, Tohoku University, 6-3 Aramaki, Aoba-ku, Sendai, Miyagi 980-8578, Japan Yuichi Terashima Graduate School of Science and Engineering, Ehime University, 2-5 Bunkyo-cho, Matsuyama, Ehime 790-8577, Japan Research Center for Space and Cosmic Evolution, Ehime University, 2-5 Bunkyo-cho, Matsuyama, Ehime 790-8577, Japan Yoshihiro Ueda Department of Astronomy, Kyoto University, Kitashirakawa-Oiwake-cho, Kyoto 606-8502, Japan Takuji Yamashita National Astronomical Observatory of Japan, 2-21-1 Osawa, Mitaka, Tokyo 181-8588, Japan ###### Abstract We report optical spectroscopic observations of four blue-excess dust-obscured galaxies (BluDOGs) identified by Subaru Hyper Suprime-Cam. BluDOGs are a sub- class of dust-obscured galaxies (DOGs, defined with the extremely red color $(i-[22])_{\rm AB}\geq 7.0$; Toba et al. 2015), showing a significant flux excess in the optical $g$\- and $r$-bands over the power-law fits to the fluxes at the longer wavelengths. Noboriguchi et al. (2019) has suggested that BluDOGs may correspond to the blowing-out phase involved in a gas-rich major merger scenario. However the detailed properties of BluDOGs are not understood because of the lack of spectroscopic information. In this work, we carry out deep optical spectroscopic observations of four BluDOGs using Subaru/FOCAS and VLT/FORS2. The obtained spectra show broad emission lines with extremely large equivalent widths, and a blue wing in the C iv line profile. The redshifts are between 2.2 and 3.3. The averaged rest-frame equivalent widths of the C iv lines are $160\pm 33$ Å, $\sim$7 times higher than the average of a typical type-1 quasar. The FWHMs of their velocity profiles are between 1990 and 4470 ${\rm km\ s^{-1}}$, and their asymmetric parameters are 0.05 and 0.25. Such strong C iv lines significantly affect the broad-band magnitudes, which is partly the origin of the blue excess seen in the spectral energy distribution of BluDOGs. Their estimated supermassive black hole masses are $1.1\times 10^{8}<M_{\rm BH}/M_{\odot}<5.5\times 10^{8}$. The inferred Eddington ratios of the BluDOGs are higher than 1 ($1.1<\lambda_{\rm Edd}<3.8$), suggesting that the BluDOGs are in a rapidly evolving phase of supermassive black holes. galaxies: active — galaxies: evolution — infrared: galaxies — quasars: general — techniques: spectroscopic ††software: Astropy, Astro-SCRAPPY, X-CIGALE, Numpy, Scipy.optimaize.curve_fit, Recipe flexible execution workbench (Reflex)††facilities: Subaru (FOCAS, HSC), VLT (FORS2), VISTA, WISE, and Herschel (PACS, SPIRE) ## 1 Introduction In the last two decades, observations of low-redshift galaxies have revealed tight correlations between the mass of supermassive black holes (SMBHs) and the host galaxy properties such as bulge mass (e.g., Magorrian et al. 1998; Ferrarese & Merritt 2000; Gebhardt et al. 2000; Tremaine et al. 2002; Marconi & Hunt 2003; Kormendy & Ho 2013; Ding et al. 2020). Such scaling relations suggest the so-called co-evolution between galaxies and SMBHs. It has been argued that a major merger of gas-rich galaxies triggers active star-forming activity and subsequent mass accretion onto SMBHs (e.g., Sanders et al. 1988; Hopkins et al. 2008; Treister et al. 2012; Goulding et al. 2018). In this scenario, the merging two galaxies first evolve into a dusty star-forming (SF) galaxy. Then it evolves into a dusty active galactic nucleus (AGN) as gas accretion to the nuclear region triggers the activity of SMBHs. Finally, a dusty AGN evolves into an optically-thin quasar after the surrounding dust is blown out by the powerful AGN outflow. The most active period of such SF and AGN activity is generally heavily obscured by dust, which prevents us from investigating these phases observationally. By combining optical, near-infrared (NIR), and mid-infrared (MIR) catalogs obtained from the Subaru Hyper Suprime-Cam (HSC; Miyazaki et al. 2018)-Subaru Strategic Program (SSP; Aihara et al. 2018), the VISTA Kilo-degree Infrared Galaxy survey (VIKING; Arnaboldi et al. 2007), and the Wide-field Infrared Survey Explorer (WISE; Wright et al. 2010) all-sky survey (ALLWISE; Cutri 2014), Toba et al. (2015, 2017b) and Noboriguchi et al. (2019) selected dusty SF galaxies and/or powerful AGNs as dust-obscured galaxies (DOGs; Dey et al. 2008; Fiore et al. 2008; Bussmann et al. 2009; Desai et al. 2009; Bussmann et al. 2011). DOGs are defined with a very red optical-MIR color ($(i-[22])_{\rm AB}\geq 7.0$; Toba et al. 2015). DOGs represent a transition phase from a gas- rich major merger to an optically-thin quasar in the gas-rich major merger scenario (Dey et al. 2008), suggesting that some DOGs are expected to have buried AGNs. Recently, eight blue-excess DOGs (BluDOGs; Noboriguchi et al. 2019) were discovered from the HSC-selected DOGs based on their optical spectral slopes (i.e., $\alpha_{\rm opt}<0.4$, where $\alpha_{\rm opt}$ is the observed-frame optical spectral index for the HSC $g$-, $r$-, $i$-, $z$-, and $y$-bands in the power-law fit, $f_{\nu}\propto\lambda^{\alpha_{\rm opt}}$), and are a very rare population (eight BluDOGs out of 571 HSC-selected DOGs). Noboriguchi et al. (2019) suggested that the BluDOGs with such blue excess may be in the blowing-out phase involved in the gas-rich major merger scenario. However, the detailed properties of BluDOGs are not well understood because of the lack of spectroscopic information. Spectroscopic observations will give us accurate redshifts, and thus reliable AGN luminosities as a measure of the accretion rates, as well as the SMBH masses. Another interesting population that may represent the transition phase between optically-thick AGNs and optically-thin quasars is extremely red quasars (ERQs; e.g., Ross et al. 2015; Hamann et al. 2017; Perrotta et al. 2019; Villar Martín et al. 2020). ERQs were identified by combining the optical photometric data of Sloan Digital Sky Survey (SDSS; York et al. 2000), the optical spectroscopic data from SDSS-III (Eisenstein et al. 2011) Baryon Oscillation Spectroscopic Survey (BOSS; Dawson et al. 2013), and the MIR photometric data of WISE catalog. They are also defined with very red optical to MIR colors ($F_{\nu}(24{\rm\mu m})/F_{\nu}(R)\geq 1000$), and their spectra show broad emission lines with extremely large equivalent widths (Ross et al. 2015; Hamann et al. 2017). Hamann et al. (2017) refined the definition of ERQs as $(i-[12])_{\rm AB}>4.6$111All of the BluDOGs also satisfy the criterion of the ERQ (see Table 1)., and reported notable blue-wing features in their C iv profiles, which suggests the presence of powerful outflow. However, the ERQ sample is limited to optically bright objects since their selection requires SDSS spectra. Detailed studies of optically-faint populations in the transition phase between the optically-thick and optically-thin stages are required to understand the whole scenario of the merger-driven evolution of SMBHs. Therefore, it is important to execute the spectroscopic observations for BluDOGs and to research their spectroscopic properties. In this work, we present the results of spectroscopic observations and subsequent analyses of four BluDOGs. This paper is organized as follows. We describe sample selection of our targets and observations in Section 2. In Section 3, we present properties of the detected emission lines, the estimated dust extinctions, bolometric luminosities of an AGN ($L^{\rm AGN}_{\rm bol}$), and SMBH masses ($M_{\rm BH}$). The discussion on the large equivalent widths of the C iv emission, their SMBH mass, and Eddington ratios is given in Section 4. Then we give a brief summary in Section 5. Throughout this paper, the adopted cosmology is a flat universe with $H_{0}=70\ {\rm km\ s^{-1}\ Mpc^{-1}}$, $\Omega_{M}=0.3$, and $\Omega_{\Lambda}=0.7$. Unless otherwise noted, all magnitudes refer to the AB system. Table 1: Photometric data of BluDOGs Name | HSC $r$-band | HSC $i$-band | WISE $W3$-band | WISE $W4$-band ---|---|---|---|--- | [AB mag] | [AB mag] | [AB mag] | [AB mag] HSC J090705.64$+$020955.8 (HSC J0907) | 22.56$\pm$0.01 | 22.59$\pm$0.01 | 16.06$\pm$0.13 | 14.89$\pm$0.34 HSC J120200.84$-$011846.4 (HSC J1202) | 20.92$\pm$0.00 | 20.87$\pm$0.00 | 14.47$\pm$0.04 | 13.46$\pm$0.10 HSC J120728.71$-$005808.4 (HSC J1207) | 22.12$\pm$0.01 | 22.31$\pm$0.01 | 16.28$\pm$0.16 | 15.01$\pm$0.36 HSC J141435.21$+$003547.4 | 23.32$\pm$0.02 | 23.11$\pm$0.02 | 17.24aaThe magnitude is a 95% confidence upper limit. https://wise2.ipac.caltech.edu/docs/release/allwise/expsup/sec2_1a.html | 15.33$\pm$0.33 HSC J143727.40$-$011726.5 | 23.17$\pm$0.02 | 23.10$\pm$0.01 | 16.94$\pm$0.23 | 15.37$\pm$0.31 HSC J144333.84$-$000830.3 (HSC J1443) | 22.34$\pm$0.01 | 22.24$\pm$0.01 | 16.14$\pm$0.10 | 15.04$\pm$0.23 HSC J144813.65$+$002244.3 | 23.55$\pm$0.02 | 23.43$\pm$0.02 | 16.83$\pm$0.16 | 15.37$\pm$0.34 HSC J144900.84$+$002350.2 | 23.95$\pm$0.03 | 23.74$\pm$0.02 | 17.15$\pm$0.22 | 15.47$\pm$0.36 Table 2: Observation log Name | Exp. time [s] | Date | Standard star | Instrument ---|---|---|---|--- HSC J0907 | 900$\times$2 | 2019 October 8 | G191-B2B | FOCAS (Subaru) | 600$\times$1 | | | HSC J1202 | 900$\times$6 | 2019 February 27 | LTT 6248 | FORS2 (VLT) HSC J1207 | 900$\times$12 | 2019 March 1, 2, 6 | LTT 4816 | FORS2 (VLT) HSC J1443 | 900$\times$12 | 2019 March 7, 8 | LTT 4816, EG 274 | FORS2 (VLT) ## 2 Sample and the data ### 2.1 Sample selection In Noboriguchi et al. (2019), 571 DOGs were selected by combining $\sim$105 deg2 imaging data obtained from the survey of HSC-SSP222 We utilize the photometric data of S16A HSC-SSP, which was released internally within the HSC survey team and is based on data obtained from 2014 March to 2016 April. ($g$, $r$, $i$, $z$, and $y$), VIKING ($Z$, $Y$, $J$, $H$, and $Ks$), and ALLWISE ($W1$, $W2$, $W3$, and $W4$). The eight BluDOGs were defined among the DOG sample with the smallest observed-frame optical slope ($\alpha_{\rm opt}\textless 0.4$, where $\alpha_{\rm opt}$ is the observed-frame optical spectral index of the power-law fits to the HSC $g$, $r$, $i$, $z$, and $y$-band fluxes, $f_{\nu}\propto\lambda^{\alpha_{\rm opt}}$). We selected the four brightest BluDOGs ($r_{\rm AB}<23$: see Table 1) as the targets of our spectroscopic observations presented in this paper. ### 2.2 Spectroscopic observations and data reductions We executed the observations by using Faint Object Camera and Spectrograph (FOCAS; Kashikawa et al. 2002) installed on the Subaru Telescope of National Astronomical Observatory of Japan, and FORS2 (Appenzeller et al. 1998) installed on Very Large Telescope (VLT-UT1) of European Southern Observatory (ESO). We present the observation log in Table 2. #### 2.2.1 Subaru FOCAS By using FOCAS, we observed HSC J090705.64$+$020955.8 (hereafter J0907) on October 8th in 2019, with airmass $\sim$1.76 and seeing $\sim$0.5 arcsec. We used the 300B grism and the SY47 filter to cover $\lambda_{\rm obs}\sim$4700–9200 Å, with the resultant spectral resolution of $R\sim$800 for the used 0″.8-width slit. To reduce the obtained data, we performed bias correction, flat fielding with dome flat, removal of cosmic-rays, spectral extraction, sky subtraction, wavelength calibration, and flux calibration with a standard star (G191-B2B) using the Python packages of `Astropy` and `Numpy`. For removing cosmic-rays, we utilized `Astro-SCRAPPY` (McCully & Tewes 2019). `Astro-SCRAPPY` is based on the algorithm of `L.A.Cosmic`, which removes cosmic-rays based on a variation of Laplacian edge detection (van Dokkum 2001). The final spectrum is an inverse-variance weighted mean of the individual shots, corrected for the Galactic extinction (Schlegel et al. 1998). #### 2.2.2 VLT FORS2 By using FORS2, we observed HSC J120200.84$-$011846.4, HSC J120728.71$-$005808.4, and HSC J144333.84$-$000830.3 (hereafter J1202, J1207, and J1443, respectively) between February 27th and March 8th, 2019. We used the GRISM_600RI$+19$ and the GG435 filter to cover $\lambda_{\rm obs}\sim$5200–8000 Å, which results in the spectral resolution of $R\sim$1500 with 0″.7-width slit. The typical airmasses of the observations for J1202, J1207, and J1443 were 1.17, 1.24, and 1.12, and the typical seeing sizes were $\sim$1.0, 0.5, and 0.5 arcsec, respectively. For the data reduction, we utilized the `Recipe flexible execution workbench` (`Reflex`; Freudling et al. 2013) software. `Reflex` performed bias correction, flat fielding with dome flat, sky subtraction, removing cosmic-rays, spectral extraction, wavelength calibration, and flux calibration with a standard star (LTT 6248, LTT 4816, and EG 274). The final spectrum of each target is the inverse-variance weighted mean of the individual shots, corrected for the Galactic extinction. Figure 1: The reduced spectra of the BluDOGs. The spectra are for J0907, J1202, J1207 and J1443 from the top to bottom. Detected lines are marked by arrows and labels. #### 2.2.3 Spectrophotometric re-calibration We re-calibrated the reduced spectra to match the HSC photometry, in order to correct for the effects of the slit loss of the flux, systematic errors in the photometric and spectroscopic calibrations, and any other possible systematic errors. In our observations, the spectra cover the wavelength range of the HSC $r$-band. We calculate the calibration factor, $f_{\rm photo\\_calib}=F_{{\rm photo}\\_r}/F_{{\rm spec}\\_r}$, where $F_{{\rm photo}\\_r}$ and $F_{{\rm spec}\\_r}$ are the photometric and spectroscopic fluxes in the HSC $r$-band. The derived calibration factors of J0907, J1202, J1207, and J1443 are 0.97, 1.50, 1.40, and 1.36, respectively. We multiply the spectra with the derived calibration factors. ## 3 Results Figure 2: Spectral fits to the C iv emission lines of the BluDOGs. The top left, top right, bottom left, and bottom right show the C iv emission lines of J0907, J1202, J1207, and J1443, respectively. The green, magenta, red, blue, and black lines represent the observed spectrum, linear fit to the continuum emission, two Gaussians for the red and blue components, and best-fit model, respectively. The orange line on the J1202 and J1443 panels represents the C iv doublet absorption line. The horizontal black bars denote the wavelength range used to fit the continuum emission. In each panel, the lower part presents the residual of the best-fit, with the same flux scale as in the upper part. ### 3.1 Emission-line measurements Figure 1 shows the reduced spectra of the four BluDOGs. In order to measure the emission-line properties, we divide emission lines into six groups as follows; (1) Ly$\alpha$1216, N v$\lambda$1240, and Si ii$\lambda$1263, (2) Si iv$\lambda$1397 and O iv]$\lambda$1402, (3) He ii$\lambda$1640 and O iii]$\lambda$1663, and (4) Al iii$\lambda$1857, Si iii]$\lambda$1892, and C iii]$\lambda$1909, (5) C iv 1549, and (6) Mg ii. We fit the emission lines in each group simultaneously, with a linear continuum model subtracted from the observed spectrum. We adopt a single-Gaussian profile for Ly ${\alpha}$, N v, Si ii, Si iv, O iv], O iii], Al iii, Si iii and Mg ii. The C iii] of J1202 is fitted with a single-Gaussian profile, while those of J0907 and J1207 are fitted with a double-Gaussian profile. For the fit around the Si iv and O iv] of J1202, we add an additional Gaussian profile to reproduce the observed broad component. We fit C iv and He ii with double-Gaussian profiles, and denote the blue and red components with the suffixes of “_B” and “_R”, respectively. Additionally, we fit the doublet absorption lines observed around the C iv emission lines of J1202 and J1443. The C iv absorption lines observed at $\lambda_{\rm obs}$ = 5916.8 Å and 5926.6 Å for J1202 and those at $\lambda_{\rm obs}$ = 6678.8 Å and 6689.9 Å for J1443 are fitted using the Voigt profile, respectively. The doublet absorption line ratio is fixed as 2:1 (Feibelman, 1983). The best values and standard deviations for emission and absorption lines parameters are estimated by using scipy.optimize.curve_fit333https://docs.scipy.org/doc/scipy/reference/, while we calculate full width at half maximum (FWHM) of emission lines with double Gaussian by using a Monte Carlo method. For this Monte Carlo simulation, we created 10,000 mock spectra using the noise arrays of the observed spectra, and calculate the mean and standard deviation of the line properties. The results for emission lines are listed in Tables 3–6. For absorption lines, the observed-frame equivalent widths and redshifts of the doublet absorption lines on the J1202 C iv emission line are 47.7$\pm$19.6 Å, 23.6$\pm$9.7 Å, and 2.822, respectively, while the observed-frame equivalent widths and redshift of the doublet absorption lines on J1443 C iv emission line are 14.0$\pm$6.3 Å, 6.94$\pm$3.14 Å, and 3.314, respectively. Therefore, the co-moving distance between J1202 and its C iv absorber is 8.73 Mpc, while that between J1443 and its C iv absorber is 2.79 Mpc. The flux ratios of N v/Ly$\alpha$ and N v/C iv for J1443 are 3.9 and 1.8, respectively, whereas the values for the typical quasar (Vanden Berk et al., 2001) are 0.02 and 0.10. One possible reason of these unusual flux ratios in J1443 is the presence of absorption lines, which absorb most of the Ly$\alpha$ and the C iv fluxes around the peak. The unusual flux ratios cannot be explained by the dust reddening, given too small wavelength separations among emission lines of Ly$\alpha$, N v, and C iv. Figure 2 showcases the best-fit models to the C iv emission lines in the four BluDOGs. We adopt the C iv redshift taking C iv_R $+$ C iv_B into account as the systemic redshift of the targets. The determined systemic redshifts of J0907, J1202, J1207, and J1443 are 2.258$\pm$0.002, 2.830$\pm$0.002, 2.511$\pm$0.001, and 3.317$\pm$0.006, respectively. ### 3.2 Emission-line contributions to the HSC $g$\- and $r$-band magnitudes Figure 1 suggests the very large equivalent width (EW) of the emission lines. The average rest-frame EW (REW) of the C iv line of the four BluDOGs is $160\pm 33$ Å, $\sim$7 times higher than the average of SDSS type-1 quasars ($23.8\pm 0.1$ Å; Vanden Berk et al. 2001). Here we investigate the effect of the large REWs on the HSC $g$\- and $r$-band magnitudes. First, we calculate the expected magnitudes at the $g$\- and $r$-bands from an extrapolation of the power-law fit to the longer wavelength bands ($i$, $z$, $y$, $Z$, $Y$, $J$, $H$, $Ks$, $W1$, $W2$, $W3$, and $W4$). Figure 3 clearly shows that the observed $g$\- and $r$-band magnitudes exceed the extrapolation of the power-law fit. The excesses of the $g$-band magnitudes for J0907, J1202, J1207, and J1443 are 1.27, 1.13, 1.48, and 0.88 mag, and those for the $r$-band excesses are 0.47, 0.44, 0.65, and 0.40 mag, respectively. Furthermore, we estimate the effect of the strong emission lines, based on their observed-frame EWs and the band widths (BW) of the HSC $g$ and $r$-bands. The BWs of the HSC $g$ and $r$-bands are 1468 and 1508 Å (Kawanomoto et al., 2018), respectively. By taking all of the emission lines (Tables 3 – 6) covered by the HSC $g$-band (4000–5500Å) and $r$-band (5500–7000Å) into account, the total observed-frame EWs for J0907, J1202, J1207, and J1443 in the $g$-band are 604, 258, 608, and 1380 Å respectively, while those in the $r$-band are 129, 836, 329, and 721 Å (Table 7). Note that the total observed-frame EWs in the $g$-band are lower limits, because our optical spectra do not cover the entire wavelength range of the band (Section 2.2) and thus some emission lines are not taken into account in the derived total observed-frame EWs. Especially Ly$\alpha$, the strongest emission line in the rest-frame UV spectrum of typical AGNs, is not covered in our spectra of J0907, J1202, and J1207, thus the total observed-frame EWs for these 3 objects are largely underestimated444The Ly$\alpha$ line of J0907 is at the shorter edge of the HSC $g$-band coverage but the flux contribution to the $g$-band magnitude is likely to be significant owing to its broad nature.. Since the magnitude excess by the emission lines is given by $\Delta{\rm mag}=2.5\log{(1+EW/BW)}$, the estimated effects in the $g$-band for J0907, J1202, J1207, and J1443 are 0.37, 0.18, 0.38, and 0.72 mag, respectively. Similarly, the estimated effects of emission lines to the $r$-band magnitudes are 0.09, 0.48, 0.21, and 0.42 mag, respectively (see Table 7 for a summary). We will discuss the implication from these estimates in Section 4.2. Table 3: The detected lines of J0907 Line name | $\lambda_{\rm rest}$ [Å] | $z_{\rm line}$ | $FWHM_{\rm rest}$ [Å] | $F_{\rm line}$ [${\rm erg\ s^{-1}\ cm^{-2}}$] | $EW_{\rm rest}$ [Å] | $v_{\rm width}$ [${\rm km\ s^{-1}}$] ---|---|---|---|---|---|--- (1) | (2) | (3) | (4) | (5) | (6) | (7) C iv_R | 1549.5 | 2.258$\pm$0.001 | 12.8$\pm$0.7 | (1.10$\pm$0.07)E$-$15 | 118$\pm$10 | 2470$\pm$130 C iv_B | 1549.5 | 2.227$\pm$0.004 | 13.8$\pm$2.5 | (2.87$\pm$0.60)E$-$16 | 31.5$\pm$6.7 | 2670$\pm$490 C iv_R + C iv_B | 1549.5 | 2.258$\pm$0.002 | 15.2$\pm$0.8 | (1.39$\pm$0.09)E$-$15 | 148$\pm$12 | 2940$\pm$150 He ii_R | 1640.4 | 2.260$\pm$0.002 | 3.63$\pm$2.32 | (3.50$\pm$2.36)E$-$17 | 4.43$\pm$3.01 | 663$\pm$425 He ii_B | 1640.4 | 2.235$\pm$0.005 | 38.1$\pm$5.9 | (2.13$\pm$0.48)E$-$16 | 26.8$\pm$6.3 | 6960$\pm$1080 He ii_R + He ii_B | 1640.4 | 2.260$\pm$0.002 | 5.54$\pm$1.29 | (2.48$\pm$0.53)E$-$16 | 31.4$\pm$7.0 | 1010$\pm$240 O iii] | 1663.5 | 2.264$\pm$0.001 | 3.29$\pm$1.69 | (4.49$\pm$1.48)E$-$17 | 5.84$\pm$1.98 | 593$\pm$305 Si iii | 1892.0 | 2.258$\pm$0.002 | 4.03$\pm$4.16 | (2.97$\pm$2.55)E$-$17 | 3.56$\pm$3.06 | 639$\pm$659 C iii]_R | 1908.7 | 2.261$\pm$0.005 | 30.3$\pm$5.3 | (2.34$\pm$0.66)E$-$16 | 28.2$\pm$8.1 | 4760$\pm$830 C iii]_R | 1908.7 | 2.258$\pm$0.002 | 6.19$\pm$3.04 | (6.50$\pm$3.21)E$-$17 | 7.83$\pm$3.89 | 971$\pm$477 C iii]_R + C iii]_B | 1908.7 | 2.258$\pm$0.003 | 11.6$\pm$1.7 | (2.99$\pm$0.73)E$-$16 | 36.0$\pm$9.0 | 1830$\pm$260 Mg ii | 2799.1 | 2.259$\pm$0.001 | 17.6$\pm$1.6 | (2.44$\pm$0.29)E$-$16 | 49.0$\pm$7.3 | 1890$\pm$170 Note. — Column (1): Line name, (2): Rest-frame wavelength of the line, (3): Line redshift, (4): Rest-frame FWHM, (5): Line flux, (6): Rest-frame EW, (7): Velocity width after the correction for the instrumental broadening. Table 4: The detected lines of J1202 Line name | $\lambda_{\rm rest}$ [Å] | $z_{\rm line}$ | $FWHM_{\rm rest}$ [Å] | $F_{\rm line}$ [${\rm erg\ s^{-1}\ cm^{-2}}$] | $EW_{\rm rest}$ [Å] | $v_{\rm width}$ [${\rm km\ s^{-1}}$] ---|---|---|---|---|---|--- (1) | (2) | (3) | (4) | (5) | (6) | (7) Si iv | 1393.8 | 2.831$\pm$0.001 | 4.27$\pm$0.73 | (1.56$\pm$0.33)E$-$16 | 5.55$\pm$1.17 | 919$\pm$157 Broad componenta | — | — | — | (1.23$\pm$0.13)E$-$15 | — | 4970$\pm$170 O iv] | 1399.9 | 2.842$\pm$0.001 | 9.90$\pm$0.87 | (5.13$\pm$0.68)E$-$16 | 18.1$\pm$2.4 | 2120$\pm$190 C iv_R | 1549.5 | 2.831$\pm$0.001 | 15.0$\pm$0.4 | (5.46$\pm$0.26)E$-$15 | 177$\pm$9 | 2900$\pm$70 C iv_B | 1549.5 | 2.793$\pm$0.002 | 13.0$\pm$1.0 | (7.98$\pm$1.03)E$-$16 | 27.7$\pm$3.6 | 2510$\pm$190 C iv_R + C iv_B | 1549.5 | 2.830$\pm$0.002 | 16.0$\pm$0.5 | (6.26$\pm$0.28)E$-$15 | 203$\pm$10 | 3100$\pm$90 He ii_R | 1640.4 | 2.838$\pm$0.002 | 11.0$\pm$1.7 | (1.68$\pm$0.47)E$-$16 | 5.06$\pm$1.43 | 2010$\pm$310 He ii_B | 1640.4 | 2.806$\pm$0.005 | 21.1$\pm$3.4 | (3.16$\pm$0.55)E$-$16 | 9.31$\pm$1.64 | 3870$\pm$620 He ii_R + He ii_B | 1640.4 | 2.834$\pm$0.004 | 25.5$\pm$2.5 | (4.84$\pm$0.73)E$-$16 | 14.5$\pm$2.2 | 4650$\pm$450 O iii] | 1663.5 | 2.852$\pm$0.003 | 7.57$\pm$2.94 | (2.06$\pm$1.07)E$-$17 | 0.661$\pm$0.343 | 1360$\pm$530 Al iii | 1858.8 | 2.839$\pm$0.002 | 21.5$\pm$2.4 | (3.05$\pm$0.41)E$-$16 | 10.5$\pm$1.4 | 3470$\pm$390 Si iii | 1892.0 | 2.819$\pm$0.004 | 13.6$\pm$5.5 | (9.40$\pm$4.81)E$-$17 | 3.19$\pm$1.64 | 2160$\pm$870 C iii] | 1908.7 | 2.835$\pm$0.002 | 30.6$\pm$1.9 | (1.01$\pm$0.07)E$-$15 | 33.5$\pm$2.4 | 4810$\pm$300 Note. — See Table 3 for the description of each column. aThe observed-frame wavelength, continuum flux, and FWHM of the broad component are 5349.5$\pm$2.0 Å, (7.36$\pm$0.12)E$-$18 ${\rm erg\ s^{-1}\ cm^{-2}}$ Å-1, and 88.9$\pm$2.9 Å, respectively. See Section 3.1 for details. Table 5: The detected lines of J1207 Line name | $\lambda_{\rm rest}$ [Å] | $z_{\rm line}$ | $FWHM_{\rm rest}$ [Å] | $F_{\rm line}$ [${\rm erg\ s^{-1}\ cm^{-2}}$] | $EW_{\rm rest}$ [Å] | $v_{\rm width}$ [${\rm km\ s^{-1}}$] ---|---|---|---|---|---|--- (1) | (2) | (3) | (4) | (5) | (6) | (7) C iv_R | 1549.5 | 2.512$\pm$0.001 | 7.87$\pm$0.43 | (1.02$\pm$0.08)E$-$15 | 74.2$\pm$6.0 | 1520$\pm$80 C iv_B | 1549.5 | 2.500$\pm$0.001 | 20.2$\pm$0.5 | (1.36$\pm$0.09)E$-$15 | 99.3$\pm$6.9 | 3900$\pm$90 C iv_R + C iv_B | 1549.5 | 2.511$\pm$0.001 | 10.3$\pm$0.4 | (2.39$\pm$0.12)E$-$15 | 173$\pm$9 | 1990$\pm$70 He ii_R | 1640.4 | 2.511$\pm$0.001 | 6.53$\pm$1.12 | (6.61$\pm$1.50)E$-$17 | 5.32$\pm$1.21 | 1190$\pm$210 He ii_B | 1640.4 | 2.499$\pm$0.002 | 17.7$\pm$1.3 | (1.47$\pm$0.24)E$-$16 | 11.7$\pm$1.9 | 3240$\pm$240 He ii_R + He ii_B | 1640.4 | 2.509$\pm$0.002 | 11.6$\pm$1.2 | (2.13$\pm$0.28)E$-$16 | 17.2$\pm$2.3 | 2120$\pm$220 O iii] | 1663.5 | 2.516$\pm$0.002 | 12.2$\pm$2.4 | (3.70$\pm$0.95)E$-$17 | 3.18$\pm$0.82 | 2200$\pm$430 Al iii | 1858.8 | 2.508$\pm$0.002 | 19.7$\pm$1.7 | (9.09$\pm$1.02)E$-$17 | 9.23$\pm$1.04 | 3180$\pm$280 Si iii | 1892.0 | 2.511$\pm$0.003 | 10.7$\pm$4.9 | (2.64$\pm$1.60)E$-$17 | 2.79$\pm$1.69 | 1690$\pm$770 C iii]_R | 1908.7 | 2.510$\pm$0.001 | 12.9$\pm$1.4 | (1.69$\pm$0.25)E$-$16 | 18.2$\pm$2.7 | 2030$\pm$220 C iii]_B | 1908.7 | 2.503$\pm$0.002 | 42.0$\pm$2.9 | (4.02$\pm$0.65)E$-$16 | 43.1$\pm$7.0 | 6600$\pm$450 C iii]_R + C iii]_B | 1908.7 | 2.509$\pm$0.002 | 19.6$\pm$1.2 | (5.71$\pm$0.70)E$-$16 | 61.4$\pm$7.5 | 3080$\pm$180 Note. — See Table 3 for the description of each column. Table 6: The detected lines of J1443 Line name | $\lambda_{\rm rest}$ [Å] | $z_{\rm line}$ | $FWHM_{\rm rest}$ [Å] | $F_{\rm line}$ [${\rm erg\ s^{-1}\ cm^{-2}}$] | $EW_{\rm rest}$ [Å] | $v_{\rm width}$ [${\rm km\ s^{-1}}$] ---|---|---|---|---|---|--- (1) | (2) | (3) | (4) | (5) | (6) | (7) Ly ${\alpha}$ | 1215.7 | 3.341$\pm$0.003 | 10.7$\pm$2.6 | (5.28$\pm$1.31)E$-$16 | 64.3$\pm$17.3 | 2650$\pm$640 N v | 1240.8 | 3.312$\pm$0.001 | 16.8$\pm$0.6 | (2.06$\pm$0.09)E$-$15 | 242$\pm$23 | 4070$\pm$160 Si ii | 1262.6 | 3.326$\pm$0.003 | 16.0$\pm$1.9 | (1.24$\pm$0.17)E$-$16 | 13.6$\pm$2.0 | 3790$\pm$450 Si iv | 1393.8 | 3.324$\pm$0.010 | 16.4$\pm$3.2 | (3.18$\pm$1.01)E$-$16 | 33.8$\pm$10.7 | 3520$\pm$700 O iv] | 1399.9 | 3.339$\pm$0.008 | 12.9$\pm$2.5 | (1.74$\pm$1.11)E$-$16 | 18.6$\pm$11.8 | 2760$\pm$540 C iv_R | 1549.5 | 3.328$\pm$0.005 | 14.2$\pm$1.6 | (6.15$\pm$1.74)E$-$16 | 60.0$\pm$17.0 | 2740$\pm$310 C iv_B | 1549.5 | 3.296$\pm$0.007 | 15.7$\pm$2.3 | (5.47$\pm$1.67)E$-$16 | 54.9$\pm$16.8 | 3030$\pm$440 C iv_R + C iv_B | 1549.5 | 3.317$\pm$0.006 | 23.1$\pm$1.8 | (1.16$\pm$0.24)E$-$15 | 114$\pm$24 | 4470$\pm$350 He ii_R | 1640.4 | 3.337$\pm$0.001 | 6.48$\pm$1.56 | (3.23$\pm$1.01)E$-$17 | 3.24$\pm$1.02 | 1180$\pm$290 He ii_B | 1640.4 | 3.307$\pm$0.004 | 20.3$\pm$2.9 | (9.15$\pm$1.53)E$-$17 | 9.25$\pm$1.56 | 3710$\pm$530 He ii_R + He ii_B | 1640.4 | 3.335$\pm$0.003 | 20.1$\pm$2.5 | (1.24$\pm$0.18)E$-$16 | 12.4$\pm$1.9 | 3680$\pm$460 Note. — See Table 3 for the description of each column. Table 7: Emission-line contribution and excess magnitude to the power-law fit in the $g$\- and $r$-band | $g$-band | $r$-band ---|---|--- | Total EWs | $\Delta$mag | Excess mag | Total EWs | $\Delta$mag | Excess mag | [Å] | [AB mag] | [AB mag] | [Å] | [AB mag] | [AB mag] HSC J0907 | $>$604 | $>$0.37 | 1.27 | 129 | 0.09 | 0.47 HSC J1202 | $>$258 | $>$0.18 | 1.13 | 836 | 0.48 | 0.44 HSC J1207 | $>$608 | $>$0.38 | 1.48 | 329 | 0.21 | 0.65 HSC J1443 | 1380 | 0.72 | 0.88 | 721 | 0.43 | 0.40 Note. — $\Delta$mag: $2.5\log{(1+EW/BW)}$, Excess mag: The excesses of the $g$\- and $r$\- band magnitudes between the observed magnitudes and the expected magnitudes from an extrapolation of the power-law fit to the longer wavelength bands. Figure 3: The SED of J0907 (top), J1202 (middle upper), J1207 (middle lower), and J1443 (bottom). The red dots denote the $g$\- and $r$-band magnitudes, while the blue dots denote the longer-wavelength optical and near-infrared magnitudes that are used for the power-law fit (black line). The green lines represent the observed spectra. Figure 4: J1207 images in the WISE $W1$-band (left) and H-ATLAS 250 $\mu$m-band (right). The orange stars, green crosses, and red cross denote source detections in the HSC-SSP, ALLWISE, and H-ATLAS catalogs, respectively. The size of each image is 60″$\times$ 60″, centered at the HSC position of J1207. ### 3.3 Estimating the dust extinction We need to estimate dust extinction, $E(B-V)$ of AGN radiation, and $L^{\rm AGN}_{\rm bol}$ to calculate the SMBH mass and Eddington ratio. Since Balmer decrement or other spectral measures of $E(B-V)$ is not available, we perform the SED fitting to the broad-band photometry to estimate the $E(B-V)$ and $L^{\rm AGN}_{\rm bol}$. In this work, we utilize the new version of Code Investigating GAlaxy Emission (CIGALE; Burgarella et al. 2005; Noll et al. 2009; Boquien et al. 2019) called X-CIGALE (Yang et al., 2020), to perform the SED fit in a self-consistent framework by considering an energy balance between the UV/optical absorption and IR emission. X-CIGALE generates the best-fit model including the stellar, AGN, and SF components that fits the photometric data in the rest-frame UV to far-infrared (FIR) bands. We utilize the Herschel Space Observatory (Pilbratt et al. 2010) Astrophysical Terahertz Large Area Survey (H-ATLAS; Eales et al. 2010; Valiante et al. 2016; Bourne et al. 2016) data observed with Photodetector Array Camera and Spectrometer (PACS; Poglitsch et al. 2010) at 100 and 160 $\mu$m and with Spectral and Photometric Imaging REceiver (SPIRE; Griffin et al. 2010) at 250, 350, and 500 $\mu$m in the FIR, in addition to optical, NIR, and MIR data obtained by Subaru HSC, VISTA, and WISE. The 1$\sigma$ limiting fluxes at 100, 160, 250, 350, and 500 $\mu$m are 44, 49, 7.4, 9.4, and 10.2 mJy, respectively (Valiante et al. 2016). To search for the H-ATLAS counterpart of the four BluDOGs, we adopt a search radius of 10 arcsec by following Toba et al. (2019) (and Toba et al. 2022). Accordingly we found the counterparts of two BluDOGs (J1202 and J1207). The separation between the HSC position and the H-ATLAS counterpart position is 0.94 arcsec for J1202, and 9.7 arcsec for J1207. The relatively large separation in the latter case suggests the counterpart being a coincidental detection. There are two WISE sources around J1207 (Figure 4); one probably corresponds to J1207 itself (the angular separation between the HSC and WISE potions is 0.65 arcsec) and another is located at 19 arcsec away to the north- east direction. The H-ATLAS source is located between these two WISE sources, and thus the FIR fluxes given in the H-ATLAS catalog are possibly attributed to the two WISE sources. Therefore we regard the H-ATLAS fluxes of J1207 as the upper limit. For the remaining two BluDOGs (J0907 and J1443), we adopt the 5$\sigma$ upper limit fluxes. As for the optical–MIR photometric data, we utilize $g$, $r$, $i$, $z$, $y$ (HSC-SSP), $Z$, $Y$, $J$, $H$, $Ks$ (VIKING DR2), $W1$, $W2$, $W3$, and $W4$ (ALLWISE) bands (see Noboriguchi et al. 2019). Note that the SNRs in these bands are more than 5, except for the $W4$-band with SNR more than 3 because we adopted such SNR cut in the selection of DOGs (Noboriguchi et al. 2019). Since the $g$\- and $r$-band photometry are significantly affected by the strong emission lines (see Figure 1 and Tables 3–6) which cannot be treated properly in X-CIGALE, we corrected for their contribution by referring to the estimates given in Table 7. Table 8: Parameters adopted in the X-CIGALE fit Parameter | Value ---|--- Delayed SFH (Ciesla et al. 2015) ${\tau}_{\rm main}$ [Myr] | 100, 250, 500 ${\tau}_{\rm burst}$ [Myr] | 10, 50 $f_{\rm burst}$ | 0.0, 0.5, 0.99 Agemain [Myr] | 500, 800, 1000 Ageburst [Myr] | 1, 5, 10 Single stellar population (Bruzual & Charlot 2003) IMF | Chabrier (2003) Metallicity | 0.02 Ageseparation [Myr] | 10 Nebular emission (Inoue 2011) $\log{U}$ | $-$2.0 $f_{\rm esc}$ | 0.0 $f_{\rm dust}$ | 0.0 Lines width [${\rm km\ s^{-1}}$] | 300.0 Dust attenuation (Calzetti et al. 2000) $E(B-V)_{\rm line}$ | 3, 4, 5, 6, 7, 8, 9, 10 $f_{E(B-V)}$ | 0.44 $\lambda_{\rm UV,bump}$ [nm] | 217.5 FWHMUV,bump [nm] | 35.0 $A_{\rm UV,bump}$ | 0.0 $\delta$ | 0.0 Extinction law of emission lines | the Milky Way $R_{V}$ | 3.1 Dust emission (Dale et al. 2014) AGN fraction | 0.0 $\alpha_{\rm IR,AGN}$ | 0.0625, 0.2500, 2.0000 AGN model (Stalevski et al. 2016) $\tau_{9.7}$ | 3, 7 $p$ | 1.0 $q$ | 1.0 $oa$ [deg] | 10, 20, 30, 40, 50, 60, 70, 80 $R_{\rm ratio}$ | 20 $M_{\rm cl}$ | 0.97 $i$ [deg] | 0, 10, 20, 30, 40, | 50, 60, 70, 80, 90 $f_{\rm AGN}$ | 0.1, 0.3, 0.5, 0.7, 0.9 Extinction law of polar dust | Calzetti et al. 2000 $E(B-V)^{\rm AGN}_{\rm polar\ dust}$ | 0.1, 0.2, 0.3, 0.4, 0.5 $T^{\rm AGN}_{\rm polar\ dust}$ [K] | 600, 700, 800, 900, | 1000, 1100, 1200, 1300, 1400 Emissivity of polar dust | 1.6 The models and parameters of X-CIGALE adopted in this work are summarized in Table 8. We assume a delayed star formation history (SFH; Ciesla et al. 2015) with the e-folding times of the main stellar population ($\tau_{\rm main}$) and late starburst population ($\tau_{\rm burst}$), mass fraction of the late burst population ($f_{\rm burst}$), and age of the main stellar population (Agemain) and the late burst (Ageburst). As the stellar population, we assume the initial mass function of Chabrier (2003), solar metallicity, and 10-Gyr separation between young and old stellar population (Ageseparation). The nebular emission model (Inoue 2011) is characterized by the ionization parameter ($U$), fractions of Lyman continuum photons escaping the galaxy ($f_{\rm esc}$) and absorbed by dust ($f_{\rm dust}$), and line width. We utilize a modified dust attenuation model presented by Boquien et al. (2019). The dust attenuation model for the continuum is taken from Calzetti et al. (2000) with the extension taken from Leitherer et al. (2002) between the Lyman break and 1500 Å. The emission lines are attenuated with a Milky Way extinction with $R_{V}=3.1$ (Cardelli et al., 1989). We assumed $E(B-V)_{\rm continuum}=0.44E(B-V)_{\rm line}$, following Calzetti et al. (2000). The $E(B-V)_{\rm line}$ is varied between 3 and 10. We utilize the SKIRTOR model as the AGN emission model, which takes geometric parameters of the AGN into account and also allows us to incorporate the effect of extinction by the polar dust. The parameters of the AGN model are the average edge-on optical depth at 9.7 $\mu$m ($\tau_{9.7}$), the torus density parameters ($p$ and $q$; Stalevski et al. 2016), the angle between the equatorial plane and the edge of the torus ($oa$), the ratio of the maximum to minimum radii of the dust torus ($R_{\rm ratio}$), the fraction of total dust mass inside clumps ($M_{\rm cl}$), the inclination ($i$), the AGN fraction ($f_{\rm AGN}$), the extinction law, color excess ($E(B-V)^{\rm AGN}_{\rm polar\ dust}$), dust temperature ($T^{\rm AGN}_{\rm polar\ dust}$), and emissivity index of the polar dust. The best-fit SED models are shown in Figure 5. The reduced $\chi^{2}$ of the fits are 1.38, 3.19, 0.93, and 1.74 for J0907, J1202, J1207, and J1443, respectively. The best-fit values and associated errors for $E(B-V)^{\rm AGN}_{\rm polar\ dust}$ and $L^{\rm AGN}_{\rm bol}$ are estimated with a Bayesian-like strategy presented in Noll et al. (2009), and are reported in Table 9. On the other hands, we cannot quantitatively constrain the parameters of the host galaxies because the $E(B-V)$ values are too large and the optical parts in their SEDs are dominated by their AGN emission (see Figure 5). Figure 5: The results of the SED fitting for the four BluDOGs. The upper left, upper right, lower left and lower right panels show the results of J0907, J1202, J1207, and J1443, respectively. The black, blue, green, red, and orange lines represent the best-fit model, stellar component (with dust attenuation), AGN component, SF component (FIR re-emission from the dust heated by SF), and nebular component, respectively. The magenta plots represent the photometric data. The arrows denote 5$\sigma$ upper limit flux. ### 3.4 Measurement of the SMBH mass Table 9: Physical properties of the four BluDOGs | HSC J0907 | HSC J1202 | HSC J1207 | HSC J1443 ---|---|---|---|--- Redshift | $2.258\pm 0.002$ | $2.830\pm 0.002$ | $2.511\pm 0.001$ | $3.317\pm 0.006$ $E(B-V)^{\rm AGN}_{\rm polar\ dust}$ | $0.26\pm 0.05$ | $0.33\pm 0.04$ | $0.30\pm 0.02$ | $0.20\pm 0.01$ $L^{\rm AGN}_{\rm bol}$/$L_{\odot}$ | $(6.11\pm 0.95)\times 10^{12}$ | $(6.11\pm 1.18)\times 10^{13}$ | $(7.95\pm 1.47)\times 10^{12}$ | $(2.52\pm 0.13)\times 10^{13}$ $M_{\rm BH}$ (C iv)/$M_{\odot}$ | $(1.69\pm 0.17)\times 10^{8}$ | $(4.95\pm 0.30)\times 10^{8}$ | $(1.11\pm 0.08)\times 10^{8}$ | $(5.48\pm 0.86)\times 10^{8}$ $M_{\rm BH}$ (Mg ii)/$M_{\odot}$ | $(9.85\pm 1.80)\times 10^{7}$ | — | — | — $\lambda_{\rm Edd}$ (C iv) | $1.10\pm 0.20$ | $3.75\pm 0.76$ | $2.19\pm 0.43$ | $1.40\pm 0.23$ We have detected the C iv emission line for all the four BluDOGs and Mg ii emission line for J0907, both of which are widely used to calculate the SMBH mass of type-1 AGNs. Note that the systematic uncertainty is larger in the C iv-based SMBH mass than in the Mg ii-based SMBH mass, due to a powerful outflow sometimes seen in the C iv velocity profile (e.g., Baskin & Laor, 2005; Netzer, 2015; Coatman et al., 2017). We calculate the single-epoch mass of SMBHs with the C iv and Mg ii emission lines, following the calibrations given in Vestergaard & Peterson (2006) and Vestergaard & Osmer (2009) respectively: $\displaystyle M_{\rm BH}$ $\displaystyle=$ $\displaystyle 10^{6.66}\Big{(}\frac{{\rm FWHM(C~{}{\sc IV})}}{10^{3}\ {\rm km\ s^{-1}}}\Big{)}^{2}\Big{(}\frac{\lambda L_{\lambda}(1350{\rm\AA})}{10^{44}\ {\rm erg\ s^{-1}}}\Big{)}^{0.53}M_{\odot},$ and $\displaystyle M_{\rm BH}$ $\displaystyle=$ $\displaystyle 10^{6.86}\Big{(}\frac{{\rm FWHM(Mg~{}{\sc II})}}{10^{3}\ {\rm km\ s^{-1}}}\Big{)}^{2}\Big{(}\frac{\lambda L_{\lambda}(3000{\rm\AA})}{10^{44}\ {\rm erg\ s^{-1}}}\Big{)}^{0.5}M_{\odot},$ where FWHM(C iv), FWHM(Mg ii), $\lambda L_{\lambda}(1350{\rm\AA})$ and $\lambda L_{\lambda}(3000{\rm\AA})$ are the FWHM of the C iv and Mg ii velocity profile, and the monochromatic luminosity at 1350 Å and 3000 Å, respectively. Note that we use the FWHM of C iv_R + C iv_B as the FWHM of the C iv. We cannot eliminate the possiblility that the estimated SMBH masses are overestimated because the C iv profiles are affected by nucleus outflows (Section 4.1). For estimating the reddening-corrected monochromatic luminosity, we use the optical spectra presented in Section 3.1. We converted the spectra to the rest-frame, de-reddened them with $E(B-V)^{\rm AGN}_{\rm polar\ dust}$ derived in the SED fit, and masked out emission and absorption lines as well as pixels with negative values. Then, we fit a power-law continuum model to the spectra and estimate the monochromatic luminosities from the best fits. The estimated $\lambda L_{\lambda}$(1350) of J0907, J1202, J1207, and J1443 are $(1.54\pm 0.05)\times 10^{45}$, $(9.64\pm 0.27)\times 10^{45}$, $(3.06\pm 0.06)\times 10^{45}$, and $(2.93\pm 0.03)\times 10^{45}$ ${\rm erg\ s^{-1}}$, respectively. The $\lambda L_{\lambda}$(3000) of J0907 is estimated to be $(1.45\pm 0.04)\times 10^{45}$ ${\rm erg\ s^{-1}}$. The resultant SMBH masses are summarized in Table 9. It should be noted that the C iv-based $M_{\rm BH}$ and Mg ii-based $M_{\rm BH}$ of J0907 is not consistent within the statistical error. This is probably attributed to a systematic error especially in the C iv-based $M_{\rm BH}$, known to be accompanied with a large systematic error ($\sim$0.5 dex; see, e.g., Shen 2013). Hereafter we use only the C iv-based $M_{\rm BH}$, since it is measured in all the four BluDOGs. ## 4 Discussion ### 4.1 Spectral features and nuclear outflows We found that the redshifts of the four BluDOGs are in the range of $2.2\lesssim z_{\rm sp}\lesssim 3.3$. They are systematically higher than the typical redshifts of DOGs ($z_{\rm sp}=1.99\pm 0.45$; Dey et al. 2008; Pope et al. 2008). One possible reason for this systematically high redshift is a selection effect related to the blue-excess criterion. When we select BluDOGs from the parent DOG sample, the $g$\- and $r$-band magnitudes show an excess of the expected magnitudes estimated by the power-law extrapolation from $i$-band to $W4$-band. Thus we may select DOGs in a preferred redshift range where strong emission lines such as Ly$\alpha$ and C iv shifts into the two bands (see Section 4.2 for more quantitative assessments). The reason for the underestimated photometric redshift ($\sim 1$; Noboriguchi et al. 2019) is the unusual emission lines with the large REW. The detected emission lines have large velocity widths, $\gtrsim 2000\ {\rm km\ s^{-1}}$ in most cases. This suggests that the broad-line region (BLR) of the BluDOGs is not completely obscured; in other words, the observed BluDOGs are classified as type-1 AGNs. This is an unexpected result, because their very red color between optical and mid-IR suggests the heavily obscured nature. One possible interpretation is that we are looking at a phase where the surrounding dust is just blown away by the nuclear activity (outflow, radiation pressure, or both), as discussed more in Section 4.3. It should be noted that the type-1 nature is seen not only in the presented BluDOGs but also in some other DOGs (e.g., Toba & Nagao, 2016; Toba et al., 2017a; Zou et al., 2020). Systematic spectroscopic observations for the whole populations of DOGs are required to study the nature of obscuration occurring in various populations of DOGs. As shown in Figure 2, the velocity profile of the observed C iv lines show a notable excess feature in the blue wing. Such an excess in the C iv velocity profile has been observed in other type-1 AGNs, and interpreted as a result of powerful nuclear outflows (e.g., Baskin & Laor, 2005; Netzer, 2015; Coatman et al., 2017). To evaluate quantitatively how the nuclear outflow in BluDOGs is strong compared to ordinary AGNs, we examine the “asymmetry parameter ($\alpha_{\beta}$)” defined by De Robertis (1985) as $\displaystyle\alpha_{\beta}$ $\displaystyle=$ $\displaystyle\frac{\lambda_{c}(3/4)-\lambda_{c}(1/4)}{\Delta\lambda(1/2)},$ (3) where $\lambda_{c}(h)$ and $\Delta\lambda(1/2)$ are the central wavelength at which the flux falls to a $h$ time the peak flux and FWHM of the broad profile, respectively. The positive and negative values of $\alpha_{\beta}$ express the blue and red excesses, respectively. The derived values of $\alpha_{\beta}$ for J0907, J1202, J1207, and J1443 are 0.216, 0.102, 0.246, and 0.051, respectively. As a reference, the C iv velocity profile in the composite spectrum of SDSS type-1 quasars given by Vanden Berk et al. (2001) shows $\alpha_{\beta}=0.110$. Thus J0907, and J1207 may possess a significant nuclear outflow that is more powerful than typical quasars. In order to compare $\alpha_{\beta}$ of the BluDOG with that of another dusty AGN population, we fitted the C iv profile of 97 “core” ERQs (ERQs with REW(C iv) $>100$ Å) in Hamann et al. (2017) and measured $\alpha_{\beta}$ by adopting a single or double Gaussian profile. The core ERQ sample consists of 80 objects without BAL and 17 objects with BAL, and we investigate the statistics of $\alpha_{\beta}$ for the two subsamples separately because the BAL feature can affect the C iv line profile. Here we exclude J1443 from the BluDOG sample when comparing the $\alpha_{\beta}$ index because its velocity profile is largely affected by narrow absorption lines (hereafter the limited- BluDOG sample to infer the 3 BluDOGs; i.e., J0907, J1202, and J1207). Figure 6 shows the cumulative fraction of $\alpha_{\beta}$ for the limited-BluDOGs, core ERQs without BAL, and core ERQs with BAL. The averaged values of the limited-BluDOGs, core ERQs without BAL, and core ERQs with BAL are 0.15$\pm$0.08, 0.02$\pm$0.13, and 0.01$\pm$0.09, respectively. We performed the Kolmogorov-Smirnov test (KS-test) to examine the statistical significance of the difference in $\alpha_{\beta}$ among the samples. The p-values of the limited-BluDOGs-core ERQs without BAL, and limited-BluDOG-core ERQs with BAL are 0.0178 and 0.0175, respectively. Thus we conclude that the distributions of $\alpha_{\beta}$ of the limited-BluDOGs and core ERQs with/without BAL are marginally different with $>2$ sigma significance. This suggests that the BluDOGs show nuclear outflow that is possibly more powerful than the nuclear outflow in core ERQs with/without BAL. Figure 6: Cumulative distribution of the $\alpha_{\beta}$ indices for the limited-BluDOGs (see the main text; blue line), core ERQs without BAL (green line), and core ERQs with BAL (orange line). The red dashed line denotes the $\alpha_{\beta}$ index measured for the composite spectrum of SDSS type-1 quasars. Figure 7: Cumulative distribution of the $kt_{80}$ indices for the limited-BluDOGs (blue line), core ERQs without BAL (green line), and core ERQs with BAL (orange line). The red dashed line denotes the $kt_{80}$ index for the single Gaussian profile. We also focus on the kurtosis index ($kt_{80}$) defined as follows (see Hamann et al. 2017 for detailes): $kt_{80}=\Delta v$(80%) $/\Delta v$(20%), where $\Delta v$($x$%) is the velocity width at $x$% of the peak flux height. In addition to $\alpha_{\beta}$, this $kt_{80}$ index is useful to characterize the C iv wing (a more prominent blue wing results in smaller $kt_{80}$). By using the best-fit double Gaussian profile of the BluDOGs, $kt_{80}$ of J0907, J1202, J1207, and J1443 are 0.276, 0.313, 0.252, and 0.440, respectively. Again we exclude J1443 from the BluDOG sample when comparing the $kt_{80}$ index as the discussion of the $\alpha_{\beta}$ index. For comparison, $kt_{80}$ of a single Gaussian is $\sqrt{1-\frac{2\ln(2)}{\ln(5)}}$ ($\sim 0.37$), whereas most quasars have $kt_{80}\sim 0.15$–$0.30$ (see Figure 7 in Hamann et al. 2017). The limited-BluDOGs, core ERQs without BAL, and core ERQs with BAL show $kt_{80}=0.28\pm 0.03$, $0.33\pm 0.06$, and $0.34\pm 0.05$, respectively (see also Figure 7). Note that the C iv profile of 41 core ERQs without BAL and 7 core ERQs with BAL is fitted by a single Gaussian, which is the reason why many objects have $kt_{80}\sim 0.37$ as shown in Figure 7. C iv velocity profiles of core ERQs with/without BAL are roughly consistent with the Gaussian without a blue wing. However, the $kt_{80}$ index of the limited- BluDOGs is less than $\sqrt{1-\frac{2\ln(2)}{\ln(5)}}$, suggesting that their C iv line profile has a wing. We performed the KS-test to examine the statistical significance of the difference in $kt_{80}$ among the samples. The p-values of the limited-BluDOGs-core ERQs without BAL, and limited-BluDOGs- core ERQs with BAL are 0.0637 and 0.0175, respectively. Therefore, we conclude that the distributions of $kt_{80}$ between the samples of the limited-BluDOGs and core ERQs with/without BAL feature are marginally different with $>2$ sigma significance. It has been reported that AGNs with a high Eddington ratio tend to show a Lorentzian-line velocity profile in BLR lines (e.g., Moran et al. 1996; Véron- Cetty et al. 2001; Collin et al. 2006; Zamfir et al. 2010). Therefore, the small $kt_{80}$ value of BluDOGs can be caused by the contribution of extended Lorentzian wings instead of the asymmetric blue wing. For a symmetric Lorentzian profile, $kt_{80}\sim 1/16$ (much smaller than a Gaussian profile, $kt_{80}\sim 0.37$) and $\alpha_{\beta}=0$ are expected. However, the BluDOGs are inconsistent with this expectation (Figure 8). This Figure 8 also shows that the BluDOGs follow the trend made by core ERQs with/without BAL in the $kt_{80}-\alpha_{\beta}$ plane, while a systematic deviation of BluDOGs toward ($\alpha_{\beta}$, $kt_{80}$) = (0, 0) is expected if a Lorentzian component significantly contributes to the C iv line of BluDOGs. Thus, we conclude that extended Lorentzian wings do not affect the C iv line profile of BluDOGs, but the small $kt_{80}$ of BluDOGs is caused by the asymmetric blue excess due to the stronger nuclear outflow than that of ERQs. Figure 8: The $kt_{80}$ vs. $\alpha_{\beta}$ plot for the limited-BluDOGs (blue dots), core ERQs without BAL (green dots), and core ERQs with BAL (orange dots). The black solid line and dashed line denote the $kt_{80}$ values of single Gaussian and single Lorentzian, respectively. Note that 41 core ERQs without BAL and 7 core ERQs with BAL are fitted with a single Gaussian, and they are plotted at ($\alpha_{\beta}$, $kt_{80}$) = (0, $\sqrt{1-\frac{2\ln(2)}{\ln(5)}}$). ### 4.2 Large equivalent widths of the CIV emission As we summarized in Table 7, the blue excess in J1443 can be almost explained by the contribution of the strong emission lines. This is also the case for J1202 by taking into account of the additional contribution of unobserved Ly$\alpha$ to $g$-band. On the other hand, the blue excess of the remaining two BluDOGs cannot be explained only by the contribution of BLR emission lines. Figure 3 strongly suggests that a part of the excess flux comes from the continuum emission, which deviates at $\lesssim$7000Å from the extrapolation of the power-law fit. These results demonstrate the complexity and diversity of BluDOGs; systematic exploration of a larger sample is required to statistically understand the origin of the blue excess. Figure 9: Rest-frame EW ratio vs. rest-frame wavelength. REW ratios are defined as REWs of objects over REWs from the composite spectrum of SDSS type-1 quasar measured by Vanden Berk et al. (2001). The red, blue, green, and orange plots show the REW ratios of J0907, J1202, J1207, and J1443, respectively. Figure 10: Rest-frame EW of the C iv vs. the monochromatic luminosity at 1350 Å. Blue and orange dots represent the four BluDOGs and WISSH quasars (Vietri et al. 2018). The grey 2D histogram represents the number density of the SDSS quasars (Shen et al., 2011). The green line represents the linear fit to the distribution of the SDSS quasars. The red plots show the mean and standard deviation in luminosity bins with 0.5 dex width. The numbers of the SDSS quasars in the individual bins are shown at the bottom of the panel. Not only REW(C iv), but the REW of other BLR emission lines are also systematically larger than observed in typical type-1 quasars (see Tables 3–6, Figure 9, and also Table 2 in Vanden Berk et al. 2001). Such a trend may be explained if the observed BluDOGs have lower UV luminosity than typical quasars owing to the Baldwin effect (Baldwin 1977; Kinney et al. 1990; Baskin & Laor 2004), i.e., the negative correlation between the REWs and the continuum luminosities of quasars. Figure 10 shows the four BluDOGs on the C iv REW vs. $\lambda L_{\lambda}$(1350Å) diagram. Note that the REW of J1443 ($114\pm 24$ Å) is somewhat smaller than that of the remaining three BluDOGs ($148\pm 12$, $203\pm 10$ and $173\pm 9$ Å for J0907, J1202 and J1207, respectively; see Tables 3–6). This is partly because of an underestimation of the C iv flux caused by the absorption features. The figure also shows SDSS type-1 quasars with reliable measurement of C iv REW (`EWCIV/e_EWCIV > 5`) and without broad absorption lines (`BAL < 1`) taken from Shen et al. (2011). Since the Baldwin effect does not significantly depend on redshift (e.g., Croom et al. 2002; Dietrich et al. 2002; Niida et al. 2016), we do not adopt any redshift criterion to select the SDSS quasars so that a wide luminosity range is covered. We also use another comparison sample taken from the WISE/SDSS selected hyper-luminous quasar sample (WISSH; Bischetti et al. 2017; Vietri et al. 2018)555The C iv REW and C iv line luminosity of WISSH quasars are given by Vietri et al. (2018). To calculate $L_{\lambda}$(1350Å) of WISSH quasars, we assume that the continuum spectrum of WISSH quasars is a power-law and adopt the following formula: $\displaystyle L_{\lambda}({\rm 1350\AA})$ $\displaystyle=$ $\displaystyle\frac{L^{\rm line}({\rm C~{}{\sc IV}})}{REW({\rm C~{}{\sc IV}})}\times\Big{(}\frac{1350}{1549}\Big{)}^{\alpha_{\lambda}},$ (4) where $L_{\lambda}({\rm 1350\AA})$, $L^{\rm line}({\rm C~{}{\sc IV}})$, and $\alpha_{\lambda}$ are the monochromatic luminosity at 1350 Å, the line luminosity of C iv, and power-law index, respectively. Here we adopt $\alpha_{\lambda}=-1.7$ (Vanden Berk et al., 2001) as the power-law index. , in order to add objects at the high-luminosity end. Figure 10 clearly shows that the C iv REWs of BluDOGs are larger than the comparison samples at a given UV luminosity. The excess REW over the average relation of the Baldwin effect (shown with a green solid line in Figure 10) for J0907, J1202, J1207, and J1443 are 0.29, 0.66, 0.44, and 0.26 dex, respectively. This excess is larger than the scatter of the comparison samples (see red plots in Figure 10). Therefore the large REW seen in the BluDOGs are not due to the Baldwin effect. The averages and standard deviations of REW(C iv) for core ERQs and ERQ-like objects are $178\pm 74$ and $86\pm 45$ Å, respectively (Hamann et al., 2017). The distributions of REW(C iv) and $(i-W3)_{\rm AB}$ color for BluDOGs are consistent with these of core ERQs although the most of core ERQs and ERQ-like objects do not show a blue-wing profile in C iv (Section 4.1). Hamann et al. (2017) proposed a scenario that the large REW of ERQs are possibly due to the spatially extended geometry of BLRs caused by the powerful nuclear outflow. If the obscuration is heavier for the accretion disk than for the BLRs which have extended geometry, the continuum emission is more heavily extinct than the BLR emission lines and thus the observed-frame EW becomes larger. Such a scenario may also apply to BluDOGs. Unfortunately it is not observationally feasible to confirm this idea by resolving the spatial structure of BLRs in ERQs or BluDOGs due to the required angular resolution, even with the JWST or exisiting ground-based interferometers. Without spatially resolving them, a possible approach is the velocity-resolved reverberation mapping of the geometry and kinematics of BLR clouds (e.g., Horne et al., 2004; Denney et al., 2009; Li et al., 2013; Kollatschny et al., 2014; Pancoast et al., 2014). ### 4.3 Possible extreme accretion and the nature of BluDOGs To understand the nature of BluDOGs especially in the context of the major- merger scenario for the quasar evolution, we compare the SMBH accretion of the four BluDOGs with other AGN populations. Figure 11 is a diagram of $L_{\rm bol}$ vs. the SMBH mass. As in Section 4.2, SDSS quasars (Shen et al., 2011) and WISSH quasars (Vietri et al., 2018) are used as comparison samples. For the SDSS quasars, we select only non-BAL quasars (`BAL < 1`) with the uncertainty of $L_{\rm bol}$ and $M_{\rm BH}$ less than 0.5 dex (`e_logBHCV < 0.5 & e_logLbol < 0.5`), and adopt the C iv-based SMBH mass for a fair comparison with those of the BluDOGs. We also plot samples of 28 ERQs (Perrotta et al., 2019), 5 Hot DOGs (Wu et al., 2018), 2 power-law DOGs (Melbourne et al., 2011), and 1 Compton-thick (CT) DOG (Toba et al., 2020). Hot DOGs are DOGs with a special color of WISE (very faint in the 3.4 $\mu$m and 4.5 $\mu$m bands, but bright in the longer bands; Eisenhardt et al. 2012; Wu et al. 2012), while power-law DOGs are DOGs with a featureless power-law SED from optical to mid-IR (e.g., Dey et al., 2008; Bussmann et al., 2012; Toba et al., 2015; Noboriguchi et al., 2019). The CT DOG was identified by Nuclear Spectroscopic Telescope Array (Harrison et al., 2013) from SDSS-WISE DOGs sample. All but the CT DOG have spectroscopic redshifts. The SMBH masses of the ERQs and the WISSH quasars are estimated from H$\beta$, while those of the Hot DOGs and the DOGs are estimated from H$\alpha$. The SMBH mass of the CT DOG was estimated by Toba et al. (2020) from the stellar mass by using an empirical relation between the stellar mass and SMBH mass (Kormendy & Ho, 2013). Since Perrotta et al. (2019) and Melbourne et al. (2011) did not correct the absorption of dust, the $M_{\rm BH}$ of ERQs and DOGs are lower limits. Figure 11: The SMBH mass vs. the bolometric luminosity diagram. The filled- blue stars and gray contour denote the BluDOGs and SDSS quasars, while the filled hexagons with green, orange, purple, cyan, and light green colors denote ERQs (Perrotta et al., 2019), WISSH quasars (Vietri et al., 2018), Hot DOGs (Wu et al., 2018), DOGs (Melbourne et al., 2011, 2012), and a CT DOG (Toba et al., 2020), respectively. The red dashed lines represent a constant Eddington ratio of $\lambda_{\rm Edd}=$ 0.01, 0.1, 1.0, and 10.0. Figure 11 shows that the four BluDOGs are more luminous than the other AGN populations at a given SMBH mass, or equivalently, they have lower-mass SMBHs than the other AGN populations at a given bolometric luminosity. This suggests that the SMBH growth in the BluDOGs is more rapid than AGNs in comparison samples. Indeed, the Eddington ratios ($\lambda_{\rm Edd}$) of J0907, J1202, J1207, and J1443 are $1.10\pm 0.20$, $3.89\pm 0.78$, $2.19\pm 0.44$, and $1.40\pm 0.23$, respectively (Table 9), with the average value of 2.26. In other words, the SMBHs in the BluDOGs are now in the stage of the Eddington- limit or super-Eddington accretion. Even if the intrinsic SMBH masses are lower than those estimated (Section 3.4), the conclusion of this study remains qualitatively unchanged. The higher Eddington ratios compared to other populations suggest that the SMBHs in BluDOGs are in the most rapidly evolving phase during the whole evolutionary history of SMBHs. In the gas-rich major merger scenario of Hopkins et al. (2008), the peak of the AGN activity (i.e., the mass growth of SMBHs) corresponds to the transition phase from the optically thick to optically thin quasars, where the surrounding dust is blown out by the powerful AGN activity. Note that optically thick quasars in the major merger scenario should be recognized as type-2 quasars in optical (the BLR cannot be observed due to the heavy dust reddening). Since optically-thick quasars in the final stage of the evolution can be recognized as both type-1 and type-2 due to the orientation effect toward the dusty torus, the observed type-1 nature suggests the object is not in the early (optically-thick) stage in the major merger evolutionary scenario. Preferentially in AGNs with high $\lambda_{\rm Edd}$, a blue-wing feature tends to be observed (e.g., Aoki et al. 2005; Komossa et al. 2008). The observed characteristics of the BluDOGs such as the type-1 nature and the blue-wing feature of the C iv velocity profile are consistent with the picture that BluDOGs are in such a peak stage of the SMBH evolution. To discuss the evolutionary relation among populations of dusty galaxies (BluDOGs, core ERQs, and hot DOGs), we focus on $E(B-V)$ and $kt_{80}$. $E(B-V)$ for BluDOGs, core ERQs (Hamann et al., 2017), and hot DOGs (Wu et al., 2018) are $0.273\pm 0.049$, $0.242\pm 0.127$, and $4.781\pm 1.986$, respectively. The $E(B-V)$ of the hot DOGs is significantly larger than that of the BluDOG and core ERQ samples, suggesting the hot DOGs are thought to be in a heavily obscured phase. Since the $kt_{80}$ of BluDOGs is smaller than that of core ERQs (Section 4.1), and the $kt_{80}$ of Mid-IR detected quasars is close to that of BluDOGs (Figure 1 of Monadi & Bird 2022), the BluDOG phase is thought to be close to the optically-thin quasar phase. Therefore, it is suggested that the evolutionary path of various AGN populations is “Hot DOGs – core ERQs – BluDOGs – optically-thin quasars”. For AGNs in general, the mass accretion efficiency ($\eta$) is defined as following: $\displaystyle\eta$ $\displaystyle=$ $\displaystyle\frac{L_{\rm bol}}{\dot{M}c^{2}},$ (5) where $\dot{M}$ is the mass accretion rate. By multiplying the $\dot{M}$ and lifetime of BluDOGs ($t_{\rm life}$), we can roughly estimate the accreted mass ($M_{\rm Acc}$) in the BluDOGs phase. Bian & Zhao (2003) estimated $\log\eta=-1.61$ of Seyfert 1 galaxies and Palomar-Green quasars by assuming that the geometrically-thin and optically-thick standard $\alpha$-prescription accretion disk model (Shakura & Sunyaev 1973). By assuming $\log\eta=-1.61$ and $t_{\rm life}=1$ Myr (Noboriguchi et al., 2019), the estimated $M_{\rm Acc}$ of J0907, J1202, J1207, and J1443 are about $1.68\times 10^{7}$, $1.68\times 10^{8}$, $2.19\times 10^{7}$, and $6.93\times 10^{7}\ M_{\odot}$. The SMBH masses reached when the SMBH masses of the BluDOGs are increased by the observed mass accretion rate during the typical lifetime of BluDOGs ($M^{+}_{\rm BH}=M_{\rm BH}+M_{\rm Acc}$) of J0907, J1202, J1207, and J1443 are $1.86\times 10^{8}$, $6.63\times 10^{8}$, $1.32\times 10^{8}$, and $6.17\times 10^{8}\ M_{\odot}$, respectively. Therefore the SMBH mass of BluDOGs increases by $\sim$20% during the short BluDOG phase, suggesting that BluDOGs are actually in a rapidly glowing phase. Figure 12 shows the Eddington ratios of various populations of AGNs as a function of redshift. The excess of $\lambda_{\rm Edd}$ of the four BluDOGs is more significant than the scatter of the $\lambda_{\rm Edd}$ distribution, suggesting that BluDOGs are a special class of AGNs that harbor SMBHs in the most actively evolving phase. Then, why such a class of AGNs is found only in a limited redshift range, $2.2<z_{\rm sp}<3.3$? A possible reason comes from their selection criteria, as briefly mentioned in Sec 4.1. Since the BluDOGs are selected by the blue excesses which are largely caused by the contribution of strong BLR emission lines, the resultant redshift distribution would be biased such that the blue bands contain strong emission lines. It is also not clear whether the whole population of DOGs have systematically larger $\lambda_{\rm Edd}$ than ordinary type-1 quasars, due to the paucity of the spectroscopic data. In order to reveal the total picture of the dust- enshrouded evolution of SMBHs, more systematic spectroscopic observations for various populations of BluDOGs and DOGs are needed. Figure 12: The redshift vs. the Eddington ratio diagram. The filled-blue stars and hexagons are the same as in Figure 11. The gray 2D histogram represents the number density of SDSS quasars (Shen et al., 2011). The red plots show the mean and standard deviation of $\lambda_{\rm Edd}$ in redshift bins with the width of $\Delta z=0.5$. The numbers shown at the upper part denote the numbers of SDSS quasars in the redshift bins. ## 5 CONCLUSION We carried out spectroscopic observations of the four BluDOGs selected by Noboriguchi et al. (2019) using Subaru/FOCAS and VLT/FORS2. The analysis of the obtained spectroscopic data revealed the following spectroscopic properties of the BluDOGs: 1. 1. The rest-frame UV spectra of the BluDOGs show broad ($\gtrsim$2000 ${\rm km\ s^{-1}}$) emission lines. This suggests that the BLRs of the BluDOGs are not completely obscured, albeit the very dusty nature inferred from their optical- IR SED. 2. 2. The C iv lines of the BluDOGs show a significant blue wing, which is more prominent than in ordinary SDSS type-1 quasars. This suggests a presence of powerful nuclear outflow at the spatial scale of the BLR in the BluDOGs. 3. 3. The REWs of their BLR lines are very large, REW(C iv)$\sim$160 Å, $\sim$7 times larger than the average of SDSS type-1 quasars. Such strong lines cause the flux excess of the two BluDOGs in the HSC $g$\- and $r$-bands, while blue continuum emission also contributes the blue excess in the remaining two objects. The large REWs are not explained by the Baldwin effect. A possible origin is a powerful nuclear outflow in the BluDOGs causing a selective obscuration of the nuclear region, as suggested for ERQs. 4. 4. The Eddington ratios of the BluDOGs are higher than 1.0 and are systematically larger than other AGN populations. The mass accretion onto the SMBH in BluDOGs is in the mode of the Eddington-limit or super-Eddington accretion. All of the above results support the scenario that BluDOGs represent a population of AGNs in the transition phase from optically thick to optically thin quasars, i.e., in the blowing-out phase of the major-merger scenario for the SMBH evolution. The spectroscopic properties of the BluDOGs are similar to those of ERQs. For further understandings of the complete picture, more systematic spectroscopic observations are crucial, not only of BluDOGs but also of the whole population of DOGs. The authors gratefully acknowledge the anonymous referee for a careful reading of the manuscript and very helpful comments. This study is based on data collected at Subaru Telescope, which is operated by the National Astronomical Observatory of Japan, and observations collected at the European Southern Observatory under a ESO programme 0102.B-0614(A). We are honored and grateful for the opportunity of observing the Universe from Maunakea, which has the cultural, historical and natural significance in Hawaii. The Hyper Suprime-Cam (HSC) collaboration includes the astronomical communities of Japan and Taiwan, and Princeton University. The HSC instrumentation and software were developed by the National Astronomical Observatory of Japan (NAOJ), the Kavli Institute for the Physics and Mathematics of the Universe (Kavli IPMU), the University of Tokyo, the High Energy Accelerator Research Organization (KEK), the Academia Sinica Institute for Astronomy and Astrophysics in Taiwan (ASIAA), and Princeton University. Funding was contributed by the FIRST program from Japanese Cabinet Office, the Ministry of Education, Culture, Sports, Science and Technology (MEXT), the Japan Society for the Promotion of Science (JSPS), Japan Science and Technology Agency (JST), the Toray Science Foundation, NAOJ, Kavli IPMU, KEK, ASIAA, and Princeton University. This paper makes use of software developed for the Large Synoptic Survey Telescope. We thank the LSST Project for making their code available as free software at http://dm.lsstcorp.org. The Pan-STARRS1 Surveys (PS1) have been made possible through contributions of the Institute for Astronomy, the University of Hawaii, the Pan-STARRS Project Office, the Max-Planck Society and its participating institutes, the Max Planck Institute for Astronomy, Heidelberg and the Max Planck Institute for Extraterrestrial Physics, Garching, The Johns Hopkins University, Durham University, the University of Edinburgh, Queen’s University Belfast, the Harvard-Smithsonian Center for Astrophysics, the Las Cumbres Observatory Global Telescope Network Incorporated, the National Central University of Taiwan, the Space Telescope Science Institute, the National Aeronautics and Space Administration under Grant No. NNX08AR22G issued through the Planetary Science Division of the NASA Science Mission Directorate, the National Science Foundation under Grant No. AST-1238877, the University of Maryland, and Eotvos Lorand University (ELTE) This publication has made use of data from the VIKING survey from VISTA at the ESO Paranal Observatory, program ID 179.A-2004. Data processing has been contributed by the VISTA Data Flow System at CASU, Cambridge and WFAU, Edinburgh. This publication makes use of data products from the Wide-field Infrared Survey Explorer, which is a joint project of the University of California, Los Angeles, and the Jet Propulsion Laboratory/California Institute of Technology, funded by the National Aeronautics and Space Administration. Herschel is an ESA space observatory with science instruments provided by European-led Principal Investigator consortia and with important participation from NASA. This research made use of Astropy,666http://www.astropy.org a community- developed core Python package for Astronomy (Astropy Collaboration et al., 2013, 2018). We would like to thank Enago (www.enago.jp) for the English language review and Kohei Iwashita for helpful discussion. This study was financially supported by the Japan Society for the Promotion of Science (JSPS) KAKENHI grant Nos. 19J10458, 21H01126 (A.N.), 16H03958, 17H01114, 19H00697, 20H01949 (T.N.), 18J01050, 19K14759, 22H01266 (Y. Toba), 20H01939 (K.I.), and 20K04014 (Y. Terashima). ## References * Aihara et al. (2018) Aihara, H., Arimoto, N., Armstrong, R., et al. 2018, PASJ, 70, S4, doi: 10.1093/pasj/psx066 * Aoki et al. (2005) Aoki, K., Kawaguchi, T., & Ohta, K. 2005, ApJ, 618, 601, doi: 10.1086/426075 * Appenzeller et al. (1998) Appenzeller, I., Fricke, K., Fürtig, W., et al. 1998, The Messenger, 94, 1 * Arnaboldi et al. (2007) Arnaboldi, M., Neeser, M. J., Parker, L. C., et al. 2007, The Messenger, 127 * Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33, doi: 10.1051/0004-6361/201322068 * Astropy Collaboration et al. (2018) Astropy Collaboration, Price-Whelan, A. M., Sipőcz, B. M., et al. 2018, AJ, 156, 123, doi: 10.3847/1538-3881/aabc4f * Baldwin (1977) Baldwin, J. A. 1977, ApJ, 214, 679, doi: 10.1086/155294 * Baskin & Laor (2004) Baskin, A., & Laor, A. 2004, MNRAS, 350, L31, doi: 10.1111/j.1365-2966.2004.07833.x * Baskin & Laor (2005) —. 2005, MNRAS, 356, 1029, doi: 10.1111/j.1365-2966.2004.08525.x * Bian & Zhao (2003) Bian, W.-H., & Zhao, Y.-H. 2003, PASJ, 55, 599, doi: 10.1093/pasj/55.3.599 * Bischetti et al. (2017) Bischetti, M., Piconcelli, E., Vietri, G., et al. 2017, A&A, 598, A122, doi: 10.1051/0004-6361/201629301 * Boquien et al. (2019) Boquien, M., Burgarella, D., Roehlly, Y., et al. 2019, A&A, 622, A103, doi: 10.1051/0004-6361/201834156 * Bourne et al. (2016) Bourne, N., Dunne, L., Maddox, S. J., et al. 2016, MNRAS, 462, 1714, doi: 10.1093/mnras/stw1654 * Bruzual & Charlot (2003) Bruzual, G., & Charlot, S. 2003, MNRAS, 344, 1000, doi: 10.1046/j.1365-8711.2003.06897.x * Burgarella et al. (2005) Burgarella, D., Buat, V., & Iglesias-Páramo, J. 2005, MNRAS, 360, 1413, doi: 10.1111/j.1365-2966.2005.09131.x * Bussmann et al. (2009) Bussmann, R. S., Dey, A., Borys, C., et al. 2009, ApJ, 705, 184, doi: 10.1088/0004-637X/705/1/184 * Bussmann et al. (2011) Bussmann, R. S., Dey, A., Lotz, J., et al. 2011, ApJ, 733, 21, doi: 10.1088/0004-637X/733/1/21 * Bussmann et al. (2012) Bussmann, R. S., Dey, A., Armus, L., et al. 2012, ApJ, 744, 150, doi: 10.1088/0004-637X/744/2/150 * Calzetti et al. (2000) Calzetti, D., Armus, L., Bohlin, R. C., et al. 2000, ApJ, 533, 682, doi: 10.1086/308692 * Cardelli et al. (1989) Cardelli, J. A., Clayton, G. C., & Mathis, J. S. 1989, ApJ, 345, 245, doi: 10.1086/167900 * Chabrier (2003) Chabrier, G. 2003, PASP, 115, 763, doi: 10.1086/376392 * Ciesla et al. (2015) Ciesla, L., Charmandaris, V., Georgakakis, A., et al. 2015, A&A, 576, A10, doi: 10.1051/0004-6361/201425252 * Coatman et al. (2017) Coatman, L., Hewett, P. C., Banerji, M., et al. 2017, MNRAS, 465, 2120, doi: 10.1093/mnras/stw2797 * Collin et al. (2006) Collin, S., Kawaguchi, T., Peterson, B. M., & Vestergaard, M. 2006, A&A, 456, 75, doi: 10.1051/0004-6361:20064878 * Croom et al. (2002) Croom, S. M., Rhook, K., Corbett, E. A., et al. 2002, MNRAS, 337, 275, doi: 10.1046/j.1365-8711.2002.05910.x * Cutri (2014) Cutri, R. M. 2014, VizieR Online Data Catalog, 2328 * Dale et al. (2014) Dale, D. A., Helou, G., Magdis, G. E., et al. 2014, ApJ, 784, 83, doi: 10.1088/0004-637X/784/1/83 * Dawson et al. (2013) Dawson, K. S., Schlegel, D. J., Ahn, C. P., et al. 2013, AJ, 145, 10, doi: 10.1088/0004-6256/145/1/10 * De Robertis (1985) De Robertis, M. 1985, ApJ, 289, 67, doi: 10.1086/162865 * Denney et al. (2009) Denney, K. D., Peterson, B. M., Pogge, R. W., et al. 2009, ApJ, 704, L80, doi: 10.1088/0004-637X/704/2/L80 * Desai et al. (2009) Desai, V., Soifer, B. T., Dey, A., et al. 2009, ApJ, 700, 1190, doi: 10.1088/0004-637X/700/2/1190 * Dey et al. (2008) Dey, A., Soifer, B. T., Desai, V., et al. 2008, ApJ, 677, 943, doi: 10.1086/529516 * Dietrich et al. (2002) Dietrich, M., Hamann, F., Shields, J. C., et al. 2002, ApJ, 581, 912, doi: 10.1086/344410 * Ding et al. (2020) Ding, X., Silverman, J., Treu, T., et al. 2020, ApJ, 888, 37, doi: 10.3847/1538-4357/ab5b90 * Eales et al. (2010) Eales, S., Dunne, L., Clements, D., et al. 2010, PASP, 122, 499, doi: 10.1086/653086 * Eisenhardt et al. (2012) Eisenhardt, P. R. M., Wu, J., Tsai, C.-W., et al. 2012, ApJ, 755, 173, doi: 10.1088/0004-637X/755/2/173 * Eisenstein et al. (2011) Eisenstein, D. J., Weinberg, D. H., Agol, E., et al. 2011, AJ, 142, 72, doi: 10.1088/0004-6256/142/3/72 * Feibelman (1983) Feibelman, W. A. 1983, A&A, 122, 335 * Ferrarese & Merritt (2000) Ferrarese, L., & Merritt, D. 2000, ApJ, 539, L9, doi: 10.1086/312838 * Fiore et al. (2008) Fiore, F., Grazian, A., Santini, P., et al. 2008, ApJ, 672, 94, doi: 10.1086/523348 * Freudling et al. (2013) Freudling, W., Romaniello, M., Bramich, D. M., et al. 2013, A&A, 559, A96, doi: 10.1051/0004-6361/201322494 * Gebhardt et al. (2000) Gebhardt, K., Bender, R., Bower, G., et al. 2000, ApJ, 539, L13, doi: 10.1086/312840 * Goulding et al. (2018) Goulding, A. D., Greene, J. E., Bezanson, R., et al. 2018, PASJ, 70, S37, doi: 10.1093/pasj/psx135 * Griffin et al. (2010) Griffin, M. J., Abergel, A., Abreu, A., et al. 2010, A&A, 518, L3, doi: 10.1051/0004-6361/201014519 * Hamann et al. (2017) Hamann, F., Zakamska, N. L., Ross, N., et al. 2017, MNRAS, 464, 3431, doi: 10.1093/mnras/stw2387 * Harrison et al. (2013) Harrison, F. A., Craig, W. W., Christensen, F. E., et al. 2013, ApJ, 770, 103, doi: 10.1088/0004-637X/770/2/103 * Hopkins et al. (2008) Hopkins, P. F., Hernquist, L., Cox, T. J., & Kereš, D. 2008, ApJS, 175, 356, doi: 10.1086/524362 * Horne et al. (2004) Horne, K., Peterson, B. M., Collier, S. J., & Netzer, H. 2004, PASP, 116, 465, doi: 10.1086/420755 * Inoue (2011) Inoue, A. K. 2011, MNRAS, 415, 2920, doi: 10.1111/j.1365-2966.2011.18906.x * Kashikawa et al. (2002) Kashikawa, N., Aoki, K., Asai, R., et al. 2002, PASJ, 54, 819, doi: 10.1093/pasj/54.6.819 * Kawanomoto et al. (2018) Kawanomoto, S., Uraguchi, F., Komiyama, Y., et al. 2018, PASJ, 70, 66, doi: 10.1093/pasj/psy056 * Kinney et al. (1990) Kinney, A. L., Rivolo, A. R., & Koratkar, A. P. 1990, ApJ, 357, 338, doi: 10.1086/168924 * Kollatschny et al. (2014) Kollatschny, W., Ulbrich, K., Zetzl, M., Kaspi, S., & Haas, M. 2014, A&A, 566, A106, doi: 10.1051/0004-6361/201423901 * Komossa et al. (2008) Komossa, S., Xu, D., Zhou, H., Storchi-Bergmann, T., & Binette, L. 2008, ApJ, 680, 926, doi: 10.1086/587932 * Kormendy & Ho (2013) Kormendy, J., & Ho, L. C. 2013, ARA&A, 51, 511, doi: 10.1146/annurev-astro-082708-101811 * Leitherer et al. (2002) Leitherer, C., Li, I. H., Calzetti, D., & Heckman, T. M. 2002, ApJS, 140, 303, doi: 10.1086/342486 * Li et al. (2013) Li, Y.-R., Wang, J.-M., Ho, L. C., Du, P., & Bai, J.-M. 2013, ApJ, 779, 110, doi: 10.1088/0004-637X/779/2/110 * Magorrian et al. (1998) Magorrian, J., Tremaine, S., Richstone, D., et al. 1998, AJ, 115, 2285, doi: 10.1086/300353 * Marconi & Hunt (2003) Marconi, A., & Hunt, L. K. 2003, ApJ, 589, L21, doi: 10.1086/375804 * McCully & Tewes (2019) McCully, C., & Tewes, M. 2019, Astro-SCRAPPY: Speedy Cosmic Ray Annihilation Package in Python. http://ascl.net/1907.032 * Melbourne et al. (2011) Melbourne, J., Peng, C. Y., Soifer, B. T., et al. 2011, AJ, 141, 141, doi: 10.1088/0004-6256/141/4/141 * Melbourne et al. (2012) Melbourne, J., Soifer, B. T., Desai, V., et al. 2012, AJ, 143, 125, doi: 10.1088/0004-6256/143/5/125 * Miyazaki et al. (2018) Miyazaki, S., Komiyama, Y., Kawanomoto, S., et al. 2018, PASJ, 70, S1, doi: 10.1093/pasj/psx063 * Monadi & Bird (2022) Monadi, R., & Bird, S. 2022, MNRAS, 511, 3501, doi: 10.1093/mnras/stac294 * Moran et al. (1996) Moran, E. C., Halpern, J. P., & Helfand, D. J. 1996, ApJS, 106, 341, doi: 10.1086/192341 * Netzer (2015) Netzer, H. 2015, ARA&A, 53, 365, doi: 10.1146/annurev-astro-082214-122302 * Niida et al. (2016) Niida, M., Nagao, T., Ikeda, H., et al. 2016, ApJ, 832, 208, doi: 10.3847/0004-637X/832/2/208 * Noboriguchi et al. (2019) Noboriguchi, A., Nagao, T., Toba, Y., et al. 2019, ApJ, 876, 132, doi: 10.3847/1538-4357/ab1754 * Noll et al. (2009) Noll, S., Burgarella, D., Giovannoli, E., et al. 2009, A&A, 507, 1793, doi: 10.1051/0004-6361/200912497 * Pancoast et al. (2014) Pancoast, A., Brewer, B. J., Treu, T., et al. 2014, MNRAS, 445, 3073, doi: 10.1093/mnras/stu1419 * Perrotta et al. (2019) Perrotta, S., Hamann, F., Zakamska, N. L., et al. 2019, MNRAS, 488, 4126, doi: 10.1093/mnras/stz1993 * Pilbratt et al. (2010) Pilbratt, G. L., Riedinger, J. R., Passvogel, T., et al. 2010, A&A, 518, L1, doi: 10.1051/0004-6361/201014759 * Poglitsch et al. (2010) Poglitsch, A., Waelkens, C., Geis, N., et al. 2010, A&A, 518, L2, doi: 10.1051/0004-6361/201014535 * Pope et al. (2008) Pope, A., Bussmann, R. S., Dey, A., et al. 2008, ApJ, 689, 127, doi: 10.1086/592739 * Ross et al. (2015) Ross, N. P., Hamann, F., Zakamska, N. L., et al. 2015, MNRAS, 453, 3932, doi: 10.1093/mnras/stv1710 * Sanders et al. (1988) Sanders, D. B., Soifer, B. T., Elias, J. H., et al. 1988, ApJ, 325, 74, doi: 10.1086/165983 * Schlegel et al. (1998) Schlegel, D. J., Finkbeiner, D. P., & Davis, M. 1998, ApJ, 500, 525, doi: 10.1086/305772 * Shakura & Sunyaev (1973) Shakura, N. I., & Sunyaev, R. A. 1973, A&A, 500, 33 * Shen (2013) Shen, Y. 2013, Bulletin of the Astronomical Society of India, 41, 61. https://arxiv.org/abs/1302.2643 * Shen et al. (2011) Shen, Y., Richards, G. T., Strauss, M. A., et al. 2011, ApJS, 194, 45, doi: 10.1088/0067-0049/194/2/45 * Stalevski et al. (2016) Stalevski, M., Ricci, C., Ueda, Y., et al. 2016, MNRAS, 458, 2288, doi: 10.1093/mnras/stw444 * Toba et al. (2017a) Toba, Y., Bae, H.-J., Nagao, T., et al. 2017a, ApJ, 850, 140, doi: 10.3847/1538-4357/aa918a * Toba & Nagao (2016) Toba, Y., & Nagao, T. 2016, ApJ, 820, 46, doi: 10.3847/0004-637X/820/1/46 * Toba et al. (2015) Toba, Y., Nagao, T., Strauss, M. A., et al. 2015, PASJ, 67, 86, doi: 10.1093/pasj/psv057 * Toba et al. (2017b) Toba, Y., Nagao, T., Kajisawa, M., et al. 2017b, ApJ, 835, 36, doi: 10.3847/1538-4357/835/1/36 * Toba et al. (2019) Toba, Y., Yamashita, T., Nagao, T., et al. 2019, ApJS, 243, 15, doi: 10.3847/1538-4365/ab238d * Toba et al. (2020) Toba, Y., Yamada, S., Ueda, Y., et al. 2020, ApJ, 888, 8, doi: 10.3847/1538-4357/ab5718 * Toba et al. (2022) Toba, Y., Liu, T., Urrutia, T., et al. 2022, A&A, 661, A15, doi: 10.1051/0004-6361/202141547 * Treister et al. (2012) Treister, E., Schawinski, K., Urry, C. M., & Simmons, B. D. 2012, ApJ, 758, L39, doi: 10.1088/2041-8205/758/2/L39 * Tremaine et al. (2002) Tremaine, S., Gebhardt, K., Bender, R., et al. 2002, ApJ, 574, 740, doi: 10.1086/341002 * Valiante et al. (2016) Valiante, E., Smith, M. W. L., Eales, S., et al. 2016, MNRAS, 462, 3146, doi: 10.1093/mnras/stw1806 * van Dokkum (2001) van Dokkum, P. G. 2001, PASP, 113, 1420, doi: 10.1086/323894 * Vanden Berk et al. (2001) Vanden Berk, D. E., Richards, G. T., Bauer, A., et al. 2001, AJ, 122, 549, doi: 10.1086/321167 * Véron-Cetty et al. (2001) Véron-Cetty, M. P., Véron, P., & Gonçalves, A. C. 2001, A&A, 372, 730, doi: 10.1051/0004-6361:20010489 * Vestergaard & Osmer (2009) Vestergaard, M., & Osmer, P. S. 2009, ApJ, 699, 800, doi: 10.1088/0004-637X/699/1/800 * Vestergaard & Peterson (2006) Vestergaard, M., & Peterson, B. M. 2006, ApJ, 641, 689, doi: 10.1086/500572 * Vietri et al. (2018) Vietri, G., Piconcelli, E., Bischetti, M., et al. 2018, A&A, 617, A81, doi: 10.1051/0004-6361/201732335 * Villar Martín et al. (2020) Villar Martín, M., Perna, M., Humphrey, A., et al. 2020, A&A, 634, A116, doi: 10.1051/0004-6361/201937086 * Wright et al. (2010) Wright, E. L., Eisenhardt, P. R. M., Mainzer, A. K., et al. 2010, AJ, 140, 1868, doi: 10.1088/0004-6256/140/6/1868 * Wu et al. (2012) Wu, J., Tsai, C.-W., Sayers, J., et al. 2012, ApJ, 756, 96, doi: 10.1088/0004-637X/756/1/96 * Wu et al. (2018) Wu, J., Jun, H. D., Assef, R. J., et al. 2018, ApJ, 852, 96, doi: 10.3847/1538-4357/aa9ff3 * Yang et al. (2020) Yang, G., Boquien, M., Buat, V., et al. 2020, MNRAS, 491, 740, doi: 10.1093/mnras/stz3001 * York et al. (2000) York, D. G., Adelman, J., Anderson, Jr., J. E., et al. 2000, AJ, 120, 1579, doi: 10.1086/301513 * Zamfir et al. (2010) Zamfir, S., Sulentic, J. W., Marziani, P., & Dultzin, D. 2010, MNRAS, 403, 1759, doi: 10.1111/j.1365-2966.2009.16236.x * Zou et al. (2020) Zou , F., Brandt, W. N., Vito, F., et al. 2020, MNRAS, 499, 1823, doi: 10.1093/mnras/staa2930
# DiProber: Using Dual Probing to Estimate Tor Relay Capacities in Underloaded Networks Hussein Darir University of Illinois at Urbana-Champaign <EMAIL_ADDRESS>Nikita Borisov University of Illinois at Urbana-Champaign <EMAIL_ADDRESS>Geir Dullerud University of Illinois at Urbana-Champaign <EMAIL_ADDRESS> ###### Abstract Tor is the most popular anonymous communication network. It has millions of daily users seeking privacy while browsing the internet. It has thousands of relays to route and anonymize the source and destinations of the users packets. To create a path, Tor authorities generate a probability distribution over relays based on the estimates of the capacities of the relays. An incoming user will then sample this probability distribution and choose three relays for their paths. The estimates are based on the bandwidths of observation probes the authority assigns to each relay in the network. Thus, in order to achieve better load balancing between users, accurate estimates are necessary. Unfortunately, the currently implemented estimation algorithm generate inaccurate estimates causing the network to be under utilized and its capacities unfairly distributed between the users paths. We propose $\mathit{DiProber}$, a new relay capacity estimation algorithm. The algorithm proposes a new measurement scheme in Tor consisting of two probes per relay and uses maximum likelihood to estimate their capacities. We show that the new technique works better in the case of under-utilized networks where users tend to have very low demand on the Tor network. ## I Introduction Tor [7] is the most popular anonymous communication network, with several million estimated daily users [17]. It offers users a way to communicate online while preserving their identity and relationship to third parties. Tor operates by using a network of volunteer _relays_ to forward an encrypted version of users’ traffic in order to obscure the source and/or the destination of network traffic. To ensure consistent performance, users’ traffic needs to be load-balanced across the relays. The network capacities of the relays are quite heterogeneous, spanning many orders of magnitude; an accurate estimate of these capacities is an important input to the load- balancing process. Initially Tor relied on relays’ own measurements of their own capacities to generate estimates ; this, however, created the possibility of low-resource attacks on the Tor network [2]. This motivated the development of $\mathit{TorFlow}$, a bandwidth monitoring system [12]. $\mathit{TorFlow}$ uses external probes to monitor the performance of individual relays and uses this value to adjust the bandwidth value reported by the relay itself. The capacity estimates produced by $\mathit{TorFlow}$, however, vary considerably over time and between different $\mathit{TorFlow}$ instance, impacting the traffic allocation algorithm’s ability to properly balance load across relays. A new estimation algorithm based on maximum likelihood estimation was also developed, $\mathit{MLEFlow}$ [5]. While the estimation accuracy of this new algorithm showed a lot of promises, the probabilistic model used to derive the estimator had many simplifying assumptions that are not necessarily true in practise. Notably, the model used assumed that users utilize all the bandwidth allocated to them while in practise users’ demand fluctuate and is generally lower than the available bandwidth. Moreover, while the estimation algorithm used had significantly lower estimation error for exit relays, it still led to relatively high error for guard and middle relays. We propose a new maximum likelihood estimator based on the work of $\mathit{MLEFlow}$, where we relax the assumption aforementioned about clients usage of the available bandwidth. The developed algorithm proposes a new measurement scheme of two probes per relay and uses the results of both measurements to identify whether a relay is bottlenecked or not during each epoch. Then, depending on the case, the algorithm uses a distinct probabilistic model relating measurements to actual capacity and performs maximum likelihood estimation. We derive analytical bounds of convergence for the estimates. We then validate the results of our analysis in flow-based Python simulations, where we simulate both loaded and under-loaded networks and show the benefits of using the new measurements scheme. ## II Path allocation in Tor The current Tor network consists of around $6000$ relays [16] that are used to forward user traffic. To create a connection, a user chooses a _path_ of three different relays to construct a circuit that forwards traffic in both directions. Only the user knows the entire path; the relays know only their predecessor and successor, obscuring the relationship between clients and destinations. The traffic is also encrypted / decrypted at each node to hide the correspondence between incoming and outgoing traffic from a network observer. Figure 1: Tor relay selection and capacity estimation. Users select three relays from collections $G$, $G\cup M$, and $E$ using respective weight vectors $w_{t}^{g},w_{t}^{m}$, and $w_{t}^{e}$, to form a _path_ (black arrows). The bandwidth authority collects measurements $m_{t}$ of each relay (red arrows), and updates the capacity estimates $C_{t+1}$ in the next consensus document, which are then used to generate new weight vectors $w^{x}_{t+1}$. Relays in Tor have heterogeneous capabilities and have network capacity111By “capacity” we refer to the smaller of upload and download bandwidth limit on the relay. This may be imposed by the ISP, the network configuration, or manually configured by the relay operator. In some cases, there may exist other bottlenecks on the path between two relays but a per-node bandwidth limit is a common and useful model of network capacity constraints. sizes that differ by orders of magnitude (see Figure 5). Relays also have different capabilities and can be divided into three classes: _exits_ , which can be used in the last position of the path, _guards_ , which can be used in the first or second position, and _middles_ which can only be used in the second position [15]. We denote the corresponding sets of relays by $E$, $G$, and $M$, respectively. To create a path, nodes are sampled from these sets with a probability proportional to their estimated capacity. For example, if we define $C[j]$ to be the estimated capacity of relay $j$, then the probability of choosing relay $j\in E$ as the last node in a path is $w^{e}[j]=C[j]/\left(\sum_{j^{\prime}\in E}C[j^{\prime}]\right)$; likewise for guard nodes being chosen in the first position. The middle position can be chosen from both guard and middle nodes; to balance bandwidth among classes, guard node capacity is adjusted by a multiplier $W_{mg}$; i.e., a guard node $j\in G$ is chosen for the middle position with probability: $w^{m}[j]=\frac{W_{mg}C[j]}{\sum_{j^{\prime}\in G}W_{mg}C[j^{\prime}]+\sum_{j^{\prime}\in M}C[j^{\prime}]}$ The multiplier is computed as: $W_{mg}=\frac{\sum_{j^{\prime}\in G}C[j^{\prime}]-\sum_{j^{\prime}\in M}C[j]}{2\sum_{j^{\prime}\in G}C[j^{\prime}]}$ This is a somewhat simplified presentation that describes the scenario where exit bandwidth is scarce and there is more guard bandwidth than middle bandwidth, as is the case in the actual Tor network. (See the Tor Directory Specification for more details on how other cases would be handled [15].) See Figure 1 for a description of this process. It is easy to see that, in this scenario, if the estimated capacities are equal to the true relay capacities, which we will call $C^{*}[j]$, the expected number of paths using each exit relay will be proportional to its bandwidth; likewise, the expected number of paths using each guard and middle node will be proportional to their bandwidth. Using $X[j]$ to denote the number of paths on relay $j$, we have: $\displaystyle E[X[j]]/C^{*}[j]=E[X[j^{\prime}]]/C^{*}[j^{\prime}]\quad$ for $j,j^{\prime}\in E$ $\displaystyle E[X[j]]/C^{*}[j]=E[X[j^{\prime}]]/C^{*}[j^{\prime}]\quad$ for $j,j^{\prime}\in G\cup M$ Thus, in expectation, each path would have the same bandwidth—$C^{*}[j]/E[X[j]]$ for $j\in E$. Our goal is therefore to estimate these capacities as accurately as possible.222Note that some research suggests allocation other than proportional to bandwidth results in better performance [14, 8]; nevertheless, an accurate capacity estimate is still needed for these alternative path allocation strategies. ### II-A Capacity Estimation Each relay estimates its own network capacity by computing the maximum sustained download and upload bandwidth over a 5-second period over the last 5 days. It reports this value (called the _observed bandwidth_) to directory authorities, who then compile it across all relays and distribute the information to the clients in a _consensus_ document, published every hour. We will use $b_{t}[j]$ to refer to the observed bandwidth of relay $j$ in the consensus document published at time $t$. Using the observed bandwidth directly for load-balancing creates the opportunity for a low-resource attack on the Tor network [2]. In particular, a relay can publish a high observed bandwidth for itself, which will cause more clients to choose it, and create more chances for it to break users anonymity. This motivated the design of $\mathit{TorFlow}$ [12], which used observations of actual relay performance to estimate capacities, rather than simply trusting the value reported by the relay itself. In $\mathit{TorFlow}$, a _bandwidth authority_ creates probe circuits through each relay and downloads a file of a certain size, measuring the realized bandwidth.333Since Tor does not allow one-hop circuits, these circuits use two relays: the relay under measurement and a high-bandwidth relay. We will call this the _measured bandwidth_ , $m_{t}[j]$. Note that in a perfectly load-balanced network, all of these observations should be equal, regardless of the chosen relay. The design of $\mathit{TorFlow}$ uses a PID controller to attempt to bring these observations into balance. More specifically, let $C^{\mathit{TF}}_{t}[j]$ be the capacity of relay $j$, as estimated by $\mathit{TorFlow}$, at time $t$. $\mathit{TorFlow}$ computes an error term, $e_{t}[j]$ as the difference of the measured bandwidth and the average measured bandwidth, normalized by the average bandwidth: $e_{t}[j]=(m_{t}[j]-\bar{m_{t}})/\bar{m_{t}}$ where $\bar{m_{t}}=\sum_{j\in G\cup M\cup E}m_{t}[j]/\left(|G|+|M|+|E|\right)$ It then computes the new estimate as [11, §3.1]: $C^{\mathit{TF}}_{t+1}[j]=C^{\mathit{TF}}_{t}[j]\left(1+K_{p}e_{t}[j]+K_{i}\int_{0}^{t}e_{t^{\prime}}[j]dt^{\prime}+K_{d}\frac{de_{t}[j]}{dt}\right)$ The constants $K_{p},K_{i},$ and $K_{d}$ control the proportional, integral, and derivative components of the PID controller. In the default configuration of $\mathit{TorFlow}$, $K_{i}=K_{d}=0$ and $K_{p}=1$, so we will call this version of $\mathit{TorFlow}$ $\mathit{TorFlow\text{-}P}$. In this case the update equation can be simplified as: $C^{\mathit{TF}}_{t+1}[j]=C^{\mathit{TF}}_{t}[j]m_{t}[j]/\bar{m_{t}}$ Since the update equation does not have a normalization step, when $\mathit{TorFlow\text{-}P}$ was enabled in late 2011 in the actual Tor network, the absolute values of estimated bandwidth grew without bound444As can be seen on this graph: https://metrics.torproject.org/totalcw.html?start=2011-06-01&end=2011-12-31. This caused the Tor network to turn off $\mathit{TorFlow\text{-}P}$ and switch to using a version of $\mathit{TorFlow}$ that uses adjusted observed bandwidth instead, called $\mathit{sbws}$. We will call estimates produced by this version $C^{A}$, with: $C^{A}_{t+1}[j]=b_{t}[j]m_{t}[j]/\bar{m_{t}}$ This version uses the observed bandwidth published by the relay itself, but adjusts it down if the observed bandwidth is below average or up if it is above. It has been in use in Tor since 2012; however, it has a number of disadvantages. A relay that is not sufficiently loaded may underestimate its observed bandwidth; this leads to a well-documented ramp-up period of new relays, where their low observed bandwidth leads to a small estimated capacity and low load, which in turn leads to low observed bandwidth [6]. But even established relays see their observed bandwidth change. Figure 2 shows the observed bandwidth of 10 randomly selected relays over the month of May 2020, demonstrating that the observed values vary significantly over time. Figure 2: Variation in observed bandwidth in Tor relays over the month of May 2020: plot of 10 randomly selected relays. A new method for estimating the capacity of relays based on maximum likelihood estimation was proposed with $\mathit{MLEFlow}$ [5]. To develop $\mathit{MLEFlow}$, a probabilistic model relating the actual relay capacities $C^{*}[j]$’s and the bandwidth measurements $m[j]$’s was derived. In order to derive this relationship, a number of assumptions was made in [5]: 1. 1. Relays fall into a single category and each user path goes through only a single relay. 2. 2. A synchronized model where time is divided into epochs and user connections all terminate at the end of each epoch. At the end of the epoch the weight vector is updated and the new vector is used by all users in the next epoch. 3. 3. Users arrive to the network randomly following a Poisson process with rate $\lambda_{s}$, denoted by $\text{Pois}(\lambda_{s})$. 4. 4. Clients circuits use all the bandwidth allocated to them and are only bottlenecked by the relays. The estimates produced by this algorithm, denoted $C^{H}$ are then computed using: $C_{t+1}^{H}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ f(\kappa,m_{[t]}[j],w_{[t]}[j]),\text{ where}$ $f(\kappa,m_{[t]}[j],w_{[t]}[j])=\prod_{i=0}^{t}\frac{e^{-\lambda_{s}w_{i}[j]}}{\big{(}\frac{\kappa}{m_{i}[j]}-1\big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa}{m_{i}[j]}-1}$ While the algorithm derived showed a lot of promises, the model used to derive the estimator assumed that users can utilize arbitrary amounts of bandwidth and are only bottlenecked on the Tor network. In practice, at times the client demand on Tor is lower than the overall available bandwidth. When simulated in a low-load scenario, $\mathit{MLEFlow}$ tended to misestimate relays capacities. Also, due to this assumption, the probabilistic model in [5] is expected to produce useful results for exit relays since those relays are generally scarce in the Tor network and are expected to be paths bottleneckes. This is not true for guard and middle relays. As shown in both Python and Shadow simulations, guard and middle relays had larger estimation errors when using $\mathit{MLEFlow}$. ## III $\mathit{DiProber}$: Maximum Likelihood Estimation of Relays Capacities using dual probing We propose a new method $\mathit{DiProber}$ for estimating the capacity of relays based on maximum likelihood estimation (MLE). We use the same assumptions used to derive $\mathit{MLEFlow}$ while relaxing the last assumption made. The paths in our model consist of single relays. We assume that time is divided into epochs and connections terminate at the end of each epoch. We denote the number of paths passing through the $j^{\mathit{th}}$ relay in the $i^{\mathit{th}}$ epoch by $x_{i}[j]$. Hence, given $w_{i}$, the number of paths using the $j^{\mathit{th}}$ relay during the $i^{\mathit{th}}$ epoch is a random variable $X_{i}[j]$ with distribution $\text{Pois}(\lambda_{s}w_{i}[j])$. We design an updated measurement mechanism based on two measurement probes assigned by the authorities to each relay instead of just one. The result of the two measurements is used to group the relays in two different groups, then derive a probabilistic model for each case and apply MLE. In this section, we will present the new measurement scheme proposed as well as the probabilistic models derived to relate the actual capacity of a relay to the measurements obtained. We then derive analytical guarantees to the new estimation algorithm proposed. ### III-A Measurement mechanism Currently, Tor authorities assign a measurement probe to each relay in the network, download a file of a certain size and measure the realized bandwidth of the probe. Instead of having only one measurement probe for each relay, we propose having two measurement probes added sequentially to each relay. The authorities start by activating the first probe and measure its realized bandwidth, denoted $m^{1}_{i}[j]$. Then, while the first probe is still active, the authority starts the second probe circuit and measure its bandwidth, denoted $m^{2}_{i}[j]$. Using both _non-noisy_ measurements of a relay $j$ at a given epoch $t$, we can divide the relays in two groups and derive a probabilistic model for each one. First, we let $C_{client,i}[j]$ be the total bandwidth used by clients using relay $j$ during epoch $i$. Case 1: Relay is not bottlenecked by clients and when adding the two probes. When the clients using relay $j$, during the $i^{th}$ epoch, are not using all the available bandwidth of the relay we have $C_{client,i}[j]<C^{*}[j]$. After adding the first probe, in the case where the capacity of the relay was equally divided between the users and the probe, each path using relay $j$ will have a bandwidth of $\frac{C^{*}[j]}{x_{i}[j]+1}$. Hence all the clients using this relay will have a total bandwidth of $x_{i}[j]\frac{C^{*}[j]}{x_{i}[j]+1}$. If $C_{client,i}[j]<x_{i}[j]\frac{C^{*}[j]}{x_{i}[j]+1}$, then client utilization is not affected by adding the first probe and the probe will use all the remaining _unused capacity_ of the relay, notably $m^{1}_{i}[j]=C^{*}[j]-C_{client,i}[j]$. The same logic is true when adding the second probe. If $C_{client,i}[j]<x_{i}[j]\frac{C^{*}[j]}{x_{i}[j]+2}$ then client utilization is not affected by adding the second probe. The remaining _unused capacity_ will then be equally divided between the two probes, notably $m^{2}_{i}[j]=\frac{C^{*}[j]-C_{client,i}[j]}{2}=\frac{m^{1}_{i}[j]}{2}$. Figure 3 illustrate the idea aforementioned. Figure 3: (Case 1) Relay not bottlenecked by users. Case 2: Relay bottlenecked by clients or when adding any of the two probes. 1. (a) If the clients are using all the available capacity of the relay there will be no remaining unused capacity. Hence, as in [5], the capacity of the relay will be divided equally between all the paths going through it, notably $m^{1}_{i}[j]=\frac{C^{*}[j]}{x_{i}[j]+1}$ and $m^{2}_{i}[j]=\frac{C^{*}[j]}{x_{i}[j]+2}$. Figure 4 illustrates this case. 2. (b) In this case, the clients are not using all the available capacity of relay. However, when adding the first probe, the client utilization can be affected if $C_{client,i}[j]>x_{i}[j]\frac{C^{*}[j]}{x_{i}[j]+1}$. In other words, clients will then be using all the capacity available to them after the first probe was added and there will be no remaining unused capacity. Thus as case (a), the capacity of the relay will be divided equally between all the paths going through it, notably $m^{1}_{i}[j]=\frac{C^{*}[j]}{x_{i}[j]+1}$ and $m^{2}_{i}[j]=\frac{C^{*}[j]}{x_{i}[j]+2}$. 3. (c) In this case, clients are not using all the available capacity of the relay, nor adding the first probe will affect their utilization. Thus, the first probe will be first assigned all the remaining _unused capacity_ , $m^{1}_{i}[j]=C^{*}[j]-C_{client,i}[j]$. However, when adding the second probe, the clients utilization is affected if $C_{client,i}[j]>x_{i}[j]\frac{C^{*}[j]}{x_{i}[j]+2}$ and there will be no remaining capacity. Thus, $m^{2}_{i}[j]=\frac{C^{*}[j]}{x_{i}[j]+2}$. Figure 4: (Case 2) Relay bottlenecked by users. Thus, if the relay falls under case 1, the second observation of the relay will be equal to half the first observation obtained and the relationship between the actual capacity of the relay $C^{*}[j]$ and the observation is $M^{2}_{i}[j]=\frac{C^{*}[j]-C_{client,i}[j]}{2}$. While for relays that fall under the second case, we can’t make the same conclusion about $m^{1}_{i}[j]$ and $m^{2}_{i}[j]$. However, for all subcases of case 2, we have a relationship between $C^{*}[j]$ and $M^{2}_{i}[j]$ with $M^{2}_{i}[j]=\frac{C^{*}[j]}{X_{i}[j]+2}$. ### III-B MLE capacities estimation In this section, we will show the method used to compute the maximum likelihood estimation of relays capacities using the sequence of pairs of non- noisy measurements and the weights published by the Tor authority. More specifically, for any relay $j\in[n]$, the MLE estimate of its actual capacity $C^{*}[j]$ is the maximizer in $\mathcal{C}\subset\mathbb{R}^{n}_{\geq 0}$ of the probability of observing the full history of measurements $(m^{1}_{[t]}[j],m^{2}_{[t]}[j])$, given the published weights $w_{[t]}[j]$ over the first $t+1$ periods. First at each epoch $i\in[t]$, and in order to use the correct probabilistic model, the algorithm compares the values of $m^{1}_{i}[j]$ and $m^{2}_{i}[j]$ and determine which case of the two cases discussed in III-A is true for the relay $j$ at the $i^{th}$ epoch. * • If $m^{1}_{i}[j]=2m^{2}_{i}[j]$, then $M^{2}_{i}[j]=\frac{C^{*}[j]-C_{client,i}[j]}{2}$. * • If $m^{1}_{i}[j]\neq 2m^{2}_{i}[j]$, then $M^{2}_{i}[j]=\frac{C^{*}[j]}{X_{i}[j]+2}$ Assuming that we know total client utilization of each relay during each epoch is not a practical assumption. In this paper, we assume that the average utilization of a client on the Tor network is known and denoted $C^{avg}_{client}$. Thus the total utilization of a relay $j$ during the $i^{th}$ epoch will be $X_{i}[j]C^{avg}_{client}$ We add the superscript $D$ to the capacity estimate to denote that the result of two probes is considered. ###### Theorem 1 (MLE estimates using dual probing). For any $j\in[n]$ and $t\in\mathbb{N}$, the MLE estimate of $C^{*}[j]$ given the weight and observation pairs vectors $w_{[t]}[j]$ and $(m^{1}_{[t]}[j],m^{2}_{[t]}[j])$ is $\displaystyle C_{t+1}^{D}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \prod_{i=0}^{t}f(\kappa,m_{i}[j],w_{i}[j]),\text{ where}$ (1) $f(\kappa,m_{i}[j],w_{i}[j])=$ $\left\\{\begin{array}[]{ll}\Pr_{X_{i}[j]\sim\text{Pois}(\lambda_{s}w_{i}[j])}\left(\frac{\kappa- X_{i}[j]C^{avg}_{client}}{2}=m^{2}_{i}[j]\right)\text{if}\hskip 2.84544ptm^{1}_{i}[j]=2m^{2}_{i}[j]\\\ \Pr_{X_{i}[j]\sim\text{Pois}(\lambda_{s}w_{i}[j])}\left(\frac{\kappa}{X_{i}[j]+2}=m^{2}_{i}[j]\right)\text{if}\hskip 2.84544ptm^{1}_{i}[j]\neq 2m^{2}_{i}[j]\end{array}\right.$ We now write the probability in equation (1) in terms of the known quantities: $\lambda_{s}$, $w_{[t]}$, and $m_{[t]}$. For any $j\in[n]$ and $t\in\mathbb{N}$ the function $f$ of equation (1) can be written as follows: $f(\kappa,m_{i}[j],w_{i}[j])=$ (5) ###### Definition 1 ($\mathit{DiProber\text{-}WH}$). For any relay $j\in[n]$ and $t\in\mathbb{N}$, $\mathit{DiProber\text{-}WH}$ updates the weight vector by: $\displaystyle w_{t+1}^{\mathit{WH}}[j]=\frac{C_{t}^{D}[j]}{\sum_{k=0}^{n}C_{t}^{D}[k]}.$ (6) where we use the supersctipt $\mathit{WH}$ in $w_{t}^{\mathit{WH}}$ to identify $\mathit{DiProber\text{-}WH}$. Equation (1) does not account for non-modeled noise in the measurements. Once noise affects $m^{1}_{i}[j]$ and $m^{2}_{i}[j]$, we will not be able to accurately discern between the two cases. ### III-C One step maximum likelihood In this section, we look at the special case of only considering the last relay measurement at each epoch instead of the whole history of measurements. We add the superscript $D1$ to the capacity estimate to denote that only the last result of the two probes is considered. Equation 1 can now be written as, $\displaystyle C_{t+1}^{D1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ f(\kappa,m_{t}[j],w_{t}[j]),\text{ where}$ (7) $f(\kappa,m_{t}[j],w_{t}[j])$ is defined in Equation 5. To approximate the maximizer of equation (1), we find $\kappa$ that sets the derivative of $f$ to zero. The result is given in the following theorem. ###### Theorem 2 (One step MLE closed form). For any $j\in[n]$ and $t\in\mathbb{N}$, the MLE estimate of $C^{*}[j]$ given the last weight and observations pair, $w_{t}[j]$ and ($m^{1}_{t}[j],m^{2}_{t}[j]$ is $\displaystyle C_{t+1}^{D1}[j]$ $\displaystyle=\left\\{\begin{array}[]{ll}\lambda_{s}w_{t}[j]C^{avg}_{client}+2m_{t}^{2}[j],\text{if}\hskip 2.84544ptm^{1}_{i}[j]=2m^{2}_{i}[j]\\\ \\\ m_{t}^{2}[j]\left(\lambda_{s}w_{t}[j]+2\right),\text{if}\hskip 2.84544ptm^{1}_{i}[j]\neq 2m^{2}_{i}[j]\end{array}\right.$ (11) From Theorem 2, in the case where the relay is not bottlenecked, i.e. $m^{1}_{i}[j]=2m^{2}_{i}[j]$, the estimated capacity is equal to the sum of the unused capacity, $2m^{2}_{t}[j]$, and the expected capacity used by clients, $\lambda_{s}w_{t}[j]C^{avg}_{client}$. $C_{t+1}^{D1}[j]=\underbrace{\lambda_{s}w_{t}[j]C^{avg}_{client}}_{\text{expected capacity used by clients}}+\underbrace{m_{t}^{1}[j]}_{\text{capacity unused by clients}}$ While in the case where the relay is bottlenecked, the estimated capacity is the result of the product of the observation, $m^{2}_{t}[j]$, and the expected number of users, $\lambda_{s}w_{t}[j]$. The added two in the denominator refers to the two probes added. $C_{t+1}^{D1}[j]=m_{t}^{2}[j]\left(\underbrace{\lambda_{s}w_{t}[j]}_{\text{expected number of users}}+2\right)$ ###### Definition 2 ($\mathit{DiProber\text{-}O}$). For any relay $j\in[n]$ and $t\in\mathbb{N}$, $\mathit{DiProber\text{-}O}$ updates the weight vector by: $\displaystyle w_{t+1}^{\mathit{O}}[j]=\frac{C_{t}^{D1}[j]}{\sum_{k=0}^{n}C_{t}^{D1}[k]}.$ (12) where we use the superscript $\mathit{O}$ in $w_{t}^{\mathit{O}}$ to identify $\mathit{DiProber\text{-}O}$. ### III-D Convergence of $\mathit{DiProber\text{-}WH}$ and $\mathit{DiProber\text{-}O}$ estimates In this section we show that, starting with any initial weight vector, the mean of the dual probing algorithm estimates for any relay capacity, whether considering the full history in every update as in $\mathit{DiProber\text{-}WH}$ (Definition 1) or only the most recent measurement in every update as in $\mathit{DiProber\text{-}O}$ (Definition 2), converges to the actual relay capacity. We also show that the variance of the estimates when using $\mathit{DiProber\text{-}WH}$ estimation goes to zero as the number of epochs increases. ###### Theorem 3 (Estimates of both methods converge). For any $j\in[n]$, $t\in\mathbb{N}$, and a method $y\in\\{\mathit{TorFlow\text{-}P},\mathit{MLEFlow\text{-}CF}\,\mathit{DiProber\text{-}WH}\\}$, $\displaystyle\mathbb{E}[C_{t}^{y}[j]]\leq C^{*}[j].$ (13) Moreover, as $t\rightarrow\infty$, $\displaystyle\mathbb{E}[C_{t}^{y}[j]]\geq C^{*}[j]\left(1-\frac{1}{\lambda_{s}w^{*}[j]}\right).$ (14) ###### Corollary 4 (More users paths leads to a better convergence). As the rate of users arrival $\lambda_{s}\rightarrow\infty$, for any $j\in[n]$, $t\in\mathbb{N}$, and method $y\in\\{\mathit{DiProber\text{-}O},\mathit{DiProber\text{-}WH}\\}$, $\mathbb{E}[C_{t}^{y}[j]]\rightarrow C^{*}[j]$. Furthermore, we show that the variance of the estimates of $\mathit{DiProber\text{-}WH}$ converge to zero. This shows that $\mathit{DiProber\text{-}WH}$ provides stable and consistent estimates. ###### Theorem 5 (Convergence of variance of $\mathit{DiProber\text{-}WH}$). As $t\rightarrow\infty$, $\mathit{Var}[C_{t}^{\mathit{WH}}[j]]\rightarrow 0$. Hence the variances of $\mathit{DiProber}$ estimates when considering the full history of weights and measurements are close to the actual capacities. We will show this experimentally in the next section. ### III-E $\mathit{DiProber}$: Implementation In order to solve the maximization of Theorem 1, we discretize the bounded capacity set $\mathcal{C}$ and iteratively find the maximizer of (1). Note that quantization requires knowing a lower and upper bound on the relay capacity, which can be estimated based on past observations. Consider a partition $\bar{\mathcal{C}}$ of $\mathcal{C}$ into bins. The set $\bar{\mathcal{C}}$ contains the centers of the bins of $\mathcal{C}$. For any $j\in[n]$, $t\in\mathbb{N}$, and $\kappa\in\bar{\mathcal{C}}$, we define ${L_{t}}(j,\kappa)$ to be: (18) the $t^{\mathit{th}}$ term of the sum when taking the $\log$ of (5). For $\mathit{DiProber\text{-}O}$, since we are only considering the last measurement, the estimate of the capacity of relay $j$ is computed by iteratively searching for the maximizer $\kappa$ over the discretized capacity set $\bar{\mathcal{C}}$ in the following equation: $\displaystyle C_{t+1}^{\mathit{O}}[j]:=\max_{\kappa\in\bar{\mathcal{C}}}{L_{t}}(j,\kappa).$ (19) For $\mathit{DiProber\text{-}WH}$, the sum of ${L_{t}}(j,\kappa)$ over measurement periods is stored in a variable $S_{t+1}(j,\kappa)$: ${S_{t+1}}(j,\kappa)={S}_{t}(j,\kappa)+{L_{t}}(j,\kappa)$, with ${S}_{0}(j,\kappa)=0$. Then, the maximum likelihood estimate is computed by iteratively searching for the maximizer $\kappa$ over the discretized capacity set $\bar{\mathcal{C}}$ in the following equation: $\displaystyle C_{t+1}^{\mathit{WH}}[j]:=\max_{\kappa\in\bar{\mathcal{C}}}{S_{t+1}}(j,\kappa).$ (20) ## IV Analytical Model Simulations: Low Fidelity Experiments To better understand the properties and performance of $\mathit{TorFlow\text{-}P}$, $\mathit{MLEFlow}$, $\mathit{sbws}$ and $\mathit{DiProber}$, we evaluated them using simulations of our analytical model of the Tor network. These simulations are implemented in Python; we will make our implementation publicly available at publication time. We evaluate the performance of the different methods using two metrics: (a) the accuracy of the relay capacity estimates in a network where clients use all the bandwidth allocated to them, (b) the accuracy of the relay capacity estimates in an underloaded network and (c) the amount of bandwidth allocated to the user paths resulting from the weight vectors generated using the capacity estimates. The simulation algorithm we have used is shown in Algorithm 1. The algorithm takes as input: the Poisson arrival rate of users $\lambda_{s}$, the total number of measurement periods $T$ to be simulated, a method $\in\\{\mathit{Actual},\mathit{TorFlow\text{-}P},$ $\mathit{MLEFlow},\mathit{DiProber\text{-}O},\mathit{DiProber\text{-}WH}\\}$ to compute the capacities of the relays from measurements. The algorithm outputs the bandwidths allocated for users paths and the weight vectors published over all periods between $0$ and $T$. The simulation algorithm iterates over measurement periods. In each period $i$, it generates the total number of users paths $N_{i}$ that will join the network by sampling a Poisson distribution with rate $\lambda_{s}$ in line 3. Then, it uses the weight vector $w_{i}$ computed in the previous period as a probability distribution for the users to choose the relays of their paths from in line 4. In line 5, it adds the first probe to each relay and uses the max-min fairness bandwidth allocation algorithm to get the bandwidth allocated for each path, and thus generate the observation vector $m^{1}_{i}$. In line 6, it adds the second probe and generates $m^{2}_{i}$. After that, it computes $w_{i+1}$ using the given method in line 7. Finally, it deletes all the paths for a fresh start of the next period. Algorithm 1 Low fidelity simulation 1:input: $\lambda_{s},T$, $\mathit{method}\in\\{\mathit{Actual}$, $\mathit{TorFlow\text{-}P},\mathit{MLEFlow},\mathit{DiProber\text{-}WH},\mathit{DiProber\text{-}O}\\}$. 2:for $i\in[0,...,T]$ do 3: Pick the number of users $N_{i}\sim Poi(\lambda_{s})$. 4: Construct users paths of $3$ relays using $w_{i}$. 5: Compute ${m^{1}_{i}}$ using max-min bandwidth allocation. 6: Compute ${m^{2}_{i}}$ using max-min bandwidth allocation. 7: Compute $w_{i+1}$ based on $m_{i}$ and $w_{i}$ using $\mathit{method}$. 8: Delete all paths in the network. 9:return: $m_{0:T},w_{0:T}$ We consider a network analogous to the current Tor network with 6037 relays as of June $23^{\mathit{rd}}$ 2020\. The relays are distributed as follows: $2351$ are guard relays, $2576$ are middle relays, and $1110$ are exit relays (this includes any relays that have both the Exit and Guard flags set). Lacking a ground truth, we used the measured capacity in the Tor consensus document as the actual capacity of the Tor relays in our simulation. The maximum capacity of all relays was $169\,000$ kb/s, while the total capacity of the guard, middle and exit relays in the network are around $42.6\times 10^{6}$, $6.7\times 10^{6}$ and $17.7\times 10^{6}$ kb/s respectively. Hence, the capacity set is $\mathcal{C}=[0,169000]$. The capacity distributions of the guard, middle, and exit relays are shown in Figure 5. The max-min bandwidth allocation algorithm [4, 3], in Python assumes that clients will use the full bandwidth allocated to them. Thus we simulated two types of networks: (1) a full utilization network using the max-min bandwidth allocation algorithm where each client path consists of three relays and (2) an under loaded network scenario taking into account client side bottlenecks. To simulate this idea, we adjusted the simulation to add a bandwidth cap to each client flow. This bandwidth cap is enforced by a fourth relay added to each client flow, with a capacity selected uniformly at random from the interval [8,18] KB/s. This interval was chosen based on the full utilization scenario bandwidth distribution. Since the average bandwidth of flows in the full utilization scenario was approximately 22 KB/s, the cap means that the flows can utilize at most about 60% of the Tor network capacity. To use $\mathit{DiProber}$, we need to partition $\mathcal{C}$ into bins. From Figure 5, we use the same technique used in [5]. We choose the bins to be intervals of the form $[a^{b-1},a^{b}]$ where $a$ is a strictly positive real number and $b\in[1,...,b_{max}]$ where $b_{max}=\lceil\frac{\log(\max(C[j]))}{\log(a)}\rceil$ for $j\in[n]$. Figure 5: relays capacity distribution. ### IV-A Low-fidelity sims results and analysis In all of the simulation runs, we chose the number of periods $t$ to be $50$, the initial weight vector $w_{0}$ to be uniform, the rate of arrival $\lambda_{s}$ to be $10^{6}$. #### $\mathit{DiProber\text{-}WH}$ performs better than $\mathit{DiProber\text{-}O}$ While both algorithms of $\mathit{DiProber}$ outperformed the other algorithms, $\mathit{DiProber\text{-}WH}$ had a lower average error than $\mathit{DiProber\text{-}O}$ for all classes of relays. The results are shown in Figure 6. Figure 6: Estimation error distribution after the $50^{th}$ measurement period in a fully utilized network for both $\mathit{DiProber}$ algorithms. #### $\mathit{DiProber}$ leads to better guard and middle relays estimates in fully loaded networks We tested the different estimation algorithms on three-relays paths networks. Both $\mathit{DiProber}$ algorithms and $\mathit{MLEFlow}$ had lower average error than $\mathit{TorFlow\text{-}P}$ and $\mathit{sbws}$. The results are shown in Figure 7. While the errors for exit relays when using $\mathit{DiProber}$ and $\mathit{MLEFlow}$ were close, $\mathit{DiProber}$ leads to lower average error for guard and middle relays. The average estimation errors for $\mathit{DiProber}$ and $\mathit{MLEFlow}$ stayed below $10\%$ for exit relays, while it was higher for $\mathit{TorFlow\text{-}P}$ with $72\%$. For guard and middle relays, the error was lowest for $\mathit{DiProber\text{-}WH}$ at $18\%$ and $14\%$ respectively. It was higher for $\mathit{MLEFlow}$ at around $25\%$ error and even higher for $\mathit{TorFlow\text{-}P}$ at $75\%$. Figure 7: Estimation error distribution after the $50^{th}$ measurement period in a fully utilized network. #### The dual probing algorithm advantage extend to the under-loaded network scenario: We simulate the case of under-loaded networks as was done in [5] and as described above by adding a forth relay to each path created. $\mathit{DiProber}$ outperforms all other algorithms. The error for all classes of relays was below $25\%$ for $\mathit{DiProber}$. While the error for exit relays remained relatively low when using $\mathit{MLEFlow}$, other classes estimates all had an average error above $45\%$. The results are shown in Figure 8. Figure 8: Estimation error distribution after the $50^{th}$ measurement period in an under-loaded network. #### $\mathit{DiProber\text{-}WH}$ and $\mathit{DiProber\text{-}O}$ give better and fairer bandwidth allocation than than the other algorithms The means of the bandwidths allocated for paths using $\mathit{MLEFlow}$ and $\mathit{DiProber}$ are equal to that of the Actual scenario, while that of $\mathit{TorFlow\text{-}P}$ is slightly smaller. The advantage of both those methods is the stability of their estimates. The standard deviation of $\mathit{DiProber}$ is a maximum of 1.4, while $\mathit{MLEFlow}$ was 1.7, when that of Actual is around 0.6 and $\mathit{TorFlow\text{-}P}$ is around 165, orders of magnitude larger. Moreover, the maximum and minimum bandwidths allocated of our methods are similar to that of Actual while $\mathit{TorFlow\text{-}P}$ had orders of magnitude larger maximum. That means that our methods distribute bandwidths more fairly than $\mathit{TorFlow\text{-}P}$. ## V Related Work Improving the performance of the Tor network has been the subject of much research; we refer the reader to the survey by AlSabah and Goldberg for an overview [1]. Here we summarize related work specifically focusing on relay capacity estimation. Snader and Borisov proposed using _opportunistic measurements_ , where each relay measures the bandwidth of each other relay it communicates with as part of normal operation, and designed EigenSpeed [13], which combines these measurements using principal component analysis to derive a single relay capacity. EigenSpeed was designed to avoid certain types of collusion and misreporting attacks; however, Johnson et al. [10] discovered that it is subject to a number of other attacks that allow colluding adversaries to inflate their bandwidth. They also designed PeerFlow, which is a more robust mechanism to combine opportunistic measurements from relays with provable limits on inflation attacks. These bounds, however, depend on having a fraction of bandwidth being on trusted nodes, and it has slow convergence properties due to its limitations on changing bandwidth values. FlashFlow [18] is a new proposal to replace TorFlow. FlashFlow uses several servers that measure a relay simultaneously, generating a large network load intended to max out its capacity. FlashFlow has a guaranteed inflation bound of only 33% but it is based on the assumption that a relay capacity is based on a hard limit that cannot be exceeded, as TorFlow uses traffic that is explicitly labeled for for bandwidth probing. In practice, it is often easier and cheaper to obtain high peak bandwidth capability than sustaining the same bandwidth continuously. At the same time, TorFlow probes, though not explicitly labeled, can also be identified by their distinct characteristics, and can be used to preferentially forward probe traffic to inflate bandwidth estimates, or to perform sophisticated denial-of-service attacks [9]. A more stealthy approach for bandwidth measurement probing remains an open research question. The simplest and most effective attack on TorFlow, however, is to inflate the observed bandwidth published by the relay [10]; this attack does not apply to $\mathit{DiProber}$ as it does not use the observed bandwidth. ## VI Conclusion We have developed a new method for estimating the relay capacities in the Tor network, $\mathit{DiProber}$. We show that $\mathit{MLEFlow}$ fails to accurately estimate relays capacities in under-loaded networks. Our detailed mathematical analysis showed that $\mathit{DiProber}$ capacity estimates converge to their true value, while the estimate variance converges to 0, as the number of observations grows. We validated the performance of $\mathit{MLEFlow}$ with extensive simulations using a our custom flow-based simulator. Our results show that $\mathit{DiProber}$ produces much more accurate estimates of relay capacities, which in turn results in much better load balancing of user traffic across the network, as compared with current methods. ## References * [1] M. AlSabah and I. Goldberg, “Performance and security improvements for Tor: A survey,” _ACM Computing Surveys (CSUR)_ , vol. 49, no. 2, p. 32, 2016. * [2] K. Bauer, D. McCoy, D. Grunwald, T. Kohno, and D. Sicker, “Low-resource routing attacks against Tor,” in _Proceedings of the 2007 ACM Workshop on Privacy in Electronic Society (WPES)_ , 2007, pp. 11–20. * [3] H. Darir, “Privacy-preserving network congestion control,” 2019. * [4] H. Darir, H. Sibai, N. Borisov, G. Dullerud, and S. Mitra, “Tightrope: Towards optimal load-balancing of paths in anonymous networks,” in _Proceedings of the 2018 Workshop on Privacy in the Electronic Society_ , 2018, pp. 76–85. * [5] H. Darir, H. Sibai, C.-Y. Cheng, N. Borisov, G. Dullerud, and S. Mitra, “Mleflow: Learning from history to improve load balancing in tor,” _Proceedings on Privacy Enhancing Technologies_ , vol. 2022, no. 1, pp. 75–104, 2022. [Online]. Available: https://doi.org/10.2478/popets-2022-0005 * [6] R. Dingledine, “The lifecycle of a new relay,” The Tor Project Blog, https://blog.torproject.org/lifecycle-new-relay, Sep. 2013. * [7] R. Dingledine, N. Mathewson, and P. F. Syverson, “Tor: The second-generation onion router,” in _USENIX Security Symposium_. USENIX, 2004, pp. 303–320. * [8] S. Herbert, S. J. Murdoch, and E. Punskaya, “Optimising node selection probabilities in multi-hop m/d/1 queuing networks to reduce latency of tor,” _Electronics letters_ , vol. 50, no. 17, pp. 1205–1207, 2014. * [9] R. Jansen, T. Vaidya, and M. Sherr, “Point break: a study of bandwidth denial-of-service attacks against Tor,” in _28th USENIX Security Symposium_ , 2019, pp. 1823–1840. * [10] A. Johnson, R. Jansen, N. Hopper, A. Segal, and P. Syverson, “PeerFlow: Secure load balancing in Tor,” _Proceedings on Privacy Enhancing Technologies_ , vol. 2017, no. 2, pp. 74–94, 2017. * [11] K. Loesing, M. Perry, and A. Gibson, “Bandwidth scanner specification,” https://gitweb.torproject.org/torflow.git/tree/NetworkScanners/BwAuthority/README.spec.txt, 2011\. * [12] M. Perry, “TorFlow: Tor network analysis,” in _Proceedings of the 2nd Workshop on Hot Topics in Privacy Enhancing Technologies (HotPETs)_ , 2009, pp. 1–14. * [13] R. Snader and N. Borisov, “EigenSpeed: Secure peer-to-peer bandwidth evaluation,” in _8th International Workshop on Peer-To-Peer Systems_ , R. Rodrigues and K. Ross, Eds. Berkeley, CA, USA: USENIX Association, Apr. 2009. * [14] ——, “Improving security and performance in the Tor network through tunable path selection,” _IEEE Transactions on Dependable and Secure Computing_ , vol. 8, no. 5, pp. 728–741, 2011. * [15] The Tor Project, “Tor directory protocol, version 3,” https://gitweb.torproject.org/torspec.git/tree/dir-spec.txt, 2020. * [16] ——, “Tor metrics: Servers,” https://metrics.torproject.org/networksize.html, 2020. * [17] ——, “Tor metrics: Users,” https://metrics.torproject.org/userstats-relay-country.html, 2020. * [18] M. Traudt, R. Jansen, and A. Johnson, “Flashflow: A secure speed test for tor,” 2020. ### -A Proofs See 1 ###### Proof. As discussed, at each epoch, a relay can fall into the two cases discussed in section III-A. We start with Case 1, where the second measurement is equal to double the first measurement of a relay. When evaluating the objective function of the MLE, the observation random variable of the $j^{\mathit{th}}$ relay $M_{i}[j]$ can be written as a function of $\kappa$, as if we are assuming $\kappa=C^{*}[j]$, and the random variable $X_{i}[j]$ for $i\in[t]$: $M^{2}_{i}[j]=\frac{\kappa-X_{i}[j]C_{Client}^{avg}}{2}.$ (21) Recall that we assume that the random variable $X_{i}[j]$ follows a Poisson distribution with parameter $\lambda_{s}w_{i}[j]$ and all users leave at the end of each epoch. Hence, given $w_{i}[j]$ for $j\in[n]$, the $M_{i}[j]$’s at different iterations are independent random variables. Thus eq. 21 can be written as the product of the probability of the independent random variables $[M_{1}[j],...,M_{t}[j]]$: $\begin{multlined}C_{t+1}^{H}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\Pr_{X_{[t]}[j]\sim\text{Pois}(\lambda_{s}W_{[t]}[j])}(M^{2}_{[t]}[j]=m^{2}_{[t]}[j]\hskip 2.84544pt|W_{[t]}[j]=w_{[t]}[j])\\\ C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}\Pr(M^{2}_{i}[j]=m^{2}_{i}[j]\hskip 2.84544pt|\hskip 2.84544ptW_{i}[j]=w_{i}[j]).\\\ \end{multlined}C_{t+1}^{H}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\Pr_{X_{[t]}[j]\sim\text{Pois}(\lambda_{s}W_{[t]}[j])}(M^{2}_{[t]}[j]=m^{2}_{[t]}[j]\hskip 2.84544pt|W_{[t]}[j]=w_{[t]}[j])\\\ C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}\Pr(M^{2}_{i}[j]=m^{2}_{i}[j]\hskip 2.84544pt|\hskip 2.84544ptW_{i}[j]=w_{i}[j]).\\\ $ Rearranging eq. 21 results in: $X_{i}[j]=\frac{\kappa-2M^{2}_{i}[j]}{C_{Client}^{avg}}.$ (22) When the measurement is made and the observation is fixed, i.e. $M^{2}_{i}[j]=m^{2}_{i}[j]$, the probability in LABEL:eq:Mal3 can be expressed in terms of the random variable $X_{i}[j]$: $x_{i}[j]=\frac{\kappa-2m^{2}_{i}[j]}{C_{Client}^{avg}}$. $C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}\Pr(X_{i}[j]=x_{i}[j]|\hskip 2.84544ptW_{i}[j]=w_{i}[j])$ (23) Using the Poisson distribution probability mass function, we can write: $\begin{multlined}C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{(x_{i}[j])!}(\lambda_{s}w_{i}[j])^{x_{i}[j]}\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{\big{(}\frac{\kappa-2m^{2}_{i}[j]}{C_{Client}^{avg}}\big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa-2m^{2}_{i}[j]}{C_{Client}^{avg}}}\\\ \end{multlined}C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{(x_{i}[j])!}(\lambda_{s}w_{i}[j])^{x_{i}[j]}\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{\big{(}\frac{\kappa-2m^{2}_{i}[j]}{C_{Client}^{avg}}\big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa-2m^{2}_{i}[j]}{C_{Client}^{avg}}}\\\ $ (24) Considering Case 2, the objective function of the MLE, the observation random variable of the $j^{\mathit{th}}$ relay $M_{i}[j]$ can be written as a function of $\kappa$, as if we are assuming $\kappa=C^{*}[j]$, and the random variable $X_{i}[j]$ for $i\in[t]$: $M^{2}_{i}[j]=\frac{\kappa}{X_{i}[j]+2}.$ (25) Recall that we assume that the random variable $X_{i}[j]$ follows a Poisson distribution with parameter $\lambda_{s}w_{i}[j]$ and all users leave at the end of each epoch. Hence, given $w_{i}[j]$ for $j\in[n]$, the $M^{2}_{i}[j]$’s at different iterations are independent random variables. Thus eq. 25 can be written as the product of the probability of the independent random variables $[M^{2}_{1}[j],...,M^{2}_{t}[j]]$: $\begin{multlined}C_{t+1}^{H}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\Pr_{X_{[t]}[j]\sim\text{Pois}(\lambda_{s}W_{[t]}[j])}(M_{[t]}[j]=m_{[t]}[j]\hskip 2.84544pt|W_{[t]}[j]=w_{[t]}[j])\\\ C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}\Pr(M_{i}[j]=m_{i}[j]\hskip 2.84544pt|\hskip 2.84544ptW_{i}[j]=w_{i}[j]).\\\ \end{multlined}C_{t+1}^{H}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\Pr_{X_{[t]}[j]\sim\text{Pois}(\lambda_{s}W_{[t]}[j])}(M_{[t]}[j]=m_{[t]}[j]\hskip 2.84544pt|W_{[t]}[j]=w_{[t]}[j])\\\ C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}\Pr(M_{i}[j]=m_{i}[j]\hskip 2.84544pt|\hskip 2.84544ptW_{i}[j]=w_{i}[j]).\\\ $ Rearranging eq. 25 results in: $X_{i}[j]=\frac{\kappa}{M^{2}_{i}[j]}-2.$ (26) When the measurement is made and the observation is fixed, i.e. $M^{2}_{i}[j]=m_{i}[j]$, the probability in LABEL:eq:Maln3 can be expressed in terms of the random variable $X_{i}[j]$: $X_{i}[j]=\frac{\kappa}{m^{2}_{i}[j]}-2$. $C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}\Pr(X_{i}[j]=x_{i}[j]|\hskip 2.84544ptW_{i}[j]=w_{i}[j])$ (27) Using the Poisson distribution probability mass function, we can write: $\begin{multlined}C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{(x_{i}[j])!}(\lambda_{s}w_{i}[j])^{x_{i}[j]}\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{\big{(}\frac{\kappa}{m^{2}_{i}[j]}-2\big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa}{m^{2}_{i}[j]}-2}\\\ \end{multlined}C^{H}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{(x_{i}[j])!}(\lambda_{s}w_{i}[j])^{x_{i}[j]}\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{\big{(}\frac{\kappa}{m^{2}_{i}[j]}-2\big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa}{m^{2}_{i}[j]}-2}\\\ $ (28) ∎ See 2 ###### Proof. Case 1: As discussed previously, we know that we are in this case if $m_{i}^{1}[j]=2m_{i}^{2}[j]$. We denote $C_{avg,i}$ the average bandwidth used by each client in the network during the $i^{th}$ epoch. Hence since the relay is not bottlenecked, we can say that $m_{i}^{1}[j]=C^{*}[j]-x_{i}[j]C_{avg,i}$; hence, given $w_{i}$, the measurement of the $j^{\mathit{th}}$ relay at the $i^{\mathit{th}}$ iteration is a random variable $M_{i}^{1}[j]=C^{*}[j]-X_{i}[j]C_{avg,i}$. W will use superscript $D1$ to indicate that we are only using the most recent value of $m_{t}[j]$ and $w_{t}[j]$. Using maximum likelihood estimation, we have $\displaystyle C_{t+1}^{R}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ f(\kappa,m_{t}[j],w_{t}[j]),\text{ where}$ (29) $\displaystyle f(\kappa,m_{t}[j],w_{t}[j])=\Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(\kappa- X_{t}[j]C_{avg,t}=m_{i}^{1}[j]\right),$ (30) $\begin{multlined}C_{t+1}^{D1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(\kappa- X_{t}[j]C_{avg,t}=m_{t}^{1}[j]\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(X_{t}[j]=\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(X_{t}[j]=\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\exp\left(-\lambda_{s}w_{t}[j]\right)\frac{1}{\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)!}\left(\lambda_{s}w_{t}[j]\right)^{\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ -\lambda_{s}w_{t}[j]-\log\left(\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)!\right)+\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\log\left(\lambda_{s}w_{t}[j]\right)\\\ \end{multlined}C_{t+1}^{D1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(\kappa- X_{t}[j]C_{avg,t}=m_{t}^{1}[j]\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(X_{t}[j]=\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(X_{t}[j]=\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\exp\left(-\lambda_{s}w_{t}[j]\right)\frac{1}{\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)!}\left(\lambda_{s}w_{t}[j]\right)^{\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ -\lambda_{s}w_{t}[j]-\log\left(\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)!\right)+\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\log\left(\lambda_{s}w_{t}[j]\right)\\\ $ Using Stirling approximation for the second term, we get $\begin{multlined}C_{t+1}^{D1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ -\lambda_{s}w_{t}[j]-\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\log\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\\\ +\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)+\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\log\left(\lambda_{s}w_{t}[j]\right)\\\ \end{multlined}C_{t+1}^{D1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ -\lambda_{s}w_{t}[j]-\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\log\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\\\ +\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)+\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\log\left(\lambda_{s}w_{t}[j]\right)\\\ $ In order to find maximum, we differentiate the right side with respect to $\kappa$ and equate it to zero: $\begin{multlined}-\frac{1}{C_{avg,t}}\log\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)-\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\frac{\frac{1}{C_{avg,t}}}{\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)}+\frac{1}{C_{avg,t}}+\frac{1}{C_{avg,t}}\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ -\frac{1}{C_{avg,t}}\log\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)-\frac{1}{C_{avg,t}}+\frac{1}{C_{avg,t}}+\frac{1}{C_{avg,t}}\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ -\log\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)+\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ \log\left(\frac{C_{avg,t}\lambda_{s}w_{t}[j]}{\kappa-m_{t}^{1}[j]}\right)=0\\\ \frac{C_{avg,t}\lambda_{s}w_{t}[j]}{\kappa-m_{t}^{1}[j]}=1\\\ \end{multlined}-\frac{1}{C_{avg,t}}\log\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)-\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)\frac{\frac{1}{C_{avg,t}}}{\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)}+\frac{1}{C_{avg,t}}+\frac{1}{C_{avg,t}}\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ -\frac{1}{C_{avg,t}}\log\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)-\frac{1}{C_{avg,t}}+\frac{1}{C_{avg,t}}+\frac{1}{C_{avg,t}}\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ -\log\left(\frac{\kappa- m_{t}^{1}[j]}{C_{avg,t}}\right)+\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ \log\left(\frac{C_{avg,t}\lambda_{s}w_{t}[j]}{\kappa-m_{t}^{1}[j]}\right)=0\\\ \frac{C_{avg,t}\lambda_{s}w_{t}[j]}{\kappa-m_{t}^{1}[j]}=1\\\ $ Thus $\kappa=\underbrace{\lambda_{s}w_{t}[j]C_{avg,t}}_{\text{expected bandwidth used by clients}}+\underbrace{m_{t}^{1}[j]}_{\text{bandwidth left unused}}$ Case 2: We know that we are in this case if $m_{i}^{1}[j]\neq 2m_{i}^{2}[j]$. Since the relay is bottlenecked when we add the second probe in all cases described previously, we can say that $m_{i}^{2}[j]=\frac{C^{*}[j]}{x_{i}[j]+2}$; hence, given $w_{i}$, the measurement of the $j^{\mathit{th}}$ relay at the $i^{\mathit{th}}$ iteration is a random variable $M_{i}^{2}[j]=\frac{C^{*}[j]}{X_{i}[j]+2}$. Using maximum likelihood estimation, we have $\displaystyle C_{t+1}^{D1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ f(\kappa,m_{t}[j],w_{t}[j]),\text{ where}$ (31) $\displaystyle f(\kappa,m_{t}[j],w_{t}[j])=\Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(\frac{\kappa}{X_{i}[j]+2}=m_{i}^{2}[j]\right),$ (32) $\begin{multlined}C_{t+1}^{D1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(\frac{\kappa}{X_{i}[j]+2}=m_{t}^{2}[j]\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(X_{t}[j]=\frac{\kappa}{m_{t}^{2}[j]}-2\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(X_{t}[j]=\frac{\kappa}{m_{t}^{2}[j]}-2\right)\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\exp\left(-\lambda_{s}w_{t}[j]\right)\frac{1}{\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)!}\left(\lambda_{s}w_{t}[j]\right)^{\frac{\kappa}{m_{t}^{2}[j]}-2}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ -\lambda_{s}w_{t}[j]-\log\left(\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)!\right)+\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)\log\left(\lambda_{s}w_{t}[j]\right)\\\ \end{multlined}C_{t+1}^{D1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(\frac{\kappa}{X_{i}[j]+2}=m_{t}^{2}[j]\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(X_{t}[j]=\frac{\kappa}{m_{t}^{2}[j]}-2\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\Pr_{X_{t}[j]\sim\text{Pois}(\lambda_{s}w_{t}[j])}\left(X_{t}[j]=\frac{\kappa}{m_{t}^{2}[j]}-2\right)\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\exp\left(-\lambda_{s}w_{t}[j]\right)\frac{1}{\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)!}\left(\lambda_{s}w_{t}[j]\right)^{\frac{\kappa}{m_{t}^{2}[j]}-2}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ -\lambda_{s}w_{t}[j]-\log\left(\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)!\right)+\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)\log\left(\lambda_{s}w_{t}[j]\right)\\\ $ Using Stirling approximation for the second term, we get $\begin{multlined}C_{t+1}^{R}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ -\lambda_{s}w_{t}[j]-\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)\log\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)+\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)+\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)\log\left(\lambda_{s}w_{t}[j]\right)\\\ \end{multlined}C_{t+1}^{R}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ -\lambda_{s}w_{t}[j]-\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)\log\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)+\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)+\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)\log\left(\lambda_{s}w_{t}[j]\right)\\\ $ In order to find the optimum, we differentiate the right side with respect to $\kappa$ and equate it to zero: $\begin{multlined}-\frac{1}{m_{t}^{2}[j]}\log\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)-\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)\frac{\frac{1}{m^{2}_{t}[j]}}{\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)}+\frac{1}{m_{t}^{2}[j]}+\frac{1}{m_{t}^{2}[j]}\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ -\frac{1}{m_{t}^{2}[j]}\log\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)-\frac{1}{m_{t}^{2}[j]}+\frac{1}{m_{t}^{2}[j]}+\frac{1}{m_{t}^{2}[j]}\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ -\log\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)+\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ \log\left(\frac{\lambda_{s}w_{t}[j]m_{t}^{2}[j]}{\kappa-2m_{t}^{2}[j]}\right)=0\\\ \frac{\lambda_{s}w_{t}[j]m_{t}^{2}[j]}{\kappa-2m_{t}^{2}[j]}=1\\\ \end{multlined}-\frac{1}{m_{t}^{2}[j]}\log\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)-\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)\frac{\frac{1}{m^{2}_{t}[j]}}{\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)}+\frac{1}{m_{t}^{2}[j]}+\frac{1}{m_{t}^{2}[j]}\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ -\frac{1}{m_{t}^{2}[j]}\log\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)-\frac{1}{m_{t}^{2}[j]}+\frac{1}{m_{t}^{2}[j]}+\frac{1}{m_{t}^{2}[j]}\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ -\log\left(\frac{\kappa}{m_{t}^{2}[j]}-2\right)+\log\left(\lambda_{s}w_{t}[j]\right)=0\\\ \log\left(\frac{\lambda_{s}w_{t}[j]m_{t}^{2}[j]}{\kappa-2m_{t}^{2}[j]}\right)=0\\\ \frac{\lambda_{s}w_{t}[j]m_{t}^{2}[j]}{\kappa-2m_{t}^{2}[j]}=1\\\ $ Thus $\kappa=m_{t}^{2}[j]\left(\underbrace{\lambda_{s}w_{t}[j]}_{\text{expected number of users}}+2\right)$ ∎ See 3 ###### Proof. We start by considering a relay that falls into case 1 for its whole operation and hence is never bottlenecked. From Theorem 1, we can write: $\begin{multlined}C_{t+1}^{D}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \prod_{i=0}^{t}\Pr_{X_{i}[j]\sim\text{Pois}(\lambda_{s}w_{i}[j])}\left(\kappa- X_{i}[j]C_{avg}=m_{i}^{1}[j]\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \prod_{i=0}^{t}\Pr_{X_{i}[j]\sim\text{Pois}(\lambda_{s}w_{i}[j])}\left(X_{i}[j]=\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\prod_{i=0}^{t}\Pr_{X_{i}[j]\sim\text{Pois}(\lambda_{s}w_{i}[j])}\left(X_{i}[j]=\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\prod_{i=0}^{t}\exp\left(-\lambda_{s}w_{i}[j]\right)\frac{1}{\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)!}\left(\lambda_{s}w_{i}[j]\right)^{\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\sum_{i=0}^{t}-\lambda_{s}w_{i}[j]-\log\left(\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)!\right)+\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\log\left(\lambda_{s}w_{i}[j]\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\sum_{i=0}^{t}-\lambda_{s}w_{i}[j]-\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\log\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)+\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)+\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\log\left(\lambda_{s}w_{i}[j]\right)\\\ \end{multlined}C_{t+1}^{D}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \prod_{i=0}^{t}\Pr_{X_{i}[j]\sim\text{Pois}(\lambda_{s}w_{i}[j])}\left(\kappa- X_{i}[j]C_{avg}=m_{i}^{1}[j]\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \prod_{i=0}^{t}\Pr_{X_{i}[j]\sim\text{Pois}(\lambda_{s}w_{i}[j])}\left(X_{i}[j]=\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\prod_{i=0}^{t}\Pr_{X_{i}[j]\sim\text{Pois}(\lambda_{s}w_{i}[j])}\left(X_{i}[j]=\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\ \log\left(\prod_{i=0}^{t}\exp\left(-\lambda_{s}w_{i}[j]\right)\frac{1}{\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)!}\left(\lambda_{s}w_{i}[j]\right)^{\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}}\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\sum_{i=0}^{t}-\lambda_{s}w_{i}[j]-\log\left(\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)!\right)+\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\log\left(\lambda_{s}w_{i}[j]\right)\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\sum_{i=0}^{t}-\lambda_{s}w_{i}[j]-\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\log\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)+\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)+\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\log\left(\lambda_{s}w_{i}[j]\right)\\\ $ We differentiate the right side with respect to $\kappa$ and equate it to zero: $\begin{multlined}\sum_{i=0}^{t}-\frac{1}{C_{avg}}\log\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)-\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\frac{\frac{1}{C_{avg}}}{\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)}+\frac{1}{C_{avg}}+\frac{1}{C_{avg}}\log\left(\lambda_{s}w_{i}[j]\right)=0\\\ \sum_{i=0}^{t}-\frac{1}{C_{avg}}\log\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)-\frac{1}{C_{avg}}+\frac{1}{C_{avg}}+\frac{1}{C_{avg}}\log\left(\lambda_{s}w_{i}[j]\right)=0\\\ \sum_{i=0}^{t}-\log\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)+\log\left(\lambda_{s}w_{i}[j]\right)=0\\\ \end{multlined}\sum_{i=0}^{t}-\frac{1}{C_{avg}}\log\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)-\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)\frac{\frac{1}{C_{avg}}}{\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)}+\frac{1}{C_{avg}}+\frac{1}{C_{avg}}\log\left(\lambda_{s}w_{i}[j]\right)=0\\\ \sum_{i=0}^{t}-\frac{1}{C_{avg}}\log\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)-\frac{1}{C_{avg}}+\frac{1}{C_{avg}}+\frac{1}{C_{avg}}\log\left(\lambda_{s}w_{i}[j]\right)=0\\\ \sum_{i=0}^{t}-\log\left(\frac{\kappa- m_{i}^{1}[j]}{C_{avg}}\right)+\log\left(\lambda_{s}w_{i}[j]\right)=0\\\ $ Using linearization to solve for the optimum $\kappa$, we have that $\begin{multlined}f^{\prime}_{t}(\kappa)=\sum_{i=0}^{t}-\log\left(\kappa- m_{i}^{1}[j]\right)+\sum_{i=0}^{t}\log\left(\lambda_{s}w_{i}[j]C_{avg}\right)\\\ =f_{t-1}^{\prime}(\kappa)+\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa- m_{t}^{1}[j]\right)\\\ \end{multlined}f^{\prime}_{t}(\kappa)=\sum_{i=0}^{t}-\log\left(\kappa- m_{i}^{1}[j]\right)+\sum_{i=0}^{t}\log\left(\lambda_{s}w_{i}[j]C_{avg}\right)\\\ =f_{t-1}^{\prime}(\kappa)+\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa- m_{t}^{1}[j]\right)\\\ $ We know that $f^{\prime}_{t}(\kappa_{t})=0$ and $f^{\prime}_{t-1}(\kappa_{t-1})=0$ thus $\begin{multlined}f^{\prime}_{t}(\kappa_{t-1})=f_{t-1}^{\prime}(\kappa_{t-1})+\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa_{t-1}-m_{t}^{1}[j]\right)\\\ =\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa_{t-1}-m_{t}^{1}[j]\right)\\\ \end{multlined}f^{\prime}_{t}(\kappa_{t-1})=f_{t-1}^{\prime}(\kappa_{t-1})+\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa_{t-1}-m_{t}^{1}[j]\right)\\\ =\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa_{t-1}-m_{t}^{1}[j]\right)\\\ $ We also have $f^{\prime\prime}_{t}(\kappa)=-\sum_{i=0}^{t}\frac{1}{\kappa- m_{i}^{1}[j]}=f^{\prime\prime}_{t-1}(\kappa)-\frac{1}{\kappa-m_{t}^{1}[j]}$. By linearization we have $f^{\prime}_{t}(\kappa_{t})=f^{\prime}_{t}(\kappa_{t-1})+f^{\prime\prime}_{t}(\kappa_{t-1})(\kappa_{t}-\kappa_{t-1})$ and thus, $\begin{multlined}\kappa_{t}=\frac{f^{\prime}_{t}(\kappa_{t})-f^{\prime}_{t}(\kappa_{t-1})}{f^{\prime\prime}_{t}(\kappa_{t-1})}+\kappa_{t-1}\end{multlined}\kappa_{t}=\frac{f^{\prime}_{t}(\kappa_{t})-f^{\prime}_{t}(\kappa_{t-1})}{f^{\prime\prime}_{t}(\kappa_{t-1})}+\kappa_{t-1}$ Thus we can find an iterative solution of the optimization, $\begin{multlined}\kappa_{0}=m_{0}^{1}[j]+\lambda_{s}w_{0}[j]C_{avg}\hskip 8.53581pt\text{and}\\\ \kappa_{t}=\frac{\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa_{t-1}-m_{t}^{1}[j]\right)}{\sum_{i=0}^{t}\frac{1}{\kappa_{t-1}-m_{i}^{1}[j]}}+\kappa_{t-1}\end{multlined}\kappa_{0}=m_{0}^{1}[j]+\lambda_{s}w_{0}[j]C_{avg}\hskip 8.53581pt\text{and}\\\ \kappa_{t}=\frac{\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa_{t-1}-m_{t}^{1}[j]\right)}{\sum_{i=0}^{t}\frac{1}{\kappa_{t-1}-m_{i}^{1}[j]}}+\kappa_{t-1}$ Finding the steady state convergence of the above iterative formulation: $\begin{multlined}\kappa=\frac{\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa- m_{t}^{1}[j]\right)}{\sum_{i=0}^{t}\frac{1}{\kappa-m_{i}^{1}[j]}}+\kappa\\\ \log(\lambda_{s}w_{t}[j]C_{avg})=\ log(\kappa-m_{t}^{1}[j])\\\ \lambda_{s}w_{t}[j]C_{avg}=\kappa-m_{t}^{1}[j]\\\ \end{multlined}\kappa=\frac{\log\left(\lambda_{s}w_{t}[j]C_{avg}\right)-\log\left(\kappa- m_{t}^{1}[j]\right)}{\sum_{i=0}^{t}\frac{1}{\kappa-m_{i}^{1}[j]}}+\kappa\\\ \log(\lambda_{s}w_{t}[j]C_{avg})=\ log(\kappa-m_{t}^{1}[j])\\\ \lambda_{s}w_{t}[j]C_{avg}=\kappa-m_{t}^{1}[j]\\\ $ (33) Hence the expected value of the estimate $E(\kappa_{t})=C^{*}[j]$ since the right hand side of Equation 33 is equal to the actual capacity of the relay if it was never bottlenecked on average. Now we consider case 2, where the relay is always bottlenecked for its whole measurement history. In this proof we will drop the upperscript $2$ from the measurement since we are only dealing with the second measurement of a relay.As we derived in eq. 28, we know that for any $j\in[n]$, the weight at iteration $(t+1)$ should satisfy the following equation: $C^{D}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{\Big{(}\frac{\kappa}{m_{i}[j]}-2\Big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa}{m_{i}[j]}-2}$ (34) Since the logarithm function is a strictly increasing function, the maximum likelihood estimate of the capacity of a relay $j\in[n]$ using full history can be found: $\begin{multlined}C^{D}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{\Big{(}\frac{\kappa}{m_{i}[j]}-2\Big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa}{m_{i}[j]}-2}\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{(\frac{\kappa}{m_{i}[j]})^{2}}{\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa}{m_{i}[j]}}(\lambda_{s}w_{i}[j])^{-2}\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\sum_{i=0}^{t}-\lambda_{s}w_{i}[j]+2\log\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}-\log\bigg{(}\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}!\bigg{)}+\frac{\kappa}{m_{i}[j]}\log(\lambda_{s}w_{i}[j])-2\log(\lambda_{s}w_{i}[j])\end{multlined}C^{D}_{t+1}[j]=\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{1}{\Big{(}\frac{\kappa}{m_{i}[j]}-2\Big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa}{m_{i}[j]}-2}\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\prod_{i=0}^{t}e^{-\lambda_{s}w_{i}[j]}\frac{(\frac{\kappa}{m_{i}[j]})^{2}}{\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}!}(\lambda_{s}w_{i}[j])^{\frac{\kappa}{m_{i}[j]}}(\lambda_{s}w_{i}[j])^{-2}\\\ =\underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\sum_{i=0}^{t}-\lambda_{s}w_{i}[j]+2\log\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}-\log\bigg{(}\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}!\bigg{)}+\frac{\kappa}{m_{i}[j]}\log(\lambda_{s}w_{i}[j])-2\log(\lambda_{s}w_{i}[j])$ (35) Using Stirling’s approximation, we have $\log(x!)\approx x\log(x)-x$. Thus substituting in eq. 35: $\begin{multlined}C^{H}_{t+1}[j]=\\\ \underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\sum_{i=0}^{t}-\lambda_{s}w_{i}[j]+2\log\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}-\frac{\kappa}{m_{i}[j]}\log\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}+\frac{\kappa}{m_{i}[j]}+\frac{\kappa}{m_{i}[j]}\log(\lambda_{s}w_{i}[j])-2\log(\lambda_{s}w_{i}[j])\end{multlined}C^{H}_{t+1}[j]=\\\ \underset{\kappa\in\mathcal{C}}{\operatorname{argmax}}\hskip 2.84544pt\sum_{i=0}^{t}-\lambda_{s}w_{i}[j]+2\log\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}-\frac{\kappa}{m_{i}[j]}\log\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}+\frac{\kappa}{m_{i}[j]}+\frac{\kappa}{m_{i}[j]}\log(\lambda_{s}w_{i}[j])-2\log(\lambda_{s}w_{i}[j])$ (36) Hence in order to find $C^{H}_{t+1}[j]$, we differentiate the right hand side of eq. 36 with respect to $\kappa$, and find the value of $C^{H}_{t+1}[j]$ for which the derivative is zero. $\begin{multlined}\sum_{i=0}^{t}\frac{2}{m_{i}[j]}\frac{1}{\frac{\kappa}{m_{i}[j]}}-\frac{1}{m_{i}[j]}\log\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}-\frac{1}{m_{i}[j]}+\frac{1}{m_{i}[j]}+\frac{1}{m_{i}[j]}\log(\lambda_{s}w_{i}[j])=0\\\ \sum_{i=0}^{t}\frac{2}{\kappa}-\frac{1}{m_{i}[j]}\log(\kappa)+\frac{1}{m_{i}[j]}\log(m_{i}[j])+\frac{1}{m_{i}[j]}\log(\lambda_{s}w_{i}[j])=0\\\ \sum_{i=0}^{t}\frac{2}{\kappa}-\frac{1}{m_{i}[j]}\log(\kappa)+\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])=0\\\ \sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\kappa)=\sum_{i=0}^{t}\frac{2}{\kappa}+\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])\\\ \log(\kappa)\Big{(}\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\Big{)}=\frac{2(t+1)}{\kappa}+\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])\\\ \log(\kappa)=\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}+\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}\\\ \log(\kappa)-\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}=\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}\\\ \kappa e^{-\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}}=e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ \frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{2(t+1)}\kappa e^{-\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}}=\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{2(t+1)}e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ \end{multlined}\sum_{i=0}^{t}\frac{2}{m_{i}[j]}\frac{1}{\frac{\kappa}{m_{i}[j]}}-\frac{1}{m_{i}[j]}\log\Big{(}\frac{\kappa}{m_{i}[j]}\Big{)}-\frac{1}{m_{i}[j]}+\frac{1}{m_{i}[j]}+\frac{1}{m_{i}[j]}\log(\lambda_{s}w_{i}[j])=0\\\ \sum_{i=0}^{t}\frac{2}{\kappa}-\frac{1}{m_{i}[j]}\log(\kappa)+\frac{1}{m_{i}[j]}\log(m_{i}[j])+\frac{1}{m_{i}[j]}\log(\lambda_{s}w_{i}[j])=0\\\ \sum_{i=0}^{t}\frac{2}{\kappa}-\frac{1}{m_{i}[j]}\log(\kappa)+\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])=0\\\ \sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\kappa)=\sum_{i=0}^{t}\frac{2}{\kappa}+\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])\\\ \log(\kappa)\Big{(}\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\Big{)}=\frac{2(t+1)}{\kappa}+\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])\\\ \log(\kappa)=\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}+\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}\\\ \log(\kappa)-\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}=\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}\\\ \kappa e^{-\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}}=e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ \frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{2(t+1)}\kappa e^{-\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}}=\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{2(t+1)}e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ $ (37) Letting $z=\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}$ in eq. 37, we have: $\begin{multlined}\frac{1}{z}e^{-z}=\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{2(t+1)}e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ \frac{1}{ze^{z}}=\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{2(t+1)}e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ ze^{z}=\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}e^{-\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ \end{multlined}\frac{1}{z}e^{-z}=\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{2(t+1)}e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ \frac{1}{ze^{z}}=\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{2(t+1)}e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ ze^{z}=\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}e^{-\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ $ (38) We know that the inverse image of the function $ze^{z}$ is the Lambert W function which has real solutions along its principal branch for $z>-\frac{1}{e}$, denoted $W_{0}$. Thus we can solve for $z$: $\begin{multlined}\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}=z=W_{0}\Bigg{(}\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}e^{-\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\Bigg{)}\\\ \end{multlined}\frac{2(t+1)}{\kappa(\sum_{i=0}^{t}\frac{1}{m_{i}[j]})}=z=W_{0}\Bigg{(}\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}e^{-\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\Bigg{)}\\\ $ (39) And hence solving for $\kappa$: $\begin{multlined}C^{D}_{t+1}[j]=\kappa=\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}\frac{1}{W_{0}\Bigg{(}\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}e^{-\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\Bigg{)}}\\\ \end{multlined}C^{D}_{t+1}[j]=\kappa=\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}\frac{1}{W_{0}\Bigg{(}\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}e^{-\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\Bigg{)}}\\\ $ (40) $\begin{multlined}C^{D}_{t+1}[j]=\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}\frac{1}{W_{0}\Bigg{(}\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}e^{-\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\Bigg{)}},\\\ \end{multlined}C^{D}_{t+1}[j]=\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}\frac{1}{W_{0}\Bigg{(}\frac{2(t+1)}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}e^{-\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\Bigg{)}},\\\ $ (41) where $W_{0}$ is the Lambert $W$ function along the principal branch. The Lambert $W$ function is the multi-valued complex function $(ze^{z})^{-1}$ and $W_{0}$ is the unique-valued real function that takes the unique real value of $W$ when $z>\frac{-1}{e}$. Implementations of Lambert function exist in multiple software libraries 555https://kite.com/python/docs/mpmath.lambertw. $W_{0}$ has the following Taylor series expansion for $z$ in the neighborhood of 0: $W_{0}(z)=z+o(z^{2})$. Moreover, the argument of $W_{0}$ is small if the rate of users arrival to the network $\lambda_{s}$ is large enough. Hence, the Taylor expansion around zero is valid and therefore: $\begin{multlined}C^{D}_{t+1}[j]\approx e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ =e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\frac{m_{i}[j]\lambda_{s}w_{i}[j]C^{*}[j]}{C^{*}[j]})}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ =e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\frac{m_{i}[j]\lambda_{s}w_{i}[j]}{C^{*}[j]})+\frac{1}{m_{i}[j]}\log(C^{*}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ =e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\frac{m_{i}[j]\lambda_{s}w_{i}[j]}{C^{*}[j]})}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}e^{\frac{\log(C^{*}[j])\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ =C^{*}[j]e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\frac{m_{i}[j]\lambda_{s}w_{i}[j]}{C^{*}[j]})}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}.\end{multlined}C^{D}_{t+1}[j]\approx e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(m_{i}[j]\lambda_{s}w_{i}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ =e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\frac{m_{i}[j]\lambda_{s}w_{i}[j]C^{*}[j]}{C^{*}[j]})}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ =e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\frac{m_{i}[j]\lambda_{s}w_{i}[j]}{C^{*}[j]})+\frac{1}{m_{i}[j]}\log(C^{*}[j])}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ =e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\frac{m_{i}[j]\lambda_{s}w_{i}[j]}{C^{*}[j]})}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}e^{\frac{\log(C^{*}[j])\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}\\\ =C^{*}[j]e^{\frac{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}\log(\frac{m_{i}[j]\lambda_{s}w_{i}[j]}{C^{*}[j]})}{\sum_{i=0}^{t}\frac{1}{m_{i}[j]}}}.$ (42) We refer the reader to the Appendix of [5] for a proof of the expected value of the closed form found above since this form exactly matches the form derived for $\mathit{MLEFlow\text{-}CF}$. ∎ See 5 ###### Proof. We refer the reader to the Appendix of [5] for a complete proof of the theorem. ∎
# Nuclear Volterra composition operators between Bloch and weighted type spaces Aakriti Sharma Department of Mathematics, Central University of Jammu, Central University of Jammu, (Bagla) Rahya-Suchani, Samba-181143, J&K, India. <EMAIL_ADDRESS>and Ajay K. Sharma Department of Mathematics, Central University of Jammu, Central University of Jammu, (Bagla) Raya- Suchani, Samba-181143, J&K, India<EMAIL_ADDRESS> ###### Abstract. In this paper, we completely characterize nuclear Volterra composition operators $T^{\phi}_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ and $S^{\phi}_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ acting between weighted type spaces in terms of the symbols $g$ and $\phi$ of $T^{\phi}_{g}$ and $S^{\phi}_{g}$ and weights $\nu$ and $\mu$, when the weights $\nu$ and $\mu$ are normal weights in the sense of Shields and Williams. Moreover, nuclear Volterra composition operators acting between little weighted type spaces and Bloch spaces of order $\beta$ are also characterized. ###### Key words and phrases: Nuclear operator; Volterra composition operator, weighted Banach space, Bergman space, weighted Bloch space, little weighted Banach space, little weighted Bloch space. ###### 2010 Mathematics Subject Classification: 47B33, 47B38, 46E10, 32A37. The first author is thankful to CSIR(India), Grant Number (File no. 09/1231(0001)/2019-EMR-I). The second author is thankful to DST(SERB) for the project grant under MATRICS scheme. Grant No. MTR/2018/000479. ## 1\. Introduction Denote by ${H}(\mathbb{D})$, the class of all analytic functions on $\mathbb{D}$, where $\mathbb{D}$ is the open unit disk in the complex plane $\mathbb{C}$. For $g\in H(\mathbb{D}),$ Ch. Pommerenke [28] introduced an integral operator $T_{g}:H(\mathbb{D})\to H(\mathbb{D})$ which maps every $f\in H(\mathbb{D})$ to a function vanishing at $0,$ defined as $T_{g}f(z)=\int_{0}^{z}f(\zeta)dg(\zeta)=\int_{0}^{z}f(\zeta)g^{\prime}(z)d\zeta,\;\;\;z\in\mathbb{D}.$ Another integral operator, generally known as a companion operator of $T_{g}$ was defined by Yoneda [31] as $S_{g}f(z)=\int_{0}^{z}f^{\prime}(\zeta)g(\zeta)d\zeta,\;\;\;z\in\mathbb{D}.$ The operators $T_{g}$ and $S_{g}$ can be viewed as cousins of Hankel and Toeplitz operators and they appeared in the literature under various names such as Volterra operators, Cesaro operators, Riemann Stieltjes operators and integration operators. For any $g\in{H}(\mathbb{D})$ and $\phi$ an analytic self-map of $\mathbb{D}$, the Volterra composition operators $T^{\phi}_{g}$ and $S^{\phi}_{g}$ are defined, respectively on ${H}(\mathbb{D})$ by $(T^{\phi}_{g}f)(z)=\int_{0}^{z}f(\phi(\zeta))g^{\prime}(\zeta)d\zeta,\;\;\;z\in\mathbb{D}$ and $(S^{\phi}_{g}f)(z)=\int_{0}^{z}f^{\prime}(\phi(\zeta))g(\zeta)d\zeta,\;\;\;z\in\mathbb{D}.$ Recently, these type of operators are considered on spaces of analytic functions by several authors, see [1]-[6], [10],[11],[16],[20]-[22] and references therein. Motivated by results in [8] and [17], in this paper, we characterize nuclear Volterra composition operators between Bloch and weighted type spaces of analytic functions, when the weights are normal. Recall that a radial weight $\nu$ is a non-increasing, non-negative continuous function defined on $\mathbb{D}$ such that $\nu(z)=\nu(|z|)$ for all $z\in\mathbb{D}.$ A radial weight $\nu$ is normal in the sense of Shields and Williams if it satisfies following two conditions: 1. ($\mathbf{I}$). There exists $\beta>0$ such that the function $r\rightarrow\frac{\nu(r)}{(1-r^{2})^{\beta}}$ is almost increasing, or equivalently, $\inf_{n}\frac{\nu(1-2^{-(n+1)})}{\nu(1-2^{-n})}>0$. 2. ($\mathbf{II}$). There exists $\gamma>0$ such that the function $r\rightarrow\frac{\nu(r)}{(1-r^{2})^{\gamma}}$ is almost decreasing, or equivalently, these is some $k\in\mathbb{N}$ such that $\limsup_{n}\frac{\nu(1-2^{-(n+k)})}{\nu(1-2^{-n})}<1$. It is clear that the standard weights $\nu_{\alpha}(z)=(1-|z|^{2})^{\alpha},(\alpha>0)$ are normal weights. Throughout this paper, we denote the set of all normal weights on $\mathbb{D}$ by $N_{W}(\mathbb{D}).$ For more about normal weights, we refer to [29],[6] and [15]. The weighted Banach spaces $H^{\infty}_{\nu}$ and $H_{\nu}^{0}$ are defined as follows: $\mathcal{H}_{\nu}^{\infty}=\bigg{\\{}{f\in{H}(\mathbb{D}):\|f\|_{\mathcal{H}_{\nu}^{\infty}}=\sup_{z\in\mathbb{D}}\nu(z)|f(z)|<\infty\bigg{\\}}}$ and $H_{\nu}^{0}=\bigg{\\{}{f\in{H}(\mathbb{D}):\lim_{|z|\rightarrow 1^{-}}\nu(z)|f(z)|=0\bigg{\\}}}.$ The weighted Bloch space $\mathcal{B}_{\nu}$ is the space of all functions $f$ in ${H}(\mathbb{D})$ such that $\|f\|_{\nu}=\sup_{z\in\mathbb{D}}\nu(z)|f^{\prime}(z)|<{\infty}.$ Note that $\|\cdot\|_{\nu}$ is a semi-norm on $\mathcal{B}_{\nu}$ and a norm can be defined on $\mathcal{B}_{\nu}$ as $\|f\|_{\mathcal{B}_{\nu}}=|f(0)|+\sup_{z\in\mathbb{D}}\nu(z)|f^{\prime}(z)|.$ Endowed with the norm $\|\cdot\|_{\mathcal{B}_{\nu}}$, $\mathcal{B}_{\nu}$ becomes a Banach space. Similarly, the little weighted Bloch space $\mathcal{B}^{0}_{\nu}$ is defined as $\lim_{|z|\rightarrow 1^{-}}\nu(z)|f^{\prime}(z)|=0.$ If $\lim_{|z|\rightarrow 1^{-}}\nu(z)=0$, then the weight $\nu$ is called typical and for typical weights $H_{\nu}^{\infty}$ is the bidual space of $H_{\nu}^{0}$. If the weight $\nu$ satisfies the condition ($\mathbf{I}$), then $\nu$ satisfies condition ($\mathbf{II}$) if and only if $H_{\nu}^{0}=\mathcal{B}^{0}_{\nu(z)(1-|z|^{2})}$, see [25]. In particular, if $\nu\in N_{W}(\mathbb{D})$, then $H_{\nu}^{0}=\mathcal{B}_{\nu(z)(1-|z|^{2})}^{0}$ and by duality, $H_{\nu}^{\infty}=\mathcal{B}_{\nu(z)(1-|z|^{2})}$, see [23]. Moreover, if $\nu\in N_{W}(\mathbb{D})$, then the topological isomorphisms $H_{\nu}^{0}\simeq c_{0}$ and $H_{\nu}^{\infty}\simeq l_{\infty}$ hold, see [8],[30] and [23]-[26]. Next, let us recall a duality pairing between $\mathcal{H}^{\infty}_{\nu}$ and a weighted Bergman space, which was established by Shields and Williams in [29]. Following [29], let $\nu$ be a typical weight and $\omega$ a radial, positive and continuous function which satisfies $\int_{0}^{1}\omega(r)dr<\infty.$ Let $A^{1}_{\omega}$ be the subspace of $H(\mathbb{D})$ consists of functions $f$ such that $\|f\|_{A^{1}_{\omega}}=\int_{\mathbb{D}}|f(z)|\omega(z)dA(z)<\infty,$ where $dA$ is the normalized Lebesgue measure on $\mathbb{D}.$ For a normal weight $\nu$ and $\omega$ a function defined as above, the pair $\\{\nu,\omega\\}$ is called a normal pair if $\nu(r)\omega(r)=(1-r^{2})^{\alpha},\;\;\;0\leq r<1$ (1.1) for some $\alpha>\beta-1$, where $\beta$ is a positive real number defined as in $(\mathbf{I}).$ For a normal pair $\\{\nu,\omega\\}$, the following pairing between $\mathcal{H}^{\infty}_{\nu}$ and $A^{1}_{\omega}$ is defined in [29], see p. 292. $\displaystyle(f,g)$ $\displaystyle=\int_{\mathbb{D}}f(z)g(\bar{z})(1-|z|^{2})^{\alpha}dA(z)=\int_{\mathbb{D}}f(\bar{z})g({z})(1-|z|^{2})^{\alpha}dA(z).$ (1.2) The pairing $(\cdot,\cdot)$ define a duality relation $(\mathcal{H}^{0}_{\nu})^{\prime}=A^{1}_{\omega},$ see [29], Theorem 2, p. 296. Thus we also have $(\mathcal{H}^{0}_{\nu})^{\prime\prime}=\mathcal{H}^{\infty}_{\nu}.$ Moreover, if $K_{\zeta}(z)=(\alpha+1)/(1-{\zeta}z)^{\alpha+2},$ $\alpha>-1$ is the kernel function and $\\{\nu,\omega\\}$ a normal pair, then $K_{\zeta}\in\mathcal{H}^{0}_{\nu}\cap A^{1}_{\omega},$ $g(\zeta)=(K_{\zeta},g)$ for all $g\in A^{1}_{\omega}$ and $f(\zeta)=(f,K_{\zeta})$ for all $f\in\mathcal{H}^{\infty}_{\nu},$ see Lemma 10 in [29]. Recall that a linear operator $T$ between two Banach space $X$ and $Y$ is 1. (a) Nuclear if there is a sequence $(x^{\prime}_{k})\subset X^{\prime}$ and a sequence $(y_{k})\subset Y$ such that $\sum_{k}\|x^{\prime}_{k}\|\|y_{k}\|<{\infty}\mbox{ and }T=\sum_{k=1}^{\infty}x^{\prime}_{k}\otimes y_{k},$ where $x^{\prime}_{k}\otimes y_{k}:X\rightarrow Y$ is the mapping defined as $x\rightarrow x^{\prime}_{k}(x)y_{k}$. 2. (b) Absolutely summing if there is $C>0$ such that for all finite sequences $(x_{i})_{i=1}^{n}\subset X$ we have $\sum_{i=1}^{n}\|Tx_{i}\|\leq C\sup_{\|x^{\prime}\|\leq 1}\sum_{i=1}^{n}|x^{\prime}(x_{i})|=C\sup_{|\eta_{i}|=1}\|\sum_{i=1}^{n}\eta_{i}x_{i}\|.$ Moreover, it is well known that any nuclear operator is compact and absolutely summing. Also if $X=c_{0}$ or $l_{\infty}$, then a linear operator $T:X\rightarrow Y$ is absolutely summing if and only if it is nuclear, see [17]. Throughout this paper, $C$ denotes a positive constant, the exact value of which may be not be same at each occurrence. The expression $A\lesssim B$ means that there is a positive constant $C$ such that $A\leq CB$ and the expression $A\asymp B$ means that there are positive constants $C$ and $D$ such that $CB\leq A\leq DB$. ## 2\. Nuclear Volterra composition operators In this section, we characterize nuclear Volterra composition operator between Bloch and weighted spaces of analytic functions. Next we present characterizations of bounded Volterra composition operators between weighted spaces of analytic functions. The arguments are standard and may possibly appear in literature in a more general sense, so we give an outline proof for completeness. ###### Theorem 1. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $g\in H(\mathbb{D})$ and $\phi$ a self-map of $\mathbb{D}$. Then 1. ($i$) The following are equivalent. 1. ($a$) $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded. 2. ($b$) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded. 3. ($c$) $g$, $\nu,$ $\mu$ and $\phi$ satisfy the following condition: $M_{1}=\sup_{z\in\mathbb{D}}(1-|z|^{2})\frac{\mu(z)}{\nu(\phi(z))}|g^{\prime}(z)|<\infty.$ (2.1) Moreover, $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}$ is bounded if and only if $g\in\mathcal{B}^{0}_{(1-|z|^{2})\mu(z)}$ and (2.1) holds. 2. ($ii$) The following are equivalent. 1. ($d$) $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded. 2. ($e$) $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded. 3. ($f$) $g$, $\nu,$ $\mu$ and $\phi$ satisfy the following condition: $\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})}{(1-|\phi(z)|^{2})}\frac{\mu(z)}{\nu(\phi(z))}|g(z)|<\infty.$ (2.2) Moreover, $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}$ is bounded if and only if $g\in\mathcal{H}^{0}_{(1-|z|^{2})\mu(z)}$ and (2.2) holds. ###### Proof. $(i)$ Since $\mu$ is a normal weight, so $\mathcal{H}_{\mu(z)}=\mathcal{B}_{(1-|z|^{2})\mu(z)}.$ Thus boundedness of $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is equivalent to the boundedness of $M_{g^{\prime}}C_{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{(1-|z|^{2})\mu(z)}.$ By Proposition 2 in [15], every normal weight is essential, that is there is a constant $C>0$ such that $\nu(z)\leq\tilde{\nu}(z)\leq C\nu(z)$ for each $z\in\mathbb{D}$. $(a)\Leftrightarrow(c)$ Then by Proposition 3.1 in [9], we have that $(a)\Leftrightarrow(c).$ It is obvious that $(a)\Rightarrow(b).$ Thus to complete the proof, we need to show that $(b)\Rightarrow(c).$ $(b)\Rightarrow(c).$ Suppose that $M_{g^{\prime}}C_{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{(1-|z|^{2})\mu(z)}$ is bounded. For a fixed $\zeta\in\mathbb{D},$ consider the function $f_{\zeta}$ defined on $\mathbb{D}$ as $f_{\zeta}(z)=\frac{(1-|\phi(\zeta)|^{2})^{\beta+1}}{\nu(\phi(\zeta))(1-\overline{\phi(\zeta)}z)^{\beta+1}},$ where $\beta>0$ is as in condition $(\mathbf{I})$ of normal weight. Then it is easy to see that $f_{\zeta}\in\mathcal{H}^{0}_{\nu}$ and $\sup_{\zeta\in\mathbb{D}}\|f_{\zeta}\|_{\mathcal{H}^{\infty}_{\nu}}\lesssim 1.$ Moreover, $f_{\zeta}(\phi(\zeta))=1/\nu(\phi(\zeta)).$ Thus by the boundedness of $M_{g^{\prime}}C_{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{(1-|z|^{2})\mu(z)}$, we have that $(1-|\zeta|^{2})\mu(\zeta)|g^{\prime}(\zeta)||f_{\zeta}(\phi(\zeta))|\leq\|M_{g^{\prime}}C_{\phi}f_{\zeta}\|_{\mathcal{H}^{\infty}_{(1-|z|^{2})\mu(z)}}\lesssim\|f_{\zeta}\|_{\mathcal{H}^{\infty}_{\nu}}\lesssim 1.$ Taking supermum over $\zeta\in\mathbb{D},$ we get (2.1). By Proposition 3.2 in [9], we have that $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}$ is bounded if and only if $g\in\mathcal{B}^{0}_{(1-|z|^{2})\mu(z)}$ and (2.1) holds. This completes the proof of $(i)$. $(ii)$ Thus boundedness of $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is equivalent to the boundedness of $M_{g}C_{\phi}D:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{(1-|z|^{2})\mu(z)},$ where $D$ is the differentiation operator. We write an outline proof of $(d)\Leftrightarrow(f)$ and $(e)\Leftrightarrow(f)$ The details of all other cases are omitted. $(d)\Leftrightarrow(f)$ By Theorem 7 in [27], we have that $(d)\Leftrightarrow(f).$ $(e)\Rightarrow(f).$ Suppose that $M_{g}C_{\phi}D:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{(1-|z|^{2})\mu(z)}$ is bounded. For a fixed $\zeta\in\mathbb{D},$ consider the function $f_{\zeta}$ defined as in case $(i).$ Then proceeding as in case $(i)$, we have that $f^{\prime}_{\zeta}(\phi(\zeta))=\overline{\phi(\zeta)}/(1-|\phi(\zeta)|^{2})\nu(\phi(\zeta)).$ Thus by the boundedness of $M_{g}C_{\phi}D:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{(1-|z|^{2})\mu(z)}$, we have that $(1-|\zeta|^{2})\mu(\zeta)|g(\zeta)||f^{\prime}_{\zeta}(\phi(\zeta))|\leq\|M_{g}C_{\phi}Df_{\zeta}\|_{\mathcal{H}^{\infty}_{(1-|z|^{2})\mu(z)}}\lesssim\|f_{\zeta}\|_{\mathcal{H}^{\infty}_{\nu}}\lesssim 1.$ Taking supermum over $\zeta\in\mathbb{D},$ we get $\sup_{z\in\mathbb{D}}\frac{(1-|\zeta|^{2})\mu(\zeta)}{(1-|\phi(\zeta)|^{2})\nu(\phi(\zeta))}|\phi(\zeta)||g(\zeta)|<\infty.$ Thus $\sup_{|\phi(\zeta)|>1/2}\frac{(1-|\zeta|^{2})\mu(\zeta)}{(1-|\phi(\zeta)|^{2})\nu(\phi(\zeta))}|g(\zeta)|<\infty.$ (2.3) The above condition easily yields (2.2) by bifurcating the supermum over $z\in\mathbb{D}$ in (2.2) as sum of supermum over $|\phi(\zeta)|\leq 1/2$ and supermum over $|\phi(\zeta)|>1/2$. Thus the proof can be finished, using the fact that the first term in this sum is definitely finite and the second term in this sum is finite by (2.3).∎ ###### Theorem 2. Let $\nu,\mu\in N_{W}(\mathbb{D}).$ If $g\in H(\mathbb{D})$ and $\phi$ a self- map of $\mathbb{D}$ such that $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded, then 1. ($i$) The following are equivalent. 1. ($a$) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is compact. 2. ($b$) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is weakly compact. 3. ($c$) The bi-transpose $(T_{g}^{\phi})^{\prime\prime}$ of $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ satisfies $(T^{\phi}_{g})^{\prime\prime}=T^{\phi}_{g}$. 2. ($ii$) If $g\in\mathcal{B}^{0}_{(1-|z|^{2})\mu(z)}$, then $T^{\phi}_{g}:\mathcal{H}_{\nu}^{0}\rightarrow\mathcal{H}_{\mu}^{0}$ is bounded and the bi-transpose $(T^{\phi}_{g})^{\prime\prime}$ of $T^{\phi}_{g}$ satisfies $(T^{\phi}_{g})^{\prime\prime}=T^{\phi}_{g}$. Moreover, $T^{\phi}_{g}:\mathcal{H}_{\nu}^{0}\rightarrow\mathcal{H}_{\mu}^{0}$ is compact if and only if $T^{\phi}_{g}(H^{\infty}_{\nu})\subset\mathcal{H}^{0}_{\mu}.$ ###### Proof. $(i)$ $(a)\Leftrightarrow(b)$ Since $\mathcal{H}^{\infty}_{\mu}=\mathcal{B}_{(1-|z|^{2})\mu(z)}$, so $T_{g}^{\phi}f\in H^{\infty}_{\mu}$ if and only if $M_{g^{\prime}}C_{\phi}f\in H^{\infty}_{(1-|z|^{2})\mu}.$ Moreover, $T_{g}^{\phi}f$ and $M_{g^{\prime}}C_{\phi}f$ have comparable norm. Thus compactness (and/or weak compactness) of $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is equivalent to compactness (and/or weak compactness) of $M_{g^{\prime}}C_{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}.$ By Theorem 1 in [7], $M_{g^{\prime}}C_{\phi}f$ is compact if and only if it is weakly compact. Therefore, it follows that $(a)$ and $(b)$ are equivalent. $(a)\Rightarrow(c)$ Assume that $(a)$ holds. Then by Gantmacher-Nakamura’s theorem, see Theorem 5.5 in [12], the bi-transpose $(T_{g}^{\phi})^{\prime\prime}$ of $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ acts continuously from $\mathcal{H}^{\infty}_{\nu}$ with $w^{*}$-topology to $\mathcal{H}^{\infty}_{\mu}$ with $w^{*}$-topology. Since pointwise topology $\tau_{p}$ is weaker than $w^{*}$-topology on $\mathcal{H}^{\infty}_{\nu}$ and $\mathcal{H}^{\infty}_{\mu}$, so $(T_{g}^{\phi})^{\prime\prime}$ and $T_{g}^{\phi}$ are both $w^{*}$ and $\tau_{p}$ continuous and they agree in $\mathcal{H}^{0}_{\nu}$ which is $w^{*}$ dense in $\mathcal{H}^{\infty}_{\nu}.$ Therefore, $(c)$ holds. $(c)\Rightarrow(b)$ Assume that the bi-transpose $(T_{g}^{\phi})^{\prime\prime}$ of $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ satisfies $(T^{\phi}_{g})^{\prime\prime}=T^{\phi}_{g}$. Then it holds that $(T^{\phi}_{g})^{\prime\prime}(\mathcal{H}^{\infty}_{\nu})\subset\mathcal{H}^{\infty}_{\mu}.$ Using the fact that a bounded linear operator $T$ acting from a Banach space $X$ to a Banach space $Y$ is weakly compact if and only if $T^{\prime\prime}(X^{\prime\prime})\subset Y,$ see p. 482 in [14], we have that $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is weakly compact. $(ii)$ If $g\in\mathcal{B}^{0}_{(1-|z|^{2})\mu(z)}$, then by Theorem 1, $T^{\phi}_{g}:\mathcal{H}_{\nu}^{0}\rightarrow\mathcal{H}_{\mu}^{0}$ is bounded. Moreover, by replacing $T_{g}$ in Lemma 1 in [6] by $T^{\phi}_{g}$, we can easily prove that $(T^{\phi}_{g})^{\prime\prime}=T^{\phi}_{g}$. Finally, by Gantmacher-Nakamura’s theorem it holds that the compactness of $T^{\phi}_{g}:\mathcal{H}_{\nu}^{0}\rightarrow\mathcal{H}_{\mu}^{0}$ is equivalent to $T^{\phi}_{g}(H^{\infty}_{\nu})\subset\mathcal{H}^{0}_{\mu}.$ This completes the proof. ∎ ###### Theorem 3. Let $\nu,\mu\in N_{W}(\mathbb{D}).$ If $g\in H(\mathbb{D})$ and $\phi$ a self- map of $\mathbb{D}$ such that $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded, then 1. ($i$) The following are equivalent. 1. ($a$) $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is compact. 2. ($b$) $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is weakly compact. 3. ($c$) The bi-transpose $(S_{g}^{\phi})^{\prime\prime}$ of $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ satisfies $(S^{\phi}_{g})^{\prime\prime}=S^{\phi}_{g}$. 2. ($ii$) If $g\in\mathcal{H}^{0}_{(1-|z|^{2})\mu(z)}$, then $S^{\phi}_{g}:\mathcal{H}_{\nu}^{0}\rightarrow\mathcal{H}_{\mu}^{0}$ is bounded and the bi-transpose $(S^{\phi}_{g})^{\prime\prime}$ of $S^{\phi}_{g}$ satisfies $(S^{\phi}_{g})^{\prime\prime}=S^{\phi}_{g}$. Moreover, $S^{\phi}_{g}:\mathcal{H}_{\nu}^{0}\rightarrow\mathcal{H}_{\mu}^{0}$ is compact if and only if $S^{\phi}_{g}(H^{\infty}_{\nu})\subset\mathcal{H}^{0}_{\mu}.$ ###### Proof. (i) Proceeding as in Theorem 2, $S_{g}^{\phi}f\in H^{\infty}_{\mu}$ if and only if $M_{g^{\prime}}C_{\phi}Df\in H^{\infty}_{(1-|z|^{2})\mu(z)}$ with comparable norm. Thus compactness (and/or weak compactness) of $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is equivalent to compactness (and/or weak compactness) of $M_{g^{\prime}}C_{\phi}D:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ which is further equivalent to is compactness (and/or weak compactness) of $M_{g^{\prime}}C_{\phi}:\mathcal{H}^{0}_{(1-|z|^{2})\mu(z)}\longrightarrow\mathcal{H}^{\infty}_{\mu}.$ Thus the first part of theorem can now be settles as the proof of first part of Theorem 2. (ii) If $g\in\mathcal{H}^{0}_{(1-|z|^{2})\mu(z)}$, then by Theorem 1, $S^{\phi}_{g}:\mathcal{H}_{\nu}^{0}\rightarrow\mathcal{H}_{\mu}^{0}$ is bounded. Moreover, by replacing $S_{g}$ in Lemma 2.12 in [21], we can easily prove that $(S^{\phi}_{g})^{\prime\prime}=S^{\phi}_{g}$. The last part can be settle by Gantmacher-Nakamura’s theorem as in the second part of Theorem 2.∎ Finally, we recall Piestsch’s theorem which plays an important role in the proof of main results of this paper. For more about the Piestsch’s theorem, we refer the readers to Theorem 2.12 of [13]. ###### Theorem 4. (Pietsch’s Theorem) Let $E$ and $F$ be Banach spaces and $1\leq p<+\infty$. A bounded linear operator $T:E\longrightarrow F$ is absolutely p-summing if and only if there is a constant $C$ and a regular Borel probability measure $\varrho$ on closed unit ball $B_{E^{*}}$ of $E^{*}$ with the weak∗-topology $\sigma(E^{*},E)$ such that for every $x\in E$, $\|T(x)\|\leq C\left(\int_{B_{E^{*}}}|\phi(x)|^{p}d\varrho(\phi)\right)^{1/p}.$ ###### Theorem 5. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in H(\mathbb{D})$ and $\phi$ a self-map of $\mathbb{D}$ be such that $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is nuclear. 2. (b) $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is nuclear. 3. (c) $g$, $\nu,$ $\mu$ and $\phi$ satisfy the following condition: $M_{\alpha}=\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+2}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Proof. $(c)\Rightarrow(a)$ Since $\mathcal{H}^{0}_{\nu}\simeq c_{0}$, so $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is nuclear if and only if $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is absolutely summing. Thus to complete the proof of the implication $(c)\Rightarrow(a)$, we need to prove that $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is absolutely summing. First note that if $\mu$ is normal, $\mathcal{H}^{\infty}_{\mu(z)}=\mathcal{B}_{\mu(z)(1-|z|^{2})}$, $\|T_{g}^{\phi}(f)\|_{\mathcal{H}^{\infty}_{\mu}}\asymp\|T_{g}^{\phi}(f)\|_{\mathcal{B}_{\mu(z)(1-|z|^{2})}}$ for any $f\in H^{\infty}_{\nu}$ and $T_{g}^{\phi}(f)$ vanishes at origin for any $f\in H(\mathbb{D})$. Now polynomials are dense in $\mathcal{H}^{0}_{\mu}$, so it is sufficient to consider polynomials $f_{1},f_{2},\cdots,f_{N}$ in the definition of absolutely summing. Now for any $C>1,$ we can select $z_{i}$, $i=1,2,\cdots N$ such that $\displaystyle\sum_{i=1}^{N}\|T_{g}^{\phi}f_{i}\|_{\mathcal{H}^{\infty}_{\mu}}$ $\displaystyle\asymp\displaystyle\sum_{i=1}^{N}\left(\big{|}(T_{g}^{\phi}f_{i})(0)\big{|}+\sup_{z\in\mathbb{D}}(1-|z|^{2})\mu(z)\big{|}(T_{g}^{\phi}f_{i})^{\prime}(z)\big{|}\right)$ $\displaystyle=\sum_{i=1}^{N}\displaystyle\sup_{z\in\mathbb{D}}(1-|z|^{2})\mu(z)\big{|}f_{i}(\phi(z))g^{\prime}(z)\big{|}$ $\displaystyle\leq C\displaystyle\sum_{i=1}^{N}(1-|z_{i}|^{2})\mu(z_{i})\big{|}f_{i}(\phi(z_{i}))g^{\prime}(z_{i})\big{|}$ (2.4) Using the duality pairing in (1.2), we have that $\displaystyle\big{|}f_{i}(\phi(z_{i}))\big{|}=\big{|}(f_{i},K_{\phi(z_{i})})\big{|}=\bigg{|}(\alpha+1)\int_{\mathbb{D}}f_{i}(\zeta)\frac{(1-|\zeta|^{2})^{\alpha}}{|1-\bar{\zeta}\phi(z_{i})|^{\alpha+2}}dA(\zeta)\bigg{|}.$ Thus from (2.4), we have that $\displaystyle\sum_{i=1}^{N}\|T_{g}^{\phi}f_{i}\|_{{}_{\mathcal{H}^{\infty}_{\mu}}}$ $\displaystyle\leq C\bigg{(}\sup_{w\in\mathbb{D}}\displaystyle\sum_{i=1}^{N}\big{|}f_{i}(w)\big{|}\nu(w)\bigg{)}$ $\displaystyle\quad\times\int_{\mathbb{D}}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)\big{|}g^{\prime}(z)\big{|}}{\big{|}1-\bar{\zeta}\phi(z)\big{|}^{\alpha+2}}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)$ $\displaystyle=C\bigg{(}\sup_{w\in\mathbb{D}}\displaystyle\sum_{i=1}^{N}\big{|}f_{i}(w)\big{|}\nu(w)\bigg{)}M_{\alpha}$ (2.5) Using the equality $\sup_{w\in\mathbb{D}}\sum_{i=1}^{N}|f_{i}(w)|\nu(w)=\sup_{|\eta_{i}|=1}\sup_{w\in\mathbb{D}}\bigg{|}\displaystyle\sum_{i=1}^{N}\eta_{i}f_{i}(w)\bigg{|}\nu(w)=\sup_{|\eta_{i}|=1}\bigg{\|}\displaystyle\sum_{i=1}^{N}\eta_{i}f_{i}\bigg{\|}_{\mathcal{H}^{\infty}_{\mu}}$ in (2.5), we have that $\displaystyle\sum_{i=1}^{N}\|T_{g}^{\phi}f_{i}\|_{{}_{\mathcal{H}^{\infty}_{\mu}}}$ $\displaystyle\leq C\sup_{|\eta_{i}|=1}\bigg{\|}\displaystyle\sum_{i=1}^{N}\eta_{i}f_{i}\bigg{\|}_{\mathcal{H}^{\infty}_{\mu}}.$ Since $C>0$ is arbitrary, so $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is absolutely summing. $(a)\Rightarrow(c)$ Since every nuclear operator is absolutely summing, therefore, $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is absolutely summing. By Theorem 4, there is a probability Borel measure $\varrho$ on $\sigma(A^{1}_{\omega},\mathcal{H}^{0}_{\mu})$-compact unit ball $\mathbb{B}_{1}$ of $A^{1}_{\omega}$, where $\sigma(A^{1}_{\omega},\mathcal{H}^{0}_{\mu})$ is the weak∗-topology on $(\mathcal{H}^{0}_{\mu})^{\prime}=A^{1}_{\omega},$ and some $\xi$ in $A^{1}_{\omega}$ such that $\|T_{g}^{\phi}f\|_{\mathcal{H}_{\mu}^{\infty}}\leq C\int_{\mathbb{B}_{1}}|\xi(f)|d\varrho(\xi)$ (2.6) for every $f\in\mathcal{H}^{0}_{\mu}$ and some constant $C>0$ independent of $f$. For every $\zeta\in\mathbb{D}$, we have that $K_{{\zeta}}\in\mathcal{H}^{0}_{\mu}$. Therefore, $\displaystyle(\alpha+1)\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g^{\prime}(z)|}{|1-{\zeta}\phi(z)|^{\alpha+2}}\leq\|T_{g}^{\phi}K_{{\zeta}}\|_{\mathcal{B}_{\mu(z)(1-|z|^{2})}}\leq C\|T_{g}^{\phi}K_{{\zeta}}\|_{\mathcal{H}^{\infty}_{\mu(z)}}.$ (2.7) Replacing $f$ in (2.6) by $K_{{\zeta}}$, and then using (2.7), we have that $\displaystyle\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g^{\prime}(z)|}{|1-{\zeta}\phi(z)|^{\alpha+2}}\leq\frac{C}{\alpha+1}\int_{\mathbb{B}_{1}}|\xi(K_{w})|d\varrho(\xi).$ Integrating over $\mathbb{D}$ with respect to $\omega dA$, using (1.1) and then applying Fubini’s theorem, we get $\displaystyle\int_{\mathbb{D}}\sup_{z\in\mathbb{D}}$ $\displaystyle\frac{(1-|z|^{2})\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+2}}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)$ $\displaystyle\quad\quad\leq\frac{C}{\alpha+1}\int_{\mathbb{B}_{1}}\int_{\mathbb{D}}|\xi(K_{\zeta})|\omega(z)dA(\zeta)d\varrho(\xi).$ Now $K_{\zeta}\in\mathcal{H}^{0}_{\mu}$ and $\xi$ in $A^{1}_{\omega}$, so $\xi(K_{\zeta})=\langle K_{\zeta},\xi\rangle=\xi(w).$ Thus using the fact that $\varrho$ is a probability Borel measure, we get $\displaystyle\int_{\mathbb{D}}\sup_{z\in\mathbb{D}}$ $\displaystyle\frac{(1-|z|^{2})\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+2}}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)$ $\displaystyle\quad\quad\leq\frac{C}{\alpha+1}\sup_{\xi\in\mathbb{B}_{1}}\int_{\mathbb{D}}|\xi(\zeta)|\omega(\zeta)dA(\zeta)$ $\displaystyle\quad\quad=\frac{C}{\alpha+1}\sup_{\xi\in\mathbb{B}_{1}}\|\xi\|_{A^{1}_{\omega}}=\frac{C}{\alpha+1}.$ This completes the proof of the implication $(a)\Rightarrow(c)$. $(a)\Rightarrow(b)$ Since transpose of a nuclear operator is nuclear, see 3.1.8 in [15], so by an application of Theorem 2(1), it follows that $(a)\Rightarrow(b)$. $(b)\Rightarrow(a)$ Recall that class of nuclear operators forms an operator ideal, that is, nuclear operators are a class of operators which do not fixes a copy of $l^{\infty},$ (see [7] and [8]), and $\mathcal{H}^{0}_{\nu}$ is a closed subspace of $\mathcal{H}^{\infty}_{\mu}$, so it is obvious that $(b)\Rightarrow(a)$. This completes the proof. ∎ ###### Theorem 6. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in\mathcal{B}^{0}_{(1-|z|^{2})\mu(z)}$ and $\phi$ a self-map of $\mathbb{D}$ such that $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is nuclear. 2. (b) $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is nuclear. 3. (c) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}$ is nuclear. 4. (d) $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}$ is nuclear. 5. (e) For $g$, $\phi$ and $\nu$ satisfy the following condition: $M_{\alpha}=\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+2}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Proof. By Theorem 4, we have that $(a)\Leftrightarrow(b)\Leftrightarrow(e).$ To complete the proof it is sufficient to prove that $(a)\Rightarrow(c)\Rightarrow(d)\Rightarrow(a).$ $(a)\Rightarrow(c)$ Suppose that $(a)$ holds. Then by Theorem 2(2) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}$ is bounded. Moreover, $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is absolutely summing if and only if $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}.$ Thus the implication $(a)\Rightarrow(d)$ follows. $(c)\Rightarrow(d)$ Using Theorem 2(ii) and the fact that transpose of a nuclear operator is nuclear the implication follows. $(d)\Rightarrow(a).$ The proof follows as the proof of implication $(b)\Rightarrow(a)$ in Theorem 4. This completes the proof.∎ ###### Corollary 1. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in H(\mathbb{D})$ and $\phi$ a self-map of $\mathbb{D}$ such that $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow B_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) For $g$, $\phi$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+2}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Corollary 2. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in\mathcal{B}^{0}_{\mu}$ and $\phi$ a self-map of $\mathbb{D}$ such that $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) $T_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 4. (d) $T_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 5. (e) For $g$, $\phi$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+2}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ We can easily obtain the following corollaries. ###### Corollary 3. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $\phi$ is a self-map of $\mathbb{D}$ such that $C_{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $C_{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $C_{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) $\phi$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|\phi^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+2}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Corollary 4. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $\phi\in\mathcal{B}^{0}_{\mu(z)}$ and $C_{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $C_{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $C_{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) $C_{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 4. (d) $C_{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 5. (e) $\phi$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|\phi^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+2}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Corollary 5. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in{H}(\mathbb{D})$ be such that $T_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow B_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $T_{g}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $T_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) For $g$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}z|^{\alpha+2}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Corollary 6. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in\mathcal{B}^{0}_{\mu(z)}$ be such that $T_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $T_{g}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $T_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) $T_{g}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 4. (d) $T_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 5. (e) For $g$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}z|^{\alpha+2}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Theorem 7. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in H(\mathbb{D})$ and $\phi$ a self-map of $\mathbb{D}$ such that $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}$ is nuclear. 2. (b) $S_{g}^{\phi}:\mathcal{H}_{\nu}\longrightarrow\mathcal{H}_{\mu}$ is nuclear. 3. (c) $g$, $\nu,$ $\mu$ and $\phi$ satisfy the following condition: $N_{\alpha}=\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Proof. That $(a)\Leftrightarrow(b)$ can be proved, proceeding as in the proof of Theorem 5. Thus to complete the proof, we need to prove that $(a)\Leftrightarrow(c)$. $(c)\Rightarrow(a)$ Proceeding as in the proof of $(c)\Rightarrow(a)$ of Theorem 5, for any $C>1,$ we can select $z_{i}$, $i=1,2,\cdots N$ such that $\displaystyle\sum_{i=1}^{N}\|T_{g}^{\phi}f_{i}\|_{\mathcal{H}_{\mu}^{\infty}}\leq C\displaystyle\sum_{i=1}^{N}(1-|z_{i}|^{2})\mu(z_{i})\big{|}f_{i}(\phi(z_{i}))g(z_{i})\big{|}$ (2.8) and $\displaystyle f_{i}(w)=\langle f_{i},K_{w}\rangle=(\alpha+1)\int_{\mathbb{D}}f_{i}(\zeta)\frac{(1-|\zeta|^{2})^{\alpha}}{(1-\bar{\zeta}w)^{\alpha+2}}dA(\zeta).$ Differentiating with respect to $w,$ we have that $\displaystyle f^{\prime}_{i}(w)=(\alpha+1)(\alpha+2)\int_{\mathbb{D}}f_{i}(\zeta)\frac{\bar{\zeta}(1-|\zeta|^{2})^{\alpha}}{(1-\bar{\zeta}w)^{\alpha+3}}dA(\zeta).$ Thus $\displaystyle\big{|}f^{\prime}_{i}(\phi(z_{i}))\big{|}\leq(\alpha+1)(\alpha+2)\int_{\mathbb{D}}|f_{i}(\zeta)|\frac{\bar{\zeta}(1-|\zeta|^{2})^{\alpha}}{|1-\bar{\zeta}\phi(z_{i})|^{\alpha+3}}dA(\zeta).$ (2.9) Using (2.8) and (2.9 ), we have that $\displaystyle\sum_{i=1}^{N}\|T_{g}^{\phi}f_{i}\|_{{}_{\mathcal{H}_{\mu}^{\infty}}}$ $\displaystyle\leq C(\alpha+1)(\alpha+2)\bigg{(}\sup_{w\in\mathbb{D}}\displaystyle\sum_{i=1}^{N}\big{|}f_{i}(w)\big{|}\nu(w)\bigg{)}$ $\displaystyle\quad\times\int_{\mathbb{D}}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)\big{|}g^{\prime}(z)\big{|}}{\big{|}1-\bar{\zeta}\phi(z)\big{|}^{\alpha+3}}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)$ $\displaystyle=C(\alpha+1)(\alpha+2)\bigg{(}\sup_{w\in\mathbb{D}}\displaystyle\sum_{i=1}^{N}\big{|}f_{i}(w)\big{|}\nu(w)\bigg{)}N_{\alpha}$ (2.10) The proof can now be completed proceeding as in the Theorem 5. We omit the details. $(a)\Rightarrow(c)$ Once again, proceeding as in the proof of Theorem 5, there exists a probability Borel measure $\varrho$ on $\sigma(A^{1}_{\omega},\mathcal{H}^{0}_{\mu})$-compact unit ball $\mathbb{B}_{1}$ of $A^{1}_{\omega}$ and $\xi$ in $A^{1}_{\omega}$ such that $\|S_{g}^{\phi}f\|_{\mathcal{H}_{\mu}^{\infty}}\leq C\int_{\mathbb{B}_{1}}|\xi(f)|d\varrho(\xi)$ (2.11) for every $f\in\mathcal{H}^{0}_{\mu}$ and some constant $C>0$ independent of $f$. For every $\zeta\in\mathbb{D}$, we have that $K_{\zeta}\in\mathcal{H}^{0}_{\mu}$. $\displaystyle(\alpha+1)(\alpha+2)\sup_{z\in\mathbb{D}}\frac{|\zeta|(1-|z|^{2})\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\leq\|S_{g}^{\phi}K_{\zeta}\|_{\mathcal{B}_{\mu(z)(1-|z|^{2})}}\leq C\|S_{g}^{\phi}K_{\zeta}\|_{\mathcal{H}_{\mu(z)}^{\infty}}.$ (2.12) Replacing $f$ in (2.11) by $K_{\zeta}$, and then using (2.12), we have that $\displaystyle\sup_{z\in\mathbb{D}}\frac{|\zeta|(1-|z|^{2})\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\leq\frac{C}{(\alpha+1)(\alpha+2)}\int_{\mathbb{B}_{1}}|\xi(K_{\zeta})|d\varrho(\xi).$ (2.13) Integrating (2.13) over $\mathbb{D}$ with respect to $\omega dA$ and proceeding as in the Theorem 5, we get $\displaystyle\int_{\mathbb{D}}\sup_{z\in\mathbb{D}}\frac{|\zeta|(1-|z|^{2})\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)\leq\frac{C}{(\alpha+1)(\alpha+2)}.$ (2.14) Now if $|\zeta|>1/2,$ then by (2.14), we have that $\displaystyle\int_{|\zeta|>1/2}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)\leq\frac{2C}{(\alpha+1)(\alpha+2)}.$ (2.15) Further, if $|\zeta|\leq 1/2,$ then we have that $\displaystyle\int_{|\zeta|\leq 1/2}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)\leq\frac{2^{\alpha+1}\|g\|_{\mathcal{H}_{\mu}^{\infty}}}{\nu(1/2)}.$ (2.16) Thus $\displaystyle N_{\alpha}$ $\displaystyle\leq\int_{|\zeta|\leq 1/2}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)$ $\displaystyle\quad+\int_{|\zeta|>1/2}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)$ $\displaystyle\leq\frac{2^{\alpha+1}\|g\|_{\mathcal{H}_{\mu}^{\infty}}}{\nu(1/2)}+\frac{2C}{(\alpha+1)(\alpha+2)}.$ This completes the proof. ∎ ###### Corollary 7. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in\mathcal{H}^{0}_{(1-|z|^{2})\mu(z)}$ and $\phi$ an analytic self-map of $\mathbb{D}$ such that $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is nuclear. 2. (b) $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{\infty}_{\mu}$ is nuclear. 3. (c) $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}$ is nuclear. 4. (d) $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{H}^{0}_{\mu}$ is nuclear. 5. (e) For $g$, $\phi$ and $\nu$ satisfy the following condition: $N_{\alpha}=\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Corollary 8. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in{H}(\mathbb{D})and$ $\phi$ an analytic self-map of $\mathbb{D}$ such that $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow B_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) For $g$, $\phi$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Corollary 9. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in\mathcal{B}^{0}_{\mu(z)}$ and $\phi$ an analytic self-map of $\mathbb{D}$ such that $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) $S_{g}^{\phi}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 4. (d) $S_{g}^{\phi}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 5. (e) For $g$, $\phi$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ We can also obtain the following corollaries. ###### Corollary 10. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $\phi$ is an analytic self-map of $\mathbb{D}$ such that $C_{\phi}D:\mathcal{H}^{\infty}_{\nu}\longrightarrow B_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $C_{\phi}D:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $C_{\phi}D:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) $\phi$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|\phi^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Corollary 11. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $\phi$ an analytic self-map of $\mathbb{D}$ such that $\phi\in\mathcal{H}^{0}_{\mu(z)}$ and $C_{\phi}D:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $C_{\phi}D:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $C_{\phi}D:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) $C_{\phi}D:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 4. (d) $C_{\phi}D:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 5. (e) $\phi$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|\phi^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Corollary 12. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If Let $g\in{H}(\mathbb{D})$ be such that $S_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow B_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $S_{g}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $S_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) For $g$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}z|^{\alpha+3}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ###### Corollary 13. Let $\nu,\mu\in N_{W}(\mathbb{D}),$ $\alpha>-1$ and $\omega$ a weight function such that $\\{\nu,\omega\\}$ is a normal pair and (1.1) holds. If $g\in\mathcal{B}^{0}_{\mu(z)}$ be such that $S_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is bounded, then the following statements are equivalent. 1. (a) $S_{g}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 2. (b) $S_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}_{\mu}$ is nuclear. 3. (c) $S_{g}:\mathcal{H}^{0}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 4. (d) $S_{g}:\mathcal{H}^{\infty}_{\nu}\longrightarrow\mathcal{B}^{0}_{\mu}$ is nuclear. 5. (e) $g$ and $\nu$ satisfy the following condition: $\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{\mu(z)|g^{\prime}(z)|}{|1-\bar{\zeta}z|^{\alpha+3}}\bigg{]}\frac{(1-|\zeta|^{2})^{\alpha}}{\nu(\zeta)}dA(\zeta)<\infty.$ ## 3\. Nuclear $T_{g}^{\phi}$ and $S_{g}^{\phi}$ between Bloch spaces of order $\beta$ Finally, we apply our results to $T_{g}^{\phi}$ and $S_{g}^{\phi}$ acting between weighted Banach spaces and Bloch spaces of order $\beta,$ where $\beta>0.$ Recall that if $\nu_{\beta}$ is a classical weight, that is $\nu_{\beta}(z)=(1-|z|^{2})^{\beta}$, then the spaces $\mathcal{H}^{\infty}_{\nu}$, $\mathcal{H}^{0}_{\nu}$, $\mathcal{B}_{\nu}$ and $\mathcal{B}^{0}_{\mu}$ reduces respectively to weighted and Bloch spaces of order $\beta$, denoted by $\mathcal{H}^{\infty}_{\beta}$, $\mathcal{H}^{0}_{\beta}$, $\mathcal{B}_{\beta}$ and $\mathcal{B}^{0}_{\beta}$. Moreover, for $\beta>1$, we have that $\mathcal{H}^{0}_{\beta-1}=\mathcal{B}^{0}_{\beta}$ and $\mathcal{H}^{\infty}_{\beta-1}=\mathcal{B}^{\infty}_{\beta}$ with equivalent norm. Thus we have the following corollary. ###### Corollary 14. Let $\beta>1$, $\gamma>0$, $g\in{H}(\mathbb{D})$ and $\phi$ an analytic self- map of $\mathbb{D}$ such that $T_{g}^{\phi}:\mathcal{B}_{\beta}\longrightarrow\mathcal{B}_{\gamma}$ is bounded. Then the following statements are equivalent. 1. (a) $T_{g}^{\phi}:\mathcal{B}^{0}_{\beta}\longrightarrow\mathcal{B}_{\gamma}$ is nuclear. 2. (b) $T_{g}^{\phi}:\mathcal{B}_{\beta}\longrightarrow\mathcal{B}_{\gamma}$ is nuclear. 3. (c) $g$, $\beta,$ $\gamma$ and $\phi$ satisfy the following condition: $P_{\alpha}=\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})^{\gamma}|g^{\prime}(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+2}}\bigg{]}{(1-|\zeta|^{2})^{\alpha-\beta+1}}dA(\zeta)<\infty.$ The case $\beta=1$ is settled in the next theorem. ###### Theorem 8. Let $g\in{H}(\mathbb{D})$ and $\phi$ an analytic self-map of $\mathbb{D}$ such that $T_{g}^{\phi}:\mathcal{B}\longrightarrow\mathcal{B}$ is bounded. Then the following statements are equivalent. 1. (a) $T_{g}^{\phi}:\mathcal{B}_{0}\longrightarrow\mathcal{B}$ is nuclear. 2. (b) $g$ and $\phi$ satisfy the following condition: $M=\int_{\mathbb{D}}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})|\phi^{\prime}(z)|}{|1-\bar{w}\phi(z)|^{3}}|g(z)|dA(w)<\infty.$ ###### Proof. $(b)\Rightarrow(a)$ Since $c_{0}\simeq\mathcal{B}_{0}$ and polynomials are dense in $\mathcal{B}_{0}$. Moreover, by Corollary $1.5$ in [18], we have that $f\in\mathcal{B},$ then we have that $f^{\prime}(z)=2\int_{\mathbb{D}}\frac{1-|w|^{2}}{(1-z\bar{w})^{3}}f^{\prime}(w)dA(w).$ (3.1) Consider polynomials $p_{1},p_{2},\cdots,p_{N}$ in the definition of absolutely summing and then using (3.1), we have $\sum_{i=1}^{N}\|T_{g}^{\phi}p_{i}\|_{\mathcal{B}}\leq C\bigg{(}\sup_{w\in\mathbb{D}}\displaystyle\sum_{i=1}^{N}\big{|}p^{\prime}_{i}(w)\big{|}(1-|w|^{2})\bigg{)}\int_{\mathbb{D}}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})\big{|}\phi^{\prime}(z)\big{|}}{\big{|}1-\bar{w}\phi(z)\big{|}^{3}}\big{|}g(z)\big{|}dA(w).$ The proof can be settled proceeding as in the proof of Theorem 5. $(a)\Rightarrow(b)$ By Theorem 1, there exists a probability Borel measure $\varrho$ on $\sigma((\mathcal{B}_{0})^{*},\mathcal{B})$-compact unit ball $B^{\prime}$ of $(\mathcal{B}_{0})^{*}=A^{1}$ such that $\|T_{g}^{\phi}f\|_{\mathcal{B}}\leq C\int_{B^{{}^{\prime}}}|\xi(f)|d\varrho(\xi)$ for every $f\in\mathcal{B}_{0}$ and some constant $C>0$ independent of $f$. Next for every $w\in\mathbb{D}$, we consider $f_{w}(z)={1}/{(1-\bar{w}z)^{2}}$ which lies in $\mathcal{B}_{0}\cap H^{\infty}$. By duality and reproducing property of $f_{w}$ in the Bergman space $(\mathcal{B}_{0})^{*}=A^{1}$, we have $\xi(f_{w})=\langle f_{w},h\rangle=\int_{\mathbb{D}}h(z)\overline{f_{w}(z)}dA(z)=\int_{\mathbb{D}}\dfrac{h(z)}{(1-\bar{z}w)^{2}}dA(z)=h(w).$ Also, $\sup_{z\in\mathbb{D}}\dfrac{2|w|(1-|z|^{2})|\phi^{\prime}(z)||g(z)|}{|1-\bar{w}\phi(z)|^{3}}=\|T_{g}^{\phi}f_{w}\|_{1}\leq\|T_{g}^{\phi}f_{w}\|_{\mathcal{B}}$ Now proceeding as in the proof of Theorem 7, we can easily see that $(b)$ holds. This completes the proof.∎ Similarly, we have the following corollary. ###### Corollary 15. Let $\beta>1$, $\gamma>0$, $g\in{H}(\mathbb{D})$ and $\phi$ an analytic self- map of $\mathbb{D}$ such that $S_{g}^{\phi}:\mathcal{B}_{\beta}\longrightarrow\mathcal{B}_{\gamma}$ is bounded. Then the following statements are equivalent. 1. (a) $S_{g}^{\phi}:\mathcal{B}^{0}_{\beta}\longrightarrow\mathcal{B}_{\gamma}$ is nuclear. 2. (b) $S_{g}^{\phi}:\mathcal{B}_{\beta}\longrightarrow\mathcal{B}_{\gamma}$ is nuclear. 3. (c) For $g$, $\phi$ and $\nu$ satisfy the following condition: $Q_{\alpha}=\int_{\mathbb{D}}\bigg{[}\sup_{z\in\mathbb{D}}\frac{(1-|z|^{2})^{\gamma}|g(z)|}{|1-\bar{\zeta}\phi(z)|^{\alpha+3}}\bigg{]}{(1-|\zeta|^{2})^{\alpha-\beta+1}}dA(\zeta)<\infty.$ ## References * [1] Aleman, A., Cima, J.: An integral operator on $H^{p}$ and Hardy’s inequality. J. Anal. Math. 85, 157–176 (2001) * [2] Aleman, A., Siskakis, A.: An integral operator on $H^{p}$. Complex Var. Theory Appl. 28(2), 149–158 (1995) * [3] Aleman, A., Siskakis, A.: Integration operators on Bergman spaces. Indiana Univ. Math. J. 46(2), 337–356 (1997) * [4] Anderson, A., Jovovic, M., Smith, W.: Some integral operators acting on $H^{\infty}$. Integral Equ. Oper. Theory 80(2), 275–291 (2014) * [5] Anderson, A.: Some closed range integral operators on spaces of analytic functions. Integral Equ. Oper. Theory 69, 87–99 (2011) * [6] Basallote, M., Contreras, M., Hernández-Mancera, C., Martín, M., Paúl, P.: Volterra operators and semigroups in weighted Banach spaces of analytic functions. Collect. Math. 65(2), 233–249 (2014) * [7] Bonet, J., Domański, P., Lindström, M.: Essential norm and weak compactness of composition operators on weighted Banach spaces of analytic functions. Canad. Math. Bull. 42, 139– 148 (1999) * [8] Bonet, J., Gómez-collado, M. C., Jordá, E., Jornet, D.: Nuclear weighted composition operators on weighted Banach spaces of analytic functions. Proc. Amer. Math. Soc., 149, 311–321 (2021) * [9] Contreras, M., Hernández-Mancera, C.: Weighted composition operators in weighted Banach spaces of analytic functions. J. Austral. Math. Soc. 69, 41–60 (2000) * [10] Contreras, M., Hernández-Mancera, C., Martín, M., Paúl, P.: Volterra operators and semigroups in weighted Banach spaces of analytic functions. Collect. Math., 65, 233–249 (2014) * [11] Contreras, M., Peláez, J., Pommerenke, C., Rättyä, J.: Integral operators mapping into the space of bounded analytic functions. J. Funct. Anal. 271(10), 2899–2943 (2016) * [12] Conway, J. B.: A course in functional analysis, second ed., Graduate Texts in Mathematics. vol. 96, Springer-Verlag, New York (1990) * [13] Diestel, J., Jarchow, H., Tonge, A.: Absolutely summing operators. Cambridge Studies in Advanced Mathematics. 43, Cambridge University Press. Cambridge (1995) * [14] Dunford, N., Schwartz, J.T.: Linear Operators I. Interscience Publishers. Wiley, New York (1958) * [15] Domański, P., Lindström, M.: Sets of interpolation and sampling for weighted Banach spaces of holomorphic functions. Ann. Polon. Math. 89, 233–264 (2002) * [16] Eklund, T., Lindström, M., Pirasteh, M., Sanatpour, A., Wikman, N.: Generalized Volterra operators mapping between Banach spaces of analytic functions. Monatsh. Math. (2018). https://doi.org/10.1007/s00605-018-1216-5 * [17] Fares, T., Lefévre, P.: Nuclear composition operators on Bloch spaces. Proc. Amer. Math. Soc.. 148, 2487-2496 (2020) * [18] Hedenmalm, H., Korenblum, B., Zhu, K.: Theory of Bergman Spaces. Springer, New York Berlin Heidelberg (2000) * [19] Lefévre, P., Rodríguez-Piazza, L.: Absolutely summing Carleson embeddings on Hardy spaces. Adv. Math. 340, 528–587, (2018) * [20] Li, S., Stević, S.: Products of Volterra type operator and composition operator from $H^{\infty}$ and Bloch spaces to Zygmund spaces. J. Math. Anal. Appl. 345, 40–52 (2008) * [21] Lin, Q.:, Volterra type operators between Bloch type spaces and weighted Banach spaces. Integral equations and operator theory. 91 (2019) Article No. 13 * [22] Lin, Q., Liu, J., Wu, Y.: Volterra type operators on $S^{p}(\mathbb{D})$ spaces. J. Math. Anal. Appl. 461, 1100–1114 (2018) * [23] Lusky, W.: On the structure of $Hv_{0}(D)$ and $hv_{0}(D)$. Math. Nachr. 159, 279–289 (1992) * [24] Lusky, W.: On weighted spaces of harmonic and holomorphic functions. J. London Math. Soc. 51, 309–320 (1995) * [25] Lusky, W.: Growth Conditions for Harmonic and Holomorphic Functions, Functional Analysis. (Trier, 1994), de Gruyter, Berlin 281–291 (1996) * [26] Lusky, W.: On the isomorphism classes of weighted spaces of harmonic and holomorphic functions. Studia Math. 175, 19–45 (2006) * [27] Manhas, J., Zhao, R.: Products of weighted composition operators and differentiation operators between Banach spaces of analytic functions. Acta Sci. Math. (Szeged) 80, 665–679 (2014) * [28] Pommerenke, Ch.: Schlichte funktionen und analytische funktionen von beschränkter mittlerer oszilation. Comment. Math. Helv. 52, 591–502 (1977) * [29] Shields, A. L., Williams, D. L.: Bounded projections, duality and multipliers in spaces of analytic functions. Trans. Amer. Math. Soc. 162, 287–302 (1971) * [30] Shields, A. L., Williams, D. L.: Bounded projections, duality, and multipliers in spaces of harmonic functions, J. Reine Angew. Math. 299/300, 256–279 (1978) * [31] Yoneda, R.: Integration operators on weighted Bloch spaces. Nihonkai Math. J. 12, 123–133 (2001)
# Split-PU: Hardness-aware Training Strategy for Positive-Unlabeled Learning Chengming Xu School of Data Science, Fudan University China <EMAIL_ADDRESS>, Chen Liu Department of Mathematics Hong Kong University of Science and Technology China<EMAIL_ADDRESS>, Siqian Yang , Yabiao Wang Tencent Youtu Lab China<EMAIL_ADDRESS><EMAIL_ADDRESS>, Shijie Zhang School of Data Science, Fudan University China<EMAIL_ADDRESS>, Lijie Jia Shanghai Jiaotong University China<EMAIL_ADDRESS>and Yanwei Fu School of Data Science, Fudan University China<EMAIL_ADDRESS> (2022) ###### Abstract. Positive-Unlabeled (PU) learning aims to learn a model with rare positive samples and abundant unlabeled samples. Compared with classical binary classification, the task of PU learning is much more challenging due to the existence of many incompletely-annotated data instances. Since only part of the most confident positive samples are available and evidence is not enough to categorize the rest samples, many of these unlabeled data may also be the positive samples. Research on this topic is particularly useful and essential to many real-world tasks which demand very expensive labelling cost. For example, the recognition tasks in disease diagnosis, recommendation system and satellite image recognition may only have few positive samples that can be annotated by the experts. While this problem is receiving increasing attention, most of the efforts have been dedicated to the design of trustworthy risk estimators such as uPU (Du Plessis et al., 2014) and nnPU (Kiryo et al., 2017) and direct knowledge distillation, e.g., Self-PU (Chen et al., 2020a). These methods mainly omit the intrinsic hardness of some unlabeled data, which can result in sub-optimal performance as a consequence of fitting the easy noisy data and not sufficiently utilizing the hard data. In this paper, we focus on improving the commonly-used nnPU (Kiryo et al., 2017) with a novel training pipeline. We highlight the intrinsic difference of hardness of samples in the dataset and the proper learning strategies for easy and hard data. By considering this fact, we propose first splitting the unlabeled dataset with an early-stop strategy. The samples that have inconsistent predictions between the temporary and base model are considered as hard samples. Then the model utilizes a noise-tolerant Jensen-Shannon divergence loss for easy data; and a dual-source consistency regularization for hard data which includes a cross-consistency between student and base model for low-level features and self-consistency for high-level features and predictions, respectively. Our method achieves much better results compared with existing methods on CIFAR10 and two medical datasets of liver cancer survival time prediction, and low blood pressure diagnosis of pregnant, individually. The experimental results validates the efficacy of our proposed method. Codes and models are available at https://github.com/loadder/SplitPU_MM2022. Positive-Unlabeled learning; Consistency Regularization ††journalyear: 2022††copyright: acmcopyright††conference: Proceedings of the 30th ACM International Conference on Multimedia; October 10–14, 2022; Lisboa, Portugal††booktitle: Proceedings of the 30th ACM International Conference on Multimedia (MM ’22), October 10–14, 2022, Lisboa, Portugal††price: 15.00††doi: 10.1145/3503161.3548045††isbn: 978-1-4503-9203-7/22/10††ccs: Computing methodologies††ccs: Computing methodologies Artificial intelligence††ccs: Computing methodologies Computer vision††ccs: Computing methodologies Computer vision problems ## 1\. Introduction The deep models of classical recognition task, especially binary classification, has been supported by plenty of positive and negative samples, which are comprehensively annotated in general. On the other hand, the collection of training data could not always be perfect. In many Multimedia applications, only part of the positive samples can be annotated, while the other samples including both positive and negative ones can only be left without labels. These cases may happen quite often. For example, the annotator source is severely limited for tasks, such as military targets recognition from satellite images (Fan et al., 2020), since expertise and confidential knowledge is required. In medical image analysis such as liver cancer diagnosis, the annotators can only provide confident judgement to few positive samples, while the other images without obvious symptom could also be positive but cannot be explicitly annotated. Essentially, the Positive-Unlabeled (PU) learning is studied to handle these tasks. With a well-trained PU learning model, many real-world applications of different modalities with only positive and unlabeled data can thus be effectively solved, e.g. medical diagnosis, recommendation system, satellite image recognition, etc. Figure 1. The binary classification versus PU learning. Left: The most commonly-seen visual data with distinguished categories. The data can be easily identified as an animal or not animal. Right: Data for low blood pressure diagnosis. Without sufficient expertise annotators, only part of the most confident positive samples can be recognized. Therefore the data in such an area is often taken as PU data. Due to the special structure of training set, PU learning is generally more challenging than standard supervised binary classification. Directly applying existing methods in both fully-supervised and semi-supervised learning would lead to great performance degradation. The classical PU learning methods can be traced back to confident negative sample selection(Liu et al., 2002, 2003) in the early stages, and focus on designing unbiased risk estimator (Du Plessis et al., 2014; Kiryo et al., 2017) recently. While some previous works like Self-PU (Chen et al., 2020a) and PUUPL (Dorigatti et al., 2022) have explored combining together the well-designed objective function and sample selection for better performance, these methods neglect an important fact in the PU data: Some samples are relatively much harder to learn by deep neural networks than most data. Such hardness is not similar to the definition of ’easiness’ depicted in Self- PU which is measured by the confidence of predictions. It is the intrinsic property of image data that we focus on, which could be resulted from many different reasons. For example, for commonly-used image datasets, those images with extremely small, distorted or camouflaged objects can be relatively harder to learn than other normal one. Similarly, in medical image analysis, some samples may have severe artifacts, affecting training and inference of models. These artifacts may be caused by the errors of medical instruments, which can hardly be avoided in all cases. By considering this fact, we advocate that the unlabeled data should be further divided into easy and hard set. The easy data dominates the unlabeled set, being easy to fit but vulnerable to noise, while the hard data is on the opposite. Specifically, inspired by Self-PU, we can annotate these unlabeled samples with pseudo-labels generated by a base model, e.g. a nnPU model. Then both sets contain wrongly-labeled (noisy) data and correctly-labeled (clean) data. We highlight two important points which imply that these different types of data serve as different roles in training. (1) Easy data is not totally reliable. Especially, since the easy noisy samples can be fitted by the model facilely, it is more likely for these samples than the hard noisy ones to hurt the generalization ability of the learned PU models. Thus potentially, it may be questionable to directly utilize all pseudo-labels from the easy set in a vanilla objective functions like cross entropy losses. (2) Hard data is more difficult to deal with and requires objective functions that exclude the pseudo-labels. As the name of hard data suggests, the model can hardly learn information from the pseudo-labels of hard data no matter whether they are clean or not. Nevertheless, the hard clean samples are essentially valuable for estimating the decision boundary. For example, as a common rule of the thumb, the hard and ambiguous training instances may also be taken as support vectors to compute the classifier margins in support vector machine. Therefore, it is a problem to make use of the hard data in a more effective way. Given these two points, the naïve selection-distillation pipeline as in Self-PU, i.e. gradually selecting confident samples and training with their pseudo-labels, may lead to sub-optimal performance, as all unlabeled samples are treated with the same form of objective functions. To this end, we propose a novel Split-PU method that employs a novel hardness- aware training strategy for positive-unlabeled learning. Our Split-PU method takes apart the easy and hard data from all PU data to deal with, respectively. Specifically, the whole method contains four steps. (1) We first train a base teacher model of nnPU (Kiryo et al., 2017) as objective function. The base model is applied to generate the initial pseudo-labels for all unlabeled data. (2) Then a temporary student model is trained via knowledge distillation from the teacher model, with recording training accuracy of each epoch. When the temporary model reaches the specified performance, we split out those training samples as hard data that have the inconsistent predictions between the temporary model and the base one, with the rest as easy data. (3) Further, different learning strategies are employed to handle the easy and hard data. For easy data, we find out that those easy noisy samples cannot be detected as effectively as the hard noisy samples by using existing noise detection methods. As an alternative, we thus for the first time introduce the Jensen-Shannon divergence loss (Englesson and Azizpour, 2021) to optimize the PU learning model for noisy-robust PU learning. For hard data, we develop a dual-source consistency regularization which covers low-level and high-level features together with predictions, thus fully utilizing the supervision from both the base and the student model. We have to regularize such consistency here: a cross-consistency between the low-level features of base model and student model, which is aimed to guide the student model with well-trained low-level features like edges and colors from base model; and a self- consistency on the high-level features and predictions of student model between different views of each image, which is used to improve the stability of training for better visual representations. With the above training strategy, we can better handle the unlabeled data with the pseudo-labels predicted by the base model, leading to a student model with better performance. (4) Finally, we replace the base model with the student model to iterate the training algorithm for further improvement. To validate the effectiveness of our proposed method, we conduct extensive experiments. First, one of the most widely-used visual benchmarks CIFAR10 is adopted, following the existing methods. Additionally, we adopt a MRI dataset on revival time prediction for liver cancer and a myometrioum image dataset of low blood pressure diagnosis for the pregnant to further testify the capability of our model on real-life applications. The results on these three datasets illustrate that our model can significantly improve the base model trained with nnPU and surpass the state-of-the-art methods. Moreover, when trained with fewer positive labeled data, our model still performs better than competitors which use more labeled data, which supports the argument that our model can solve the data scarcity problem to some extent. In summary, our paper has the following contributions: * • We analyze the essence of the PU data and find that those hard data could make the existing pseudo-labeling methods less effective. * • To take advantage of the analysis, we propose to use an early-stop splitting strategy to excavate the easy and hard data from the unlabeled set. * • We leverage the Jensen-Shannon divergence loss to learn the easy data, and propose a novel dual-source consistency regularization to efficiently utilize the hard unlabeled data. * • With the help of our proposed techniques, we consistently improve the base model among different datasets when using various amount of labeled samples. In the following context, we first briefly introduce the related works in Sec. 2, which includes PU learning, learning with noisy labels and consistency in deep learning. Then in Sec. 3 we give the problem formulation of PU learning and our proposed method. Experiment results among multiple datasets are listed in Sec. A. ## 2\. Related Works Positive-unlabeled Learning Positive-unlabeled learning is a surging research topic that aims to learn patterns on a training set with only few labeled positive data and sufficient unlabeled data. The research on PU learning can be dated back to (Liu et al., 2002, 2003; Yu et al., 2002) in which reliable negative samples are selected and used to train a set of classifiers. The recent effort on PU learning with deep deep learning can be generally categorized into two buckets. One is about the design of objective function that can unbiasedly estimate the risk. uPU (Du Plessis et al., 2014) shows that a cost-sensitive classifier can be used by reform the risk in original classification, with known class prior. nnPU (Kiryo et al., 2017) points out the tendency of overfitting complex models when using uPU. As a solution they impose a non-negative constraint on the objective function of uPU, which leads to better generalization. Some other papers follow these two works to find powerful objective functions in different settings such as imbalanced data (Su et al., 2021), biased negative data (Hsieh et al., 2019), data selection bias (Kato et al., 2018), etc. Some other works focus on sample selection from the unlabeled samples based on the former methods. For example, (Xu et al., 2019) analyze the reasonability of only selecting positive data, Self-PU (Chen et al., 2020a) leverages self-paced learning to gradually update the base model with newly-learned knowledge, PUUPL (Dorigatti et al., 2022) strengthens the calibration of base model so that the pseudo-labels are all uncertainty-aware. Note that the concept of hardness has also been mentioned in Self-PU. However, they refer to the this concept as the reliability of pseudo-labels, which is reflected by the confidence of model predictions. Our method split the data into easy and hard set based on the intrinsic property of data, which is different with the existing ideas. Moreover, we point out that directly using cross entropy on pseudo-labeled data and PU loss on unlabeled data is inappropriate. Instead, we propose to conquer the easy and hard unlabeled data based on noise-tolerant learning and dual-source consistency regularization respectively. Additionally, PU learning is different with semi-supervised learning which also uses the unlabeled data as the regularizer. The unlabeled data in PU learning plays a more important role of estimating the data distribution since no labeled negative samples are available. Learning with noisy labels Since the pseudo-labeling strategy in PU learning inevitably generates wrongly-labeled samples, i.e. noisy samples, such methods are correlated with the task of learning with noisy labels, which is aimed to train a more robust model from the noisy dataset. Methods in learning with noisy labels mainly belong to two groups: noise robust algorithm and noise detection. Noise robust algorithms target designing network architectures (Chen and Gupta, 2015; Goldberger and Ben-Reuven, 2016) or objective function (Zhang and Sabuncu, 2018; Ghosh et al., 2017; Englesson and Azizpour, 2021) that are robust against side effect of noisy samples. Noise detection methods involve a pipeline for selecting and processing the noisy data from the training set. Typical selection basis includes large error (Shen and Sanghavi, 2019), incoherent gradient (Chatterjee, 2020), inconsistency among different networks or different views of the same input (Yu et al., 2019; Zhou et al., 2020) and so on. As for the strategies for handling noisy samples, label- correction modules are proposed in many works (Li et al., 2017; Tanaka et al., 2018; Vahdat, 2017). We in this paper leverage the technique in this topic to solve PU learning, which has never been studied in the previous works. Moreover, while directly applying these advantaged noise detection methods may also help dealing with the noisy pseudo-labels in PU learning, we will show in the experiments that these methods could easily fail when facing with some of the training data. Consistency Regularization Recently consistency regularization has been widely used in deep learning as extra supervision besides the annotations. Specifically, contrastive learning methods (He et al., 2020; Chen et al., 2020b; Caron et al., 2020; Chen and He, 2021a) take as objective function the similarity among features of different views of the same image, which can be seen as the feature level consistency. Such objectives are utilized to enhance the instance discriminative ability for networks. Moreover, for semi- supervised learning, many works (Sohn et al., 2020; Verma et al., 2019; Jeong et al., 2019) adopt prediction level consistency to increase the stability of training. Compared with these methods, our proposed dual-source consistency regularization takes into account the various properties of features and predictions, which leads to consistency with more capacity from multiple sources. ## 3\. Methodology Problem Formulation Suppose we have a data distribution $\mathcal{T}$ which can be represented as a joint distribution of image distribution $\mathcal{I}$ and label distribution $\mathcal{Y}$, i.e., $\mathcal{T}=\mathcal{I}\times\mathcal{Y}$. In PU learning, the training set $\mathcal{T}^{train}$ is composed of two parts including $\mathcal{T}^{p}$ which only contains few labeled positive samples and $\mathcal{T}^{u}$ which contains abundant unlabeled samples. The labeled positive $\mathcal{T}^{p}=\\{(I_{i}^{p},y_{i}^{p})\\}_{i=1}^{n_{p}}$, where $I_{i}^{p}\sim\mathcal{I},y_{i}^{p}=1$. The unlabeled $\mathcal{T}^{u}=\\{I_{i}^{u}\\}_{i=1}^{n_{u}}$, where $I_{i}^{u}\sim\mathcal{I}$. Our goal is to learn a network $\phi$, which can well generalize to the test set $\mathcal{T}^{test}=\\{(I_{i},y_{i})\\}_{i=1}^{n_{t}}$. Figure 2. Overview of our model. We first train a base teacher model using nnPU as objective function. Then the unlabeled data is split into easy and hard set via a temporary model which is trained with knowledge distillation given the base model. Then we iterate a new training process applying different learning strategies to the easy and hard data to train the student model. Overview. We propose a hardness-based splitting strategy to deal with the PU learning, which is schematically shown in Fig. 2. Concretely, we first train a base model $\phi_{base}$ with nnPU (Kiryo et al., 2017) (Sec. 3.1), from which all of the following supervision is generated. After attaining this base model, we split the unlabeled set $\mathcal{T}^{u}$ into easy set $\mathcal{T}^{e}$ and hard set $\mathcal{T}^{h}$ by the training accuracy of a temporary student network $\phi_{tmp}$ which is trained with naive knowledge distillation from $\phi_{base}$ (Sec. 3.2). Then we re-train a network $\phi$ with different learning strategies on $\mathcal{T}^{easy}$ and $\mathcal{T}^{hard}$. For the easy set $\mathcal{T}^{easy}$, we utilize a noise-tolerant objective function to make use of the pseudo-labels from $\phi_{base}$ (Sec. 3.3). For the hard set $\mathcal{T}^{hard}$, we adopt a dual-source consistency regularization to take advantage of the helpful knowledge from $\phi_{base}$ without using the pseudo-labels (Sec. 3.4). Such a learning strategy can be repeated by replacing the teacher model $\phi_{base}$ with the newly-learned $\phi$ to further improve the performance not only for commonly-used visual datasets, but also for PU learning settings in real scenario. ### 3.1. Preliminary: Base Model Training We first briefly introduce the nnPU (Kiryo et al., 2017) as our model. The nnPU is a non-negative version of an unbiased risk estimator as following: (1) $\displaystyle L_{nnPU}$ $\displaystyle=\frac{\pi}{n_{p}}\sum_{i=1}^{n_{p}}L(\phi(I_{i}^{p}),1)+$ (2) $\displaystyle max\\{0,\frac{1}{n_{u}}\sum_{i=1}^{n_{u}}L(\phi(I_{i}^{u}),-1)-\frac{\pi}{n_{u}}\sum_{i=1}^{n_{u}}L(\phi(I_{i}^{p}),-1)\\}$ where $\pi$ is the class prior of positive samples, i.e. $\pi=P(I_{i}|y_{i}=1)$, which is supposed to be known as (Kiryo et al., 2017). By training with $L_{nnPU}$ on $\mathcal{T}^{train}$, we can get our base model $\phi_{base}$ and correspondingly a pseudo-labeled extension of $\mathcal{T}^{u}$: (3) $\displaystyle\hat{\mathcal{T}}^{u}$ $\displaystyle=\\{(I_{i}^{u},\hat{y_{i}}^{u})\\}_{i=1}^{n_{u}}$ (4) $\displaystyle\hat{y_{i}}^{u}$ $\displaystyle=\phi_{base}(I_{i}^{u})$ Note that $\hat{y_{i}}^{u}$ can be either soft-label or hard-label. In this way, $\hat{\mathcal{T}}^{u}$ is composed of correctly-labeled data (i.e. clean data), and wrongly-labeled data (i.e., noisy data). A straightforward way, like in Self-PU (Chen et al., 2020a), to make use of $\hat{\mathcal{T}}^{u}$ is to train a supervised learning model based on these pseudo-labels and then iteratively rectify the noisy or wrongly pseudo-labeled samples. However, as we will show in the experiments (see Fig. 8), some training samples are intrinsically hard to learn. This implies that the models can barely receive knowledge from these samples no matter they are affiliated with correct pseudo-labels or not. Thus such a learning strategy will lead to sub-optimal performance. As a solution, we propose to use the following method to separate these hard samples from $\mathcal{T}^{u}$ and then learn the easy and hard data respectively. ### 3.2. Early-stop Splitting Strategy Given the base model, we consider splitting $\mathcal{T}^{u}$ apart into easy and hard data, and then conquer them respectively. Exactly, we learn a new student model $\phi_{tmp}$ based on the pseudo-labels generated by $\phi_{base}$ with soft-label cross entropy as objective function: (5) $L_{tmp}=\sum_{i=1}^{n_{u}}-\hat{y}_{i}^{u}log(\phi_{tmp}(I_{i}^{u}))$ We set a small learning rate for this optimization process to track the variety of $\phi_{tmp}$ in detail. After each epoch, we record the training accuracy. Once the accuracy is over a threshold $\tau$, we stop the training and split the unlabeled set $\hat{\mathcal{T}}^{u}$ based on the prediction coherence between $\phi_{tmp}$ and $\phi_{base}$. The samples that are still not learned by $\phi_{tmp}$ are categorized as hard samples, which compose the hard set $\mathcal{T}^{hard}$. The rest are called easy set, denoted as $\mathcal{T}^{easy}$. Since the noisy samples generally have different patterns from clean samples as in existing works (Chatterjee, 2020; Arpit et al., 2017), most of them would be learned more slowly than clean ones. As a consequence, $\mathcal{T}^{hard}$ can have a much larger proportion of noisy samples than $\mathcal{T}^{u}$, and the noise rate of $\mathcal{T}^{easy}$ is much smaller. Meanwhile, the roles of these two subsets are totally different: (1) The easy clean samples take up the most part of $\mathcal{T}^{u}$, from which the model learns its patterns. (2) The easy noisy samples, despite with a small sample size, can be easily fitted by the model and hurt the generalization ability. (3) The hard samples, while can hardly be learned by $\phi_{tmp}$ and have high noise rate, can regularize the model for better performance. Especially, the hard clean samples can provide data knowledge which is not contained in the easy set. Therefore, it is essential to utilize a learning strategy robust to noise on $\mathcal{T}^{easy}$ and a new supervision other than pseudo-labels on $\mathcal{T}^{hard}$. To this end, we propose the following noise-tolerant learning on easy samples and dual-source consistency regularization on hard samples. ### 3.3. Noise-tolerant Learning on Easy Samples As mentioned above, the easy set $\mathcal{T}^{easy}$ has two noticeable properties: (1) The noise rate is much smaller, thus most pseudo-labels are reliable. (2) The negative effect of easy noisy samples on generalization ability cannot be neglected. Directly learning the easy samples with plain objective like cross entropy will make the model be trapped by the noisy samples. A straightforward thought is if we can use noise detection methods to further cull the noisy easy samples. However, due to the fact that the slowly- learned hard noisy samples have already been taken out from $\mathcal{T}^{easy}$, most methods based on the difference of fitting speed or fitting performance between noisy and clean data could fail in our case, as we will show in experiments. As an alternative solution, we adopt a noise-robust objective function to process the easy data, which was proposed in (Englesson and Azizpour, 2021). Concretely, for each easy sample $I_{i}^{easy}\in\mathcal{T}^{easy}$ and its corresponding pseudo-label $\hat{y}_{i}$, the following objective function is used: (6) $L_{easy}=\frac{(\rho D_{KL}(\phi(I_{i}^{easy})\|m)+(1-\rho)D_{KL}(\hat{y}_{i}\|m))}{-(1-\rho)log(1-\rho)}$ where $m=\rho\phi(I_{i}^{easy})+(1-\rho)\hat{y}_{i}$ which is an interpolation between the pseudo-label and prediction from $\phi$, $D_{KL}$ denotes the Kullback-Leibler divergence, $\rho$ is a hyper-parameter. Accordingly, Eq. 6 can take into account both robustness to noise and the convergence speed. As we will show in our experiments, this term is more effective than plain knowledge distillation using pseudo-labels. ### 3.4. Dual-source Consistency Regularization on Hard Samples The hard set $\mathcal{T}^{hard}$ has a totally different characteristic compared with $\mathcal{T}^{easy}$ and $\hat{\mathcal{T}}^{u}$. $\mathcal{T}^{hard}$ contains much less samples than the other two sets, while its proportion of noisy samples are much larger as we have explained above. Besides, these samples can be hardly learned with the pseudo-labels, which is reflected by the low training accuracy in the splitting phase. Therefore, it is not proper to adopt the same training strategy as for the easy samples. To this end, we explore to use these data via a dual-source consistency regularization without using the pseudo-labels. Specifically, we use the following objective function to learn the hard samples: (7) $L_{hard}=L_{low}+L_{high}$ where $L_{low}$ is a low-level cross-consistency between $\phi_{base}$ and $\phi$, and $L_{high}$ is a high-level self-consistency of $\phi$, which will be introduced in the following. Low-level cross-consistency. While the pseudo-label generated by $\phi_{base}$ is not usable due to the high noisy rate of $\mathcal{T}^{hard}$, the teacher model can still provide some guidance of visual knowledge embedded in the low- level features. According to the existing study (Lee et al., 2016), compared with high-level features that is correlated with the semantic category of each sample, these low-level features contain more basic and class-agnostic visual patterns, such as edges and colors. Such patterns are still reusable no matter if it leads to noisy pseudo-prediction in the $\mathcal{T}^{hard}$. Therefore, we regularize $\phi$ with the consistency between low-level features of $\phi$ and $\phi_{base}$ on the same sample. Specifically, given a hard sample $I_{i}^{hard}\in\mathcal{T}^{hard}$, we simultaneously extract the corresponding features from $\phi$ and $\phi_{base}$ and compute their similarity as the objective function: (8) $L_{cross}=\|\phi^{1}(I_{i}^{hard})-\phi_{base}^{1}(I_{i}^{hard})\|_{2}$ where $\phi^{1}$ denotes the first convolutional layer of $\phi$, and the same for $\phi_{base}^{1}$. High-level self-consistency. The guidance from $\phi_{base}$ is sub-optimal for the high-level features and the predictions. Therefore, inspired by recent research on self-supervised learning (Chen and He, 2021b) and semi-supervised learning (Sohn et al., 2020), we propose to utilize self-consistency on these high-level information. Specifically, for a hard sample $I_{i}^{hard}$, we first process it via both weak and strong augmentation pipelines $A^{weak},A^{strong}$, which leads to two different views of this image $\hat{I}_{i}^{hard}=A^{weak}(I_{i}^{hard}),\tilde{I}_{i}^{hard}=A^{strong}(I_{i}^{hard})$. Then the high-level features are generated from the last convolutional layer of $\phi$, together with the classification prediction from which the consistency can be calculated: (9) $\displaystyle\hat{X}_{i}$ $\displaystyle=\phi^{l}(\hat{I}_{i}^{hard}),\hat{y}_{i}=\phi(\hat{I}_{i}^{hard})$ (10) $\displaystyle\tilde{X}_{i}$ $\displaystyle=\phi^{l}(\tilde{I}_{i}^{hard}),\tilde{y}_{i}=\phi(\tilde{I}_{i}^{hard})$ (11) $\displaystyle L_{pred}$ $\displaystyle=D_{KL}(\hat{y}_{i}\|\tilde{y}_{i})$ After that a new predictor $f$ is introduced to transform the high-level features into another embedding space. Based on that we can regularize $\phi$ with high-level feature consistency: (12) $L_{feat}=\frac{1}{2}D(\hat{X}_{i},f(\tilde{X}_{i}))+\frac{1}{2}D(\tilde{X}_{i},f(\hat{X}_{i}))$ where $D$ denotes the normalized negative cosine similarity. The high-level self-consistency can then be summarized as the combination of $L_{pred}$ and $L_{feat}$ with tunable weights $\alpha,\beta$: (13) $L_{high}=\alpha L_{pred}+\beta L_{feat}$ This objective function can help the model with better instance discrimination and to be more stable against disturbance on the samples. Compared with the self-supervised consistency used in Self-PU (Chen et al., 2020a), our method considers the intrinsically high noise-rate of hard data, thus making better use of teacher model and leading to more effective regularization. Meanwhile, our proposed consistency does not require training multiple networks at one time so that we can reduce the computation cost. Remark: why not use nnPU on $\mathcal{T}^{hard}$. One would ask if a simple solution works for the hard samples, i.e., still taking them as unlabeled samples and learning them by nnPU loss together with $\mathcal{T}^{p}$ as in Eq. 1. We here provide an intuitive explanation about this problem. There are two key components that makes $L_{nnPU}$ works: (1) The $\mathcal{T}^{u}$ can well represent the data distribution. (2) We can have an accurate estimation of class prior $\pi$. However, since the the sample size of $\mathcal{T}^{hard}$ is much smaller than that of $\mathcal{T}^{u}$, there is little chance that it can provide as much information to estimate the data distribution as $\mathcal{T}^{u}$ can. On the other hand, if we think of $\mathcal{T}^{p}$ and $\mathcal{T}^{hard}$ as another training set in PU setting, then using $L_{nnPU}$ requires a new class prior $\pi^{\prime}$, which cannot be easily estimated. Therefore it is inappropriate to use $L_{nnPU}$ on $\mathcal{T}^{hard}$, which will be empirically verified in Sec. A. ### 3.5. Iterative Training With the above pipeline, we can get a $\phi$ that not only inherits the useful knowledge from $\phi_{base}$, but also eliminate the side-effect of noisy samples. Moreover, $\phi$ can be used as the teacher model again to train another new student model for further improvement. ## 4\. Experiments Figure 3. Image examples in the liver cancer dataset and myometrium dataset. Dataset To validate the efficacy of our model, we first follow Self-PU (Chen et al., 2020a) to conduct extensive experiments on CIFAR10 (Krizhevsky et al., 2009). CIFAR10 is a basic visual dataset containing 50000 training images and 10000 testing images from 10 categories. We follow the setting in Self-PU to take 4 categories of vehicles (i.e., ’airplane’, ’automobile’, ’ship’, ’truck’) to compose the positive class and the other animal classes (i.e., ’bird’, ’cat’, ’deer’, ’dog’, ’frog’, ’horse’) for the negative class. In each experiment, 500/1000/3000 positive samples are randomly chosen and the rest are set as unlabeled samples. Moreover, as shown in Fig. 3, we further adopt two new benchmarks of liver cancer data and myometrium data, which would be publicly released. The liver cancer dataset contains MRI data of 259 patients and their corresponding survival time. We split the dataset by setting samples with survival time less than 90 as positive. This leads to totally 145 positive samples, from which we randomly choose 40 as labeled samples, and 114 negative samples. For test set, we have 65 negative samples and 35 positive samples. The Myometrium data is used to diagnose the low blood pressure of the pregnant, with 327 positive images and 233 negative images in total. We split the dataset into 395 training images, 53 validation images and 112 testing images, and then randomly sample 40 positive images as labeled set. Implementation Details For CIFAR10 we use the same network architecture as in Self-PU, i.e. a 13-layer CNN for CIFAR10. About Liver, we adopt a multi-branch ResNet18, with the detail in supplementary. For Myometrium we finetune the whole layers of an ImageNet-pretrained ResNet18. For all datasets, we first train a base nnPU model for 50 epoches, using Adam (Kingma and Ba, 2014) with learning rate of $10^{-4}$. Batch size is set to 512 for CIFAR10 and 64 for other datasets. After training the base model, we fix the splitting of labeled and unlabeled samples and then further train $\phi_{tmp}$ using SGD as optimizer with learning rate of $10^{-3}$ to avoid converging to quickly. $\phi$ is trained for 100 epoches using Adam with learning rate of $0.00005$. As for the hyper-parameters, $\tau,\rho,\alpha,\beta$ are set as 92%, 0.7, 0.3, 0.1, respectively. We use random cropping and flipping as weak augmentation. Color jittering is adopted as strong augmentation for CIFAR10 and CutOut is adopted for the other two medical datasets. Metrics For all results, we report the mean accuracies and standard deviations among 5 repetitions of each experiment. Competitors We choose several former methods in PU learning as our competitors including uPU (Du Plessis et al., 2014), nnPU (Kiryo et al., 2017), DAN (Liu et al., 2019), Self-PU (Chen et al., 2020a), P3MIX-C (Li et al., 2021), PUUPL (Dorigatti et al., 2022). Model | CIFAR10 ---|--- 500 | 1000 | 3000 uPU (Du Plessis et al., 2014) | — | 88.00$\pm$0.62 | — nnPU (Kiryo et al., 2017) | — | 88.60$\pm$0.40 | — nnPU∗ | 87.22$\pm$0.46 | 89.02$\pm$0.39 | 90.53$\pm$0.36 Self-PU (Chen et al., 2020a) | — | 89.68$\pm$0.22 | 90.77$\pm$0.21 PUbN (Hsieh et al., 2019) | — | 89.30$\pm$0.57 | — P3MIX-C (Li et al., 2021) | — | 87.90$\pm$0.50 | — VPU (Chen et al., 2020c) | — | 85.06$\pm$0.55 | 87.50$\pm$1.05 DAN (Liu et al., 2019) | — | — | 89.70$\pm$0.40 PUUPL (Dorigatti et al., 2022) | — | 89.84$\pm$0.13 | 91.37$\pm$0.05 Ours | 89.18$\pm$0.12 | 90.51$\pm$0.10 | 92.51$\pm$0.10 Table 1. Test accuracies and standard deviation on CIFAR10 with three positive size, * denotes results reproduced by us. Model | Myometrium | Liver ---|---|--- uPU (Du Plessis et al., 2014) | 60.18$\pm$2.33 | 40.40$\pm$5.50 nnPU (Kiryo et al., 2017) | 62.32$\pm$2.71 | 56.60$\pm$5.90 Self-PU (Chen et al., 2020a) | 59.11$\pm$1.16 | 55.60$\pm$8.17 Ours | 65.89$\pm$2.13 | 63.80$\pm$1.64 Supervised | 72.32 | 68.00 Table 2. Test accuracies and standard deviation on Myometrium and Liver. All results of competitors are reproduced by us. ### 4.1. Main Results We compare our model with the competitors in Tab. 1 and Tab. 2. Note that for nnPU on CIFAR10 we both report the results in the original paper and the one reproduced by us. CIFAR10 result. For CIFAR10, we train our model among three different positive sample size. When $n_{p}=500$, the test accuracy of our model is 1.96% better than that of nnPU, and comparable with nnPU model trained with 1000 labeled positive samples. This means that support nnPU with our proposed method is able to solve the problem of data scarcity for positive samples. When $n_{p}=1000$, which is the most common setting, our proposed model attains 1.49% accuracy increasement against nnPU and performs 0.67% better than the best counterpart PUUPL. When $n_{p}=3000$, the improvement on nnPU of our model reaches 1.6%, leading to a much better result than all of the competitors. Such an improvement is consistently significant among these three settings, which validates the effectiveness of our proposed method and shows the potential that our model can be used as a fixed suppplement for nnPU. As for the standard deviation, our model is on par with the competitors, which shows our method is stable enough to be safely utilized. Myometrium and Liver results. We report results of different methods and one oracle model trained with whole labeled training set together on the two medical datasets in Tab. 2. For Myometrioum, our model is better than nnPU by 3.75% with comparable standard deviation. For Liver, our model improves nnPU by 7.2% with remarkably more stability. It is noticeable that Self-PU is worse than nnPU on these datasets. We suspect this is due to the poor performance of nnPU leads to unreliable pseudo-labels, thus increasing the size of clean noisy samples. In this way, since Self-PU cannot well handle such data, the performance could be damaged. ### 4.2. Ablation Study To further verify the efficacy of our contributions, we conduct several ablation studies on CIFAR10, including each part of our model and the choice of hyper-parameters. Figure 4. (a) Size of noisy samples in easy and hard set with different $\tau$. (b) Proportion of noisy samples in easy and hard set with different $\tau$. (c) Test accuracy when trained with the easy and hard set that are generated with different $\tau$. (d) Number of clean and noisy samples that are selected from the clean set by SPR in each epoch. Early stop | Acc. ---|--- 500 | 1000 | 3000 w/o | 88.57 | 89.86 | 91.73 w | 89.28 | 90.51 | 92.51 Table 3. Test accuracies on CIFAR10 of models trained with and without early stop splitting. Loss | label type | Acc. ---|---|--- 500 | 1000 | 3000 CE | hard | 87.51 | 89.02 | 91.55 soft | 88.15 | 89.48 | 92.06 DJS | hard | 88.61 | 89.58 | 92.11 soft | 89.28 | 90.51 | 92.51 Table 4. Test accuracies on CIFAR10 of our model trained with soft cross entropy, hard cross entropy, soft Jensen-Shannon Divergence and hard Jensen-Shannon Divergence. Loss | Acc. ---|--- 500 | 1000 | 3000 no | 88.52 | 89.71 | 91.46 nnPU | 87.91 | 89.43 | 90.66 self-consistency | 89.03 | 90.09 | 92.11 cross consistency | 89.17 | 90.42 | 92.27 dual-source consistency | 89.28 | 90.51 | 92.51 Table 5. Test accuracies on CIFAR10 of our model trained with different losses on hard data. What can early-stop splitting do? First of all, we would like to show the effectiveness of the early-stop splitting strategy. To do so, we conduct an oracle experiment where all training set labels are available for indicating which samples are wrongly labeled by $\phi_{base}$. Then we gradually increase the accuracy threshold $\tau$ and visualize the sample size and proportion of noisy samples that are categorized as easy and hard ones respectively. The results are shown in Fig. 4. We can find that (1) As $\tau$ increases, less samples can be filtered out from $\mathcal{T}^{u}$, but the noise rate of $\mathcal{T}^{hard}$ gets larger. (2) When $\tau$ is relatively small as 70%, the proportion of noisy sample in $\mathcal{T}^{hard}$ is twice as larger than that of $\mathcal{T}^{u}$, which coincides with the common sense that noisy samples are generally learned more slowly than clean samples. Also, such a strategy can effectively reduce the noise rate of $\mathcal{T}^{easy}$. (3) There is a balance between the threshold and the test accuracy. The benefit from increasing $\tau$ saturates at about 92%, then the test accuracy gradually decreases. This is reasonable since larger $\tau$ means smaller hard set, which leads to a less representative hard set. (4) As shown in 3, performance is worse when early-stop strategy is not used , i.e., $L_{easy}$ is applied to all unlabeled data, which indicates the efficacy of this strategy. Can noise detection methods fit in our case? As analyzed in Sec. 3.3, noise detection methods may not be proper for further detecting noisy samples in $\mathcal{T}^{easy}$. To verify this, we adopt two state-of-the-art methods in noisy label learning, SPR (Wang et al., 2022) and CGH (Chatterjee, 2020). SPR is based on penalized regression to build a regularization path, which is used to detect the most confident noisy samples. Fig. 4(d) shows the result using SPR. With about 9% noisy samples in $\mathcal{T}^{easy}$, SPR can only filter out a set with noise rate of 10% in each epoch, which is even worse than uniform sampling. The result of CGH which is provided in the supplementary is also unsatisfactory, failing to detect enough noisy samples. Considering these results, we speculate that the easy set is too challenging for these noise detection methods to work. Choices of objective functions for easy data. In Tab. 4 we compare the noise- tolerant DJS loss $L_{easy}$ with cross entropy, using both hard and soft labels. The results demonstrate that: (1) DJS loss is more effective than cross entropy. When using soft label, DJS loss is better than cross entropy by 1.13%, 1.03% and 0.45% when using 500, 1000 and 3000 positive samples. (2) Using soft label is better than hard label no matter which objective function is utilized in training. We think this is due to the base model is calibrated enough. Therefore transferring low-confidence soft labels to hard labels could hurt the model training. #Iter | Acc. ---|--- 500 | 1000 | 3000 1 | 88.86 | 90.07 | 92.02 2 | 89.28 | 90.51 | 92.51 3 | 89.19 | 90.55 | 92.64 4 | 89.12 | 90.49 | 92.93 Table 6. Test accuracies on CIFAR10 of our model trained with different number of iterations. Choices of objective functions for hard data. In Tab. 5 we compare several choices to optimize hard data, including (1) not using hard data. (2) nnPU loss on hard data and positive labeled data. (3) cross consistency as in Eq. 8. (4) self-consistency as in Eq. 13. (5) dual-source consistency as in Eq. 7. The results reflect several important points: (1) Either forms of consistency can improve the model. Specifically, the model using self-consistency is better than the model without using hard data by 0.51%, 0.32% and 0.65% among three settings, and the model using cross consistency is better by 0.65%, 0.71% and 0.81%. (2) Using $L_{nnPU}$ worsens the performance by 0.61%, 0.32% and 0.80%, which is compatible with our analysis in Sec. 3.4. (3) Cross consistency is more effective than self-consistency, which is reasonable because cross consistency can guide the model to learn better low-level feature. Based on such features, the model can then learn to build useful high-level features easily. Relationship between test accuracy and number of iterations. Since our method utilizes an iterative training strategy, we investigate how the number of iterations could affect the final result. Tab. 6 shows that while training for two iterations is much better than single iteration, i.e. 0.44% improvement, repeating the algorithm more times does not significantly boost the performance when using 500 or 1000 positive samples. Under the setting of 3000 positive samples, training 4 iterations leads to the best result. The drawback of using more iterations is the longer training time, which could make the algorithm less practical and not comparable with the existing methods. Also, since the variation caused by the randomness can stack during each iteration, the final result with high iteration number would be more unstable. Usage | Acc. ---|--- 500 | 1000 | 3000 all data | 88.96 | 90.49 | 92.25 hard data | 89.28 | 90.51 | 92.51 Table 7. Test accuracies on CIFAR10 of our model trained with dual-source consistency applied to all unlabeled data or only hard data. Usage of consistency regularization As a straightforward thought, our proposed dual-source consistency regularization can also be applied to easy data. We testify this idea with results in Tab. 7. By applying dual-source consistency on all unlabeled data, the model gets accuracy of 88.96%, 90.49% and 92.25%, receiving no marginal advancement. We think the reason is that since most pseudo-labels of easy data is reliable and easy to fit, the supervision with regard to semantic labels would be stronger than other forms of guidance. Therefore, using DJS loss is enough for the easy data. What kind of data is selected by early-stop splitting? We visualize some positive and negative data from easy and hard set of three datasets in Fig. 8. We can find that the easy data is discriminative enough, with obvious patterns and symptoms. On the other side, the hard samples are generally much harder to distinguish. For example the hard negative sample of CIFAR10 is actually an aeroplane but also looks like a bird, and the hard positive sample in Myometrium does not have the same symptoms like apparent blood vessels with dark red color. ## 5\. Conclusion In this paper we explore the task of PU learning. The problem of intrinsic hardness of some unlabeled samples is highlighted which may affect the existing methods’ performance. We analyze the different role of easy and hard data in training and propose a novel training strategy as a solution to this problem. Our model includes an early-stop splitting strategy, a noise-tolerant learning strategy for easy data and a dual-source consistency regularization for hard data. We achieve remarkable results on CIFAR10 and two medical datasets, which shows the strength of our method as a promising choice for solving PU learning both in scientific research and in real-life applications. ## References * (1) * Arpit et al. (2017) Devansh Arpit, Stanisław Jastrzebski, Nicolas Ballas, David Krueger, Emmanuel Bengio, Maxinder S Kanwal, Tegan Maharaj, Asja Fischer, Aaron Courville, Yoshua Bengio, et al. 2017\. A closer look at memorization in deep networks. In _International conference on machine learning_. PMLR, 233–242. * Caron et al. (2020) Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. 2020\. Unsupervised learning of visual features by contrasting cluster assignments. _Advances in Neural Information Processing Systems_ 33 (2020), 9912–9924. * Chatterjee (2020) Satrajit Chatterjee. 2020\. Coherent gradients: An approach to understanding generalization in gradient descent-based optimization. _arXiv preprint arXiv:2002.10657_ (2020). * Chen et al. (2020c) Hui Chen, Fangqing Liu, Yin Wang, Liyue Zhao, and Hao Wu. 2020c. A Variational Approach for Learning from Positive and Unlabeled Data. _Advances in Neural Information Processing Systems_ 33 (2020), 14844–14854. * Chen et al. (2020b) Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. 2020b. A simple framework for contrastive learning of visual representations. In _International conference on machine learning_. PMLR, 1597–1607. * Chen et al. (2020a) Xuxi Chen, Wuyang Chen, Tianlong Chen, Ye Yuan, Chen Gong, Kewei Chen, and Zhangyang Wang. 2020a. Self-pu: Self boosted and calibrated positive-unlabeled training. In _International Conference on Machine Learning_. PMLR, 1510–1519. * Chen and Gupta (2015) Xinlei Chen and Abhinav Gupta. 2015. Webly supervised learning of convolutional networks. In _Proceedings of the IEEE international conference on computer vision_. 1431–1439. * Chen and He (2021a) Xinlei Chen and Kaiming He. 2021a. Exploring simple siamese representation learning. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_. 15750–15758. * Chen and He (2021b) Xinlei Chen and Kaiming He. 2021b. Exploring simple siamese representation learning. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_. 15750–15758. * Dorigatti et al. (2022) Emilio Dorigatti, Jann Goschenhofer, Benjamin Schubert, Mina Rezaei, and Bernd Bischl. 2022\. Positive-Unlabeled Learning with Uncertainty-aware Pseudo-label Selection. _arXiv preprint arXiv:2201.13192_ (2022). * Du Plessis et al. (2014) Marthinus C Du Plessis, Gang Niu, and Masashi Sugiyama. 2014\. Analysis of learning from positive and unlabeled data. _Advances in neural information processing systems_ 27 (2014). * Englesson and Azizpour (2021) Erik Englesson and Hossein Azizpour. 2021. Generalized Jensen-Shannon Divergence Loss for Learning with Noisy Labels. _Advances in Neural Information Processing Systems_ 34 (2021). * Fan et al. (2020) Deng-Ping Fan, Ge-Peng Ji, Guolei Sun, Ming-Ming Cheng, Jianbing Shen, and Ling Shao. 2020\. Camouflaged object detection. In _Proceedings of the IEEE/CVF conference on computer vision and pattern recognition_. 2777–2787. * Ghosh et al. (2017) Aritra Ghosh, Himanshu Kumar, and PS Sastry. 2017. Robust loss functions under label noise for deep neural networks. In _Proceedings of the AAAI conference on artificial intelligence_ , Vol. 31. * Goldberger and Ben-Reuven (2016) Jacob Goldberger and Ehud Ben-Reuven. 2016. Training deep neural-networks using a noise adaptation layer. (2016). * He et al. (2020) Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. 2020. Momentum contrast for unsupervised visual representation learning. In _Proceedings of the IEEE/CVF conference on computer vision and pattern recognition_. 9729–9738. * Hsieh et al. (2019) Yu-Guan Hsieh, Gang Niu, and Masashi Sugiyama. 2019. Classification from positive, unlabeled and biased negative data. In _International Conference on Machine Learning_. PMLR, 2820–2829. * Jeong et al. (2019) Jisoo Jeong, Seungeui Lee, Jeesoo Kim, and Nojun Kwak. 2019\. Consistency-based semi-supervised learning for object detection. _Advances in neural information processing systems_ 32 (2019). * Kato et al. (2018) Masahiro Kato, Takeshi Teshima, and Junya Honda. 2018\. Learning from positive and unlabeled data with a selection bias. In _International conference on learning representations_. * Kingma and Ba (2014) Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. _arXiv preprint arXiv:1412.6980_ (2014). * Kiryo et al. (2017) Ryuichi Kiryo, Gang Niu, Marthinus C Du Plessis, and Masashi Sugiyama. 2017. Positive-unlabeled learning with non-negative risk estimator. _Advances in neural information processing systems_ 30 (2017). * Krizhevsky et al. (2009) Alex Krizhevsky, Geoffrey Hinton, et al. 2009\. Learning multiple layers of features from tiny images. (2009). * Lee et al. (2016) Gayoung Lee, Yu-Wing Tai, and Junmo Kim. 2016. Deep saliency with encoded low level distance map and high level features. In _Proceedings of the IEEE conference on computer vision and pattern recognition_. 660–668. * Li et al. (2021) Changchun Li, Ximing Li, Lei Feng, and Jihong Ouyang. 2021\. Who Is Your Right Mixup Partner in Positive and Unlabeled Learning. In _International Conference on Learning Representations_. * Li et al. (2017) Yuncheng Li, Jianchao Yang, Yale Song, Liangliang Cao, Jiebo Luo, and Li-Jia Li. 2017\. Learning from noisy labels with distillation. In _Proceedings of the IEEE International Conference on Computer Vision_. 1910–1918. * Liu et al. (2003) Bing Liu, Yang Dai, Xiaoli Li, Wee Sun Lee, and Philip S Yu. 2003. Building text classifiers using positive and unlabeled examples. In _Third IEEE international conference on data mining_. IEEE, 179–186. * Liu et al. (2002) Bing Liu, Wee Sun Lee, Philip S Yu, and Xiaoli Li. 2002\. Partially supervised classification of text documents. In _ICML_ , Vol. 2. Sydney, NSW, 387–394. * Liu et al. (2019) Fangqing Liu, Hui Chen, Liyue Zhao, and Hao Wu. 2019\. Discriminative adversarial networks for positive-unlabeled learning. _arXiv_ (2019), arXiv–1906. * Shen and Sanghavi (2019) Yanyao Shen and Sujay Sanghavi. 2019. Learning with bad training data via iterative trimmed loss minimization. In _International Conference on Machine Learning_. PMLR, 5739–5748. * Sohn et al. (2020) Kihyuk Sohn, David Berthelot, Nicholas Carlini, Zizhao Zhang, Han Zhang, Colin A Raffel, Ekin Dogus Cubuk, Alexey Kurakin, and Chun-Liang Li. 2020. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. _Advances in Neural Information Processing Systems_ 33 (2020), 596–608. * Su et al. (2021) Guangxin Su, Weitong Chen, and Miao Xu. 2021. Positive-unlabeled learning from imbalanced data. In _Proceedings of the 30th International Joint Conference on Artificial Intelligence, Virtual Event_. * Tanaka et al. (2018) Daiki Tanaka, Daiki Ikami, Toshihiko Yamasaki, and Kiyoharu Aizawa. 2018. Joint optimization framework for learning with noisy labels. In _Proceedings of the IEEE conference on computer vision and pattern recognition_. 5552–5560. * Vahdat (2017) Arash Vahdat. 2017\. Toward robustness against label noise in training deep discriminative neural networks. _Advances in Neural Information Processing Systems_ 30 (2017). * Verma et al. (2019) Vikas Verma, Kenji Kawaguchi, Alex Lamb, Juho Kannala, Yoshua Bengio, and David Lopez-Paz. 2019\. Interpolation consistency training for semi-supervised learning. _arXiv preprint arXiv:1903.03825_ (2019). * Wang et al. (2022) Yikai Wang, Xinwei Sun, and Yanwei Fu. 2022. Scalable Penalized Regression for Noise Detection in Learning with Noisy Labels. _arXiv preprint arXiv:2203.07788_ (2022). * Xu et al. (2019) Miao Xu, Bingcong Li, Gang Niu, Bo Han, and Masashi Sugiyama. 2019. Revisiting sample selection approach to positive-unlabeled learning: Turning unlabeled data into positive rather than negative. _arXiv preprint arXiv:1901.10155_ (2019). * Yu et al. (2002) Hwanjo Yu, Jiawei Han, and Kevin Chen-Chuan Chang. 2002\. PEBL: positive example based learning for web page classification using SVM. In _Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining_. 239–248. * Yu et al. (2019) Xingrui Yu, Bo Han, Jiangchao Yao, Gang Niu, Ivor Tsang, and Masashi Sugiyama. 2019\. How does disagreement help generalization against label corruption?. In _International Conference on Machine Learning_. PMLR, 7164–7173. * Zhang and Sabuncu (2018) Zhilu Zhang and Mert Sabuncu. 2018. Generalized cross entropy loss for training deep neural networks with noisy labels. _Advances in neural information processing systems_ 31 (2018). * Zhou et al. (2020) Tianyi Zhou, Shengjie Wang, and Jeff Bilmes. 2020. Robust curriculum learning: from clean label detection to noisy label self-correction. In _International Conference on Learning Representations_. ## Appendix A Experimental Detail ### A.1. Detailed Information for Liver dataset Each sample in the Liver dataset denotes a patient with liver cancer. The data consists of several modalities of MRI images from all patient, including T1WI, T1CE1, T1IP and T2WI. We simply adopt T1CE1 as the input modality of our experiment and leave the full usage of all modalities as future works. the model reads all of the slices of each sample with Pydicom and resize them to $256\times 256$ without additional normalization. The labels are transformed from the survival time of each patient, which ranges from 0 to 120. Figure 5. Detailed architecture of our network for Liver dataset. N, C, H, W denotes number of slices, number of channels, image height and image width respectively. ### A.2. Detailed Architecture for Liver dataset As we have mentioned in our main context, since each liver cancer sample is composed of several slices, we adopt a special architecture to deal with such data. Essentially, as shown in Fig. 5, the basic ResNet18 is used to extract features for all slices. The network weights are shared among different slices. Then for slices from one sample, a multi-head attention module is used to aggregate these features. Then a pooling layer and a classifier are used, as the same in original ResNet18. ## Appendix B Additional Experiment Results Figure 6. Training accuracy of clean and noisy data respectively in easy set. Figure 7. Gradient visualization of both clean and noisy samples on CIFAR10. ### B.1. Results on noise detection methods To support our claim in the main context that noise detection methods may not be proper for further detecting noisy samples in $\mathcal{T}^{easy}$, we first plot the training accuracy of noisy and clean samples in the easy set. As depicted in Fig. 6, while the training accuracy of clean samples is higher than that of noisy ones for each epoch, the actual size of wrongly-predicted clean samples is also much higher than noisy set. The proportion of wrongly- predicted noisy samples among all samples that are wrongly predicted is much smaller than that in the original unlabeled data as. This means these easy noisy samples do not enjoy the slow-learning property to the same extend as the hard noisy ones. Moreover, the result of CGH (Chatterjee, 2020) is shown in Fig. 7, in which we visualize the gradient of the first weight in the first convolutional layer of the whole network with regard to both easy noisy samples and easy clean samples. We can find that both the scale and distribution are similar between these two kinds of data, which means the coherent gradient hypothesis also does not hold so well on the easy data. $\rho$ | Acc. ---|--- 500 | 1000 | 3000 0.3 | 88.26 | 89.47 | 91.75 0.4 | 88.68 | 89.51 | 92.03 0.5 | 88.55 | 90.02 | 92.06 0.6 | 88.92 | 90.24 | 92.03 0.7 | 89.28 | 90.51 | 92.51 0.8 | 89.18 | 90.31 | 92.13 0.9 | 89.16 | 90.23 | 92.17 Table 8. Test accuracies on CIFAR10 of our model trained with different $\rho$ in Eq. (6) in the main context. ### B.2. Choices of $\rho$ We compare different $\rho$ in Eq. (6) in the main context, with the results in Tab. 8. As mentioned in the original paper (Englesson and Azizpour, 2021), smaller $\rho$ leads to faster convergence and less robustness, which is roughly consistent with our results. When $\rho=0.3$, the model gets 88.26%, 89.47% and 91.75%, which is the worst among all $\rho$ and almost similar to the results when using soft cross entropy. As $\rho$ gets larger, the performance gets better and saturates at $\rho=0.7$. Model | Acc. ---|--- 500 | 1000 | 3000 uPU | 85.98$\pm$1.09 | 88.22$\pm$0.28 | 89.83$\pm$0.42 uPU+Split-PU | 87.44$\pm$0.18 | 90.57$\pm$0.26 | 91.51$\pm$0.13 Table 9. Test accuracies on CIFAR10 of models trained with and without early stop splitting. ### B.3. Choices of $\alpha$ and $\beta$ We conduct the comparison on $\alpha$ and $\beta$ with the same setting as in the main paper, i.e. 500/1000/3000 positive labeled samples on CIFAR10. As shown in Tab. 10 and Tab. 11, the performance of our model among different hyper-parameter settings is relatively robust. Meanwhile, the trend of performance is generally consistent with few exceptions. For example, $\alpha=0.5$ or $\beta=0.2$ lead to the highest accuracy when using 3000 positive labeled samples. We advocate that these two hyper-parameters control the balance among different loss terms adopted by our method. In particular, larger $\alpha$ and $\beta$ lead to the scale of $L_{pred}$ and $L_{feat}$, as depicted in Eq. (11) and Eq. (12), larger than that of $L_{easy}$. The strength of annotations is thus weakened in this way. Consequently, the model becomes poorer at learning label information from the dataset, which leads to worse performance. On the other hand, smaller $\alpha$ makes the effect of prediction consistency negligible, hence resulting in lower accuracy. We will add these results to our paper. $\alpha$ | Acc. ---|--- 500 | 1000 | 3000 0.1 | 87.68 | 89.87 | 90.84 0.2 | 88.05 | 90.08 | 92.05 0.3 | 89.18 | 90.51 | 92.51 0.4 | 88.61 | 90.03 | 92.27 0.5 | 88.52 | 90.08 | 92.55 0.6 | 88.48 | 89.93 | 92.41 0.7 | 88.60 | 90.09 | 92.13 0.8 | 88.68 | 89.56 | 91.17 0.9 | 88.30 | 89.30 | 90.53 Table 10. Test accuracies on CIFAR10 of our model trained with different $\alpha$ in Eq. (13) in the main context. $\beta$ | Acc. ---|--- 500 | 1000 | 3000 0.1 | 89.18 | 90.51 | 92.51 0.2 | 88.96 | 90.31 | 92.63 0.3 | 89.02 | 90.39 | 92.37 0.4 | 88.67 | 90.11 | 92.43 0.5 | 88.84 | 89.87 | 92.14 0.6 | 88.81 | 90.09 | 92.37 0.7 | 89.07 | 90.08 | 91.71 0.8 | 88.48 | 89.14 | 91.24 0.9 | 88.26 | 89.49 | 91.54 Table 11. Test accuracies on CIFAR10 of our model trained with different $\beta$ in Eq. (13) in the main context. ### B.4. Application of Split-PU on other base models To further verify the efficacy of our proposed method, we use another popular objective function uPU (Du Plessis et al., 2014) as our base model, i.e. replacing $L_{nnPU}$ with the following $L_{uPU}$ when training $\phi_{base}$: (14) $\displaystyle L_{uPU}$ $\displaystyle=\frac{\pi}{n_{p}}\sum_{i=1}^{n_{p}}L(\phi(I_{i}^{p}),1)+$ (15) $\displaystyle\frac{1}{n_{u}}\sum_{i=1}^{n_{u}}L(\phi(I_{i}^{u}),-1)-\frac{\pi}{n_{u}}\sum_{i=1}^{n_{u}}L(\phi(I_{i}^{p}),-1)$ The results on CIFAR10 are shown in Tab. 9, which indicates that the improvement of our Split-PU on both nnPU and uPU is consistent. It is noticeable that when using 1000 positive labeled samples, using uPU with Split-PU is even better than using nnPU with Split-PU, receiving accuracy of 90.57% with comparable standard deviation. Figure 8. Visualization of the easy and hard sets.
# Taming Hyperparameter Tuning in Continuous Normalizing Flows Using the JKO Scheme Alexander Vidal Dept. of Applied Mathematics and Statistics Colorado School of Mines &Samy Wu Fung Dept. of Applied Mathematics and Statistics Dept. of Computer Science Colorado School of Mines &Luis Tenorio Dept. of Applied Mathematics and Statistics Colorado School of Mines &Stanley Osher Dept. of Mathematics University of California, Los Angeles &Levon Nurbekyan Dept. of Mathematics University of California, Los Angeles <EMAIL_ADDRESS> ###### Abstract A normalizing flow (NF) is a mapping that transforms a chosen probability distribution to a normal distribution. Such flows are a common technique used for data generation and density estimation in machine learning and data science. The density estimate obtained with a NF requires a change of variables formula that involves the computation of the Jacobian determinant of the NF transformation. In order to tractably compute this determinant, continuous normalizing flows (CNF) estimate the mapping and its Jacobian determinant using a neural ODE. Optimal transport (OT) theory has been successfully used to assist in finding CNFs by formulating them as OT problems with a soft penalty for enforcing the standard normal distribution as a target measure. A drawback of OT-based CNFs is the addition of a hyperparameter, $\alpha$, that controls the strength of the soft penalty and requires significant tuning. We present JKO-Flow, an algorithm to solve OT-based CNF without the need of tuning $\alpha$. This is achieved by integrating the OT CNF framework into a Wasserstein gradient flow framework, also known as the JKO scheme. Instead of tuning $\alpha$, we repeatedly solve the optimization problem for a fixed $\alpha$ effectively performing a JKO update with a time- step $\alpha$. Hence we obtain a "divide and conquer" algorithm by repeatedly solving simpler problems instead of solving a potentially harder problem with large $\alpha$. ## 1 Introduction A normalizing flow (NF) is a type of generative modeling technique that has shown great promise in applications arising in physics noe2019boltzmann ; brehmer2020madminer ; carleo2019machine as a general framework to construct probability densities for continuous random variables in high-dimensional spaces papamakarios2019normalizing ; kobyzev2019normalizing ; rezende2015 . An NF provides a $\mathcal{C}^{1}$-diffeomorphism $f$ (i.e., a normalizing transformation) that transforms the density $\rho_{0}$ of an initial distribution $P_{0}$ to the density $\rho_{1}$ of the standard multivariate normal distribution $P_{1}$ – hence the term "normalizing." Given such mapping $f$, the density $\rho_{0}$ can be recovered from the Gaussian density via the change of variables formula, $\log\rho_{0}(x)=\log\rho_{1}\left(f(x)\right)+\log|\det J_{f}(x)|,$ (1) where $J_{f}\in\mathbb{R}^{d\times d}$ is the Jacobian of $f$. Moreover, one can obtain samples with density $\rho_{0}$ by pushing forward Gaussian samples via $f^{-1}$. ###### Remark 1. Throughout the paper we slightly abuse the notation, using the same notation for probability distributions and their density functions. Additionally, given a probability distribution $P_{0}$ on $\mathbb{R}^{d}$ and a measurable mapping $f:\mathbb{R}^{d}\to\mathbb{R}^{d}$, we define the pushforward distribution of $P_{0}$ through $f$ as $(f\sharp P_{0})(B)=P_{0}(f^{-1}(B))$ for all Borel measurable $B\subseteq\mathbb{R}^{d}$ villani2003topics . There are two classes of normalizing flows: finite and continuous. A finite flow is defined as a composition of a finite number of $\mathcal{C}^{1}$-diffeomorphisms: $f=f_{1}\circ f_{2}\circ\cdots\circ f_{n}$. To make finite flows computationally tractable, each $f_{i}$ is chosen to have some regularity properties such as a Jacobian with a tractable determinant; for example, $J_{f_{i}}$ may have a triangular structure papamakarios2017masked ; baptista2021learning ; zech2022sparse . On the other hand, continuous normalizing flows (CNFs) estimate $f$ using a neural ODE of the form chen2017continuous : $\partial_{t}z(x,t)=v_{\theta}(z(x,t),t),\qquad z(x,0)=x,\qquad 0\leq t\leq T,$ (2) where $\theta$ are the parameters of the neural ODE. In this case, $f$ is defined as $f(x)=z(x,T)$ (for simplicity, we remove the dependence of $z$ on $\theta$). One of the main advantages of CNFs is that we can tractably estimate the log-determinant of the Jacobian using Jacobi’s identity, which is commonly used in fluid mechanics (see, e.g., (villani2003topics, , p.114)): $\begin{split}\partial_{t}\log|\det\nabla_{x}z(x,t)|&=\nabla_{z}\cdot v_{\theta}(z(x,t),t)=\operatorname{trace}\left(\nabla_{z}v_{\theta}(z(x,t),t)\right).\end{split}$ (3) This is computationally appealing as one can replace the expensive determinant calculation by a more tractable trace computation of $\nabla_{z}v_{\theta}(z(x,t),t)$. Importantly, no restrictions on $\nabla_{z}v_{\theta}(z(x,t),t)$ (e.g., diagonal or triangular structure) are needed; thus, these Jacobians are also referred to as “free-form Jacobians” grathwohl2019ffjord . The goal in training a CNF is to find parameters, $\theta$, such that $f=z(\,\cdot,T)$ leads to a good approximation of $\rho_{1}$ or, assuming $f$ is invertible, the pushforward of $\rho_{1}$ through $f^{-1}$ is a good approximation of $\rho_{0}$ rezende2015 ; papamakarios2017masked ; papamakarios2019normalizing ; grathwohl2019ffjord . Indeed, let $\widehat{\rho}_{0}$ be this pushforward density obtained with a CNF $f$; that is, $\widehat{\rho}_{0}=f^{-1}\sharp\rho_{1}$. We then minimize the Kullback- Leibler (KL) divergence from $\widehat{\rho}_{0}$ to $\rho_{0}$ given by $\min_{\theta}~{}\mathbb{E}_{x\sim\rho_{0}}\log(\rho_{0}(x)/\widehat{\rho}_{0}(x))=\min_{\theta}~{}\mathbb{E}_{x\sim\rho_{0}}\left[\log\rho_{0}(x)-\log\rho_{1}(z(x,T))-\ell(x,T)\right],$ where $\ell(x,T)=\log|\det\nabla z(x,T)|$. Dropping the $\theta$-independent term $\log\rho_{0}$ and using (2) and (3), this previous optimization problem reduces to the minimization problem $\min_{\theta}\;\;\mathbb{E}_{x\sim\rho_{0}}\;C(x,T),\quad C(x,T):=-\log\rho_{1}(z(x,T))-\ell(x,T)$ (4) subject to ODE constraints $\partial_{t}\begin{bmatrix}z(x,t)\\\ \ell(x,t)\end{bmatrix}=\begin{bmatrix}v_{\theta}(z(x,t),t)\\\ \operatorname{trace}\left(\nabla_{z}v_{\theta}(z(x,t),t)\right)\end{bmatrix},\qquad\begin{bmatrix}z(x,0)\\\ \ell(x,0)\end{bmatrix}=\begin{bmatrix}x\\\ 0\end{bmatrix}.$ (5) The ODE (5) might be stiff for certain values of $\theta$, leading to extremely long computation times. Indeed, the dependence of $v$ on $\theta$ is highly nonlinear and might generate vector fields that lead to highly oscillatory trajectories with complex geometry. Some recent work leverages optimal transport theory to find the CNF finlay2020train ; onken2021ot . In particular, a kinetic energy regularization term (among others) is added to the loss to “encourage straight trajectories” $z(x,t)$. That is, the flow is trained by minimizing the following instead of (4): $\begin{split}\min_{\theta}\;\;\mathbb{E}_{x\sim\rho_{0}}\;\int_{0}^{T}\frac{1}{2}\|v_{\theta}(z(x,t),t)\|^{2}dt+\alpha C(x,T)\end{split}$ (6) subject to (5). The key insight in finlay2020train ; onken2021ot is that (4) is an example of a degenerate OT problem with a soft terminal penalty and without a transportation cost. The first objective term in (6) given by the time integral is the transportation cost term, whereas $\alpha$ is a hyperparameter that balances the soft penalty and the transportation cost. Including this cost makes the problem well-posed by forcing the solution to be unique villani2008optimal . Additionally, it enforces straight trajectories so that (5) is not stiff. Indeed finlay2020train ; onken2021ot empirically demonstrate that including optimal transport theory leads to faster and more stable training of CNFs. Intuitively, we minimize the KL divergence _and_ the arclength of the trajectories. Although including optimal transport theory into CNFs has been very successful onken2021ot ; finlay2020train ; yang2019 ; zhang2018monge , there are two key challenges that render them difficult to train. First, estimating the log- determinant in (4) via the trace in (5) is still computationally taxing and commonly used methods rely on stochastic approximations grathwohl2019ffjord ; finlay2020train , which adds extra error. Second, including the kinetic energy regularization requires tuning of the hyperparameter $\alpha$. Indeed, if $\alpha$ is chosen too small in (6), then the kinetic regularization term dominates the training process, and the optimal solution consists of not moving, i.e., $f(x)=x$. On the other hand, if $\alpha$ is chosen too large, we return to the original setting where the problem is ill-posed, i.e., there are infinitely many solutions. Finally, finding an "optimal" $\alpha$ is problem dependent and requires tuning on a cases-by-case basis. ### 1.1 Our Contribution We present JKO-Flow, an optimal transport-based algorithm for training CNFs without the need to tune the hyperparameter $\alpha$ in (6). Our approach also leverages fast numerical methods for exact trace estimation from the recently developed optimal transport flow (OT-Flow) onken2021ot ; ruthotto2020machine . The key idea is to integrate the OT-Flow approach into a Wasserstein gradient flow framework, also known as the Jordan, Kinderlehrer, and Otto (JKO) scheme jordan1998variational . Rather than tuning the hyperparameter $\alpha$ (commonly done using a grid search), the idea is to simply pick any $\alpha$ and solve a sequence of "easier" OT problems that gradually approach the target distribution. Each solve is precisely a gradient descent in the space of distributions, a Wasserstein gradient descent, and the scheme provably converges to the desired distribution for all $\alpha>0$ salim20Wasserstein . Our experiments show that our proposed approach is effective in generating higher quality samples (and density estimates) and also allows us to reduce the number of parameters required to estimate the desired flow. Our strategy is reminiscent of debiasing techniques commonly used in inverse problems. Indeed, the transportation cost that serves as a regularizer in (6) introduces a bias – the smaller $\alpha$ the more bias is introduced (see, e.g., tenIP ), so good choices of $\alpha$ tend to be larger. One way of removing the bias and the necessity of tuning the regularization strength is to perform a sequence of Bregman iterations osher05iterative ; burger05nonlinear also known as nonlinear proximal steps. Hence our approach reduces to debiasing via Bregman or proximal steps in the Wasserstein space. In the context of CNF training, Bregman iterations are advantageous due to the flexibility of the choice for $\alpha$. Indeed, the resulting loss function is non-convex and its optimization tends to get harder for large $\alpha$. Thus, instead of solving one harder problem we solve several “easier” problems. ## 2 Optimal Transport Background and Connections to CNFs Denote by $\mathcal{P}_{2}(\mathbb{R}^{d})$ the space of Borel probability measures on $\mathbb{R}^{d}$ with finite second-order moments, and let $\rho_{0},\rho_{1}\in\mathcal{P}_{2}(\mathbb{R}^{d})$. The quadratic optimal transportation (OT) problem (which also defines the Wasserstain metric $W_{2}$) is then formulated as $W_{2}^{2}(\rho_{0},\rho_{1})=\inf_{\pi\in\Gamma(\rho_{0},\rho_{1})}\int_{\mathbb{R}^{2d}}\|x-y\|^{2}d\pi(x,y),$ (7) where $\Gamma(\rho_{0},\rho_{1})$ is the set of probability measures $\pi\in\mathcal{P}(\mathbb{R}^{2d})$ with fixed $x$ and $y$-marginals densities $\rho_{0}$ and $\rho_{1}$, respectively. Hence the cost of transporting a unit mass from $x$ to $y$ is $\|x-y\|^{2}$, and one attempts to transport $\rho_{0}$ to $\rho_{1}$ as cheaply as possible. In (7), $\pi$ represents a transportation plan, and $\pi(x,y)$ is the mass being transported from $x$ to $y$. One can prove that $(\mathcal{P}_{2}(\mathbb{R}^{d}),W_{2})$ is a complete separable metric space villani2003topics . OT has recently become a very active research area in PDE, geometry, functional inequalities, economics, data science and elsewhere partly due to equipping the space of probability measures with a (Riemannian) metric villani2003topics ; villani2008optimal ; peyre2018computational ; santambrogio2015optimal . As observed in prior works, there are many similarities between OT and NFs onken2021ot ; finlay2020train ; yang2020potential ; zhang2018monge . This connection becomes more transparent when considering the dynamic formulation of (7). More precisely, the Benamou-Brenier formulation of the OT problem is given by benamou2000computational : $\begin{split}\frac{T}{2}W_{2}^{2}(\rho_{0},\rho_{1})=\inf_{v,\rho}\;\;&\int_{0}^{T}\int_{\mathbb{R}^{d}}\frac{1}{2}\|v(x,t)\|_{2}^{2}\rho(x,t)dxdt\\\ \mbox{ s.t. }\;\;&\partial_{t}\rho(x,t)+\nabla\cdot(\rho(x,t)v(x,t))=0\\\ &\rho(x,0)=\rho_{0}(x),\;\;\rho(x,T)=\rho_{1}(x).\end{split}$ (8) Hence, the OT problem can be formulated as a problem of flowing $\rho_{0}$ to $\rho_{1}$ with a velocity field $v$ that achieves minimal kinetic energy. The optimal velocity field $v$ has several appealing properties. First, particles induced by the optimal flow $v$ travel in straight lines. Second, particles travel with constant speed. Moreover, under suitable conditions on $\rho_{0}$ and $\rho_{1}$, the optimal velocity field is unique villani2003topics . Given a velocity field $v$, denote by $z(x,t)$ the solution of the ODE $\partial_{t}z(x,t)=v(z(x,t),t),\qquad z(x,0)=x,\qquad 0\leq t\leq T.$ Then, under suitable regularity conditions, we have that the solution of the continuity equation is given by $\rho(\cdot,t)=z(\cdot,t)\sharp\rho_{0}$. Thus the optimization problem in (8) can be written as $\begin{split}\inf_{v}\;\;&\int_{0}^{T}\int_{\mathbb{R}^{d}}\frac{1}{2}\|v(z(x,t),t)\|_{2}^{2}\rho_{0}(x)dxdt\\\ \mbox{ s.t. }\;\;&\partial_{t}z(x,t)=v(z(x,t),t),~{}z(x,0)=x,~{}z(\cdot,T)\sharp\rho_{0}=\rho_{1}.\end{split}$ (9) This previous problem is very similar to (4) with the following differences: * • the objective function in (4) does not have the kinetic energy of trajectories, * • the terminal constraint is imposed as a soft constraint in (4) and as a hard constraint in (9), and * • $v$ in (4) is $\theta$-dependent, whereas the formulation in (9) is in the non-parametric regime. So the NF (4) can be thought of as an approximation to a degenerate transportation problem that lacks transportation cost. Based on this insight one can regularize (4) by adding the transportation cost and arrive at (6) or some closely related version of it onken2021ot ; finlay2020train ; yang2020potential ; zhang2018monge . It has been observed that the transportation cost (kinetic energy) regularization significantly improves the training of NFs. ## 3 JKO-Flow: Wasserstein Gradient Flows for CNFs While the OT-based formulation of CNFs in (6) has been found successful in some applications onken2021ot ; finlay2020train ; yang2020potential ; zhang2018monge , a key difficulty arises in choosing how to balance the kinetic energy term and the KL-divergence, i.e., on selecting $\alpha$. This difficulty is typical in problems where the constraints are imposed in a soft fashion. Standard training of CNFs typically involves tuning for a “large but hopefully stable enough” step size $\alpha$ so that the KL divergence term is sufficiently small after training. To this end, we propose an approach that avoids the need to tune $\alpha$ by using the fact that the solution to (6) is an approximation to a backward Euler (or proximal point) algorithm when discretizing the Wasserstein gradient flow using the Jordan-Kinderlehrer-Otto (JKO) scheme jordan1998variational . The seminal work in jordan1998variational provides a gradient flow structure of the Fokker-Planck equation using an implicit time discretization. That is, given $\alpha>0$, density at $k^{\text{th}}$ iteration, $\rho^{(k)}$, and terminal density $\rho_{1}$, one finds $\begin{split}\rho^{(k+1)}=&\operatorname*{arg\,min}_{\rho\in\mathcal{P}_{2}(\mathbb{R}^{d})}\;\frac{1}{2\alpha}W_{2}^{2}(\rho,\rho^{(k)})+KL(\rho||\rho_{1})\\\ =&\operatorname*{arg\,min}_{v}\;\frac{1}{\alpha}\int_{0}^{1}\int_{\mathbb{R}^{d}}\frac{1}{2}\|v(z(x,t),t)\|_{2}^{2}\rho_{0}(x)dxdt+KL(z(\cdot,1)\sharp\rho^{(k)}||\rho_{1})\\\ \mbox{ s.t. }\;\;&\partial_{t}z(x,t)=v(z(x,t),t),~{}z(x,0)=x,~{}z(\cdot,T)\sharp\rho_{0}=\rho_{1},\end{split}$ (10) for $k=0,1,\ldots$, and $\rho^{(0)}=\rho_{0}$. Here, $\alpha$ takes the role of a step size when applying a proximal point method to the KL divergence using the Wasserstein-2 metric, and $\\{\rho^{(k)}\\}$ provably converges to $\rho_{1}$ jordan1998variational ; salim20Wasserstein . Hence, repeatedly solving (9) with the KL penalty acting as a soft constraint yields an arbitrarily accurate approximation of $\rho_{1}$. In the parametric regime each iteration takes the form $\operatorname*{arg\,min}_{\theta}\;\;\mathbb{E}_{x\sim\rho^{(k)}}\;\int_{0}^{T}\frac{1}{2}\|v_{\theta}(x,t)\|^{2}dt+\alpha C(x,T)\quad\text{subject to}\quad\eqref{eq: min_likelihood_constraints}.$ (11) Thus we solve a sequence of problems (6), where the initial density of the current subproblem is given by the pushforward of the density generated in the previous subproblem. Algorithm 1 Proposed Algorithm 1: Input: Samples from $\rho_{0}$, step size $\alpha>0$, number of steps $K$ 2: Initialize $\theta_{1}$ at random 3: for $k=1,\ldots,K$ do 4: Solve for $\theta_{k}$ using samples by solving (11) 5: Update distribution of samples using $v_{\theta_{k}}$ 6: end for 7: Output: saved weights $\theta_{1},\ldots,\theta_{K}$ Importantly, our proposed approach _does not require tuning $\alpha$_. Instead, we solve a sequence of subproblems that is guaranteed to converge to $\rho_{1}$ jordan1998variational prior to the neural network parameterization; see Alg. 1. While our proposed methodology can be used in tandem with any algorithm used to solve (11), an important numerical aspect in our approach is to leverage fast computational methods that use _exact_ trace estimation in (5); this approach is called OT-Flow onken2021ot . Consequently, we avoid the use of stochastic approximation methods for the trace, e.g., Hutchinson’s estimator avron2011 ; hutchinson1990stochastic ; tenIP , as is typically done in CNF methods grathwohl2019ffjord ; finlay2020train . A surprising result of our proposed method is that it empirically shows improved performance even with fewer number of parameters (see Fig. 3). ## 4 Related Works Density Estimation. One of the main advantages of NFs over other generative models is that they provide density estimates of probability distributions using (1). That is, we do not need to apply a separate density estimation technique after generating samples from a distribution, e.g., as in GANs goodfellow2020generative . Multivariate density estimation is a fundamental problem in statistics silverman1986density ; scott2015multivariate , High Energy Physics (HEP) cranmer2001kernel and in other fields of science dealing with multivariate data. For instance, particle physicists in HEP study possible distributions from a set of high energy data. Another application of density estimation is in confidence level calculations of particles in Higgs searches at Large Electron Positron Colliders (LEP) opal2000search and discriminant methods used in the search for new particles cranmer2001kernel . Finite Flows. Finite normalizing flows tabak2013family ; rezende2015 ; papamakarios2019normalizing ; kobyzev2019normalizing use a composition of discrete transformations, where specific architectures are chosen to allow for efficient inverse and Jacobian determinant computations. NICE dinh2014nice , RealNVP dinh2016density , IAF kingma2016improved , and MAF papamakarios2017masked use either autoregressive or coupling flows where the Jacobian is triangular, so the Jacobian determinant can be tractably computed. GLOW kingma2018glow expands upon RealNVP by introducing an additional invertible convolution step. These flows are based on either coupling layers or autoregressive transformations, whose tractable invertibility allows for density evaluation and generative sampling. Neural Spline Flows durkan2019neural use splines instead of the coupling layers used in GLOW and RealNVP. Using monotonic neural networks, NAF huang2018neural require positivity of the weights, which UMNN wehenkel2019unconstrained circumvent this requirement by parameterizing the Jacobian and then integrating numerically. Continuous and Optimal Transport-based Flows. Modeling flows with differential equations is a natural and commonly used method suykens1998 ; welling2011bayesian ; neal2011mcmc ; salimans2015markov ; ruthotto2021introduction ; huang2020convex . In particular, CNFs model their flow via a neural ordinary differential equation chen2017continuous ; chen2018neural ; grathwohl2019ffjord . Among the most well-known CNFs are FFJORD grathwohl2019ffjord , which estimates the determinant of the Jacobian by accumulating its trace along the trajectories, and the trace is estimated using Hutchinson’s estimator avron2011 ; hutchinson1990stochastic ; tenIP . To promote straight trajectories, RNODE finlay2020train regularizes FFJORD with a transport cost $L(\boldsymbol{x},T)$. RNODE also includes the Frobenius norm of the Jacobian $\|\nabla\mathbf{v}\|_{F}^{2}$ to stabilize training. The trace and the Frobenius norm are estimated using a stochastic estimator and report speedup by a factor of 2.8. Monge-Ampère Flows zhang2018monge and Potential Flow Generators yang2019 similarly draw from OT theory but parameterize a potential function instead of the dynamics directly. OT is also used in other generative models sanjabi2018 ; salimans2018improving ; lei2018geometric ; lin2019fluid ; avraham2019parallel ; tanaka2019discriminator . OT-Flow onken2021ot is based on a discretize-then-optimize approach onken2020do that also parameterizes the potential function. To evaluate the KL divergence, OT-Flow estimates the density using an _exact_ trace computation following the work of ruthotto2020machine . Wasserstein Gradient Flows. Our proposed method is most closely related to fan2021variational , which also employs a JKO-based scheme to perform generative modeling. But a key difference is that fan2021variational reformulates the KL-divergence as an optimization over difference of expectations (See (fan2021variational, , Prop. 3.1)); this makes their approach akin to GANs, where the density cannot be obtained without using a separate density estimation technique. Our proposed method is also closely related to methods that use input-convex CNNs mokrov2021large ; bunne2022proximal ; alvarez2021optimizing . mokrov2021large focuses on the special case with KL divergence as objective function. alvarez2021optimizing solve a sequence of subproblems different from the fluid flow formulation presented in (11). They also require an end-to-end training scheme that backpropagates to the initial distribution; this can become a computational burden when the number of time discretizations is large. bunne2022proximal utilizes a JKO-based scheme to approximate a population dynamics given an observed trajectory and focus on applications in computational biology. Other related works include natural gradient methods nurbekyan2022efficient and implicit schemes based on the Wasserstein-1 distance heaton2020wasserstein . ## 5 Numerical Experiments We demonstrate the effectiveness of our proposed JKO-Flow on a series of synthetic and real-world datasets. As previously mentioned, we compute each update in (10) by solving (6) using the OT-Flow solver onken2021ot , which leverages fast and exact trace computations. We also use the same architecture provided in onken2021ot . Henceforth, we shall also call the traditional CNF approach the “single-shot” approach. $\alpha=1$ | $\alpha=5$ | $\alpha=10$ | $\alpha=50$ ---|---|---|--- MMD2 Values, Single Shot 3.58e-2 | 3.56e-3 | 1.42e-3 | 1.26e-3 | | | MMD2 Values, JKO-Flow (five iterations) 4.90e-4 | 5.70e-4 | 6.40e-4 | 9.00e-4 | | | Figure 1: Checkerboard dataset: Generated samples of $\widehat{\rho}_{0}$ using the standard one-shot approach (top row). Generated using our proposed JKO-Flow using five iterations (bottom row). Here, we use $\alpha=$ 1, 5, 10, 50. JKO-Flow returns consistent results _regardless of the value of $\alpha$_. Data $x$ | | | | | | | ---|---|---|---|---|---|---|--- Estimate $\rho_{0}$ | | | | | | | Generation $f^{-1}(y)$ | | | | | | | Figure 2: Density estimation on 2D toy problems using five JKO-Flow iterations. Top: samples from the unknown distribution $\rho_{0}$. Middle: density estimate for $\rho_{0}$ computed by inverting the flow through the five iterations of JKO-Flow from $\rho_{1}$ via (2). Bottom: samples generated by inverse JKO-Flow through five iterations where $y$ has density $\rho_{1}$. Maximum Mean Discrepancy Metric (MMD). Our density estimation problem requires approximating a density $\rho_{0}$ by finding a transformation $f$ such that $f^{-1}\sharp\rho_{1}$ has density $\widehat{\rho}_{0}$ close to $\rho_{0}$ where $\rho_{1}$ is the standard Gaussian. However, $\rho_{0}$ is not known in real-world density estimation scenarios, such as in physics applications, all we have are samples $X=\\{x_{i}\\}_{i=1}^{n}$ from the unknown distribution. Consequently, we use the observed samples $X$ and samples $\widehat{X}=\\{\widehat{x}_{j}\\}_{j=1}^{m}$, $\widehat{x}_{j}=f^{-1}(q_{j})$, generated by the CNF and samples $Q=\\{q_{j}\\}_{j=1}^{m}$ from $\rho_{1}$ to determine if their corresponding distributions are close in some sense. To measure the discrepancy we use a particular integral probability metric zolotarev1976metric ; rachev2013methods ; muller1997integral known as maximum mean discrepancy (MMD) defined as follows gretton2012mmd : Let $x$ and $y$ be random vectors in $\mathbb{R}^{d}$ with distributions $\mu_{x}$ and $\mu_{y}$, respectively, and let $\mathcal{H}$ be a reproducing kernel Hilbert space of functions on $\mathbb{R}^{d}$ with Gaussian kernel (see paulsen2016introduction for an introduction) $k(x_{i},x_{j})=\exp{\left(-\frac{1}{2}\|x_{i}-x_{j}\|^{2}\right)}.$ (12) Then the MMD of $\mu_{x}$ and $\mu_{y}$ is given by $\mathrm{MMD}_{\mathcal{H}}(\mu_{x},\mu_{y})=\sup_{\|f\|_{\mathcal{H}}\leq 1}\,|\,\mathbb{E}\,f(x)-\mathbb{E}\,f(y)\,|.$ It can be shown that $\mathrm{MMD}_{\mathcal{H}}$ defines a metric on the class of probability measures on $\mathbb{R}^{d}$ gretton2012mmd ; fukumizu2007kernel . The squared-MMD can be written in terms of the kernel as follows: $\mathrm{MMD}^{2}_{\mathcal{H}}(\mu_{x},\mu_{y})=\mathbb{E}\,k(x,x^{\prime})+\mathbb{E}\,k(y,y^{\prime})-2\,\mathbb{E}\,k(x,y),$ where $x,x^{\prime}$ are iid $\mu_{x}$ independent of $y,y^{\prime}$ which are iid $\mu_{y}$. An unbiased estimate of the squared-MMD based on the samples $X$ and $\widehat{X}$ defined above is given by gretton2012mmd : $\mathrm{MMD}^{2}_{\mathcal{H}}(X,\widehat{X})=\frac{1}{n(n-1)}\sum_{i\neq j}k(x_{i},x_{j})+\frac{1}{m(m-1)}\sum_{k\neq\ell}k(\widehat{x}_{k},\widehat{x}_{\ell})-\frac{2}{nm}\sum_{i,\ell}k(x_{i},\widehat{x}_{\ell}).$ Note that the MMD is not used for algorithmic training of the CNF, it is only used to compare the densities $\rho_{0}$ and $\widehat{\rho}_{0}$ based on the samples $X$ and $\widehat{X}$. Synthetic 2D Data Set. We begin by testing our method on seven two-dimensional (2D) benchmark datasets for density estimation algorithms commonly used in machine learning grathwohl2019ffjord ; wehenkel2019unconstrained ; see Fig. 2. We generate results with JKO-Flow for different values of $\alpha$ and for different number of iterations. We use $\alpha=1,\,5,\,10,\,$ and $50$, and for each $\alpha$ we use the single shot approach $k=1$ and JKO-Flow with $k=5$ iterations from (10). Note that in CNFs, we are interested in estimating the density (and generating samples) from $\rho_{0}$; consequently, once we have the optimal weights $\theta^{(1)},\theta^{(2)},\ldots,\theta^{(5)}$, we must “flow backwards” starting with samples from the normal distribution $\rho_{1}$. Fig. 1 shows that JKO-Flow outperforms the single shot approach for different values of $\alpha$. In particular, the performance for the single shot approach varies drastically for different values of $\alpha$, with $\alpha=1$ being an order of magnitude higher in MMD than $\alpha=5$. On the other hand, JKO-Flow is _consistent regardless of the value of $\alpha$_. As previously mentioned, this is expected as JKO-Flow is a proximal point algorithm that converges regardless of the step size $\alpha$. In this case, five JKO-Flow iterations are enough to obtain this consistency. Additional plots and hyperparameter setups for different benchmark datasets with similar performance results are shown in the Appendix. Tab. 1 summarizes the comparison between the single shot and JKO-Flow on all synthetic 2D datasets for different values of $\alpha$. We also show an illustration of all the datasets, estimated densities, and generated samples with JKO-Flow in Fig. 2. Varying Network Size. In addition to obtaining consistent results for different values of $\alpha$, we also _empirically_ observe that JKO-Flow outperforms the single shot approach for different numbers of network parameters, i.e., network size. We illustrate this in Fig. 3. This is also intuitive as we reformulate the problem of finding a single “difficult” optimal transportation problem as a sequence of “smaller and easier” OT problems. In this setup, we vary the width of a two-layer ResNet he2016deep . In particular, we choose the widths to be $m=3,4,5,8$, and $16$. These correspond to $40,53,68,125$, and $365$ parameters. The hyperparameter $\alpha$ is chosen to be the best performing value for each synthetic dataset. All datasets vary $m$ for fixed $\alpha=5$, except the 2 Spiral dataset, which uses $\alpha=50$; we chose these $\alpha$ values as they performed the best in the fixed $m$ experiments. Similar results are also shown for the remaining synthetic datasets in the appendix. Tab. 2 summarizes the comparison between the single shot and JKO-Flow on all synthetic 2D datasets. $m=3$ | $m=4$ | $m=5$ | $m=8$ | $m=16$ ---|---|---|---|--- MMD2 Values, Single Shot 1.1e-2 | 5.6e-3 | 2.46e-3 | 3.03e-3 | 2.7e-3 | | | | MMD2 Values, JKO-Flow (five iterations) 5.6e-3 | 1.07e-3 | 2.7e-4 | 2.32e-4 | 4.16e-4 | | | | Figure 3: Checkerboard dataset: Generated samples of $\widehat{\rho}_{0}$ using the standard single shot approach (top row). Generated samples using our proposed JKO-Flow using five iterations (bottom row). Here, we fix $\alpha=5$ and vary the network width $m=3,4,5,8$, and $16$. JKO-Flow performs competitively even with fewer parameters. Density Estimation on a Physics Dataset We train JKO-Flow on the 43-dimensional Miniboone dataset which is a high-dimensional, real-world physics dataset used as benchmark for high-dimensional density estimation algorithms in physics misc_miniboone_particle_identification_199 . For this physics problem, our method is trained for $\alpha=0.5,\,1,\,5,\,10,\,50$ and using $10$ JKO-Flow iterations. Fig 4 shows generated samples with JKO-Flow and the standard single-shot approach for $\alpha=5$. Since Miniboone is a high-dimensional dataset, we follow onken2021ot and plot two-dimensional slices. JKO-Flow generates better quality samples. Similar experiments for $\alpha=1,10$, and $50$ are shown in Figs 17, 18, and 19 in the Appendix. Tab. 3 summarizes the results for all values of $\alpha$. Note that we compute MMD values for all the dimensions as well as 2D slices; this is because we only have limited data ( 3000 testing samples) and the 2D slice MMD give a better indication on the improvement of the generated samples. Results show that _the MMD is consistent across all $\alpha$ values for JKO-Flow_. We also show the convergence (in MMD2) of the miniboone dataset across each 2D slice in Fig. 5. As expected, smaller step size $\alpha$ values converge slower (see $\alpha=0.5)$, but all converge to similar accuracy (unlike the single-shot). Samples | Single Shot | JKO-Flow ---|---|--- $x\sim\rho_{0}(x)$ | $f(x)$ | $f(x)$ | | $y\sim\rho_{1}(y)$ | $f^{-1}(y)$ | $f^{-1}(y)$ | | (a) Miniboone dimension 16 vs 17 Samples | Single Shot | JKO-Flow ---|---|--- $x\sim\rho_{0}(x)$ | $f(x)$ | $f(x)$ | | $y\sim\rho_{1}(y)$ | $f^{-1}(y)$ | $f^{-1}(y)$ | | (b) Miniboone dimension 28 vs 29 Figure 4: Generated samples for the 43-dimensional Miniboone dataset using the single shot approach and JKO-Flow with 10 iterations. To visualize the dataset, we show 2-dimensional slices. We show the forward flow $f(x)$ where $x\sim\rho_{0}$ and the genereated samples $f^{-1}(y)$ where $y\sim\rho_{1}$. Dimension 16 vs. 17 | Dimension 28 vs. 29 ---|--- | Figure 5: MMD2 per iteration when using JKO-Flow after training. MMD2 values vs. JKO-Flow iteration for 2-dimensional slice (dimensions 16-17 on the left and dimensions 28-29 on the right) of the Miniboone dataset. JKO-Flow achieves same accuracy _regardless_ of the value of $\alpha$. ## 6 Conclusion We propose a new approach we call JKO-Flow to train OT-regularized CNFs without having to tune the regularization parameter $\alpha$. The key idea is to embed an underlying OT-based CNF solver into a Wasserstein gradient flow framework, also known as the JKO scheme; this approach makes the regularization parameter act as a “time” variable. Thus, instead of tuning $\alpha$, we repeatedly solve proximal updates for a fixed (time variable) $\alpha$. In our setting, we choose OT-Flow onken2021ot , which leverages exact trace estimation for fast CNF training. Our numerical experiments show that JKO-Flow leads to improved performance over the traditional approach. Moreover, _JKO-Flow achieves similar results regardless of the choice of $\alpha$_. We also empirically observe improved performance when varying the size of the neural network. Future work will investigate JKO-Flow on similar problems such as deep learning-based methods for optimal control fleming06controlled ; onken2022neural ; onken2021neural and mean field games lin2021alternating ; ruthotto2020machine ; agrawal2022random . ## Acknowledgments LN and SO were partially funded by AFOSR MURI FA9550-18-502, ONR N00014-18-1-2527, N00014-18-20-1-2093 and N00014-20-1-2787. Table 1: Synthetic 2D Data: JKO-Flow performance for different values of $\alpha$. JKO-Flow returns consistent performance for different $\alpha$. $\alpha$ | | 1 | 5 | 10 | 50 ---|---|---|---|---|--- Dataset | Approach | MMD 2 Checkerboard | single Shot | 3.58e-2 | 3.56e-3 | 1.42e-3 | 1.26e-3 JKO-Flow (5 iters) | 4.9e-4 | 5.67e-4 | 6.40e-4 | 9.00e-4 2 Spirals | Single Shot | 7.21e-2 | 2.30e-2 | 1.84e-2 | 7.73e-4 JKO-Flow (5 iters) | 2.10e-2 | 4.62e-4 | 1.02e-4 | 5.37e-5 Swiss Roll | Single Shot | 4.74e-3 | 7.33e-4 | 2.86e-4 | 7.03e-4 JKO-Flow (5 iters) | 5.16e-4 | 8.3e-5 | 3.27e-5 | 6.07e-4 8 Gaussians | Single Shot | 9.18e-3 | 2.69e-4 | 3.94e-4 | 7.10e-4 JKO-Flow (5 iters) | 1.07e-4 | 4.13e-5 | 2.67e-4 | 7.27e-6 Circles | Single Shot | 9.84e-3 | 2.24e-4 | 6.51e-4 | 1.04e-4 JKO-Flow (5 iters) | 9.49e-4 | 9.97e-6 | 2.38e-5 | 9.28e-5 Pinwheel | Single Shot | 1.18e-2 | 1.8e-3 | 1.37e-3 | 2.2e-5 JKO-Flow (5 iters) | 4.7e-4 | 2.63e-4 | 3.84e-4 | 4.30e-4 Moons | Single Shot | 1.45e-3 | 2.05e-3 | 2.49e-4 | 2.42e-4 JKO-Flow (5 iters) | 1.92e-4 | 4.65e-5 | 4.3e-5 | 1.08e-4 Table 2: Synthetic 2D Data: Network width comparison for 1 and 5 iterations given a fixed, best performing $\alpha$. JKO-Flow performs better than the single shot approach for different network sizes. [2mm] $m$ 3 4 5 8 16 Dataset Approach MMD 2 Checkerboard Single Shot 1.10e-2 5.60e-3 2.46e-3 3.03e-3 2.70e-3 JKO-Flow (5 iters) 5.60e-3 1.07e-3 2.7e-4 2.32e-4 4.16e-4 2 Spirals Single Shot 5.98e-3 4.54e-3 5.47e-3 1.19e-3 3.96e-3 JKO-Flow (5 iters) 1.42e-3 1.49e-5 6.11e-4 3.93e-5 2.19e-3 Swiss Roll Single Shot 8.89e-3 7.71e-3 1.41e-3 1.37e-3 1.52e-3 JKO-Flow (5 iters) 1.49e-3 2.90e-4 6.13e-4 2.29e-4 8.40e-5 8 Gaussians Single Shot 2.20e-3 1.05e-3 1.04e-3 2.3e-4 5.05e-4 JKO-Flow (5 iters) 1.33e-4 9.85e-4 2.40e-5 3.96e-4 1.07e-4 Circles Single Shot 2.06e-3 1.72e-3 1.37e-3 1.69e-3 1.34e-3 JKO-Flow (5 iters) 1.94e-3 3.24e-4 7.71e-4 5.9e-5 1.01e-4 Pinwheel Single Shot 1.10e-2 4.03e-3 2.27e-3 3.80e-3 5.43e-4 JKO-Flow (5 iters) 1.20e-3 8.23e-4 1.60e-3 7.00e-5 2.69e-4 Moons Single Shot 4.98e-3 4.54e-3 5.47e-3 1.2e-3 3.96e-3 JKO- Flow (5 iters) 1.42e-3 1.5e-5 6.11e-4 3.90e-5 2.19e-3 Table 3: Miniboone: Comparison of Single Shot and JKO-Flow for different values of $\alpha$. [2mm] $\alpha$ 0.5 1 5 10 50 Dimensions Approach MMD 2 2d: 16 vs. 17 Single Shot 4.62e-3 4.50e-3 5.88e-3 4.89e-3 5.2e-3 JKO-Flow (10 iters) 3.42e-4 2.94e-4 1.27e-4 1.43e-4 1.71e-4 2d: 28 vs. 29 Single Shot 4.33e-2 4.60e-2 5.02e-2 4.97e-2 4.74e-2 JKO-Flow (10 iters) 8.02e-5 3.31e-5 4.43e-5 6.33e-5 9.97e-5 Full Single Shot 4.7e-2 4.51e-3 4.75e-3 4.21e-3 4.27e-3 JKO-Flow (10 iters) 4.72e-4 4.72e-4 4.71e-4 4.72e-4 4.72e-4 ## References * [1] S. Agrawal, W. Lee, S. W. Fung, and L. Nurbekyan. Random features for high-dimensional nonlocal mean-field games. Journal of Computational Physics, 459:111136, 2022. * [2] D. Alvarez-Melis, Y. Schiff, and Y. Mroueh. Optimizing functionals on the space of probabilities with input convex neural networks. arXiv preprint arXiv:2106.00774, 2021. * [3] G. Avraham, Y. Zuo, and T. Drummond. Parallel optimal transport GAN. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4406–4415, 2019. * [4] H. Avron and S. Toledo. Randomized algorithms for estimating the trace of an implicit symmetric positive semi-definite matrix. Journal of the ACM (JACM), 58(2):1–34, 2011. * [5] R. Baptista, Y. Marzouk, R. E. Morrison, and O. Zahm. Learning non-Gaussian graphical models via Hessian scores and triangular transport. arXiv preprint arXiv:2101.03093, 2021. * [6] J.-D. Benamou and Y. Brenier. A computational fluid mechanics solution to the Monge-Kantorovich mass transfer problem. Numerische Mathematik, 84(3):375–393, 2000. * [7] J. Brehmer, F. Kling, I. Espejo, and K. Cranmer. Madminer: Machine learning-based inference for particle physics. Computing and Software for Big Science, 4(1):1–25, 2020. * [8] C. Bunne, L. Papaxanthos, A. Krause, and M. Cuturi. Proximal optimal transport modeling of population dynamics. In International Conference on Artificial Intelligence and Statistics, pages 6511–6528. PMLR, 2022. * [9] M. Burger, S. Osher, J. Xu, and G. Gilboa. Nonlinear inverse scale space methods for image restoration. In N. Paragios, O. Faugeras, T. Chan, and C. Schnörr, editors, Variational, Geometric, and Level Set Methods in Computer Vision, pages 25–36, Berlin, Heidelberg, 2005. Springer Berlin Heidelberg. * [10] G. Carleo, I. Cirac, K. Cranmer, L. Daudet, M. Schuld, N. Tishby, L. Vogt-Maranto, and L. Zdeborová. Machine learning and the physical sciences. Reviews of Modern Physics, 91(4):045002, 2019. * [11] C. Chen, C. Li, L. Chen, W. Wang, Y. Pu, and L. C. Duke. Continuous-time flows for efficient inference and density estimation. In International Conference on Machine Learning (ICML), pages 824–833, 2018. * [12] T. Q. Chen, Y. Rubanova, J. Bettencourt, and D. K. Duvenaud. Neural ordinary differential equations. In Advances in Neural Information Processing Systems (NeurIPS), pages 6571–6583, 2018. * [13] O. collaboration and G. Abbiendi. Search for neutral Higgs bosons in collisions at 189 gev. The European Physical Journal C-Particles and Fields, 12(4):567–586, 2000. * [14] K. Cranmer. Kernel estimation in high-energy physics. Computer Physics Communications, 136(3):198–207, 2001. * [15] L. Dinh, D. Krueger, and Y. Bengio. NICE: non-linear independent components estimation. In Y. Bengio and Y. LeCun, editors, International Conference on Learning Representations (ICLR), 2015. * [16] L. Dinh, J. Sohl-Dickstein, and S. Bengio. Density estimation using real NVP. In International Conference on Learning Representations (ICLR), 2017\. * [17] C. Durkan, A. Bekasov, I. Murray, and G. Papamakarios. Neural spline flows. In Advances in Neural Information Processing Systems (NeurIPS), pages 7509–7520, 2019. * [18] J. Fan, A. Taghvaei, and Y. Chen. Variational Wasserstein gradient flow. arXiv preprint arXiv:2112.02424, 2021. * [19] C. Finlay, J.-H. Jacobsen, L. Nurbekyan, and A. M. Oberman. How to train your neural ODE: the world of Jacobian and kinetic regularization. In International Conference on Machine Learning (ICML), pages 3154–3164, 2020. * [20] W. H. Fleming and H. M. Soner. Controlled Markov processes and viscosity solutions, volume 25 of Stochastic Modelling and Applied Probability. Springer, New York, second edition, 2006. * [21] K. Fukumizu, A. Gretton, X. Sun, and B. Schölkopf. Kernel measures of conditional dependence. Advances in Neural Information Processing Systems, 20, 2007\. * [22] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020. * [23] W. Grathwohl, R. T. Chen, J. Betterncourt, I. Sutskever, and D. Duvenaud. FFJORD: Free-form continuous dynamics for scalable reversible generative models. In International Conference on Learning Representations (ICLR), 2019\. * [24] A. Gretton, K. M. Borgwardt, M. J. Rasch, B. Schölkopf, and A. Smola. A kernel two-sample test. Journal of Machine Learning Research (JMLR), 13(25):723–773, 2012\. * [25] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770–778, 2016. * [26] H. Heaton, S. W. Fung, A. T. Lin, S. Osher, and W. Yin. Wasserstein-based projections with applications to inverse problems. arXiv preprint arXiv:2008.02200, 2020. * [27] C.-W. Huang, R. T. Chen, C. Tsirigotis, and A. Courville. Convex potential flows: Universal probability distributions with optimal transport and convex optimization. arXiv preprint arXiv:2012.05942, 2020. * [28] C.-W. Huang, D. Krueger, A. Lacoste, and A. Courville. Neural autoregressive flows. In International Conference on Machine Learning (ICML), pages 2078–2087, 2018. * [29] M. F. Hutchinson. A stochastic estimator of the trace of the influence matrix for Laplacian smoothing splines. Communications in Statistics-Simulation and Computation, 19(2):433–450, 1990. * [30] R. Jordan, D. Kinderlehrer, and F. Otto. The variational formulation of the Fokker–Planck equation. SIAM Journal on Mathematical Analysis, 29(1):1–17, 1998. * [31] D. P. Kingma and P. Dhariwal. Glow: Generative flow with invertible 1x1 convolutions. In Advances in Neural Information Processing Systems (NeurIPS), pages 10215–10224, 2018. * [32] D. P. Kingma, T. Salimans, R. Jozefowicz, X. Chen, I. Sutskever, and M. Welling. Improved variational inference with inverse autoregressive flow. In Advances in Neural Information Processing Systems (NeurIPS), pages 4743–4751, 2016. * [33] I. Kobyzev, S. Prince, and M. Brubaker. Normalizing flows: An introduction and review of current methods. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020. * [34] N. Lei, K. Su, L. Cui, S.-T. Yau, and X. D. Gu. A geometric view of optimal transportation and generative model. Computer Aided Geometric Design, 68:1–21, 2019. * [35] A. T. Lin, S. W. Fung, W. Li, L. Nurbekyan, and S. J. Osher. Alternating the population and control neural networks to solve high-dimensional stochastic mean-field games. Proceedings of the National Academy of Sciences, 118(31):e2024713118, 2021. * [36] J. Lin, K. Lensink, and E. Haber. Fluid flow mass transport for generative networks. arXiv:1910.01694, 2019. * [37] P. Mokrov, A. Korotin, L. Li, A. Genevay, J. M. Solomon, and E. Burnaev. Large-scale Wasserstein gradient flows. Advances in Neural Information Processing Systems, 34:15243–15256, 2021. * [38] A. Müller. Integral probability metrics and their generating classes of functions. Advances in Applied Probability, 29(2):429–443, 1997. * [39] R. M. Neal. MCMC using Hamiltonian dynamics. Handbook of Markov Chain Monte Carlo, 2(11):2, 2011. * [40] F. Noé, S. Olsson, J. Köhler, and H. Wu. Boltzmann generators: Sampling equilibrium states of many-body systems with deep learning. Science, 365(6457), 2019. * [41] L. Nurbekyan, W. Lei, and Y. Yang. Efficient natural gradient descent methods for large-scale optimization problems. arXiv preprint arXiv:2202.06236, 2022. * [42] D. Onken, L. Nurbekyan, X. Li, S. W. Fung, S. Osher, and L. Ruthotto. A neural network approach applied to multi-agent optimal control. In 2021 European Control Conference (ECC), pages 1036–1041. IEEE, 2021. * [43] D. Onken, L. Nurbekyan, X. Li, S. W. Fung, S. Osher, and L. Ruthotto. A neural network approach for high-dimensional optimal control applied to multiagent path finding. IEEE Transactions on Control Systems Technology, 2022. * [44] D. Onken and L. Ruthotto. Discretize-optimize vs. optimize-discretize for time-series regression and continuous normalizing flows. arXiv:2005.13420, 2020. * [45] D. Onken, S. Wu Fung, X. Li, and L. Ruthotto. Ot-flow: Fast and accurate continuous normalizing flows via optimal transport. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, 2021. * [46] S. Osher, M. Burger, D. Goldfarb, J. Xu, and W. Yin. An iterative regularization method for total variation-based image restoration. Multiscale Modeling & Simulation, 4(2):460–489, 2005. * [47] G. Papamakarios, E. Nalisnick, D. J. Rezende, S. Mohamed, and B. Lakshminarayanan. Normalizing flows for probabilistic modeling and inference. arXiv:1912.02762, 2019. * [48] G. Papamakarios, T. Pavlakou, and I. Murray. Masked autoregressive flow for density estimation. In Advances in Neural Information Processing Systems (NeurIPS), pages 2338–2347, 2017. * [49] V. I. Paulsen and M. Raghupathi. An introduction to the theory of reproducing kernel Hilbert spaces, volume 152. Cambridge University Press, 2016. * [50] G. Peyré and M. Cuturi. Computational optimal transport. Foundations and Trends in Machine Learning, 11(5-6):355–607, 2019\. * [51] S. T. Rachev, L. B. Klebanov, S. V. Stoyanov, and F. Fabozzi. The Methods of Distances in the Theory of Probability and Statistics, volume 10. Springer, 2013. * [52] D. J. Rezende and S. Mohamed. Variational inference with normalizing flows. In International Conference on Machine Learning (ICML), pages 1530–1538, 2015. * [53] B. Roe. MiniBooNE particle identification. UCI Machine Learning Repository, 2010. * [54] L. Ruthotto and E. Haber. An introduction to deep generative modeling. GAMM-Mitteilungen, 44(2):e202100008, 2021. * [55] L. Ruthotto, S. J. Osher, W. Li, L. Nurbekyan, and S. W. Fung. A machine learning framework for solving high-dimensional mean field game and mean field control problems. Proceedings of the National Academy of Sciences, 117(17):9183–9193, 2020. * [56] A. Salim, A. Korba, and G. Luise. The Wasserstein proximal gradient algorithm. In H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 12356–12366. Curran Associates, Inc., 2020. * [57] T. Salimans, D. Kingma, and M. Welling. Markov chain Monte Carlo and variational inference: Bridging the gap. In International Conference on Machine Learning (ICML), pages 1218–1226, 2015. * [58] T. Salimans, H. Zhang, A. Radford, and D. N. Metaxas. Improving GANs using optimal transport. In International Conference on Learning Representations (ICLR), 2018\. * [59] M. Sanjabi, J. Ba, M. Razaviyayn, and J. D. Lee. On the convergence and robustness of training gans with regularized optimal transport. In Advances in Neural Information Processing Systems (NeurIPS), pages 7091–7101, 2018. * [60] F. Santambrogio. Optimal Transport for aAplied Mathematicians, volume 87 of Progress in Nonlinear Differential Equations and their Applications. Birkhäuser/Springer, Cham, 2015. Calculus of variations, PDEs, and modeling. * [61] D. W. Scott. Multivariate Density Estimation: Theory, Practice, and Visualization. John Wiley & Sons, 2015. * [62] B. W. Silverman. Density Estimation for Statistics and Data Analysis. CRC Press, 1986. * [63] J. Suykens, H. Verrelst, and J. Vandewalle. On-line learning Fokker-Planck machine. Neural Processing Letters, 7:81–89, 04 1998. * [64] E. G. Tabak and C. V. Turner. A family of nonparametric density estimation algorithms. Communications on Pure and Applied Mathematics, 66(2):145–164, 2013\. * [65] A. Tanaka. Discriminator optimal transport. In Advances in Neural Information Processing Systems (NeurIPS), pages 6816–6826, 2019. * [66] L. Tenorio. An Introduction to Data Analysis and Uncertainty Quantification for Inverse Problems. SIAM, 2017. * [67] C. Villani. Topics in optimal transportation, volume 58 of Graduate Studies in Mathematics. American Mathematical Society, Providence, RI, 2003. * [68] C. Villani. Optimal Transport: Old and New, volume 338. Springer Science & Business Media, 2008. * [69] A. Wehenkel and G. Louppe. Unconstrained monotonic neural networks. In Advances in Neural Information Processing Systems (NeurIPS), pages 1543–1553, 2019. * [70] M. Welling and Y. W. Teh. Bayesian learning via stochastic gradient Langevin dynamics. In International Conference on Machine Learning (ICML), pages 681–688, 2011. * [71] L. Yang and G. E. Karniadakis. Potential flow generator with ${L}_{2}$ optimal transport regularity for generative models. IEEE Transactions on Neural Networks and Learning Systems, 2020\. * [72] L. Yang and G. E. Karniadakis. Potential flow generator with l2 optimal transport regularity for generative models. IEEE Transactions on Neural Networks and Learning Systems, 2020\. * [73] J. Zech and Y. Marzouk. Sparse approximation of triangular transports, part II: The infinite-dimensional case. Constructive Approximation, 55(3):987–1036, 2022. * [74] L. Zhang, W. E, and L. Wang. Monge-Ampère flow for generative modeling. arXiv:1809.10188, 2018. * [75] V. M. Zolotarev. Metric distances in spaces of random variables and their distributions. Mathematics of the USSR-Sbornik, 30(3):373, 1976. ## Appendix A Appendix $\alpha=1$ | $\alpha=5$ | $\alpha=10$ | $\alpha=50$ ---|---|---|--- MMD2 Values, Single Shot 7.21e-2 | 2.30e-2 | 1.84e-2 | 7.73e-4 | | | MMD2 Values, JKO-Flow (five iterations) 2.10e-2 | 4.62e-4 | 1.02e-4 | 5.37e-5 | | | Figure 6: 2 Spirals dataset: Generated samples of $\hat{\rho}_{0}$ using the standard one-shot approach (top row). Generated using our proposed JKO-Flow using five iterations (bottom row). Here, we use $\alpha=$ 1, 5, 10, 50. JKO-Flow returns consistent results _regardless of the value of $\alpha$_. $\alpha=1$ | $\alpha=5$ | $\alpha=10$ | $\alpha=50$ ---|---|---|--- MMD2 Values, Single Shot 4.73e-3 | 7.33e-4 | 2.86e-4 | 7.03e-4 | | | MMD2 Values, JKO-Flow (five iterations) 5.16e-4 | 8.3e-5 | 3.27e-5 | 6.07e-4 | | | Figure 7: Swiss Roll dataset: Generated samples of $\hat{\rho}_{0}$ using the standard one-shot approach (top row). Generated using our proposed JKO-Flow using five iterations (bottom row). Here, we use $\alpha=$ 1, 5, 10, 50. JKO-Flow returns consistent results _regardless of the value of $\alpha$_. $\alpha=1$ | $\alpha=5$ | $\alpha=10$ | $\alpha=50$ ---|---|---|--- MMD2 Values, Single Shot 9.18e-3 | 2.69e-4 | 3.94e-4 | 7.10e-4 | | | MMD2 Values, JKO-Flow (five iterations) 1.07e-4 | 4.13e-4 | 2.67e-4 | 7.27e-6 | | | Figure 8: 8 Gaussians dataset: Generated samples of $\hat{\rho}_{0}$ using the standard one-shot approach (top row). Generated using our proposed JKO-Flow using five iterations (bottom row). Here, we use $\alpha=$ 1, 5, 10, 50. JKO-Flow returns consistent results _regardless of the value of $\alpha$_. $\alpha=1$ | $\alpha=5$ | $\alpha=10$ | $\alpha=50$ ---|---|---|--- MMD2 Values, Single Shot 9.84e-3 | 2.24e-4 | 6.51e-4 | 1.04e-4 | | | MMD2 Values, JKO-Flow (five iterations) 9.49e-4 | 1.0e-5 | 2.4e-5 | 9.3e-5 | | | Figure 9: Circles dataset: Generated samples of $\hat{\rho}_{0}$ using the standard one-shot approach (top row). Generated using our proposed JKO-Flow using five iterations (bottom row). Here, we use $\alpha=$ 1, 5, 10, 50. JKO-Flow returns consistent results _regardless of the value of $\alpha$_. $\alpha=1$ | $\alpha=5$ | $\alpha=10$ | $\alpha=50$ ---|---|---|--- MMD2 Values, Single Shot 1.18e-2 | 1.8e-3 | 1.37e-4 | 2.2e-5 | | | MMD2 Values, JKO-Flow (five iterations) 4.78e-4 | 2.63e-4 | 3.84e-4 | 4.30e-4 | | | Table 4: Pinwheel dataset: Generated samples of $\hat{\rho}_{0}$ using the standard one-shot approach (top row). Generated using our proposed JKO-Flow using five iterations (bottom row). Here, we use $\alpha=$ 1, 5, 10, 50. JKO-Flow returns consistent results _regardless of the value of $\alpha$_. $\alpha=1$ | $\alpha=5$ | $\alpha=10$ | $\alpha=50$ ---|---|---|--- MMD2 Values, Single Shot 1.45e-3 | 2.05e-3 | 2.49e-4 | 2.42e-4 | | | MMD2 Values, JKO-Flow (five iterations) 1.9e-5 | 5.0e-5 | 4.3e-5 | 1.08e-4 | | | Figure 10: Moons dataset: Generated samples of $\hat{\rho}_{0}$ using the standard one-shot approach (top row). Generated using our proposed JKO-Flow using five iterations (bottom row). Here, we use $\alpha=$ 1, 5, 10, 50. JKO-Flow returns consistent results _regardless of the value of $\alpha$_. $m=3$ | $m=4$ | $m=5$ | $m=8$ | $m=16$ ---|---|---|---|--- MMD2 Values, Single Shot 5.98e-3 | 4.54e-3 | 5.47e-3 | 1.19e-3 | 3.96e-3 | | | | MMD2 Values, JKO-Flow (five iterations) 1.42e-3 | 1.49e-5 | 6.11e-4 | 3.93e-5 | 2.19e-3 | | | | Figure 11: 2 Spirals dataset: Generated samples of $\hat{\rho}_{0}$ using the standard single shot approach (top row). Generated samples using our proposed JKO-Flow using five iterations (bottom row). Here, we fix $\alpha=50$ and vary the network width $m=3,4,5,8$, and $16$. JKO-Flow performs competitively even with lower number of parameters. $m=3$ | $m=4$ | $m=5$ | $m=8$ | $m=16$ ---|---|---|---|--- MMD2 Values, Single Shot 8.89e-3 | 7.71e-3 | 1.41e-3 | 1.37e-3 | 1.52e-3 | | | | MMD2 Values, JKO-Flow (five iterations) 1.49e-3 | 2.90e-4 | 6.13e-4 | 2.29e-4 | 8.4e-5 | | | | Figure 12: Swiss Roll dataset: Generated samples of $\hat{\rho}_{0}$ using the standard single shot approach (top row). Generated samples using our proposed JKO-Flow using five iterations (bottom row). Here, we fix $\alpha=5$ and vary the network width $m=3,4,5,8$, and $16$. JKO-Flow performs competitively even with lower number of parameters. $m=3$ | $m=4$ | $m=5$ | $m=8$ | $m=16$ ---|---|---|---|--- MMD2 Values, Single Shot 2.2e-3 | 1.05e-3 | 1.04e-3 | 2.3e-4 | 5.05e-4 | | | | MMD2 Values, JKO-Flow (five iterations) 1.33e-4 | 9.85e-4 | 2.4e-5 | 3.96e-4 | 1.07e-4 | | | | Figure 13: 8 Gaussians dataset: Generated samples of $\hat{\rho}_{0}$ using the standard single shot approach (top row). Generated samples using our proposed JKO-Flow using five iterations (bottom row). Here, we fix $\alpha=5$ and vary the network width $m=3,4,5,8$, and $16$. JKO-Flow performs competitively even with lower number of parameters. $m=3$ | $m=4$ | $m=5$ | $m=8$ | $m=16$ ---|---|---|---|--- MMD2 Values, Single Shot 2.06e-3 | 1.72e-3 | 1.37e-3 | 1.69e-3 | 1.34e-3 | | | | MMD2 Values, JKO-Flow (five iterations) 1.94e-3 | 3.24e-4 | 7.71e-4 | 5.9e-5 | 1.01e-4 | | | | Figure 14: Circles dataset: Generated samples of $\hat{\rho}_{0}$ using the standard single shot approach (top row). Generated samples using our proposed JKO-Flow using five iterations (bottom row). Here, we fix $\alpha=5$ and vary the network width $m=3,4,5,8$, and $16$. JKO-Flow performs competitively even with lower number of parameters. $m=3$ | $m=4$ | $m=5$ | $m=8$ | $m=16$ ---|---|---|---|--- MMD2 Values, Single Shot 1.10e-2 | 4.03e-3 | 2.27e-3 | 3.8e-3 | 5.43e-4 | | | | MMD2 Values, JKO-Flow (five iterations) 1.20e-3 | 8.23e-4 | 1.60e-3 | 7e-5 | 2.69e-4 | | | | Figure 15: Pinwheel dataset: Generated samples of $\hat{\rho}_{0}$ using the standard single shot approach (top row). Generated samples using our proposed JKO-Flow using five iterations (bottom row). Here, we fix $\alpha=5$ and vary the network width $m=3,4,5,8$, and $16$. JKO-Flow performs competitively even with lower number of parameters. $m=3$ | $m=4$ | $m=5$ | $m=8$ | $m=16$ ---|---|---|---|--- MMD2 Values, Single Shot 5.98e-3 | 4.54e-3 | 5.47e-3 | 1.2e-3 | 3.96e-3 | | | | MMD2 Values, JKO-Flow (five iterations) 1.42e-3 | 1.5e-5 | 6.11e-4 | 3.9e-5 | 2.19e-3 | | | | Figure 16: Moons dataset: Generated samples of $\hat{\rho}_{0}$ using the standard single shot approach (top row). Generated samples using our proposed JKO-Flow using five iterations (bottom row). Here, we fix $\alpha=5$ and vary the network width $m=3,4,5,8$, and $16$. JKO-Flow performs competitively even with lower number of parameters. Samples | Single Shot | JKO Flow ---|---|--- $\boldsymbol{x}\sim\rho_{0}(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | | $\boldsymbol{y}\sim\rho_{1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | | (a) Miniboone dimension 16 vs 17 Samples | Single Shot | JKO Flow ---|---|--- $\boldsymbol{x}\sim\rho_{0}(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | | $\boldsymbol{y}\sim\rho_{1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | | (b) Miniboone dimension 28 vs 29 Figure 17: Generated samples for the 43-dimensional Miniboone dataset using the single shot approach and JKO-Flow with 10 iterations for $\alpha=1$. To visualize the dataset, we show 2-dimensional slices. We show the forward flow $f(x)$ where $x\sim\rho_{0}$ and the genereated samples $f^{-1}(y)$ where $y\sim\rho_{1}$. Samples | Single Shot | JKO Flow ---|---|--- $\boldsymbol{x}\sim\rho_{0}(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | | $\boldsymbol{y}\sim\rho_{1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | | (a) Miniboone dimension 16 vs 17 Samples | Single Shot | JKO Flow ---|---|--- $\boldsymbol{x}\sim\rho_{0}(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | | $\boldsymbol{y}\sim\rho_{1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | | (b) Miniboone dimension 28 vs 29 Figure 18: Generated samples for the 43-dimensional Miniboone dataset using the single shot approach and JKO-Flow with 10 iterations for $\alpha=10$. To visualize the dataset, we show 2-dimensional slices. We show the forward flow $f(x)$ where $x\sim\rho_{0}$ and the genereated samples $f^{-1}(y)$ where $y\sim\rho_{1}$. Samples | Single Shot | JKO Flow ---|---|--- $\boldsymbol{x}\sim\rho_{0}(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | | $\boldsymbol{y}\sim\rho_{1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | | (a) Miniboone dimension 16 vs 17 Samples | Single Shot | JKO Flow ---|---|--- $\boldsymbol{x}\sim\rho_{0}(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | $f(\boldsymbol{x})$ | | $\boldsymbol{y}\sim\rho_{1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | $f^{-1}(\boldsymbol{y})$ | | (b) Miniboone dimension 28 vs 29 Figure 19: Generated samples for the 43-dimensional Miniboone dataset using the single shot approach and JKO-Flow with 10 iterations for $\alpha=50$. To visualize the dataset, we show 2-dimensional slices. We show the forward flow $f(x)$ where $x\sim\rho_{0}$ and the genereated samples $f^{-1}(y)$ where $y\sim\rho_{1}$.
# Extracting maximal entanglement from linear cluster states J. de Jong1, F. Hahn1, N. Tcholtchev2, M. Hauswirth1,2, and A. Pappa1,2 1 Electrical Engineering and Computer Science Department, Technische Universität Berlin, 10587 Berlin, Germany 2 Fraunhofer Institute for Open Communication Systems - FOKUS, 10589 Berlin, Germany (August 28, 2024) ###### Abstract Most quantum information processing architectures only allow for nearest- neighbour entanglement creation. In many cases, this prevents the direct generation of maximally entangled states, which are commonly used for many communication and computation tasks. Here we show how to obtain maximally entangled $\mathrm{GHZ}$ states between vertices initially connected by a minimum number of connections, which specifically allows them to share linear cluster states. We prove that the largest $\mathrm{GHZ}$ state that a linear cluster state on $n$ qubits can be transformed into by means of local Clifford unitaries, local Pauli measurements and classical corrections, is of size $\lfloor(n+3)/2\rfloor$. We demonstrate exactly which qubit selection patterns are possible below this threshold and which are not, and implement the transformation on the IBMQ Montreal quantum device for linear cluster states of up to $n=19$ qubits. ††preprint: APS/123-QED ## I Introduction Recent years have seen exciting developments in quantum computation and communication, both in theory and experiment. Building upon the year-long research on bipartite settings, focus has now also turned towards multipartite settings, where multiple vertices in a network share quantum resources between them. While the correlations of maximally entangled states, in particular Greenberger-Horne-Zeilinger ($\mathrm{GHZ}$) states [1] have naturally been the first to explore, other types of graph states have also been extensively examined [2, 3, 4, 5]. The possible transformations between quantum states is a topic that is heavily studied [6, 7, 8, 9, 10], and while several hardness results have emerged [11, 12], a lot of practical questions remain unanswered. But why should we be interested in transforming one quantum state to another in the first place? For one, the underlying network architecture might not allow for the direct distribution of large maximally entangled states. Here we show that there is an indirect remedy for this deficiency using suitable transformations of the distributed quantum states. In particular, we focus on the transformation of linear cluster states [13] to $\mathrm{GHZ}$ states, i.e. on transforming to maximally entangled states the graph states that are naturally distributable in linear quantum networks. Such transformations require the removal of some of the qubits from the state by measuring them, such that only a selected subset of the qubits of the resource linear cluster state form the final $\mathrm{GHZ}$ state; these transformations are therefore also called $\mathrm{GHZ}$ extractions. A previous study [14] by some of the authors showed how to extract three- and four-partite $\mathrm{GHZ}$ states from linear cluster states. Here, we conclude this study by providing a complete characterisation of which $\mathrm{GHZ}$ extractions are possible and which are not. Very importantly, we provide a tight upper bound to the size of the largest $\mathrm{GHZ}$ state that can be extracted, equal to $\lfloor(n+3)/2\rfloor$; interestingly this is slightly higher than the one of $n/2$ conjectured in Ref. [13]. In addition to our theoretical analysis, we perform experimental implementations of the $\mathrm{GHZ}$ extractions from linear cluster states with $n\in\\{5,7,\dots,19\\}$ qubits on the IBMQ Montreal device. Our manuscript is organized as follows: The notation, technical terminology and main definitions are introduced in Section II. Section III contains the main theoretical results. In Section IV the experiments are introduced, discussed, and their results presented. Finally, Section V discusses the results presented and opportunities for future research. Various technical details are diverted to topical appendices: Appendix A contains the proof of an important lemma stated in the theoretical section, Appendix B contains technical details regarding the post-processing steps during the extractions, and Appendix C contains technical details regarding the data analysis of the experimental section. Figure 1: Example of extracting GHZ states from a linear cluster state with seven qubits: The only $5$-partite GHZ state that can be extracted from this resource is on the qubits corresponding to $1,2,4,6,7$ and is highlighted in green. For $4$-partite $\mathrm{GHZ}$ states, we also highlight all $15$ possible extraction patterns in green, while the patterns in brown are impossible due to Lemma 1 and the patterns shown in violet are impossible due to both Corollary 1 and Lemma 1. Note that due to Theorem 1 it is impossible to extract $\mathrm{GHZ}$ states with six or more qubits from this resource – it is however trivially possible to extract all combinations of three-partite $\mathrm{GHZ}$ states. ## II Notation and terminology In this work, two quantum graph states play a central role: we define linear cluster states $\ket{\mathrm{L}}$ and GHZ states $\ket{\mathrm{GHZ}}$ as $\begin{split}\ket{\mathrm{L}}_{\\{1,\dots,n\\}}&\mathrel{\mathop{:}}=\frac{1}{2^{n}}\bigotimes_{i=1}^{n}\left(\ket{0}+\ket{1}\sigma_{z}^{i+1}\right)\\\ \ket{\mathrm{GHZ}}_{\\{1,\dots,m\\}}&\mathrel{\mathop{:}}=\frac{1}{\sqrt{2}}\left(\bigotimes_{i=1}^{m}\ket{0}+\bigotimes_{i=1}^{m}\ket{1}\right)\end{split}$ (1) and $\ket{\mathrm{L}}_{V}$, $\ket{\mathrm{GHZ}}_{V}$ as corresponding to the vertex set $V$. When context permits, with e.g. $\ket{\mathrm{L}}_{n}$ we denote the linear cluster state of size $n$. Our resource state is the $n$-partite linear cluster state $\ket{\mathrm{L}}_{V_{L}}$. As a graph state it corresponds to a line graph on the vertices $V_{L}{}\mathrel{\mathop{:}}=\\{1,2,\ldots,n\\}$. Here, each vertex $i$ corresponds to the $i$-th qubit of $\ket{\mathrm{L}}_{V_{L}}$ and the edges of the graph correspond to nearest-neighbour entangling controlled phase gates. This structure allows us to use the terms left and right neighbours of $i$ to indicate any vertices $h$, $j$ with $h<i$, $i<j$, respectively; e.g. the direct left and right neighbours of $i$ are $i\pm 1$. Let $V_{G}{}\subset V_{L}{}$ be a set of vertices for which we can extract a $\mathrm{GHZ}$ state from the linear cluster resource state. Performing Pauli measurements on the qubits corresponding to $V_{M}\mathrel{\mathop{:}}=V_{L}\setminus V_{G}{}$, we obtain a post- measurement state which is local-Clifford equivalent to the $\ket{\mathrm{GHZ}}_{V_{G}{}}$ state. By performing local operations based on the measurement outcomes, the state can then be locally transformed into this $\mathrm{GHZ}$ state. This construction allows for $V_{G}$ to inherit the neighbour structure from the linear network $V_{L}{}$: For a vertex $j\in V_{G}{}$, we use $j_{-}$ and $j_{+}$ to indicate the left and right neighbour of $j$ in $V_{G}$, respectively. We refer to the smallest and largest element of $V_{G}$ as the boundaries of the $\mathrm{GHZ}$ state. We finally define as a $k$-island any selection of consecutive vertices $i,i+1,\dots,i+k\in V_{G}{}$. ## III Main results We now examine what are the different types of GHZ states one can obtain from a given linear cluster state. We first provide an upper bound for the size $\absolutevalue{V_{G}{}}$ of the extracted GHZ state and we then show how to saturate it. In order to achieve this, we use Lemma 1, which provides an impossibility result for $2$-islands (the proof can be found in App. A). ###### Lemma 1. No $2$-island can have both a left and a right neighbour in $V_{G}$. If two vertices $i,i+1$ are in $V_{G}{}$, then there is either no vertex to the left of $i$ or no vertex to the right of $i+1$. Lemma 1 implies that all vertices $i$ in the target $\mathrm{GHZ}$ state must be ‘isolated’ in the linear cluster state; $i-1$ and $i+1$ cannot be in $V_{G}{}$ (with the exception of the boundaries). A corollary for $3$-islands follows directly: ###### Corollary 1. If $V_{G}$ contains a $3$-island, then $\absolutevalue{V_{G}{}}{}=3$. ###### Proof. Let $i,i+1,i+2$ be a $3$-island in $V_{G}{}$ and assume that $|V_{G}{}|\geq 4$, i.e. that we have $h<i$ or $j>i+2$ in $V_{G}$. Now, either $i,i+1$ form a $2$-island with both left-neighbour $h$ and right-neighbour $i+2$ or $i+1,i+2$ form a $2$-island with both left-neighbour $i$ and right-neighbour $j$. Both are in direct contradiction to Lemma 1. ∎ By the same argument, $k$-islands with $k\geq 4$ are impossible; ultimately, each such $k$-island would contain several $3$-islands in contradiction to Corollary 1. This allows us to calculate the upper bound to $\absolutevalue{V_{G}{}}{}$. ###### Theorem 1. The size of a GHZ state extractable from an $n$-partite linear cluster state via local Clifford operations, local Pauli measurements, and local unitary corrections, is upper-bounded as $\absolutevalue{V_{G}{}}\leq\left\lfloor\frac{n+3}{2}\right\rfloor$. ###### Proof. As there are at most two $2$-islands, for every other $i$ in $V_{G}{}$ both $i\pm 1$ were measured. Thus, to maximize $\absolutevalue{V_{G}{}}$, we may have $1,2,n-1,n$ in $V_{G}{}$, and $V_{M}{}$ containing every other vertex in between: For $n$ odd, $V_{M}{}=\\{3,5,\dots,n-2\\}$; for $n$ even $V_{M}{}=\\{3,5,\dots,n-5,n-3,n-2\\}$ (see 111 In the case of $n$ being even, there is more than one such pattern. While we have chosen here to measure the two consecutive vertices, $n-3$ and $n-2$, other possibilities would have been to measure consecutive vertices further to the left and measure only the even vertices to the right. Another option would have been to measure not two consecutive vertices, but a vertex of one of the $2$ islands, i.e. either $1$,$2$,$n-1$ or $n$. It is important to note that all resulting sets $V_{M}{}$ have the same size. ). In the even case, $n-2$ must be measured due to Corollary 1. In both cases $\absolutevalue{V_{G}{}}{}=n-|V_{M}{}|$ is upper bounded by $\left\lfloor\frac{n+3}{2}\right\rfloor$. ∎ We now show that there is a set of measurements that saturates the bound of Theorem 1 by explicitly giving such a measurement pattern. For $n\leq 5$ this pattern was shown in Ref. [14]; the special case $n=7$ is visualized in Figure 1. For the general case, let us consider a case distinction with respect to the parity of $n$: For odd $n$, we can choose $V_{M}=\\{2i+1\\}_{i=1}^{\frac{n-3}{2}}$ and every corresponding qubit to be measured in the $\sigma_{x}$-basis; we refer to this measurement pattern as the maximal pattern. The linear cluster state is a stabilizer state, i.e. it is an element of the shared $+1$ eigenspace of the operators $\\{l_{i}=\sigma_{z}^{i-1}\sigma_{x}^{i}\sigma_{z}^{i+1}\\}_{i\in V_{L}}$, where $\sigma_{z}^{0}$ and $\sigma_{z}^{n+1}$ are set equal to the identity. This set of operators forms the set of canonical generators of an Abelian subgroup of the $n$-qubit Pauli group known as the stabilizer of the linear cluster states. For an overview of the stabilizer formalism and stabilizer measurements in particular see [16], [17]. Consider the generator transformation $l_{2}\rightarrow l_{2}^{{}^{\prime}}=l_{2}l_{4}\dots l_{n-4}l_{n-2}\mathrel{\mathop{:}}=\sigma_{z}^{1}\sigma_{z}^{n}\prod_{2i\in V_{L}}\sigma_{x}^{2i},$ (2) which ensures that $l^{{}^{\prime}}_{2}$ and all odd-indexed generators commute with all measurement operators $\\{\sigma_{x}^{j_{0}}\\}_{j_{0}\in V_{M}{}}$. The post-measurement state is determined by replacing the other $\absolutevalue{V_{M}{}}$ generators $\\{l_{2i}\\}_{i=2}^{\frac{n-1}{2}}$ with the measurement operators –together with a multiplicative phase depending on the respective measurement outcome. Then (after removing the support on the measured qubits and applying a Hadamard transformation to $1$ and $n$) the post measurement state on $V_{G}$ is characterized by the generators $\sigma_{x}^{V_{G}{}}$ and $\\{m_{j_{0}}\sigma_{z}^{j_{0}}\sigma_{z}^{j_{0+}}\\}_{j_{0}\in V_{G}{}\setminus\\{n\\}}$, where the $m_{j_{0}}=\pm 1$ are phases due to the measurement outcomes. These phases can be accounted for by applying $\sigma_{x}$-operations to a selection of the nodes, recovering the generators of the $\ket{\mathrm{GHZ}}_{V_{G}{}}$ state. The number of measurements implies $\absolutevalue{V_{G}{}}{}=n-\absolutevalue{\\{3,5,\dots,n-2\\}}=\frac{n+3}{2}$ which saturates the bound for odd $n$. For even $n$, it suffices to observe that a $\sigma_{z}$-measurement on the qubit corresponding to $n$ yields a linear cluster state $\ket{\mathrm{L}}_{\\{1,2,\dots,n-1\\}}$ up to a randomized $\sigma_{z}^{n-1}$-correction depending on the measurement outcome. In analogy to the odd parity case we then obtain $V_{M}=\\{3,5,\dots,n-3,n\\}$ such that $\absolutevalue{V_{G}{}}{}=n-\absolutevalue{V_{M}{}}{}=\frac{n+2}{2}=\left\lfloor\frac{n+3}{2}\right\rfloor$ for even $n$. Note that the even-case-analysis above also applies for measuring an ‘internal’ node in the $\sigma_{y}$-basis, rather than the first or last; this does introduce a Clifford rotation on the two neighbours of the node which needs to be accounted for [18]. The resulting state is then also LOCC equivalent to an ($n-1$)-partite linear cluster state on the remaining nodes, from which in turn a $\ket{\mathrm{GHZ}}_{\frac{n+2}{2}}$ state can be extracted through the maximal pattern. This approach can be extended to more measurements, where additional ‘inside’ nodes are measured in the $\sigma_{y}$-basis, and ‘outside’ nodes are measured in the $\sigma_{z}$-basis. It is straightforward to see that any choice $V_{G}$ allowed by Lemma 1 can be seen as arising from such a setting. Finally, note that while Lemma 1 does allow $2$-islands on the boundaries of the extracted $\mathrm{GHZ}$ states, they do not necessarily have to be contained in them. For example, $\ket{\mathrm{GHZ}}_{\\{1,3,5,7\\}}$ can be extracted from $\ket{\mathrm{L}}_{\\{1,\ldots,7\\}}$ as shown in Figure 1. Rigorously stated, this pattern does not arise from one of the maximal patterns defined above, but can instead be considered as a maximal pattern $\ket{\mathrm{GHZ}}_{\\{0,1,3,5,7,8\\}}$ extracted from $\ket{\mathrm{L}}_{\\{0,\ldots,8\\}}$. Here, the additional qubits corresponding to $0$ and $8$ are just “virtual” and not really there; they simply help visualize all possible patterns: $\ket{\mathrm{GHZ}}_{\\{1,3,5,7\\}}$ can be extracted from the “virtual” state $\ket{\mathrm{GHZ}}_{0,1,3,5,7,8}$ by measuring qubits $0$ and $8$ in the $\sigma_{x}$-basis. The measurements on the other qubits are unaffected by this; the physical measurements of $2,4,6$ to obtain $\ket{\mathrm{GHZ}}_{\\{1,3,5,7\\}}$ from $\ket{\mathrm{L}}_{\\{1,\ldots,7\\}}$ are exactly the same as the ones that would be required to obtain $\ket{\mathrm{GHZ}}_{\\{0,1,3,5,7,8\\}}$ from the “virtual” $\ket{\mathrm{L}}_{\\{0,\ldots,8\\}}$. In this sense, all possible selections of $V_{G}$ can be seen as subsets of the maximal measurement patterns defined above. ## IV Experimental implementation We used the IBMQ Montreal device to experimentally demonstrate our protocol for the maximal extraction of $\mathrm{GHZ}$ states from resource linear cluster states. For odd $n\in\\{5,7,\dots,19\\}$ we prepared the state $\ket{\psi}_{n}=\bigotimes_{i\text{ odd}}H^{i}\ket{\mathrm{L}}_{\\{1,\ldots,n\\}},$ (3) i.e. the linear cluster state with every odd qubit rotated to the $\sigma_{x}$-basis. We then extract $\mathrm{GHZ}$ states for $m\in\\{\frac{n+3}{2}\\}_{n}=\\{4,5,\dots,11\\}$ using the maximal pattern described in the previous section. Implementing $\ket{\psi}_{n}$ instead of $\ket{\mathrm{L}}_{\\{1,\ldots,n\\}}$ allows us to reduce the circuit depth of the preparation circuit by one, when compiling for the gateset of the IBMQ Montreal device (Pauli-basis rotations, $CNOT$; see Figure 2). When considering the $\mathrm{GHZ}$ extraction, this approach has further benefits: The necessary Hadamard transformations on the first and last qubit have, in essence, been applied ‘in advance’, and the $\sigma_{x}$-measurements prescribed by the maximal pattern become $\sigma_{z}$-measurements, which are native to the device. The Pauli-based flips due to the measurement outcomes that are necessary to obtain the $\mathrm{GHZ}$ state can be performed in post-processing, as all the subsequent measurements on the $\mathrm{GHZ}$ state itself are in the Pauli basis. Figure 2: The circuit on the left and on the right are equal; the circuit on the left is the standard preparation of the $\ket{\mathrm{L}}_{3}$ state. The generalization to $\ket{\mathrm{L}}_{n}$ for higher (odd) $n$ follows naturally. The circuit on the right is the compiled version for the IBMQ Montreal chip. By not implementing the single-qubit gates in the grey box, the circuit depth is reduced by $25\%$. This results in a rotated linear cluster state $\ket{\psi}_{n}$ as defined in Eq.(3). This state carries the same entanglement properties as the $\ket{\mathrm{L}}_{n}$ state, and most notably can still be used to extract $\ket{\mathrm{GHZ}}$ states by adapting the measurement bases. Figure 3: A lower bound of the fidelity of the rotated linear cluster and $\mathrm{GHZ}$ states prepared on the IBMQ Montreal device. We prepared rotated (see Eq. 3) linear cluster states $\ket{\psi}_{n}$ for $n\in\\{5,7,\dots,19\\}$ and extracted $\ket{\mathrm{GHZ}}_{m}$ states for $m\in\\{4,5,\dots,11\\}$using the maximal pattern introduced in Sec. III. For states with a higher number of qubits, the lower bound on the linear cluster state is increasingly worse due to a technical aspect of the estimation method (see App. C for details). The results are ordered such that each linear cluster state $\ket{\mathrm{L}}_{n}$ is paired with the $\ket{\mathrm{GHZ}}_{m}$ state extracted from it. To benchmark the results, we compute an estimate for the lower bound of the fidelity for both the linear cluster states and the GHZ states extracted from them. For the linear cluster states we use methods adapted from [19] using insights originally presented in [20]; two measurement settings suffice to estimate the lower bound –one in which all qubits are measured in the $\sigma_{z}$-basis, and one in which all qubits are measured in the $\sigma_{x}$-basis. For the $\mathrm{GHZ}$ states we derive a similar technique. Again two measurement settings suffice –one where all the qubits of the $\mathrm{GHZ}$ state are measured in the $\sigma_{z}$-basis, and one where all the qubits are measured in the $\sigma_{x}$-basis. These measurements are performed in parallel to the $\sigma_{z}$-measurements of the qubits not included in the $\mathrm{GHZ}$ state that are required for the extraction. All measurements are repeated $32000$ times to obtain estimates for the expectation values. Figure 3 shows the lower bounds on the fidelity for all linear cluster states that we generated with the IBMQ Montreal device –as well as for the $\mathrm{GHZ}$ states we extracted from them. Note that our estimation method imposes a relative penalty for linear cluster state fidelity estimations compared to $\mathrm{GHZ}$ state fidelity estimations. However, this does not mean that the fidelity of the $\mathrm{GHZ}$ states is truly higher than that of the linear cluster states from which they were extracted: It simply means that we have used a method of bounding the fidelity from below, which works comparatively better for $\mathrm{GHZ}$ states than it does for linear cluster states. For the details of the estimation method we refer to App. C. ## V Discussion $\mathrm{GHZ}$ states are an indispensable resource in quantum communication settings. Our results show that their extraction in linear architectures is costly, since about half of the qubits in a linear cluster state must be measured to obtain a $\mathrm{GHZ}$ state on the remaining qubits. We give an exhaustive characterization of all possible $\mathrm{GHZ}$ states that can be extracted and provide a constructive method to obtain them, including calculations for the necessary local rotations on the remaining qubits. Finally, extending our methods to other simple graph states does not seem to be straightforward and requires more research. We note however that deriving a similar characterization for ring graph states, i.e. , those where the leftmost and rightmost qubits of the linear cluster state are connected, can be easily achieved using our methods. In this case, only a single $2$-island is possible, so the upper bound for $\absolutevalue{V_{G}{}}{}$ becomes $\lfloor\frac{n+1}{2}\rfloor$. ## VI Acknowledgements A.P., J.d.J. and F.H. acknowledge support from the German Research Foundation (DFG, Emmy Noether Grant No. 418294583). Upon completion of this work, we became aware of work similarly motivated [21]. ## References * Greenberger _et al._ [1989] D. M. Greenberger, M. A. Horne, and A. Zeilinger, Going beyond bell’s theorem, in _Bell’s Theorem, Quantum Theory and Conceptions of the Universe_, edited by M. Kafatos (Springer Netherlands, Dordrecht, 1989) pp. 69–72. * Schlingemann [2004] D.-M. Schlingemann, Error syndrome calculation for graph codes on a one-way quantum computer: Towards a quantum memory, Journal of Mathematical Physics 45, 4322 (2004). * Hein _et al._ [2004] M. Hein, J. Eisert, and H. J. Briegel, Multiparty entanglement in graph states, Physical Review A 69, 062311 (2004). * Bouchet [1988] A. Bouchet, Graphic presentations of isotropic systems, Journal of Combinatorial Theory, Series B 45, 58 (1988). * Hahn _et al._ [2022] F. Hahn, A. Dahlberg, J. Eisert, and A. Pappa, Limitations of nearest-neighbor quantum networks, Phys. Rev. A 106, L010401 (2022). * Briegel and Raussendorf [2001a] H. J. Briegel and R. Raussendorf, Persistent entanglement in arrays of interacting particles, Phys. Rev. Lett. 86, 910 (2001a). * Van den Nest _et al._ [2004] M. Van den Nest, J. Dehaene, and B. De Moor, Graphical description of the action of local Clifford transformations on graph states, Physical Review A 69, 022316 (2004). * Van den Nest _et al._ [2005] M. Van den Nest, J. Dehaene, and B. De Moor, Local unitary versus local Clifford equivalence of stabilizer states, Physical Review A 71, 062323 (2005). * Dahlberg and Wehner [2018] A. Dahlberg and S. Wehner, Transforming graph states using single-qubit operations, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 376, 20170325 (2018). * Dahlberg _et al._ [2020a] A. Dahlberg, J. Helsen, and S. Wehner, How to transform graph states using single-qubit operations: computational complexity and algorithms, Quantum Science and Technology 5, 045016 (2020a). * Dahlberg _et al._ [2020b] A. Dahlberg, J. Helsen, and S. Wehner, Counting single-qubit clifford equivalent graph states is# p-complete, Journal of Mathematical Physics 61, 022202 (2020b). * Dahlberg _et al._ [2020c] A. Dahlberg, J. Helsen, and S. Wehner, Transforming graph states to Bell-pairs is NP-Complete, Quantum 4, 348 (2020c). * Briegel and Raussendorf [2001b] H. J. Briegel and R. Raussendorf, Persistent Entanglement in Arrays of Interacting Particles, Physical Review Letters 86, 910 (2001b). * Hahn _et al._ [2019] F. Hahn, A. Pappa, and J. Eisert, Quantum network routing and local complementation, npj Quantum Information 5, 76 (2019). * Note [1] In the case of $n$ being even, there is more than one such pattern. While we have chosen here to measure the two consecutive vertices, $n-3$ and $n-2$, other possibilities would have been to measure consecutive vertices further to the left and measure only the even vertices to the right. Another option would have been to measure not two consecutive vertices, but a vertex of one of the $2$ islands, i.e. either $1$,$2$,$n-1$ or $n$. It is important to note that all resulting sets $V_{M}{}$ have the same size. * Gottesman [1997] D. Gottesman, Stabilizer codes and quantum error correction (1997). * Gottesman [1998] D. Gottesman, The heisenberg representation of quantum computers (1998). * git [2022] Supplementary information to the paper “Extracting maximal entanglement from linear cluster states”, https://github.com/hahnfrederik/Extracting-maximal-entanglement-from-linear-cluster-states (2022), [Online; accessed 25-November-2022]. * Tiurev and Sørensen [2022] K. Tiurev and A. S. Sørensen, Fidelity measurement of a multiqubit cluster state with minimal effort, Phys. Rev. Research 4, 033162 (2022). * Tóth and Gühne [2005] G. Tóth and O. Gühne, Detecting genuine multipartite entanglement with two local measurements, Phys. Rev. Lett. 94, 060501 (2005). * Mannalath and Pathak [2022] V. Mannalath and A. Pathak, Multiparty Entanglement Routing in Quantum Networks (2022), arXiv:2211.06690 [math-ph, physics:quant-ph]. * Dehaene and De Moor [2003] J. Dehaene and B. De Moor, Clifford group, stabilizer states, and linear and quadratic operations over gf(2), Phys. Rev. A 68, 042318 (2003). ## Appendix A Proof of Lemma 1 In this section we proof Lemma 1. We prove the theorem by contradiction. Fix a set $V_{G}$ such that $\\{i,i+1\\}\subset V_{G}{}$ and let the post- measurement state $\ket{\psi}_{V_{G}{}}$ be locally equivalent to $\ket{\mathrm{GHZ}}_{V_{G}{}}$. Assume now that there are both $j<i$ and $k>i+1$ for which both $j,k\in V_{G}{}$. W.l.o.g. assume that $j$ and $k$ are the direct left- and right neighbour of $i$ and $i+1$, respectively. Recall that a set of generators for the linear cluster state is $\\{l_{i_{0}}=\sigma_{z}^{i_{-}}\sigma_{x}^{i_{0}}\sigma_{z}^{i_{+}}\\}_{i_{0}\in V_{L}}$. If the post-measurement state is locally equivalent to the $\mathrm{GHZ}$ state then there must exist a (reversible) generator transformation such that their support on $i$ and $i+1$ coincides exactly with (the generators of) the $\mathrm{GHZ}$ state - up to local Clifford rotations. We will now show that, from a reversible transformation of the $\\{l_{i_{0}}\\}$, it is impossible to obtain such a set of generators when $j,i,i+1,k\in V_{G}{}$. This directly implies that a measurement pattern such that the $\mathrm{GHZ}$ state can be obtained is not possible. (A set of) generators for the $\mathrm{GHZ}$ state are, $\\{\sigma_{x}^{V_{G}{}}\\}\cup\\{\sigma_{z}^{i_{0}}\sigma_{z}^{i_{+}}\\}_{i_{0}\in V_{G}{}}$, where it is implied that $\sigma_{z}^{i_{+}}=1$ when $i_{+}\not\in V_{G}{}$. Focusing on $i$ and $i+1$, the only generators with non-trivial support are $\\{\alpha,\beta,\gamma,\delta\\}=\\{\sigma_{a_{i}}^{i},\sigma_{a_{i}}^{i}\sigma_{a_{i+1}}^{i+1},\sigma_{a_{i+1}}^{i+1},\sigma_{b_{i}}^{i}\sigma_{b_{i+1}}^{i+1}\\}$, where $a_{i},a_{i+1},b_{i},b_{i+1}\in\\{x,y,z\\}$ reflect the fact that the state is locally equivalent to the $\mathrm{GHZ}$ state. This implies that $a_{i}\not=b_{i}$ and $a_{i+1}\not=b_{i+1}$. Similarly, only the generators $l_{i-1},l_{i},l_{i+1}$ and $l_{i+2}$ of the linear cluster state (i.e. those with support on $i$ or $i+1$) can have a non- trivial contribution to the generator transformation on the vertices in question. Therefore, w.l.o.g., we can focus on just these four generators and restrict our attention to vertices $i$ and $i+1$. If we show that there is no reversible transformation of $\\{l_{k}\\}_{k=\\{i-1,i,i+1,i+2\\}}$ to obtain $\\{\alpha,\beta,\gamma,\delta\\}$ when only considering these nodes, the lemma follows. We show there is no such transformation by exhaustive contradiction. There are three different ways of creating generator $\alpha$: i) $\alpha\propto l_{i-1}=\sigma_{z}^{i}$, ii) $\alpha\propto l_{i}\circ l_{i+2}=\sigma_{x}^{i}$, iii) $\alpha\propto l_{i-1}\circ l_{i}\circ l_{i+2}=\sigma_{y}^{i}$, where ‘$\alpha\propto l_{i-1}$’ should be read as ‘$l_{i-1}$ takes the role of $\alpha$’, and $\circ$ denotes the (qubit-wise) product (e.g. $l_{i}\circ l_{i+1}=\sigma_{x}^{i}\sigma_{z}^{i+1}\circ\sigma_{z}^{i}\sigma_{x}^{i+1}\hat{=}\sigma_{y}^{i}\sigma_{y}^{i+1}$, where the last equality is up to an irrelevant global phase). Similarly, there are three different ways of creating generator $\gamma$: j) $\gamma\propto l_{i+2}=\sigma+{z}^{i+2}$, jj) $\gamma\propto l_{i-1}\circ l_{i+1}=\sigma_{x}^{i+2}$, jjj $\gamma\propto l_{i-1}\circ l_{i+1}\circ l_{i+2}=\sigma_{y}^{i+2}$. Picking e.g. i) and j) one sees that $\beta$ is fixed at $\propto\sigma_{z}^{i}\sigma_{z}^{i+1}$. But this is $l_{i-1}\circ l_{i+2}\propto\alpha\circ\gamma$, which would not be a reversible transformation of the generators $l_{i-1},l_{i},l_{i+1}$ and $l_{i+2}$. Any other pair from $\\{\textbf{i)},\textbf{ii)},\textbf{iii)}\\}$ and $\\{\textbf{j)},\textbf{jj)},\textbf{jjj)}\\}$ would also necessitate such a non-reversible transformation. In essence, when viewing the generators as vectors over $\mathbf{F}^{2n}_{2}$ through the binary representation [22], the argument follows from the observation that (the vector associated with) $\beta$ lies in the subspace spanned by (the vectors associated with) $\alpha$ and $\gamma$. As such there can never be a reversible (i.e. basis-preserving) operation on (the vectors associated with) $l_{i-1},l_{i},l_{i+1}$ and $l_{i+2}$ that obtains $\alpha,\beta$ and $\gamma$. ## Appendix B Local-Clifford corrections to obtain GHZ states. We provide a jupyter notebook for determining the required correction operations under [18]. ## Appendix C Estimation of lower bound for fidelity in experimental implementation. We here provide details for the method of estimation of the lower bound of the fidelity of both the linear cluster state and $\mathrm{GHZ}$ state, that has been used in the experimental implementation. The method is presented in and adapted from [19] using insights originally presented in [20]. The state that is prepared is $\ket{\psi}_{n}=\bigotimes_{i\text{ odd}}H^{i}\ket{\mathrm{L}}_{n},$ (4) which is a linear cluster state rotated by Clifford operations and thus a stabilizer state. Note that the generators $G^{L}$ for the stabilizer of $\ket{\psi}_{n}$ can be grouped into ‘odd’ generators $G^{L}_{o}=\\{\sigma_{z}^{i-1}\sigma_{z}^{i}\sigma_{z}^{i+1}\\}_{i\text{ odd}}$ and ‘even’ generators $G^{L}_{e}=\\{\sigma_{x}^{i-1}\sigma_{x}^{i}\sigma_{x}^{i+1}\\}_{i\text{ even}}$, where again $\sigma_{z}^{0}=\sigma_{z}^{n+1}=1$. The fidelity of the prepared state $\rho$ with the rotated linear cluster state is $F(\rho,\ket{\psi}_{n})=\tr\left[\rho\outerproduct{\psi}{\psi}_{n}\right]$. Writing $G_{o(e)}=\prod_{g\in G^{L}_{o(e)}}\frac{\mathbb{I}+g}{2}$, and using $\outerproduct{\psi}{\psi}_{n}=\prod_{g\in G}\frac{\mathbb{I}+g}{2}=G_{o}G_{e}$, we can write $\displaystyle F(\rho,\ket{\psi}_{n})$ $\displaystyle=\tr\left[G_{o}G_{e}\rho\right]$ (5) $\displaystyle=\tr\left[G_{o}\rho\right]+\tr\left[G_{e}\rho\right]-\tr\left[\mathbb{I}\rho\right]+\tr\left[K\rho\right],$ (6) where $K=\left(\mathbb{I}-G_{o}\right)\left(\mathbb{I}-G_{e}\right)$. $K$ is positive semidefinite and thus we can discard the last term to obtain a lower bound for the fidelity: $F(\rho,\ket{\psi_{n}})\geq\mathbb{E}\left[G_{o}\right]+\mathbb{E}\left[G_{e}\right]-1,$ (7) where $\mathbb{E}\left[G_{o(e)}\right]=\frac{1}{2^{\absolutevalue{\mathcal{S}_{o(e)}}}}\sum_{\sigma\in\mathcal{S}_{o(e)}}\tr\left[\rho\sigma\right]$ with $\mathcal{S}_{o(e)}=\langle G^{L}_{o(e)}\rangle\subset\mathcal{S}$ the subgroup generated by the ‘odd’ (‘even’) generators of $\ket{\psi}_{n}$. Notably, all terms $\tr\left[\rho\sigma\right]$ comprise of only $\sigma_{z}$-basis ($\sigma\in\mathcal{S}_{o}$) or $\sigma_{x}$-basis ($\sigma\in\mathcal{S}_{e}$) measurements. This means that just two measurement settings suffice to estimate the lower bound: measuring all vertices in the $\sigma_{z}$-basis, and measuring all vertices in the $\sigma_{x}$-basis. By repeating these measurements $32000$ times and obtaining the outcome statistics, we estimate all terms $\tr\left[\rho\sigma\right]$ by selecting the outcomes associated with the $+1$ and $-1$ eigenspaces of all different observables. For the $\mathrm{GHZ}$ state we use a similar method, where we now group the generators $G^{G}$ of the $\mathrm{GHZ}$ state into $G_{o}^{G}=\\{\sigma_{x}^{V_{G}{}}\\}$ and $G_{e}^{G}=\\{\sigma_{z}^{j}\sigma_{z}^{j_{+}}\\}_{j\in V_{G}{}}$, which again allows for an estimate of the lower bound with just two measurement settings. A caveat is that now there is only one ‘odd’ generator and thus $\mathbb{E}\left[G^{G}_{o}\right]=\frac{1}{2}\tr\left[\rho\mathbb{I}\right]+\frac{1}{2}\tr\left[\rho\sigma_{x}^{V_{G}{}}\right]$. By definition $\tr\left[\rho\mathbb{I}\right]=1$ and therefore the expectation value is more skewed towards $1$ than for the linear cluster state estimation. In other words it gives a higher bound on the fidelity when compared to the linear cluster state, since $G^{L}_{o}=O(2^{n})$ and as such the identity does not have such a strong impact on the estimate, especially for larger linear cluster states. To give another comparison between the two states, Figure 4 contains the same results as Figure 3 from the main text, but with the identity-term omitted. This gives a lower but more equal estimate for both classes of states. Figure 4: Lower bound on the fidelity using an adapted estimate method. In comparison with Figure 3, positive terms that favour the $\mathrm{GHZ}$ states are dropped, which renders a lower but more equal estimate on the fidelities for all states.
# Ferromagnetic Crossover within the Ferromagnetic Order of U7Te12 Petr Opletal Hironori Sakai Yoshinori Haga Yoshifumi Tokiwa Etsuji Yamamoto Shinsaku Kambe Yo Tokunaga Advanced Science Research Center, Japan Atomic Energy Agency, Tokai, Ibaraki 319-1195, Japan ###### Abstract We investigate the physical properties of a single crystal of uranium telluride U7Te12. We have confirmed that U7Te12 crystallizes in the hexagonal structure with three nonequivalent crystallographic uranium sites. The paramagnetic moments are estimated to be approximately 1 $\mu_{\rm B}$ per the uranium site, assuming a uniform moment on all the sites. A ferromagnetic phase transition occurs at $T_{\rm C}=48$ K, where the in-plane magnetization increases sharply, whereas the out-of-plane component does not increase significantly. With decreasing temperature further below $T_{\rm C}$ under field-cooling conditions, the out-of-plane component increases rapidly around $T^{\star}=26$ K. In contrast, the in-plane component hardly changes at $T^{\star}$. Specific heat measurement indicates no $\lambda$-type anomaly around $T^{\star}$, so this is a cross-over suggesting a reorientation of the ordering moments or successive magnetic ordering on the part of the multiple uranium sites. ††preprint: APS/123-QED ## I Introduction Uranium 5$f$ electrons play an important role in bringing out various solid- state properties. In comparison to 4$f$ electron states in lanthanide compounds, which are strongly localized, 5$f$ electron states extend into space [1]. This leads to inter-site interaction with direct overlap by 5$f$ electron states of the nearest uranium atoms, and/or hybridization with conduction electron states of ligand atoms. Meanwhile, the intrasite coupling with 6$d$ electron states is also important for microscopic understanding. To investigate the role of 5$f$ electrons in physical properties, it is important to systematically study a group of compounds constructed from the same elements. In this sense, uranium chalcogenides are promising research targets, which exhibit various physical properties, such as, piezomagnetism in UO2 [2], metal-insulator crossover induced by magnetic field in $\beta$-US2 [3, 4], and spin-triplet superconductivity recently discovered in UTe2 [5]. Especially, there has been a growing interest in the adjacent phase of UTe2, which is not a line phase, including nonstoichiometry caused by U deficiency [6]. In addition, U7Te12 can be grown with stoichiometric UTe2 by the molten salt flux method [7]. U7Te12 is reported to crystallize in non-centrosymmetric hexagonal Cr12P7-like structure (space group $P\overline{6}$, No. 174) [8, 9]. This structure is characterized by three nonequivalent uranium sites – U1 with multiplicity 1 and U2 and U3 with multiplicity 3. Only reported data for U7Te12 were obtained on polycrystalline samples showing ferromagnetic transition at 52 K [10, 9]. Although anisotropic magnetism is expected from the crystal structure, the details of this magnetism are unknown. Moreover, it was reported as a semimetallic compound [9]. However, no data on temperature- dependent resistivity has been presented until now. In this study, we prepared single crystals of U7Te12 by chemical vapor transport (CVT) and molten salt flux methods. Single crystals enable us to examine the anisotropic physical properties of this compound. Two different magnetic anomalies are observed through magnetic, electrical transport, and heat capacity measurements: one is a ferromagnetic transition at $T_{\rm C}=48$ K, and another is a ferromagnetic crossover at $T^{\star}=26$ K. The former ferromagnetic ordering is characterized by an increase of the $a$-axis magnetization, while the latter crossover is characterized by an increase of $c$-axis magnetization. We have also revealed that the material exhibits semimetallic conductivity, of which resistivity slightly decreases below $T_{\rm C}$ and conversely increases below $T^{\star}$. It may be a half- gapped semimetallic state induced by the ferromagnetic crossover. We discuss the possible nature of these magnetic anomalies and their connection to nonequal uranium sites. ## II Experimental methods Single crystals of U7Te12 were prepared by the CVT method with NH4Cl as a transport agent. Precursor elements of etched U metal and Te (Te 6N purity produced by Rare Metallic Co. Ltd.) in a 1:1.5 ratio were used and the quartz tube for growth was etched and smoothed inside by hydrofluoric acid. Precursor elements were sealed inside the quartz tube under high vacuum (10-5 mbar) with approx. 0.04 mg/cm3 of NH4Cl. During the evacuation, the whole quartz tube was heated up to 100∘C to desorb any water in NH4Cl. The sealed tube was placed in a two-zone horizontal electric tube furnace (ARF2-370-50KC, Asahi-rika Co., Ltd.). The entire tube was heated up to 950∘C for 24 hours to allow the reaction of contents. An opposite gradient was applied to clean the growth zone for three days. Following that, the gradient was flipped with the lower temperature - 850∘C at the growth zone (etched by hydrofluoric acid) and higher temperature - 950∘C at the charge zone and was left for 10 days. We were also able to prepare a single crystal by the molten-salt method as described in [7]. Single crystals grown by the molten-salt method tend to be prism-like with a hexagonal base growing along the c-axis with a thickness of 0.5 to 1 mm and a length of 2 to 6 mm. Single crystals obtained by the CVT method tend to have a block shape with dimensions ranging from 2 to 4 mm. Samples obtained by both methods have dark metallic color. The crystal structure was determined by single crystal diffraction at room temperature using graphite monochromated Mo $K_{\alpha}$ radiation. The scattered X-ray beam was recorded on an image plate detector (R-AXIS RAPID, Rigaku Corp.). Absorption correction with the empirical method was applied prior to the structural solution. Structural solution and crystallographic parameters fitting were performed by the SHELX program. The absence of any different phases was checked by electron-probe microanalysis using wavelength- dispersive spectrometers installed in a scanning electron microscope (JXA-8900, JEOL Ltd.). To measure the physical properties, samples were cut from larger single crystals by an electric spark saw and polished. Magnetic properties were measured by a commercial superconducting quantum interference device magnetometer (MPMS, Quantum Design) in the temperature range of 4.2 K to 300 K and a maximum magnetic field of 7 T. The obtained data were not corrected for the demagnetization effect. We estimate the demagnetization field to be below 0.05 T for the highest magnetic moment. The heat capacity was measured by relaxation method by two $\tau$ models in Physical Properties Measurement System (PPMS, Quantum Design) from 1.8 K up to 300 K. Electrical resistivity was measured by the four-probe method in PPMS. ## III Results ### III.1 Crystal structure Figure 1: Crystal structure of U7Te12 with coordination polyhedra for (a) U1 in tricapped trigonal prismatic and (b) U2 and (c) U3 in bicapped trigonal prismatic geometry. (d) Crystal structure only with uranium (for better clarity) highlighting triangular prism of U2 and U3. Distances between uranium atoms are shown. Table 1: Wyckoff notation, local symmetry, atomic coordinates, equivalent isotropic displacement parameter, and site occupancy for each atomic position in U7Te12 which are obtained from the single crystal X-ray refinement. atom | site | local symm. | $x$ | $y$ | $z$ | $B_{\mathrm{e}q}$ (Å2) | occupancy ---|---|---|---|---|---|---|--- U1 | $1a$ | $\overline{6}..$ | 0 | 0 | 0 | 0.53(3) | 1 U2 | $3k$ | $m..$ | 0.17296(11) | 0.46524(10) | 1/2 | 0.58(3) | 0.983(6) U3 | $3j$ | $m..$ | 0.43295(10) | 0.26651(11) | 0 | 0.56(2) | 1 Te1 | $3k$ | $m..$ | 0.21589(19) | 0.2081(2) | 1/2 | 0.57(4) | 1 Te2 | $3k$ | $m..$ | 0.52583(19) | 0.1345(2) | 1/2 | 0.59(3) | 1 Te3 | $3j$ | $m..$ | 0.3765(2) | 0.4949(2) | 0 | 0.71(4) | 1 Te4 | $3j$ | $m..$ | 0.0173(2) | 0.2649(2) | 0 | 0.72(5) | 0.955(10) We have confirmed U7Te12 crystallizes in Cr12P7-like structure (space group $P\overline{6}$, No. 174) with crystal lattice parameters $a=12.277(3)$ Å and $c=4.2646(5)$ Å. The resulting coordinates of the atoms are presented in Table 1. The structure was determined with final agreement factor $R_{1}$ = 3.05%. If the atomic occupancies on the U2 and Te4 sites were fixed as 1, the obtained equivalent displacement coefficients ($B_{\rm eq}$) were too large. Alternatively, if these occupancies were treated as fitting parameters, the occupancies were slightly reduced. However the $B_{\rm eq}$ values become reasonable, and then the $R_{1}$-value became slightly reduced. The resulting displacement factor of U2 sites remains large and qualitatively consistent with the result by Tougait et al. [9] As shown in Fig. 1(b), U2 is located in the polyhedron formed by Te3 and Te4, where U2 is much closer to Te4 leaving more space in the opposite direction. The atomic displacement would likely be larger in this direction, resulting in a large displacement factor. As shown in Fig. 1, two different coordination polyhedra of uranium and tellurium are seen in U7Te12. The coordination polyhedra for U1 sites is tricapped trigonal prismatic (Fig. 1(c)), while those for U2 and U3 sites are bicapped trigonal prismatic geometry (Figs. 1(d) and 1(e)). Therefore, two types (at least) of crystalline electric field (CEF) effects are expected to exist for different polyhedral U sites. At first glance, the polyhedral arrays of U2 and U3 appear similar, but their polyhedral sequences are quite different as follows. In the U7Te12 structure, the shortest distance between uranium sites is the U2-U2 distance of 3.917 Å which is longer than Hill’s limit [11]. This shortest U2-U2 corresponds to a side of the U2 triangle in the basal plane (Fig. 1). Since each U2 site is aligned in a straight line in the $c$-axis direction, every $c=4.2646$ Å of the lattice constant, a geometrical frustration effect may exist in the U2 prism if antiferromagnetic interactions were dominant. Meanwhile, the shortest distance in the U3 prisms corresponds to that along the $c$-axis, i.e., it is equal to the lattice parameter of $c$, whereas the side in the U3 triangle is equal to 4.45 Å. The shortest distance in U1 prism is the same as in U3, but for U1 no triangle exists. In between different uranium sites, the shortest one is found between U1-U3 equal to 4.67 $\mathrm{\AA}$. The distances between U1-U2 and between U2-U3 are more than 5 $\mathrm{\AA}$, as a result of the stacking sequence of U2 being shifted by $c/2$ as shown in Fig. 1(a). ### III.2 Magnetic properties Figure 2: Temperature dependence of magnetic susceptibility defined as $M/H$ for U7Te12 with an application of $\mu_{0}H=0.01$ T along the $a$-axis and $c$-axis. The data were taken under field-cooled (FC) conditions. Closed and open squares show data obtained from the field applied along the $a$-axis and $c$-axis, respectively. The inset shows detailed behavior close to ferromagnetic transition at 48 K for the field applied along the $c$-axis. Figure 3: Left figure shows magnetic curves at different temperatures for the field applied along the $a$-axis. Right figure shows magnetic curves at different temperatures for the field applied along the $c$-axis with an inset showing detail of magnetic curves around the zero field region. Figure 4: Temperature dependence of inverse magnetic susceptibility for magnetic field applied along the $a$-axis and $c$-axis. The fit to data by Curie-Weiss law and modified Curie-Weiss law are shown as dashed lines (for more information see text). The inset shows the low-temperature data of $M/H$ for different magnetic fields applied along the $a$-axis. Data at 1 T and 7 T are scaled and shifted for clarity. Figure 2 shows the temperature dependence of magnetic susceptibility $\chi$ which is defined as the magnetization ($M$) divided by the applied magnetic field ($H$) along the $a$\- and $c$-axes for U7Te12. The data shown in Fig. 2 were taken in the field cooled (FC) condition. As seen in Fig. 2, it appears as if U7Te12 underwent two ferromagnetic transitions along the $a$-axis and in the $c$-axis, respectively. For the magnetic susceptibility defined as $M/H$ along the $a$-axis, a ferromagnetic transition is observed at $T_{\mathrm{C}}$ = 48 K with a steep increase in magnetization. Below $T_{\mathrm{C}}$, a local maximum is observed in susceptibility at $T^{\star}\approx$ 26 K. A large increase on $c$-axis magnetic susceptibility is seen below $T_{\mathrm{C}}$ and an inflection point is seen at $T^{\star}$. This increase is observed even in higher magnetic fields indicating a ferromagnetic-like origin. By increasing $H$ along the $a$-axis, the local maximum around $T^{\star}$ in magnetic susceptibility is suppressed (inset of Fig. 4). The polarization of ferromagnetically ordered magnetic moments in high magnetic fields can result in an additional contribution to magnetic susceptibility which may appear as an additional shift of the local maximum at $T^{\star}$. U7Te12 is a soft ferromagnet with a small magnetic coercivity for a magnetic field applied along the $a$-axis. Magnetization curves for a magnetic field applied along the $a$-axis (Fig. 3) show almost no hysteresis down to 4.2 K, and quickly become nearly saturated, with a magnetization of 1.41 $\mu_{\mathrm{B}}$/U at 7 T. Spontaneous magnetization is observed below $T_{\mathrm{C}}$, but the magnetization curve at 50 K shows an ‘S’-shape characteristic just above $T_{\rm C}$. Although the $M$-$H$ data near $T_{\rm C}$ were analyzed by the the generalized Arrott–Noakes relationship $\left(H/M\right)^{1/\gamma}=(T-T_{\rm C})/T_{1}+(M/M_{1})^{1/\beta}$ with free parameters of $T_{1}$, and $M_{1}$. The fitting could not yield reasonable critical exponents in the temperature region. It may be due to the geometrically frustrated magnetism or competing interactions between in-plane and out-of-plane magnetism. In the case of $H\parallel c$, the magnetization curve (Fig. 3) exhibits a rectangular shape below $T^{\star}$. The saturation magnetization at 4.2 K is observed to be 0.89 $\mu_{\mathrm{B}}$/U at 7 T, which is nearly half the value for the $a$-axis. Similarly, the magnetization curve at 30 K above $T^{\star}$ shows an ‘S’-like shape again (Fig. 3(b)), except for the low- field region near zero field. In the field region above $T^{\star}$ but below $T_{\rm C}$, the $M$-$H$ curve exhibits a tiny remanent magnetization and a small coercive field (inset of the left figure in Fig. 3), which completely disappears above $T_{\rm C}$. The small remanent magnetization suggests a minor contribution from the $a$-axis due to misalignment. Paramagnetic susceptibility in the high-temperature region above 150 K for the $a$-axis obeys the Curie-Weiss law, as shown in Fig. 4. The Curie-Weiss fitting yields an effective moment $\mu_{\rm eff}$ of 1.23 $\mu_{\mathrm{B}}$/U assuming a uniform moment and Weiss temperature $\Theta_{a}$ of 63 K. In the case of $H\parallel c$, since the inverse magnetic susceptibility is not linear, the modified Curie-Weiss law is required, which yields $\chi_{0}=3.7\times 10^{-8}$ m3/mol, $\mu_{\mathrm{eff}}$ = 0.92 $\mu_{\mathrm{B}}$/U, and $\Theta_{c}$ = 36 K. Magnetic susceptibility for the $c$-axis follows the modified Curie-Weiss law down to $\approx$ 60 K. For both directions, the effective magnetic moment is much smaller than that for free U3+/U4+ ion. Positive $\Theta$ means that ferromagnetic interactions are dominant in the system, and $\Theta_{a}>\Theta_{c}$ suggests the in-plane interactions are stronger. Suski [10] previously discussed the possibility of different magnetic moments by assuming 3H4 term for U4+ ($5f^{2}$) for the uranium but different splitting based on the two types of polyhedral U sites, shown in Fig. 1. However, because U7Te12 can be formally considered as U${}^{4+}_{3}$U${}^{3+}_{4}$Te${}_{12}^{2-}$ from the charge neutral principle, valence conditions are probably needed, e.g. U3+ state on the U1 site, and U3+ or U4+ state assigned on the U2 and U3 sites, respectively. Additional experiments are desirable to determine the CEF and valence states. The magnetic anisotropy is evidently expected to be site-dependent in such a situation, but based on the ratio of magnetic moments $M_{a}/M_{c}$(2 K, 7 T) $\approx$ 4/3 and multiplicity of uranium sites we expect uraniums on U1 and U2 or U3 to be ordering in-plane at $T_{\mathrm{C}}$ while uraniums on U2 or U3 site are responsible for the $c$-axis magnetization increase at $T^{\star}$. ### III.3 Heat capacity Figure 5: Heat capacity of U7Te12 in different magnetic applied along the $c$-axis and $a$-axis. Data for $H\parallel a$ (empty symbols) are shifted downward by 40 J/mol$\cdot$K for clarity. Figure 6: Plot of $C/T$ vs. $T$ of U7Te12 in the magnetic field applied along the $c$-axis. The inset shows the low-temperature part of $C/T$ vs. $T^{2}$ to determine the Sommerfeld coefficient. Specific heat measurements were performed to examine the magnetic anomalies at $T_{\rm C}$ and $T^{\star}$. The results for various magnetic fields applied in both directions are shown in Figs. 5 and 6. Only one phase transition is observed at 48.6 K corresponding to $T_{\mathrm{C}}$. Concurrently, no $\lambda$-type anomaly is observed around $T^{\star}$, indicating that the anomaly at $T^{\star}$ is a magnetic crossover. As shown in Fig. 5, in the case of $H\parallel a$, the $\lambda$-type anomaly at $T_{\rm C}$ is easily blurred by a tiny external field of 0.1 T. Meanwhile, in the case of $H\parallel c$, the anomaly remains at 0.1 T, despite being blurred by an $a$-projected field due to a field misalignment within several degrees. The Sommerfeld coefficient well below $T_{\rm C}$ is extrapolated to be $\gamma=48$ mJ/mol$\cdot\mathrm{K}^{2}$ from data below 10 K, which is similar to the $\gamma$ values well below $T_{\rm C}$ in other ferromagnetic uranium compounds with itinerant $5f$ electrons (e.g., UCoAl $\gamma=48$ mJ/mol$\cdot\mathrm{K}^{2}$ [12], UCu2Ge2 $\gamma=30$ mJ/mol$\cdot\mathrm{K}^{2}$ [13], UGe2 $\gamma=35$ mJ/mol$\cdot\mathrm{K}^{2}$ [14], UIr $\gamma=50$ mJ/mol$\cdot\mathrm{K}^{2}$ [15], U5Sb4 $\gamma=37$ mJ/mol$\cdot\mathrm{K}^{2}$ [16]). In the $C/T$ vs. $T$ plot as shown in Fig. 6, instead of a $\lambda$-type anomaly, there is a sign of hump anomaly that spreads widely around $T^{\star}=26$ K, which does not seem to be affected by external fields along the $c$ axis and no change was observed for field applied along the $a$-axis. For a crossover from in-plane ferromagnetism to $c$-axis ferromagnetism of uniform moments, applying an external magnetic field will change $C/T$ vs. $T$ at the crossover, but it is not the case in U7Te12. Therefore, the crossover at $T^{\star}$ is most likely a crossover from a paramagnetic state to a ferromagnetic state with an out-of-plane component, in the background of in- plane ferromagnetism. Namely, it suggests the presence of partially disordered (paramagnetic) uranium sites in the intermediate region of $T^{\star}<T<T_{\rm C}$. It should be also noted that the $T^{\star}$ is reduced by the fields applied along the $a$-axis in electrical resistivity and magnetization. As shown in the inset of Fig. 6, the $\gamma$ value is reduced by an external field along the $c$ axis, but it is almost independent of $H$ along the $a$-axis. This indicates that the low-temperature state is only sensitive to the magnetic field in the $c$ direction. ### III.4 Electrical resistivity Figure 7: (a) Temperature dependence of electrical resistivity $\rho_{c}$ and $\rho_{a}$ for U7Te12 measured with the electrical current ($I$) applied along the $c$\- and $a$-axes, respectively. (b) $\rho_{c}(T)$ with applying fields along the $a$-axis. (c) $\rho_{a}(T)$ with applying fields along the $c$-axis. The inset shows a magnified $\rho$-$T$ plot around $T_{\mathrm{C}}$. $T_{\mathrm{C}}$ and $T^{\star}$ are highlighted by red and blue lines, respectively. As shown in Fig. 7(a), U7Te12 exhibits a semimetallic behavior. The electrical resistivity $\rho(T)$ is much lower than the reported value of $\rho({\rm 300\ K})\simeq$3 m$\Omega\cdot$cm [9], indicating the high-quality of our crystal. The $\rho_{c}(T)$ in the $c$-axis direction is smaller than $\rho_{a}(T)$ in the $a$-direction, confirming the high anisotropy of U7Te12. Both $\rho_{a}(T)$ and $\rho_{c}(T)$ show a gradual increase with decreasing temperature from 300 K. At $T_{\rm C}=48$ K, $\rho_{a}(T)$ shows a local maximum and $\rho_{c}(T)$ shows a kink anomaly. Notably, around $\sim$60 K above $T_{\rm C}$, a broad maximum is seen. Below 48 K, $\rho_{a}(T)$ keeps increasing without any other anomalies. For $\rho_{c}(T)$, a sharp decrease is observed below $T_{\rm C}$, but increases sharply below $T^{\star}=26$ K. The $\rho_{c}(T)$ below $T^{\star}$ does not follow the Arrhenius-type $T$-dependence, i.e., it is not a semiconducting behavior with a fixed energy gap. It may be a half-gapped semimetallic state induced by the ferromagnetic crossover. Figures 7(b) and 7(c) show the magnetic field dependence of $\rho_{c}(T)$ and $\rho_{a}(T)$ measured with applying magnetic fields along the perpendicular directions to the current ($I$) directions. For $\rho_{c}$, as shown in Fig. 7(b), when the magnetic field is applied along the $a$-axis, the ferromagnetic transition at $T_{\rm C}(0)=48$ K moves to higher temperatures, and it is easily smeared out as magnetic fluctuations are suppressed. In other words, the negative magnetoresistance effect of $\rho_{c}$ above $T_{\rm C}$ is significant in the in-plane field direction. The $T^{\star}$ slightly moves to lower temperatures with an increasing magnetic field, which is similar to the behavior observed in magnetic susceptibility in $H\parallel a$. The $\rho_{c}(T)$ below $T^{\star}$ does not show any significant magnetoresistance effect by applying fields along the $a$ axis. Alternatively, in $H\parallel c$ shown in Fig. 7(c), the $T_{\rm C}$ does not shift significantly below $\sim$3 T. There is no significant magnetoresistance effect in this field direction. The sample was easily moved from the fixed position in the higher fields along the hard magnetic axis due to the strong magnetic torque. Therefore, the data above 3 T were unable to be collected under these conditions. ## IV Discussions To begin with the discussion, the main experimental results are summarized as follows. Two anomalies at $T_{\rm C}=48$ K and around $T^{\star}=26$ K are observed in U7Te12. A ferromagnetic ordering occurs at $T_{\rm C}$ where the ferromagnetic component emerges in the basal plane. In the ferromagnetic state below $T_{\rm C}$, an additional ferromagnetic crossover along the $c$-axis occurs around $T^{\star}$. In the specific heat measurement, a clear $\lambda$-type anomaly is observed at $T_{\rm C}$, whereas a broad hump is barely visible around $T^{\star}$. Such an anisotropic ferromagnetic response along the $a$\- and $c$-axes can be due to the multiple uranium sites of U1, U2, and U3. As shown in Sec. IIIB, the in-plane ferromagnetism below $T_{\rm C}$ can be ascribed to the four out of the seven uranium moments, while the rest uranium moments would be responsible for the ferromagnetic crossover along the $c$ axis below $T^{\star}$. Based on these observations, the nature and origin of the anomaly at $T^{\star}$ would be discussed. In general, a crossover between paramagnetic and ferromagnetic states is often induced by external magnetic fields. This occurs in U7Te12 even at zero magnetic field. The ferromagnetic transition at $T_{\rm C}$ is a second-order transition, breaking the time-reversal symmetry. Considering the in-plane ferromagnetic ordering at the $T_{\mathrm{C}}$, the symmetry of U1, U2, and U3 positions denoted in Table 1 seems to be reduced to $m^{\prime}..$ by time-reversal symmetry breaking. If this were the case, the appearance of ferromagnetic moments along the $c$-axis below $T^{\star}$ broke the $m^{\prime}$ symmetry to the lowest symmetry 1, which should induce a phase transition at $T^{\star}$. This contradicts the observed crossover behavior below $T^{\star}$, indicating that the symmetry at the uranium sites is already lowered to 1 at $T_{\mathrm{C}}$. At the present, the origin of this symmetry lowering to 1 at $T_{\mathrm{C}}$ is unknown. Let us consider the in-plane ferromagnetic structure below $T_{\rm C}$ again. Considering the in-plane arrangement of uranium atoms, the U1 moments can form a collinear ferromagnetic structure without the cost of anisotropy energy. However for U2 sites, because the local easy axis is not parallel between the neighboring sites, a collinear ferromagnetic structure costs the anisotropy energy. Thus, non-colinear ferromagnetic states, such as 2-in-1-out (1-in-2-out), are considered more favorable. The same situation might be realized for the U3 site. In reality, the interaction between the out-of-plane site and the uranium moments of the other sites must be considered further. As a result, the overall ferromagnetic structure seems to be more complex. To identify the magnetic structure, it is necessary to perform neutron scattering experiments. ## V Conclusion We have succeeded in growing the single crystals of U7Te12 and characterized the physical properties. The ferromagnetic transition is observed at $T_{\rm C}=48$ K in different measurements with magnetic moment parallel to the basal plane. The crossover is observed around $T^{\star}=26$ K and is characterized by an increase of magnetic moment along the $c$-axis and no dependence on the magnetic field. We explain the low-temperature state below $T^{\star}$ as the ferromagnetic state with magnetic moment along the $c$-axis or canted from the $c$-axis. We propose that ordered magnetic moments originate from uranium at different crystallographic positions giving rise to ordering at $T_{\rm C}$ and $T^{\star}$. ## VI Acknowledgement We wish to thank C. Tabata, K. Kaneko, and M.-T. Suzuki, for the helpful discussion. A part of this work was supported by JSPS KAKENHI Grant Nos. JP20KK0061, and JP20K20905. This work was also partly supported by the JAEA Fund for Exploratory Researches (Houga fund) and REIMEI Research Program. ## References * Sechovsky and Havela [1998] V. Sechovsky and L. Havela, Chapter 1 Magnetism of ternary intermetallic compounds of uranium, in _Handbook of Magnetic Materials_ , Vol. Volume 11, edited by K. H. J. Buschow (Elsevier, Amsterdam, 1998) pp. 1–289. * Antonio _et al._ [2021] D. J. Antonio, J. T. Weiss, K. S. Shanks, J. P. Ruff, M. Jaime, A. Saul, T. Swinburne, M. Salamon, K. Shrestha, B. Lavina, D. Koury, S. M. Gruner, D. A. Andersson, C. R. Stanek, T. Durakiewicz, J. L. Smith, Z. Islam, and K. Gofryk, Piezomagnetic switching and complex phase equilibria in uranium dioxide, Communications Materials 2, 1 (2021). * Ikeda _et al._ [2009] S. Ikeda, H. Sakai, N. Tateiwa, T. D. Matsuda, D. Aoki, Y. Homma, E. Yamamoto, A. Nakamura, Y. Shiokawa, Y. Ota, K. Sugiyama, M. Hagiwara, K. Kindo, K. Matsubayashi, M. Hedo, Y. Uwatoko, Y. Haga, and Y. Ōnuki, Possible existence of magnetic polaron in nearly ferromagnetic semiconductor $\beta$-US2, J. Phys. Soc. Japan 78, 114704 (2009). * Sugiyama _et al._ [2011] K. Sugiyama, Y. Hirose, K. Enoki, S. Ikeda, E. Yamamoto, N. Tateiwa, Y. Haga, T. Kida, M. Hagiwara, K. Kindo, F. Honda, R. Settai, and Y. Onuki, Magnetic-field-induced metallic state in $\beta$-US2, in _Journal of the Physical Society of Japan_, Vol. 80 (The Physical Society of Japan, 2011) p. 104. * Ran _et al._ [2019] S. Ran, C. Eckberg, Q.-P. Ding, Y. Furukawa, T. Metz, S. R. Saha, I.-L. Liu, M. Zic, H. Kim, J. Paglione, and N. P. Butch, Nearly ferromagnetic spin-triplet superconductivity, Science 365, 684 (2019). * Haga _et al._ [2022] Y. Haga, P. Opletal, Y. Tokiwa, E. Yamamoto, Y. Tokunaga, S. Kambe, and H. Sakai, Effect of uranium deficiency on normal and superconducting properties in unconventional superconductor UTe2, Journal of Physics Condensed Matter 34, 175601 (2022). * Sakai _et al._ [2022] H. Sakai, P. Opletal, Y. Tokiwa, E. Yamamoto, Y. Tokunaga, S. Kambe, and Y. Haga, Single crystal growth of superconducting ${\mathrm{ute}}_{2}$ by molten salt flux method, Phys. Rev. Materials 6, 073401 (2022). * Breeze and Brett [1971] E. W. Breeze and N. H. Brett, The crystal structure of U7Te12 and U2O2Te (1971). * Tougait _et al._ [1998] O. Tougait, M. Potel, and H. Noël, Characterization of the Binary Uranium and Thorium Tellurides U7Te12 and Th7Te12, Inorganic Chemistry 37, 5088 (1998). * Suski [1972] W. Suski, Magnetic properties of UTe2 and U7Te12, physica status solidi (a) 13, 675 (1972). * Hill [1970] H. H. Hill, Early actinides: The periodic system’s f electron transition metal series., Nucl. Met., Met. Soc. AIME 17, 2 (1970). * Matsuda _et al._ [1999] T. Matsuda, Nbsp, D, Y. Aoki, H. Sugawara, H. Sato, A. Andreev, V, and V. Sechovsky, Specific-Heat Anomaly around Metamagnetic Transition in UCoAl, J. Phys. Soc. Japan 68, 3922 (1999). * Matsuda _et al._ [2007] T. D. Matsuda, S. Ikeda, E. Yamamoto, Y. Haga, and Y. Onuki, Magnetic property of a single crystal UCu2Ge2, J. Phys. Soc. Japan 76, 074708 (2007). * Ōnuki _et al._ [1992] Y. Ōnuki, I. Ukon, S. W. Yun, I. Umehara, K. Satoh, T. Fukuhara, H. Sato, S. Takayanagi, M. Shikama, and A. Ochiai, Magnetic and Electrical Properties of U-Ge Intermetallic Compounds, J. Phys. Soc. Japan 61, 293 (1992). * Galatanu _et al._ [2004] A. Galatanu, Y. Haga, E. Yamamoto, T. D. Matsuda, S. Ikeda, and Y. Onuki, High Temperature Magnetic Properties of UIr Single Crystals, J. Phys. Soc. Japan 73, 766 (2004). * Paixão _et al._ [1994] J. A. Paixão, J. Rebizant, A. Blaise, A. Delapalme, J. P. Sanchez, G. H. Lander, H. Nakotte, P. Burlet, and M. Bonnet, Magnetism of a new U-Sb phase: U5Sb4, Physica B: Physics of Condensed Matter 203, 137 (1994).
# Improving Cross-Modal Retrieval with Set of Diverse Embeddings Dongwon Kim1 Namyup Kim1 Suha Kwak1,2 1Dept. of CSE, POSTECH 2 Graduate School of AI, POSTECH {kdwon, namyup<EMAIL_ADDRESS> ###### Abstract Cross-modal retrieval across image and text modalities is a challenging task due to its inherent ambiguity: An image often exhibits various situations, and a caption can be coupled with diverse images. Set-based embedding has been studied as a solution to this problem. It seeks to encode a sample into a set of different embedding vectors that capture different semantics of the sample. In this paper, we present a novel set-based embedding method, which is distinct from previous work in two aspects. First, we present a new similarity function called smooth-Chamfer similarity, which is designed to alleviate the side effects of existing similarity functions for set-based embedding. Second, we propose a novel set prediction module to produce a set of embedding vectors that effectively captures diverse semantics of input by the slot attention mechanism. Our method is evaluated on the COCO and Flickr30K datasets across different visual backbones, where it outperforms existing methods including ones that demand substantially larger computation at inference. ## 1 Introduction Cross-modal retrieval is the task of searching for data relevant to a query from a database when the query and database have different modalities. While it has been studied for various pairs of modalities such as video-text [19, 3, 7] and audio-text [5, 16], the most representative setting for the task is the retrieval across image and text modalities [30, 47, 10, 38]. A naïve solution to cross-modal retrieval is a straightforward extension of the conventional unimodal retrieval framework [18, 17, 27], _i.e._ , learning a joint embedding space of the different modalities with known ranking losses (_e.g._ , contrastive loss [9] and triplet loss [45]). In this framework, each sample is represented as a single embedding vector and the task reduces to neighbor search on the joint embedding space. However, this naïve approach has trouble in handling the inherent ambiguity of the cross-modal retrieval across image and text modalities [30, 47, 10]. A cause of the ambiguity is the fact that even a single image often contains various situations and contexts. Consider an image in Figure 1, which illustrates a group of children in a skate park. One of the captions coupled with it could be about children carrying up a bike, while another may describe the child riding a skateboard. Indeed, different local features of the image are matched to different captions. Similarly, visual manifestations of a caption could vary significantly as text descriptions are highly abstract. This ambiguity issue suggests that a sample should be embedded while reflecting its varying semantics in cross-modal retrieval. Embedding models dedicated to the uni-modal retrieval do not meet this requirement since they represent a sample as a single embedding vector. Figure 1: An example of the ambiguity problem introduced in the cross-modal retrieval task; an image region and a word corresponding to each other are highlighted in the same color. This example demonstrates that a single image can be coupled with multiple heterogeneous captions. Various methods have been studied to mitigate the ambiguity issue of cross- modal retrieval. Most of them adopt cross-attention networks that directly predict the similarity of an input image-caption pair [31, 38, 51, 12, 25, 39, 50, 54, 14]. These models successfully address the ambiguity since they explicitly infer relations across the modalities by drawing attentions on both modalities at once. However, they inevitably impose a large computation burden on retrieval systems since they demand both image and caption to be processed together for computing their similarity; all data in a database thus have to be reprocessed whenever a query arrives. In contrast, methods using separate textual and visual encoders [30, 17, 21, 24] seek to find samples relevant to query through nearest-neighbor search on pre-computed embedding vectors, which are more suitable to nowadays retrieval systems working on huge databases. However, most of the previous arts in this direction cannot address the ambiguity issue since their encoders return a single embedding vector for a given input. Recent studies [47, 10] have advanced embedding model architectures to tackle the ambiguity issue even with separate encoders for the two modalities. To be specific, their models compute a set of heterogeneous embedding vectors for a given sample using self-attention layers on top of visual or textual features; such a set of embedding vectors is called _embedding set_ in the remainder of this paper. Then the ambiguity issue is addressed through elements of the embedding set that encode diverse semantics of input. Although the set-based embedding models enable a retrieval system to be powerful yet efficient, however, similarity functions used for their training do not consider the ambiguity of the data. Hence, training the models with the similarity functions often causes the following two side effects: (1) _Sparse supervision_ –An embedding set most of whose elements remain untrained, or (2) _Set collapsing_ –An embedding set with a small variance where elements do not encode sufficient ambiguity. Further, self-attention modules used for set prediction in the previous work do not explicitly consider disentanglement between set elements. These limitations lead to an embedding set whose elements encode redundant semantics of input, which also causes the _set collapsing_ and degrades the capability of learned embedding models. To address the aforementioned limitations of previous work, we propose a novel set-based embedding method for cross-modal retrieval. The proposed method is distinct from previous work in mainly two aspects. First, we design a novel similarity function for sets, called smooth-Chamfer similarity, that is employed for both training and evaluation of our model. In particular, our loss based on the smooth-Chamfer similarity addresses both limitations of the existing similarity functions, _i.e._ , sparse supervision and set collapsing. Second, we propose a model with a novel set prediction module motivated by slot attention [34]. In the proposed module, learnable embeddings called element slots compete with each other for aggregating input data while being transformed into an embedding set by progressive update. Therefore, our model captures the diverse semantic ambiguity of input successfully, with little redundancy between elements of the embedding set. The proposed method is evaluated and compared with previous work on two realistic cross-modal retrieval benchmarks, COCO [33] and Flickr30K [43], where it outperforms the previous state of the art in most settings. In summary, our contribution is three-fold as follows: * • We address issues on previous set-based embedding methods by proposing a novel similarity function for sets, named smooth-Chamfer similarity. * • We introduce a slot attention based set prediction module where elements of embedding set iteratively compete with each other for aggregating input data, which can capture semantic ambiguity of input without redundancy. * • Our model achieved state-of-the-art performance on the COCO and Flickr30K datasets, two standard benchmarks for cross-modal retrieval. ## 2 Related Work Cross-modal retrieval: Previous work can be categorized into two classes, methods using two separate encoders and those using a cross-attention network. Models of the first class consist of disjoint visual and textual encoders and learn an embedding space where embedding vectors of matching pairs are located nearby. Retrieval is then performed by finding nearest neighbors of query on the embedding space, which is efficient since their embedding vectors are pre- computed. To improve the quality of the embedding space, loss functions [17, 10], model architectures [24, 32, 28], and embedding methods [47, 10] have been proposed. Though the separate encoders enable simple and fast retrieval, they often failed to handle the inherent ambiguity of cross-modal retrieval, leading to inferior retrieval performance. On the other hand, models of the other class adopt a cross-attention network that directly predicts the similarity score between an image-caption pair [31, 51, 50, 54, 14, 53]. Though these methods make their models explicitly consider relations across modalities, they are not suitable for real-world retrieval scenarios due to the large computation burden imposed during evaluation. We thus focus on improving the separate encoders in this paper. Figure 2: An overview of our model. (a) The overall framework of our model. The model consists of three parts: visual feature extractor, textual feature extractor, and set-prediction modules $f^{\mathcal{V}}$ and $f^{\mathcal{T}}$. First, the feature extractors of each modality extract local and global features from input samples. Then, the features are fed to the set prediction modules to produce embedding sets $\mathbf{S}^{\mathcal{V}}$ and $\mathbf{S}^{\mathcal{T}}$. The model is trained with the loss using our smooth-Chamfer similarity. (b) Details of our set prediction module and attention maps that slots of each iteration produce. A set prediction module consists of multiple aggregation blocks that share their weights. Note that $f^{\mathcal{V}}$ and $f^{\mathcal{T}}$ have the same model architecture. Embedding beyond vector representation: Most of existing work projects each sample into a single embedding vector [17, 21, 24]. However, it has been empirically proven that such single vector representation is not sufficient to deal with inherent ambiguity of data [49, 46]. To mitigate this issue for the cross-modal retrieval task, PVSE [47] and PCME [10] introduce novel embedding methods. To be specific, PVSE represents each sample as a set of embedding vectors and PCME introduces probabilistic embedding where each sample is represented as a set of vectors sampled from a normal distribution. Slot attention: Recently, slot attention [34] is proposed to learn object- centric representation, which is especially beneficial for tasks that require perception of objectness, such as object discovery and set property prediction. Slots, which are embedding vectors sampled from a random distribution, compete with each other for explaining the input in an iterative manner. This attention mechanism enables the final outputs to encode heterogeneous semantics that appear in the input without any explicit supervision. To produce informative embedding sets with sufficient within-set variance, our set prediction module employs the slot attention mechanism. However, the ability of slot attention to discover individual objects is only verified on synthetic datasets [29] and is known to fail on real-world images [4]. We make three architectural modifications to resolve this issue: (1) using learnable embeddings for initial element slots instead of random vectors, (2) replacing GRU [8] with a residual sum, and (3) adding a global feature to the final element slots. We will detail these differences in Section 3.2. Without them, we observed that the loss does not converge, and thus training fails. ## 3 Proposed Method This section first describes the overall model architecture and elaborates on the proposed set prediction module. Then we introduce our smooth-Chamfer similarity and illustrate how it differs from existing set similarity functions. Finally, details of the training and inference of our model are provided. ### 3.1 Overall Model Architecture The overall framework of our method is illustrated in Figure 2. Our model architecture consists of a visual feature extractor, a textual feature extractor, and a set prediction module of each modality: $f^{\mathcal{V}}$ and $f^{\mathcal{T}}$. The feature extractors have two branches that compute local features and global features from the input sample, respectively. The extracted features are given as input to the set prediction modules, each of which fuses local and global features to encode an embedding set. For the visual and textual feature extractors, we followed the conventional setting of the previous work [47, 10, 28]. Visual feature extractor: We consider two different types of visual feature extractors for fair comparisons with previous work. One employs a flattened convolutional feature map as local features and their average pooled feature as the global feature. The other uses ROI features of a pretrained object detector as local features and their max pooled feature as the global feature. In any of these cases, the local and global features are denoted by $\psi^{\mathcal{V}}(\mathbf{x})\in\mathbb{R}^{N\times D}$ and $\phi^{\mathcal{V}}(\mathbf{x})\in\mathbb{R}^{D}$, respectively, where $\mathbf{x}$ is input image. Textual feature extractor: For the textual feature, we also employ two different types of extractors: bi-GRU [8] and BERT [13]. When using bi-GRU, for $L$-words input caption $\mathbf{y}$, we take GloVe [42] word embedding of each word as a local feature $\psi^{\mathcal{T}}(\mathbf{y})\in\mathbb{R}^{L\times 300}$. Then we apply bi- GRU with $D$ dimensional hidden states on the top of the $\psi^{\mathcal{T}}(\mathbf{y})$. The last hidden state is used as the global feature $\phi^{\mathcal{T}}(\mathbf{y})\in\mathbb{R}^{D}$. Similarly, when using BERT, output hidden states of BERT and their max pooled feature are used as a $\psi^{\mathcal{T}}(\mathbf{y})\in\mathbb{R}^{L\times D}$ and $\phi^{\mathcal{T}}(\mathbf{y})\in\mathbb{R}^{D}$, respectively. ### 3.2 Set Prediction Module Elements of an embedding set should encode heterogeneous semantics that appear in the input data. Otherwise, it will degenerate into a set with small variations that cannot handle the ambiguity of the input. Inspired by slot attention [34], we introduce an aggregation block, where element slots compete with each other for aggregating input data. The slots are progressively updated to capture various semantics of the input through multiple iterations of the aggregation block, and then the slots are fused with a global feature to be used as elements of the output embedding set. In this manner, the proposed module produces an embedding set whose elements encode substantially different semantics, meanwhile preserving the global context of the input data. In this section, we only present how the visual set prediction module $f^{\mathcal{V}}$ works; $f^{\mathcal{T}}$ has the same model architecture and works in the same manner with $f^{\mathcal{V}}$. As shown in Figure 2(b), the proposed set prediction module consists of multiple aggregation blocks, which share their weights. We define the initial element slots as learnable embedding vectors $\mathbf{E}^{0}\in\mathbb{R}^{K\times D}$, where $K$ is the cardinality of the embedding set. Then, the element slots of the $t$-th iteration $\mathbf{E}^{t}$ are computed by $\displaystyle\mathbf{E}^{t}=\mathrm{AggBlock}(\psi^{\mathcal{V}}(\mathbf{x});\mathbf{E}^{t-1})\in\mathbb{R}^{K\times D}.$ (1) The aggregation block first layer-normalizes [2] inputs and then linearly projects input local feature $\psi^{\mathcal{V}}(\mathbf{x})$ to $\mathbf{k}\in\mathbb{R}^{N\times D_{h}}$ and $\mathbf{v}\in\mathbb{R}^{N\times D_{h}}$, and $\mathbf{E}^{t-1}$ to $\mathbf{q}\in\mathbb{R}^{K\times D_{h}}$. 111When using convolutional features for visual feature extraction, we add sinusoidal positional encoding before projecting $\psi^{\mathcal{V}}(\mathbf{x})$ to $\mathbf{k}$. Then, the attention map $A$ between $\psi^{\mathcal{V}}(\mathbf{x})$ and $\mathbf{E}^{t-1}$ is obtained via $\displaystyle A_{n,k}=\frac{e^{M_{n,k}}}{\sum_{k=1}^{K}e^{M_{n,k}}},\text{ where }M=\frac{\mathbf{k}\mathbf{q}^{\top}}{\sqrt{D_{h}}}.$ (2) It is worth noting that the attention map is normalized over slots, not keys as in the transformer [48]. Since this way of normalization lets slots compete with each other, each slot attends to nearly disjoint sets of local features, and these sets will correspond to the distinctive semantics of the input. Using the obtained attention map, we update the element slot with the weighted mean of local features and then feed it to a multi-layer perceptron (MLP) with layer normalization, residual connection, and GELU [23] activation: $\displaystyle\hat{A}_{n,k}=\frac{A_{n,k}}{\sum_{n=1}^{N}A_{n,k}},\text{ }\overline{\mathbf{E}^{t}}=\hat{A}^{\top}\mathbf{v}W_{o}+\mathbf{E}^{t-1},$ (3) $\displaystyle\mathbf{E}^{t}=\mathrm{AggBlock}(\psi^{\mathcal{V}}(\mathbf{x});\mathbf{E}^{t-1})=\mathrm{MLP}(\overline{\mathbf{E}^{t}})+\overline{\mathbf{E}^{t}},$ (4) where $W_{0}\in\mathbb{R}^{D_{h}\times D}$ is a learnable linear projection. The output of the aggregation block is used as an element slot for the $t$-th iteration. As shown in Figure 2, attention maps that slots of early stage produce are sparse and noisy, but refined to aggregate local features with distinctive semantics as proceeds. Finally, at $T$-th iteration, the model predicts an embedding set $\mathbf{S}$ by adding the global feature $\phi^{\mathcal{V}}(\mathbf{x})$ to each element slot $\mathbf{E}^{T}$: $\displaystyle\mathbf{S}=\mathrm{LN}(\mathbf{E}^{T})+\big{[}\mathrm{LN}(\phi^{\mathcal{V}}(\mathbf{x}))\stackrel{{\scriptstyle\scriptscriptstyle\times K}}{{\cdots}}\big{]}),$ (5) where $\mathrm{LN}$ is a layer normalization, and $\big{[}\phi^{\mathcal{V}}(\mathbf{x})\stackrel{{\scriptstyle\scriptscriptstyle\times K}}{{\cdots}}\big{]}\in\mathbb{R}^{K\times D}$ is $K$ repetitions of $\phi^{\mathcal{V}}(\mathbf{x})$. The module benefits from the global feature when treating samples with little ambiguity since the global feature allows them to be represented as an embedding set of small within-set variance. ### 3.3 Smooth-Chamfer Similarity Figure 3: Comparison between our similarity function and existing ones used for embedding set. Illustrations of embedding space before and after optimizing each similarity function are presented. Presented two sets are matching pairs, and thus optimized to maximize their similarity. Lines indicate the association that similarity functions consider, where their intensities represent the weights given to each association. Our smooth-Chamfer similarity is proposed to resolve the drawbacks of the existing set-based embedding approaches. Before presenting smooth-Chamfer similarity, we first review the similarity functions used in PVSE [47] and PCME [10]. Figure 3 illustrates embedding spaces trained with these set similarity functions. Let $\mathbf{S}_{1}$ and $\mathbf{S}_{2}$ be the embedding sets, which are the set of vectors. $c(x,y)$ and $d(x,y)$ denote the cosine similarity and $\ell_{2}$-normalized Euclidean distance between vectors $x$ and $y$, respectively. PVSE adopts the multiple instance learning (MIL) framework [15] during training and inference. Its similarity function, which we call MIL similarity, is given by $s_{\mathrm{MIL}}(\mathbf{S}_{1},\mathbf{S}_{2})=\max_{x\in\mathbf{S}_{1},y\in\mathbf{S}_{2}}c(x,y)$. MIL similarity only takes account of the closest pair of elements, as shown in Figure 3(a). While this behavior simplifies similarity measurement, MIL similarity suffers from the _sparse supervision_ problem. In other words, the majority of elements are not sampled as the closest pair and thus remain untrained. On the other hand, PCME uses the match probability (MP) as a similarity function, which is defined as $s_{\mathrm{MP}}(\mathbf{S}_{1},\mathbf{S}_{2})=\sum_{x\in\mathbf{S}_{1},y\in\mathbf{S}_{2}}\sigma(\alpha d(x,y)+\beta),$ where $\alpha$ and $\beta$ are learnable parameters and $\sigma$ is the sigmoid function. MP takes average on every pairwise distance between elements. Though MP resolves the sparse supervision, it introduces the _set collapsing_ problem. As presented in Figure 3(b), directly optimizing MP leads to a collapsed embedding set since elements of embedding sets pull each other without the consideration of their relative proximity. To mitigate this issue, MP requires a specialized loss function [10, 40], which makes MP incompatible with other losses or frameworks. The aforementioned problems stem from the way how similarity functions associate set elements. Associating only the nearest pair of elements like MIL leads to insufficient supervisory signals, while equally taking account of every possible association like MP results in a collapsed set. Unlike these similarity functions, our smooth-Chamfer similarity associates every possible pair with different degrees of weight, as illustrated in Figure 3(c). Smooth- Chamfer similarity is formulated as $\begin{split}s_{\mathrm{SC}}(\mathbf{S}_{1},\mathbf{S}_{2})=\frac{1}{2\alpha|\mathbf{S}_{1}|}\sum_{x\in\mathbf{S}_{1}}\log\bigg{(}\sum_{y\in\mathbf{S}_{2}}e^{\alpha c(x,y)}\bigg{)}\\\ +\frac{1}{2\alpha|\mathbf{S}_{2}|}\sum_{y\in\mathbf{S}_{2}}\log\bigg{(}\sum_{x\in\mathbf{S}_{1}}e^{\alpha c(x,y)}\bigg{)},\end{split}$ (6) where $|\cdot|$ is the set cardinality and $\alpha>0$ is a scaling parameter. Using Log-Sum-Exp, it can be reformulated in a more interpretable form as shown below: $\small\begin{split}s_{\mathrm{SC}}(\mathbf{S}_{1},\mathbf{S}_{2})=\frac{1}{2\alpha|\mathbf{S}_{1}|}\sum_{x\in\mathbf{S}_{1}}\underset{y\in\mathbf{S}_{2}}{\mathrm{LSE}}\bigg{(}\alpha c(x,y)\bigg{)}\\\ +\frac{1}{2\alpha|\mathbf{S}_{2}|}\sum_{y\in\mathbf{S}_{2}}\underset{x\in\mathbf{S}_{1}}{\mathrm{LSE}}\bigg{(}\alpha c(x,y)\bigg{)}.\end{split}$ (7) LSE indicates Log-Sum-Exp, which is the smooth approximation of the max function. Thanks to the property of LSE, smooth-Chamfer similarity softly assigns elements in one set to those in the other set, where the weight for a pair of elements is determined by their relative proximity. Hence, smooth- Chamfer similarity can provide dense supervision like MP but without the collapsing issue. By replacing LSE with the max function, we can consider a non-smooth version of smooth-Chamfer similarity, which we refer to as Chamfer similarity. Chamfer similarity clears up the problem of set collapsing by assigning individual elements of $\mathbf{S}_{1}$ to their closest neighbor in $\mathbf{S}_{2}$, and vice-versa. However, unlike smooth-Chamfer similarity, it does not provide supervision that is dense as MP since most associations are not considered. The behavior of smooth-Chamfer similarity can be clearly demonstrated by its gradient with respect to $c(x,y)$, which is given by $\begin{split}\frac{\partial s_{\mathrm{SC}}(\mathbf{S}_{1},\mathbf{S}_{2})}{\partial c(x^{\prime},y^{\prime})}=\frac{1}{2|\mathbf{S}_{1}|}\frac{e^{\alpha c(x^{\prime},y^{\prime})}}{\sum_{y\in\mathbf{S}_{2}}e^{\alpha c(x^{\prime},y)}}\\\ +\frac{1}{2|\mathbf{S}_{2}|}\frac{e^{\alpha c(x^{\prime},y^{\prime})}}{\sum_{x\in\mathbf{S}_{1}}e^{\alpha c(x,y^{\prime})}},\end{split}$ (8) where $x^{\prime}$ and $y^{\prime}$ are elements of $\mathbf{S}_{1}$ and $\mathbf{S}_{2}$, respectively. One can see that the gradient is the sum of two relative similarity scores, suggesting that the gradient for $c(x^{\prime},y^{\prime})$ is emphasized when $x^{\prime}$ and $y^{\prime}$ are close to each other. Using the weighting scheme based on the relative proximity, we can give denser supervision during training, while preserving sufficient within-set variance. ### 3.4 Training and Inference Training: Our model is trained with the objective functions presented in [47], which consists of the triplet loss, diversity regularizer, and Maximum Mean Discrepancy (MMD) [20] regularizer. Following previous work [17, 47, 51], we adopt the triplet loss with hard negative mining. Let $B=\\{(S^{\mathcal{V}}_{i},S^{\mathcal{T}}_{i})\\}^{N}_{i=1}$ be a batch of embedding sets. Then the triplet loss is given by $\displaystyle\begin{split}\mathcal{L}_{\mathrm{tri}}(B)=\sum_{i=1}^{N}\max_{j}[\delta+s_{\mathrm{SC}}(\mathbf{S}^{\mathcal{V}}_{i},\mathbf{S}^{\mathcal{T}}_{j})-s_{\mathrm{SC}}(\mathbf{S}^{\mathcal{V}}_{i},\mathbf{S}^{\mathcal{T}}_{i})]_{+}\\\ +\sum_{i=1}^{N}\max_{j}[\delta+s_{\mathrm{SC}}(\mathbf{S}^{\mathcal{T}}_{i},\mathbf{S}^{\mathcal{V}}_{j})-s_{\mathrm{SC}}(\mathbf{S}^{\mathcal{T}}_{i},\mathbf{S}^{\mathcal{V}}_{i})]_{+},\end{split}$ (9) where $[\cdot]_{+}$ indicates the hinge function and $\delta>0$ is a margin. The MMD regularizer minimizes MMD between embedding sets of image and text. It prevents embeddings of different modalities from diverging at the early stage of training. The diversity regularizer penalizes similar element slots, which helps the model produce a set of diverse embeddings. The two regularizers are formulated as $\mathcal{L}_{\mathrm{mmd}}=\mathrm{MMD}(B^{\mathcal{V}},B^{\mathcal{T}})$ and $\mathcal{L}_{\mathrm{div}}=\sum_{x,x^{\prime}\in\mathbf{E}^{T}}e^{-2||x-x^{\prime}||^{2}_{2}}$, where $B^{\mathcal{V}}$ and $B^{\mathcal{T}}$ denote subsets of batch $B$ consisting of each modality embeddings. Inference: We pre-compute the embedding set of size $K$ for every sample in the database. Then, the sample most relevant to a query is retrieved via nearest-neighbor search on embedding sets, using smooth-Chamfer similarity. ## 4 Experiments Method | CA | 1K Test Images | 5K Test Images ---|---|---|--- Image-to-Text | Text-to-Image | RSUM | Image-to-Text | Text-to-Image | RSUM R@1 | R@5 | R@10 | R@1 | R@5 | R@10 | R@1 | R@5 | R@10 | R@1 | R@5 | R@10 ResNet-152 + Bi-GRU VSE++ [17] | ✗ | 64.6 | 90.0 | 95.7 | 52.0 | 84.3 | 92.0 | 478.6 | 41.3 | 71.1 | 81.2 | 30.3 | 59.4 | 72.4 | 355.7 PVSE [47] | ✗ | 69.2 | 91.6 | 96.6 | 55.2 | 86.5 | 93.7 | 492.8 | 45.2 | 74.3 | 84.5 | 32.4 | 63.0 | 75.0 | 374.4 PCME [10] | ✗ | 68.8 | - | - | 54.6 | - | - | - | 44.2 | - | - | 31.9 | - | - | - Ours | ✗ | 70.3 | 91.5 | 96.3 | 56.0 | 85.8 | 93.3 | 493.2 | 47.2 | 74.8 | 84.1 | 33.8 | 63.1 | 74.7 | 377.7 Faster R-CNN + Bi-GRU SCAN† [31] | ✓ | 72.7 | 94.8 | 98.4 | 58.8 | 88.4 | 94.8 | 507.9 | 50.4 | 82.2 | 90.0 | 38.6 | 69.3 | 80.4 | 410.9 VSRN† [32] | ✗ | 76.2 | 94.8 | 98.2 | 62.8 | 89.7 | 95.1 | 516.8 | 53.0 | 81.1 | 89.4 | 40.5 | 70.6 | 81.1 | 415.7 CAAN [54] | ✓ | 75.5 | 95.4 | 98.5 | 61.3 | 89.7 | 95.2 | 515.6 | 52.5 | 83.3 | 90.9 | 41.2 | 70.3 | 82.9 | 421.1 IMRAM† [6] | ✓ | 76.7 | 95.6 | 98.5 | 61.7 | 89.1 | 95.0 | 516.6 | 53.7 | 83.2 | 91.0 | 39.7 | 69.1 | 79.8 | 416.5 SGRAF† [14] | ✓ | 79.6 | 96.2 | 98.5 | 63.2 | 90.7 | 96.1 | 524.3 | 57.8 | - | 91.6 | 41.9 | - | 81.3 | - VSE∞ [28] | ✗ | 78.5 | 96.0 | 98.7 | 61.7 | 90.3 | 95.6 | 520.8 | 56.6 | 83.6 | 91.4 | 39.3 | 69.9 | 81.1 | 421.9 NAAF† [53] | ✓ | 80.5 | 96.5 | 98.8 | 64.1 | 90.7 | 96.5 | 527.2 | 58.9 | 85.2 | 92.0 | 42.5 | 70.9 | 81.4 | 430.9 Ours | ✗ | 79.8 | 96.2 | 98.6 | 63.6 | 90.7 | 95.7 | 524.6 | 58.8 | 84.9 | 91.5 | 41.1 | 72.0 | 82.4 | 430.7 Ours† | ✗ | 80.6 | 96.3 | 98.8 | 64.7 | 91.4 | 96.2 | 528.0 | 60.4 | 86.2 | 92.4 | 42.6 | 73.1 | 83.1 | 437.8 ResNeXt-101 + BERT VSE∞ [28] | ✗ | 84.5 | 98.1 | 99.4 | 72.0 | 93.9 | 97.5 | 545.4 | 66.4 | 89.3 | 94.6 | 51.6 | 79.3 | 87.6 | 468.9 VSE∞† [28] | ✗ | 85.6 | 98.0 | 99.4 | 73.1 | 94.3 | 97.7 | 548.1 | 68.1 | 90.2 | 95.2 | 52.7 | 80.2 | 88.3 | 474.8 Ours | ✗ | 86.3 | 97.8 | 99.4 | 72.4 | 94.0 | 97.6 | 547.5 | 69.1 | 90.7 | 95.6 | 52.1 | 79.6 | 87.8 | 474.9 Ours† | ✗ | 86.6 | 98.2 | 99.4 | 73.4 | 94.5 | 97.8 | 549.9 | 71.0 | 91.8 | 96.3 | 53.4 | 80.9 | 88.6 | 482.0 Table 1: Recall@K (%) and RSUM on the COCO dataset. Evaluation results on both 1K test setting (average of 5-fold test dataset) and 5K test setting are presented. The best RSUM scores are marked in bold. CA and $\dagger$ indicate models using cross-attention and ensemble models of two hypotheses, respectively. Method | CA | Image-to-text | Text-to-image | RSUM ---|---|---|---|--- R@1 | R@5 | R@10 | R@1 | R@5 | R@10 ResNet-152 + Bi-GRU VSE++ | ✗ | 52.9 | 80.5 | 87.2 | 39.6 | 70.1 | 79.5 | 409.8 PVSE∗ | ✗ | 59.1 | 84.5 | 91.0 | 43.4 | 73.1 | 81.5 | 432.6 PCME∗ | ✗ | 58.5 | 81.4 | 89.3 | 44.3 | 72.7 | 81.9 | 428.1 Ours | ✗ | 61.8 | 85.5 | 91.1 | 46.1 | 74.8 | 83.3 | 442.6 Faster R-CNN + Bi-GRU SCAN† | ✓ | 67.4 | 90.3 | 95.8 | 48.6 | 77.7 | 85.2 | 465.0 VSRN† | ✗ | 71.3 | 90.6 | 96.0 | 54.7 | 81.8 | 88.2 | 482.6 CAAN | ✓ | 70.1 | 91.6 | 97.2 | 52.8 | 79.0 | 87.9 | 478.6 IMRAM† | ✓ | 74.1 | 93.0 | 96.6 | 53.9 | 79.4 | 87.2 | 484.2 SGRAF† | ✓ | 77.8 | 94.1 | 97.4 | 58.5 | 83.0 | 88.8 | 499.6 VSE∞ | ✗ | 76.5 | 94.2 | 97.7 | 56.4 | 83.4 | 89.9 | 498.1 NAAF† | ✓ | 81.9 | 96.1 | 98.3 | 61.0 | 85.3 | 90.6 | 513.2 Ours | ✗ | 77.8 | 94.0 | 97.5 | 57.5 | 84.0 | 90.0 | 500.8 Ours† | ✗ | 80.9 | 94.7 | 97.6 | 59.4 | 85.6 | 91.1 | 509.3 ResNeXt-101 + BERT VSE∞ | ✗ | 88.4 | 98.3 | 99.5 | 74.2 | 93.7 | 96.8 | 550.9 VSE∞† | ✗ | 88.7 | 98.9 | 99.8 | 76.1 | 94.5 | 97.1 | 555.1 Ours | ✗ | 88.8 | 98.5 | 99.6 | 74.3 | 94.0 | 96.7 | 551.9 Ours† | ✗ | 90.6 | 99.0 | 99.6 | 75.9 | 94.7 | 97.3 | 557.1 Table 2: Recall@K(%) and RSUM on the Flickr30K dataset. CA, $\dagger$, and * indicate models using cross-attention, ensemble models of two hypotheses, and models we reproduce, respectively. ### 4.1 Datasets and Evaluation Metric We validate the effectiveness of our method on COCO [33] and Flickr30K [43] datasets. In both datasets, we follow the split proposed by [30]. COCO dataset consists of a train split of 113,287 images, a validation split of 5,000 images, and a test split of 5,000 images. Retrieval results of our method on 1K test setting and 5K test setting are both reported, following [30]. In the 1K test setting, the average retrieval performance on the 5-fold test split is reported. For the Flickr30K dataset, we use 28,000 images for training, 1,000 images for validation, and 1,000 images for testing. We report the retrieval results on the Flick30K using a test split of 1,000 images. In both datasets, each image is given with five matching captions. For evaluation, we use the Recall@$K$, which is the percentage of the queries that have matching samples among top-$K$ retrieval results. Following [28], We also report the RSUM, which is the sum of the Recall@$K$ at $K\in\\{1,5,10\\}$ in the image-to-text and text-to-image retrieval settings. ### 4.2 Implementation Details Feature extractor: For the visual feature extractor, convolutional visual features are obtained by applying $1\times 1$ convolution to the last feature map of CNN. We obtain ROI visual features by feeding the pre-extracted features [1] from a Faster R-CNN [44] to the 2-layer MLP with residual connection, following [28]. In every model, we set $D$ to 1024 and $K$ to 4. Figure 4: For each element of the image embedding set, we present its attention map and the caption nearest to the element in the embedding space. Matching captions are colored in green. Entities corresponding to the attention maps are underlined. Set prediction module: In every experiment, we set $T$ to 4. When using convolutional visual features, $D_{h}$ is set to 1024 and otherwise set to 2048. Similarity and loss function: For smooth-Chamfer similarity, we set scaling parameter $\alpha$ to 16. In $\mathcal{L}_{\mathrm{tri}}$, we use margin $\delta$ of 0.1 for convolutional visual feature and 0.2 for ROI visual feature. We multiply the factor of $0.1$ and $0.01$ to $\mathcal{L}_{\mathrm{div}}$ and $\mathcal{L}_{\mathrm{mmd}}$, respectively. Training: The model is trained with the AdamW optimizer [36]. We construct the input batch with 200 images and all of their captions. When using ROI visual features, the model is trained for 80 epochs with the initial learning rate of 1e-3, which is scheduled with cosine annealing [35]. For the models employing convolutional visual features, we follow the training setting of previous work [47, 28], which will be discussed in detail in the supplementary material. ### 4.3 Comparisons with Other Methods Results on COCO and Flickr30K are summarized in Table 1 and Table 2, respectively. For fair comparisons with previous work, our method is evaluated under three different visual extractors: ResNet-152 [22] pretrained on ImageNet [11], ROI features [1] pre-extracted by Faster R-CNN [44], and ResNeXt-101 [52] pretrained on the Instagram [37] dataset. For ResNet-152 and ResNeXt-101, input image resolution is set to respectively $224\times 224$ and $512\times 512$, following previous work [47, 28]. For the textual feature extractor, we use either bi-GRU or BERT. The ensemble results are obtained by averaging the similarity scores of two models trained with different random seeds. Our method outperforms previous methods in terms of RSUM in every setting except for Flickr30K with ROI visual features. Even in this setting, ours achieves the second-best. Note that NAAF, currently the best in this setting, requires two orders of magnitude heavier computations than ours at inference since it relies on cross-attention, as will be shown in the next section. The ensemble of ours with ResNeXt-101 and BERT clearly outperforms all existing records, improving the previous best RSUM by $7.2\%$p and $2.0\%$p on COCO 5K and Flickr30K. Comparisons with PVSE and PCME demonstrate the superiority of our set-based embedding framework. Our model outperforms both of them in every case. Specifically, our model improves RSUM by $3.3\%$p on COCO 5K and $10.0\%$p on Flickr30K, compared to PVSE. Also, our method shows better results than models involving cross-attention networks [31, 6, 14], which impose substantial computation as noted. We finally emphasize that ours using a single model often outperforms previous work using ensemble [31, 32, 6, 14]. ### 4.4 Computation Cost Analysis To demonstrate the efficiency of our model during evaluation time, we measure FLOPs required for computing a similarity score between representations of an image and a caption. Ours demands $16.4$K FLOPs, while SCAN [31], a representative cross-attention method, requires $1.24$M FLOPs. This result shows that our method demands about $80$ times less floating-point operations than SCAN since the cross-attention requires re-processing of the image or caption representations for each query. Also, it is worth noting that our method outperforms SCAN in terms of RSUM, as shown in Table 1 and 2. ### 4.5 Analysis of Attention In Figure 4, attentions from the last aggregation block in the $f^{\mathcal{V}}$ are visualized. Each attention is used to encode individual elements of the image embedding set. For each element, its nearest caption in the embedding space is presented together. Thanks to slot attention, the set prediction module produces heterogeneous attention maps that focus on individual objects or contexts, which enable retrieving diverse captions. In particular, on the top-right, elements are matched with the captions that describe different semantics (helmet, children, park, and bicycle). Interestingly, we observed that one of the element slots attends to regions left out after other slots capture distinctive semantics, which wholly charges for retrieving related but highly abstract captions. Similarity | Arch. | RSUM ---|---|--- MIL | Ours | 491.7 MP | Ours | 490.5 Ours (Chamfer) | Ours | 499.6 Ours (S-Chamfer) | PIE-Net | 483.3 Ours (S-Chamfer) | Ours | 500.8 Setting | $\log$(Var.) | RSUM ---|---|--- PIE-Net | -7.35 | 483.3 Ours \w MP | -5.27 | 490.5 Transformer | -2.27 | 496.1 Ours | -2.13 | 500.8 Table 3: Ablation studies of the proposed similarity function and set prediction module. ### 4.6 Ablation Study We perform ablation studies to investigate the contributions of the proposed similarity function and set prediction module. For all ablation experiments, we use the models employing ROI visual features. Importance of set similarity function: In the left of Table 3, we ablate the proposed similarity function and replace it with the MP, MIL, and Chamfer similarity. The results show that the models trained with smooth-Chamfer similarity surpass those trained with MIL or MP similarity under the same set prediction architecture. Though Chamfer similarity improves the performance by resolving set collapsing, smooth-Chamfer similarity further improves performance by providing dense supervision during training. Importance of set prediction module: We also ablate the proposed set prediction module and substitute it with the PIE-Net, which is the baseline set prediction architecture of the previous set-based embedding methods, PVSE and PCME. The results in the left of Table 3 suggest that the impact of our set prediction module is significant. Verification of set collapsing: In the right of Table 3, we report the average circular variance of embedding set. Formally, the circular variance of embedding set $\mathbf{S}$ is denoted as $1-\|\sum_{x\in\mathbf{S}}x/|\mathbf{S}|\|_{2}$. As discussed earlier, results show that employing PIE-Net or MP leads to set collapsing. Moreover, we report the result when aggregation block is replaced to transformer. This variant results in a 13$\%$ lower variance on a linear scale and inferior performance since it does not guarantee disentanglement between elements. $K$ | 1 | 2 | 3 | 4 | 5 | 6 ---|---|---|---|---|---|--- RSUM | 492.6 | 495.5 | 497.4 | 500.8 | 498.4 | 499.3 $T$ | 1 | 2 | 3 | 4 | 5 | 6 ---|---|---|---|---|---|--- RSUM | 492.6 | 497.8 | 499.3 | 500.8 | 499.4 | 499.1 $\alpha$ | 1 | 2 | 4 | 8 | 16 | 32 | 64 ---|---|---|---|---|---|---|--- RSUM | 495.2 | 497.2 | 497.7 | 498.9 | 500.8 | 499.3 | 499.0 Table 4: Impact of hyperparameters for the proposed similarity function and set prediction module. ### 4.7 Impact of Hyperparameters In Table 4, we report RSUM of the model while varying the cardinality of embedding sets $K$, the total number of iterations $T$, and the scaling parameter $\alpha$. Results demonstrate that the model shows consistently high RSUM when $K>3$, $T>2$, and $\alpha>4$. Specifically, we notice that when $K=1$, which is identical to using a single embedding vector, the model shows significantly lower accuracy compared to others. It suggests that embedding sets enable a more accurate retrieval by addressing the semantic ambiguity. When $T=1$, we observe substantial degradation of accuracy, which underpins that progressive refinement done by multiple aggregation blocks helps element slots aggregate semantic entities, as shown in Figure 2. ### 4.8 Analysis of Embedding Set Elements For further analysis on the embedding set elements, we use only one of the elements during the evaluation of the model. Let $\mathbf{S}(i)$ be an element of $\mathbf{S}$ produced by the $i$-th element slot. Table 5 summarizes the accuracy of the model in terms of RSUM when elements of $\mathbf{S}^{\mathcal{V}}$ and $\mathbf{S}^{\mathcal{T}}$ are ablated during evaluation. Results present that using only one of the elements degrades accuracy. Specifically, using only $\mathbf{S}^{\mathcal{V}}(2)$ during evaluation leads to far lower accuracy compared to using the other elements. To check whether $\mathbf{S}^{\mathcal{V}}(2)$ is just a noisy element that hinders final performance, we also report RSUM when only $\mathbf{S}^{\mathcal{V}}(2)$ is ablated. However, interestingly, we still observed the degraded accuracy compared to using a complete embedding set. Though one of the elements often retrieves non-matching samples, the results demonstrate that using them together during evaluation helps the model find accurate matching samples. We hypothesize that $\mathbf{S}^{\mathcal{V}}(2)$ is trained to capture highly ambiguous semantics. Therefore, using them alone during evaluation leads to less accurate retrievals, while it improves performance when used together by successfully representing ambiguous situations and context. A similar tendency could be observed in Figure 4, where one of the slots often encodes an element that is located nearby to a related but highly ambiguous caption in the embedding space. Evaluation | ---|--- $\mathbf{S}^{\mathcal{V}}(1)$ | $\mathbf{S}^{\mathcal{V}}(2)$ | $\mathbf{S}^{\mathcal{V}}(3)$ | $\mathbf{S}^{\mathcal{V}}(4)$ | RSUM ✓ | ✓ | ✓ | ✓ | 500.8 ✓ | | | | 491.1 | ✓ | | | 309.6 | | ✓ | | 484.9 | | | ✓ | 486.0 ✓ | | ✓ | ✓ | 500.2 Evaluation | ---|--- $\mathbf{S}^{\mathcal{T}}(1)$ | $\mathbf{S}^{\mathcal{T}}(2)$ | $\mathbf{S}^{\mathcal{T}}(3)$ | $\mathbf{S}^{\mathcal{T}}(4)$ | RSUM ✓ | ✓ | ✓ | ✓ | 500.8 ✓ | | | | 481.9 | ✓ | | | 483.0 | | ✓ | | 481.7 | | | ✓ | 497.2 Table 5: RSUM on the Flickr30K dataset when elements of the $\mathbf{S}^{\mathcal{V}}$(left) and $\mathbf{S}^{\mathcal{T}}$(right) are ablated during evaluation. ## 5 Conclusion We propose the novel set embedding framework for cross-modal retrieval, consisting of the set prediction module and smooth-Chamfer similarity. The proposed set prediction module outputs an embedding set that successfully captures ambiguity, while smooth-Chamfer similarity successfully resolves undesirable effects of the existing similarity functions. As a result, our model surpasses most of the previous methods, including methods involving higher computation costs, on COCO and Flickr30K datasets. We will extend our model for the different modalities and tasks in the future. ## References * [1] Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. Bottom-up and top-down attention for image captioning and visual question answering. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018. * [2] Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. Layer normalization. arXiv preprint arXiv:1607.06450, 2016. * [3] Max Bain, Arsha Nagrani, Gül Varol, and Andrew Zisserman. Frozen in time: A joint video and image encoder for end-to-end retrieval. In Proc. IEEE International Conference on Computer Vision (ICCV), 2021. * [4] Michael Chang, Thomas L. Griffiths, and Sergey Levine. Object representations as fixed points: Training iterative inference algorithms with implicit differentiation. In ICLR Workshop on Deep Generative Models for Highly Structured Data, 2022. * [5] Gal Chechik, Eugene Ie, Martin Rehn, Samy Bengio, and Dick Lyon. Large-scale content-based audio retrieval from text queries. In Proceedings of the 1st ACM international conference on Multimedia information retrieval, 2008. * [6] Hui Chen, Guiguang Ding, Xudong Liu, Zijia Lin, Ji Liu, and Jungong Han. Imram: Iterative matching with recurrent attention memory for cross-modal image-text retrieval. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. * [7] Shizhe Chen, Yida Zhao, Qin Jin, and Qi Wu. Fine-grained video-text retrieval with hierarchical graph reasoning. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. * [8] Kyunghyun Cho, Bart Van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. Learning phrase representations using rnn encoder-decoder for statistical machine translation. In Proc. Empirical Methods in Natural Language Processing (EMNLP), 2014. * [9] S. Chopra, R. Hadsell, and Y. LeCun. Learning a similarity metric discriminatively, with application to face verification. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2005. * [10] Sanghyuk Chun, Seong Joon Oh, Rafael Sampaio De Rezende, Yannis Kalantidis, and Diane Larlus. Probabilistic embeddings for cross-modal retrieval. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. * [11] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. ImageNet: a large-scale hierarchical image database. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2009. * [12] Karan Desai and Justin Johnson. Virtex: Learning visual representations from textual annotations. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. * [13] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proc. North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019. * [14] Haiwen Diao, Ying Zhang, Lin Ma, and Huchuan Lu. Similarity reasoning and filtration for image-text matching. In Proc. AAAI Conference on Artificial Intelligence (AAAI), 2021\. * [15] Thomas G. Dietterich, Richard H. Lathrop, and Tomás Lozano-Pérez. Solving the multiple instance problem with axis-parallel rectangles. Artificial Intelligence, 89(1-2):31–71, 1997. * [16] Benjamin Elizalde, Shuayb Zarar, and Bhiksha Raj. Cross modal audio search and retrieval with joint embeddings based on text and audio. In ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019. * [17] Fartash Faghri, David J Fleet, Jamie Ryan Kiros, and Sanja Fidler. Vse++: Improving visual-semantic embeddings with hard negatives. In Proc. British Machine Vision Conference (BMVC), 2018. * [18] Andrea Frome, Greg Corrado, Jonathon Shlens, Samy Bengio, Jeffrey Dean, Marc’Aurelio Ranzato, and Tomas Mikolov. Devise: A deep visual-semantic embedding model. In Proc. Neural Information Processing Systems (NeurIPS), 2013. * [19] Simon Ging, Mohammadreza Zolfaghari, Hamed Pirsiavash, and Thomas Brox. Coot: Cooperative hierarchical transformer for video-text representation learning. In Proc. Neural Information Processing Systems (NeurIPS), 2020. * [20] Arthur Gretton, Karsten Borgwardt, Malte Rasch, Bernhard Schölkopf, and Alex Smola. A kernel method for the two-sample-problem. In Proc. Neural Information Processing Systems (NeurIPS), 2006. * [21] Jiuxiang Gu, Jianfei Cai, Shafiq R Joty, Li Niu, and Gang Wang. Look, imagine and match: Improving textual-visual cross-modal retrieval with generative models. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 7181–7189, 2018. * [22] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016. * [23] Dan Hendrycks and Kevin Gimpel. Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415, 2016. * [24] Yan Huang, Qi Wu, Chunfeng Song, and Liang Wang. Learning semantic concepts and order for image and sentence matching. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018. * [25] Zhicheng Huang, Zhaoyang Zeng, Bei Liu, Dongmei Fu, and Jianlong Fu. Pixel-bert: Aligning image pixels with text by deep multi-modal transformers. In arXiv preprint arXiv:2004.00849, 2020. * [26] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proc. International Conference on Machine Learning (ICML), 2015\. * [27] Boseung Jeong, Jicheol Park, and Suha Kwak. Asmr: Learning attribute-based person search with adaptive semantic margin regularizer. In Proc. IEEE International Conference on Computer Vision (ICCV), pages 12016–12025, 2021. * [28] Chen Jiacheng, Hu Hexiang, Wu Hao, Jiang Yuning, and Wang Changhu. Learning the best pooling strategy for visual semantic embedding. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. * [29] Justin Johnson, Bharath Hariharan, Laurens Van Der Maaten, Li Fei-Fei, C Lawrence Zitnick, and Ross Girshick. Clevr: A diagnostic dataset for compositional language and elementary visual reasoning. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017. * [30] Andrej Karpathy and Li Fei-Fei. Deep visual-semantic alignments for generating image descriptions. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015. * [31] Kuang-Huei Lee, Xi Chen, Gang Hua, Houdong Hu, and Xiaodong He. Stacked cross attention for image-text matching. In Proc. European Conference on Computer Vision (ECCV), 2018. * [32] Kunpeng Li, Yulun Zhang, Kai Li, Yuanyuan Li, and Yun Fu. Visual semantic reasoning for image-text matching. In Proc. IEEE International Conference on Computer Vision (ICCV), 2019. * [33] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft COCO: common objects in context. In Proc. European Conference on Computer Vision (ECCV), 2014. * [34] Francesco Locatello, Dirk Weissenborn, Thomas Unterthiner, Aravindh Mahendran, Georg Heigold, Jakob Uszkoreit, Alexey Dosovitskiy, and Thomas Kipf. Object-centric learning with slot attention. In Proc. Neural Information Processing Systems (NeurIPS), 2020. * [35] Ilya Loshchilov and Frank Hutter. Sgdr: Stochastic gradient descent with warm restarts. In Proc. International Conference on Learning Representations (ICLR), 2017. * [36] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In Proc. International Conference on Learning Representations (ICLR), 2019. * [37] Dhruv Mahajan, Ross Girshick, Vignesh Ramanathan, Kaiming He, Manohar Paluri, Yixuan Li, Ashwin Bharambe, and Laurens Van Der Maaten. Exploring the limits of weakly supervised pretraining. In Proc. European Conference on Computer Vision (ECCV), 2018. * [38] Antoine Miech, Jean-Baptiste Alayrac, Ivan Laptev, Josef Sivic, and Andrew Zisserman. Thinking fast and slow: Efficient text-to-visual retrieval with transformers. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. * [39] Hyeonseob Nam, Jung-Woo Ha, and Jeonghee Kim. Dual attention networks for multimodal reasoning and matching. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017. * [40] Seong Joon Oh, Kevin Murphy, Jiyan Pan, Joseph Roth, Florian Schroff, and Andrew Gallagher. Modeling uncertainty with hedged instance embedding. arXiv preprint arXiv:1810.00319, 2018. * [41] Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer. Automatic differentiation in pytorch. In AutoDiff, NIPS Workshop, 2017. * [42] Jeffrey Pennington, Richard Socher, and Christopher D Manning. Glove: Global vectors for word representation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), 2014. * [43] Bryan A. Plummer, Liwei Wang, Chris M. Cervantes, Juan C. Caicedo, Julia Hockenmaier, and Svetlana Lazebnik. Flickr30k entities: Collecting region-to-phrase correspondences for richer image-to-sentence models. In Proc. IEEE International Conference on Computer Vision (ICCV), 2015. * [44] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proc. Neural Information Processing Systems (NeurIPS), pages 91–99, 2015. * [45] Florian Schroff, Dmitry Kalenichenko, and James Philbin. FaceNet: A unified embedding for face recognition and clustering. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015. * [46] Yichun Shi and Anil K Jain. Probabilistic face embeddings. In Proc. IEEE International Conference on Computer Vision (ICCV), 2019. * [47] Yale Song and Mohammad Soleymani. Polysemous visual-semantic embedding for cross-modal retrieval. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019. * [48] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Proc. Neural Information Processing Systems (NeurIPS), 2017. * [49] Luke Vilnis and Andrew McCallum. Word representations via gaussian embedding. In Proc. International Conference on Learning Representations (ICLR), 2015. * [50] Zihao Wang, Xihui Liu, Hongsheng Li, Lu Sheng, Junjie Yan, Xiaogang Wang, and Jing Shao. Camp: Cross-modal adaptive message passing for text-image retrieval. In Proc. IEEE International Conference on Computer Vision (ICCV), 2019. * [51] Jiwei Wei, Xing Xu, Yang Yang, Yanli Ji, Zheng Wang, and Heng Tao Shen. Universal weighting metric learning for cross-modal matching. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. * [52] Saining Xie, Ross Girshick, Piotr Dollár, Zhuowen Tu, and Kaiming He. Aggregated residual transformations for deep neural networks. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017. * [53] Kun Zhang, Zhendong Mao, Quan Wang, and Yongdong Zhang. Negative-aware attention framework for image-text matching. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2022. * [54] Qi Zhang, Zhen Lei, Zhaoxiang Zhang, and Stan Z Li. Context-aware attention network for image-text retrieval. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 3536–3545, 2020. Appendix ## Appendix A Implementation details Our model is implemented with PyTorch [41] v 1.8.1. Automatic mixed precision is used for faster and more efficient training. Additional to implementation details provided in the main paper, settings for training differ by which feature extractors are used. ResNet-152 + bi-GRU: In this setting, the model is trained for 120 epochs, and the learning rate for the set prediction module is scaled with $0.1$ on both datasets. The learning rate decays by a multiplicative factor of $0.1$ for every 10 epochs. Following [47], the entire model is trained in an end-to-end manner, but CNN is fixed for the first 50 epochs. Training is done with a single RTX 3090 GPU. Faster-RCNN + bi-GRU: The learning rate for the set prediction module is scaled with $0.1$ and $0.05$ on COCO and Flickr30K, respectively. Following [28], size augmentation is used, which drops 20% of ROI features and words during training. Training is done with a single RTX 3090 GPU. ResNeXt-101 + BERT: In this setting, we construct a batch with 128 images and their entire matching captions. The learning rate for the set prediction module is scaled with $0.1$. The model is trained for 50 epochs, where the learning rate decays by a multiplicative factor of $0.1$ for every 20 epochs. Following [28], the learning rates for CNN and BERT are scaled with $0.01$ and $0.1$, respectively. The statistics of the batch normalization [26] layer are fixed during training. The entire model is trained in an end-to-end manner except for CNN in the first epoch. In the first epoch, triplet loss without mining is used, whereas the hardest negative mining is used for later epochs. Training is done with two A100 PCIe GPU. ## Appendix B Additional qualitative results In Figure A1 and Figure A2, we present additional visualization of attention map from the visual set prediction module $f^{\mathcal{V}}$. Visualizations of attention maps, including ones presented in the main paper, are obtained from the model using ResNeXt + BERT feature extractors. Attention maps from each iteration are presented together, where $t=4$ is the last iteration. For each attention map from the last iteration, its corresponding element of embedding set and nearest caption are provided together. Results show that the aggregation block produces heterogeneous attention maps capturing various semantics, such as different objects (1st row of Figure A1) and action (2nd row of Figure A2). Moreover, in every case, we can observe that element slots are progressively updated to capture distinctive semantics, starting from sparse and noisy attention maps. Specifically, in Figure A2, we present the examples where the nearest captions of multiple elements are the same. For instance, in the 2nd row of Figure A2, each element attends to individual entities (sky, larger giraffe, grassy area, and baby giraffe), but their nearest captions, which describe the entire scene, are the same. Results imply that by fusing element slots with the global feature, elements of the embedding set can preserve the global context while focusing on distinctive semantics. These characteristics help model when samples with little ambiguity are given, such as a caption describing the entire scene (2nd row of Figure A2) or an image containing a single iconic entity (3rd row of Figure A2). Figure A1: For each element of the image embedding set, we present its attention map and the caption nearest to the element in the embedding space. Matching captions are colored in green. Entities corresponding to the attention maps are underlined. Figure A2: For each element of the image embedding set, we present its attention map and the caption nearest to the element in the embedding space. Matching captions are colored in green. Entities corresponding to the attention maps are underlined.
# GeoUDF: Surface Reconstruction from 3D Point Clouds via Geometry-guided Distance Representation Siyu Ren1,2 Junhui Hou1 Xiaodong Chen2 Ying He3 Wenping Wang4 1City University of Hong Kong 2Tianjin University 3 Nanyang Technological University 4Texas A&M University <EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS> <EMAIL_ADDRESS><EMAIL_ADDRESS>Corresponding author. This work was supported by the Hong Kong Research Grants Council under Grants 11202320 and 11219422. ###### Abstract The recent neural implicit representation-based methods have greatly advanced the state of the art for solving the long-standing and challenging problem of reconstructing a discrete surface from a sparse point cloud. These methods generally learn either a binary occupancy or signed/unsigned distance field (SDF/UDF) as surface representation. However, existing SDF/UDF-based methods usually use neural networks to implicitly regress the distance in a purely data-driven manner, thus limiting the accuracy and generalizability to some extent. In contrast, we propose a geometry-guided method for UDF and its gradient estimation that explicitly formulates the unsigned distance of a query point as the learnable affine averaging of its distances to the tangent planes of neighbouring points. Besides, we model the local geometric structure of the input point clouds by explicitly learning a quadratic polynomial for each point. This not only facilitates upsampling the input sparse point cloud but also naturally induces unoriented normal, which further augments UDF estimation. Finally, to extract triangle meshes from the predicted UDF we propose a customized edge-based marching cube module. We conduct extensive experiments and ablation studies to demonstrate the significant advantages of our method over state-of-the-art methods in terms of reconstruction accuracy, efficiency, and generalizability. The source code is publicly available at https://github.com/rsy6318/GeoUDF. ## 1 Introduction Surface reconstruction from point clouds is a fundamental and challenging problem in 3D vision, graphics and robotics [17, 26, 9]. Traditional approaches, such as Poisson surface reconstruction [18], compute an occupancy or signed distance field (SDF) by solving Poisson’s equation, resulting in a sparse line system. Then they utilize the Marching Cubes (MC) [22] algorithm to extract iso-surfaces as the reconstructed meshes. These methods are computationally efficient, scalable, resilient to noisy data and tolerant to registration artifacts. However, they work only for oriented point samples. Recent deep learning methods can learn implicit fields, such as binary occupancy fields (BOFs) and SDFs, directly from raw point clouds, making them an ideal tool for real-world scans, of which orientation is often unavailable. In spite of many advantages, BOF and SDF-based approaches can only deal with watertight and manifold surfaces, lacking the ability to represent more general ones, such as those with open boundaries or interior structures. In contrast, the unsigned distance field (UDF) is more powerful in terms of representation ability, and can overcome the above-mentioned limitations of BOFs and SDFs. The existing deep UDF-based methods, such as [11, 38, 16], employ neural networks to regress UDFs from input point clouds. Due to their reliance on the training data, these methods have limited generalizability and cannot deal with unseen objects well. Another technical challenge diminishing UDFs’ practical usage is that one cannot use the conventional MC to extract iso-surfaces directly because UDFs do not have zero-crossings and thus do not provide the required sign-change information of inside or outside. Some methods [16, 40] locally convert a UDF to an SDF in a cubic cell and then apply MC to extract the surface. However, they often fail at corners or edges. Recently, GIFS [38] modifies MC by predicting whether the surface interacts with each cube edge via a neural network. In this paper, we propose GeoUDF, a new learning-based framework reconstructing the underlying surface of a sparse point cloud. We particularly aim at addressing two fundamental issues of UDFs, namely, (1) how to predict an accurate UDF and its gradients for arbitrary 3D inputs? and (2) how to extract triangle meshes, given a UDF? More importantly, the constructed methods should be well generalized to unseen data. To solve the first issue, we propose a geometry-guided learning module, adaptively approximating the UDF and its gradient of a point cloud by leveraging its local geometry in a decoupled manner. As for the second issue, we propose an edge-based MC algorithm, extracting triangle mesh from any UDF according to the intersection relationship between the surface and the connections of any two vertices of the cubes. In addition, we propose to enhance input point cloud quality by explicitly and concisely modeling its local geometry structure, which is beneficial for surface reconstruction. The elegant formulations distinguish our GeoUDF from existing methods, making it more lightweight, efficient, and accurate and better generalizability. Besides, the proposed modules contained in GeoUDF can be independently used as a general method in its own right. Extensive experiments on various datasets show the significant superiority of our method over state-of-the-art methods. In summary, the main contributions of this paper are three-fold: * • novel accurate, compact, efficient, and explainable learning-based UDF and its gradient estimation methods that are decoupled; * • a concise yet effective learning-based point cloud upsampling and representation method; and * • a general method for extracting triangle meshes from any unsigned distance field. ## 2 Related Work Figure 1: The pipeline of GeoUDF, a lightweight, explainable, and efficient learning-based surface reconstruction framework that produces 3D meshes with high accuracy and strong generalizability. Learning Implicit Surface Representation. With the development of deep learning, many advances have been made for the implicit representation of 3D shapes in recent years. For example, Binary Occupancy Field (BOF) casts the problem into binary classification, i.e., any point in the 3D space can be classified as inside or outside of a closed shape. ONet [25] uses a latent vector to represent the whole input, then uses a decoder to get the occupancy of a query point. Subsequently, IF-NET [10] and CONet [29] adopt convolutions on voxelized features, which increase the reconstruction accuracy. Referring to SPSR [18], SAP [28] introduces a differentiable formulation of SPSR and then incorporates it into a neural network to reconstruct the surfaces from point clouds. Recently, POCO [5] designs an attention-based convolution and further improves the reconstruction quality. SDFs, distinguishing interior and exterior by assigning a sign to the distance, are a more accurate representation. DeepSDF [27] uses an auto-decoder to optimize the latent vector in order to refine the SDF. Based on DeepSDF, DeepLS [7] uses a grid to store the independent latent codes, each only responsible for a local area. Neural-Pull [1] and OSP [23] concentrate on the differential property and use the gradient of the SDF to pull the point onto the surface, where they estimate and refine the SDF. Some traditional methods can also be adapted through neural networks, such as DeepIMLS [21] and DOG [34], which incorporate implicit moving least-squares (IMLS) [19] into neural networks to estimate the SDF of a point cloud. BOF and SDF can only deal with the watertight shapes because they assume a surfacee that partition the whole space into inside and outside, making it impossible to represent general shapes, such as those with open surface with boundaries or non-manifold surfaces. UDF can represent general shapes since it represents the (unsigned) distance from a query point to the surface. NDF [11] and GIFS [38] use 3D convolution to regress the UDF from the voxelized point cloud. CAP-UDF [40] employs a field consistency constraint to get consistency- aware UDF. To our best knowledge, there are no methods learning a UDF from geometric information explicitly. Surface Extraction from Implicit Fields. After learning implicit fields, such as BOFs or SDFs, the MC algorithm [22] is commonly used to obtain a triangle mesh. However, MC cannot be applied to UDFs because there is no inside/outside information. NDF [11] uses the gradient of the UDF to generate a much denser point cloud and then uses ball-pivoting algorithm [3] to extract the triangle mesh, which inefficient with poor surface quality. GIFS [38] not only predicts the UDF, but also determines whether any two vertices of the cubes are separated by the surface, which may be regarded as a binary classification. Then it modifies the MC to extract a triangle mesh from the UDF. Note that the Marching Cubes method modified by GIFS cannot be applied to any UDF because it needs an extra network to achieve do the binary classification on each edge of the cube. MeshUDF [16] uses one vertex of the cube as a reference and computes the inner product of the gradients of UDF with the other seven points, thus converting UDF to SDF in this cube according to the sign of the inner product. But such a criterion is not robust enough and often predicts wrong faces at the edges or corners. Learning-based Point Cloud Upsampling. Point cloud upsampling (PU) aims to densify sparse point clouds to recover more geometry details, which will be beneficial to downstream surface reconstruction. As the first learning-based method for point cloud upsampling, PU-Net [39] employs PointNet++ [30] to extract features and then expands the features through multi-branch MLPs to achieve the upsampling. Furthermore, the generative adversarial network-based PU-GAN [20] achieves impressive results on non-uniform data, but the network is difficult to train. PU-GCN [31], a graph convolutional network-based approach, introduces the NodeShuffle module into the existing upsampling pipeline. The first geometry-centric method PUGeo-Net [32] links deep learning with differentiable geometry to learn the local geometry structure of point clouds. MPU [36] appends a 1D code $\\{-1,~{}1\\}$ to separate replicated the point-wise features. Built upon the linear approximation theorem, MAFU [33] adaptively learns the interpolation weights in a flexible manner, as well as high-order approximation errors. Alternatively, PU-Flow [24] conducts learnable interpolation in feature space regularized by normalizing flows. Recently, NP [15] adopts a neural network to represent the local geometric shape in a continuous manner, embedding more shape information. ## 3 Proposed Method Overview. As shown in Fig. 1, our GeoUDF consists of three modules, i.e., local geometry representation (LGR), geometry-guided UDF estimation (GUE), and edge-based marching cube (E-MC). Specifically, given a sparse 3D point cloud, we first model its local geometry through LGR, producing a dense point cloud associated with un-oriented normal vectors. Then, we predict the unsigned distance field of the resulting dense point cloud via GUE from which we customize an E-MC module to extract the triangle mesh for the zero level set. Note that our GeoUDF works for both watertight and open surfaces with interior structures. Besides, each of the three modules can be independently used as a general method in its own right. In what follows, we will detail each module. ### 3.1 Local Geometry Representation Problem Formulation. Let $\mathcal{P}=\\{\mathbf{p}_{i}|\mathbf{p}_{i}\in\mathbb{R}^{3}\\}_{i=1}^{N}$ be an input sparse point cloud of $N$ points sampled from a surface $\mathcal{S}$ to be reconstructed. Based on differential geometry theory of surfaces [14], the local neighborhood of a regular point of a surface is completely determined by the first and second fundamental forms. Therefore we use a quadratic polynomial surface to approximate the local surface centered at each point: $f_{i}(\mathbf{u})=\mathbf{p}_{i}+\mathbf{A}_{i}\texttt{E}(\mathbf{u}),\vspace{-0.15cm}$ (1) where $\mathbf{u}=[u_{1},u_{2}]^{\mathsf{T}}\in\mathbb{R}^{2}$ is the coordinate in the 2D local parameter domain, $\texttt{E}(\mathbf{u}):=[1\ u_{1}\ u_{2}\ u_{1}^{2}\ u_{1}u_{2}\ u_{2}^{2}]^{\mathsf{T}}\in\mathbb{R}^{6}$, and $\mathbf{A}_{i}\in\mathbb{R}^{3\times 6}$ is the coefficient matrix. With this representation, we can perform the following two operations, which will bring benefits to the subsequent module. (1) Densifying $\mathcal{P}$. For each point $\mathbf{p}_{i}\in\mathcal{P}$, we can uniformly sample $M$ 2D coordinates from a pre-defined local paramterization $\mathcal{D}=[-\delta,\delta]^{2}\subset\mathbb{R}^{2}$, which are then substituted into Eq. (1) to generate additional $M$ 3D points around $\mathbf{p}_{i}$, producing a dense point cloud $\mathcal{P}_{M}=\\{\mathbf{p}_{j}|\mathbf{p}_{j}\in\mathbb{R}^{3}\\}_{j=1}^{NM}$. (2) Inducing un-oriented normal vectors of $\mathcal{P}_{M}$. From the explicit parameterization function in Eq. (1), we can calculate the un- oriented normal vector for each point of $\mathcal{P}_{M}$, denoted as $\mathcal{N}_{M}=\\{\mathbf{n}_{j}|\mathbf{n}_{j}\in\mathbb{R}^{3},\|\mathbf{n}_{j}\|=1\\}$, based on the differential geometry property [2]. We refer the readers to the supplementary material for details. Learning-based Implementation. As depicted in Fig. 2, we design a sub-network to realize this representation process in a data-driven manner, which is boiled down to predicting the coefficient matrix $\\{\mathbf{A}_{i}\\}_{i=1}^{N}$. Specifically, we adopt 3-layer EdgeConvs [35], which can capture structural information of point cloud data, to embed $\mathcal{P}$ into a high-dimensional feature space, generating point-wise features $\\{\mathbf{c}^{(l)}_{i}\in\mathbb{R}^{d_{1}}\\}_{i=1}^{N}$ at the $l$-th layer. Then, we concatenate the features of three layers and feed them into an MLP to predict 18-dimension vectors, which are further reshaped into $\\{\mathbf{A}_{i}\\}_{i=1}^{N}$. Besides, in Sec. 4.3, we experimentally demonstrate the advantages of this concise operation over state-of-the-art point cloud upsampling methods. Figure 2: Flowchart of the LGR module. Note that the 2D coordinates are randomly sampled from $\mathcal{D}$ via PDS [6] at each iteration. We refer readers to the supplementary material for the detailed network architecture. ### 3.2 Geometry-guided UDF Estimation This module aims to estimate the unsigned distance field of $\mathcal{P}_{M}$, where the iso-surface with the value of zero indicates the surface. As aforementioned, existing learning-based UDF estimation methods, such as NDF [11] and GIFS [38], utilize a neural network to implicitly regress the UDFs from point clouds, thus limiting their accuracy. Besides, the trained models often yield poor results on unseen data. In sharp contrast to the regression- based methods, GUE leverages the inherent geometric property of input point clouds, leading to a more accurate UDF estimation method with better generalizability. (a) (b) (c) (d) Figure 3: Illustration of the UDF and its gradient estimation processes. (a) Un-oriented normal vectors. (b) Aligned normal vectors. (c) UDF estimation. (d) UDF gradient estimation. Formulation of UDF and its Gradient Estimation. Given a query point $\mathbf{q}\in\mathbb{R}^{3}$, we can find its $K$ nearest points from $\mathcal{P}_{M}$ in a Euclidean distance sense, denoted as $\Omega(\mathbf{q})=\\{\mathbf{p}_{k}|\mathbf{p}_{k}\in\mathcal{P}_{R}\\}_{k=1}^{K}$. Denote by $\mathbf{n}_{k}$ the un-oriented normal of $\mathbf{p}_{k}$. Before calculating the UDF, as shown in Fig. 3(b), we first align $\\{\mathbf{n}_{k}\\}_{k=1}^{K}$ with reference to $\mathbf{q}$ via $\mathbf{\tilde{n}}_{k}(\mathbf{q})=\texttt{sgn}\left(\langle\mathbf{n}_{k},\mathbf{\tilde{p}}_{k}\rangle\right)\mathbf{n}_{k},\vspace{-0.15cm}$ (2) where $\mathbf{\tilde{p}}_{k}:=\mathbf{q}-\mathbf{p}_{k}$, $\langle\cdot,\cdot\rangle$ computes the inner product of two vectors, and $\texttt{sgn}(\cdot)$ extracts the sign of the input. Eq. (2) makes $\mathbf{\tilde{n}}_{k}$ align with $\mathbf{\tilde{p}}_{k}$, i.e., $\langle\mathbf{\tilde{n}}_{k},\mathbf{\tilde{p}}_{k}\rangle>0$. Denote by $\mathbf{p}^{*}$ the point on surface $\mathcal{S}$ closest to $\mathbf{q}$, and $\mathbf{\tilde{n}}^{*}$ its aligned normal vector. Assume the points of $\Omega(\mathbf{q})$ are exactly distributed on $\mathcal{S}$, and $\\{\tilde{\mathbf{n}}_{k}\\}_{k=1}^{K}$ are accurate, i.e, they are perpendicular to $\mathcal{S}$. As shown in Fig. 3(c), as $\Omega(\mathbf{q})$ generally covers a small region/area, $\mathbf{p}^{*}$ should be close to the tangent plane of any $\mathbf{p}_{k}$. Based on this observation, we propose to approximate the UDF of $\mathbf{q}$, denoted as $\texttt{U}(\mathbf{q})$, using the weighted sum of the distances from $\mathbf{q}$ to the tangent planes: $\texttt{U}(\mathbf{q})\approx\Phi(\mathbf{q}):=\sum_{\mathbf{p}_{k}\in\Omega(\mathbf{q})}w_{1}(\mathbf{q},\mathbf{p}_{k})\cdot\langle\mathbf{\tilde{n}}_{k},\mathbf{\tilde{p}}_{k}\rangle,\vspace{-0.2cm}$ (3) where the weights $\\{w_{1}(\mathbf{q},\mathbf{p}_{k})\\}_{k=1}^{K}$ are non- negative and satisfy partition of unity $\sum_{k=1}^{K}w_{1}(\mathbf{q},\mathbf{p}_{k})=1$. Figure 4: Flowchart of the GUE module. We refer readers to the supplementary material for the detailed network architecture. Similarly, we propose to approximate the gradient of $\texttt{U}(\mathbf{q})$, denoted as $\nabla\texttt{U}(\mathbf{q})$, using the continuity property of the normal vectors of a surface and the fact $\nabla\texttt{U}(\mathbf{q})=\mathbf{\tilde{n}}^{*}$. See Fig. 3(d). We use $\mathbf{\tilde{n}}_{k}$, each of which is around $\mathbf{\tilde{n}}^{*}$, to approximate $\nabla\texttt{U}(\mathbf{q})$ as $\nabla\texttt{U}(\mathbf{q})\approx\Theta(\mathbf{q}):=\frac{\sum_{\mathbf{p}_{k}\in\Omega(\mathbf{q})}w_{2}(\mathbf{q},\mathbf{p}_{k})\cdot\tilde{\mathbf{n}}_{k}}{\|\sum_{\mathbf{p}_{k}\in\Omega(\mathbf{q})}w_{2}(\mathbf{q},\mathbf{p}_{k})\cdot\tilde{\mathbf{n}}_{k}\|},\vspace{-0.15cm}$ (4) where the weights $\\{w_{2}(\mathbf{q},\mathbf{p}_{k})\\}_{k=1}^{K}$ are non- negative and satisfy $\sum_{k=1}^{K}w_{2}(\mathbf{q},\mathbf{p}_{k})=1$. We refer readers to the supplementary material for the detailed derivation process. Learning the Weights. Based on the above formulation, the problem of UDF estimation is boiled down to obtaining $\\{w_{1}(\mathbf{q},\mathbf{p}_{k}),~{}w_{2}(\mathbf{q},\mathbf{p}_{k})\\}_{k=1}^{K}$. As illustrated in Fig. 4, we construct a sub-network to learn them adaptively. Intuitively, the values of the weights should be relevant to the relative position between $\mathbf{q}$ and $\mathbf{p}_{k}$ and the overall shape of $\Omega(\mathbf{q})$. Thus, we embed $\mathbf{\tilde{p}}_{k}$ and $\mathbf{\tilde{n}}_{k}$ via an MLP to obtain point-wise features for all neighbors, $\mathbf{F}_{k}(\mathbf{q})$, which are then maxpooled, leading to a global feature, $\mathbf{F}(\mathbf{q})$, to encode the overall shape of $\Omega(\mathbf{q})$. Finally, we concatenate the above-mentioned features and feed them into two separated MLPs followed with the Softmax layer to regress the two sets of weights. Remark. Although $\Theta(\mathbf{q})$ could be calculated through the back- propagation (BP) of the neural network (the NDF method [11] adopts this manner), such a process is time-consuming. Compared with BP-based UDF gradient estimation, our method is more efficient. Besides, owing to the separate estimation processes, the UDF error would not be transferred to its gradient, making it more accurate. To the best of our knowledge, this is the first to decouple the estimation processes of UDF and its gradient. See the experimental comparison in Sec. 4.3. ### 3.3 Edge-based Marching Cube Inspired by the classic marching cube algorithm [22] that extracts triangles in a cube according to the occupancy of its eight vertices, we propose edge- based marching cube (E-MC) to extract triangle meshes from the predicted unsigned distance field in the preceding module. Generally, we first detect whether the connection between any pair of vertices of a cube intersects with the surface, then find the most matched condition with the detection results from the lookup table of Marching Cube. (a) (b) (c) (d) Figure 5: The position relationship between the surface and edge. (a) The surface interacts with the edge. (b)(c) The surface does not interact with the edge. (d) One vertex of the edge is extremely close to the surface. See the supplementary material for comprehensive examination of different cases. Edge Intersection Detection. As illustrated in Fig. 5, if the surface interacts with the connection between vertices $\mathbf{q}_{1}$ and $\mathbf{q}_{2}$, denoted as $\mathbf{q}_{1}\mathbf{q}_{2}$, the following constraints must be satisfied: $\displaystyle\langle\Theta(\mathbf{q}_{1}),\Theta(\mathbf{q}_{2})\rangle<0,$ $\displaystyle\langle\Theta(\mathbf{q}_{1}),\overrightarrow{\mathbf{o}\mathbf{q}_{2}}\rangle>0,~{}~{}\langle\Theta(\mathbf{q}_{2}),\overrightarrow{\mathbf{o}\mathbf{q}_{1}}\rangle>0,\vspace{-0.3cm}$ (5) where $\mathbf{o}$ is the midpoint between $\mathbf{q}_{1}$ and $\mathbf{q}_{2}$. Specifically, the first constraint indicates the directions of the UDF gradients of $\mathbf{q}_{1}$ and $\mathbf{q}_{2}$ must be opposite, i.e., the angle between them is larger than $90^{\circ}$. The last two constraints ensure that $\mathbf{q}_{1}$ and $\mathbf{q}_{2}$ are located at different sides of an identical part of a surface to eliminate the case shown in Fig. 5(c). Once the above three constraints are satisfied, the intersection of the surface with $\mathbf{q}_{1}\mathbf{q}_{2}$, denoted as $\mathbf{v}(\mathbf{q}_{1},\mathbf{q}_{2})$, can be calculate via $\mathbf{v}(\mathbf{q}_{1},\mathbf{q}_{2})=\frac{\mathbf{q}_{2}\Phi(\mathbf{q}_{1})+\mathbf{q}_{1}\Phi(\mathbf{q}_{2})}{\Phi(\mathbf{q}_{1})+\Phi(\mathbf{q}_{2})}.\vspace{-0.15cm}$ (6) In addition, if at least one of $\Phi(\mathbf{q}_{1})$ and $\Phi(\mathbf{q}_{2})$ is less than a small threshold $\tau$, their connection is determined to interact with the surface, and the vertex with the smaller UDF is the intersection. Triangle Extraction. After utilizing the edge intersection detection on the connections between any two vertices of a cube, we find the most similar condition to the detection results in the lookup table of Marching Cube to extract the triangles. We refer the readers to the supplementary material for more details. Remark. With reference to a typical vertex of the 3D cube, MeshUDF [16] converts UDFs to SDFs in a small cube by using only the first constraint of Eq. (3.3). GIFS [38] utilizes a sub-network to determine whether $\mathbf{q}_{1}\mathbf{q}_{2}$ intersects the surface, which is not as general as ours. In Sec. 4.3, we experimentally demonstrate the advantage of our E-MC over MeshUDF. ### 3.4 Loss Function We design a joint loss function to train GeoUDF in an end-to-end manner. (1) Loss for upsampling $\mathcal{L}_{\text{PU}}$. We compute the Chamfer Distance (CD) between the upsampled point cloud $\mathcal{P}_{M}$ and its ground-truth dense point cloud to drive the learning of the LGR module. Note that we do not supervise the normal vectors. (2) Losses for UDF and its gradient learning $\mathcal{L}_{\rm UDF}$ and $\mathcal{L}_{\rm Grad}$. We use the mean absolute error between predicted and ground-truth UDFs and the cosine distance between predicted and ground-truth gradients to supervise the learning of the GUE module. The overall loss function is finally written as $\mathcal{L}=\lambda_{1}\mathcal{L}_{\rm PU}+\lambda_{2}\mathcal{L}_{\rm UDF}+\lambda_{3}\mathcal{L}_{\rm Grad},\vspace{-0.15cm}$ (7) where $\lambda_{1}$, $\lambda_{2}$, and $\lambda_{3}$ are hyper-parameters for balancing the three terms. ## 4 Experiments Implementation Details. During training, we set the upsampling factor $M=16$. For each shape, we sampled 2048 query points near the surface in one training iteration. We set the size of neighbourhood $K=10$. We trained our framework in two stages: first, we only trained LGR, i.e., $\lambda_{1}=100$ and $\lambda_{2}=\lambda_{3}=0$, with the learning rate $10^{-3}$ for 100 epochs; second, we trained the whole network with $\lambda_{1}=100$, $\lambda_{2}=1$, and $\lambda_{3}=0.1$ for 300 epochs with the learning rate $10^{-4}$. We conducted all experiments on an NVIDIA RTX 3090 GPU with Intel(R) Xeon(R) CPU. Method | Clean | Noisy (0.005) ---|---|--- CD $(\times 10^{-2})$ $\downarrow$ | F-Score $\uparrow$ | CD $(\times 10^{-2})$ $\downarrow$ | F-Score $\uparrow$ Mean | Median | $\text{F1}^{0.5\%}$ | $\text{F1}^{1\%}$ | Mean | Median | $\text{F1}^{0.5\%}$ | $\text{F1}^{1\%}$ ONet [25] | $0.894$ | $0.716$ | $0.535$ | $0.756$ | $1.175$ | $1.005$ | $0.256$ | $0.624$ CONet [29] | $0.549$ | $0.489$ | $0.624$ | $0.910$ | $0.454$ | $0.398$ | $0.766$ | $0.936$ SAP [28] | $0.411$ | $0.370$ | $0.841$ | $0.953$ | $0.393$ | $0.358$ | $0.821$ | $0.959$ POCO [5] | 0.321 | 0.283 | 0.913 | 0.971 | 0.369 | 0.337 | 0.863 | 0.965 DOG [34] | $0.429$ | $0.353$ | $0.792$ | $0.956$ | $0.381$ | $0.337$ | $0.814$ | $0.967$ NDF [11] | $0.341$ | $0.320$ | $0.840$ | $0.976$ | $0.431$ | $0.419$ | $0.685$ | $0.961$ GIFS [38] | $0.328$ | $0.276$ | $0.860$ | $0.974$ | $0.418$ | $0.358$ | $0.731$ | $0.958$ Ours | $\boldsymbol{0.234}$ | $\boldsymbol{0.226}$ | $\boldsymbol{0.938}$ | $\boldsymbol{0.992}$ | $\boldsymbol{0.289}$ | $\boldsymbol{0.278}$ | $\boldsymbol{0.893}$ | $\boldsymbol{0.987}$ Table 1: Quantitative comparison of different methods on the 13 classes of the ShapeNet dataset. (a) Input(b) GT (c) ONet [25] (d) CONet [29] (e) SAP [28] (f) POCO [5] (g) DOG [34] (h) GIFS [38] (i) Ours Figure 6: Visual comparisons on the ShapeNet dataset [8]. We refer readers to the supplementary material for more visual results. ### 4.1 Watertight Surface Reconstruction Dataset and Metrics. Following previous works [25, 29, 28, 5], we chose 13 classes of the ShapeNet dataset [8] and followed DISN [37] to remove the non- manifold structures of the original shapes. We split the data into the train/val/test sets according to 3D-R2N2 [12]. For each shape, we randomly sampled 3000 points from the surface as the input sparse point cloud. Besides, we also added the Gaussian noise of standard deviation 0.005 to generate noisy inputs. To measure reconstruction quality, we followed previous works [25, 29, 28, 5, 38] to randomly sample $10^{5}$ points from each reconstructed surface to compute CD and F-Score with the threshold of 0.5% and 1% with respect to ground-truths. Comparisons. We compared our GeoUDF with state-of-the-art methods, including ONet [25], CONet [29], SAP [28], POCO [5], DOG [34], NDF [11], and GIFS [38]. For fair comparisons, we trained all methods with the same dataset until the minimum validation error was reached. We quantitatively compared those methods in Table 1, where it can be seen that our method outperforms the other methods in terms of all metrics. Besides, as illustrated in Fig. 6, our GeoUDF can reconstruct surfaces with sharp edges and correct structures that are closer to GTs. However, the methods, i.e., ONet, CONet, SAP, and POCO, which divide the whole space into inside and outside, fail to recover the thin structures, such as the back of a chair. Although GIFS can reconstruct such thin structures, its resulting surfaces are not smooth. (d) Ours (c) GIFS [38] (b) GT (a) Input Figure 7: Visual comparisons on the MGN dataset [4]. ### 4.2 Unseen Non-watertight Surface Reconstruction Dataset and Metrics. Following GIFS [38], we evaluated our method on the MGN [4] and the raw ShapeNet car111These raw data were not used in the training of the experiment in Sec. 4.1. [8] datasets, containing the shapes with boundaries and interior structures. Besides, we also evaluated our method on the ScanNet dataset [13], in which the point clouds were collected through an RGB-D camera from the real scenes. For the MGN dataset, we randomly sampled 3000 points from the surface as the input. As for the shapes and scenes in the ShapeNet car and ScanNet dataset, we randomly sampled 6000 points from the surface or the dense point cloud as the input. Comparisons. To verify the generalizability, for all methods, we used the trained models on the ShapeNet dataset (watertight shapes) to perform inference directly. As shown in Tables 2 and 3 and Figs. 7 and 8, our GeoUDF exceeds GIFS [38] both quantitatively and visually on the MGN and ShapeNet car datasets. Besides, Fig. 9 visualizes the results of different methods on the ScanNet dataset, where it can be seen that CONet [29] and POCO [5] fail to reconstruct the whole scene; GIFS can reconstruct the objects in the scene, but the details are not well preserved. By contrast, the surfaces by our GeoUDF contain more details and are closer to GT ones. (a) Input (b) GT (c) GIFS [38] (d) Ours Figure 8: Visual comparisons on the ShapeNet cars. Method | CD ($\times 10^{-2}$) $\downarrow$ | F-Score $\uparrow$ ---|---|--- Mean | Media | $\text{F1}^{0.5\%}$ | $\text{F1}^{1\%}$ GIFS [38] | $0.248$ | $0.234$ | $0.951$ | $0.991$ Ours | $\boldsymbol{0.194}$ | $\boldsymbol{0.190}$ | $\boldsymbol{0.975}$ | $\boldsymbol{0.996}$ Table 2: Quantitative comparisons on the MGN dataset. Method | CD ($\times 10^{-2}$) $\downarrow$ | F-Score $\uparrow$ ---|---|--- Mean | Media | $\text{F1}^{0.5\%}$ | $\text{F1}^{1\%}$ GIFS [38] | $0.408$ | $0.346$ | $0.758$ | $0.954$ Ours | $\boldsymbol{0.310}$ | $\boldsymbol{0.271}$ | $\boldsymbol{0.870}$ | $\boldsymbol{0.986}$ Table 3: Quantitative comparisons on the ShapeNet Car dataset. (f) Ours (e) GIFS [38] (d) POCO [5] (c) CONet [29] (b) GT (a) Input Figure 9: Visual comparisons of different surface reconstruction methods on the ScanNet dataset [13]. ### 4.3 Ablation Study Upsampling process. We evaluated the upsampling function of our LGR module on the ShapeNet [8] dataset with $M=16$ and compared with recent representative PU methods, i.e., PUGeo-Net [32], MAFU [33], and NP [15]. For a fair comparison, we retrained the official codes of the compared methods with the same training data as ours. As listed in Table 4, it can be seen that our method with the fewest number of parameters achieves the best performance. NP is much worse than others because its modeling manner makes it hard to deal with very sparse point clouds. Besides, Fig. 10 visually compares the results of different methods, further demonstrating the advantage of our method. Method | # Param | Normal Supervision | CD ($\downarrow$) ($\times 10^{-2}$) | P2F ($\downarrow$) ($\times 10^{-3}$) ---|---|---|---|--- PUGeo [32] | 1.287M | Yes | $0.287$ | $1.309$ MAFU [33] | 1.390M | No | $0.289$ | $1.220$ NP [15] | 0.637M | Yes | $0.493$ | $4.572$ Ours | 0.522M | No | $\boldsymbol{0.245}$ | $\boldsymbol{0.512}$ Table 4: Comparisons of different PU methods ($M=16$). P2F: point-to-face distance. (a) (b) (c) (d) (e) (f) Figure 10: Visual comparison of the results by different upsampling methods on the ShapeNet dataset [8] . Effectiveness of LGR. We removed LGR from our GeoUDF to reconstruct surfaces from input sparse point clouds directly. Under this scenario, we estimated the normal vectors by using the principal component analysis (PCA). Besides, we also equipped GIFS [38] with our LGR, i.e., input point clouds were upsampled by our LGR before voxelization. From Table 5, it can be seen that LRG can boost the reconstruction accuracy of both GIFS and ours, validating its effectiveness. Besides, as visualized in Fig. 11, without LGR, the reconstructed surfaces are much worse, i.e., there are many holes caused by the sparsity of the input point cloud. Method | L1-CD $(\times 10^{-2})$ $\downarrow$ | F-Score $\uparrow$ ---|---|--- Mean | Median | $\text{F1}^{0.5\%}$ | $\text{F1}^{1\%}$ GIFS | 0.328 | 0.276 | 0.860 | 0.974 GIFS w/ LGR | 0.316 | 0.292 | 0.894 | 0.981 Ours w/o LGR | 0.284 | 0.264 | 0.882 | 0.967 Ours | 0.234 | 0.226 | 0.938 | 0.992 Table 5: The reconstruction accuracy of GIFS [38] and our method with and without LGR. (a) (b) (c) (d) Figure 11: Visual results of our method w/ and w/o LGR. Accuracy of UDF and its gradient estimation. The accuracy of reconstructed surfaces cannot exactly reflect that of UDFs and corresponding gradients due to the discrete cubes of MC. Thus, we directly compared the accuracy of UDFs and gradients estimated by different methods. In addition to NDF [11] and GIFS [38] for comparison, we also set another baseline (named Regress) by replacing Eqs. (3) and (4) of our GUE with an MLP to regress UDFs and gradients. As listed in Table 6, our GUE produces the most accurate UDFs and gradients. Besides, BP can obtain gradients that are slightly worse than those of our GUE, but it is much slower. Finally, we validated that the shape encoding (SE) of $\Omega(\mathbf{q})$ is essential to our GUE, i.e., GUE w/o SE produces UDFs and gradients with much larger error than GUE. UDF Estimation | UDF Error $\downarrow$ | Time $\downarrow$ ($\upmu$s/point) ---|---|--- UDF $(\times 10^{-3})$ | Grad $(^{\circ})$ NDF [11] | 2.536 | 13.930 | 1.173 GIFS [38] | 2.439 | - | 4.228 Regress | 0.994 | 8.371 | 1.762 BP | 0.615 | 7.877 | 12.125 GUE w/o SE | 1.453 | 10.707 | 1.271 GUE | 0.615 | 7.237 | 3.405 Table 6: UDF accuracy of different methods. Note that GIFS does not rely on the gradients of UDFs. “BP” means in our GUE, the gradients of UDFs are directly obtained by the back-propagation of the UDF estimation network. Size of $\Omega(\mathbf{q})$. From Table 7, we can conclude that our GeoUDF is robust to the size of neighborhood used in Eqs. (3) and (4). This is credited to the manner of learning adaptive weights, i.e., very smaller weights would be predicted for not important neighbours. $K$ | CD ($\times 10^{-2}$) $\downarrow$ | F-Score $\uparrow$ ---|---|--- Mean | Median | $\text{F1}^{0.5\%}$ | $\text{F1}^{1\%}$ 5 | $0.235$ | $0.227$ | $0.937$ | $0.992$ 10 | $0.234$ | $0.226$ | $0.938$ | $0.992$ 20 | $0.232$ | $0.224$ | $0.939$ | $0.993$ Table 7: The reconstruction accuracy of different $K$-NN sizes. Superiority of E-MC. We also compared our E-MC with MeshUDF [16], the SOTA method for extracting triangle meshes from unsigned distance fields. For a fair comparison, we fed the two methods with identical unsigned distance fields estimated by our GUE. From Table 8, it can be seen that our E-MC can extract triangle meshes with higher quality than MeshUDF. Besides, we also studied how the resolution of the 3D grid affects reconstruction accuracy. From Table 9, we can see that the reconstruction accuracy gradually improves with the resolution increasing, which is consistent with the visual results in Fig. 12, but more times are consumed. Such an observation is fundamentally credited to the highly-accurate UDFs by our method. Method | CD ($\times 10^{-2}$) $\downarrow$ | F-Score $\uparrow$ ---|---|--- Mean | Median | $\text{F1}^{0.5\%}$ | $\text{F1}^{1\%}$ GUE+MeshUDF [16] | $0.266$ | $0.251$ | $0.916$ | $0.978$ GUE+E-MC | $\boldsymbol{0.234}$ | $\boldsymbol{0.226}$ | $\boldsymbol{0.938}$ | $\boldsymbol{0.992}$ Table 8: Quantitative comparison of our E-MC with MeshUDF [16]. (a) (b) (c) (d) (e) (f) Figure 12: Visual illustration of the effect of the resolution of E-MC on the reconstructed surface. Res. | CD ($\times 10^{-2}$) $\downarrow$ | F-Score $\uparrow$ | Time (s) $\downarrow$ ---|---|---|--- Mean | Median | $\text{F1}^{0.5\%}$ | $\text{F1}^{1\%}$ | UDF | E-MC 32 | $0.584$ | $0.554$ | $0.646$ | $0.824$ | 0.051 | 3.295 64 | $0.312$ | $0.285$ | $0.860$ | $0.958$ | 0.138 | 4.744 128 | $0.234$ | $0.226$ | $0.938$ | $0.992$ | 0.759 | 15.282 192 | $0.223$ | $0.218$ | $0.949$ | $0.995$ | 2.264 | 32.052 Table 9: The reconstruction accuracy under various 3D grid resolutions used in E-MC. ### 4.4 Efficiency Analysis We also evaluated the efficiency of our GeoUDF on the ShapeNet dataset. As listed in Table 10, our method has the fewest number of parameters while achieving the highest reconstruction accuracy among all methods, which is credited to our explicit and elegant formulations to this problem. We also evaluated the time consumption of our method, demonstrating that all modules before surface extraction are very efficient. As for the surface extraction process, due to the global optimization in our E-MC, our method is slower than ONet, CONet and SAP, but faster than GIFS. Method | # Param | Time (s) $\downarrow$ ---|---|--- Inference | Surface Extraction ONet [25] | 10.373M | 0.653 | 0.219 CONet [29] | 1.978M | 0.941 | 0.589 SAP [28] | 1.085M | 0.247 | 0.384 POCO [5] | 12.790M | 4.652 | 8.560 DOG [34] | 2.182M | 0.302 | 0.228 GIFS [38] | 3.682M | 27.771 | 26.915 Ours | 0.775M | 0.759 | 15.282 Table 10: Efficiency comparisons. The time of “Inference” refers to the overall time minus the time for iso-surface extraction. ## 5 Conclusion We introduced GeoUDF, a new learning-based framework for reconstructing general surfaces from sparse 3D point clouds. GeoUDF is featured with lightweight, efficient, accurate, explainable, and generalizability properties, which are validated by extensive experiments and ablation studies. The various advantages of our framework are fundamentally credited to the explicit and elegant formulations of the problems of UDF and its gradient estimation and local structure representation from the geometric perspective, as well as the edge-based marching cube. ## References * [1] Ma Baorui, Han Zhizhong, Liu Yu-Shen, and Zwicker Matthias. Neural-pull: Learning signed distance functions from point clouds by learning to pull space onto surfaces. In International Conference on Machine Learning (ICML), 2021. * [2] Jan Bednarik, Shaifali Parashar, Erhan Gundogdu, Mathieu Salzmann, and Pascal Fua. Shape reconstruction by learning differentiable surface representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020. * [3] Fausto Bernardini, Joshua Mittleman, Holly Rushmeier, Cláudio Silva, and Gabriel Taubin. The ball-pivoting algorithm for surface reconstruction. IEEE transactions on visualization and computer graphics, 5(4):349–359, 1999. * [4] Bharat Lal Bhatnagar, Garvita Tiwari, Christian Theobalt, and Gerard Pons-Moll. Multi-garment net: Learning to dress 3d people from images. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2019. * [5] Alexandre Boulch and Renaud Marlet. Poco: Point convolution for surface reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 6302–6314, June 2022. * [6] Robert Bridson. Fast poisson disk sampling in arbitrary dimensions. In ACM SIGGRAPH 2007 sketches, pages 22–es. 2007. * [7] Rohan Chabra, Jan E Lenssen, Eddy Ilg, Tanner Schmidt, Julian Straub, Steven Lovegrove, and Richard Newcombe. Deep local shapes: Learning local sdf priors for detailed 3d reconstruction. In European Conference on Computer Vision, pages 608–625. Springer, 2020. * [8] Angel X Chang, Thomas Funkhouser, Leonidas Guibas, Pat Hanrahan, Qixing Huang, Zimo Li, Silvio Savarese, Manolis Savva, Shuran Song, Hao Su, et al. Shapenet: An information-rich 3d model repository. arXiv preprint arXiv:1512.03012, 2015. * [9] Jiawen Chen, Dennis Bautembach, and Shahram Izadi. Scalable real-time volumetric surface reconstruction. ACM Transactions on Graphics (ToG), 32(4):1–16, 2013. * [10] Julian Chibane, Thiemo Alldieck, and Gerard Pons-Moll. Implicit functions in feature space for 3d shape reconstruction and completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020. * [11] Julian Chibane, Gerard Pons-Moll, et al. Neural unsigned distance fields for implicit function learning. Advances in Neural Information Processing Systems, 33:21638–21652, 2020. * [12] Christopher B Choy, Danfei Xu, JunYoung Gwak, Kevin Chen, and Silvio Savarese. 3d-r2n2: A unified approach for single and multi-view 3d object reconstruction. In European conference on computer vision, pages 628–644. Springer, 2016. * [13] Angela Dai, Angel X. Chang, Manolis Savva, Maciej Halber, Thomas Funkhouser, and Matthias Niessner. Scannet: Richly-annotated 3d reconstructions of indoor scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017. * [14] Manfredo P Do Carmo. Differential geometry of curves and surfaces: revised and updated second edition. Courier Dover Publications, 2016. * [15] Wanquan Feng, Jin Li, Hongrui Cai, Xiaonan Luo, and Juyong Zhang. Neural points: Point cloud representation with neural fields for arbitrary upsampling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 18633–18642, June 2022. * [16] Benoit Guillard, Federico Stella, and Pascal Fua. Meshudf: Fast and differentiable meshing of unsigned distance field networks. In European Conference on Computer Vision, 2022. * [17] Hugues Hoppe. Poisson surface reconstruction and its applications. In Proceedings of the 2008 ACM symposium on Solid and physical modeling, pages 10–10, 2008. * [18] Michael Kazhdan and Hugues Hoppe. Screened poisson surface reconstruction. ACM Transactions on Graphics (ToG), 32(3):1–13, 2013. * [19] Ravikrishna Kolluri. Provably good moving least squares. ACM Transactions on Algorithms (TALG), 4(2):1–25, 2008. * [20] Ruihui Li, Xianzhi Li, Chi-Wing Fu, Daniel Cohen-Or, and Pheng-Ann Heng. Pu-gan: A point cloud upsampling adversarial network. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2019. * [21] Shi-Lin Liu, Hao-Xiang Guo, Hao Pan, Peng-Shuai Wang, Xin Tong, and Yang Liu. Deep implicit moving least-squares functions for 3d reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 1788–1797, June 2021. * [22] William E Lorensen and Harvey E Cline. Marching cubes: A high resolution 3d surface construction algorithm. ACM siggraph computer graphics, 21(4):163–169, 1987. * [23] Baorui Ma, Yu-Shen Liu, and Zhizhong Han. Reconstructing surfaces for sparse point clouds with on-surface priors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 6315–6325, June 2022. * [24] Aihua Mao, Zihui Du, Junhui Hou, Yaqi Duan, Yong-jin Liu, and Ying He. Pu-flow: A point cloud upsampling network with normalizing flows. IEEE Transactions on Visualization and Computer Graphics, 2022. * [25] Lars Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. Occupancy networks: Learning 3d reconstruction in function space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. * [26] Philipp Mittendorfer and Gordon Cheng. 3d surface reconstruction for robotic body parts with artificial skins. In 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 4505–4510, 2012. * [27] Jeong Joon Park, Peter Florence, Julian Straub, Richard Newcombe, and Steven Lovegrove. Deepsdf: Learning continuous signed distance functions for shape representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. * [28] Songyou Peng, Chiyu Jiang, Yiyi Liao, Michael Niemeyer, Marc Pollefeys, and Andreas Geiger. Shape as points: A differentiable poisson solver. Advances in Neural Information Processing Systems, 34:13032–13044, 2021. * [29] Songyou Peng, Michael Niemeyer, Lars Mescheder, Marc Pollefeys, and Andreas Geiger. Convolutional occupancy networks. In European Conference on Computer Vision, pages 523–540. Springer, 2020. * [30] Charles Ruizhongtai Qi, Li Yi, Hao Su, and Leonidas J Guibas. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Advances in neural information processing systems, 30, 2017. * [31] Guocheng Qian, Abdulellah Abualshour, Guohao Li, Ali Thabet, and Bernard Ghanem. Pu-gcn: Point cloud upsampling using graph convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 11683–11692, June 2021. * [32] Yue Qian, Junhui Hou, Sam Kwong, and Ying He. PUGeo-Net: A geometry-centric network for 3D point cloud upsampling. In European conference on computer vision, pages 752–769. Springer, 2020. * [33] Yue Qian, Junhui Hou, Sam Kwong, and Ying He. Deep magnification-flexible upsampling over 3d point clouds. IEEE Transactions on Image Processing, 30:8354–8367, 2021. * [34] Peng-Shuai Wang, Yang Liu, and Xin Tong. Dual octree graph networks for learning adaptive volumetric shape representations. ACM Transactions on Graphics (TOG), 41(4):1–15, 2022. * [35] Yue Wang, Yongbin Sun, Ziwei Liu, Sanjay E Sarma, Michael M Bronstein, and Justin M Solomon. Dynamic graph cnn for learning on point clouds. Acm Transactions On Graphics (tog), 38(5):1–12, 2019. * [36] Yifan Wang, Shihao Wu, Hui Huang, Daniel Cohen-Or, and Olga Sorkine-Hornung. Patch-based progressive 3d point set upsampling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. * [37] Qiangeng Xu, Weiyue Wang, Duygu Ceylan, Radomir Mech, and Ulrich Neumann. Disn: Deep implicit surface network for high-quality single-view 3d reconstruction. Advances in Neural Information Processing Systems, 32, 2019. * [38] Jianglong Ye, Yuntao Chen, Naiyan Wang, and Xiaolong Wang. Gifs: Neural implicit function for general shape representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 12829–12839, June 2022. * [39] Lequan Yu, Xianzhi Li, Chi-Wing Fu, Daniel Cohen-Or, and Pheng-Ann Heng. Pu-net: Point cloud upsampling network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018. * [40] Junsheng Zhou, Baorui Ma, Yu-Shen Liu, Yi Fang, and Zhizhong Han. Learning consistency-aware unsigned distance functions progressively from raw point clouds. arXiv preprint arXiv:2210.02757, 2022.
# Inverse molecular design and parameter optimization with Hückel theory using automatic differentiation Rodrigo A. Vargas–Hernández Chemical Physics Theory Group, Department of Chemistry, University of Toronto, Toronto, Ontario, M5S 3H6, Canada. Vector Institute for Artificial Intelligence, 661 University Ave. Suite 710, Toronto, Ontario M5G 1M1, Canada. current affiliation: Department of Chemistry and Chemical Biology, McMaster University, 1280 Main Street West, Hamilton, Ontario, L8S 4M1 Canada<EMAIL_ADDRESS>Kjell Jorner Chemical Physics Theory Group, Department of Chemistry, University of Toronto, Toronto, Ontario, M5S 3H6, Canada. Department of Computer Science, University of Toronto, 40 St. George St, Ontario M5S 2E4, Canada. Department of Chemistry and Chemical Engineering, Chalmers University of Technology, Kemigården 4, SE-41258, Gothenburg, Sweden. Robert Pollice Chemical Physics Theory Group, Department of Chemistry, University of Toronto, Toronto, Ontario, M5S 3H6, Canada. Department of Computer Science, University of Toronto, 40 St. George St, Ontario M5S 2E4, Canada. current affiliation: Stratingh Institute for Chemistry, University of Groningen, Nijenborgh 4, Groningen, 9747 AG, The Netherlands. Alán Aspuru–Guzik Chemical Physics Theory Group, Department of Chemistry, University of Toronto, Toronto, Ontario, M5S 3H6, Canada. Department of Computer Science, University of Toronto, 40 St. George St, Ontario M5S 2E4, Canada. Department of Chemical Engineering & Applied Chemistry, 200 College St., University of Toronto, Ontario M5S 3E5, Canada. Department of Materials Science & Engineering, 184 College St., University of Toronto, Ontario M5S 3E4, Canada. Vector Institute for Artificial Intelligence, 661 University Ave. Suite 710, Toronto, Ontario M5G 1M1, Canada. Lebovic Fellow, Canadian Institute for Advanced Research (CIFAR), 661 University Ave., Toronto, Ontario M5G 1M1, Canada. ###### Abstract Semi-empirical quantum chemistry has recently seen a renaissance with applications in high-throughput virtual screening and machine learning. The simplest semi-empirical model still in widespread use in chemistry is Hückel’s $\pi$-electron molecular orbital theory. In this work, we implemented a Hückel program using differentiable programming with the JAX framework, based on limited modifications of a pre-existing NumPy version. The auto-differentiable Hückel code enabled efficient gradient-based optimization of model parameters tuned for excitation energies and molecular polarizabilities, respectively, based on as few as 100 data points from density functional theory simulations. In particular, the facile computation of the polarizability, a second-order derivative, via auto-differentiation shows the potential of differentiable programming to bypass the need for numeric differentiation or derivation of analytical expressions. Finally, we employ gradient-based optimization of atom identity for inverse design of organic electronic materials with targeted orbital energy gaps and polarizabilities. Optimized structures are obtained after as little as 15 iterations, using standard gradient-based optimization algorithms. ## 1 Introduction Mathematical models that are both predictive and provide insight are a cornerstone of the physical sciences. However, accurate models for complicated processes often have no analytical solution and require large computational resources to solve numerically. At the same time, they also tend to be hard to interpret, as highlighted by Mulliken’s famous quote ”the more accurate the calculations became, the more the concepts tended to vanish into thin air” [1]. Approximate models with problem-specific parameters are therefore used in practice, but finding optimal values for these parameters can be non-trivial. Parameter optimization normally requires considerable amounts of reference data and is done either manually or with algorithms that do not take advantage of first or higher order derivatives as the corresponding analytical expressions are often unavailable. In chemistry, the Schrödinger equation is an archetype of such a mathematical model that describes the interactions between nuclei and electrons in both atoms and molecules. However, (near) exact solutions are too computationally expensive for most molecules of interest. Quantum chemistry is an entire research field dedicated to finding computationally efficient solutions to the Schrödinger equation by introducing prudent approximations or reformulations [2]. One approach that was extremely successful in the early days of quantum chemistry is the use of so-called semiempirical (SE) approximations [3]. The central idea is the use of problem-specific parameters to simplify the mathematical form of the Schrödinger equation. One of the earliest SE models was Hückel’s method to treat the $\pi$-electrons in organic molecules [4, 5, 6, 7]. Traditionally, the parameters in the Hückel method were derived manually by human scientists with the aim to reproduce properties for well- known reference molecules [8, 9], or they were derived from more accurate calculations [10]. Over the years, the Hückel method has been used for pedagogical purposes and for obtaining physical insight into problems in organic chemistry [11] and photochemistry and photophysics [12, 13]. However, it can also be used as a fast method for the prediction of molecular properties [14], and for inverse design of molecules with desired target properties [15, 16]. The recent upsurge in machine learning (ML), and specifically deep neural networks, created a need for robust and efficient algorithms to co-optimize a very large number of model parameters for various architectures. This problem is now solved by automatic differentiation (AD), a technique to evaluate the derivatives of mathematical expressions via the chain rule [17]. Importantly, AD removes the need to determine analytic expressions for derivatives and makes complicated mathematical models amenable to gradient-based optimization, allowing them to be applied in the same way as general supervised machine learning models. Regular machine learning approaches like deep neural networks are meant to be very general mathematical models with a large number of parameters. Through learning, they can adapt to essentially any problem provided sufficient training data is available. In contrast, physics-based mathematical models have expressions that are specific to a certain type of problem to be solved and feature a much smaller number of parameters. Implementing physical models such as quantum chemistry within AD frameworks enables the use of default learning algorithms for parameter optimization with a potentially much smaller training data requirement. Along these lines, autodifferentiable versions of Hartree-Fock [18, 19], density functional theory (DFT) [20, 21, 22, 19], excited state mean-field theory [23]. For semi- empirical methods [24, 25, 26, 27], and other applications [28, 20, 29, 30, 30, 31, 32, 33, 34, 35, 36, 37], AD has been used to accelerate the calculation of gradients physical methods and to blend with ML algorithms. In this work, we developed an auto-differentiable implementation of the Hückel method, by minimal adaptation of an initially developed NumPy [38] version into the JAX [39] AD framework. We use this model to demonstrate the ease and efficiency of parameter fitting based on computational reference data sets for both excitation energies and molecular polarizabilities, a property calculated via a second order derivative. Additionally, we demonstrate that our AD model allows for gradient-based inverse design by regarding the atomic composition of a molecular system as an adjustable parameter to find molecules with targeted properties.[16] The corresponding code is made publicly available, allowing it to be applied to a large variety of chemical problems. As the Hückel calculations are extremely fast, our workflow allows for facile development of property-specific models that can be readily used in molecular generative models that require on the order of 105–106 property evaluations. The paper is structured as follows: we first present a short introduction on automatic differentiation and the Hückel model (Sections 2.1 and 2.2). Following that, we execute inverse design of molecules as a fully differentiable procedure (Section 3.1) and perform optimization of the Hückel model parameters using modern gradient-based methods (Section 3.2). ## 2 Methods ### 2.1 Automatic differentiation Gradients and high-order derivatives are at the core of physical simulations. For physical models, common approaches to evaluate derivatives of any order are closed-form solutions, symbolic differentiation, and numerical differentiation, i.e., finite differences [40, 17]. For any function represented as a computer program, AD [17] is an alternative way to compute gradients and higher order derivatives. AD makes use of the chain rule for differentiation to create a program that computes the gradients during evaluation. There are two main modes in AD, forward and reverse mode. For scalar functions, reverse mode is more efficient as differentiation requires a single evaluation of the function to fully compute the Jacobian. An example of reverse mode differentiation is the backpropagation algorithm that is used for training neural networks. For more details about AD, we refer the reader to Ref. [17]. The optimization of ML models is mostly done with methods that require the gradient of the loss or error function (${\cal L}$) with respect to the model parameters ($\boldsymbol{\theta}$), $\nabla_{\boldsymbol{\theta}}{\cal L}(\boldsymbol{\theta})$. All contemporary ML libraries, e.g., Tensorflow [41], PyTorch [42] and JAX [39], are built on top of an AD engine which computes $\nabla_{\boldsymbol{\theta}}{\cal L}(\boldsymbol{\theta})$ for any ML model. Given the robustness of AD libraries, differentiating physical models [40] could be done similarly to modern ML algorithms. ### 2.2 Hückel model The Hückel model, a well-known semi-empirical quantum chemistry model [4, 5, 6, 7], was first proposed to describe the interactions of $\pi$-electrons in conjugated unsaturated hydrocarbons. In the Hückel model, this interaction is restricted to electrons centered at nearest neighbour atoms. Generally, the Hückel model is considered a tight-binding type Hamiltonian (Eq. 1) where the on-site and hopping parameters are commonly denoted in the literature as $\alpha_{\ell}$ and $\beta_{\ell,k}$, respectively. The matrix elements of the Hückel Hamiltonian are given by $H_{\ell,k}=\begin{cases}\alpha_{\ell},&\ell=k\\\ \beta_{\ell,k},&\ell\text{ and }k\text{ are adjacent }\\\ 0,&\text{ otherwise, }\end{cases}$ (1) where the $\alpha_{\ell}$ parameters roughly represent the energy of an electron in a 2p orbital, and the $\beta_{\ell,k}$ parameters describe the energy of an electron in the bond $\ell-k$. Extensions of the Hückel Hamiltonian are possible and can, for instance, incorporate distance- dependence via $\beta_{\ell,k}=\beta^{0}_{\ell,k}\;g(\mathbf{R}_{\ell,k})$ (cf. Section 3.2). For more details, we refer the reader to standard quantum chemistry textbooks [43, 44]. Notably, any molecular property computed with the Hückel method depends directly on the $\alpha_{\ell}$ and $\beta_{\ell,k}$ parameters. Therefore, by tuning their values, one can either construct a more accurate Hückel model for a given molecule and property (cf. Section 3.2), or search for atomic compositions that optimize target properties given a preset connectivity (cf. Section 3.1). In the following sections, we demonstrate how AD can be used to facilitate both these types of problems. ## 3 Results and Discussion ### 3.1 Inverse molecular design Inverse molecular design can be carried out via gradient-based optimization methods, as shown in Ref. [16]. The Hückel model can be extended to search for the molecular structure with a desired property. Both the diagonal and off- diagonal elements of the Hückel Hamiltonian matrix can be described by a weighted average of atom types at each site, $H_{\ell,k}=\begin{cases}\sum_{i}^{M}b_{\ell}^{i}\alpha^{i}_{\ell},&\ell=k\\\ \sum_{i}^{M}\sum_{j}^{M}b_{\ell}^{i}b_{k}^{j}\beta^{ij}_{\ell,k},&\ell\text{ and }k\text{ are adjacent }\\\ 0,&\text{ otherwise },\end{cases}$ (2) where $b_{\ell}^{i}$ is the weight of the atom of type $i$ for site $\ell$. For a meaningful description, the weights of each site must be normalized, i.e., $\sum_{i}^{M}b_{\ell}^{i}=1$. $M$ is the total number of atom types considered in the search. As a proof of concept, we consider eight different molecular frameworks [16], which are displayed in Fig. 1. The $x$-symbol indicates the sites with variable atom types to be optimized. We only considered carbon ($C$), nitrogen ($N$) and phosphorus ($P$), i.e., $M=3$, as these atom types each contribute one electron, assuming that the remaining valences of carbon will be satisfied with a bond to an implicit hydrogen atom, and can be incorporated interchangeably at all sites with two neighbors in the $\pi$-framework (Fig. 1). Therefore we defined the following vector of atom type weight parameters: $\mathbf{b}_{\ell}=[b_{\ell}^{C},b_{\ell}^{N},b_{\ell}^{P}]$. For clarity, $\mathbf{b}$ jointly describes the $\mathbf{b}_{\ell}$ parameters for all search sites in a molecule, i.e., $\mathbf{b}=\\{\mathbf{b}_{\ell}\\}^{N}$. For all results presented, the values of the $\alpha_{\ell}$ and $\beta^{0}_{\ell,k}$ parameters were previously optimized with respect to the desired property (cf. Section 3.2). For the set of eight molecules considered (cf. Fig. 1), we search for the type of atom $\mathbf{b}$ at each site that gives the lowest HOMO-LUMO gap (Eq. 3), denoted as $\epsilon_{HL}$, and the maximum polarizability denoted as $\langle\tilde{\alpha}\rangle$, (Eq, 4). $\epsilon_{HL}$ is defined as, $\epsilon_{HL}=\epsilon_{LUMO}-\epsilon_{HOMO},$ (3) where $\epsilon_{HOMO}$ and $\epsilon_{LUMO}$ are the eigenvalues of the highest occupied molecular orbital (HOMO), and lowest unoccupied molecular orbital (LUMO), respectively. The polarizability function is defined as $\langle\tilde{\alpha}\rangle=\frac{1}{3}\left(\tilde{\alpha}_{xx}+\tilde{\alpha}_{yy}+\tilde{\alpha}_{zz}\right),$ (4) where the $\tilde{\alpha}_{ij}$ elements are the polarizability components defined as $\tilde{\alpha}_{ij}=-\frac{\partial^{2}E}{\partial F_{i}\partial F_{j}}.$ (5) The $F_{i}$ terms are the components of the electric field, $\vec{\boldsymbol{F}}=[F_{x},F_{y},F_{z}]$, and $E$ is the electronic energy of the system. The elements of the polarizability tensor are usually computed using a finite-difference (FD) approach [16], $\tilde{\alpha}_{ii}=\frac{2E(0)-\left[E(-F_{i})+E(+F_{i})\right]}{F_{i}^{2}},$ (6) where the electronic energy is evaluated several times, typically three times for each diagonal element, and four times for each cross term. Notably, if the parameters $\mathbf{b}$ are to be optimized using a gradient-based method, the Jacobians $\nabla_{\mathbf{b}}\epsilon_{HL}$ and $\nabla_{\mathbf{b}}\langle\tilde{\alpha}\rangle$ will also be constructed using an FD approach. However, this will increase the number of energy evaluations needed, especially for $\langle\tilde{\alpha}\rangle$ as the elements of $\nabla_{\mathbf{b}}\langle\tilde{\alpha}\rangle$ are third-order derivatives: $\frac{\partial}{\partial\mathbf{b}^{i}}\frac{\partial^{2}\tilde{\alpha}_{kk}}{\partial F^{2}_{k}}$. Thus, for a single element of $\nabla_{\mathbf{b}}\langle\tilde{\alpha}\rangle$, using FD will require 18 energy calculations, ${\cal O}(18\times\|\mathbf{b}\|)$, where $\|\mathbf{b}\|$ is the total number of parameters in $\mathbf{b}$. For $\epsilon_{HL}$, using FD, we only require ${\cal O}(2\times\|\mathbf{b}\|)$, as $\nabla_{\mathbf{b}}\epsilon_{HL}$ is a first-order derivative. In contrast, using modern AD frameworks, we can efficiently compute, with a single energy calculation (i.e., one forward pass), the Jacobian of $\epsilon_{HL}$ with respect to $\mathbf{b}$. For $\langle\tilde{\alpha}\rangle$, the number of total energy evaluations depends on the dimension of the external field to construct the diagonal elements of the Hessian (Eq. 5), which we also computed using AD. The Jacobian of $\langle\tilde{\alpha}\rangle$ with respect to $\mathbf{b}$, a third order derivative, can be constructed from only three energy evaluations using AD [17], a drastic reduction from the 18 required for FD. After implementation of the Hückel model using the JAX ecosystem [39], we could fully differentiate both observables, $\epsilon_{HL}$ and $\langle\tilde{\alpha}\rangle$. Importantly, using JAX allowed us to convert our existing Python-based Hückel code very easily by replacing calls to NumPy with almost equivalent calls to the JAX.Numpy package. For optimization, we used the Broyden–Fletcher–Goldfarb–Shanno (BFGS) algorithm via the JaxOpt library [45]. Instead of using a constrained optimization scheme to satisfy the site-normalization restriction for $\mathbf{b}$, we used the softmax function, $b^{i}=\frac{\exp(\underline{b}^{i})}{\sum_{i}^{M}\exp(\underline{b}^{i})},$ (7) where $\underline{b}$ represents the unnormalized $b$ parameters. Given the flexibility of the JAX ecosystem, we were able to test other gradient optimization algorithms such as Adam [46] and canonical gradient- descent, but found the BFGS to be most efficient as it required, on average, fifteen or less total iterations to reach convergence (cf. Fig. 2). The Adam and gradient-descent algorithms, with an exponential learning rate decay, each needed more than thirty iterations to minimize $\epsilon_{HL}$ or $\langle\tilde{\alpha}\rangle$. Notably, we initialized the values for all $\underline{b}^{i}$ parameters by sampling a uniform distribution, $\underline{b}^{i}\sim{\cal U}(-1,1)$. Instead of using literature Hückel parameters, we used our optimized parameters for each target observable where 5,000 training molecules were used. More details are described in Section 3.2. As seen in Figs. 3 and 4, we found that our random initialization of $\mathbf{\underline{b}}$ allows us to sample a wide range of molecules with a broad range of values for both objectives, $\epsilon_{HL}^{\mathrm{initial}}$ and $\langle\tilde{\alpha}\rangle_{\mathrm{initial}}$. Because of the statistical description of the molecules by the $\mathbf{b}$ parameters, the optimal parameters ($\mathbf{b}^{*}$) found by optimization are not one-hot vectors that correspond to only one atom type per site in the molecule but rather a linear combination of multiple atom types. We define the observable value for this unphysical molecule as $y_{\mathrm{virtual}}$ and that for the real molecule as $y_{\mathrm{feasible}}$ (i.e., the most probable atom is picked for each site to define the real molecule). An example of this is displayed for framework 3 in Fig. 5 where we show the change in $\mathbf{b}$ throughout the optimization for both objectives ($\epsilon_{HL}$ and $\langle\tilde{\alpha}\rangle$). As we can observe, the change in $\epsilon_{HL}$ from the initial random molecule to the optimal one is close to 1 eV. The optimizations in Fig. 5 were done using Adam only to properly illustrate the change in $\mathbf{b}$ as the change between iterations is smoother and more readily discernable. Importantly, this shows that the generative model can shift the distribution of properties towards the target property with little dependence on the random initialization of $\mathbf{b}$ (Figs. 3 and 4). We also observe that, for the majority of the optimized molecules, $y_{virtual}$ and $y_{feasible}$ are linearly correlated, indicating that the optimization converged essentially to feasible molecules. The property distributions of $y_{\mathrm{virtual}}$ and $y_{\mathrm{feasible}}$ are reasonably close, even in the few cases when the correlations are poor. Fig. 6 displays the molecules with the lowest $\epsilon_{HL}$ and maximum $\langle\tilde{\alpha}\rangle$ from the ensemble of different optimizations. First, we notice that there is a higher amount of phosphorus atoms in the molecules when $\langle\tilde{\alpha}\rangle$ was the target property. This is not unexpected as molecular polarizabilities, while not simply a sum of the atomic polarizabilities, are strongly influenced by the atomic polarizabilities of the constituent atoms [47, 48]. As phosphorus is a third- row element in the same group with nitrogen and atomic polarizability increases significantly when increasing the row number, its atomic polarizability, both in free atoms [49] and in molecules [50], is significantly larger than both nitrogen and carbon. Therefore, incorporating a large number of phosphorus atoms is expected to be a viable strategy to maximize the molecular polarizability in all of the molecular frameworks considered. For the molecules with the lowest $\epsilon_{HL}$, we see extensive incorporation of both nitrogen and phosphorous atoms. This can be understood in terms of the effect of heteroatom substitution on $\epsilon_{HL}$ within the HMO framework [51]. For alternant compounds such as 1–8, $\epsilon_{HL}$ is unaffected by changing $\alpha_{\ell}$. The main effect comes from changing $\beta_{\ell,k}$, and is expected to be largest for bonds that feature a bonding interaction in the HOMO and an anti-bonding interaction in the LUMO. A lowered $\beta_{\ell,k}$ leads to decreased bonding interactions in the HOMO and consequently, a higher HOMO energy. For the LUMO, decreasing the antibonding interactions by a lowered $\beta_{\ell,k}$ leads to a lowering of the energy. The net effect by raising the HOMO and lowering the LUMO is a decrease in the $\epsilon_{HL}$ gap. We can therefore expect optimization to favor atom pairs with a low $\beta_{\ell,k}$ for bonds that feature a bonding interaction in the HOMO and an antibonding interaction in the LUMO. Inspection of the molecular orbitals of the optimized frameworks (cf. Figure S1) indeed reveals that these bonds are dominantly between two N atoms, which feature the by far lowest $\beta_{\ell,k}$ at 0.159 (the next lowest is for P–P at 0.539). Control optimizations with the original parameter set by Van-Catledge [10] instead gives molecules with P–P at those bonds (cf. Figure S2), consistent with the fact that $\beta_{P,P}$ = 0.63 is the lowest for this parameter set. For this proof of principle work, we picked two distinct molecular target properties. However, based on the framework employed, a significant number of alternative properties could also be predicted and, thus, used for inverse design via gradient-based optimization. Additionally, any combined objective that is derived from multiple target properties can equally be optimized for via the same types of algorithms out of the box. This is particularly interesting for properties where Hückel models are known to provide reasonable prediction accuracies such as HOMO-LUMO gaps. The use of gradient-based optimization algorithms enables fast convergence towards the closest local optimum solution reducing the number of evaluations and leading to significantly increased computation time. This is particularly important as one of the main bottlenecks in current approaches to inverse molecular design is the number of property evaluations needed to find an optimal structure [52]. Going beyond single-objective optimization, one possible extension of our presented approach is targeting multiple objectives via genuine gradient- based multi-objective optimization, for example, both $\epsilon_{HL}$ and $\langle\tilde{\alpha}\rangle$. The standard approach to perform multi- objective optimization is via via property concatenation into a single function to use standard single-objective algorithms, where algorithms like Bayesian optimization are used [53, 54, 55, 56, 57]. However, gradient-based multi-objective optimization algorithms [58] have been developed and they, together with automatic differentiation, could be employed for both parameter optimization and inverse molecular design in order to explore the corresponding Pareto fronts in a systematic manner. From a conceptual point of view, representing chemical structure subspaces in a parameterized form can greatly facilitate inverse design [16] as it allows the use of well-established approaches for parameter optimization to be used for the design of molecules. This is particularly effective when used in combination with AD due to its numerical stability and computational efficiency compared to alternative means to compute gradients. Consequently, this also makes the molecular size that can still be feasibly treated in such an approach larger and thus, essentially, expands the chemical subspace the generative model can explore. However, one of the main downsides of the approach implemented in this work is the reliance on fixed molecular frameworks, which is common for alchemical formulations [59] strongly limiting the structural space considered in the optimization. Simple extensions would be i) the combination of methods to change the molecular framework without relying on gradients with the method presented here to modify the atom identities within the respective framework, or ii) differentiable supermatrix structure where atom vacancies are allowed [60]. Ideally, future extensions should aim to find prudent ways allowing for framework modifications based on gradients as this potentially can lead to a dramatic reduction in the number of structure optimization steps and, thus, the number of property evaluations necessary. The extended Hückel model is also compatible with the proposed methodology, even with ML learned parameters [25], by considering a description of the overlap integrals between different atoms types, similar to Eq. 2. Figure 1: The eight molecular frameworks considered for inverse molecular design in this work [16]. The $x$-symbol represents atomic sites whose identity is optimized. Figure 2: Average learning curve for $\epsilon_{HL}$ (left panel) and $\langle\tilde{\alpha}\rangle$ (right panel) for 250 random initial molecules based on the eight different molecular frameworks. We use the BFGS algorithm to optimize both observables. More details about the random initialization are provided in the text. Figure 3: Histograms of the optimized HOMO-LUMO gap, Eq. 3, for 250 random initial molecules ($\epsilon_{HL}^{\mathrm{initial}}$). The inset panels compare the similarity between $\epsilon_{HL}$ computed with the value of $\mathbf{b}$ at the end of the optimization protocol ($\epsilon_{HL}^{\mathrm{virtual}}$), and the values of $\epsilon_{HL}$ selecting the most probable atoms given $\mathbf{b}^{*}$ ($\epsilon_{HL}^{\mathrm{feasible}}$). Curves represent the derived histograms using kernel density estimation, (solid) $\epsilon_{HL}^{\mathrm{feasible}}$, (dashed) $\epsilon_{HL}^{\mathrm{virtual}}$, and (dotdash) $\epsilon_{HL}^{\mathrm{initial}}$. All molecules were optimized using the BFGS algorithm. Molecular frameworks are displayed in Fig. 1. Figure 4: Histograms of the optimized $\langle\tilde{\alpha}\rangle$, Eq. 4, for 250 random initial molecules ($\langle\tilde{\alpha}\rangle^{\mathrm{initial}}$). The inset panels compare the similarity between $\langle\tilde{\alpha}\rangle$ computed with the value of $\mathbf{b}$ at the end of the optimization protocol ($\langle\tilde{\alpha}\rangle^{\mathrm{virtual}}$), and the values of $\langle\tilde{\alpha}\rangle$ selecting the most probable atoms given $\mathbf{b}^{*}$ ($\langle\tilde{\alpha}\rangle^{\mathrm{feasible}}$). Curves represent the derived histograms using kernel density estimation, (solid) $\langle\tilde{\alpha}\rangle^{\mathrm{feasible}}$, (dashed) $\langle\tilde{\alpha}\rangle^{\mathrm{virtual}}$, and (dotdash) $\langle\tilde{\alpha}\rangle^{\mathrm{initial}}$. All molecules were optimized using the BFGS algorithm. Molecular frameworks are displayed in Fig. 1. Figure 5: Change in the parameters $\mathbf{b}$ for $\epsilon_{HL}$ (left panel) and $\langle\tilde{\alpha}\rangle$ (right panel) during the optimization of a single random initial virtual molecule based on framework 3. For the initial, one intermediate and the final $\mathbf{b}$, we plot the values of $\mathbf{b}$ for molecular framework 3. For both panels, the filled markers represent the values of $\epsilon_{HL}$ and $\langle\tilde{\alpha}\rangle$ computed with $\mathbf{b}$ at each iteration (virtual), and the empty markers represent the observable values computed only with the most probable atoms given $\mathbf{b}$ at each iteration (feasible). For each search site we only considered three different atom types, namely $C$, $N$ and $P$; $\mathbf{b}_{\ell}=[b_{\ell}^{C},b_{\ell}^{N},b_{\ell}^{P}]$. The optimization of $\epsilon_{HL}$ and $\langle\tilde{\alpha}\rangle$ w.r.t. $\mathbf{b}$ was carried out with Adam using a learning rate of 0.2. More details are provided in the main text. Figure 6: The molecular structures with the lowest HOMO-LUMO gap and maximum polarizability for each of the eighth molecules considered (Fig. 1). ### 3.2 Parameter optimization Another important task that might sometimes be underappreciated in computational chemistry is model parameter optimization. Here, we leverage the flexibility of AD and optimize all free parameters of the Hückel model in the same way as it is done for modern ML algorithms. Originally, the Hückel model is solely based on electronic interactions between nearest-neighbour atoms, which is typically also referred to as the tight-binding approximation (cf. $\beta_{\ell,k}$ parameters in Eq. 1). Beyond the standard Hückel model, one can introduce atomic distance-dependence of the corresponding interactions via $\beta_{\ell,k}=\beta^{0}_{\ell,k}g(\mathbf{R}_{\ell,k})$. For example, based on previous work by Longuett-Higgins and Salem [61], $\beta_{\ell,k}$ has an exponential dependence on $\mathbf{R}_{\ell,k}$, $\beta_{\ell,k}^{exp}=-\beta^{0}_{\ell,k}\exp^{-\frac{\Delta R_{\ell,k}}{y_{\ell,k}}}.$ (8) A second functional form, which is based on the work of Su, Schrieffer and Heege [62, 63] on conducting polymers, uses a linear distance-dependence of the interactions, $\beta_{\ell,k}^{lr}=-\beta^{0}_{\ell,k}\left(1-y^{-1}_{\ell,k}\Delta R_{\ell,k}\right).$ (9) For both expressions (see Eqs. 8–9), $\Delta R_{\ell,k}$ is the difference with respect to the reference bond length distance $R^{0}_{\ell,k}$, and $y_{\ell,k}$ is a length scale parameter. By including $R^{0}_{\ell,k}$ and $y_{\ell,k}$ in the set of parameters for the Hückel model, the complete set of parameters becomes $\boldsymbol{\theta}=[\alpha_{\ell},\beta^{0}_{\ell,k},y_{\ell,k},R^{0}_{\ell,k}]$. For this work, all initial $\alpha_{\ell}$ and $\beta^{0}_{\ell,k}$ parameters were taken from Van-Catledge [10], and the initial $R^{0}_{\ell,k}$ parameters were approximated from tables of standard bond lengths [64]. The length scale parameters ($y_{\ell,k}$) were initially set to 0.3 Å, which corresponds to the value that has been used for C–C in the literature [65, 66]. We used a subset of the GDB-13 data [67] set that only consists of molecules with $\pi$-systems for fitting our model parameters (see Supplementary Material for details on how the dataset was generated). Note that some molecules in the dataset could have n-$\pi$* transitions as their lowest excited state. We used a pool of 60,000 molecules and randomly sampled 100, 1,000, and 5,000 molecules from this set, and used 1,000 additional molecules as validation set to monitor the optimization procedure. To optimize $\boldsymbol{\theta}$, we used the mean squared error as a loss function, ${\cal L}(\boldsymbol{\theta})=\frac{1}{2}\sum_{i}^{N}\left(\hat{\epsilon}_{HL}({\cal M}_{i})-\epsilon_{HL}(\boldsymbol{\theta};{\cal M}_{i})\right)^{2},$ (10) where ${\cal M}_{i}$ is a single molecule of the training set, and $\hat{\epsilon}_{HL}$ is the vertical excitation energy between the ground state and the first excited singlet state computed at the TDA- SCS-$\omega$PBEPP86/def2-SVP level of theory [68, 69]. At the Hückel level of theory, this excitation energy simply corresponds to the HOMO-LUMO gap ($\epsilon_{HL}$) due to the disregard for electron correlation. To compare the prediction of $\epsilon_{HL}$ with the DFT reference values properly, we linearly transformed the results of the Hückel model using two additional parameters, $w_{0}$ and $w_{1}$ (Eq. 11), $\epsilon_{HL}(\boldsymbol{\theta};{\cal M}_{i})=w_{1}\times\epsilon_{HL}(\alpha_{\ell},\beta^{0}_{\ell,k},y_{\ell,k},R^{0}_{\ell,k};{\cal M}_{i})+w_{0},$ (11) where $\boldsymbol{\theta}$ jointly represents all parameters of the model, i.e., $\boldsymbol{\theta}=[\alpha_{\ell},\beta^{0}_{\ell,k},y_{\ell,k},R^{0}_{\ell,k},w_{0},w_{1}]$. For the optimization of all free parameters, we used the AdamW optimization algorithm [46], as implemented in the Optax library [70], with a learning rate of $0.02$, and a weight decay of $10^{-4}$. Notably, we considered various training scenarios that included different values for the weight decay, and the regularization of different sets of the Hückel parameters. However, we found no impact on the accuracy of the model. The initial model parameters were gathered from Refs. [10, 64]. We optimized the parameters of three different Hückel models, i) the original one where $\beta_{\ell,k}$ is distance-independent ($\beta_{\ell,k}=\beta^{0}_{\ell,k}$), and both ii) the exponential (Eq. 8), and iii) the linear (Eq. 9) distance-dependence functional forms. We want to emphasize that any other analytic form for $\beta_{\ell,k}$ could be considered as well as AD makes any of these expressions fully differentiable. Following the convention in the literature, we scaled the parameters $\beta^{0}_{\ell,k}$ and $\alpha_{\ell}$ with respect to the carbon atom parameters according to $\alpha_{\ell}=\alpha_{\ell}-\alpha_{C}$, and $\beta^{0}_{\ell,k}=\beta^{0}_{\ell,k}/\beta^{0}_{C,C}$. Notably, at least for our results in this work, we found that including a regularization term in the loss function did not impact the accuracy of the model. Finally, we found 20 epochs to be enough to minimize the loss function when the parameters are initialized with values from Refs. [10, 64]. In Figures 7–10 we display the optimized values of the parameters for the three different Hückel models considered. From the optimized parameters, we observe that $\alpha_{O}$ (i.e., the 2p orbital energy parameter for oxygen), for all three models, is the one that differs the most from the literature [10, 64]. While there is no good reference data for $y_{\ell,k}$ to compare to, we observe, nevertheless, that the C–C parameter value changes considerably from the initial value of 0.3. For the $\beta^{0}_{\ell,k}$ parameters, only the values for N–C resemble the literature values. Furthermore, the optimal values of $R^{0}_{\ell,k}$ change the least from the values in Ref. [64]. Using these optimized parameters, we predicted $\epsilon_{HL}$ for 40,000 additional test set molecules and compare the results with DFT reference data. The direct comparison is depicted in Figures 11–13. By optimizing all parameters there is a significant improvement in the prediction of $\epsilon_{HL}$ using our semi-empirical model. Notably, we also found that considering a larger training data set does not impact the accuracy of the Hückel model which suggests that either the corresponding molecules do not provide any additional information with respect to the relevant interaction parameters or that the model already is close to its best expected performance and cannot be improved further. Another important observation in that regard is that the analytical form of $g(\mathbf{R}_{\ell,k})$ in $\beta_{\ell,k}$ does not impact the accuracy of the model when optimized parameters are used. Next, we also optimized the parameters with respect to the polarizability for the distance-independent Hückel model. Here, the gradients needed for training are of third-order, e.g., $\frac{\partial}{\partial\alpha_{\ell}}\frac{\partial^{2}\tilde{\alpha}_{kk}}{\partial F^{2}_{k}}$ or $\frac{\partial}{\partial\beta_{k,\ell}}\frac{\partial^{2}\tilde{\alpha}_{kk}}{\partial F^{2}_{k}}$, and can be computed more efficiently via AD, illustrating the potential of this approach. The molecular polarizabilities that were used as reference data were computed using dftd4 (version 3.4.0) [71, 72, 73] via the default methodology summing atomic polarizabilities. Even though we observe a higher prediction accuracy when the optimized parameters are used compared to the model before parameter refinement, as depicted in Fig. 14, the accuracy of the model still remains relatively low and does not improve anymore when more training data is used. We suspect that this prediction task is particularly challenging for the Huc̈kel model as the simulated polarizability only corresponds to the contribution from $\pi$-electrons, while that of the reference data accounts for all the electrons in the molecules. Even though we expect a significant portion of the molecular polarizabilities to stem from the $\pi$-electrons, the contributions of the $\sigma$-electrons cannot be neglected and can dominate this property. Nevertheless, this proof-of-concept application example demonstrates the operational ease of conducting parameter refinement of a given physics-based prediction model based on reference data, even when derivative properties are targeted. Figure 7: Optimized $\alpha_{\ell}$ parameters for different Hückel models, a) $\beta^{0}_{\ell,k}$, b) $\beta^{lr}_{\ell,k}$ (Eq. 9), and c) $\beta^{exp}_{\ell,k}$ (Eq. 8). All parameters reported were averaged over ten different training data sets, and different number of training molecules. Colored symbols and bars represent the mean and standard deviation of the optimized parameters averaged over ten different data sets. The reference parameters ($\blacksquare$-symbol) were taken from Refs. [10], and used as the initial parameters for all optimizations. We refer the reader to the main text for the optimization details. Figure 8: Optimized $\beta^{0}_{\ell,k}$ parameters for different Hückel models, a) $\beta^{0}_{\ell,k}$, b) $\beta^{lr}_{\ell,k}$ (Eq. 9), and c) $\beta^{exp}_{\ell,k}$ (Eq. 8). All parameters reported were averaged over ten different training data sets, and different number of training molecules. Colored symbols and bars represent the mean and standard deviation of the optimized parameters averaged over ten different data sets. The reference parameters ($\blacksquare$-symbol) were taken from Refs. [10], and used as the initial parameters for all optimizations. We refer the reader to the main text for the optimization details. Figure 9: Optimized $y_{\ell,k}$ parameters for different Hückel models, a) $\beta^{lr}_{\ell,k}$ (Eq. 9), and b) $\beta^{exp}_{\ell,k}$ (Eq. 8). All parameters reported were averaged over ten different training data sets, and different number of training molecules. Colored symbols and bars represent the mean and standard deviation of the optimized parameters averaged over ten different data sets.The reference parameters ($\blacksquare$-symbol) were set to 0.3 Å, and used as the initial parameters for all optimizations. We refer the reader to the main text for the optimization details. Figure 10: Optimized $R^{0}_{\ell,k}$ parameters for different Hückel models, a) $\beta^{lr}_{\ell,k}$ (Eq. 9), and b) $\beta^{exp}_{\ell,k}$ (Eq. 8). All parameters reported were averaged over ten different training data sets, and different number of training molecules. Colored symbols and bars represent the mean and standard deviation of the optimized parameters averaged over ten different data sets. The reference parameters ($\blacksquare$-symbol) were taken from Refs. [64], and used as the initial parameters for all optimizations. We refer the reader to the main text for the optimization details. Figure 11: $\epsilon_{HL}$ predicted with the Hückel models and with DFT level for 40K molecules not considered during training. The atom-atom interaction of the Hückel model is described by a distance-independent parameter, $\beta^{0}_{\ell,k}$. Results computed with a Hückel with parameters taken from the literature (a), and parameters optimized with N=100 (b) and N=5,000 (c) data points. We refer the reader to the main text for the optimization details. Figure 12: $\epsilon_{HL}$ predicted with the Hückel models and with DFT level for 40K molecules not considered during training. The atom-atom interaction of the Hückel model is described by a distance-dependent parameter, $\beta^{lr}_{\ell,k}$ (Eq. 9). Results computed with a Hückel with parameters taken from the literature (a), and parameters optimized with N=100 (b) and N=5,000 (c) data points. We refer the reader to the main text for the optimization details. Figure 13: $\epsilon_{HL}$ predicted with the Hückel models and with DFT level for 40K molecules not considered during training. The atom-atom interaction of the Hückel model is described by a distance-dependent parameter, $\beta^{exp}_{\ell,k}$ (Eq. 8). Results computed with a Hückel model with parameters taken from the literature (a), and parameters optimized with N=100 (b) and N=5,000 (c) data points. We refer the reader to the main text for the optimization details. Figure 14: $\langle\tilde{\alpha}\rangle$ predicted with the Hückel models and with DFT level for 40K molecules not considered during training. The atom-atom interaction of the Hückel model is described by a distance-independent parameter, $\beta^{0}_{\ell,k}$. Results computed with a Hückel model with parameters taken from the literature (a), and (b) parameters optimized with N=100 data points. We refer the reader to the main text for the optimization details. ## 4 Conclusions In this work, we demonstrate the power of automatic differentiation to enable the efficient use of physics-inspired models for gradient-based optimization problems in the realm of molecular chemistry via semi-empirical Hückel models. In particular, we showcase inverse molecular design via an alchemical problem formulation using fixed molecular frameworks. This allows us to perform structure optimization requiring only a very small number of intermediate structures to find local minima with respect to the properties of interest utilizing gradients with respect to atom identities at specific sites. While our approach is currently limited to a fixed molecular framework, performing optimizations over the molecular composition space alone is far from trivial. Compared to various alternative approaches, our implementation shows a remarkably high molecular sampling efficiency due to efficient utilization of gradient information in combination with powerful gradient-based optimization algorithms. Additionally, we showcase the ease of generating calibrated physics-based property prediction models using high quality reference training data of relatively modest size, again allowing for quick convergence of model parameters. This is particularly important as most physical models that rely on empirical parameters such as semi-empirical quantum chemistry models and density functional approximations are still largely optimized by hand, making the corresponding procedures tedious. Thus, we believe that our work will serve as an inspiration for the field of computational chemistry in order to adopt the readily available AD capabilities of mature ML programming frameworks allowing to accelerate the construction of ever more accurate physics-based property simulation models. ## Acknowledgments R.P. acknowledges funding through a Postdoc.Mobility fellowship by the Swiss National Science Foundation (SNSF, Project No. 191127). A.A.-G. thanks Anders G. Frøseth for his generous support. A.A.-G. acknowledges the generous support of Natural Resources Canada and the Canada 150 Research Chairs program. We also thank the SciNet HPC Consortium for support regarding the use of the Niagara supercomputer. SciNet is funded by the Canada Foundation for Innovation, the Government of Ontario, Ontario Research Fund - Research Excellence, and the University of Toronto. ## Data Availability The data that support the findings of this study are available within the article and its supplementary material and in GitHub (inverse molecular design) https://github.com/RodrigoAVargasHdz/huxel_molecule_desing and (parameter optimization) https://github.com/RodrigoAVargasHdz/huxel. ## References ## References * [1] R. S. Mulliken, “Molecular scientists and molecular science: Some reminiscences,” Journal of Chemical Physics, vol. 43, pp. S2–S11, Nov. 1965. * [2] P.-O. Löwdin, “Present situation of quantum chemistry,” Journal of Physical Chemistry, vol. 61, pp. 55–68, Jan. 1957. * [3] W. Thiel, “Semiempirical quantum–chemical methods,” WIREs Computational Molecular Science, vol. 4, no. 2, pp. 145–157, 2014. * [4] E. Hückel, “Quantentheoretische beiträge zum benzolproblem,” Zeitschrift für Physik, vol. 70, pp. 204–286, Mar 1931. * [5] E. Hückel, “Quanstentheoretische beiträge zum benzolproblem,” Zeitschrift für Physik, vol. 72, pp. 310–337, May 1931. * [6] E. Hückel, “Quantentheoretische beiträge zum problem der aromatischen und ungesättigten verbindungen. iii,” Zeitschrift für Physik, vol. 76, pp. 628–648, Sep 1932. * [7] E. Hückel, “Die freien radikale der organischen chemie,” Zeitschrift für Physik, vol. 83, pp. 632–668, Sep 1933. * [8] A. Streitwieser, Molecular Orbital Theory for Organic Chemists. New York,: Wiley, 1961. * [9] B. Andes Hess, L. Schaad, and C. Holyoke, “On the aromaticity of annulenones,” Tetrahedron, vol. 28, pp. 5299–5305, Jan. 1972. * [10] F. A. Van-Catledge, “A Pariser-Parr-Pople-based set of Hückel molecular orbital parameters,” The Journal of Organic Chemistry, vol. 45, pp. 4801–4802, Nov. 1980. * [11] I. Fleming, Molecular Orbitals and Organic Chemical Reactions: Reference Edition. Chichester, United Kingdom: Wiley, 2010. * [12] P. Klán and J. Wirz, Photochemistry of Organic Compounds: From Concepts to Practice. Wiley, 2009. * [13] C. S. Anstöter and J. R. R. Verlet, “A Hückel model for the excited-sate dynamics of a protein chromophore developed using photoelectron imaging,” Accounts of Chemical Research, vol. 55, no. 9, pp. 1205–1213, 2022. * [14] J. A. N. F. Gomes and R. B. Mallion, “Aromaticity and ring currents,” Chemical Reviews, vol. 101, no. 5, pp. 1349–1384, 2001. * [15] B. Sanchez-Lengeling and A. Aspuru-Guzik, “Inverse molecular design using machine learning: Generative models for matter engineering,” Science, vol. 361, pp. 360–365, July 2018. * [16] D. Xiao, W. Yang, and D. N. Beratan, “Inverse molecular design in a tight-binding framework,” The Journal of Chemical Physics, vol. 129, no. 4, p. 044106, 2008. * [17] A. G. Baydin, B. A. Pearlmutter, A. A. Radul, and J. M. Siskind, “Automatic differentiation in machine learning: a survey,” J. Mach. Learn. Res., vol. 18, 2018. * [18] T. Tamayo-Mendoza, C. Kreisbeck, R. Lindh, and A. Aspuru-Guzik, “Automatic differentiation in quantum chemistry with applications to fully variational hartree–fock,” ACS Central Science, vol. 4, no. 5, pp. 559–566, 2018\. * [19] X. Zhang and G. K.-L. Chan, “Differentiable quantum chemistry with pyscf for molecules and materials at the mean-field level and beyond,” 2022. * [20] M. F. Kasim, S. Lehtola, and S. M. Vinko, “Dqc: a python program package for differentiable quantum chemistry,” 2021. * [21] L. Li, S. Hoyer, R. Pederson, R. Sun, E. D. Cubuk, P. Riley, and K. Burke, “Kohn-sham equations as regularizer: Building prior knowledge into machine-learned physics,” Phys. Rev. Lett., vol. 126, p. 036401, Jan 2021\. * [22] M. F. Kasim and S. M. Vinko, “Learning the exchange-correlation functional from nature with fully differentiable density functional theory,” Phys. Rev. Lett., vol. 127, p. 126403, Sep 2021. * [23] L. Zhao and E. Neuscamman, “Excited state mean-field theory without automatic differentiation,” The Journal of Chemical Physics, vol. 152, no. 20, p. 204112, 2020. * [24] G. Zhou, B. Nebgen, N. Lubbers, W. Malone, A. M. N. Niklasson, and S. Tretiak, “Graphics processing unit-accelerated semiempirical Born Oppenheimer molecular dynamics using PyTorch,” Journal of Chemical Theory and Computation, vol. 16, pp. 4951–4962, Aug. 2020. * [25] T. Zubatiuk, B. Nebgen, N. Lubbers, J. S. Smith, R. Zubatyuk, G. Zhou, C. Koh, K. Barros, O. Isayev, and S. Tretiak, “Machine learned hückel theory: Interfacing physics and deep neural networks,” The Journal of Chemical Physics, vol. 154, no. 24, p. 244108, 2021. * [26] C. H. Pham, R. K. Lindsey, L. E. Fried, and N. Goldman, “High-accuracy semiempirical quantum models based on a minimal training set,” The Journal of Physical Chemistry Letters, vol. 13, no. 13, pp. 2934–2942, 2022\. PMID: 35343698. * [27] G. Zhou, N. Lubbers, K. Barros, S. Tretiak, and B. Nebgen, “Deep learning of dynamically responsive chemical hamiltonians with semiempirical quantum mechanics,” Proceedings of the National Academy of Sciences, vol. 119, no. 27, p. e2120333119, 2022. * [28] A. Yachmenev and S. N. Yurchenko, “Automatic differentiation method for numerical construction of the rotational-vibrational hamiltonian as a power series in the curvilinear internal coordinates using the eckart frame,” The Journal of Chemical Physics, vol. 143, no. 1, p. 014105, 2015. * [29] A. S. Abbott, B. Z. Abbott, J. M. Turney, and H. F. Schaefer, “Arbitrary-order derivatives of quantum chemical methods via automatic differentiation,” The Journal of Physical Chemistry Letters, vol. 12, no. 12, pp. 3232–3239, 2021\. PMID: 33764068. * [30] V. Bergholm, J. Izaac, M. Schuld, C. Gogolin, M. S. Alam, S. Ahmed, J. M. Arrazola, C. Blank, A. Delgado, S. Jahangiri, K. McKiernan, J. J. Meyer, Z. Niu, A. Száva, and N. Killoran, “Pennylane: Automatic differentiation of hybrid quantum-classical computations,” 2020. * [31] F. Pavošević and S. Hammes-Schiffer, “Automatic differentiation for coupled cluster methods,” 2020. * [32] G. F. von Rudorff, “Arbitrarily accurate quantum alchemy,” The Journal of Chemical Physics, vol. 155, no. 22, p. 224103, 2021. * [33] N. Fedik, R. Zubatyuk, M. Kulichenko, N. Lubbers, J. S. Smith, B. Nebgen, R. Messerly, Y. W. Li, A. I. Boldyrev, K. Barros, O. Isayev, and S. Tretiak, “Extending machine learning beyond interatomic potentials for predicting molecular properties,” Nature Reviews Chemistry, Aug 2022. * [34] R. A. Vargas-Hernández, R. T. Q. Chen, K. A. Jung, and P. Brumer, “Fully differentiable optimization protocols for non-equilibrium steady states,” New Journal of Physics, vol. 23, p. 123006, dec 2021. * [35] R. A. Vargas-Hernández, R. T. Q. Chen, K. A. Jung, and P. Brumer, “Inverse design of dissipative quantum steady-states with implicit differentiation,” 2020\. * [36] N. Yoshikawa and M. Sumita, “Automatic differentiation for the direct minimization approach to the hartree–fock method,” The Journal of Physical Chemistry A, vol. 126, no. 45, pp. 8487–8493, 2022. PMID: 36346835. * [37] S. Bac, A. Patra, K. J. Kron, and S. Mallikarjun Sharada, “Recent advances toward efficient calculation of higher nuclear derivatives in quantum chemistry,” The Journal of Physical Chemistry A, vol. 126, no. 43, pp. 7795–7805, 2022. PMID: 36282088. * [38] C. R. Harris, K. J. Millman, S. J. van der Walt, R. Gommers, P. Virtanen, D. Cournapeau, E. Wieser, J. Taylor, S. Berg, N. J. Smith, R. Kern, M. Picus, S. Hoyer, M. H. van Kerkwijk, M. Brett, A. Haldane, J. F. del Río, M. Wiebe, P. Peterson, P. Gérard-Marchant, K. Sheppard, T. Reddy, W. Weckesser, H. Abbasi, C. Gohlke, and T. E. Oliphant, “Array programming with NumPy,” Nature, vol. 585, pp. 357–362, Sept. 2020. * [39] J. Bradbury, R. Frostig, P. Hawkins, M. J. Johnson, C. Leary, D. Maclaurin, and S. Wanderman-Milne, “JAX: composable transformations of Python+NumPy programs,” 2018. * [40] B. Ramsundar, D. Krishnamurthy, and V. Viswanathan, “Differentiable physics: A position piece,” 2021. * [41] M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro, G. S. Corrado, A. Davis, J. Dean, M. Devin, S. Ghemawat, I. Goodfellow, A. Harp, G. Irving, M. Isard, Y. Jia, R. Jozefowicz, L. Kaiser, M. Kudlur, J. Levenberg, D. Mané, R. Monga, S. Moore, D. Murray, C. Olah, M. Schuster, J. Shlens, B. Steiner, I. Sutskever, K. Talwar, P. Tucker, V. Vanhoucke, V. Vasudevan, F. Viégas, O. Vinyals, P. Warden, M. Wattenberg, M. Wicke, Y. Yu, and X. Zheng, “TensorFlow: Large-scale machine learning on heterogeneous systems,” 2015. Software available from tensorflow.org. * [42] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, A. Desmaison, A. Köpf, E. Yang, Z. DeVito, M. Raison, A. Tejani, S. Chilamkurthy, B. Steiner, L. Fang, J. Bai, and S. Chintala, “Pytorch: An imperative style, high-performance deep learning library,” 2019. * [43] K. YATES, “Ii - hÜckel molecular orbital theory,” in Hückel Molecular Orbital Theory (K. YATES, ed.), pp. 27–87, Academic Press, 1978. * [44] J. Pople and D. Beveridge, Approximate Molecular Orbital Theory. McGraw-Hill series in advanced chemistry, McGraw-Hill, 1970. * [45] M. Blondel, Q. Berthet, M. Cuturi, R. Frostig, S. Hoyer, F. Llinares-López, F. Pedregosa, and J.-P. Vert, “Efficient and modular implicit differentiation,” arXiv preprint arXiv:2105.15183, 2021. * [46] I. Loshchilov and F. Hutter, “Decoupled weight decay regularization,” in International Conference on Learning Representations, 2019. * [47] F. Eisenlohr, “Eine neuberechnung der atomrefraktionen. i,” Zeitschrift für Physikalische Chemie, vol. 75U, no. 1, pp. 585–607, 1911. * [48] A. L. von Steiger, “Ein beitrag zur summationsmethodik der molekularrefraktionen, besonders bei aromatischen kohlenwasserstoffen,” Berichte der deutschen chemischen Gesellschaft (A and B Series), vol. 54, no. 6, pp. 1381–1393, 1921. * [49] P. Schwerdtfeger and J. K. Nagle, “2018 table of static dipole polarizabilities of the neutral elements in the periodic table,” Molecular Physics, vol. 117, no. 9-12, pp. 1200–1225, 2019. * [50] A. Krawczuk, D. Pérez, and P. Macchi, “PolaBer: a program to calculate and visualize distributed atomic polarizabilities based on electron density partitioning,” Journal of Applied Crystallography, vol. 47, pp. 1452–1458, Aug 2014. * [51] E. Heilbronner and H. Bock, The HMO Model and Its Application: Basis and Manipulation. Wiley, 1976. * [52] W. Gao, T. Fu, J. Sun, and C. W. Coley, “Sample efficiency matters: A benchmark for practical molecular optimization,” arXiv preprint arXiv:2206.12411, 2022. * [53] F. Häse, L. M. Roch, and A. Aspuru-Guzik, “Chimera: enabling hierarchy based multi-objective optimization for self-driving laboratories,” Chem. Sci., vol. 9, pp. 7642–7655, 2018. * [54] R. J. Hickman, M. Aldeghi, F. Häse, and A. Aspuru-Guzik, “Bayesian optimization with known experimental and design constraints for chemistry applications,” 2022. * [55] M. Seifrid, R. J. Hickman, A. Aguilar-Granda, C. Lavigne, J. Vestfrid, T. C. Wu, T. Gaudin, E. J. Hopkins, and A. Aspuru-Guzik, “Routescore: Punching the ticket to more efficient materials development,” ACS Central Science, vol. 8, no. 1, pp. 122–131, 2022. * [56] H. Tao, T. Wu, S. Kheiri, M. Aldeghi, A. Aspuru-Guzik, and E. Kumacheva, “Self-driving platform for metal nanoparticle synthesis: Combining microfluidics and machine learning,” Advanced Functional Materials, vol. 31, no. 51, p. 2106725, 2021. * [57] R. A. Vargas-Hernández, C. Chuang, and P. Brumer, “Multi-objective optimization for retinal photoisomerization models with respect to experimental observables,” The Journal of Chemical Physics, vol. 155, no. 23, p. 234109, 2021. * [58] J.-A. Désidéri, “Multiple-gradient descent algorithm (mgda) for multiobjective optimization,” Comptes Rendus Mathematique, vol. 350, no. 5, pp. 313–318, 2012. * [59] G. F. von Rudorff and O. A. von Lilienfeld, “Alchemical perturbation density functional theory,” Phys. Rev. Research, vol. 2, p. 023220, May 2020. * [60] C. Shen, M. Krenn, S. Eppel, and A. Aspuru-Guzik, “Deep molecular dreaming: inverse machine learning for de-novo molecular design and interpretability with surjective representations,” Machine Learning: Science and Technology, vol. 2, p. 03LT02, jul 2021. * [61] H. Longuet-Higgins and L. Salem, “The alternation of bond lengths in long conjugated chain molecules,” Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences, vol. 251, pp. 172–185, May 1959. * [62] W. P. Su, J. R. Schrieffer, and A. J. Heeger, “Solitons in polyacetylene,” Phys. Rev. Lett., vol. 42, pp. 1698–1701, Jun 1979. * [63] A. J. Heeger, S. Kivelson, J. R. Schrieffer, and W. P. Su, “Solitons in conducting polymers,” Rev. Mod. Phys., vol. 60, pp. 781–850, Jul 1988\. * [64] F. H. Allen, O. Kennard, D. G. Watson, L. Brammer, A. G. Orpen, and R. Taylor, “Tables of bond lengths determined by x-ray and neutron diffraction. part 1. bond lengths in organic compounds,” J. Chem. Soc., Perkin Trans. 2, pp. S1–S19, 1987. * [65] L. Z. Stolarczyk and T. M. Krygowski, “Augmented hückel molecular orbital model of $\pi$-electron systems: from topology to metric. i. general theory,” Journal of Physical Organic Chemistry, vol. 34, no. 3, p. e4154, 2021. e4154 poc.4154. * [66] J. H. Kwapisz and L. Z. Stolarczyk, “Applications of hückel-su-schrieffer-heeger method,” Structural Chemistry, vol. 32, pp. 1393–1406, Aug 2021. * [67] L. C. Blum and J.-L. Reymond, “970 million druglike small molecules for virtual screening in the chemical universe database gdb-13,” Journal of the American Chemical Society, vol. 131, no. 25, pp. 8732–8733, 2009. * [68] M. Casanova-Páez and L. Goerigk, “Time-dependent long-range-corrected double-hybrid density functionals with spin-component and spin-opposite scaling: A comprehensive analysis of singlet–singlet and singlet–triplet excitation energies,” Journal of Chemical Theory and Computation, vol. 17, no. 8, pp. 5165–5186, 2021. PMID: 34291643. * [69] F. Weigend and R. Ahlrichs, “Balanced basis sets of split valence, triple zeta valence and quadruple zeta valence quality for h to rn: Design and assessment of accuracy,” Phys. Chem. Chem. Phys., vol. 7, pp. 3297–3305, 2005. * [70] M. Hessel, D. Budden, F. Viola, M. Rosca, E. Sezener, and T. Hennigan, “Optax: composable gradient transformation and optimisation, in jax!,” 2020. * [71] E. Caldeweyher, C. Bannwarth, and S. Grimme, “Extension of the d3 dispersion coefficient model,” The Journal of Chemical Physics, vol. 147, no. 3, p. 034112, 2017. * [72] E. Caldeweyher, S. Ehlert, A. Hansen, H. Neugebauer, S. Spicher, C. Bannwarth, and S. Grimme, “A generally applicable atomic-charge dependent london dispersion correction,” The Journal of Chemical Physics, vol. 150, no. 15, p. 154122, 2019. * [73] E. Caldeweyher, J.-M. Mewes, S. Ehlert, and S. Grimme, “Extension and evaluation of the d4 london-dispersion model for periodic systems,” Phys. Chem. Chem. Phys., vol. 22, pp. 8499–8512, 2020.
# Electron-phonon interactions in the Andreev Bound States of aluminum nanobridge Josephson junctions James T. Farmer<EMAIL_ADDRESS>Azarin Zarassi Sadman Shanto Darian Hartsell Eli M. Levenson-Falk<EMAIL_ADDRESS>Department of Physics, University of Southern California Center for Quantum Information Science and Technology, University of Southern California ###### Abstract We report continuous measurements of quasiparticles trapping and clearing from Andreev Bound States in aluminum nanobridge Josephson junctions integrated into a superconducting-qubit-like device. We find that trapping is well modeled by independent spontaneous emission events. Above 80 mK the clearing process is well described by absorption of thermal phonons, but other temperature-independent mechanisms dominate at low temperature. We find complex structure in the dependence of the low-temperature clearing rate on the Andreev Bound State energy. Our results shed light on quasiparticle behavior in qubit-like circuits. ††preprint: APS/123-QED Non-equilibrium quasiparticles (QPs) in superconducting quantum circuits can hinder device operation, limiting coherence in most qubit architectures [1, 2] and inducing correlated, difficult-to-correct errors across multiple qubits on the same chip [3, 4]. The QPs are generated by non-thermal mechanisms such as pair-breaking infrared photons [5] or energy dissipation from local radioactivity and cosmic rays [2]. Significant non-equilibrium QP populations with fractional densities $x_{qp}\sim 10^{-9}-10^{-5}$ are ubiquitously observed [5, 6, 7, 8] and have proven difficult to eliminate. Mitigation strategies such as improved light-tight shielding [5], input/output filtering with infrared absorbers [9, 10, 7], and device engineering [11, 12, 13, 14, 15, 16, 17] have reduced QP densities over the last decade. Many works have probed QP populations by detecting single charge tunneling across Josephson junctions [18, 19, 20, 21, 22, 1, 7, 15] or observing QPs trapped inside the Andreev Bound States (ABS) of a junction [23, 24, 25, 26, 27]. These ABS provide a complementary measurement of QP behavior, and can be used as qubit modes themselves [28, 29]. In many implementations the ABS qubit relies on a non-equilibrium QP trapping in order to initialize the state; such qubits are vulnerable to additional trapping events and to accidental clearing of the QP from the ABS. There is thus a great need to better understand the behavior of QPs in ABS and the mechanisms for QPs transitions between ABS and bulk continuum states. In this letter we investigate the electron-phonon interactions involved in trapping a quasiparticle into / clearing a quasiparticle from an Andreev Bound State. We show continuous, real-time measurements of ABS trapping dynamics as a function of ABS energy and device temperature in a superconducting-qubit- like device. We find that QP trapping is consistent with independent spontaneous emission events from a bulk QP population that is a combination of a temperature-independent non-equilibrium background and a thermal equilibrium density. We further find that QP clearing from an ABS is consistent with a process dominated by absorption of a thermal phonon at temperatures above 80 mK. At low temperatures, we find evidence that absorption of microwave photons by trapped QPs is the dominant clearing mechanism, even at low drive powers. We analyze the mean QP occupancy of our ABS device and find independent confirmation of our trapping and clearing models. Our results shed light on quasiparticle behavior in ABS and in qubit-like circuits in general. To study QP trapping, we require a circuit element which is sensitive to the occupation of single electron states with tunability below the superconducting gap. We find such an element in the aluminum nanobridge Josephson junction, an all-superconducting junction which was shown [30, 31, 32] to follow the KO-1 current-phase relation [33] while providing several hundred conduction channels. Each conduction channel hosts a pair of ABS with energies $E_{A}(\delta)=\pm\Delta\sqrt{1-\tau\sin{\frac{\delta}{2}}}$ (1) measured from the Fermi energy. The transparency $\tau$ is the probability that an incident Cooper pair is transmitted across the junction and $\delta$ is the phase bias across the junction; $\Delta$ is the superconducting gap. For short ($\lesssim 100$ nm) aluminum nanobridges, $\tau$ approximately follows a Dorokhov distribution with a strong preference to be 0 or 1 [31, 34]. When occupied, each ABS in a given channel carries equal and opposite contributions to the supercurrent. The negative state is usually occupied while the positive state is unoccupied. However, the positive ABS dips below the gap $\Delta$ when both $\delta$ and $\tau$ are nonzero, making it energetically favorable for a quasiparticle above the gap (i.e. in the bulk continuum) to relax into the ABS and become trapped. When this occurs the supercurrent contribution of the given channel is cancelled and the channel is “poisoned". This is the mechanism of our detection: the Josephson inductance becomes a function of the number of trapped quasiparticles. By embedding a DC SQUID with symmetric aluminum nanobridge junctions in a $\lambda/4$ coplanar waveguide resonator, we are able to measure the trapping of a QP as a resonant frequency shift of the resonator. This allows for a high bandwidth, continuous measurement of the ABS occupation in a qubit-like circuit using a standard dispersive measurement setup [27]. A constant flux bias on the SQUID introduces a constant, symmetric phase bias to the junctions $\delta=\pi\phi$ (where $\phi$ is the applied flux in units of flux quanta), tuning the ABS energies. The fundamental mode $f_{0}(\phi)$ of our resonator is flux tunable from 4.301 GHz to 4̃.25 GHz with a linewidth $\kappa=2\pi\times 250$ kHz and the shift due to trapping a single quasiparticle $\chi(\phi)/2\pi$ ranges from 100 kHz to 400 kHz. We perform continuous microwave reflection measurements on our device which is mounted in a dilution refrigerator with a base temperature of 30 mK. The reflected signal is homodyne demodulated with an IQ mixer and the two quadratures of signal are recorded as a gapless voltage record in 3 s segments by an Alazar 9371 digitizer operating at 300 MHz sample rate. This is down sampled to 1 MHz sample rate before fitting each record to a Hidden Markov Model (HMM) [35, 36] with further down sampling if required to maintain signal to noise ratio greater than 3. The HMM analysis models the tranisition rates between each trapped QP number and is used to extract a time series of the number of trapped QPs from the continuous voltage record. Data was collected and processed in this way over a range of parameters: the dilution refrigerator temperature, the ABS energy, and the applied microwave power. For brevity, we restrict ourselves in this analysis to a constant power of -133 dBm ($\sim$ 25 photons) at the device. More details are given in the Supplement [37]. We present three quantities of interest: $\Gamma_{trap}$ is the rate of QPs relaxing from the bulk into available ABS of the junction, $\Gamma_{release}$ is the rate of clearing QPs from ABS to the bulk, and $\bar{n}$ which we call the mean occupation is the time average of the number of trapped QPs. $\Gamma_{trap}$ and $\Gamma_{release}$ are found from the off-diagonal elements of the HMM transition matrix—that is, they are parameters of the model used to extract the ABS occupation time series—while $\bar{n}$ is found from averaging the extracted occupation over the full 3 second record. We begin our modeling with the trap rate. Assuming trapping events are independent of each other and spontaneous emission dominates the QP relaxation into the ABS, each QP in the bulk has a temperature-independent trapping rate. This implies the overall trap rate is separable: $\Gamma_{trap}(\Delta_{A},T)=f(\Delta_{A})x(T)$, where $x(T)$ is the fractional quasiparticle density and $\Delta_{A}\equiv\Delta-E_{A}$ is the trap depth. We take the limit $\tau\rightarrow 1$ as the Dorkhov distribution $\rho(\tau)$ is sharply peaked at 0 and 1, and channels with 0 transmittivity do not contribute to the transport. The fractional quasiparticle density should be the sum of a non-equilibrium background $x_{ne}$ and a thermal population: $x(T)=x_{ne}+\sqrt{\frac{2\pi k_{B}T}{\Delta}}\exp{\left(\frac{-\Delta}{k_{B}T}\right)}.$ (2) We expect that most bulk QPs are near the gap energy, so for spontaneous emission we take $f(\Delta_{A})\propto\Delta_{A}^{3}$. Putting this together, we obtain a model for the trap rate $\Gamma_{trap}=\beta\Delta_{A}^{3}\left(x_{ne}+\sqrt{\frac{2\pi k_{B}T}{\Delta}}\exp\left(\frac{-\Delta}{k_{B}T}\right)\right)$ (3) where $\beta$, $\Delta$, and $x_{ne}$ are the free parameters. To improve the quality of our fit, we take advantage of the low temperature saturation of trap rate $\Gamma_{trap}^{0}(\Delta_{A})\approx\beta\Delta_{A}^{3}x_{ne}$ for $T\leq 120$ mK. We first subtract $\Gamma_{trap}^{0}(\Delta_{A})$ from Eq. 3 and fit the resulting quantity to find the gap $\Delta$ and scaling factor $\beta$. Next we divide Eq. 3 by $\Gamma_{trap}^{0}(\Delta_{A})$ and fit this normalized rate with the fractional non-equilibrium density $x_{ne}$ as the only free parameter. This fitting procedure is covered in detail in the supplement [37]. In Figure 1, we show the full model (Eq. 3) using the combined results of this fitting procedure. Figure 1: Measured trap rate (circles) and model (solid lines). The dependence on the trap depth $\Delta_{A}$ is shown on the left, while temperature dependence is shown on the right. We note the peak in 30 mK data around 9 GHz on the left was observed as a period of significantly larger than normal mean occupation which lasted approximately 1 hour in laboratory time. The source of this peak has not been found and it is not reproducible. We find $\beta=8.73\pm 0.68\times 10^{15}$ MHz/eV3, $x_{ne}=8.50\pm 0.10\times 10^{-7}$, and $\Delta=185.0\pm 1.5\mu\text{eV}$. We note that the fractional non-equilibrium density $x_{ne}$ is quite high compared to recent works [7, 2, 8] which show a fractional density on the order of $10^{-9}$. Our setup uses light-tight radiation shields on all stages of the fridge, with Berkeley Black infrared-absorbing coating [38] on the interior of the 100 mK and mixing chamber shields. In addition, the sample package is mounted inside of an Amumetal 4K shield with a tin-plated copper can nested inside, also with a Berkeley Black interior coating. We use custom-made Eccosorb filters as well as K&L 12 Ghz low-pass filters on all inputs and outputs. A full diagram is available in Supplementary Figure 1 of the supplement [37]. We suspect that our device geometry may contribute to the higher-than-expected density, as large areas of superconducting aluminum are galvanically coupled to the SQUID. The left panel of Figure 1 shows a peak in the 30 mK data near 9 GHz. This anomaly was present in the trap rate and mean occupation, while the release rate was marginally increased. We attribute this to a temporary increase in the bulk QP density, as repeated measurement under nearly identical conditions did not show this effect. The period of increased trapping lasted for approximately one hour with no change in fridge conditions and no obvious environment factors to blame. We note the duration of the effect is too long to be caused by adhesive strain [39] or a strong cosmic ray [40]. We now turn our attention to $\Gamma_{release}$. To promote a trapped QP from ABS to the continuum of states above the gap, sufficient energy (at least $\Delta_{A}$) must be absorbed. In a well shielded dilution refrigerator, we expect this energy to come from the absorption of phonons. The clearing rate due to electron-phonon interactions should be linear in the phonon density, $\Gamma_{phonon}(\Delta_{A},T)\propto\rho_{\epsilon\geq\Delta_{A}}(T),$ (4) where $\rho_{\epsilon\geq\Delta_{A}}(T)$ is the density of phonons with energy exceeding the trap depth. In the supplement [37], we integrate the Debye density of states and Bose-Einstien distribution over energies exceeding the trap depth to obtain the model for QP clearing due to phonons: $\Gamma_{phonon}(\Delta_{A},T)=\alpha T^{3}\left[-\left(\frac{\Delta_{A}}{k_{B}T}\right)^{2}\ln{\left(1-e^{\frac{-\Delta_{A}}{k_{B}T}}\right)}+\frac{2\Delta_{A}}{k_{B}T}\text{Li}_{2}\left(e^{\frac{-\Delta_{A}}{k_{B}T}}\right)+2\text{Li}_{3}\left(e^{\frac{-\Delta_{A}}{k_{B}T}}\right)\right].$ (5) In the above, $\text{Li}_{n}(x)$ is the polylogarithm function of order $n$ and $\alpha=C_{ABS\rightarrow bulk}k_{B}^{3}/2\pi^{2}\hbar^{3}\nu^{3}$ is an overall scaling factor; $\nu$ is the speed of sound in our sample and $C_{ABS\rightarrow bulk}$ relates the ABS clearing rate to the phonon density. The formal foundation for $C_{ABS\rightarrow bulk}$ is a matter worthy of study as it represents the coupling between ABS and an incoherent bath. In our measurements, we observe that the release rate saturates at $T\leq$ 60 mK to a value which depends on the power of our microwave readout tone, suggesting that low-temperature clearing is dominated by driven electron- photon interactions. This is surprising because a single readout photon ($\approx$ 4.27 GHz) has insufficient energy to clear the ABS trap ($\Delta_{A}(\phi)>5$ GHz $\forall$ measured $\phi$). Accounting for this readout-dominated electron-photon clearing, we can model the total release rate as $\Gamma_{release}(\Delta_{A},T)=\Gamma_{RO}(\Delta_{A})+\Gamma_{phonon}(\Delta_{A},T),$ (6) where the electron-photon clearing rate $\Gamma_{RO}$ is the subject of future work. For now, we take advantage of the low temperature saturation $\Gamma_{release}^{0}\approx\Gamma_{RO}$ to eliminate this photon contribution and maintain focus on the electron-phonon clearing rate. Our model is $\Gamma_{release}(\Delta_{A},T)-\Gamma_{release}^{0}(\Delta_{A})\approx\Gamma_{phonon}(\Delta_{A},T)$ (7) which is equivalent to the right hand side of Eq. 5. Figure 2: (top) The measured release rate vs trap depth and temperature. The top left panel shows structure in the trap depth dependence which is attributed to the driven electron-photon interactions which dominate at low temperature. In the top right panel, the low temperature saturation is visible. The grey dashed line indicates the cutoff temperature (90 mK) for the fit. (bottom) The measured release rate minus the low temperature saturation is shown as circles, while the phonon clearing model (Eq. 7) is shown as solid curves. We keep $\Delta=185\mu$eV and fit Eq. 7 with $\alpha$ as the only free parameter as shown in Figure 2. We obtain $\alpha=38.51\pm 0.36\text{ MHz/K}^{3}$. Clearly the high temperature release rate is dominated by a thermal distribution of phonons, but this result shows that non-thermal sources may dominate at typical qubit operating temperatures. We point out that the 240 mK and 260 mK data in the top left panel show some clipping of the release rate data to the 1 MHz sample rate – A limitation of our measurement rather than a physical effect. Our last feature of interest is the mean occupation $\bar{n}$, which is taken directly from the extracted time series of ABS occupations, not from HMM parameters. We start with a simple sum over weighted probabilities: $\bar{n}=\sum_{i}iP(i),$ (8) where $P(i)$ is the probability of having $i$ trapped QPs. In this analysis, we are only distinguishing between 1 trapped QP and 0 trapped QPs, as the incidence of 2 or more trapped QPs is quite rare. We can therefore assume a stationary distribution to obtain $P(0)\Gamma_{trap}=P(1)\Gamma_{release}.$ (9) Plugging (9) into (8), we obtain the model for the mean occupation: $\bar{n}(\Delta_{A},T)=P(0)\frac{\Gamma_{trap}(\Delta_{A},T)}{\Gamma_{release}(\Delta_{A},T)}.$ (10) Unfortunately, we are unable to eliminate the driven electron-photon contribution as we did in Eq. (7) so we simply leave $\Gamma_{RO}(\Delta_{A})$ as a free parameter and fit each line cut along temperature separately. We normalize by dividing out the low-temperature saturation ($T\leq 60$ mK) to obtain the model $\|\bar{n}_{\Delta_{A}}(T)\|=\frac{1+\frac{1}{x_{ne}}\sqrt{\frac{2\pi k_{B}T}{\Delta}}e^{\frac{-\Delta}{k_{B}T}}}{1+\alpha_{M}T^{3}\left[-\left(\frac{\Delta_{A}}{k_{B}T}\right)^{2}\ln{\left(1-e^{\frac{-\Delta_{A}}{k_{B}T}}\right)}+\frac{2\Delta_{A}}{k_{B}T}\text{Li}_{2}\left(e^{\frac{-\Delta_{A}}{k_{B}T}}\right)+2\text{Li}_{3}\left(e^{\frac{-\Delta_{A}}{k_{B}T}}\right)\right]}.$ (11) We fit this independently for each trap depth, while holding $x_{ne}=8.5\times 10^{-7}$ and $\Delta=185\mu\text{eV}$ fixed. The only fit parameter is $\alpha_{M}\equiv\alpha/\Gamma_{RO}(\Delta_{A})$. The results are shown in Figure 3. Figure 3: (top) The measured mean occupation (circles) and the corresponding fit (solid) are shown against temperature. Note that a different fit is performed at each value of $\Delta_{A}$. (bottom) The fit parameter $\alpha_{M}$ vs trap depth. Stars indicate the value of $\alpha_{M}$ for the three curves of the same color displayed in the top panel. We note the characteristic dip in mean occupation for $T\in[80,150]$ mK arises from an increased phonon population leading to faster clearing of ABS, while the rise for $T\geq 150$ mK is due to large population of thermal QPs. We may check for self-consistency in our description by examining the relationship between $\alpha_{M}(\Delta_{A})$ and the driven electron-photon clearing rate $\Gamma_{RO}(\Delta_{A})$. We directly measure $\Gamma_{RO}(\Delta_{A})$ as the low-temperature saturation of the release rate and compare this with the estimate obtained from $\alpha/\alpha_{M}$, as shown in Figure 4. Note that the former quantity comes entirely from the HMM parameters, while the latter quantity comes from direct analysis of the ABS occupation time series. These quantities agree very closely, indicating that our analysis is robust. Figure 4: Two sources of estimate for the rate of readout photons clearing QPs from the ABS traps. The measured low temperature release rate (blue) and the fit parameter from the mean occupation, shown as $\alpha/\alpha_{M}$ (orange), where $\alpha=38.51$ is found from fitting the phonon contribution to the release rate as shown in Figure 2. We point out that these agree in shape and magnitude despite coming from different sources. The driven electron-photon clearing rate has significant structure in its dependence on $\Delta_{A}$ which is repeatable. There is additional structure when one looks at the dependence on the microwave power, which is the focus of our future work with this system. By utilizing the many ABS of aluminum nanobridge Josephson junctions, we are able to measure and explain the behavior of quasiparticle trapping in qubit- like circuits over a range of trap depth and temperature. We show that QPs relax into traps primarily by spontaneous emission of a phonon. The close agreement between our data and our model suggests that most QPs entering the trap are originally at or near the superconducting gap $\Delta$. This indicates that any “hot” non-equilibrium quasiparticles are first relaxing to the gap in an independent process before trapping or that the majority of non- equilibrium quasiparticles exist at the gap edge, in agreement with past results [41]. We do not see any evidence of “photon-assisted trapping” (in analogy to the photon-assisted tunneling observed in tunnel junctions) where an infrared photon breaks a Cooper pair, promoting a QP directly into an ABS. This process may occur at lower rates, and is the subject of future work. We also show that clearing of QPs from ABS traps at temperatures above 90 mK occurs primarily through absorption of phonons which are distributed according to the Debye model. Other sources, such as microwave photons, are the dominant source of ABS-clearing energy at qubit operating temperatures. Our results further elucidate the behavior of equilibrium and non-equilibrium quasiparticles in superconducting circuits. ###### Acknowledgements. We thank Leonid Glazman for useful discussions and the MIT Lincoln Laboratory for providing the TWPA used in these measurements. This work was funded by the AFOSR under FA9550-19-1-0060 and the NSF under DMR-1900135. JF, AZ, DH fabricated the device. JF, AZ, SS performed measurements. JF and AZ performed the analysis. JF wrote the manuscript with input from all authors. ELF was the PI supervising all aspects of the work. ## References * Serniak _et al._ [2018] K. Serniak, M. Hays, G. de Lange, S. Diamond, S. Shankar, L. Burkhart, L. Frunzio, M. Houzet, and M. Devoret, Hot Nonequilibrium Quasiparticles in Transmon Qubits, Phys. Rev. Lett. 121, 157701 (2018), publisher: American Physical Society. * Vepsäläinen _et al._ [2020] A. P. Vepsäläinen, A. H. Karamlou, J. L. Orrell, A. S. Dogra, B. Loer, F. Vasconcelos, D. K. Kim, A. J. Melville, B. M. Niedzielski, J. L. Yoder, S. Gustavsson, J. A. Formaggio, B. A. VanDevender, and W. D. Oliver, Impact of ionizing radiation on superconducting qubit coherence, Nature 584, 551 (2020), number: 7822 Publisher: Nature Publishing Group. * Wilen _et al._ [2021] C. D. Wilen, S. Abdullah, N. A. Kurinsky, C. Stanford, L. Cardani, G. D’Imperio, C. Tomei, L. Faoro, L. B. Ioffe, C. H. Liu, A. Opremcak, B. G. Christensen, J. L. DuBois, and R. McDermott, Correlated charge noise and relaxation errors in superconducting qubits, Nature 594, 369 (2021), number: 7863 Publisher: Nature Publishing Group. * Martinis [2021] J. M. Martinis, Saving superconducting quantum processors from decay and correlated errors generated by gamma and cosmic rays, npj Quantum Inf 7, 1 (2021), number: 1 Publisher: Nature Publishing Group. * Barends _et al._ [2011] R. Barends, J. Wenner, M. Lenander, Y. Chen, R. C. Bialczak, J. Kelly, E. Lucero, P. O’Malley, M. Mariantoni, D. Sank, H. Wang, T. C. White, Y. Yin, J. Zhao, A. N. Cleland, J. M. Martinis, and J. J. A. Baselmans, Minimizing quasiparticle generation from stray infrared light in superconducting quantum circuits, Appl. Phys. Lett. 99, 113507 (2011), publisher: American Institute of Physics. * de Visser _et al._ [2011] P. J. de Visser, J. J. A. Baselmans, P. Diener, S. J. C. Yates, A. Endo, and T. M. Klapwijk, Number Fluctuations of Sparse Quasiparticles in a Superconductor, Phys. Rev. Lett. 106, 167004 (2011), publisher: American Physical Society. * Serniak _et al._ [2019] K. Serniak, S. Diamond, M. Hays, V. Fatemi, S. Shankar, L. Frunzio, R. Schoelkopf, and M. Devoret, Direct Dispersive Monitoring of Charge Parity in Offset-Charge-Sensitive Transmons, Phys. Rev. Applied 12, 014052 (2019), publisher: American Physical Society. * Mannila _et al._ [2022] E. T. Mannila, P. Samuelsson, S. Simbierowicz, J. T. Peltonen, V. Vesterinen, L. Grönberg, J. Hassel, V. F. Maisi, and J. P. Pekola, A superconductor free of quasiparticles for seconds, Nat. Phys. 18, 145 (2022), number: 2 Publisher: Nature Publishing Group. * Córcoles _et al._ [2011] A. D. Córcoles, J. M. Chow, J. M. Gambetta, C. Rigetti, J. R. Rozen, G. A. Keefe, M. Beth Rothwell, M. B. Ketchen, and M. Steffen, Protecting superconducting qubits from radiation, Appl. Phys. Lett. 99, 181906 (2011), publisher: American Institute of Physics. * Pop _et al._ [2014] I. M. Pop, K. Geerlings, G. Catelani, R. J. Schoelkopf, L. I. Glazman, and M. H. Devoret, Coherent suppression of electromagnetic dissipation due to superconducting quasiparticles, Nature 508, 369 (2014), number: 7496 Publisher: Nature Publishing Group. * Court _et al._ [2008] N. A. Court, A. J. Ferguson, R. Lutchyn, and R. G. Clark, Quantitative study of quasiparticle traps using the single-Cooper-pair transistor, Phys. Rev. B 77, 100501 (2008), publisher: American Physical Society. * Riwar _et al._ [2016] R.-P. Riwar, A. Hosseinkhani, L. D. Burkhart, Y. Y. Gao, R. J. Schoelkopf, L. I. Glazman, and G. Catelani, Normal-metal quasiparticle traps for superconducting qubits, Phys. Rev. B 94, 104516 (2016), publisher: American Physical Society. * Henriques _et al._ [2019] F. Henriques, F. Valenti, T. Charpentier, M. Lagoin, C. Gouriou, M. Martínez, L. Cardani, M. Vignati, L. Grünhaupt, D. Gusenkova, J. Ferrero, S. T. Skacel, W. Wernsdorfer, A. V. Ustinov, G. Catelani, O. Sander, and I. M. Pop, Phonon traps reduce the quasiparticle density in superconducting circuits, Appl. Phys. Lett. 115, 212601 (2019), publisher: American Institute of Physics. * Rafferty _et al._ [2021] O. Rafferty, S. Patel, C. H. Liu, S. Abdullah, C. D. Wilen, D. C. Harrison, and R. McDermott, Spurious Antenna Modes of the Transmon Qubit (2021), arXiv:2103.06803 [cond-mat, physics:quant-ph]. * Kurter _et al._ [2022] C. Kurter, C. E. Murray, R. T. Gordon, B. B. Wymore, M. Sandberg, R. M. Shelby, A. Eddins, V. P. Adiga, A. D. K. Finck, E. Rivera, A. A. Stabile, B. Trimm, B. Wacaser, K. Balakrishnan, A. Pyzyna, J. Sleight, M. Steffen, and K. Rodbell, Quasiparticle tunneling as a probe of Josephson junction barrier and capacitor material in superconducting qubits, npj Quantum Inf 8, 1 (2022), number: 1 Publisher: Nature Publishing Group. * Pan _et al._ [2022] X. Pan, H. Yuan, Y. Zhou, L. Zhang, J. Li, S. Liu, Z. H. Jiang, G. Catelani, L. Hu, and F. Yan, Engineering superconducting qubits to reduce quasiparticles and charge noise (2022), arXiv:2202.01435 [cond-mat, physics:quant-ph]. * Bargerbos _et al._ [2022] A. Bargerbos, L. J. Splitthoff, M. Pita-Vidal, J. J. Wesdorp, Y. Liu, P. Krogstrup, L. P. Kouwenhoven, C. K. Andersen, and L. Grünhaupt, Mitigation of quasiparticle loss in superconducting qubits by phonon scattering (2022), arXiv:2207.12754 [cond-mat, physics:quant-ph]. * Aumentado _et al._ [2004] J. Aumentado, M. W. Keller, J. M. Martinis, and M. H. Devoret, Nonequilibrium Quasiparticles and $2e$ Periodicity in Single-Cooper-Pair Transistors, Phys. Rev. Lett. 92, 066802 (2004), publisher: American Physical Society. * Naaman and Aumentado [2006] O. Naaman and J. Aumentado, Time-domain measurements of quasiparticle tunneling rates in a single-Cooper-pair transistor, Phys. Rev. B 73, 172504 (2006), publisher: American Physical Society. * Lutchyn and Glazman [2007] R. M. Lutchyn and L. I. Glazman, Kinetics of quasiparticle trapping in a Cooper-pair box, Phys. Rev. B 75, 184520 (2007), publisher: American Physical Society. * Shaw _et al._ [2008] M. D. Shaw, R. M. Lutchyn, P. Delsing, and P. M. Echternach, Kinetics of nonequilibrium quasiparticle tunneling in superconducting charge qubits, Phys. Rev. B 78, 024503 (2008), publisher: American Physical Society. * Sun _et al._ [2012] L. Sun, L. DiCarlo, M. D. Reed, G. Catelani, L. S. Bishop, D. I. Schuster, B. R. Johnson, G. A. Yang, L. Frunzio, L. Glazman, M. H. Devoret, and R. J. Schoelkopf, Measurements of Quasiparticle Tunneling Dynamics in a Band-Gap-Engineered Transmon Qubit, Phys. Rev. Lett. 108, 230509 (2012), publisher: American Physical Society. * Zgirski _et al._ [2011] M. Zgirski, L. Bretheau, Q. Le Masne, H. Pothier, D. Esteve, and C. Urbina, Evidence for Long-Lived Quasiparticles Trapped in Superconducting Point Contacts, Phys. Rev. Lett. 106, 257003 (2011), publisher: American Physical Society. * Levenson-Falk _et al._ [2014] E. Levenson-Falk, F. Kos, R. Vijay, L. Glazman, and I. Siddiqi, Single-Quasiparticle Trapping in Aluminum Nanobridge Josephson Junctions, Phys. Rev. Lett. 112, 047002 (2014), publisher: American Physical Society. * Olivares _et al._ [2014] D. G. Olivares, A. L. Yeyati, L. Bretheau, Ç. Ö. Girit, H. Pothier, and C. Urbina, Dynamics of quasiparticle trapping in Andreev levels, Phys. Rev. B 89, 104504 (2014), publisher: American Physical Society. * Hays _et al._ [2018] M. Hays, G. de Lange, K. Serniak, D. van Woerkom, D. Bouman, P. Krogstrup, J. Nygård, A. Geresdi, and M. Devoret, Direct Microwave Measurement of Andreev-Bound-State Dynamics in a Semiconductor-Nanowire Josephson Junction, Phys. Rev. Lett. 121, 047001 (2018), publisher: American Physical Society. * Farmer _et al._ [2021] J. T. Farmer, A. Zarassi, D. M. Hartsell, E. Vlachos, H. Zhang, and E. M. Levenson-Falk, Continuous real-time detection of quasiparticle trapping in aluminum nanobridge Josephson junctions, Appl. Phys. Lett. 119, 122601 (2021), publisher: American Institute of Physics. * Janvier _et al._ [2015] C. Janvier, L. Tosi, L. Bretheau, Ç. Ö. Girit, M. Stern, P. Bertet, P. Joyez, D. Vion, D. Esteve, M. F. Goffman, H. Pothier, and C. Urbina, Coherent manipulation of Andreev states in superconducting atomic contacts, Science 349, 1199 (2015), publisher: American Association for the Advancement of Science. * Hays _et al._ [2021] M. Hays, V. Fatemi, D. Bouman, J. Cerrillo, S. Diamond, K. Serniak, T. Connolly, P. Krogstrup, J. Nygård, A. Levy Yeyati, A. Geresdi, and M. H. Devoret, Coherent manipulation of an Andreev spin qubit, Science 373, 430 (2021), publisher: American Association for the Advancement of Science. * Vijay _et al._ [2009] R. Vijay, J. D. Sau, M. L. Cohen, and I. Siddiqi, Optimizing Anharmonicity in Nanoscale Weak Link Josephson Junction Oscillators, Phys. Rev. Lett. 103, 087003 (2009), publisher: American Physical Society. * Vijay _et al._ [2010] R. Vijay, E. M. Levenson-Falk, D. H. Slichter, and I. Siddiqi, Approaching ideal weak link behavior with three dimensional aluminum nanobridges, Appl. Phys. Lett. 96, 223112 (2010), publisher: American Institute of Physics. * Levenson-Falk _et al._ [2011] E. M. Levenson-Falk, R. Vijay, and I. Siddiqi, Nonlinear microwave response of aluminum weak-link Josephson oscillators, Appl. Phys. Lett. 98, 123115 (2011), publisher: American Institute of Physics. * Kulik and Omel’Yanchuk [1975] I. O. Kulik and A. N. Omel’Yanchuk, Contribution to the microscopic theory of the Josephson effect in superconducting bridges, ZhETF Pisma Redaktsiiu 21, 216 (1975), aDS Bibcode: 1975ZhPmR..21..216K. * Dorokhov [1982] O. N. Dorokhov, Transmission coefficient and the localization length of an electron in N bound disordered chains, Soviet Journal of Experimental and Theoretical Physics Letters 36, 318 (1982), aDS Bibcode: 1982JETPL..36..318D. * [35] hmmlearn — hmmlearn 0.2.8 documentation. * Rabiner [1989] L. Rabiner, A tutorial on hidden Markov models and selected applications in speech recognition, Proceedings of the IEEE 77, 257 (1989), conference Name: Proceedings of the IEEE. * [37] See supplemental material at [url] for the full experimental diagram, detailed derivation of models, and intermediate fitting procedures. * Persky [1999] M. J. Persky, Review of black surfaces for space-borne infrared systems, Review of Scientific Instruments 70, 2193 (1999), publisher: American Institute of Physics. * Anthony-Petersen _et al._ [2022] R. Anthony-Petersen, A. Biekert, R. Bunker, C. L. Chang, Y.-Y. Chang, L. Chaplinsky, E. Fascione, C. W. Fink, M. Garcia-Sciveres, R. Germond, W. Guo, S. A. Hertel, Z. Hong, N. Kurinsky, X. Li, J. Lin, M. Lisovenko, R. Mahapatra, A. Mayer, D. McKinsey, S. Mehrotra, N. Mirabolfathi, B. Neblosky, W. A. Page, P. K. Patel, B. Penning, H. D. Pinckney, M. Platt, M. Pyle, M. Reed, R. K. Romani, H. S. Queiroz, B. Sadoulet, B. Serfass, R. Smith, P. F. Sorensen, B. Suerfu, A. Suzuki, R. Underwood, V. Velan, G. Wang, Y. Wang, S. L. Watkins, M. R. Williams, V. Yefremenko, and J. Zhang, A Stress Induced Source of Phonon Bursts and Quasiparticle Poisoning (2022), arXiv:2208.02790 [cond-mat, physics:physics]. * Cardani _et al._ [2021] L. Cardani, F. Valenti, N. Casali, G. Catelani, T. Charpentier, M. Clemenza, I. Colantoni, A. Cruciani, G. D’Imperio, L. Gironi, L. Grünhaupt, D. Gusenkova, F. Henriques, M. Lagoin, M. Martinez, G. Pettinari, C. Rusconi, O. Sander, C. Tomei, A. V. Ustinov, M. Weber, W. Wernsdorfer, M. Vignati, S. Pirro, and I. M. Pop, Reducing the impact of radioactivity on quantum circuits in a deep-underground facility, Nat Commun 12, 2733 (2021), number: 1 Publisher: Nature Publishing Group. * Diamond _et al._ [2022] S. Diamond, V. Fatemi, M. Hays, H. Nho, P. D. Kurilovich, T. Connolly, V. R. Joshi, K. Serniak, L. Frunzio, L. I. Glazman, and M. H. Devoret, Distinguishing Parity-Switching Mechanisms in a Superconducting Qubit, PRX Quantum 3, 040304 (2022), publisher: American Physical Society.
# Arithmetic autocorrelation distribution of binary $m$-sequences Xiaoyan Jing, Aixian Zhang and Keqin Feng Xiaoyan Jing is with Research Center for Number Theory and Its Applications, Northwest University, Xi’an 710127, China (e-mail: jxymg@126.com). Aixian Zhang is with Department of Mathematical Sciences, Xi’an University of Technology, Xi’an 710054, China (e-mail: zhangaixian1008@126.com). Keqin Feng is with Department of Mathematical Sciences, Tsinghua University, Beijing 100084, China (e-mail: fengkq@tsinghua.edu.cn).This work is supported by the National Natural Science Foundation of China under Grant No. 12031011 and the Natural Science Basic Research Plan in Shaanxi Province of China under Grant No. 2022JM-017. ###### Abstract Binary $m$-sequences are ones with the largest period $n=2^{m}-1$ among the binary sequences produced by linear shift registers with length $m$. They have a wide range of applications in communication since they have several desirable pseudorandomness such as balance, uniform pattern distribution and ideal (classical) autocorrelation. In his reseach on arithmetic codes, Mandelbaum [9] introduces a 2-adic version of classical autocorrelation of binary sequences, called arithmetic autocorrelation. Later, Goresky and Klapper [3, 4, 5, 6] generalize this notion to nonbinary case and develop several properties of arithmetic autocorrelation related to linear shift registers with carry. Recently, Z. Chen et al. [1] show an upper bound on arithmetic autocorrelation of binary $m$-sequences and raise a conjecture on absolute value distribution on arithmetic autocorrelation of binary $m$-sequences. In this paper we present a general formula for computing arithmetic autocorrelations, from which we totally determine the arithmetic autocorrelation distribution of arbitrary binary $m$-sequences. Particularly, the conjecture raised in [1] is verified. ###### Index Terms: arithmetic autocorrelation, binary $m$-sequences, 2-adic expression ## I Introduction Let $q=2^{m}\ (m\geq 2)$, $\mathbb{F}_{q}$ be the finite field with $q$ elements, $T:\mathbb{F}_{q}\rightarrow\mathbb{F}_{2}$ be the trace mapping defined by $T(a)=a+a^{2}+a^{2^{2}}+\cdots+a^{2^{m-1}}\quad(a\in\mathbb{F}_{q}).$ Let $\pi$ be a primitive element of $\mathbb{F}_{q}$, $\mathbb{F}^{\ast}_{q}=\mathbb{F}_{q}\setminus\\{0\\}=\langle\pi\rangle.$ The binary $m$-sequence with period $n=2^{m}-1$ is defined by $s=(s_{\lambda}=T(\pi^{\lambda}))^{n-1}_{\lambda=0}$ and shift sequences $s^{(\tau)}=(s^{(\tau)}_{\lambda})^{n-1}_{\lambda=0}\ (1\leq\tau\leq n-1)$ are defined by $s_{\lambda}^{(\tau)}=s_{\lambda+\tau}=T(\pi^{\lambda+\tau})\ (s_{\lambda}^{(0)}=s_{\lambda})$. Binary $m$-sequences have a wide range of applications in the field of communications such as error-correcting coding, spread spectrum communications, code divison multiple access systems and cryptography since such sequences have the longest period $n=2^{m}-1$ among all binary sequences generated by linear shift registers with length $m$ and have good pseudorandomness. We list several basic properties of binary $m$-sequences which we need in this paper. ###### Lemma 1. Let $s=(s_{\lambda})^{n-1}_{\lambda=0}$ be a binary $m$-sequence with period $n=2^{m}-1$. Then (1). (linearity) All shift sequences $s^{(\tau)}\ (0\leq\tau\leq n-1)$ plus zero sequence form a $\mathbb{F}_{2}$-vector space of dimension $m$. (2). (uniform pattern distribution) For $1\leq l\leq m$, $a=(a_{1},\ldots,a_{l})\in\mathbb{F}_{2}^{l}$, let $N(a)=\sharp\\{0\leq i\leq n-1\mid(s_{i},s_{i+1},\ldots,s_{i+l-1})=a\\}\quad(\text{where}\ s_{n+\lambda}=s_{\lambda}).$ Then $\displaystyle N(a)=\left\\{\begin{array}[]{ll}2^{m-l}-1,&\mbox{ if}\ a_{1}=a_{2}=\cdots=a_{l}=0\\\ 2^{m-l},&\mbox{otherwise}\end{array}\right.$ (3). (ideal classical autocorrelation) $\sum\limits_{\lambda=0}^{n-1}(-1)^{s_{\lambda}+s_{\lambda+\tau}}=-1$ for all $\tau$, $1\leq\tau\leq n-1.$ In 1967, Mandelbaum [9] introduced a new version of autocorrelation on binary (periodical) sequences, called arithmetic autocorrelation, in his research on coding theory (arithmetic codes). Later, such notion is generalized to non- binary case and investigated by Goresky and Klapper [3, 4, 5, 6] related to their research on sequences generated by linear shift registers with carry. Now we state the definition of the arithmetic autocorrelation of (periodical) binary sequences. For any integer $a\geq 1$, we have the 2-adic expression of $a$ as $a=a_{0}+a_{1}2+a_{2}2^{2}+\cdots+a_{r}2^{r}\quad(a_{i}\in\\{0,1\\},\ r\geq 0).$ The 2-adic weight of $a$ is defined by $w_{2}(a)=\sum_{\lambda=0}^{r}a_{\lambda}\ (=\text{the number of}\ a_{\lambda}=1).$ It is obvious that $1\leq w_{2}(a)\leq r$ and $w_{2}(2^{l}a)=w_{2}(a)$ for all $l\geq 1$. Let $a=(a_{\lambda})_{\lambda=0}^{n-1}$ be a binary sequence with period $n\geq 2$, $a_{\lambda}\in\\{0,1\\}$, $a_{n+\lambda}=a_{\lambda}$. For any $\tau$, $0\leq\tau\leq n-1$, the shift sequence $a^{(\tau)}=(a^{(\tau)}_{\lambda})_{\lambda=0}^{n-1}$ of $a$ is defined by $a^{(\tau)}=a_{\lambda+\tau}$. Then $a^{(0)}=a$ and $a^{(\tau)}\ (1\leq\tau\leq n-1)$ are distinct sequences. Let $\sigma(a)=\sum_{\lambda=0}^{n-1}a_{\lambda}2^{\lambda}\in\mathbb{Z}.$ By assumption $n\geq 2$, we know that $(a_{0},\ \ldots,a_{n-1})\neq(0,\ldots,0),\ (1,\ldots,1)$ and $1\leq\sigma(a)\leq 2^{n}-2$. For $1\leq\tau\leq n-1$, let $\sigma(a,\tau)=\sigma(a)-\sigma(a^{(\tau)})=\sum_{\lambda=0}^{n-1}(a_{\lambda}-a_{\lambda+\tau})2^{\lambda}\in\mathbb{Z}.$ From $1\leq\sigma(a),\sigma(a^{\tau})\leq 2^{n}-2$ and $\sigma(a)\neq\sigma(a^{\tau})$ we get $1\leq\arrowvert\sigma(a)-\sigma(a^{\tau})\arrowvert\leq 2^{n}-3$. Let $\sigma(a)-\sigma(a^{\tau})=\varepsilon\cdot\sum_{\lambda=0}^{n-1}c_{\lambda}2^{\lambda}\quad(c_{\lambda}\in\\{0,1\\},\ \varepsilon\in{\\{\pm 1\\}}).$ Then $1\leq\sum\limits_{\lambda=0}^{n-1}c_{\lambda}2^{\lambda}=\arrowvert\sigma(a)-\sigma(a^{\tau})\arrowvert\leq 2^{n}-3$ and $1\leq w_{2}(\arrowvert\sigma(a)-\sigma(a^{\tau})\arrowvert)\leq n-1.$ ###### Definition 1. Let $a=(a_{\lambda})_{\lambda=0}^{n-1}$ be a binary sequence with period $n\geq 2$. For $1\leq\tau\leq n-1$, let $\sigma(a,\tau)=\sigma(a)-\sigma(a^{(\tau)})=\varepsilon\cdot\sum_{\lambda=0}^{n-1}c_{\lambda}2^{\lambda}\quad(c_{\lambda}\in\\{0,1\\},\ \varepsilon\in{\\{\pm 1\\}}).$ The arithmetic autocorrelation $A_{a}(\tau)$ is defined by (1). For $\varepsilon=1\ (\sigma(a)>\sigma(a^{(\tau)})),$ $\displaystyle A_{a}(\tau)$ $\displaystyle=\text{``number of $c_{\lambda}=1$"-``number of $c_{\lambda}=0$"}$ $\displaystyle=2w_{2}(\sum_{\lambda=0}^{n-1}c_{\lambda}2^{\lambda})-n=2w_{2}(\arrowvert\sigma(a,\tau)\arrowvert)-n.$ (2). For $\varepsilon=-1\ (\sigma(a)<\sigma(a^{(\tau)})),$ $\displaystyle A_{a}(\tau)$ $\displaystyle=\text{``number of $c_{\lambda}=0$"-``number of $c_{\lambda}=1$"}$ $\displaystyle=n-2w_{2}(\sum_{\lambda=0}^{n-1}c_{\lambda}2^{\lambda})=n-2w_{2}(\arrowvert\sigma(a,\tau)\arrowvert).$ From $1\leq w_{2}(\sum_{\lambda=0}^{n-1}c_{\lambda}2^{\lambda})\leq n-1$ we know that $\arrowvert A_{a}(\tau)\arrowvert\leq n-2$ for all $\tau$, $1\leq\tau\leq n-1$. The arithmetic autocorrelation distribution of $a$ is the following multiset $D(a)=\\{A_{a}(\tau)\arrowvert 1\leq\tau\leq n-1\\}.$ Comparing with the classical autocorrelation, there exist few results currently on arithmetic auto (and cross) correlation on binary sequences ([1, 2, 3, 4, 5, 6, 7, 8, 9, 10]). It is desirable that the absolute values $\arrowvert A_{a}(\tau)\arrowvert\ (1\leq\tau\leq n-1)$ are as small as possible. Hofer, Merai and Winterhof [8] proved that for a truly random binary sequence with period $n$, $\arrowvert A_{a}(\tau)\arrowvert=O(n^{\frac{3}{4}}(\log_{2}n)^{\frac{1}{2}}$. There exist a family of binary sequences, called $l$-sequences, having the ideal arithmetic autocorrelation which means that $A_{a}(\tau)=0\ (1\leq\tau\leq n-1)$ [4]. Hofer and Winterhof [7] present nontrivial bounds of arithmetic autocorrelations on Legendre sequences. Recently Chen et al. proved that $\arrowvert A_{s}(\tau)\arrowvert\leq 2^{m-1}-1$ for binary $m$-sequences with period $n=2^{m}-1$ ([1], Theorem 1). Based on examples in cases $m$=3 and 4, they raise the following conjecture. ###### Conjecture. ([1]. Conjecture 1) Let $s$ be a binary $m$-sequence with period $n=2^{m}-1,\ m\geq 2$. Then 1\. the arithmetic autocorrelation of $s$ satisfies $A_{s}(\tau)\in\\{\pm(2^{k}-1):\ 1\leq k\leq m-1\\}\quad\text{for}\ 1\leq\tau\leq n-1.$ 2\. for each $k,\ 1\leq k\leq m-1$, the total number of $\tau\ (1\leq\tau\leq n-1)$ with $\arrowvert A_{s}(\tau)\arrowvert=2^{k}-1$ is $2^{n-k}.$ In this paper we prove this conjecture. Firstly we show several preliminary results on $A_{a}(\tau)$ and present a formula of $A_{a}(\tau)$ for general (periodic) binary sequences in Section 2 (Theorem 1). Then we totally determine the arithmetic autocorrelation distribution of all binary $m$-sequences (Theorem 2) in Section 3. As a direct consequence, the above conjecture is verified. ## II A general formula of arithmetic autocorrelation In this section we present a formula on arithmetic autocorrelation $A_{a}(\tau)\ (1\leq\tau\leq n-1)$ for arbitrary binary sequence $a=(a_{\lambda})_{\lambda=0}^{n-1}$ with period $n$. Firstly we need some observations. Let $a=(a_{\lambda})_{\lambda=0}^{n-1}$ and $b=(b_{\lambda})_{\lambda=0}^{n-1}$ be two distinct binary sequences with period $n\geq 2$. We have $2\times n\ \\{0,1\\}$-matrix $M=\left(\begin{array}[]{cc}a_{0}\ a_{1}\ \ldots\ a_{n-1}\\\ b_{0}\ b_{1}\ \ldots\ b_{n-1}\end{array}\right)\ \ \quad(a_{\lambda},\ b_{\lambda}\in\\{0,1\\})$ and (cyclic) shift matrices of $M$ $M^{(t)}=\left(\begin{array}[]{cc}a_{t}\ a_{t+1}\ \ldots\ a_{t+n-1}\\\ b_{t}\ b_{t+1}\ \ldots\ b_{t+n-1}\end{array}\right)\ \ \quad(0\leq t\leq n-1)$ where $a_{n+l}=a_{l}$, $b_{n+l}=b_{l}$ and $M^{(0)}=M$. Let $\sigma(a^{(t)})=\sum_{\lambda=0}^{n-1}a_{\lambda+t}2^{\lambda},\ \sigma(b^{(t)})=\sum_{\lambda=0}^{n-1}b_{\lambda+t}2^{\lambda}\in\mathbb{Z}.$ From $a\neq b$ we know that $\sigma(a^{(t)})\neq\sigma(b^{(t)}).$ Let $\displaystyle A(M^{(t)})=\left\\{\begin{array}[]{ll}n-2w_{2}(\arrowvert\sigma(a^{(t)})-\sigma(b^{(t)})\arrowvert),&\mbox{ if}\ \sigma(a^{(t)})>\sigma(b^{(t)})\\\ 2w_{2}(\arrowvert\sigma(a^{(t)})-\sigma(b^{(t)})\arrowvert)-n,&\mbox{ if}\ \sigma(a^{(t)})<\sigma(b^{(t)}).\end{array}\right.$ (3) ###### Lemma 2. With above assumptions and notations, we have $A(M)=A(M^{(t)})\quad\text{for all}\ 1\leq t\leq n-1.$ ###### Proof. We just need to show that $A(M)=A(M^{(1)})$ since the general case can be done from $M^{(t+1)}=(M^{(t)})^{(1)}$. We have $M=\left(\begin{array}[]{cc}a_{0}\ a_{1}\ \ldots\ a_{n-1}\\\ b_{0}\ b_{1}\ \ldots\ b_{n-1}\end{array}\right),\ \ \quad M^{(1)}=\left(\begin{array}[]{cc}a_{1}\ \ldots\ a_{n-1}\ a_{0}\\\ b_{1}\ \ldots\ b_{n-1}\ b_{0}\end{array}\right)$ and $\displaystyle\left.\begin{array}[]{ll}~{}~{}~{}~{}~{}~{}\sigma(a)-\sigma(b)&=\sum\limits_{\lambda=0}^{n-1}(a_{\lambda}-b_{\lambda})2^{\lambda}=(a_{0}-b_{0})+\sum\limits_{\lambda=1}^{n-1}(a_{\lambda}-b_{\lambda})2^{\lambda}\\\ &=(a_{0}-b_{0})+2\cdot\sum\limits_{\lambda=0}^{n-2}(a_{\lambda+1}-b_{\lambda+1})2^{\lambda}\\\ \sigma(a^{(1)})-\sigma(b^{(1)})&=\sum\limits_{\lambda=0}^{n-1}(a_{\lambda+1}-b_{\lambda+1})2^{\lambda}=\sum\limits_{\lambda=0}^{n-2}(a_{\lambda+1}-b_{\lambda+1})2^{\lambda}+(a_{0}-b_{0})2^{n-1}\end{array}\right\\}$ (7) $(\@slowromancap i@)$ Case $a_{0}=b_{0}.$ By (7) we have $\sigma(a)-\sigma(b)=2\cdot\sum\limits_{\lambda=0}^{n-2}(a_{\lambda+1}-b_{\lambda+1})2^{\lambda}=2(\sigma(a^{(1)})-\sigma(b^{(1)})).$ Therefore $\sigma(a)-\sigma(b)$ and $\sigma(a^{(1)})-\sigma(b^{(1)})$ have the same sign and $w_{2}(\arrowvert\sigma(a)-\sigma(b)\arrowvert)=w_{2}(2\arrowvert\sigma(a)^{(1)}-\sigma(b)^{(1)}\arrowvert)=w_{2}(\arrowvert\sigma(a)^{(1)}-\sigma(b)^{(1)}\arrowvert).$ From (3) we get $A(M)=A(M^{(1)})$. $(\@slowromancap ii@)$ Case $(a_{0},b_{0})=(1,0).$ From (7) we have $\displaystyle\begin{array}[]{ll}\sigma(a^{(1)})-\sigma(b^{(1)})=\sum\limits_{\lambda=0}^{n-2}a_{\lambda+1}2^{\lambda}-\sum\limits_{\lambda=0}^{n-2}b_{\lambda+1}2^{\lambda}+2^{n-1}>0,\\\ \sigma(a)-\sigma(b)=1+\sum\limits_{\lambda=1}^{n-1}(a_{\lambda}-b_{\lambda})2^{\lambda}.\end{array}$ $(\@slowromancap ii@.1)$ If $\sum\limits_{\lambda=1}^{n-1}(a_{\lambda}-b_{\lambda})2^{\lambda}\geq 0$, then $\sigma(a)-\sigma(b)>0$. From (7) we have $\displaystyle w_{2}(\sigma(a)-\sigma(b))=w_{2}(1+\sum\limits_{\lambda=1}^{n-1}(a_{\lambda}-b_{\lambda})2^{\lambda})=1+w_{2}(\sum\limits_{\lambda=1}^{n-1}(a_{\lambda}-b_{\lambda})2^{\lambda}),$ $\displaystyle w_{2}(\sigma(a^{(1)})-\sigma(b^{(1)}))=w_{2}(\sum\limits_{\lambda=0}^{n-2}(a_{\lambda+1}-b_{\lambda+1})2^{\lambda}+2^{n-1})$ $\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}=w_{2}(\sum\limits_{\lambda=0}^{n-2}(a_{\lambda+1}-b_{\lambda+1})2^{\lambda})+1=w_{2}(\sigma(a)-\sigma(b))$ From (3) we get $A(M)=A(M^{(1)})$. $(\@slowromancap ii@.2)$ If $\sum\limits_{\lambda=1}^{n-1}(a_{\lambda}-b_{\lambda})2^{\lambda}<0$, then $\sum\limits_{\lambda=1}^{n-1}(a_{\lambda}-b_{\lambda})2^{\lambda}=2\sum\limits_{\lambda=1}^{n-1}(a_{\lambda}-b_{\lambda})2^{\lambda}\leq-2$ and $\sigma(a)-\sigma(b)\leq 1-2<0$. From (7) we have $\displaystyle w_{2}(\arrowvert\sigma(a)-\sigma(b)\arrowvert)=w_{2}(-1+\sum\limits_{\lambda=1}^{n-1}(b_{\lambda}-a_{\lambda})2^{\lambda})$ $\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}=n-w_{2}(2^{n}-1-(-1+\sum\limits_{\lambda=1}^{n-1}(b_{\lambda}-a_{\lambda})2^{\lambda}))$ $\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}=n-w_{2}(2^{n}-(\sum\limits_{\lambda=1}^{n-1}(b_{\lambda}-a_{\lambda})2^{\lambda}))$ $\displaystyle w_{2}(\arrowvert\sigma(a^{(1)})-\sigma(b^{(1)})\arrowvert)=w_{2}(2^{n-1}-\sum\limits_{\lambda=0}^{n-2}(b_{\lambda+1}-a_{\lambda+1})2^{\lambda})$ $\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}=w_{2}(2^{n}-\sum\limits_{\lambda=1}^{n-1}(b_{\lambda}-a_{\lambda})2^{\lambda})=n-w_{2}(\arrowvert\sigma(a)-\sigma(b)\arrowvert).$ From (3) we get $\displaystyle A(M^{(1)})$ $\displaystyle=n-2w_{2}(\arrowvert\sigma(a^{(1)})-\sigma(b^{(1)})\arrowvert)=n-2(n-w_{2}(\arrowvert\sigma(a)-\sigma(b)\arrowvert))$ $\displaystyle=2w_{2}(\arrowvert\sigma(a)-\sigma(b)\arrowvert)-n=A(M).$ $(\@slowromancap iii@)$ Case $(a_{0},b_{0})=(0,1).$ Let $\widetilde{M}=\left(\begin{array}[]{cc}b_{0}\ \ldots\ b_{n-1}\\\ a_{0}\ \ldots\ a_{n-1}\end{array}\right),\ \ \quad\widetilde{M}^{(1)}=\left(\begin{array}[]{cc}b_{1}\ \ldots\ b_{n-1}\ b_{0}\\\ a_{1}\ \ldots\ a_{n-1}\ a_{0}\end{array}\right).$ From the formula (3) we know that $A(M)=-A(\widetilde{M})$, $A(M^{(1)})=-A(\widetilde{M}^{(1)})$. By Case $(\@slowromancap ii@)$ we get that $A(\widetilde{M})=A(\widetilde{M}^{(1)})$. Therefore $A(M)=A(M^{(1)})$. This completes the proof of Lemma 2. ∎ As a direct consequence of Lemma 2, we obtain the following result which means that for fixed $\tau$, the value $A_{a}(\tau)$ is a shift-invariant of binary sequence $a$. ###### Lemma 3. Let $a=(a_{\lambda})_{\lambda=0}^{n-1}$ be a binary sequence with period $n\geq 2$. Then for any $t$, $1\leq t\leq n-1$, $A_{a}(\tau)=A_{a^{(t)}}(\tau)$. ###### Proof. Take $b=a^{(\tau)}$ in Lemma 2. It is easy to see that $A(M)=A_{a}(\tau)$ and $A(M^{(t)})=A_{a^{(t)}}(\tau)$. Then from $A(M)=A(M^{(t)})$ (Lemma 2) we get $A_{a}(\tau)=A_{a^{(t)}}(\tau)$. ∎ We continue to compute the value of $A(M)$ where $M=\left(\begin{array}[]{cc}a_{0}\ \ldots\ a_{n-1}\\\ b_{0}\ \ldots\ b_{n-1}\end{array}\right)$ and $a=(a_{\lambda})_{\lambda=0}^{n-1}$, $b=(b_{\lambda})_{\lambda=0}^{n-1}$ are distinct binary sequences with period $n\geq 2$. Now we put a weak assumption. $(*)\ \text{There\ exists}\ t,\ 0\leq t\leq n-1\ \text{such\ that}\ (a_{t},b_{t})=(1,0).$ Namely, the matrix $M$ have at least one column ${a_{t}\choose b_{t}}={1\choose 0}$. Then the shift matrix $W=M^{(t+1)}$ is $W=\left(\begin{array}[]{cc}a_{t+1}\ \ldots\ a_{t-1}\ a_{t}\\\ b_{t+1}\ \ldots\ b_{t-1}\ b_{t}\end{array}\right)=\left(\begin{array}[]{cc}a_{t+1}\ \ldots\ a_{t-1}\ 1\\\ b_{t+1}\ \ldots\ b_{t-1}\ 0\end{array}\right).$ In this case, $\sigma=\sum\limits_{\lambda=0}^{n-1}a_{\lambda+t+1}2^{\lambda}=\sum\limits_{\lambda=0}^{n-2}a_{\lambda+t+1}2^{\lambda}+2^{n-1}$ is bigger than $\sigma^{\prime}=\sum\limits_{\lambda=0}^{n-1}b_{\lambda+t+1}2^{\lambda}=\sum\limits_{\lambda=0}^{n-2}b_{\lambda+t+1}2^{\lambda}$, we get $A(W)=n-2w_{2}(\sigma-\sigma^{\prime})$ and $A(M)=A(M^{(t+1)})=A(W)$ by Lemma 2. In order to give a closed formula for $A(W)$, we introduce some notations. ###### Definition 2. For $\alpha,\ \beta,\ \gamma_{1},\ \ldots,\ \gamma_{l}\in\\{0,1\\}$ and $l\geq 0$, we call the matrix $\left(\begin{array}[]{cc}\alpha\ \ \ \gamma_{1}\ \ldots\ \gamma_{l}\ \ \ \ \beta\\\ 1-\alpha\ \gamma_{1}\ \ldots\ \gamma_{l}\ 1-\beta\end{array}\right)$ as one with type $[\alpha,\ \beta;l]$. ###### Theorem 4. Let $a=(a_{\lambda})_{\lambda=0}^{n-1}$ and $b=(b_{\lambda})_{\lambda=0}^{n-1}$ be distinct binary sequences with period $n\geq 2$, and $M=\left(\begin{array}[]{cc}a_{0}\ \ldots\ a_{n-2}\ a_{n-1}\\\ b_{0}\ \ldots\ b_{n-2}\ b_{n-1}\end{array}\right).$ Suppose that there exists $\lambda$, $0\leq\lambda\leq n-1$ such that $(a_{\lambda},b_{\lambda})=(1,0)$. Then $A(M)=n-2g(M)$ where $g(M)=\sum\limits_{l=0}^{n-1}l(N(0,\ 0;l)+N(0,\ 1;l))+\sum\limits_{l=0}^{n-1}(N(1,\ 0;l)+N(1,\ 1;l))$ and for $\alpha,\ \beta\in\\{0,1\\}$, $N(\alpha,\ \beta;l)$ is the number of matrices $\left(\begin{array}[]{cc}a_{\lambda}\ a_{\lambda+1}\ \ldots\ a_{\lambda+l+1}\\\ b_{\lambda}\ b_{\lambda+1}\ \ldots\ b_{\lambda+l+1}\end{array}\right)$ $(0\leq\lambda\leq n-1)$ having type $[\alpha,\ \beta;l]$ $($namely, $\left(\begin{array}[]{cc}a_{\lambda}\ a_{\lambda+1}\ \ldots\ a_{\lambda+l+1}\\\ b_{\lambda}\ b_{\lambda+1}\ \ldots\ b_{\lambda+l+1}\end{array}\right)=\left(\begin{array}[]{cc}\alpha\ \ \ \gamma_{1}\ \ldots\ \gamma_{l}\ \ \ \ \beta\\\ 1-\alpha\ \gamma_{1}\ \ldots\ \gamma_{l}\ 1-\beta\end{array}\right)$ for some $\gamma_{1},\ \ldots,\ \gamma_{l}\in\\{0,1\\})$. ###### Proof. By assumption $(a_{\lambda},b_{\lambda})=(1,0)$ for some $\lambda$, we have a (cyclic) shift matrix $W$ of $M$ $W=\left(\begin{array}[]{ll}\alpha_{0}\ldots\ \alpha_{n-1}\\\ \varepsilon_{0}\ \ldots\ \varepsilon_{n-1}\end{array}\right)$ such that $\left(\begin{array}[]{cc}\alpha_{n-1}\\\ \varepsilon_{n-1}\end{array}\right)=\left(\begin{array}[]{cc}a_{\lambda}\\\ b_{\lambda}\end{array}\right)=\left(\begin{array}[]{cc}1\\\ 0\end{array}\right)$. Let $\sigma(\alpha)=\sum\limits_{\lambda=0}^{n-1}\alpha_{\lambda}2^{\lambda}$, $\sigma(\varepsilon)=\sum\limits_{\lambda=0}^{n-1}\varepsilon_{\lambda}2^{\lambda}$. From $\alpha_{n-1}=1,\ \varepsilon_{n-1}=0$ we know that $\sigma(\alpha)\geq 2^{n-1}>\sigma(\varepsilon).$ Since $W$ is shift matrix of $M$, we get $A(M)=A(W)=n-2w_{2}(\sigma(\alpha)-\sigma(\varepsilon)).$ We compute $\sigma(\alpha)-\sigma(\varepsilon)$ by 2-adic subtraction on the digits of $\sigma(\alpha)$ and $\sigma(\varepsilon)$ as shown in Figure 1. Figure 1 2-adic subtraction $\sigma(\alpha)-\sigma(\varepsilon)$ $\sigma(\alpha)$ | $\alpha_{0}$ | $\alpha_{1}$ | $\ldots$ | $\alpha_{n-1}$ | $(\alpha_{n-1}=1)$ ---|---|---|---|---|--- $\sigma(\varepsilon)$ | $\varepsilon_{0}$ | $\varepsilon_{1}$ | $\ldots$ | $\varepsilon_{n-1}$ | $(\varepsilon_{n-1}=0)$ $\sigma(\alpha)-\sigma(\varepsilon)$ | $\delta_{0}$ | $\delta_{1}$ | $\ldots$ | $\delta_{n-1}$ | $(\delta_{\lambda}\in\\{0,1\\})$ Then $\sigma(\alpha)-\sigma(\varepsilon)=\sum\limits_{\lambda=0}^{n-1}\delta_{\lambda}2^{\lambda}$ and $w_{2}(\sigma(\alpha)-\sigma(\varepsilon))$ is the number of $\delta_{\lambda}=1\ (0\leq\lambda\leq n-1).$ Now we write the matrix $W$ in the following block form $W=\left(W_{1}\ W_{2}\ \ldots\ W_{m}\right)\quad m\geq 1.$ Each bloack $W_{i}$ has form $W_{i}=\left(\begin{array}[]{ll}\gamma_{1}^{(i)}\ \ldots\ \gamma_{l_{i}}^{(i)}\ \ \ \ \beta_{i}\\\ \gamma_{1}^{(i)}\ \ldots\ \gamma_{l_{i}}^{(i)}\ 1-\beta_{i}\end{array}\right)\quad(l_{i}\geq 0)$ where $\gamma_{1}^{(i)},\ldots,\gamma_{l_{i}}^{(i)}\in\\{0,1\\}$ and the last block is $W_{m}=\left(\begin{array}[]{cc}\gamma_{1}^{(m)}\ \ldots\ \gamma_{l_{m}}^{(m)}\ 1\\\ \gamma_{1}^{(m)}\ \ldots\ \gamma_{l_{m}}^{(m)}\ 0\end{array}\right)\ (\beta_{m}=\alpha_{n-1}=1).$ Put the last column of $W_{i-1}$ to the left-hand side of $W_{i}$, we get $\widetilde{W_{i}}=\left(\begin{array}[]{cc}\beta_{i-1}\ \ \ \gamma_{1}^{(i)}\ \ldots\ \gamma_{l_{i}}^{(i)}\ \ \ \ \beta_{i}\\\ 1-\beta_{i-1}\ \gamma_{1}^{(i)}\ \ldots\ \gamma_{l_{i}}^{(i)}\ 1-\beta_{i}\end{array}\right)=\left(\begin{array}[]{cc}\beta_{i-1}\\\ 1-\beta_{i-1}\end{array}W_{i}\right)$ which has type $[\beta_{i-1},\ \beta_{i};l_{i}]$. For the first block $W_{1}$, we assume $\widetilde{W_{1}}=\left(\begin{array}[]{cc}\beta_{m}\\\ 1-\beta_{m}\end{array}W\right)=\left(\begin{array}[]{ll}1\ \gamma_{1}^{(1)}\ \ldots\ \gamma_{l_{1}}^{(1)}\ \ \ \ \beta_{1}\\\ 0\ \gamma_{1}^{(1)}\ \ldots\ \gamma_{l_{1}}^{(1)}\ 1-\beta_{1}\end{array}\right)$. Then we can see the 2-adic subtraction given in Figure 1 locally. For each $\widetilde{W_{i}}$ with type $[\beta_{i-1},\ \beta_{i};l_{i}]$, the digits $\delta_{\lambda}$ (in Figure 1) under the columns of $W_{i}$ can be determined as shown in Figure 2. Figure 2 2-adic subtraction at $W_{i}$ $(\beta_{i-1},\beta_{i})=(1,0)$ (1) | $\gamma_{1}^{(i)}$ | $\ldots$ | $\gamma_{l_{i}}^{(i)}$ | 0 | $\ldots$ ---|---|---|---|---|--- (0) | $\gamma_{1}^{(i)}$ | $\ldots$ | $\gamma_{l_{i}}^{(i)}$ | 1 | $\ldots$ $\ldots$ | 0 | $\ldots$ | 0 | 1 | $\ldots$ $(\beta_{i-1},\beta_{i})=(0,0)$ (0) | $\gamma_{1}^{(i)}$ | $\ldots$ | $\gamma_{l_{i}}^{(i)}$ | 0 | $\ldots$ ---|---|---|---|---|--- (1) | $\gamma_{1}^{(i)}$ | $\ldots$ | $\gamma_{l_{i}}^{(i)}$ | 1 | $\ldots$ $\ldots$ | 1 | $\ldots$ | 1 | 0 | $\ldots$ $(\beta_{i-1},\beta_{i})=(1,1)$ (1) | $\gamma_{1}^{(i)}$ | $\ldots$ | $\gamma_{l_{i}}^{(i)}$ | 1 | $\ldots$ ---|---|---|---|---|--- (0) | $\gamma_{1}^{(i)}$ | $\ldots$ | $\gamma_{l_{i}}^{(i)}$ | 0 | $\ldots$ $\ldots$ | 0 | $\ldots$ | 0 | 1 | $\ldots$ $(\beta_{i-1},\beta_{i})=(0,1)$ (0) | $\gamma_{1}^{(i)}$ | $\ldots$ | $\gamma_{l_{i}}^{(i)}$ | 1 | $\ldots$ ---|---|---|---|---|--- (1) | $\gamma_{1}^{(i)}$ | $\ldots$ | $\gamma_{l_{i}}^{(i)}$ | 0 | $\ldots$ $\ldots$ | 1 | $\ldots$ | 1 | 0 | $\ldots$ From Figure 2 we know that the number $\delta_{\lambda}=1$ given by $W_{i}$ is 1 if the type of $\widetilde{W_{i}}$ is $[\beta_{i-1},\ \beta_{i};l_{i}]=[1,\ 0;l_{i}]$ or $[1,\ 1;l_{i}]$; or $l_{i}$ if the type of $\widetilde{W_{i}}$ is $[\beta_{i-1},\ \beta_{i};l_{i}]=[0,\ 0;l_{i}]$ or $[0,\ 1;l_{i}]$. Put all local data together, we get $\displaystyle w_{2}(\sigma(\alpha)-\sigma(\varepsilon))=\sum_{\begin{subarray}{c}i=1\\\ \text{``type of $\widetilde{W_{i}}$"=$[1,\ \beta_{i};l_{i}]$}\\\ \beta_{i}\in\\{0,1\\}\end{subarray}}^{m}1+\sum_{\begin{subarray}{c}i=1\\\ \text{``type of $\widetilde{W_{i}}$"=$[0,\ \beta_{i};l_{i}]$}\\\ \beta_{i}\in\\{0,1\\}\end{subarray}}^{m}l_{i}$ $\displaystyle=$ $\displaystyle\sum_{l\geq 0}l(N_{W}(0,\ 0;l)+N_{W}(0,\ 1;l))+\sum_{l\geq 0}(N_{W}(1,\ 0;l)+N_{W}(1,\ 1;l)),$ where for $l\geq 0$, $\alpha,\ \beta\in\\{0,1\\}$, $N_{W}(\alpha,\ \beta;l)$ is the number of $\left(\begin{array}[]{cc}\alpha_{\lambda}\ \alpha_{\lambda+1}\ \ldots\ \alpha_{\lambda+l+1}\\\ \varepsilon_{\lambda}\ \varepsilon_{\lambda+1}\ \ldots\ \varepsilon_{\lambda+l+1}\end{array}\right)\ (0\leq\lambda\leq n-1)$ having type $[\alpha,\ \beta;l]$. Since $W$ is a shift of $M$, the number $N_{W}(\alpha,\ \beta;l)$ is the same as the number $N(\alpha,\ \beta;l)=N_{M}(\alpha,\ \beta;l)$. Therefore $w_{2}(\sigma(\alpha)-\sigma(\varepsilon))=\sum_{l\geq 0}l(N(0,\ 0;l)+N(0,\ 1;l))+\sum_{l\geq 0}(N(1,\ 0;l)+N(1,\ 1;l)).$ At last, if $l\geq n$ and there exists a matrix $\left(\begin{array}[]{cc}a_{\lambda}\ a_{\lambda+1}\ \ldots\ a_{\lambda+l}\ a_{\lambda+l+1}\\\ b_{\lambda}\ b_{\lambda+1}\ \ldots\ b_{\lambda+l}\ b_{\lambda+l+1}\end{array}\right)$ of type $[\alpha,\ \beta;l]$ for some $\lambda$, then $a_{\lambda+i}=b_{\lambda+i}\ (i=1,\ldots,n)$ which means $a=b$ since the period of both sequences $a$ and $b$ is $n$. This contradicts to the assumption $a\neq b$. Therefore $N(\alpha,\ \beta;l)=0$ for any $\alpha,\ \beta\in\\{0,1\\}$ and $l\geq n$. We get the final formula $\displaystyle\sum_{l\geq 0}l(N(0,\ 0;l)+N(0,\ 1;l))+\sum_{l\geq 0}(N(1,\ 0;l)+N(1,\ 1;l))$ $\displaystyle=$ $\displaystyle\sum_{l=0}^{n-1}l(N(0,\ 0;l)+N(0,\ 1;l))+\sum_{l=0}^{n-1}(N(1,\ 0;l)+N(1,\ 1;l))=g(M)$ and $A(M)=n-2g(M)$. This completes the proof of Theorem 1. ∎ By Theorem 4, the value of $A(M)$ is expressed by the numbers $N(1,\ 0;l)+N(1,\ 1;l)$ and $N(0,\ 0;l)+N(0,\ 1;l)$ ($0\leq l\leq n-1,\ N(\alpha,\ \beta;l)=N_{M}(\alpha,\ \beta;l)$) for $M=\left(\begin{array}[]{cc}a_{0}\ \ldots\ a_{n-1}\\\ b_{0}\ \ldots\ b_{n-1}\end{array}\right)$. To determine these numbers is not easy for arbitrary binary sequences $a=(a_{\lambda})_{\lambda=0}^{n-1}$ and $b=(b_{\lambda})_{\lambda=0}^{n-1}$ with period $n\geq 2$ in general case. Fortunately, if $a$ is a binary $m$-sequence and $b$ is a shift sequence of $a$, these numbers can be computed and then the arithmetic autocorrelation distribution of all binary $m$-sequences can be determined. We do this in next section. ## III Arithmetic autocorrelation distribution of binary $m$-sequences Let $q=2^{m}\ (m\geq 2)$, $\pi$ be a primitive element of the finite field $\mathbb{F}_{q}$, which means that $\mathbb{F}^{\ast}_{q}=\mathbb{F}_{q}\setminus\\{0\\}=\langle\pi\rangle$, $\pi$ is the generator of the cyclic multiplicative group $\mathbb{F}^{\ast}_{q}$. The binary $m$-sequence $s=(s_{\lambda})^{n-1}_{\lambda=0}$ with period $n=2^{m}-1$ is defined by $s_{\lambda}=T(\pi^{\lambda})\in\mathbb{F}_{2}\ (0\leq\lambda\leq n)$, where $T:\ \mathbb{F}_{q}\rightarrow\mathbb{F}_{2},\ T(\alpha)=\alpha+\alpha^{2}+\alpha^{2^{2}}+\cdots+\alpha^{2^{m-1}}$ is the trace mapping. In Theorem 4, we take $a=s$, $b=s^{(\tau)}=(s^{(\tau)}_{\lambda})_{\lambda=0}^{n-1}$, the shift sequence of $s$ where $s^{(\tau)}_{\lambda}=s_{\lambda+\tau}$. Then $M=\left(\begin{array}[]{ll}s_{0}\ \ \ s_{1}\ \ \ \ldots\ s_{n-1}\\\ s^{(\tau)}_{0}\ s^{(\tau)}_{1}\ \ldots\ s^{(\tau)}_{n-1}\end{array}\right)=\left(\begin{array}[]{ll}s_{0}\ \ \ s_{1}\ \ \ldots\ s_{n-1}\\\ s_{\tau}\ s_{\tau+1}\ \ldots\ s_{n-1+\tau}\end{array}\right)\ (s_{n+t}=s_{t})$ and the arithmetic autocorrelation $A_{s}(\tau)$ is $A(M)$ for $1\leq\tau\leq n-1$. From Theorem 4 we know that $A_{s}(\tau)=n-2g(\tau)$ where $\displaystyle g(\tau)=\sum\limits_{l=0}^{n-1}l(N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l))+\sum\limits_{l=0}^{n-1}(N^{(\tau)}(1,\ 0;l)+N^{(\tau)}(1,\ 1;l))$ (8) and for $\alpha,\ \beta\in\mathbb{F}_{2}=\\{0,1\\}$ $\displaystyle N^{(\tau)}(\alpha,\ \beta;l)$ $\displaystyle=$ $\displaystyle\sharp\left\\{0\leq t\leq n-1\Bigg{\arrowvert}\begin{array}[]{ll}\left(\begin{array}[]{ll}s_{t}\ \ \ s_{t+1}\ \ldots\ s_{t+l+1}\\\ s^{(\tau)}_{t}\ s^{(\tau)}_{t+1}\ \ldots\ s^{(\tau)}_{t+l+1}\end{array}\right)=\left(\begin{array}[]{cc}\alpha\ \ \ \gamma_{1}\ \ldots\ \gamma_{l}\ \ \ \ \beta\\\ 1-\alpha\ \gamma_{1}\ \ldots\ \gamma_{l}\ 1-\beta\end{array}\right),\ \gamma_{1},\ldots,\gamma_{l}\in\mathbb{F}_{2}\end{array}\right\\}$ $\displaystyle=$ $\displaystyle\sharp\left\\{0\leq t\leq n-1\Bigg{\arrowvert}\begin{array}[]{ll}T(\pi^{t})=\alpha,\ T(\pi^{t+\tau})=1+\alpha,\ T(\pi^{t+l+1})=\beta,\ T(\pi^{t+l+1+\tau})=1+\beta,\\\ T(\pi^{t+\lambda})=T(\pi^{t+\lambda+\tau})\ \text{for}\ 1\leq\lambda\leq l\end{array}\right\\}$ $\displaystyle=$ $\displaystyle\sharp\left\\{x\in\mathbb{F}^{\ast}_{q}\Bigg{\arrowvert}\begin{array}[]{ll}T(x)=\alpha,\ T(\pi^{\tau}x)=1+\alpha,\ T(\pi^{l+1}x)=\beta,\ T(\pi^{l+1+\tau}x)=1+\beta,\\\ T(\pi^{\lambda}x)=T(\pi^{\lambda+\tau}x)\ \text{for}\ 1\leq\lambda\leq l\end{array}\right\\}.$ In order to determine $A_{s}(\tau)$ by formula (8), we need to compute $\displaystyle N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l)$ $\displaystyle=$ $\displaystyle\sharp\left\\{x\in\mathbb{F}^{\ast}_{q}\Bigg{\arrowvert}\begin{array}[]{ll}T(x)=0,\ T(\pi^{\tau}x)=1,\ T(\pi^{l+1}x)+T(\pi^{l+1+\tau}x)=1,\\\ T(\pi^{\lambda}x)+T(\pi^{\lambda+\tau}x)=0\ \text{for}\ 1\leq\lambda\leq l\end{array}\right\\}$ (11) and $\displaystyle N^{(\tau)}(1,\ 0;l)+N^{(\tau)}(1,\ 1;l)$ $\displaystyle=$ $\displaystyle\sharp\left\\{x\in\mathbb{F}^{\ast}_{q}\Bigg{\arrowvert}\begin{array}[]{ll}T(x)=1,\ T(\pi^{\tau}x)=0,\ T(\pi^{l+1}x)+T(\pi^{l+1+\tau}x)=1,\\\ T(\pi^{\lambda}x)+T(\pi^{\lambda+\tau}x)=0\ \text{for}\ 1\leq\lambda\leq l\end{array}\right\\}.$ (14) ###### Lemma 5. Let $1\leq\tau\leq n-1.$ Then $(\@slowromancap i@).$ If $l\geq m$, then $N^{(\tau)}(\alpha,\ \beta;l)=0$ for any $\alpha,\ \beta\in\mathbb{F}_{2}.$ $(\@slowromancap ii@).\begin{array}[]{ll}\sum\limits_{l=0}^{m-1}(N^{(\tau)}(1,\ 0;l)+N^{(\tau)}(1,\ 1;l))=\sum\limits_{l=0}^{m-1}(N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l))=2^{m-2}.\end{array}$ $(\@slowromancap iii@).$ $g(\tau)=2^{m-2}+\sum\limits_{l=1}^{m-1}l(N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l)).$ ###### Proof. $(\@slowromancap i@)$ Suppose that $l\geq m$ and $N^{(\tau)}(\alpha,\ \beta;l)\geq 1$. Then there exists $t$ such that $s_{t+i}=s_{t+i}^{(\tau)}$ for $1\leq i\leq l$. From $1\leq\tau\leq n-1$ we know that $s\neq s^{(\tau)}$ and $s+s^{\tau}=(s_{\lambda}+s_{\lambda}^{(\tau)})_{\lambda=0}^{n-1}$ is a shift sequence of $s$ (Lemma 1. (1)). But there is no pattern $(s_{t+1}+s_{t+1}^{(\tau)},\ldots,s_{t+l}+s_{t+l}^{(\tau)})=(0,\ldots,0)$ in sequence $s+s^{\tau}$ for $l\geq m$ (Lemma 1. (2)). This contradiction shows that $N(\alpha,\ \beta;l)=0$ for $l\geq m$ and any $\alpha,\ \beta\in\mathbb{F}_{2}.$ $(\@slowromancap ii@)$ Since the $m$-sequence $s$ has no pattern $(0,\ldots,0)\in\mathbb{F}_{2}^{m}$, from $(\@slowromancap i@)$ and formula (III) we can see that $\displaystyle\sum_{l=0}^{m-1}N^{(\tau)}(1,\ 0;l)+N^{(\tau)}(1,\ 1;l)$ $\displaystyle=$ $\displaystyle\sharp\left\\{x\in\mathbb{F}^{\ast}_{q}\Bigg{\arrowvert}T(x)=1,\ T(\pi^{\tau}x)=0\right\\}$ $\displaystyle=$ $\displaystyle\frac{1}{4}\sum_{x\in\mathbb{F}_{q}^{\ast}}\big{(}1-(-1)^{T(x)}\big{)}\big{(}1+(-1)^{T(\pi^{\tau}x)}\big{)}$ $\displaystyle=$ $\displaystyle\frac{1}{4}\sum_{x\in\mathbb{F}_{q}}\big{(}1-(-1)^{T(x)}\big{)}\big{(}1+(-1)^{T(\pi^{\tau}x)}\big{)}\quad(\text{since}\ 1-(-1)^{T(0)}=1-1=0)$ $\displaystyle=$ $\displaystyle\frac{1}{4}\left(q+\sum_{x\in\mathbb{F}_{q}}\left[(-1)^{T(\pi^{\tau}x)}-(-1)^{T(x)}-(-1)^{T((1+\pi^{\tau})x)}\right]\right)$ (15) It is well-known that $\sum\limits_{x\in\mathbb{F}_{q}}(-1)^{T(\alpha x)}=\sum\limits_{x\in\mathbb{F}_{q}}(-1)^{T(x)}=0$ if $\alpha\in\mathbb{F}_{q}^{\ast}$. Then by $1+\pi^{\tau}\neq 0$ (for $1\leq\tau\leq n-1$) we know that the summation at right-hand side of (III) is zero. Therefore $\sum\limits_{l=0}^{m-1}(N^{(\tau)}(1,\ 0;l)+N^{(\tau)}(1,\ 1;l))=\frac{q}{4}=2^{m-2}$. We can show $\sum\limits_{l=0}^{m-1}(N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l))=2^{m-2}$ in similar way. $(\@slowromancap iii@)$ Directly from $(\@slowromancap i@)$, $(\@slowromancap ii@)$ and formula (8). ∎ Now we come to place of computing $N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l)$ for each $l$, $1\leq l\leq m-1$ and $1\leq\tau\leq n-1$. For $1\leq\tau\leq n-1$, $1+\pi^{\tau}\neq 0,\ 1$, because the order of $\pi$ is $n=q-1$. Since $\\{1,\pi,\ldots,\pi^{m-1}\\}$ is an $\mathbb{F}_{2}$-basis of $\mathbb{F}_{q}$, $(1+\pi^{\tau})^{-1}(\neq 0,\ 1)$ has the following unique expression $(1+\pi^{\tau})^{-1}=b_{0}+b_{1}\pi+\cdots+b_{s-1}\pi^{e-1}+\pi^{e}$ where $b_{0},\cdots,b_{s-1}\in\mathbb{F}_{2}$, $1\leq e\leq m-1$. Another thing we need is that the minimal polynomial of $\pi$ over $\mathbb{F}_{2}$ is an irreducible (primitive) polynomial $f(x)=1+a_{1}x+\cdots+a_{m-1}x^{m-1}+x^{m}\ (a_{1},\ldots,a_{m-1}\in\mathbb{F}_{2})$ and $f(\pi)=0$. Therefore $\pi^{m}=1+a_{1}\pi+\cdots+a_{m-1}\pi^{m-1}$. ###### Lemma 6. Let $q=2^{m}\ (m\geq 2)$, $\pi$ be a primitive element of $\mathbb{F}_{q}$, $s=(s_{\lambda})^{n-1}_{\lambda=0}$ is the binary $m$-sequence with period $n=q-1$ defined by $s_{\lambda}=T(\pi^{\lambda})$, where $T:\ \mathbb{F}_{q}\rightarrow\mathbb{F}_{2}$ is the trace mapping. Let $1\leq\tau\leq n-1$ and $(1+\pi^{e})^{-1}=b_{0}+b_{1}\pi+\cdots+b_{s-1}\pi^{e-1}+\pi^{e}\quad(b_{0},\ldots,b_{s-1}\in\mathbb{F}_{2},\ 1\leq e\leq m-1)$ Then $(\@slowromancap i@).$ For $1\leq l\leq m-2$, $\displaystyle N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l)=\left\\{\begin{array}[]{ll}2^{m-l-3},&\text{if}\ l\leq e-2\\\ 2^{m-l-3}(1-(-1)^{b_{0}}),&\text{if}\ l=e-1(\geq 1)\\\ 2^{m-l-3}(1+(-1)^{b_{0}}),&\text{if}\ l\geq e.\end{array}\right.$ $(\@slowromancap ii@).$ For $l=m-1$, $N^{(\tau)}(0,\ 0;m-1)+N^{(\tau)}(0,\ 1;m-1)=\frac{1}{2}(1+(-1)^{b_{0}}).$ ###### Proof. From formula (III) we know that $\displaystyle N^{(\tau)}(0,\ 0;$ $\displaystyle l)+N^{(\tau)}(0,\ 1;l)$ $\displaystyle=\frac{1}{2^{l+3}}\sum_{x\in\mathbb{F}_{q}^{\ast}}$ $\displaystyle\big{(}(1+(-1)^{T(x)})(1-(-1)^{T(\pi^{\tau}x)})(1-(-1)^{T(\pi^{l+1}x+\pi^{l+1+\tau}x)})\prod\limits_{\lambda=1}^{l}(1+(-1)^{T(\pi^{\lambda}x+\pi^{\lambda+\tau}x)})\big{)}$ $\displaystyle=\frac{1}{2^{l+3}}\sum_{x\in\mathbb{F}_{q}}$ $\displaystyle\big{(}(1+(-1)^{T(x)})(1-(-1)^{T(\pi^{\tau}x)})(1-(-1)^{T(\pi^{l+1}(1+\pi^{\tau})x)})\prod\limits_{\lambda=1}^{l}(1+(-1)^{T(\pi^{\lambda}(1+\pi^{\tau})x)})\big{)}$ $\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\quad(\text{for}\ x=0,\ 1-(-1)^{T(\pi^{\tau}x)}=0)$ $\displaystyle=\frac{1}{2^{l+3}}\sum_{x\in\mathbb{F}_{q}}$ $\displaystyle\sum_{c^{\prime},c_{0},\ldots,c_{l+1}=0}^{1}(-1)^{T[x(c^{\prime}+c_{0}\pi^{\tau}+(1+\pi^{\tau})(c_{1}\pi+c_{2}\pi^{2}+\cdots+c_{l+1}\pi^{l+1}))]}(-1)^{c_{0}+c_{l+1}}$ $\displaystyle=\frac{1}{2^{l+3}}\sum_{x\in\mathbb{F}_{q}}$ $\displaystyle\sum_{c,c_{0},\ldots,c_{l+1}=0}^{1}(-1)^{T[x(c+(1+\pi^{\tau})(c_{0}+c_{1}\pi+c_{2}\pi^{2}+\cdots+c_{l+1}\pi^{l+1}))]}(-1)^{c_{0}+c_{l+1}}\quad(\text{let}\ c=c^{\prime}+c_{0})$ If $\delta=c+(1+\pi^{\tau})(c_{0}+c_{1}\pi+c_{2}\pi^{2}+\cdots+c_{l+1}\pi^{l+1})\neq 0$, then $\sum_{x\in\mathbb{F}_{q}}(-1)^{T(\delta x)}=0$. Therefore $\displaystyle N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l)=\frac{q}{2^{l+3}}\sum_{\begin{subarray}{c}c,c_{0},\ldots,c_{l+1}=0\\\ c_{0}+c_{1}\pi+\cdots+c_{l+1}\pi^{l+1}=c(1+\pi^{\tau})^{-1}\end{subarray}}^{1}(-1)^{c_{0}+c_{l+1}}$ (16) where the summation in the right-hand side is under the condition $\displaystyle c_{0}+c_{1}\pi+\cdots+c_{l+1}\pi^{l+1}=c(1+\pi^{\tau})^{-1}=c(b_{0}+b_{1}\pi+\cdots+b_{e-1}\pi^{e-1}+\pi^{e}).$ (17) From now on we write $N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l)$ by $N^{(\tau)}(l)$ briefly. $(\@slowromancap i@)$ For $1\leq l\leq m-2$, $l+1\leq m-1$ and $e\leq m-1$. Since $\\{1,\pi,\ldots,\pi^{m-1}\\}$ is a basis of $\mathbb{F}_{q}$ over $\mathbb{F}_{2}$, the coefficients of $\pi^{\lambda}$ in both sides of formula (17) should be equal for each $\lambda$, $0\leq\lambda\leq m-1$. $(\@slowromancap i@.1)$ If $l\leq e-2$, then $e\geq l+2$. From the coefficients of $\pi^{e}$ we get $0=c$. Then $c_{\lambda}=cb_{\lambda}=0$ for all $\lambda$, $0\leq\lambda\leq l+1\ (\leq e-1)$. By formula (16) we get $N^{(\tau)}(l)=\frac{q}{2^{l+3}}\cdot 1=2^{m-l-3}$. $(\@slowromancap i@.2)$ For $l=e-1$, formula (17) becomes $c_{0}+c_{1}\pi+\cdots+c_{l+1}\pi^{l+1}=c(b_{0}+b_{1}\pi+\cdots+b_{l}\pi^{l}+\pi^{l+1})$. Therefore $c_{l+1}=c$ and $c_{\lambda}=cb_{\lambda}\ (0\leq\lambda\leq l)$. For $c=0$, we have $c_{0}=c_{1}=\cdots=c_{l+1}=0$ and $c_{0}+c_{l+1}=0$. For $c=1$, we have $c_{l+1}=1$, $c_{\lambda}=b_{\lambda}\ (0\leq\lambda\leq l)$ and $c_{0}+c_{l+1}=b_{0}+1$. We obtain $N^{(\tau)}(e-1)=2^{m-l-3}(1+(-1)^{b_{0}+1})=2^{m-l-3}(1-(-1)^{b_{0}})$. $(\@slowromancap i@.3)$ For $l\geq e$, then $c_{\lambda}=cb_{\lambda}\ (0\leq\lambda\leq e-1)$, $c_{e}=c$, $c_{\lambda}=0\ (e+1\leq\lambda\leq l+1)$ and $c_{0}+c_{l+1}=cb_{0}$. From formula (16) we get $\begin{array}[]{ll}N^{(\tau)}(l)=2^{m-l-3}(&~{}~{}~{}~{}~{}1~{}~{}~{}~{}~{}+~{}~{}(-1)^{b_{0}}~{}~{}).\\\ &\small{\text{(for\ c=0)}~{}~{}~{}~{}~{}~{}~{}~{}\text{(for\ c=1)}}\end{array}$ $(\@slowromancap ii@)$ For $l=m-1$, formula (17) becomes $c(b_{0}+b_{1}\pi+\cdots+b_{e-1}\pi^{e-1}+\pi^{e})=c_{0}+c_{1}\pi+\cdots+c_{m-1}\pi^{m-1}+c_{m}\pi^{m}.$ Let $f(x)=1+a_{1}x+\cdots+a_{m-1}x^{m-1}+x^{m}\ (a_{i}\in\mathbb{F}_{2})$ be the minimal polynomial of $\pi$ over $\mathbb{F}_{2}$. Then $\pi^{m}=1+a_{1}\pi+\cdots+a_{m-1}\pi^{m-1}$ and $\begin{array}[]{ll}c(b_{0}+b_{1}\pi+\cdots+b_{e-1}\pi^{e-1}+\pi^{e})=c_{0}+c_{1}\pi+\cdots+c_{m-1}\pi^{m-1}+c_{m}(1+a_{1}\pi+\cdots+a_{m-1}\pi^{m-1})\end{array}$ which imply that $\begin{array}[]{ll}cb_{0}=c_{0}+c_{m}=c_{0}+c_{l+1},\ cb_{\lambda}=c_{\lambda}+c_{l+1}a_{\lambda}\ (1\leq\lambda\leq e-1),\ c=c_{e}+c_{l+1}a_{e},\\\ 0=c_{\lambda}+c_{l+1}a_{\lambda}\ (e+1\leq\lambda\leq m-1=l).\end{array}$ For $c=0$ we have $c_{0}=c_{l+1}$, $c_{\lambda}=c_{l+1}a_{\lambda}\ (1\leq\lambda\leq l)$. For $c=1$ we have $c_{0}=b_{0}+c_{l+1}$, $c_{\lambda}=b_{\lambda}+c_{l+1}a_{\lambda}\ (1\leq\lambda\leq e-1)$, $c_{e}=1+c_{l+1}a_{e}$, $c_{\lambda}=c_{l+1}a_{\lambda}\ (e+1\leq\lambda\leq l)$. By formula (16) we get $\displaystyle\begin{array}[]{ll}N^{(\tau)}(m-1)=2^{m-l-3}(\sum\limits_{c_{l+1}=0}^{1}(-1)^{c_{l+1}+c_{l+1}}+\sum\limits_{c_{l+1}=0}^{1}(-1)^{b_{0}+c_{l+1}+c_{l+1}})\\\ ~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\small{\text{(for\ c=0)}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\text{(for\ c=1)}}\end{array}=2^{m-l-2}(1+(-1)^{b_{0}})=\frac{1}{2}(1+(-1)^{b_{0}}).$ This completes the proof of Lemma 6. ∎ Now we show our main result on arithmetic autocorrelation distribution $\\{A_{s}(\tau)\arrowvert 1\leq\tau\leq n\\}$ of any binary $m$-sequence with period $n=2^{m}-1$. ###### Theorem 7. Let $m\geq 2$, $n=2^{m}-1$, $s=(s_{\lambda}=T(\pi^{\lambda}))^{n-1}_{\lambda=0}$ be the binary $m$-sequence with period $n$ where $\pi$ is a primitive element of $\mathbb{F}_{q}\ (q=2^{m})$ and $T:\ \mathbb{F}_{q}\rightarrow\mathbb{F}_{2}$ is the trace mapping. For $1\leq\tau\leq n-1$, let $A(\tau)=A_{s}(\tau)$ be the arithmetic autocorrelation of $s$ and $(1+\pi^{\tau})^{-1}=b_{0}+b_{1}\pi+\cdots+b_{e-1}\pi^{e-1}+\pi^{e}\quad(1\leq e\leq m-1,\ b_{0},b_{1},\ldots,b_{e-1}\in\mathbb{F}_{2}).$ Then $(\@slowromancap i@)$. For $N^{(\tau)}(l)=N^{(\tau)}(0,\ 0;l)+N^{(\tau)}(0,\ 1;l)$, $\displaystyle\sum_{l=0}^{m-1}lN^{(\tau)}(l)=\left\\{\begin{array}[]{ll}2^{m-2}+2^{m-e-1}-1,&\text{if}\ b_{0}=0\\\ 2^{m-2}-2^{m-e-1},&\text{if}\ b_{0}=1.\\\ \end{array}\right.$ $(\@slowromancap ii@)$. (arithmetic autocorrelation distribution) For all $\tau$, $1\leq\tau\leq n-1$, $A_{s}(\tau)\in\\{\pm(2^{k}-1)\arrowvert 1\leq k\leq m-1\\}$. Moreover, for each $\varepsilon\in\\{0,1\\}$ and $0\leq k\leq n-1$, the number of $\tau\ (1\leq\tau\leq n-1)$ such that $A_{s}(\tau)=\varepsilon(2^{k}-1)$ is $2^{m-k-1}$. ###### Proof. For $1\leq\tau\leq n-1$ and $(1+\pi^{\tau})^{-1}=b_{0}+b_{1}\pi+\cdots+b_{e-1}\pi^{e-1}+\pi^{e}\quad(1\leq e\leq m-1,\ b_{0},\ldots,b_{e-1}\in\mathbb{F}_{2})$, by Lemma 6 we know that $\displaystyle\sum_{l=1}^{m-1}lN^{(\tau)}(l)=$ $\displaystyle\sum_{1\leq l\leq e-2}l\cdot 2^{m-l-3}+(e-1)(1-(-1)^{b_{0}})2^{m-(e-1)-3}$ $\displaystyle+\sum_{e\leq l\leq m-2}l\cdot 2^{m-l-3}(1+(-1)^{b_{0}})+\frac{1}{2}(m-1)(1+(-1)^{b_{0}})$ $\displaystyle=$ $\displaystyle 2^{m-3}[\sum_{1\leq l\leq e-2}l\cdot 2^{-l}+(e-1)(1-(-1)^{b_{0}})2^{-e+1}$ $\displaystyle+\sum_{e\leq l\leq m-1}(1+(-1)^{b_{0}})l\cdot 2^{-l}]+\frac{1}{4}(m-1)(1+(-1)^{b_{0}})$ $(\@slowromancap i@.1)$ If $b_{0}=0$, then $\displaystyle\sum_{l=1}^{m-1}lN^{(\tau)}(l)$ $\displaystyle=$ $\displaystyle 2^{m-3}[\sum_{1\leq l\leq e-2}l\cdot 2^{-l}+\sum_{e\leq l\leq m-1}l\cdot 2^{-(l-1)}]+\frac{1}{2}(m-1)$ $\displaystyle=$ $\displaystyle 2^{m-3}[\sum_{1\leq l\leq e-2}l\cdot 2^{-l}+\sum_{e-1\leq l\leq m-2}(l+1)\cdot 2^{-l}]+\frac{1}{2}(m-1)$ $\displaystyle=$ $\displaystyle 2^{m-3}[\sum_{l=1}^{m-2}l\cdot 2^{-l}+\sum_{l=e-1}^{m-2}2^{-l}]+\frac{1}{2}(m-1).$ Now we use the identity $\sum\limits_{l=1}^{d}l\cdot 2^{-l}=2-\frac{d+2}{2^{d}}\ (d\geq 1)$ which can be proved by induction. We obtain $\sum_{l=1}^{m-1}lN^{(\tau)}(l)=2^{m-3}[2-\frac{m}{2^{m}-2}+\frac{1}{2^{e-2}}-\frac{1}{2^{m-2}}]+\frac{1}{2}(m-1)=2^{m-2}+2^{m-e-1}-1.$ $(\@slowromancap i@.2)$ If $b_{0}=1$, then $\sum_{l=1}^{m-1}lN^{(\tau)}(l)=2^{m-3}[\sum_{1\leq l\leq e-2}l\cdot 2^{-l}+(e-1)2^{-e+2}].$ For $e\geq 3$, $\sum_{l=1}^{m-1}lN^{(\tau)}(l)=2^{m-3}[2-\frac{e}{2^{e-2}}+\frac{e-1}{2^{e-2}}]=2^{m-3}[2-\frac{1}{2^{e-2}}]=2^{m-2}+2^{m-e-1}.$ For $e=2$, there is no $l$ satisfing $1\leq l\leq e-2=0$, $\sum\limits_{l=1}^{m-1}lN^{(\tau)}(l)=2^{m-3}[0+2^{0}]=2^{m-3}=2^{m-2}-2^{m-e-1}\ (\text{for}\ e=2).$ For $e=1$, $\sum\limits_{l=1}^{m-1}lN^{(\tau)}(l)=2^{m-3}[0+0]=0=2^{m-2}-2^{m-e-1}\ (\text{for}\ e=1).$ $(\@slowromancap ii@)$ By Lemma 5, for $1\leq\tau\leq n-1$, $A_{s}(\tau)=n-2g(\tau)=n-2(2^{m-2}+\sum_{l=1}^{m-1}lN^{(\tau)}(l)).$ For $b_{0}=0$, $A_{s}(\tau)=2^{m}-1-2(2^{m-2}+2^{m-2}+2^{m-e-1}-1)=-(2^{m-e}-1).$ For $b_{0}=1$, $A_{s}(\tau)=2^{m}-1-2(2^{m-2}+2^{m-2}-2^{m-e-1})=2^{m-e}-1.$ Therefore $A_{s}(\tau)\in\\{\pm(2^{m-e}-1)\arrowvert 1\leq e\leq m-1\\}=\\{\pm(2^{k}-1)\arrowvert 1\leq k\leq m-1\\}.$ Moreover, $\displaystyle\\{(1+\pi^{\tau})^{-1}\arrowvert 1\leq\tau\leq n-1\\}=\mathbb{F}_{2^{m}}\setminus\\{0,1\\}$ $\displaystyle=$ $\displaystyle\\{b_{0}+b_{1}\pi+\cdots+b_{m-1}\pi^{m-1}\neq 0,\ 1\arrowvert b_{0},\ldots,b_{m-1}\in\mathbb{F}_{2}\\}$ $\displaystyle=$ $\displaystyle\\{b_{0}+b_{1}\pi+\cdots+b_{e-1}\pi^{e-1}+\pi^{e}\arrowvert b_{0},\ldots,b_{e-1}\in\mathbb{F}_{2},\ 1\leq e\leq m-1\\}.$ From part $(\@slowromancap i@)$ we know that for $1\leq\tau\leq n-1$ and $1\leq k\leq m-1$, $A_{s}(\tau)=2^{k}-1\Leftrightarrow(1+\pi^{\tau})^{-1}=1+b_{1}\pi+\cdots+b_{e-1}\pi^{e-1}+\pi^{e},\ e=m-k.$ Therefore $\displaystyle\sharp\\{1\leq\tau\leq n-1\arrowvert A_{s}(\tau)=2^{k}-1\\}$ $\displaystyle=$ $\displaystyle\sharp\\{(b_{1},\ldots,b_{e-1})\arrowvert b_{1},\ldots,b_{e-1}\in\mathbb{F}_{2}\\}$ $\displaystyle=$ $\displaystyle 2^{e-1}=2^{m-k-1}.$ Similarly, $A_{s}(\tau)=-(2^{k}-1)\Leftrightarrow(1+\pi^{\tau})^{-1}=0+b_{1}\pi+\cdots+b_{e-1}\pi^{e-1}+\pi^{e}$ and $\sharp\\{1\leq\tau\leq n-1\arrowvert A_{s}(\tau)=-(2^{k}-1)\\}=2^{e-1}=2^{m-k-1}.$ This completes Theorem 7. ∎ ###### Remark. From Theorem 7 we know that the conjecture, stated in Section I, is true. The conjecture is concern on the distribution on absolute values $\lvert A_{s}(\tau)\rvert\ (1\leq\tau\leq n-1)$. In fact, from Theorem 7 and the proof we get more informations as following. $(\@slowromancap i@)$. $\lvert A_{s}(\tau)\rvert=2^{k}-1$ if and only if $(1+\pi^{\tau})^{-1}=b_{0}+b_{1}\pi+\cdots+b_{m-k-1}\pi^{m-k-1}+\pi^{m-k}\ (1\leq k\leq m-1).$ $(\@slowromancap ii@)$. $A_{s}(\tau)=2^{k}-1\Leftrightarrow b_{0}=1$ and $A_{s}(\tau)=-(2^{k}-1)\Leftrightarrow b_{0}=0$. $(\@slowromancap iii@)$. Both of multiplicity of $\tau\ (1\leq\tau\leq n-1)$ satisfying $A_{s}(\tau)=2^{k}-1$ and $A_{s}(\tau)=-(2^{k}-1)$ are the same number $2^{m-k-1}$. ## IV Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study. ## References * [1] Z. Chen, Z. Niu, Y. Sang, C. Wu, Arithmetic autocorrelation of binary $m$-sequences, Cryptologia, DOI: 10.1080/01611194.2022.207116, 2022. * [2] Z. Chen, Z. Niu, A. Winterhof, Arithmetic crosscorrelation of pseudorandom binary sequences of coprime periods, IEEE Trans. on Inform. Theory, 68(11) (2022) 7538-7544. * [3] M. Goresky, A.Klapper, Arithmetic crosscorrelations of feedback with carry shift register sequences, IEEE Trans. on Inform. Theory, 43(4) (1997), 1342-1345. * [4] M. Goresky, A.Klapper, Some results on the arithmetic correlation of sequences. In $\langle$Sequences and Their Applications-SETA 2008$\rangle$, Lecture Notes in Computer Science, vol.5203, (2008), 71-80. * [5] M. Goresky, A.Klapper, Statistical properties of the arithmetic correlation of sequences, International Jour. of Foundations of Computer Science, 22(6) (2011), 1297-1315. * [6] M. Goresky, A.Klapper, $\langle$Algebraic Shift Register Sequences$\rangle$, Combiridege University Press, 2012. * [7] R. Hofer, A. Winterhof, On the arithmetic autocorrelation of the Legendre sequence, Advances in Mathematics of Communications, 11(1)(2017), 237-244. * [8] R. Hofer, L. Merai, A. Winterhof, Measures of pseudorandomness arithmetic autocorrelation and correlation measure, In $\langle$Number theory-Diophantine probiems, uniform distribution and applications$\rangle$, C. Elsholtz and P. Grabner edit. Springer, 303-312, 2017. * [9] D. M. Manddelbaum, Arithmetic codes with large distance, IEEE Trans. on Inform. Theory, 13(2) (1967), 237-242. * [10] H. Wang, Q. Wen and J. Zhang, GLS: New class of generalized Legendre sequences with optimal arithmetic crosscorrelation, RAIRO-Theor. Inform. Appl., 47(2013), 371-388.
# NMR evidence for a Peierls transition in the layered square-net compound LaAgSb2 Seung-Ho Baek Department of Physics, Changwon National University, Changwon 51139, Korea Department of Materials Convergence and System Engineering, Changwon National University, Changwon 51139, Korea Sergey L. Bud’ko Ames Laboratory, U.S. Department of Energy and Department of Physics and Astronomy, Iowa State University, Ames, Iowa 50011, USA Paul C. Canfield Ames Laboratory, U.S. Department of Energy and Department of Physics and Astronomy, Iowa State University, Ames, Iowa 50011, USA F. Borsa Ames Laboratory, U.S. Department of Energy and Department of Physics and Astronomy, Iowa State University, Ames, Iowa 50011, USA Byoung Jin Suh<EMAIL_ADDRESS>Department of Physics,The Catholic University of Korea, Bucheon 14662, Korea ###### Abstract We measured the central ($1/2\leftrightarrow-1/2$) and first satellite ($\pm 3/2\leftrightarrow\pm 1/2$) lines of the 139La NMR spectra as a function of temperature in LaAgSb2, in order to elucidate the origin and nature of the charge-density-wave (CDW) transitions at $T_{\text{CDW1}}=207$ K and $T_{\text{CDW2}}=186$ K. In the normal state, the Knight shift $\mathcal{K}$ reveals a fairly linear relationship with decreasing temperature, which is ascribed to a pseudogap in the spin excitation spectrum, pointing towards the material being an unconventional metal. Upon further cooling, $\mathcal{K}$ decreases more steeply below $T_{\text{CDW1}}$, indicative of the partial Fermi surface gap opening on top of the pseudogap. The most remarkable finding in our study is a clear splitting of the satellite lines at $T_{\text{CDW1}}$ observed for $H\parallel c$, whose temperature dependence behaves as the BCS order parameter in the weak-coupling limit, evidencing that the CDW transition induces the periodic lattice distortion. Our NMR findings therefore demonstrate that the CDW transition in LaAgSb2 is of Peierls type, being driven by the electronic instability in the vicinity of the Fermi level. ## I Introduction The charge-density-wave (CDW) phenomenon has been observed in numerous two- dimensional (2D) layered metallic compounds such as transition-metal dichalcogenides (TMDs) and high-$T_{c}$ cuprates, as well as quasi-one- dimensional (1D) metals [1, 2, 3, 4, 5]. However, whereas CDW in 1D is well understood by the conventional Peierls’ model [6] in which the Fermi surface nesting (FSN) drives a periodic lattice distortion (PLD), the origin of CDW in higher dimensions is much more complicated than in 1D [7] and strongly material dependent [8]. As examples, rather than FSN, strong electron correlations appear to be a crucial ingredient for charge ordering in high-$T_{c}$ cuprates [5], or local electron-phonon coupling (EPC) may be a main driving force for the CDW in TMDs [9, 10]. LaAgSb2 is a unique member of the layered Sb square-net family $R$AgSb2 ($R$=rare earth) in space group $P4/nmm$ [11, 12], isostructural to the Fe- oxypnictide LaFeAsO [13]. The square-net is featured by a shorter in-plane atomic distance ($a/\sqrt{2}$) than the lattice constant $a$, resulting in a twice larger unit cell in real space [see Fig. 1(a)] or a folding of the Brillouin zone in reciprocal space [14]. The resultant linear band crossings may occur near the Fermi level, allowing to host Dirac fermions [15], as indeed observed in the Bi square-net, SrMnBi2 [16]. Interestingly, unlike other members of $R$AgSb2, most of which have a magnetically ordered ground state, LaAgSb2 is nonmagnetic and undergoes two successive CDW transitions (CDW1 at 207 K and CDW2 at 186 K) [17] while exhibiting unusual large linear magnetoresistance (MR) [11], which may be attributed to the quantum limit of the Dirac fermions [18]. LaAgSb2 reveals not only a Dirac-cone-like band structure near the Fermi level [18, 19], but also well-nested segments of the Fermi surface directly linked to the Dirac cone, suggesting a close relationship between the Dirac fermions and the CDW ordering [19]. In this work we have performed 139La NMR to investigate the microscopic origin of the CDW transitions in LaAgSb2. As 139La has a large quadrupole moment, it could probe changes of both the local spin susceptibility and charge environments, via the Knight shift $\mathcal{K}$ and the quadrupole frequency $\nu_{Q}$. Our NMR data indicate that the CDW1 transition opens up a Fermi surface gap and causes the periodic lattice distortion, whose order parameter is well described by BCS mean-field theory, at the same time. Together with the observation of phonon softening at the CDW wave vector [20, 21], we conclude that LaAgSb2 is a rare quasi-2D material that undergoes a nesting- driven weak-coupling CDW transition. ## II Experimental details High-quality LaAgSb2 single crystals were flux grown from a typical Sb rich self-flux, La0.045Ag0.091Sb0.864, as described in Ref. [11]. The size of the crystal used in this NMR measurement is roughly $1\times 2\times 4$ mm3. 139La (nuclear spin $I=7/2$) NMR measurements were carried out at the external field of 4.7 T in the range of temperature 100–300 K. The signal intensity turned out very weak in our experiments, which is ascribed largely to the low filling factor owing to the plate-like single crystal within the NMR coil, and partly to small rf penetration in a metal that is limited by the skin depth, $\delta=\sqrt{2\rho/\omega\mu}$ where $\rho$ is the resistivity, $\omega$ is the angular frequency of the rf, and $\mu$ is the magnetic permeability [22]. Due to the weak signal intensity, we focused solely on measuring the temperature dependence of the NMR spectra, without attempting to measure the spin-lattice relaxation rate, $T_{1}^{-1}$. Further, among the three pairs of satellites expected for $I=7/2$, we have measured the first satellite pair only, because it yields sufficient information regarding the electric field gradient and its temperature evolution at the 139La in an axial symmetry. The 139La NMR spectra were obtained by a conventional Hahn spin-echo technique with a typical $\pi/2$ pulse length of 3 $\mu$s. ## III Results ### III.1 Quadrupole-perturbed 139La NMR spectra For a nuclear spin $I>1/2$, there are central ($\frac{1}{2}\leftrightarrow-\frac{1}{2}$) and satellite transitions between the $m$th and $(m-1)$th levels ($m=-I,-I+1,\cdots,+I$). In an axial symmetric surrounding (asymmetry parameter $\eta=0$), the allowed transitions to first order are given by [22] $\begin{split}\nu(m\leftrightarrow m-1)=\;&\nu_{0}(1+\mathcal{K})\\\ &+\frac{1}{2}\nu_{Q}(3\cos^{2}\theta-1)\left(m-\frac{1}{2}\right),\end{split}$ (1) where $\mathcal{K}$ is the Knight shift, $\nu_{0}$ is the unshifted Larmor frequency, $\nu_{Q}$ is the nuclear quadrupole frequency, and $\theta$ is the angle between the principal axis $z$ of the electric field gradient (EFG) and an external field $\mathbf{H}$. Figure 1(b) shows the central ($1/2\leftrightarrow-1/2$) and first satellite ($\pm 3/2\leftrightarrow\pm 1/2$) lines of the 139La NMR spectra measured at $H=4.7$ T for $H\perp c$ and $H\parallel c$, respectively, at 292 K. The observed NMR spectra are excellently described by Eq. (1), in which the distance between the first satellite lines is given by $\nu_{Q}$ and $2\nu_{Q}$ for $H\perp c$ and $H\parallel c$, respectively, indicating that the local symmetry at 139La is indeed axial with respect to the principal axis of the EFG which lies along the $c$ axis. Interestingly, the line shape of each satellites appears to be asymmetric while the satellite pairs are symmetric about the central transition, similarly for both field orientations (see Fig. 2). It should be noted that the FWHM of the NMR lines are very narrow, less than 10 kHz for the satellites which corresponds to only 1% of $\nu_{Q}$ (see Fig. 3), evidencing the very high homogeneity of our sample. Therefore, the asymmetric line shape of the satellites seems to reflect a feature of the satellite transitions ($\pm 1/2\leftrightarrow\pm 3/2$). Figure 1: (a) Crystal structure of LaAgSb2 viewed along the $b$ (upper) and $c$ (lower) axes. The La atoms above and below the Ag-Sb2 plane are staggered, and the Ag and Sb1 atoms form an identical square-net, resulting in the $P4/nmm$ structure. The dashed lines denote the unit cell. (b) 139La central and first satellite transitions for $H\perp c$ (upper) and $H\parallel c$ (lower) measured at 292 K and an external field of 4.7 T. The separations of the satellites for both directions are well described in terms of the single quadrupole frequency $\nu_{Q}=1.339$ MHz, as expected for an axial symmetry about the $c$ axis. ### III.2 Knight shift and Fermi surface gap The temperature dependence of the NMR lines is presented in Fig. 2. To begin with, we discuss the temperature dependence of the 139La central line, shown in the middle panels of Figs. 2(a) and 2(b) for $H\parallel c$ and $H\perp c$, respectively. While the FWHM of the central line remains very narrow down to 100 K, the line clearly shifts to lower frequency with decreasing temperature, involving a weak line broadening. One can note that a small shoulder peak, which persists up to room temperature, is detected at the high-frequency side for both field directions. The small peak suggests that a small portion of the 139La sites experiences slightly different hyperfine fields from the majority. Regardless of its origin and nature, the peak is so small that it could hardly affect the satellites, and it will be ignored for our purpose. Figure 2: Temperature dependence of 139La NMR spectra for the central and first satellite transitions measured at $H=4.7$ T applied (a) parallel and (b) perpendicular to the $c$ axis. The left axis refers to the temperature at which each spectrum was obtained, whereas the amplitude of the spectra is plotted in arbitrary units. The central lines for both field directions similarly shift to lower frequency, i.e., the Knight shift is decreased, with decreasing temperature. Whereas the satellite lines for $H\parallel c$ abruptly split below $T_{\text{CDW1}}=207$ K (marked by the red arrows), those for $H\perp c$ significantly broaden below $T_{\text{CDW2}}=186$ K (marked by the cyan arrows). The Knight shift $\mathcal{K}\equiv(\nu-\nu_{0})/\nu_{0}\times 100$ % as a function of temperature is plotted in Fig. 3(a). For $H\perp c$ ($\theta=90^{\circ}$), the second order quadrupole shift of the central line given by $15\nu_{Q}^{2}/16\nu_{0}(1-\cos^{2}\theta)(1-9\cos^{2}\theta)$ for $I=7/2$ [22], was subtracted from the total shift. In the normal state, we find that $\mathcal{K}$ decreases in an almost linear fashion with temperature, albeit very small, for both field orientations. For a simple Pauli metal, $\mathcal{K}$ or the local spin susceptibility is proportional to the density of states (DOS) at the Fermi level, $n(\epsilon_{F})$, so it would have been expected to be independent of temperature in the normal state. Therefore, the unusual decrease of $\mathcal{K}$ may be ascribed to a pseudogap, which is frequently observed in many layered strange metals such as high-$T_{c}$ cuprates [23], Fe-based superconductors [24, 25], and TMDs [26, 27]. In fact, LaAgSb2 shows a rather atypical linear temperature dependence of resistivity in the normal state [11, 17], and is believed to be a topological semimetal that hosts Dirac fermions [18, 19], being likely responsible for the pseudogap-like behavior. Upon further lowering temperature, we find that $\mathcal{K}$ decreases more rapidly just below $T_{\text{CDW1}}$, deviating from the linear temperature dependence in the normal state. The small but clear additional reduction of $\mathcal{K}$ at $T_{\text{CDW1}}$ observed for the two field directions suggests that $n(\epsilon_{F})$ is further suppressed by the CDW1 transition, greatly supporting that the CDW1 transition opens up a gap at the Fermi surface [20]. Interestingly, no additional notable change of $\mathcal{K}$ was observed at $T_{\text{CDW2}}$. This may suggest either that the CDW2 transition is too weak to yield a sizable gap, or that the nature of the CDW2 is different from that of the CDW1, as conjectured by the very different CDW wave vectors [17]. ### III.3 Periodic lattice distortion in the CDW state Now we evaluate how the charge environment surrounding the 139La nuclei evolves through the CDW transitions, via analysis on the temperature dependence of the satellites, giving rise to the information of the EFG and its spatial distribution. Most remarkably, we observed a clear splitting of the satellites for $H\parallel c$ just below $T_{\text{CDW1}}$, as shown in Fig. 2(a). The $\nu_{Q}$ values extracted from the distance between the central and lower satellite lines are plotted in Fig. 3(b), which reveals the sharp splitting of the satellite lines. This means that the CDW1 transition induces two inequivalent 139La sites which experience different EFG on average. Moreover, we confirmed that the difference of the $\nu_{Q}$ values between the split satellites, $\Delta\nu_{Q}$, serves as an order parameter for the CDW1 phase, which is consistent with the BCS mean field theory in the weak coupling limit, i.e., $\Delta\nu_{Q}\propto(1-T/T_{\text{CDW1}})^{1/2}$ [see the inset of Fig. 3(b)]. This result bears strong resemblance to the behavior of the CDW modulation intensity below $T_{\text{CDW1}}$ observed by x-ray scattering measurements [17], indicating that $\Delta\nu_{Q}$ is directly related to the lattice modulation associated with CDW order. That is, the local periodic lattice distortion emerges as a direct consequence of the charge modulation. Nevertheless, the presence of the two distinct 139La sites is surprising, because the extremely long wavelength of the CDW modulation ($\lambda_{1}\sim 39a$) would result in a spatial distribution of the EFG at the 139La sites, leading to an inhomogeneous quadrupole broadening rather than the split lines. This suggests that the PLD/CDW is formed in a way that discriminates the staggered La atoms above and below Ag-Sb2 plane in the unit cell [see Fig. 1(a)]. Figure 3: (a) Temperature dependence of the Knight shift $\mathcal{K}$. For both field directions, $\mathcal{K}$ decreases linearly with lowering temperature. Below $T_{\text{CDW1}}$ the slope becomes larger suggesting a partial gap opening at the Fermi surface. (b) The quadrupole frequency $\nu_{Q}$ as a function of temperature. $\nu_{Q}$ sharply splits into two at $T_{\text{CDW1}}$. Inset shows an order parameter-like behavior of $\Delta\nu_{Q}$. The solid line is given by the weak coupling BCS theory. (c) and (d) FWHM for the central and first satellite lines as a function of temperature. Whereas the FWHM of all the NMR lines increases weakly for both field directions at $T_{\text{CDW1}}$, a significant line broadening was observed below $T_{\text{CDW2}}$ only for $H\perp c$. In contrast, the satellites for $H\perp c$ are nearly unchanged through $T_{\text{CDW1}}$. The anisotropic changes of the spectra in the CDW1 phase may be understood by the strong dependence of the quadrupole effect on the angle between the direction of the EFG and $H$. Since the lattice modulation for the CDW1 is developed along the $a$ axis [17], $H$ is perpendicular to the EFG modulation which generates the two 139La sites. In other words, $H$ has the same effects on the two 139La sites, revealing the difference between them. However, the situation is quite different for $H\perp c$. Namely, the otherwise differentiated 139La sites may feel a similar quadrupole-perturbed Zeeman field for $H\perp c$ because $H$ is applied along an arbitrary direction in the $ab$ plane in our experimental setup. Note that the similar weak broadening observed for all the NMR lines below $T_{\text{CDW1}}$, as shown in Figs. 3(c) and 3(d), is what would be expected from a spatial distribution of the EFG. Another prominent feature found in Fig. 2 is a significant line broadening of the NMR lines below the CDW2 transition temperature $T_{\text{CDW2}}$ that occurs only for $H\perp c$ [see Figs. 3(c) and 3(d)]. Further, we find that the satellite lines for $H\perp c$ rapidly spread out in the CDW2 phase, becoming vanishingly small below $\sim 170$ K. While the external magnetic field anisotropy is naturally ascribed to the EFG modulation induced by the CDW2 modulation that is along the $c$ axis [17], the considerable broadening of the satellite lines for $H\perp c$ suggests that the EFG distribution is much wider for the CDW2, probably due to the much shorter modulation wavelength of the CDW2 ($\lambda_{2}\sim 6.3c$) than the CDW1 ($\lambda_{1}\sim 39a$) [17]. ## IV Discussion We have established that the CDW transition at 207 K causes a (partial) gap opening at the Fermi level and generates two inequivalent 139La sites, the difference in quadrupole frequency $\Delta\nu_{Q}$ of which is well described by the BCS mean field theory. These NMR findings microscopically prove that the CDW1 transition in LaAgSb2 is governed by the conventional (Peierls type) weak-coupling mechanism in which the Fermi surface nesting instability drives the periodic lattice distortion as well as the Fermi surface gap opening via electron-phonon interactions. This is consistent with the observation of acoustic phonon softening and the corresponding Kohn anomalies at the CDW wave vector [20, 21]. It is worthwhile to compare the aforementioned with the recent NMR results obtained in the transition metal dichalcogenide, $2H$-TaSe2, within which a strong-coupling mechanism driven by local electron-phonon coupling appears to underlie the CDW transition [28]. The seemingly different CDW mechanism in the two layered 2D materials may be related to the structural motif of the TMDs and the square-net materials. Namely, unlike TMDs whose structural symmetry is lower than tetragonal, the square-net has an inherent instability against a Peierls transition, similar to 1D metals which are always unstable toward a lattice distortion [6]. Nevertheless, the square-net could be stabilized, e.g., by the overlap of $d$-orbitals from the transition-metal elements with the $s$ and $p$ orbitals in the square-net [14]. As to the CDW2 transition at 186 K, the anomalous anisotropic line broadening of the 139La NMR spectra below 186 K reveals that the EFG modulation is formed along the $c$ axis due to the CDW2 ordering. However, neither a Fermi surface gap opening, although it is suggested by the resistivity anomaly for $H\parallel c$ [17], nor the order parameter associated with the CDW2 was observed in our work, thus leaving the origin and nature of the CDW2 still an open question. ## V Summary We carried out the 139La NMR investigation in the Sb square-net LaAgSb2. In the normal state, we found an unusual pseudogap behavior, reflecting the unconventional metallic state of the material. At the CDW transition at 207 K, the 139La NMR spectra undergo sharp changes, establishing that a CDW gap opens up at the Fermi surface and two inequivalent 139La sites are generated as a result of the periodic lattice distortion. These provide microscopic evidence that LaAgSb2 undergoes a Peierls transition, which is quite rare in non-1D materials. Our NMR studies strongly suggest that the square-net not only induces the band inversions being a good structural motif for topological semimetals that host 2D Dirac fermions, it could be also an excellent platform to study the CDW phenomenon in quasi-two dimensions and its relationship with Dirac fermions. ###### Acknowledgements. We thank Changyong Song for useful discussions. This work was supported by a National Research Foundation of Korea (NRF) grant funded by the Korea Government(MSIT) (Grant No. NRF-2020R1A2C1003817). Work at the Ames Laboratory was supported by the U.S. Department of Energy, Office of Science, Basic Energy Sciences, Materials Sciences and Engineering Division. The Ames Laboratory is operated for the U.S. Department of Energy by Iowa State University under Contract No. DE-AC02-07CH11358. ## References * Wilson _et al._ [1975] J. Wilson, F. D. Salvo, and S. Mahajan, Charge-density waves and superlattices in the metallic layered transition metal dichalcogenides, Adv. Phys. 24, 117 (1975). * Salvo and Rice [1979] F. J. D. Salvo and T. M. Rice, Charge-density waves in transition-metal compounds, Phys. Today 32, 32 (1979). * Grüner [1994] G. Grüner, _Density waves in solids_ (Perseus Publishing, Cambridge, Massachusetts, 1994). * Rossnagel [2011] K. Rossnagel, On the origin of charge-density waves in select layered transition-metal dichalcogenides, J. Phys.: Condens. Matter 23, 213001 (2011). * Chen _et al._ [2016] C.-W. Chen, J. Choe, and E. Morosan, Charge density waves in strongly correlated electron systems, Rep. Prog. Phys. 79, 084505 (2016). * Peierls [1955] R. E. Peierls, _Quantum theory of solids_ (Oxford University Press, Oxford, 1955). * Johannes and Mazin [2008] M. D. Johannes and I. I. Mazin, Fermi surface nesting and the origin of charge density waves in metals, Phys. Rev. B 77, 165135 (2008). * Zhu _et al._ [2017] X. Zhu, J. Guo, J. Zhang, and E. W. Plummer, Misconceptions associated with the origin of charge density waves, Adv. Phys. -X 2, 622 (2017). * Johannes _et al._ [2006] M. D. Johannes, I. I. Mazin, and C. A. Howells, Fermi-surface nesting and the origin of the charge-density wave in $\mathrm{Nb}{\mathrm{Se}}_{2}$, Phys. Rev. B 73, 205102 (2006). * Calandra _et al._ [2009] M. Calandra, I. I. Mazin, and F. Mauri, Effect of dimensionality on the charge-density wave in few-layer $2{H\text{-NbSe}}_{2}$, Phys. Rev. B 80, 241108 (2009). * Myers _et al._ [1999a] K. D. Myers, S. L. Bud’ko, I. R. Fisher, Z. Islam, H. Kleinke, A. H. Lacerda, and P. C. Canfield, Systematic study of anisotropic transport and magnetic properties of RAgSb2 (R=Y, La–Nd, Sm, Gd–Tm), J. Magn. Magn. Mater. 205, 27 (1999a). * Myers _et al._ [1999b] K. D. Myers, S. L. Bud’ko, V. P. Antropov, B. N. Harmon, P. C. Canfield, and A. H. Lacerda, de Haas–van Alphen and Shubnikov–de Haas oscillations in $R{\mathrm{AgSb}}_{2}$ ($R=\mathrm{Y},$ La-Nd, Sm), Phys. Rev. B 60, 13371 (1999b). * Kamihara _et al._ [2008] Y. Kamihara, T. Watanabe, M. Hirano, and H. Hosono, Iron-based layered superconductor La(O1-xFx)FeAs ($x=0.05-0.12$) with $T_{c}=26$ K, J. Am. Chem. Soc. 130, 3296 (2008). * Klemenz _et al._ [2019] S. Klemenz, S. Lei, and L. M. Schoop, Topological semimetals in square-net materials, Annu. Rev. Mater. Res. 49, 185 (2019). * Young and Kane [2015] S. M. Young and C. L. Kane, Dirac semimetals in two dimensions, Phys. Rev. Lett. 115, 126803 (2015). * Park _et al._ [2011] J. Park, G. Lee, F. Wolff-Fabris, Y. Y. Koh, M. J. Eom, Y. K. Kim, M. A. Farhan, Y. J. Jo, C. Kim, J. H. Shim, and J. S. Kim, Anisotropic Dirac Fermions in a Bi Square Net of SrMnBi2, Phys. Rev. Lett. 107, 126402 (2011). * Song _et al._ [2003] C. Song, J. Park, J. Koo, K.-B. Lee, J. Y. Rhee, S. L. Bud’ko, P. C. Canfield, B. N. Harmon, and A. I. Goldman, Charge-density-wave orderings in ${\mathrm{LaAgSb}}_{2}:$ An x-ray scattering study, Phys. Rev. B 68, 035113 (2003). * Wang and Petrovic [2012] K. Wang and C. Petrovic, Multiband effects and possible Dirac states in LaAgSb2, Phys. Rev. B 86, 155213 (2012). * Shi _et al._ [2016] X. Shi, P. Richard, K. Wang, M. Liu, C. E. Matt, N. Xu, R. S. Dhaka, Z. Ristic, T. Qian, Y.-F. Yang, C. Petrovic, M. Shi, and H. Ding, Observation of Dirac-like band dispersion in ${\mathrm{LaAgSb}}_{2}$, Phys. Rev. B 93, 081105 (2016). * Chen _et al._ [2017] R. Y. Chen, S. J. Zhang, M. Y. Zhang, T. Dong, and N. L. Wang, Revealing Extremely Low Energy Amplitude Modes in the Charge-Density-Wave Compound ${\mathrm{LaAgSb}}_{2}$, Phys. Rev. Lett. 118, 107402 (2017). * Bosak _et al._ [2021] A. Bosak, S.-M. Souliou, C. Faugeras, R. Heid, M. R. Molas, R.-Y. Chen, N.-L. Wang, M. Potemski, and M. Le Tacon, Evidence for nesting-driven charge density wave instabilities in the quasi-two-dimensional material ${\mathrm{LaAgSb}}_{2}$, Phys. Rev. Research 3, 033020 (2021). * Carter _et al._ [1977] G. C. Carter, L. H. Bennett, and D. J. Kahan, _Metallic shift in NMR_ (Pergamon, New York, 1977). * Norman _et al._ [2005] M. R. Norman, D. Pines, and C. Kallin, The pseudogap: friend or foe of high $T_{c}$?, Adv. Phys. 54, 715 (2005). * Xu _et al._ [2011] Y.-M. Xu, P. Richard, K. Nakayama, T. Kawahara, Y. Sekiba, T. Qian, M. Neupane, S. Souma, T. Sato, T. Takahashi, H.-Q. Luo, H.-H. Wen, G.-F. Chen, N.-L. Wang, Z. Wang, Z. Fang, X. Dai, and H. Ding, Fermi surface dichotomy of the superconducting gap and pseudogap in underdoped pnictides, Nat Commun 2, 392 (2011). * Moon _et al._ [2012] S. J. Moon, A. A. Schafgans, S. Kasahara, T. Shibauchi, T. Terashima, Y. Matsuda, M. A. Tanatar, R. Prozorov, A. Thaler, P. C. Canfield, A. S. Sefat, D. Mandrus, and D. N. Basov, Infrared Measurement of the Pseudogap of P-Doped and Co-Doped High-Temperature BaFe2As2 Superconductors, Phys. Rev. Lett. 109, 027006 (2012). * Klemm [2000] R. Klemm, Striking similarities between the pseudogap phenomena in cuprates and in layered organic and dichalcogenide superconductors, Physica C 341-348, 839 (2000). * Borisenko _et al._ [2008] S. V. Borisenko, A. A. Kordyuk, A. N. Yaresko, V. B. Zabolotnyy, D. S. Inosov, R. Schuster, B. Büchner, R. Weber, R. Follath, L. Patthey, and H. Berger, Pseudogap and charge density waves in two dimensions, Phys. Rev. Lett. 100, 196402 (2008). * Baek _et al._ [2022] S.-H. Baek, Y. Sur, K. H. Kim, M. Vojta, and B. Büchner, Interplay of charge density waves, disorder, and superconductivity in 2H-TaSe2 elucidated by NMR, New J. Phys. 24, 043008 (2022).
# Non-stationary difference equation, affine Laumon space and quantization of discrete Painlevé equation Hidetoshi Awata , Koji Hasegawa , Hiroaki Kanno , Ryo Ohkawa , Shamil Shakirov , Jun’ichi Shiraishi and Yasuhiko Yamada H.Awata: Graduate School of Mathematics, Nagoya University, Nagoya 464-8602, Japan <EMAIL_ADDRESS>K.Hasegawa: Mathematical Institute, Tohoku University, Sendai 980-8578, Japan<EMAIL_ADDRESS>H.Kanno: Graduate School of Mathematics and Kobayashi-Maskawa Institute, Nagoya University, Nagoya 464-8602, Japan<EMAIL_ADDRESS>R.Ohkawa: Osaka Central Advanced Mathematical Institute, Osaka Metropolitan University, Osaka 558-8585, Japan; Research Institute for Mathematical Sciences, Kyoto University, Kyoto 606-8502, Japan<EMAIL_ADDRESS><EMAIL_ADDRESS>Sh.Shakirov: University of Geneva, Switzerland; Institute for Information Transmission Problems, Moscow, Russia <EMAIL_ADDRESS>J.Shiraishi: Graduate School of Mathematical Sciences, University of Tokyo, Komaba, Tokyo 153-8914, Japan <EMAIL_ADDRESS>Y.Yamada: Department of Mathematics, Kobe University, Rokko, Kobe 657-8501, Japan<EMAIL_ADDRESS> ###### Abstract. We show the relation of the non-stationary difference equation proposed by one of the authors and the quantized discrete Painlevé VI equation. The five dimensional Seiberg-Witten curve associated with the difference equation has a consistent four dimensional limit. We also show that the original equation can be factorized as a coupled system for a pair of functions $(\mathcal{F}^{(1)},\mathcal{F}^{(2)})$, which is a consequence of the identification of the Hamiltonian as a translation element in the extended affine Weyl group. We conjecture that the instanton partition function coming from the affine Laumon space provides a solution to the coupled system. Dedicated to the memories of our friends, Omar Foda and Yaroslav Pugai ###### Contents 1. 1 Introduction 2. 2 Non-stationary difference equation 3. 3 Quantized discrete Painlevé VI equation 4. 4 Quantum Seiberg-Witten curve 5. 5 Relation to affine Laumon space 6. A Difference analogue of Painlevé VI equation 7. B Evidences for the Conjecture 8. C Four dimensional limit in a factorized form ## 1\. Introduction The conformal blocks, more precisely, the matrix elements or the traces of the intertwiners between the Verma modules of the Virasoro algebra, or the chiral algebra in general define special functions which are ubiquitous in mathematics and physics. As special functions originated from the representation theory of the symmetry, the hypergeometric series and Nekrasov function [36] to mention a few, the conformal blocks should satisfy sufficiently simple but significant equations.The Belavin-Polyakov- Zamolodchikov (BPZ) equation for the Virasoro conformal block with degenerate field insertion is a celebrated example [7]. For the deformed Virasoro algebra [46] such an equation is expected to be some (non-stationary) difference equation. Though the conformal blocks allow integral representations of Dotsenko-Fateev type or the deformed matrix model type, the desired equations were not known for a long time, more than a quarter century after the discovery of the algebra. While most attempts to work out explicit form of the expected difference equation were not successful, the non-stationary Ruijsenaar function has been proposed [45] (see also [32]). In an appropriate limit, we can see that it satisfies the non- stationary affine Toda equation which is a difference equation that involves the $q$-exponent of the Laplace operator (the $q$-Borel transformation $\mathcal{B}$, see Definition 2.2). Recently a remarkable progress has been made by one of the authors [44] based on AGT correspondence [2]. Namely a non- stationary difference equation for the Nekrasov partition function of the five dimensional gauge theory with a surface defect was discovered. The AGT correspondence tells that if the theory has four matter hypermultiplets in the fundamental representation, the partition function agrees with the genus zero five point function with one degenerate field coming from the defect [1],[47],[33]. In [44] the five dimensional lift of AGT correspondence [5] was applied, where the surface defect is realized by imposing the Higgsing condition on the $SU(2)\times SU(2)$ quiver gauge theory. Hence the non- stationary difference equation proposed in [44] is regarded as a $q$-deformed version of BPZ equation. In the decoupling limit of the matter multiplets it reproduces the non-stationary affine Toda equation for the non-stationary Ruijsenaar function. A distinguished feature of these equations is the appearance of the $q$-Borel transformation $\mathcal{B}$. The quantization of the Painlevé equations (or isomonodromic deformations more generally) has been studied for many years. One of the motivation of such studies was its relation to the conformal field theories. This relation plays a key role in the recent studies of AGT correspondence (see [40] and references therein). Also, there has been a large progress in the study of discrete (or difference) analog of Painlevé equations in the last decades. The discrete Painlevé equations are classified into additive, multiplicative (i.e. $q$-), and elliptic-difference equations [43], and each class corresponds to gauge theories in four, five and six dimensions; ${\begin{array}[]{lcccccccccccccccccc}{\rm elliptic}&E_{8}^{(1)}\\\ &&&&&&&&&&&&&&&&A_{1}^{(1)}\\\\[-8.53581pt] &&&&&&&&&&&&&&&\nearrow\\\ {\rm multiplicative}&E_{8}^{(1)}&\rightarrow&E_{7}^{(1)}&\rightarrow&E_{6}^{(1)}&\rightarrow&{\color[rgb]{1,0,0}D_{5}^{(1)}}&\rightarrow&A_{4}^{(1)}&&\rightarrow&A_{2+1}^{(1)}&\rightarrow&A_{1+1}^{(1)}&\rightarrow&A_{1}^{(1)}&\rightarrow&A_{0}^{(1)}\\\\[14.22636pt] {\rm additive}&E_{8}^{(1)}&\rightarrow&E_{7}^{(1)}&\rightarrow&E_{6}^{(1)}&&\rightarrow&&D_{4}^{(1)}&&\rightarrow&A_{3}^{(1)}&\rightarrow&A_{1+1}^{(1)}&\rightarrow&A_{1}^{(1)}&\rightarrow&A_{0}^{(1)}\\\ &&&&&&&&&&&&&\searrow&&\searrow\\\ &&&&&&&&&&&&&&A_{2}^{(1)}&\rightarrow&A_{1}^{(1)}&\rightarrow&A_{0}^{(1)}\\\ \end{array}}$ Such a correspondence can be easily seen at classical level, however, the understanding at quantum level is incomplete so far. Fortunately, for the $q$-difference Painlevé VI equation [26] relevant for this paper, a natural quantization was worked out in [23] (see also [24], [31]) based on the extended affine Weyl group symmetry of type $D_{5}^{(1)}$. Recall that the Bäcklund transformations for the discrete Painlevé equations are generated by the affine Weyl group and the automorphisms of the Dynkin diagram, which act on the dynamical variables as birational transformations. In this paper, the discrete Painlevé VI equation always (except for Appendix A) means the quantized one in the sense of [23], where the dynamical variables $(F,G)$ are non-commutative and the time evolution is defined by the adjoint action of the Hamiltonian. Since the prefix $q$\- is already used for classical $q$-difference analogue of Painlevé VI equation, we will call the quantization of the equation $qq$-Painlevé VI for short, namely the double ‘$q$’ for $q$-difference and quantization. We warn the reader that the use of $qq$\- does not mean any direct connection to the $qq$-character introduced in [37]. But there is a similarity in the sense that the full $\Omega$-background parameters $(q,t)$ [36] are turned on in both cases. The fact that one of the $\Omega$-background parameters, say $q$, plays the role of the quantization parameter of integrable systems is the same as the Nekrasov-Shatashvili limit [39] , which corresponds to the autonomous limit of the Painlevé equations. The recent paper [44] gives us valuable lessons on the problem of the quantization of the discrete Painlevé equation. The problem is also discussed from the viewpoint of the cluster integrable system [8]. The relevant cluster algebras are associated with the BPS quiver of five dimensional superconformal field theories [10]. As we have mentioned above, in [23] by constructing a representation of the extended Weyl group $\widetilde{W}(D_{5}^{(1)})$ on the non-commutative variables $(F,G)$, explicit forms of the $qq$-Painlevé equation were derived both in the Heisenberg and the Schrödinger forms; ###### Definition 1.1 (Heisenberg form of $qq$-Painlevé VI). $\displaystyle\overline{F}F=qb_{7}b_{8}{G+b_{5}\over G+b_{7}}{G+b_{6}\over G+b_{8}},\qquad G\underline{G}=qb_{3}b_{4}{F+b_{1}\over F+b_{3}}{F+b_{2}\over F+b_{4}},$ where $\overline{F}=T\cdot F,~{}\underline{G}=T^{-1}\cdot G$ and $T$ is a translation element in $\widetilde{W}(D_{5}^{(1)})$. $\mathbf{b}=(b_{1},\cdots,b_{8})$ are the standard parameters for the Painlevé VI equation (See Appendix A). ###### Definition 1.2 (Schrödinger form of $qq$-Painlevé VI). $\displaystyle H_{\mathrm{VI}}\cdot\mathsf{u}(b,G,Q|q,t^{-1})=\mathsf{u}(b,G,Q|q,t^{-1}),$ (1.1) with the Hamiltonian given by $\displaystyle H_{\mathrm{VI}}:=\frac{1}{\varphi(-qb_{5}G^{-1})\varphi(-qb_{6}G^{-1})\varphi(-b_{7}^{-1}G)\varphi(-b_{8}^{-1}G)}\theta(F^{-1}G;q)^{-1}$ $\displaystyle\times\frac{1}{\varphi(qt^{-1/2}b_{1}G^{-1})\varphi(qt^{-1/2}b_{2}G^{-1})\varphi(t^{-1/2}b_{3}^{-1}G)\varphi(t^{-1/2}b_{4}^{-1}G)}\theta(F^{-1}G;q)^{-1}T_{t^{1/4},\mathbf{b}},$ where $\varphi(x):=(x;q)_{\infty}$ and $\theta(X;q)=(X;q)_{\infty}(q/X;q)_{\infty}$. $T_{t^{1/4},\mathbf{b}}$ is a shift operator of the parameters $\mathbf{b}$ and $Q$ appears as a parameter of wave functions. In this paper we first show that the non-stationary equation proposed in [44] is successfully identified with the $qq$-Painlevé VI equation. Namely we prove ###### Proposition 1.3. By an appropriate gauge transformation from $\mathsf{U}(\Lambda,x)$ to $\mathsf{u}(b,G,Q|q,t^{-1})$, the non-stationary difference equation in [44] $\displaystyle\mathsf{U}(t\Lambda,x)=\mathcal{A}_{1}(\Lambda,x)\cdot\mathcal{B}\cdot\mathcal{A}_{2}(\Lambda,x)\cdot\mathcal{B}\cdot\mathcal{A}_{3}(\Lambda,x)\mathsf{U}(\Lambda,{x\over tqQ})$ (1.2) is transformed to the $qq$-Painlevé VI equation (1.1), where $\mathcal{B}$ is the $q$-Borel transformation and $\mathcal{A}_{i}(\Lambda,x)$ are multiplications of $\varphi(x)$ and $\Phi(x):=(x;q,t)_{\infty}$ (see section 2). In contrast to the original form of the $qq$-Painlevé equation the double infinite product $\Phi(x):=(x;q,t)_{\infty}$ arises as a consequence of the gauge transformation. In (1.2) the parameter $x$ is related to the dynamical variable $G$ and $\Lambda$ plays the role of the time variable. (See subsection 2.2 for the dictionary of variables between Painlevé side and the gauge theory side). $\mathsf{U}(\Lambda,x)$ is regarded as a formal power series in $x$ and $\Lambda/x$, which is motivated by the following conjecture that the Nekrasov partition function solves the above equation. Hence the virtue of the gauge transformation in Proposition 1.3 is due to the conjecture that the Nekrasov partition functions, which allow a combinatorial description, provide solutions to (1.2). ###### Conjecture 1.4 ([44]; Conjecture 2.4). Let $Z(\Lambda,x)$ be the the Nekrasov partition function of five dimensional $SU(2)\times SU(2)$ gauge theory (See Definition 2.3). If we define a function $\Psi(\Lambda,x)$ by imposing the Higgsing condition on $Z(\Lambda,x)$, then it gives a solution $\mathsf{U}(\Lambda,x)=\Psi(\Lambda,x)$ to the equation (1.2). The time evolution of the discrete Painlevé VI equation is given by a translation by the element in $D_{5}^{(1)}$ root lattice, which is orthogonal to the symmetry $D_{4}^{(1)}$ of discrete $P_{\rm VI}$. If we write the translation element in terms of the generators of the extended affine Weyl groups, it is a product of two factors which are exchanged by the automorphism $\tau$ of $D_{5}^{(1)}$ Dynkin graph. The factorization of the original difference equation as the coupled system reflects this fact. The decomposition of the discrete time evolution (or the Hamiltonian) by the Bäcklund transformations implies that the original equation, which is the second order in the $q$-exponent of the Laplace operator $\mathcal{B}$, can be rewritten as a coupled system of the first order difference equations in $\mathcal{B}$, as was already suggested in the original paper [44]. ###### Proposition 1.5. The following coupled system is gauge equivalent to the non-stationary difference equation of [44] and hence, to the $qq$-Painlevé VI equation; $\displaystyle\mathsf{V}^{(1)}$ $\displaystyle={\Phi(qt^{-1}b_{2}/b_{4})\Phi(b_{1}/b_{3})\over\Phi(tb_{6}/b_{8})\Phi(qb_{5}/b_{7})}{1\over\varphi(-qb_{6}/G)\varphi(-G/b_{8})}$ $\displaystyle\qquad\cdot(\widetilde{\mathcal{B}}\cdot\widetilde{T}_{\mathsf{p},b}){1\over\varphi(-\mathsf{p}^{-1}qb_{2}/G)\varphi(-\mathsf{p}^{-1}G/b_{4})}\widetilde{T}_{\mathsf{p},b}\cdot\mathsf{V}^{(2)},$ $\displaystyle\widetilde{T}_{\mathsf{p},b}\cdot\mathsf{V}^{(2)}$ $\displaystyle={\Phi(\mathsf{p}^{-2}tb_{6}/b_{8})\Phi(\mathsf{p}^{-2}qb_{5}/b_{7})\over\Phi(\mathsf{p}^{-2}qb_{2}/b_{4})\Phi(\mathsf{p}^{-2}tb_{1}/b_{3})}{1\over\varphi(-\mathsf{p}^{-1}qb_{1}/G)\varphi(-\mathsf{p}^{-1}G/b_{3})}$ $\displaystyle\cdot(\widetilde{\mathcal{B}}\cdot\widetilde{T}_{\mathsf{p},b})\cdot{1\over\varphi(-qb_{5}/G)\varphi(-G/b_{7})}\mathsf{V}^{(1)}.$ To motivate an analogous conjecture to Conjecture 1.4, let us recall that the instanton counting with a surface defect allows another description in terms of the affine Laumon spaces [18],[3],[4]. In this method the partition functions are identified with the conformal blocks of the affine Kac-Moody algebra (the current algebra) without degenerate fields [30]. For example, in the present case the parameter $x$ which originally comes from the insertion point of the degenerate field is replaced by the $SU(2)$ spin variable of $\widehat{\mathfrak{sl}}_{2}$. The existence of the surface defect is taken into account by introducing the orbifold action [19],[27]. The relation of two methods for incorporating a surface defect is discussed in [20] from the viewpoint of integrable systems. In fact the role of the affine Kac-Moody algebra was already revealed in [12],[13], where the conjecture that the prepotential of the Seiberg-Witten theory is obtained from the leading term of the Nekrasov partition function was proved. We conjecture that the solutions to the coupled system are provided by the $K$-theoretic instanton partition function derived from the equivariant character of the affine Laumon space [18]. ###### Conjecture 1.6 (Conjecture 5.4). The partition function (5.17) gives a solution to the coupled system in Proposition 1.5 by the following specialization of parameters; $\displaystyle\mathcal{F}^{(1)}=f\left(\left.\left.\begin{array}[]{ll}q^{1/2}b_{4}/b_{8},&q^{1/2}b_{6}/b_{2}\\\ (\mathsf{p}^{2}Q)^{-1/2},&(\mathsf{p}^{2}Q)^{1/2}\\\ q^{-1/2}b_{2}/b_{5},&q^{-1/2}b_{7}/b_{4}\end{array}\right|q^{1/2}\mathsf{p}^{-1}\mathsf{t}G^{-1},q^{-1/2}\mathsf{p}^{-1}\mathsf{t}G\right|q,t^{-1/2}\right),$ $\displaystyle\mathcal{F}^{(2)}=f\left(\left.\left.\begin{array}[]{ll}q^{1/2}b_{4}/b_{8},&q^{1/2}b_{6}/b_{2}\\\ (\mathsf{p}^{2}Q)^{-1/2},&(\mathsf{p}^{2}Q)^{1/2}\\\ q^{-1/2}b_{1}/b_{6},&q^{-1/2}b_{8}/b_{3}\end{array}\right|-q^{1/2}\mathsf{t}G^{-1},-q^{-1/2}\mathsf{t}G\right|q,t^{-1/2}\right).$ The point here is that due to the symmetry of the translation element $T$ which defines the discrete time evolution of the $qq$-Painlevé VI equation, the pair of solutions $(\mathcal{F}^{(1)},\mathcal{F}^{(2)})$ comes from the common instanton partition function of the affine Laumon space with different specialization of parameters. An action of the quantum toroidal algebra of $A_{r}$ type on the equivariant cohomology group and the equivariant $K$ group of the affine Laumon space can be defined geometrically [35]. In four dimensional case (cohomological version) it has been shown that the instanton partition function satisfies the Knizhnik-Zamolodchikov (KZ) equation for the affine Kac-Moody algebra [38],[40]. Hence, the non-stationary difference equation should be derived as a KZ type equation for the quantum affine algebra $U_{q}(\widehat{\mathfrak{sl}}_{2})$ or more likely $U_{q}(\widehat{\mathfrak{gl}}_{2})$. Moreover, recalling that the affine Laumon space has elliptic cohomology, we believe it is an interesting problem to show our non-stationary difference equation can be generalized to elliptic case. The paper is organized as follows; In the next section we first summarize the non-stationary difference equation proposed in [44]. We make a gauge transformation to rewrite it in a form which is natural from the viewpoint of the $qq$-Painlevé VI equation. We also propose a dictionary between the variables on the gauge theory side and those on the Painlevé side. In section 3, following [23], we describe the quantization of the discrete Painlevé VI equation and explain the relation of the non-stationary difference equation and the $qq$-Painlevé VI equation. In section 4, we introduce the five dimensional quantum Seiberg-Witten curve. We show the quantum Seiberg-Witten allows a four dimensional limit and it is consistent with our previous result [4]. This is regarded as a good support for the conjecture in [44]. In section 5, we propose a coupled system which is gauge equivalent to the $qq$-Painlevé VI equation. Finally we conjecture the instanton partition functions of the affine Laumon space are solutions $(\mathcal{F}^{(1)},\mathcal{F}^{(2)})$ to the coupled system. As a consequence of the Dynkin automorphism symmetry of the factorization, $\mathcal{F}^{(1)}$ and $\mathcal{F}^{(2)}$ are obtained from a common instanton partition function with two kinds of specialization of parameters, which are related by the automorphism $\tau$. A summary of the discrete Painlevé VI equation is provided in Appendix A. Some of notations and conventions for the discrete Painlevé VI equation are fixed there. A few examples for supporting our conjecture in section 5 are presented in Appendix B. The four dimensinal limit for a factorized form of the Hamiltonian is discussed in Appendix C. ## 2\. Non-stationary difference equation ###### Definition 2.1. Let $T_{a,\Lambda}$ and $T_{b,x}$ be the shift operators acting on the variables $\Lambda$ and $x$ by $T_{a,\Lambda}f(\Lambda,x)\mathbin{:=}f(a\Lambda,x)$, $T_{b,x}f(\Lambda,x)\mathbin{:=}f(\Lambda,bx)$. Let $\vartheta_{x}\mathbin{:=}x\partial_{x}$ be the Euler operator in $x$. We have $\vartheta_{x}x=x(\vartheta_{x}+1)$, indicating that $p\mathbin{:=}q^{\vartheta_{x}}$ acts as the $q$-shift operator $q^{\vartheta_{x}}=T_{q,x}$. ###### Definition 2.2. Set $\mathcal{B}\mathbin{:=}q^{\vartheta_{x}(\vartheta_{x}+1)/2}$. We define the action of $\mathcal{B}$ on a formal Laurent series in $x$ as the $q$-Borel transformation: $\mathcal{B}(\sum_{n}c_{n}x^{n})=\sum_{n}q^{n(n+1)/2}c_{n}x^{n}.$ (2.1) The fundamental relations among $x,~{}p=q^{\vartheta_{x}}$ and $\mathcal{B}=q^{\vartheta_{x}(\vartheta_{x}+1)/2}$ are $px=qxp,\qquad\mathcal{B}p=p\mathcal{B},\qquad\mathcal{B}x=px\mathcal{B}.$ (2.2) One can see the last relation by looking at the action on $x^{n}$. In fact, both sides give the same result; $q^{\frac{1}{2}(n+1)(n+2)}x^{n+1}$. The $q$-Borel transformation $\mathcal{B}$ (see [21] section 2 and references therein) plays a significant role in the non-stationary difference equation in [44]. It is convenient to introduce the notations $\varphi(x)\in\mathbb{Q}(q)[[x]]$ and $\Phi(x)\in\mathbb{Q}(q,t)[[x]]$ for the infinite products; $\displaystyle\varphi(x):=$ $\displaystyle(x;q)_{\infty}=\prod_{n=0}^{\infty}(1-q^{n}x)=\exp\left(-\sum_{n=1}^{\infty}{1\over n}{1\over 1-q^{n}}x^{n}\right),$ (2.3) $\displaystyle\Phi(x):=$ $\displaystyle(x;q,t)_{\infty}=\prod_{n,m=0}^{\infty}(1-q^{n}t^{m}x)=\exp\left(-\sum_{n=1}^{\infty}{1\over n}{1\over(1-q^{n})(1-t^{n})}x^{n}\right).$ (2.4) They satisfy $\frac{\Phi(x)}{\Phi(tx)}=\varphi(x),\qquad\Phi(ta\Lambda)^{-1}T_{t,\Lambda}^{-1}\Phi(ta\Lambda)=\varphi(a\Lambda)T_{t,\Lambda}^{-1}.$ (2.5) We also use the standard notation for the $q$-shifted factorial $\displaystyle(u;q)_{n}=\prod_{i=0}^{n-1}(1-uq^{i})\qquad(n\in\mathbb{Z}_{\geq 0}).$ See [22] for useful formulas for $(u;q)_{n}$. ### 2.1. Five point function with a degenerate field The correlation functions of the chiral primary fields $\Phi_{\Delta}(z)$ are the most fundamental objects in two dimensional conformal field theory with the energy-momentum tensor $T(z)$ (the generating currents of the Virasoro algebra). The BPZ equation describes the response of the correlation functions under the insertion of the descendant fields created by the action of the Virasoro algebra. The BPZ equation for the five point function on $\mathbb{P}^{1}$111The instanton partition function we are going to discuss is expanded in $x$ and $\Lambda/x$. Hence here we assume the radial ordering with $|x|<1$ and $|\Lambda/x|<1$. $\Psi_{\mathrm{CFT}}(\Lambda,x):=\langle\Phi_{\Delta_{4}}(\infty)\Phi_{\Delta_{3}}(1)\varphi(x)\Phi_{\Delta_{2}}(\Lambda)\Phi_{\Delta_{1}}(0)\rangle_{\mathbb{P}^{1}}$ (2.6) with a level two degenerate field $\varphi(x)$ at $x$ is the linear differential equation of the form $\left(\partial_{x}^{2}+a(\Lambda,x)\partial_{x}+b(\Lambda,x)+c(\Lambda,x)\partial_{\Lambda}\right)\Psi_{\mathrm{CFT}}(\Lambda,x)=0,$ (2.7) which has regular singularities at $\\{0,\Lambda,x,1\\}$ and hence, is identified with the non-stationary Heun equation. We will reserve $t$ for one of the equivariant parameters ($\Omega$ background) of the torus action on $\mathbb{C}^{2}$ and $\Lambda$ plays the role of “time” variable in (2.7). In the non-stationary case, the constant part of the Heun operator involves the time derivative $\partial_{\Lambda}$. Up to the gauge transformation with the factor $x^{\alpha}(x-1)^{\beta}(x-\Lambda)^{\gamma}$ (2.8) the equation (2.7) agrees with the quantization of the Painlevé VI equation. The BPZ equation is also obtained from the deformed Seiberg-Witten curve of four dimensional supersymmetric gauge theory in the Nekrasov-Shatashvili limit as the non-stationary Schrödinger equation [14],[42]. This is also regarded as the quantization of (continuous, additive) Painlevé VI equation [4]. What we are going to discuss in this paper is an uplift of these stories to the triality of the deformed Virasoro algebra, discrete Painlevé equation and five dimensional supersymmetric gauge theories. Recall that in the AGT correspondence $(r+3)$ point conformal blocks on the genus zero curve are identified with the instanton partition functions of the linear quiver gauge theory of type $A_{r}$. Let us consider the five dimensional uplift of the AGT correspondence. . The instanton partition function is expressed in terms of the $K$-theoretic Nekrasov factor ${\mathsf{N}}_{\lambda,\mu}(u)={\mathsf{N}}_{\lambda,\mu}(u|q,\kappa)$ defined by $\displaystyle{\mathsf{N}}_{\lambda,\mu}(u|q,\kappa)$ $\displaystyle=\prod_{(i,j)\in\lambda}(1-uq^{\lambda_{i}-j}\kappa^{-\mu^{\prime}_{j}+i-1})\cdot\prod_{(k,l)\in\mu}(1-uq^{-\mu_{k}+l-1}\kappa^{\lambda^{\prime}_{l}-k}),$ (2.9) or equivalently $\displaystyle{\mathsf{N}}_{\lambda,\mu}(u|q,\kappa)=\prod_{j\geq i\geq 1}(uq^{-\mu_{i}+\lambda_{j+1}}\kappa^{-i+j};q)_{\lambda_{j}-\lambda_{j+1}}\cdot\prod_{\beta\geq\alpha\geq 1}(uq^{\lambda_{\alpha}-\mu_{\beta}}\kappa^{\alpha-\beta-1};q)_{\mu_{\beta}-\mu_{\beta+1}}.$ Here $q$ and $\kappa$ are regarded independent indeterminates. The Nekrasov factor ${\mathsf{N}}_{\lambda,\mu}(u)$ depends on a pair of partitions $(\lambda,\mu)$, namely $\lambda=(\lambda_{1},\lambda_{2},\ldots)$ is non- increasing non-negative integers with finitely many positive parts. $\lambda^{\prime}$ denotes the conjugate of $\lambda$. In [44] the instanton partition function of five dimensional $SU(2)\times SU(2)$ theory with four fundamental matter multiplets and one bi-fundamental matter multiplet is considered. On the deformed conformal block side this corresponds to the five point function on $\mathbb{P}^{1}$. ###### Definition 2.3. $\displaystyle\mathcal{Z}(\Lambda,x)$ $\displaystyle:=\sum_{\nu_{1},\nu_{2},\mu_{1},\mu_{2}\in\mathsf{P}}\mathfrak{p}_{1}^{|\nu_{1}|+|\nu_{2}|}\mathfrak{p}_{2}^{|\mu_{1}|+|\mu_{2}|}$ (2.10) $\displaystyle\cdot\prod_{1\leq a,b\leq 2}{{\mathsf{N}}_{\emptyset,\nu_{b}}(v\mathfrak{f}_{a}^{+}/\mathfrak{n}_{b}|q,t^{-1}){\mathsf{N}}_{\nu_{a},\mu_{b}}(w\mathfrak{n}_{a}/\mathfrak{m}_{b}|q,t^{-1}){\mathsf{N}}_{\mu_{a},\emptyset}(v\mathfrak{m}_{a}/\mathfrak{f}^{-}_{b}|q,t^{-1})\over{\mathsf{N}}_{\nu_{a},\nu_{b}}(\mathfrak{n}_{a}/\mathfrak{n}_{b}|q,t^{-1}){\mathsf{N}}_{\mu_{a},\mu_{b}}(\mathfrak{m}_{a}/\mathfrak{m}_{b}|q,t^{-1})},$ where ${\mathsf{P}}$ denotes the set of all partitions and $\emptyset$ is the empty partition. The following parametrization was used in [44]. $\displaystyle v=q^{1/2}t^{-1/2},\qquad w=v\phi_{1},\qquad\mathfrak{p}_{1}=v^{-2}T_{2}\phi_{2}x,\qquad\mathfrak{p}_{2}=v^{-2}{T_{4}\Lambda\over\phi_{1}x},$ $\displaystyle\mathfrak{n}_{1}=1,\qquad\mathfrak{n}_{2}=Q,\qquad\mathfrak{m}_{1}=1,\qquad\mathfrak{m}_{2}=\phi_{1}\phi_{2}Q,$ (2.11) $\displaystyle\mathfrak{f}_{1}^{+}=T_{1}Q,\qquad\mathfrak{f}_{2}^{+}=T_{2}^{-1},\qquad\mathfrak{f}_{1}^{-}=T_{3}^{-1},\qquad\mathfrak{f}_{2}^{-}=T_{4}\phi_{1}\phi_{2}Q.$ The coefficients of the expansion depend on parameters $(Q,\phi_{1},\phi_{2},T_{1},\cdots,T_{4})$ and the equivariant parameters $(q,t)$ of the torus action on $\mathbb{C}^{2}$. The parameters $(Q,\phi_{1},\phi_{2})$ correspond to the equivariant parameters of the Cartan subalgebra $U(1)\times U(1)\subset SU(2)\times SU(2)$ of the gauge group and the mass of the bi-fundamental matter222If we extend the gauge group to $U(2)\times U(2)$, the (exponentiated) mass parameter of the bi-fundamental matter may be identified with the equivariant parameter of the relative $U(1)$ factor of the gauge group $U(2)\times U(2)$. Note that the diagonal $U(1)$ factor decouples.. On the other hand the mass parameters $M_{i}$ of the fundamental hypermultiplets are related to $\log T_{i}$ up to the appropriate shifts of $\log q=\epsilon_{1},~{}\log t=-\epsilon_{2},~{}\log Q=-2a$. The instanton partition function $\mathcal{Z}(\Lambda,x)$ is a formal power series in $(x,\Lambda/x)$, where they are related to the insertion points of the intertwiners up the $SL(2,\mathbb{R})$ transformation. Let us consider the degenerate conformal block with the insertion of a level two degenerate field. Then one of the external Liouville momentum has a special value and the degenerate fusion rule tells that there are two allowed values for the intermediate momentum. According to the AGT correspondence this imposes two conditions on parameters $(Q,\phi_{1},\phi_{2})$ on the quiver gauge theory side, which is often referred to the Higgsing condition. In the present case the conditions are explicitly $\phi_{1}=q^{-1/2}t^{3/2},\phi_{2}=q^{1/2}t^{-1/2}$ (or $\phi_{1}=q^{-1/2}t^{1/2},\phi_{2}=q^{3/2}t^{-1/2}$). Recall that there are two possibilities for the intermediate momentum. As a consequence, one of the equivariant parameters, say $t$, is transmuted to the Higgsing mass parameter. Later we will see $t^{1/4}=\mathsf{p}$ becomes a basic shift parameter, or the non-autonomous parameter ($t\to 1$ is the autonomous limit) on the Painlevé side. In the five brane web realization of the surface defect, the parameter $t$ is identified with the volume of $S^{3}$ connecting $\mathrm{NS}5$ brane and $\mathrm{D}5$ brane which are non-intersecting. Note that $\mathrm{D}3$ brane can wrap $S^{3}$ in type IIB string theory. After imposing the Higgsing condition the instanton partition function $\Psi(\Lambda,x)=\mathcal{Z}(\Lambda,x)|_{\phi_{1}=q^{-1/2}t^{3/2},\phi_{2}=q^{1/2}t^{-1/2}}$ (2.12) has parameters $(Q;T_{1},\cdots,T_{4};\Lambda,x;q,t)$. Physically this is the instanton partition function of the $SU(2)$ gauge theory with a surface defect333In general the gauge theory allows several types of the surface defect according to the breaking patters of the total gauge group $SU(N)$ which are labelled by partitions of $N$. But for $SU(2)$ the breaking pattern is unique $SU(2)\to U(1)$. Mathematically the breaking pattern defines a parabolic structure along a defect (or a divisor).. The parameter $x$ is the insertion point of the degenerate field. Then, the conjecture in [44] says ###### Conjecture 2.4 ([44]). $\displaystyle\Psi(t\Lambda,x)=\mathcal{A}_{1}(\Lambda,x)\cdot\mathcal{B}\cdot\mathcal{A}_{2}(\Lambda,x)\cdot\mathcal{B}\cdot\mathcal{A}_{3}(\Lambda,x)\Psi(\Lambda,{x\over tqQ}),$ (2.13) where $\displaystyle\mathcal{A}_{1}(\Lambda,x)={1\over\varphi(T_{1}tvx)}{\Phi(T_{3}t^{2}v\Lambda x^{-1})\over\Phi(T_{3}qv\Lambda x^{-1})}{\Phi(T_{4}t^{2}v\Lambda x^{-1})\over\Phi(T_{4}t^{2}v^{-1}\Lambda x^{-1})},$ $\displaystyle\mathcal{A}_{2}(\Lambda,x)={\varphi(qT_{2}T_{3}\Lambda)\varphi(tT_{1}T_{4}\Lambda)\over\varphi(-T_{1}T_{2}x)\varphi(-Q^{-1}x)\varphi(-T_{3}T_{4}Qqt\Lambda x^{-1})\varphi(-q\Lambda x^{-1})},$ $\displaystyle\mathcal{A}_{3}(\Lambda,x)={1\over\varphi(T_{2}Q^{-1}q^{-1}vx)}{\Phi(T_{3}Qq^{2}v\Lambda x^{-1})\over\Phi(T_{3}Qq^{2}v^{-1}\Lambda x^{-1})}{\Phi(T_{4}Qt^{3}v\Lambda x^{-1})\over\Phi(T_{4}Qq^{2}v^{-1}\Lambda x^{-1})},$ with $v:=q^{1/2}t^{-1/2}$ and $\mathcal{B}$ is the $q$-Borel transformation. In [44] several evidences for the conjecture have been provided. For example it was proved for a special choice of mass parameters (external Liouville momenta) $(T_{1},T_{2},T_{3},T_{4})=(vt^{-1},v^{-1},v^{-1},vt^{-1})$444Recall that the Higgsing condition is $(\phi_{1},\phi_{2})=(tv^{-1},v)$., where the solution is expressed in terms of the Macdonald polynomials. In this paper we will investigate the structure of the difference equation (2.13), which is independent of the validity of Conjecture 2.4. Hence, let us replace $\Psi(\Lambda,x)$, which was defined by the Nekrasov partition function, with a generic function $\mathsf{U}(\Lambda,x)$ and write the equation as follows; $\displaystyle\mathsf{U}(t\Lambda,x)=\mathcal{A}_{1}(\Lambda,x)\cdot\mathcal{B}\cdot\mathcal{A}_{2}(\Lambda,x)\cdot\mathcal{B}\cdot\mathcal{A}_{3}(\Lambda,x)\mathsf{U}(\Lambda,{x\over tqQ}).$ (2.14) The difference equation (2.14) is called non-stationary, since there appears the shift $\Lambda\to t\Lambda$ on the left hand side. It was pointed out [44] that in the mass decoupling limit $T_{1},T_{2},T_{3},T_{4}\to 0$, the difference equation (2.14) reduces to the non-stationary relativistic affine Toda equation [45]. ###### Remark 2.5. In the mass decoupling limit the difference equation (2.14)degenerates to $\mathsf{U}_{\mathrm{Toda}}(t\Lambda,x)=\widehat{H}\mathsf{U}_{\mathrm{Toda}}(\Lambda,x/tqQ),\qquad\widehat{H}=\mathcal{B}\cdot\frac{1}{\varphi(-Q^{-1}x)\varphi(-q\Lambda x^{-1})}\cdot\mathcal{B}.$ (2.15) The name ‘Toda’ comes from the following fact. ###### Proposition 2.6. The operator $\widehat{H}$ commutes with the quantum Hamiltonian of the two- particle relativistic affine Toda system; $\hat{H}_{\mathrm{Toda}}=T_{q,x}+tQT_{q,x}^{-1}+tx+\Lambda x^{-1}.$ (2.16) One can confirm the existence and uniqueness of the solution to the non- stationary difference equation (2.14): ###### Proposition 2.7. The equation (2.14) has a formal series solution of the form $\mathsf{U}(\Lambda,x)=\sum_{i,j=0}^{\infty}c_{i,j}x^{i}(\Lambda/x)^{j},$ (2.17) and it is unique up to a normalization. ###### Proof. It is easy to see that the operator $T_{t,\Lambda}-{\mathcal{A}}_{1}\cdot\mathcal{B}\cdot{\mathcal{A}}_{2}\cdot\mathcal{B}\cdot{\mathcal{A}}_{3}T_{tqQ,x}^{-1}$ is “triangular” in the sense that it sends the monomials $x^{i}(\Lambda/x)^{j}$ to linear combinations of $x^{i+m}(\Lambda/x)^{j+n}$ $(m,n\geq 0)$. Moreover the leading coefficient with $(m,n)=(0,0)$ is $t^{j}-q^{(i-j)(i-j+1)}(tqQ)^{j-i}$, which is non vanishing for generic parameters $t,q$ and $Q$. Hence the coefficients $c_{i,j}$ are uniquely solved order by order with respect to $i+j$, once the initial value $c_{0,0}$ is fixed. ∎ We can eliminate $\Phi$-factors (the double infinite products) completely from the non-stationary difference equation (2.14), by the the following gauge transformation; $\mathsf{u}(\Lambda,x)=\Phi(qtT_{2}T_{3}\Lambda)\Phi(t^{2}T_{1}T_{4}\Lambda)\mathcal{A}_{3}(t\Lambda,tqQx)\mathsf{U}(t\Lambda,x).$ (2.18) Using the relations $\displaystyle\mathcal{A}_{3}(t\Lambda,tqQx)\mathcal{A}_{1}(\Lambda,x)={1\over\varphi(T_{1}vtx)\varphi(T_{2}vtx)\varphi(T_{3}vt\Lambda x^{-1})\varphi(T_{4}vt\Lambda x^{-1})},$ $\displaystyle{\Phi(qT_{2}T_{3}\Lambda)\Phi(tT_{1}T_{4}\Lambda)\over\Phi(qtT_{2}T_{3}\Lambda)\Phi(t^{2}T_{1}T_{4}\Lambda)}=\varphi(qT_{2}T_{3}\Lambda)\varphi(tT_{1}T_{4}\Lambda),$ which follow from (2.5), we obtain ###### Proposition 2.8. The difference equation (2.14) is gauge equivalent to $\displaystyle H\mathsf{u}(\Lambda,x)=\mathsf{u}(\Lambda,x),$ (2.19) with the Hamiltonian $\displaystyle H=$ $\displaystyle{1\over\varphi(T_{1}vtx)\varphi(T_{2}vtx)\varphi(T_{3}vt\Lambda x^{-1})\varphi(T_{4}vt\Lambda x^{-1})}\cdot\mathcal{B}$ $\displaystyle\cdot{1\over\varphi(-T_{1}T_{2}x)\varphi(-Q^{-1}x)\varphi(-T_{3}T_{4}Qqt\Lambda x^{-1})\varphi(-q\Lambda x^{-1})}\cdot\mathcal{B}\cdot T_{qtQ,x}^{-1}T_{t,\Lambda}^{-1}.$ (2.20) ### 2.2. Dictionary of five dimensional gauge/Painlevé correspondence In this paper we will employ the following relations of the parameters in [44] and the root variables for the discrete Painlevé VI equation (see Appendix A); $\displaystyle a_{0}=(T_{4}/T_{3})^{1/4},\quad a_{1}=(T_{1}/T_{2})^{1/4},\quad a_{2}=(t\Lambda T_{1}T_{4})^{-1/4},$ $\displaystyle a_{3}=(t^{2}\Lambda T_{3}T_{4})^{1/4},\quad a_{4}=(QT_{1}T_{2})^{1/4},\quad a_{5}=(tQT_{3}T_{4})^{-1/4}.$ (2.21) Namely $\sqrt{Q}T_{1}={a_{1}^{2}a_{4}^{2}},\quad\sqrt{Q}T_{2}={a_{4}^{2}\over a_{1}^{2}},\quad\sqrt{tQ}T_{3}={1\over a_{0}^{2}a_{5}^{2}},\quad\sqrt{tQ}T_{4}={a_{0}^{2}\over a_{5}^{2}},$ (2.22) and $\displaystyle\mathsf{p}:=e^{\delta}=a_{0}a_{1}a_{2}^{2}a_{3}^{2}a_{4}a_{5}=t^{1/4},\quad\mathsf{t}:=a_{3}^{2}a_{4}a_{5}=t^{1/2}\Lambda(T_{1}T_{2}T_{3}T_{4})^{1/4},$ $\displaystyle{t\Lambda\over Q}=a_{3}^{4}a_{5}^{4}.$ (2.23) We assume that the parameter $Q$ is invariant under the action of the extended affine Weyl group. In [23], $\mathsf{t}$ is identified with the “time” variable. The parameter $\mathsf{p}$ defines the shift parameter of the discrete time evolution. On the gauge theory side $\Lambda$ plays the role of the corresponding time variable. The dictionary (2.21) tells that the exchange of $T_{1}$ and $T_{2}$ corresponds to the Weyl reflection $r_{1}:(a_{1},a_{2})\to(a_{1}^{-1},a_{1}a_{2})$. Similarly the exchange of $T_{3}$ and $T_{4}$ gives the action of $r_{0}:(a_{0},a_{2})\to(a_{0}^{-1},a_{0}a_{2})$ on the Painlevé side. We have to accept that (2.21) is not invertible. This is due to the fact that the root variables $a_{i}$ and the parameters $(T_{i},\Lambda,Q)$ on the gauge theory side are two kinds of variables with constraints in the ambient ten dimensional Picard lattice of $\mathbb{P}^{1}\times\mathbb{P}^{1}$ with eight points blow-ups (or $\mathbb{P}^{2}$ with nine points blow-ups). The parameter $x$ is related to a pair of $q$-commuting dynamical variables $(F,G)$ as $\displaystyle G=-\xi\qquad\mbox{\rm({\it i.e.} as multiplication by $-\xi$)}$ (2.24) $\displaystyle F=(qt^{1/2}Q)^{-1/2}\xi q^{-\vartheta_{\xi}},$ (2.25) where $\xi=\left({q^{2}\Lambda^{2}T_{3}T_{4}\over T_{1}T_{2}}\right)^{1/4}x^{-1}$ (2.26) and $\vartheta_{\xi}=\xi\partial/\partial\xi$. Note that the dynamical variables $(F,G)$ satisfy $FG=q^{-1}GF.$ (2.27) Parameters | Higgsed quiver theory [44] | Painlevé VI ---|---|--- $T_{i}$ | (dressed) mass parameters | root variables of the outer nodes $Q$ | $SU(2)$ Coulomb modulus | parameter of solutions $\Lambda$ | instanton expansion parameter | root variable of the inner node $x$ | position of degenerate field insertion | dynamical variable (coordinate) $q$ | $\Omega$ background along the defect | quantization parameter $t$ | $\Omega$ background orthogonal to the defect | non-autonomous parameter Table 1. Dictionary between the quiver gauge theory and Painlevé VI equation By the above dictionary the Hamiltonian (2.8) in terms of the variables of Painlevé VI equation takes the following formula (for the definition of the variables $b_{i}$ see Definition A.5 in Appendix A); $\displaystyle H_{\mathrm{VI}}=$ $\displaystyle{1\over\varphi(-qb_{5}G^{-1})\varphi(-qb_{6}G^{-1})\varphi(-b_{7}^{-1}G)\varphi(-b_{8}^{-1}G)}\cdot\widetilde{\mathcal{B}}$ $\displaystyle\cdot{1\over\varphi(q\mathsf{p}^{-2}b_{1}G^{-1})\varphi(q\mathsf{p}^{-2}b_{2}G^{-1})\varphi(\mathsf{p}^{-2}b_{3}^{-1}G)\varphi(\mathsf{p}^{-2}b_{4}^{-1}G)}\cdot\widetilde{\mathcal{B}}\cdot T_{t^{1/2},x}^{-1}T_{t,\Lambda}^{-1}.$ (2.28) For later convenience, by writing $T_{qtQ,x}^{-1}=(T_{qt^{1/2}Q,x}^{-1/2})^{2}T_{t^{1/2},x}^{-1}$, we have distributed $T_{qt^{1/2}Q}^{-1/2}$ and combined it with $\mathcal{B}$ to define $\widetilde{\mathcal{B}}=\mathcal{B}\cdot T_{(qt^{1/2}Q)^{-1/2},x}=T_{(qt^{1/2}Q)^{-1/2},x}\cdot\mathcal{B}$. Since $G\sim-x^{-1}$, we have the $(qt^{1/2}Q)^{\pm 1/2}$-shift of the arguments of $\varphi$ involving $b_{1},\cdots,b_{4}$ after commuting them with $T_{qt^{1/2}Q,x}^{-1/2}$ in (2.2), compared with (2.8). In terms of the variables $b_{i}$, the symmetry of the Hamiltonian (2.2), exchanging $(b_{1},\cdots,b_{4})$ and $(b_{5},\cdots,b_{8})$, becomes more manifest555The additional factor $\mathsf{p}^{-2}=t^{-1/2}$ for $b_{1},\cdots,b_{4}$ should be related to the redefinition of $\mathsf{t}$ and $\mathsf{p}$ to be discussed in the next section (see (3.1)).. The symmetry is related to the diagram automorphism $\tau$ of $D_{5}^{(1)}$, see Fig.1 in Appendix A. Geometrically this is the exchange of $\mathbb{P}^{1}$ in the product $\mathbb{P}^{1}\times\mathbb{P}^{1}$ that appears in the description of the space of initial conditions, see Fig.2 in Appendix A. We will get back to this point in section 5. ## 3\. Quantized discrete Painlevé VI equation ### 3.1. Heisenberg form of discrete $P_{\mathrm{VI}}$ We recapitulate a formulation of the $qq$-Painlevé VI equation. Consider a linear operator ${\bf H}$ of the form ${\bf H}=A(x)\cdot\mathcal{B}\cdot B(x)\cdot\mathcal{B},$ (3.1) where $A(x),B(x)$ are some formal Laurent series in $x$. ###### Proposition 3.1. Putting $a(x)\mathbin{:=}A(qx)A(x)^{-1}$, $b(x)\mathbin{:=}B(qx)B(x)^{-1}$, we have ${\bf H}^{-1}x^{-1}{\bf H}=b(p^{-1}x)x^{-1}p^{2},\quad{\bf H}x^{-1}p{\bf H}^{-1}=x^{-1}p^{-1}a(x).$ (3.2) ###### Proof. Since $\mathcal{B}^{-1}x^{-1}\mathcal{B}=x^{-1}p$, $\mathcal{B}^{-1}x\mathcal{B}=p^{-1}x$, we have $\displaystyle{\bf H}^{-1}x^{-1}{\bf H}$ $\displaystyle=$ $\displaystyle\mathcal{B}^{-1}B(x)^{-1}\mathcal{B}^{-1}A(x)^{-1}x^{-1}A(x)\mathcal{B}B(x)\mathcal{B}$ $\displaystyle=$ $\displaystyle\mathcal{B}^{-1}B(x)^{-1}\mathcal{B}^{-1}x^{-1}\mathcal{B}B(x)\mathcal{B}$ $\displaystyle=$ $\displaystyle\mathcal{B}^{-1}B(x)^{-1}x^{-1}pB(x)\mathcal{B}=\mathcal{B}^{-1}B(x)^{-1}B(qx)x^{-1}\mathcal{B}p$ $\displaystyle=$ $\displaystyle\mathcal{B}^{-1}b(x)x^{-1}\mathcal{B}p=b(p^{-1}x)x^{-1}p^{2}.$ Similarly, using $\mathcal{B}x^{-1}p\mathcal{B}^{-1}=x^{-1}$. $\mathcal{B}x^{-1}\mathcal{B}^{-1}=x^{-1}p^{-1}$, we have $\displaystyle{\bf H}x^{-1}p{\bf H}^{-1}$ $\displaystyle=$ $\displaystyle A(x)\mathcal{B}B(x)\mathcal{B}x^{-1}p\mathcal{B}^{-1}B(x)^{-1}\mathcal{B}^{-1}A(x)^{-1}$ $\displaystyle=$ $\displaystyle A(x)\mathcal{B}B(x)x^{-1}B(x)^{-1}\mathcal{B}^{-1}A(x)^{-1}$ $\displaystyle=$ $\displaystyle A(x)\mathcal{B}x^{-1}\mathcal{B}^{-1}A(x)^{-1}=A(x)x^{-1}p^{-1}A(x)^{-1}$ $\displaystyle=$ $\displaystyle x^{-1}p^{-1}A(qx)A(x)^{-1}=x^{-1}p^{-1}a(x).$ Hence the desired equations are proved. ∎ We specify the functions $A(x),B(x)$ as $A(x)=\frac{1}{\varphi(\mathsf{a}_{1}x)\varphi(\mathsf{a}_{2}x)\varphi(\frac{q}{\mathsf{a}_{3}x})\varphi(\frac{q}{\mathsf{a}_{4}x})},\quad B(x)=\frac{1}{\varphi(-\mathsf{b}_{1}x)\varphi(-\mathsf{b}_{2}x)\varphi(-\frac{q}{\mathsf{b}_{3}x})\varphi(-\frac{q}{\mathsf{b}_{4}x})},$ (3.3) where the infinite products are considered as formal Laurent series in $x$ by the expansion $\frac{1}{\varphi(z)}=\frac{1}{(z;q)_{\infty}}=\sum_{n=0}^{\infty}\frac{z^{n}}{(q)_{n}}.$ (3.4) We note that the gauge transformed Hamiltonian (2.8) is reproduced by the choice of parameters; $\displaystyle\mathsf{a}_{1}=q^{1/2}t^{1/2}T_{1},\qquad\mathsf{a}_{2}=q^{1/2}t^{1/2}T_{2},$ $\displaystyle\mathsf{a}_{3}=q^{1/2}t^{-1/2}T_{3}^{-1}\Lambda^{-1},\qquad\mathsf{a}_{4}=q^{1/2}t^{-1/2}T_{4}^{-1}\Lambda^{-1},$ (3.5) $\displaystyle\mathsf{b}_{1}=T_{1}T_{2},\qquad\mathsf{b}_{2}=Q^{-1},$ $\displaystyle\mathsf{b}_{3}=t^{-1}Q^{-1}(T_{3}T_{4})^{-1}\Lambda^{-1},\qquad\mathsf{b}_{4}=\Lambda^{-1}.$ (3.6) ###### Corollary 3.2. We put $\tilde{F}\mathbin{:=}x^{-1}p,\quad\tilde{G}\mathbin{:=}-x^{-1}\quad{\rm i.e.}\quad x=-\tilde{G}^{-1},\quad p=\tilde{G}^{-1}\tilde{F},\quad\tilde{G}\tilde{F}=q\tilde{F}\tilde{G},$ (3.7) then we have ${\bf H}\tilde{F}{\bf H}^{-1}=\frac{\mathsf{a}_{3}\mathsf{a}_{4}}{q\tilde{F}}{\displaystyle\frac{(\tilde{G}+\mathsf{a}_{1})(\tilde{G}+\mathsf{a}_{2})}{(\tilde{G}+\mathsf{a}_{3})(\tilde{G}+\mathsf{a}_{4})}},\quad{\bf H}^{-1}\tilde{G}{\bf H}={\displaystyle\frac{(\tilde{F}+\mathsf{b}_{1})(\tilde{F}+\mathsf{b}_{2})}{(\tilde{F}+\mathsf{b}_{3})(\tilde{F}+\mathsf{b}_{4})}}\frac{\mathsf{b}_{3}\mathsf{b}_{4}}{q\tilde{G}}.$ (3.8) ###### Proof. For the function $A(x),B(x)$ in (3.3), we see $a(x)={\displaystyle\frac{(1-\mathsf{a}_{1}x)(1-\mathsf{a}_{2}x)}{(1-\mathsf{a}_{3}x)(1-\mathsf{a}_{4}x)}}\mathsf{a}_{3}\mathsf{a}_{4}x^{2},\quad b(x)={\displaystyle\frac{(1+\mathsf{b}_{1}x)(1+\mathsf{b}_{2}x)}{(1+\mathsf{b}_{3}x)(1+\mathsf{b}_{4}x)}}\mathsf{b}_{3}\mathsf{b}_{4}x^{2}.$ (3.9) Then from eq. (3.2), we have $\displaystyle{\bf H}\tilde{F}{\bf H}^{-1}$ $\displaystyle=$ $\displaystyle x^{-1}p^{-1}a(x)=\frac{\mathsf{a}_{3}\mathsf{a}_{4}}{q\tilde{F}}{\displaystyle\frac{(1+\mathsf{a}_{1}/\tilde{G})(1+\mathsf{a}_{2}/\tilde{G})}{(1+\mathsf{a}_{3}/\tilde{G})(1+\mathsf{a}_{4}/\tilde{G)}}},$ (3.10) $\displaystyle{\bf H}^{-1}\tilde{G}{\bf H}$ $\displaystyle=$ $\displaystyle b(p^{-1}x)x^{-1}p^{2}={\displaystyle\frac{(1-\mathsf{b}_{1}/\tilde{F})(1-\mathsf{b}_{2}/\tilde{F})}{(1-\mathsf{b}_{3}/\tilde{F})(1-\mathsf{b}_{4}/\tilde{F})}}\frac{\mathsf{b}_{3}\mathsf{b}_{4}}{q\tilde{G}},$ (3.11) as desired. ∎ The equation (3.8) can be viewed as the Heisenberg form of the $qq$-Painlevé VI equation. Now let us check by an appropriate rescaling $\displaystyle x^{-1}p=\alpha^{-1}F,\quad x^{-1}=-\beta^{-1}G,$ (3.12) the equation (3.8) agrees with the Heisenberg equations first presented in [23]. Combining the dictionary (3.1)– (3.6) with (3.9), we can rewrite (3.2) as follows; $\displaystyle{\bf H}^{-1}x^{-1}{\bf H}=$ $\displaystyle{(1+T_{1}T_{2}p^{-1}x)(1+Q^{-1}p^{-1}x)\over(1+T_{3}T_{4}Qt\Lambda x^{-1}p)(1+\Lambda x^{-1}p)}x^{-1}p^{2}$ $\displaystyle=$ $\displaystyle q^{-1}x\mathsf{b}_{3}\mathsf{b}_{4}{(x^{-1}p+q^{-1}\mathsf{b}_{1})(x^{-1}p+q^{-1}\mathsf{b}_{2})\over(x^{-1}p+q^{-1}\mathsf{b}_{3})(x^{-1}p+q^{-1}\mathsf{b}_{4})},$ $\displaystyle{\bf H}x^{-1}p{\bf H}^{-1}=$ $\displaystyle x^{-1}p^{-1}{(1-q^{1/2}t^{1/2}T_{1}x)(1-q^{1/2}t^{1/2}T_{2}x)\over(1-q^{-1/2}t^{1/2}T_{3}\Lambda x^{-1})(1-q^{-1/2}t^{1/2}T_{4}\Lambda x^{-1})}$ $\displaystyle=$ $\displaystyle q^{-1}\mathsf{a}_{3}\mathsf{a}_{4}\frac{(x^{-1}-q^{-1}\mathsf{a}_{1})(x^{-1}-q^{-1}\mathsf{a}_{2})}{(x^{-1}-q^{-1}\mathsf{a}_{3})(x^{-1}-q^{-1}\mathsf{a}_{4})}p^{-1}x.$ We see an agreement with (3.8), which confirms the consistency of the dictionary (3.1)– (3.6). Under the scaling (3.12), we have $\displaystyle{\bf H}F{\bf H}^{-1}$ $\displaystyle={\alpha^{2}\over T_{3}T_{4}\Lambda^{2}t}{(G+\mathsf{t}\mathsf{p}^{-1}a_{1}^{2})(G+\mathsf{t}\mathsf{p}^{-1}a_{1}^{-2})\over(G+\mathsf{t}^{-1}\mathsf{p}a_{0}^{2})(G+\mathsf{t}^{-1}\mathsf{p}a_{0}^{-2})}F^{-1},$ (3.13) $\displaystyle{\bf H}^{-1}G{\bf H}=$ $\displaystyle{\beta^{2}\over T_{3}T_{4}Qqt\Lambda^{2}}G^{-1}{(F+\mathsf{t}a_{4}^{2})(F+\mathsf{t}a_{4}^{-2})\over(F+\mathsf{t}^{-1}a_{5}^{2})(F+\mathsf{t}^{-1}a_{5}^{-2})},$ (3.14) where we have defined $\displaystyle\mathsf{t}\mathsf{p}^{-1}a_{1}^{2}=\beta q^{-1/2}t^{1/2}T_{1},\quad\mathsf{t}\mathsf{p}^{-1}a_{1}^{-2}=\beta q^{-1/2}t^{1/2}T_{2},$ $\displaystyle\mathsf{t}^{-1}\mathsf{p}a_{0}^{2}={\beta\over q^{1/2}t^{1/2}T_{3}\Lambda},\quad\mathsf{t}^{-1}\mathsf{p}a_{0}^{-2}={\beta\over q^{1/2}t^{1/2}T_{4}\Lambda},$ (3.15) and $\mathsf{t}a_{4}^{2}=\alpha q^{-1}T_{1}T_{2},\quad\mathsf{t}a_{4}^{-2}=\alpha q^{-1}Q^{-1},\quad\mathsf{t}^{-1}a_{5}^{2}={\alpha\over T_{3}T_{4}Qqt\Lambda},\quad\mathsf{t}^{-1}a_{5}^{-2}={\alpha\over q\Lambda}.$ (3.16) From (3.1) and (3.16) we find $\displaystyle\beta^{4}{T_{1}T_{2}\over T_{3}T_{4}\Lambda^{2}q^{2}}=1,\quad\alpha^{4}{T_{1}T_{2}\over T_{3}T_{4}Q^{2}q^{4}t\Lambda^{2}}=1,$ which implies $\alpha^{2}=(qt^{1/2}Q)\cdot\beta^{2}.$ (3.17) Hence, we can eliminate $\alpha$ and $\beta$ from (3.13) and (3.14) as follows; $\displaystyle{\bf H}F{\bf H}^{-1}=(qt^{1/2}Q)\times q\mathsf{p}^{2}\mathsf{t}^{-2}{(G+\mathsf{t}\mathsf{p}^{-1}a_{1}^{2})(G+\mathsf{t}\mathsf{p}^{-1}a_{1}^{-2})\over(G+\mathsf{t}^{-1}\mathsf{p}a_{0}^{2})(G+\mathsf{t}^{-1}\mathsf{p}a_{0}^{-2})}F^{-1},$ $\displaystyle{\bf H}^{-1}G{\bf H}=(qt^{1/2}Q)^{-1}\times q\mathsf{t}^{-2}G^{-1}{(F+\mathsf{t}a_{4}^{2})(F+\mathsf{t}a_{4}^{-2})\over(F+\mathsf{t}^{-1}a_{5}^{2})(F+\mathsf{t}^{-1}a_{5}^{-2})}.$ The final step to a complete matching with [23],[24] requires a further redefinition $\displaystyle\widetilde{F}=(qt^{1/2}Q)^{-1}F,\qquad\widetilde{G}=G,$ $\displaystyle\widetilde{\mathsf{t}}=t^{1/2}\mathsf{t},\qquad\widetilde{\mathsf{p}}=t^{1/2}\mathsf{p},\qquad\widetilde{\mathsf{p}}^{4}=t.$ (3.18) Then we have ###### Proposition 3.3. $\displaystyle{\bf H}\widetilde{F}{\bf H}^{-1}=q\widetilde{\mathsf{p}}^{2}\widetilde{\mathsf{t}}^{-2}{(\widetilde{G}+\widetilde{\mathsf{t}}\widetilde{\mathsf{p}}^{-1}a_{1}^{2})(\widetilde{G}+\widetilde{\mathsf{t}}\widetilde{\mathsf{p}}^{-1}a_{1}^{-2})\over(\widetilde{G}+\widetilde{\mathsf{t}}^{-1}\widetilde{\mathsf{p}}a_{0}^{2})(\widetilde{G}+\widetilde{\mathsf{t}}^{-1}\widetilde{\mathsf{p}}a_{0}^{-2})}\widetilde{F}^{-1},$ (3.19) $\displaystyle{\bf H}^{-1}\widetilde{G}{\bf H}=q\widetilde{\mathsf{t}}^{-2}\widetilde{G}^{-1}{(\widetilde{F}+\widetilde{\mathsf{t}}a_{4}^{2})(\widetilde{F}+\widetilde{\mathsf{t}}a_{4}^{-2})\over(\widetilde{F}+\widetilde{\mathsf{t}}^{-1}a_{5}^{2})(\widetilde{F}+\widetilde{\mathsf{t}}^{-1}a_{5}^{-2})}.$ (3.20) Summarizing all the above definitions and rescalings, we arrive at the dictionary presented in section 2.2. ### 3.2. Coxeter relations for quantum variables To formulate the quantization of the discrete $P_{\rm VI}$ Equation reviewed in Appendix A, we first quantize the commutative canonical pair of variables $f,g$. Let $F$ and $G$ be non commutative variables satisfying the $q$-commutation relation666$F$ and $G$ are exponentiated variables of $f$ and $g$. $\displaystyle FG=q^{-1}GF.$ (3.21) Recall the notation $\mathbb{K}=\mathbb{C}(a)=\mathbb{C}(a_{0},\ldots,a_{5})$ for the rational function field in the root variables $a_{i}$. Let $\mathbb{K}\langle F,G\rangle$ be the $\mathbb{K}$-algebra generated by $F,G$ with the relation (3.21). It is known that $\mathbb{K}\langle F,G\rangle$ is an Ore domain (see [31] section 2 and references therein). Denote by $\mathbb{F}=\mathbb{K}(F,G)$ the quotient skew field of $\mathbb{K}\langle F,G\rangle$. Note that $\mathbb{F}$ is generated by $a_{0},\ldots,a_{5},F$ and $G$. For any formal power series $h(z)$ in $z$, we use the formula; ${\rm Ad}(F)\cdot h(G)=h(q^{-1}G),\qquad{\rm Ad}(G)\cdot h(F)=h(qF),$ (3.22) in our computations. As is summarized in Appendix A, the time evolution of the discrete Painlevé VI equation is derived from the translation element $T$ in the extended affine Weyl group $\widetilde{W}$. Hence, we need an action of $\widetilde{W}$ on the quantum pair of dynamical variables $(F,G)$. ###### Definition 3.4. Define the actions of $r_{0},\ldots,r_{5},\sigma_{01},\sigma_{45},\tau\in\widetilde{W}$ on the generators of $\mathbb{F}$ by the rules: $\displaystyle r_{0}$ $\displaystyle:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\mapsto(a_{0}^{-1},a_{1},a_{0}a_{2},a_{3},a_{4},a_{5},F,G),$ $\displaystyle r_{1}$ $\displaystyle:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\mapsto(a_{0},a_{1}^{-1},a_{1}a_{2},a_{3},a_{4},a_{5},F,G),$ $\displaystyle r_{2}$ $\displaystyle:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\mapsto(a_{0}a_{2},a_{1}a_{2},a_{2}^{-1},a_{2}a_{3},a_{4},a_{5},F{a_{0}a_{1}^{-1}G+a_{2}^{2}\over a_{0}a_{1}^{-1}a_{2}^{2}G+1},G),$ $\displaystyle r_{3}$ $\displaystyle:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\mapsto(a_{0},a_{1},a_{2}a_{3},a_{3}^{-1},a_{3}a_{4},a_{3}a_{5},F,{a_{3}^{2}a_{4}a_{5}^{-1}F+1\over a_{4}a_{5}^{-1}F+a_{3}^{2}}G),$ $\displaystyle r_{4}$ $\displaystyle:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\mapsto(a_{0},a_{1},a_{2},a_{3}a_{4},a_{4}^{-1},a_{5},F,G),$ $\displaystyle r_{5}$ $\displaystyle:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\mapsto(a_{0},a_{1},a_{2},a_{3}a_{5},a_{4},a_{5}^{-1},F,G),$ $\displaystyle\sigma_{01}$ $\displaystyle:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\mapsto(a_{1}^{-1},a_{0}^{-1},a_{2}^{-1},a_{3}^{-1},a_{4}^{-1},a_{5}^{-1},qF^{-1},G),$ $\displaystyle\sigma_{45}$ $\displaystyle:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\mapsto(a_{0}^{-1},a_{1}^{-1},a_{2}^{-1},a_{3}^{-1},a_{5}^{-1},a_{4}^{-1},F,qG^{-1}),$ $\displaystyle\tau$ $\displaystyle:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\mapsto(a_{5}^{-1},a_{4}^{-1},a_{3}^{-1},a_{2}^{-1},a_{1}^{-1},a_{0}^{-1},G,F).$ Then extend them to the actions on $\mathbb{F}$ by the requirements: (1) $r_{0},\ldots,r_{5}$ are the ring homomorphisms, (2) $\sigma_{01},\sigma_{45},\tau$ are the ring anti-homomorphisms. Note that we have defined $\sigma_{01},\sigma_{45},\tau$ as ring anti- homomorphisms. Compare it with Definition A.1 in Appendix A, where they are ring homomorphisms. ###### Proposition 3.5. These actions on $\mathbb{F}$ are compatible with the group structure of the extended affine Weyl group $\widetilde{W}$. Namely they satisfy the Coxeter relations given in Proposition A.2. One can check the Coxeter relations by straightforward calculations. We only present two nontrivial cases. ###### Lemma 3.6. We have $\displaystyle r_{3}(F)=F,\qquad r_{2}r_{3}r_{2}(F)=r_{3}r_{2}(F),$ $\displaystyle r_{2}(G)=G,\qquad r_{3}r_{2}r_{3}(G)=r_{2}r_{3}(G).$ Hence we have the Coxeter relations $\displaystyle r_{2}r_{3}r_{2}(F)=r_{3}r_{2}r_{3}(F),\qquad r_{2}r_{3}r_{2}(G)=r_{3}r_{2}r_{3}(G).$ ###### Proof. We have777The variables $b_{i}$ are defined in Appendix A (see Definition A.5). $\displaystyle r_{2}(F)=\sqrt{b_{5}\over b_{7}}F{G+b_{7}\over G+b_{5}},$ $\displaystyle r_{3}r_{2}(F)=\sqrt{b_{3}b_{5}\over b_{1}b_{7}}F{{F+b_{3}\over F+b_{1}}G+b_{7}\over{F+b_{3}\over F+b_{1}}G+b_{1}^{-1}b_{3}b_{5}},$ $\displaystyle r_{2}r_{3}r_{2}(F)=\sqrt{b_{3}b_{5}\over b_{1}b_{7}}F{G+b_{7}\over G+b_{5}}{A\over B},$ where $\displaystyle A={F{G+b_{7}\over G+b_{5}}+b_{3}\over F{G+b_{7}\over G+b_{5}}+b_{1}b_{5}^{-1}b_{7}}G+b_{5},\qquad B={F{G+b_{7}\over G+b_{5}}+b_{3}\over F{G+b_{7}\over G+b_{5}}+b_{1}b_{5}^{-1}b_{7}}G+b_{1}^{-1}b_{3}b_{5}.$ Note that we have $AB=BA$. Using $\displaystyle{F+b_{3}\over F+b_{1}}G+b_{7}={1\over F+b_{1}}(F(G+b_{7})+b_{3}(G+b_{1}b_{3}^{-1}b_{7})),$ $\displaystyle{F+b_{3}\over F+b_{1}}G+b_{1}^{-1}b_{3}b_{5}={1\over F+b_{1}}(F(G+b_{1}^{-1}b_{3}b_{5})+b_{3}(G+b_{5})),$ we obtain $\displaystyle r_{3}r_{2}(F)=\sqrt{b_{3}b_{5}\over b_{1}b_{7}}F{1\over F(G+b_{1}^{-1}b_{3}b_{5})+b_{3}(G+b_{5})}(F(G+b_{7})+b_{3}(G+b_{1}b_{3}^{-1}b_{7})).$ On the other hand, from $\displaystyle A=C^{-1}(F(G+b_{7})+b_{3}(G+b_{1}b_{3}^{-1}b_{7})),$ $\displaystyle B=C^{-1}(F(G+b_{1}^{-1}b_{3}b_{5})+b_{3}(G+b_{5})){G+b_{7}\over G+b_{5}},$ $\displaystyle C=F{G+b_{7}\over G+b_{5}}+b_{1}b_{5}^{-1}b_{7},$ we have $\displaystyle r_{2}r_{3}r_{2}(F)=\sqrt{b_{3}b_{5}\over b_{1}b_{7}}F{1\over F(G+b_{1}^{-1}b_{3}b_{5})+b_{3}(G+b_{5})}(F(G+b_{7})+b_{3}(G+b_{1}b_{3}^{-1}b_{7})),$ indicating that we have $r_{2}r_{3}r_{2}(F)=r_{3}r_{2}(F)$. ∎ ###### Lemma 3.7. We have $\displaystyle\sigma_{01}r_{2}(F,G)=\sigma_{01}(F{a_{0}a_{1}^{-1}G+a_{2}^{2}\over a_{0}a_{1}^{-1}a_{2}^{2}G+1},G)=({a_{0}a_{1}^{-1}G+a_{2}^{-2}\over a_{0}a_{1}^{-1}a_{2}^{-2}G+1}qF^{-1},G),$ $\displaystyle r_{2}\sigma_{01}(F,G)=r_{2}(qF^{-1},G)=(q{a_{0}a_{1}^{-1}a_{2}^{2}G+1\over a_{0}a_{1}^{-1}G+a_{2}^{2}}F^{-1},G),$ and $\displaystyle\sigma_{01}r_{3}(F,G)=\sigma_{01}(F,{a_{3}^{2}a_{4}a_{5}^{-1}F+1\over a_{4}a_{5}^{-1}F+a_{3}^{2}}G)=(qF^{-1},G{a_{3}^{-2}a_{4}^{-1}a_{5}qF^{-1}+1\over a_{4}^{-1}a_{5}qF^{-1}+a_{3}^{-2}}),$ $\displaystyle r_{3}\sigma_{01}(F,G)=r_{3}(qF^{-1},G)=(qF^{-1},{a_{3}^{2}a_{4}a_{5}^{-1}F+1\over a_{4}a_{5}^{-1}F+a_{3}^{2}}G).$ indicating that we have $\sigma_{01}r_{2}(F,G)=r_{2}\sigma_{01}(F,G),\sigma_{01}r_{3}(F,G)=r_{3}\sigma_{01}(F,G)$. ###### Proposition 3.8. Let $T=r_{2}r_{1}r_{0}r_{2}\sigma_{01}r_{3}r_{4}r_{5}r_{3}\sigma_{45}$ be the translation element in $\widetilde{W}$. Writing $\overline{F}=T\cdot F,~{}\underline{G}=T^{-1}\cdot G$ for short, we have the $qq$-Painlevé VI equation; $\displaystyle\overline{F}F$ $\displaystyle=q\mathsf{p}^{2}\mathsf{t}^{-2}{G+\mathsf{t}\mathsf{p}^{-1}a_{1}^{2}\over G+\mathsf{t}^{-1}\mathsf{p}a_{0}^{2}}{G+\mathsf{t}\mathsf{p}^{-1}a_{1}^{-2}\over G+\mathsf{t}^{-1}\mathsf{p}a_{0}^{-2}},$ (3.23) $\displaystyle G\underline{G}$ $\displaystyle=q\mathsf{t}^{-2}{F+\mathsf{t}a_{4}^{2}\over F+\mathsf{t}^{-1}a_{5}^{2}}{F+\mathsf{t}a_{4}^{-2}\over F+\mathsf{t}^{-1}a_{5}^{-2}},$ (3.24) where the discrete shift of the time variable is $T(\mathsf{t})=\mathsf{p}^{-2}\mathsf{t}$. Note that compared with the classical version, there appears the factor $q$ on the right hand side. ###### Proof. We can compute the action of the reflections and the diagram automorphisms as follows; $\displaystyle F\mathop{\longmapsto}^{\sigma_{45}}F\mathop{\longmapsto}^{r_{3}}F\mathop{\longmapsto}^{r_{5}}F\mathop{\longmapsto}^{r_{4}}F\mathop{\longmapsto}^{r_{3}}F$ $\displaystyle\mathop{\longmapsto}^{\sigma_{01}}qF^{-1}$ $\displaystyle\mathop{\longmapsto}^{r_{2}}q{a_{0}a_{1}^{-1}a_{2}^{2}G+1\over a_{0}a_{1}^{-1}G+a_{2}^{2}}F^{-1}$ $\displaystyle\mathop{\longmapsto}^{r_{0}}q{a_{0}a_{1}^{-1}a_{2}^{2}G+1\over a_{0}^{-1}a_{1}^{-1}G+a_{0}^{2}a_{2}^{2}}F^{-1}$ $\displaystyle\mathop{\longmapsto}^{r_{1}}q{a_{0}a_{1}^{3}a_{2}^{2}G+1\over a_{0}^{-1}a_{1}G+a_{0}^{2}a_{1}^{2}a_{2}^{2}}F^{-1}$ $\displaystyle\mathop{\longmapsto}^{r_{2}}q{a_{0}a_{1}^{3}a_{2}^{2}G+1\over a_{0}^{-1}a_{1}G+a_{0}^{2}a_{1}^{2}a_{2}^{2}}{a_{0}a_{1}^{-1}a_{2}^{2}G+1\over a_{0}a_{1}^{-1}G+a_{2}^{2}}F^{-1}$ $\displaystyle=q\mathsf{p}^{2}\mathsf{t}^{-2}{G+\mathsf{t}\mathsf{p}^{-1}a_{1}^{2}\over G+\mathsf{t}^{-1}\mathsf{p}a_{0}^{2}}{G+\mathsf{t}\mathsf{p}^{-1}a_{1}^{-2}\over G+\mathsf{t}^{-1}\mathsf{p}a_{0}^{-2}}F^{-1}.$ Similarly for $G$; $\displaystyle G\mathop{\longmapsto}^{r_{2}}G\mathop{\longmapsto}^{r_{1}}G\mathop{\longmapsto}^{r_{0}}G\mathop{\longmapsto}^{r_{2}}G\mathop{\longmapsto}^{\sigma_{01}}G$ $\displaystyle\mathop{\longmapsto}^{r_{3}}{a_{3}^{2}a_{4}a_{5}^{-1}F+1\over a_{4}a_{5}^{-1}F+a_{3}^{2}}G$ $\displaystyle\mathop{\longmapsto}^{r_{4}}{a_{3}^{2}a_{4}a_{5}^{-1}F+1\over a_{4}^{-1}a_{5}^{-1}F+a_{3}^{2}a_{4}^{2}}G$ $\displaystyle\mathop{\longmapsto}^{r_{5}}{a_{3}^{2}a_{4}a_{5}^{3}F+1\over a_{4}^{-1}a_{5}F+a_{3}^{2}a_{4}^{2}a_{5}^{2}}G$ $\displaystyle\mathop{\longmapsto}^{r_{3}}{a_{3}^{2}a_{4}a_{5}^{3}F+1\over a_{4}^{-1}a_{5}F+a_{3}^{2}a_{4}^{2}a_{5}^{2}}{a_{3}^{2}a_{4}a_{5}^{-1}F+1\over a_{4}a_{5}^{-1}F+a_{3}^{2}}G$ $\displaystyle\mathop{\longmapsto}^{\sigma_{45}}qG^{-1}{a_{3}^{-2}a_{4}a_{5}^{-1}F+1\over a_{4}a_{5}^{-1}F+a_{3}^{-2}}{a_{3}^{-2}a_{4}^{-3}a_{5}^{-1}F+1\over a_{4}^{-1}a_{5}F+a_{3}^{-2}a_{4}^{-2}a_{5}^{-2}}$ $\displaystyle=qG^{-1}\mathsf{t}^{-2}{F+\mathsf{t}a_{4}^{2}\over F+\mathsf{t}^{-1}a_{5}^{2}}{F+\mathsf{t}a_{4}^{-2}\over F+\mathsf{t}^{-1}a_{5}^{-2}}.$ ∎ In terms of the variables $b_{i}$ defined by Definition A.5, we can write the equations (3.23) and (3.24) as follows; $\displaystyle\overline{F}F=qb_{7}b_{8}{G+b_{5}\over G+b_{7}}{G+b_{6}\over G+b_{8}},\qquad G\underline{G}=qb_{3}b_{4}{F+b_{1}\over F+b_{3}}{F+b_{2}\over F+b_{4}}.$ By taking the conjugation (the adjoint action) by $F$ or $G^{-1}$ (see (3.22)), we can also write the equations in the following manner; $\displaystyle F\overline{F}=q^{-1}\tilde{b}_{7}\tilde{b}_{8}{G+\tilde{b}_{5}\over G+\tilde{b}_{7}}{G+\tilde{b}_{6}\over G+\tilde{b}_{8}},\qquad\underline{G}G=q^{-1}\tilde{b}_{3}\tilde{b}_{4}{F+\tilde{b}_{1}\over F+\tilde{b}_{3}}{F+\tilde{b}_{2}\over F+\tilde{b}_{4}},$ where $\tilde{b}_{i}=qb_{i}$. ### 3.3. Adjoint action and Yang-Baxter relation In the Heisenberg form of the discrete Painlevé equations the Hamiltonian acts on the dynamical variables $(F,G)$ by the adjoint action. Hence, we have to work out the adjoint action of the affine Weyl group generators including the diagram automorphism $\sigma=\sigma_{01}\sigma_{45}$ in the translation (See Appendix A). The fundamental part is to realize the birational transformation of the non-commutative variables $(F,G)$ by the Weyl reflections $r_{2}$ and $r_{3}$ as the adjoint action. We can achieve it by using the following function [23]; ###### Definition 3.9. For $X\in\mathbb{F}$, and $z\in\mathbb{K}$, i.e. when we have $zX=Xz$, set $\displaystyle\theta(X;q)=(X;q)_{\infty}(q/X;q)_{\infty},$ (3.25) $\displaystyle R(z,X)={(-X;q)_{\infty}(-qX^{-1};q)_{\infty}\over(-z^{-1}X;q)_{\infty}(-z^{-1}qX^{-1};q)_{\infty}}={\theta(-X;q)\over(-z^{-1}X;q)_{\infty}(-z^{-1}qX^{-1};q)_{\infty}}.$ (3.26) Recall that $\mathbb{F}=\mathbb{K}(F,G)$ is the quotient skew field of $\mathbb{K}\langle F,G\rangle$. The function $R(z,X)$ satisfies the following formulas; ###### Lemma 3.10. $\displaystyle R(z,X)=R(z,qX^{-1}),$ (3.27) $\displaystyle R(z,X)R(z^{-1},X)={\theta(-X;q)\theta(-X;q)\over\theta(-zX;q)\theta(-z^{-1}X;q)}=R(z,qX)R(z^{-1},qX).$ (3.28) We will need the following Yang-Baxter relation to check the Coxeter relations among the adjoint actions of $R_{i}$ to be defined shortly (see Prop 3.17) [23]. ###### Proposition 3.11 ([17]). We have the Yang-Baxter equation $\displaystyle R(x,F)R(xy,G)R(y,F)=R(y,G)R(xy,F)R(x,G).$ (3.29) The proof given in [16] was based on the Ramanujan’s summation formula, which implies the expansion formula of $R(z,X)$; ###### Remark 3.12. Ramanujan’s summation formula for the bilateral basic hypergeometric series $\displaystyle{}_{1}\psi_{1}(a;b;q;z)=\sum_{n=-\infty}^{\infty}{(a;q)_{n}\over(b;q)_{n}}z^{n}$ $\displaystyle={(q;q)_{\infty}(b/a;q)_{\infty}(az;q)_{\infty}(q/az;q)_{\infty}\over(b;q)_{\infty}(q/a;q)_{\infty}(z;q)_{\infty}(b/az;q)_{\infty}}\qquad(|b/a|<z<1),$ (3.30) gives $\displaystyle{(q;q)_{\infty}(q/z^{2};q)_{\infty}\over(q/z;q)_{\infty}(q/z;q)_{\infty}}R(z,X)=\sum_{n=-\infty}^{\infty}{(z;q)_{n}\over(q/z;q)_{n}}(-z^{-1}X)^{n}.$ (3.31) Essentially the same relation as (3.29) is proved in [34], where an elementary proof by the Heine’s formula is provided. It is also worth mentioning that the Yang-Baxter relation (3.29) is closely related to the quantum dilogarithmic identities; ###### Proposition 3.13 ([29]). We have the five term identity $\displaystyle(-F;q)_{\infty}(-G;q)_{\infty}=(-G;q)_{\infty}(-FG;q)_{\infty}(-F;q)_{\infty}.$ (3.32) Let us begin with the action of the affine Weyl group. The action on the root variables is easily obtained by introducing the dual letters. ###### Definition 3.14. Let $\partial_{0},\ldots,\partial_{5}$ be the dual letters associated with the simple roots satisfying $[\partial_{j},\alpha_{k}]=a_{jk}$. Set $\displaystyle\rho_{i}=e^{{\pi\over 2}\sqrt{-1}\alpha_{i}\partial_{i}}.$ (3.33) Later we also use the dual letters associated with the fundamental weights satisfying $[\partial_{j}^{\prime},\alpha_{k}]=\delta_{jk}$ for a realization of the adjoint action of the diagram automorphism. ###### Lemma 3.15. The action of the affine Weyl group on $\mathbb{K}=\mathbb{C}(a)$ is realized by the adjoint action $\displaystyle r_{i}\cdot a_{j}=\rho_{i}a_{j}\rho_{i}^{-1}.$ (3.34) Among the Weyl reflections $r_{i}$, only $r_{2}$ and $r_{3}$ act on $(F,G)$ non-trivially. We can show they are realized by the adjoint action of $R(z,X)$. ###### Lemma 3.16. $\displaystyle R\left(a_{2}^{2},{a_{0}\over a_{1}}G\right)F\,R\left(a_{2}^{2},{a_{0}\over a_{1}}G\right)^{-1}=F{{a_{0}\over a_{1}}G+a_{2}^{2}\over{a_{0}\over a_{1}}a_{2}^{2}G+1}=r_{2}\cdot F,$ (3.35) $\displaystyle R\left(a_{3}^{2},{a_{5}\over a_{4}}F\right)G\,R\left(a_{3}^{2},{a_{5}\over a_{4}}F\right)^{-1}={{a_{5}\over a_{4}}a_{3}^{2}F+1\over{a_{5}\over a_{4}}F+a_{3}^{2}}G=r_{3}\cdot G.$ (3.36) ###### Proof. For any parameter $z$, the computation goes as follows; $\displaystyle R(z,G)FR(z,G)^{-1}=FR(z,qG)R(z,G)^{-1}=F{1+z^{-1}G\over 1+z^{-1}G^{-1}}G^{-1}=F{G+z\over zG+1},$ $\displaystyle R(z,F)GR(z,F)^{-1}=R(z,F)R(z,qF)^{-1}G={1+z^{-1}F^{-1}\over 1+z^{-1}F}FG={zF+1\over F+z}G.$ ∎ Combining these two lemmas, we obtain the following result. ###### Proposition 3.17. If we define $\displaystyle R_{i}=\rho_{i}\qquad(i=0,1,4,5),$ (3.37) $\displaystyle R_{2}=R\left(a_{2}^{2},{a_{0}\over a_{1}}G\right)\rho_{2},\qquad R_{3}=R\left(a_{3}^{2},{a_{4}\over a_{5}}F\right)\rho_{3},$ (3.38) the action of the affine Weyl group on the skew field $\mathbb{F}$ is realized by the adjoint action $\displaystyle r_{i}\cdot u={\rm Ad}(R_{i})u=R_{i}uR_{i}^{-1}\qquad(0\leq i\leq 5,u\in\mathbb{F}).$ (3.39) As we have mentioned before, one can check ${\rm Ad}(R_{i})$ satisfy the Coxeter relation by using the Yang-Baxter relation (3.29). Next let us consider the adjoint representation of the diagram automorphism $\sigma:=\sigma_{01}\sigma_{45}$888Though $\sigma_{01}$ and $\sigma_{45}$ are anti-homomorphisms, the composition $\sigma$ is a ring homomorphism.. Since $\sigma:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\longmapsto(a_{1},a_{0},a_{2},a_{3},a_{5},a_{4},qF^{-1},qG^{-1}),$ (3.40) the action on the root variables $a_{i}$ is realized by ${\rm Ad}(\rho_{2}\rho_{1}\rho_{0}\rho_{2}\rho_{3}\rho_{4}\rho_{5}\rho_{3}):(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5},F,G)\longmapsto(a_{1},a_{0},\mathsf{p}a_{2},\mathsf{p}^{-1}a_{3},a_{5},a_{4},F,G).$ (3.41) up to the scaling of $(a_{2},a_{3})$, which is nothing but the action of the translation element $T$. On the other hand, the adjoint representation of $\sigma$ on the dynamical variables $(F,G)$ is given by suitable combinations of the theta functions with non-commutative variables in $\mathbb{F}$; ###### Lemma 3.18. We have ${\rm Ad}\Bigl{(}\theta(F^{-1}G;q)^{-1}\Bigr{)}F=-G,\qquad{\rm Ad}\Bigl{(}\theta(F^{-1}G;q)^{-1}\Bigr{)}G=-q^{-1}F^{-1}G^{2}.$ (3.42) ###### Proof. Similarly to the proof of Lemma 3.16, we can compute; $\displaystyle\theta(F^{-1}G;q)^{-1}F$ $\displaystyle=F\frac{1}{(qF^{-1}G;q)_{\infty}(G^{-1}F;q)_{\infty}}$ $\displaystyle=F\frac{1-F^{-1}G}{1-G^{-1}F}\theta(F^{-1}G;q)^{-1}=-G\cdot\theta(F^{-1}G;q)^{-1},$ $\displaystyle\theta(F^{-1}G;q)^{-1}G$ $\displaystyle=G\frac{1}{(qF^{-1}G;q)_{\infty}(G^{-1}F;q)_{\infty}}$ $\displaystyle=-GF^{-1}G\theta(F^{-1}G;q)^{-1}=-q^{-1}F^{-1}G^{2}\cdot\theta(F^{-1}G;q)^{-1}.$ ∎ ###### Lemma 3.19. For any $z\in\mathbb{K}$, i.e. when $zF=Fz$ and $zG=Gz$, we have $\displaystyle{\rm Ad}\Bigl{(}\left(\theta(zG;q)\theta(z^{-1}G;q)\right)^{-1}\Bigr{)}F=FG^{2},\quad{\rm Ad}\Bigl{(}\left(\theta(zG;q)\theta(z^{-1}G;q)\right)^{-1}\Bigr{)}F^{-1}=G^{-2}F^{-1},$ $\displaystyle{\rm Ad}\Bigl{(}\left(\theta(zG;q)\theta(z^{-1}G;q)\right)^{-1}\Bigr{)}G=G.$ Note that the right hand sides of the above relations are independent of $z$. ###### Proof. The second equation follows from the first. The third one is trivial. We can check the first equation as follows; $\displaystyle\left(\theta(zG;q)\theta(z^{-1}G;q)\right)^{-1}F$ $\displaystyle=F\frac{(1-zG)(1-z^{-1}G)}{(1-z^{-1}G^{-1})(1-zG^{-1})}\left(\theta(zG;q)\theta(z^{-1}G;q)\right)^{-1}$ $\displaystyle=FG^{2}\cdot\left(\theta(zG;q)\theta(z^{-1}G;q)\right)^{-1}.$ ∎ Combining these lemmas we find $\displaystyle{\rm Ad}\left(\left(\theta(zG;q)\theta(z^{-1}G;q)\right)^{-1}\theta(F^{-1}G;q)^{-1}\right)F=-G,$ (3.43) $\displaystyle{\rm Ad}\left(\left(\theta(zG;q)\theta(z^{-1}G;q)\right)^{-1}\theta(F^{-1}G;q)^{-1}\right)G=-qF^{-1}.$ (3.44) Hence, the square of this adjoint action999We can specialize the parameter $z$ of the first and the second actions at our disposal. agrees with the action of $\sigma$ on $(F,G)\mapsto(qF^{-1},qG^{-1})$. On the other hand, if we define $S:=e^{\frac{\pi}{2}\sqrt{-1}(\alpha_{0}-\alpha_{1})(\partial^{\prime}_{0}-\partial^{\prime}_{1})}e^{\frac{\pi}{2}\sqrt{-1}(\alpha_{4}-\alpha_{5})(\partial^{\prime}_{4}-\partial^{\prime}_{4})},$ (3.45) the adjoint action $\mathrm{Ad}(S)$ gives the same action as $(\rho_{2}\rho_{1}\rho_{0}\rho_{2}\rho_{3}\rho_{4}\rho_{5}\rho_{3})^{-1}T_{\mathsf{p},a_{2}}T_{\mathsf{p},a_{3}}^{-1}:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5})\mapsto(a_{1},a_{0},a_{2},a_{3},a_{5},a_{4})$ on the root variables. Hence, the desired realization of $\sigma$ by the adjoint action is ###### Proposition 3.20. If we define $\displaystyle\Sigma:=\left(\theta\left(-{a_{0}\over a_{1}}G;q\right)\theta\left(-{a_{1}\over a_{0}}G;q\right)\right)^{-1}\theta(F^{-1}G;q)^{-1}$ $\displaystyle\qquad\cdot\left(\theta\left({a_{4}\over a_{5}}G;q\right)\theta\left({a_{5}\over a_{4}}G;q\right)\right)^{-1}\theta(F^{-1}G;q)^{-1}\cdot S,$ the action of the element $\sigma=\sigma_{01}\sigma_{45}$ on the skew field $\mathbb{F}$ is realized by the adjoint action $\displaystyle\sigma\cdot u={\rm Ad}(\Sigma)u\qquad(u\in\mathbb{F}).$ ### 3.4. Hasegawa operator ###### Proposition 3.21. Define the Hasegawa operator $H$ for the $qq$-Painlevé VI equation as $\displaystyle H_{\mathrm{VI}}=R_{2}R_{1}R_{0}R_{2}R_{3}R_{4}R_{5}R_{3}\Sigma,$ (3.46) then $\displaystyle H_{\mathrm{VI}}=\frac{1}{\varphi(-qb_{5}G^{-1})\varphi(-qb_{6}G^{-1})\varphi(-b_{7}^{-1}G)\varphi(-b_{8}^{-1}G)}\theta(F^{-1}G;q)^{-1}$ $\displaystyle\times\frac{1}{\varphi(q\mathsf{p}^{-2}b_{1}G^{-1})\varphi(q\mathsf{p}^{-2}b_{2}G^{-1})\varphi(\mathsf{p}^{-2}b_{3}^{-1}G)\varphi(\mathsf{p}^{-2}b_{4}^{-1}G)}\theta(F^{-1}G;q)^{-1}T_{\mathsf{p},a_{2}}T_{\mathsf{p},a_{3}}^{-1}$ (3.47) and the non-stationary difference equation (2.14) is equivalent to $\displaystyle H_{\mathrm{VI}}\cdot\mathsf{u}(b,G,Q|q,t^{-1})=\mathsf{u}(b,G,Q|q,t^{-1}).$ ###### Proof. We proceed as follows; $\displaystyle R_{2}R_{1}R_{0}R_{2}R_{3}R_{4}R_{5}R_{3}\Sigma$ $\displaystyle=$ $\displaystyle R\left(a_{2}^{2},{a_{0}\over a_{1}}G\right)\rho_{2}\rho_{1}\rho_{0}R\left(a_{2}^{2},{a_{0}\over a_{1}}G\right)\rho_{2}R\left(a_{3}^{2},{a_{4}\over a_{5}}F\right)\rho_{3}\rho_{4}\rho_{5}R\left(a_{3}^{2},{a_{4}\over a_{5}}F\right)\rho_{3}\Sigma$ $\displaystyle=$ $\displaystyle R\left(a_{2}^{2},{a_{0}\over a_{1}}G\right)R\left((a_{0}a_{1}a_{2})^{2},{a_{1}\over a_{0}}G\right)R\left(\mathsf{p}^{2}(a_{3}a_{4}a_{5})^{-2},{a_{4}\over a_{5}}F\right)R\left(\mathsf{p}^{2}a_{3}^{-2},{a_{5}\over a_{4}}F\right)$ $\displaystyle\cdot\left(\theta\left(-{a_{0}\over a_{1}}G;q\right)\theta\left(-{a_{1}\over a_{0}}G;q\right)\right)^{-1}\theta(F^{-1}G;q)^{-1}\left(\theta\left({a_{4}\over a_{5}}G;q\right)\theta\left({a_{5}\over a_{4}}G;q\right)\right)^{-1}$ $\displaystyle\theta(F^{-1}G;q)^{-1}\cdot T_{\mathsf{p},a_{2}}T_{\mathsf{p},a_{3}}^{-1}$ $\displaystyle=$ $\displaystyle R\left(a_{2}^{2},{a_{0}\over a_{1}}G\right)R\left((a_{0}a_{1}a_{2})^{2},{a_{1}\over a_{0}}G\right)\left(\theta\left(-{a_{0}\over a_{1}}G;q\right)\theta\left(-{a_{1}\over a_{0}}G;q\right)\right)^{-1}\theta(F^{-1}G;q)^{-1}$ $\displaystyle\cdot R\left(\mathsf{p}^{2}(a_{3}a_{4}a_{5})^{-2},-{a_{4}\over a_{5}}G\right)R\left(\mathsf{p}^{2}a_{3}^{-2},-{a_{5}\over a_{4}}G\right)\left(\theta\left({a_{4}\over a_{5}}G;q\right)\theta\left({a_{5}\over a_{4}}G;q\right)\right)^{-1}$ $\displaystyle\theta(F^{-1}G;q)^{-1}\cdot T_{\mathsf{p},a_{2}}T_{\mathsf{p},a_{3}}^{-1}$ $\displaystyle=$ $\displaystyle{1\over(-{a_{0}\over a_{1}a_{2}^{2}}G;q)_{\infty}(-{a_{1}\over a_{0}a_{2}^{2}}qG^{-1};q)_{\infty}}{1\over(-{1\over a_{0}^{3}a_{1}a_{2}^{2}}G;q)_{\infty}(-{1\over a_{0}a_{1}^{3}a_{2}^{2}}qG^{-1};q)_{\infty}}\theta(F^{-1}G;q)^{-1}$ $\displaystyle\cdot{1\over({a_{3}^{2}a_{4}a_{5}^{3}\over\mathsf{p}^{2}}G;q)_{\infty}({a_{3}^{2}a_{4}^{3}a_{5}\over\mathsf{p}^{2}}qG^{-1};q)_{\infty}}{1\over({a_{3}^{2}a_{4}\over\mathsf{p}^{2}a_{5}}G;q)_{\infty}({a_{3}^{2}a_{5}\over\mathsf{p}^{2}a_{4}}qG^{-1};q)_{\infty}}.\theta(F^{-1}G;q)^{-1}T_{\mathsf{p},a_{2}}T_{\mathsf{p},a_{3}}^{-1}.$ ∎ We note that the combination of the shift operators $T_{\mathsf{p},a_{2}}T_{\mathsf{p},a_{3}}^{-1}$ keeps the constraints $a_{0}a_{1}a_{2}^{2}a_{3}^{2}a_{4}a_{5}=t^{1/4}$ intact. Hence, the desired identification is completed, if we can show the adjoint action of $\theta(F^{-1}G;q)^{-1}$ and $\widetilde{\mathcal{B}}=T_{(qt^{1/2}Q)^{-1/2},x}\cdot\mathcal{B}$ are the same. The adjoint action of $\theta(F^{-1}G;q)^{-1}$ is already given in Lemma 3.18. On the other hand the fundamental commutation relation (2.2) for $\mathcal{B}$ implies ${\rm Ad}\mathcal{B}(\tilde{F})=-\tilde{G},\qquad{\rm Ad}\mathcal{B}(\tilde{F})=-q^{-1}\tilde{F}^{-1}\tilde{G}^{2},$ (3.48) for $\tilde{F}=x^{-1}p,~{}\tilde{G}=-x^{-1}$. Recall that we have rescaled $\tilde{F}$ and $\tilde{G}$. This is the reason why we have to combine $\mathcal{B}$ with the shift operator $T_{(qt^{1/2}Q)^{-1/2},x}$ whose adjoint action produces the necessary multiplication factor for $\tilde{F}$ and $\tilde{G}$. As is well-known, the Painlevé equations are derived from the isomonodromic deformation of linear system. It is an interesting problem to find the Lax operators which give rise to the $qq$-Painlevé VI equation. Note that the monodromy problem is naturally associated with the Yang-Baxter relation of the universal $R$ matrix. In fact in [24] the universal $R$ matrix of $U_{q}(\widehat{\mathfrak{sl}}_{2})$ was used to define local Lax matrices for the $qq$-Painlevé VI equation, which is a good starting point to work out the problem completely. ## 4\. Quantum Seiberg-Witten curve ### 4.1. Gauge transformation and $U(1)$ factor By the following gauge transformation; $\displaystyle{1\over\Phi(t^{2}T_{1}T_{3}\Lambda)\Phi(qtT_{2}T_{4}\Lambda)}\varphi(q^{1/2}t^{1/2}T_{2}x)\varphi(q^{1/2}t^{1/2}T_{4}\Lambda/x)\mathsf{U}(\Lambda,x)=\mathsf{v}^{(1)}(\Lambda,x),$ (4.1) the $qq$-Painlevé VI equation is recast to $\widetilde{H}\mathsf{v}^{(1)}(\Lambda,x)=\mathsf{v}^{(1)}(\Lambda,x)$ with $\displaystyle\widetilde{H}={1\over\varphi(T_{1}q^{1/2}t^{1/2}x)\varphi(T_{3}q^{1/2}t^{1/2}\Lambda x^{-1})}\cdot\mathcal{B}$ $\displaystyle\qquad\cdot{\varphi(tT_{1}T_{3}\Lambda)\varphi(qT_{2}T_{4}\Lambda)\over\varphi(-T_{1}T_{2}x)\varphi(-Q^{-1}x)\varphi(-T_{3}T_{4}Qqt\Lambda x^{-1})\varphi(-q\Lambda x^{-1})}\cdot\mathcal{B}$ $\displaystyle\qquad\cdot T_{qtQ,x}^{-1}T_{t,\Lambda}^{-1}\cdot{1\over\varphi(T_{2}q^{1/2}t^{1/2}x)\varphi(T_{4}q^{1/2}t^{1/2}\Lambda x^{-1})}.$ (4.2) Recall that we have made the gauge transformation twice; $\mathsf{U}(\Lambda,x)\to\mathsf{u}(\Lambda,x)\to\mathsf{v}^{(1)}(\Lambda,x)$. We can see the total gauge factor from the original system of Higgsed quiver gauge theory is simply $\frac{\Phi(qtT_{2}T_{3}\Lambda)\Phi(t^{2}T_{1}T_{4}\Lambda)}{\Phi(t^{2}T_{1}T_{3}\Lambda)\Phi(qtT_{2}T_{4}\Lambda)}\frac{\Phi(T_{3}q^{3/2}t^{-1/2}\Lambda/x)\Phi(T_{4}q^{-1/2}t^{5/2}\Lambda/x)}{\Phi(T_{3}q^{1/2}t^{1/2}\Lambda/x)\Phi(T_{4}q^{1/2}t^{3/2}\Lambda/x)}.$ (4.3) Note that in the original definition of $\mathcal{A}_{3}$ in [44] the four dimensional limit is not well-defined. But after the above gauge transformation we can take the four dimensional limit. In fact let us look at the gauge transformation factor in the plethystic form; $\displaystyle\frac{\Phi(qtT_{2}T_{3}\Lambda)\Phi(t^{2}T_{1}T_{4}\Lambda)}{\Phi(t^{2}T_{1}T_{3}\Lambda)\Phi(qtT_{2}T_{4}\Lambda)}$ $\displaystyle=\exp\left(-\sum_{n=1}^{\infty}\frac{(q^{n}T_{2}^{n}-t^{n}T_{1}^{n})(T_{3}^{n}-T_{4}^{n})}{n(1-q^{n})(1-t^{n})}(t\Lambda)^{n}\right)$ $\displaystyle\to(1-\Lambda)^{\frac{(\mu_{3}-\mu_{4})(\mu_{2}-\mu_{1}-\epsilon_{1}-\epsilon_{2})}{\epsilon_{1}\epsilon_{2}}},\qquad\hbar\to 0,$ (4.4) $\displaystyle\frac{\Phi(T_{3}q^{3/2}t^{-1/2}\Lambda/x)\Phi(T_{4}q^{-1/2}t^{5/2}\Lambda/x)}{\Phi(T_{3}q^{1/2}t^{1/2}\Lambda/x)\Phi(T_{4}q^{1/2}t^{3/2}\Lambda/x)}$ $\displaystyle=\exp\left(-\sum_{n=1}^{\infty}\frac{(1-q^{n}t^{-n})(T_{4}^{n}q^{-n/2}t^{5n/2}-T_{3}^{n}q^{n/2}t^{n/2})}{n(1-q^{n})(1-t^{n})}(\Lambda/x)^{n}\right)$ $\displaystyle\to(1-\Lambda/x)^{\frac{(\epsilon_{1}+\epsilon_{2})(\mu_{4}-\mu_{3}+\epsilon_{1}+2\epsilon_{2})}{\epsilon_{1}\epsilon_{2}}},\qquad\hbar\to 0,$ (4.5) where we set $q=e^{\hbar\epsilon_{1}},t^{-\hbar\epsilon_{2}}$ and $T_{i}=e^{-\hbar\mu_{i}}$. This four dimensional limit completely matches with what we have found for the four dimensional instanton partition functions. This gauge transformation is also regarded as what is called the $U(1)$ factor in the AGT correspondence [2]. Thus the gauge transformation (4.3) is a five dimensional uplift of the $U(1)$ factor. The five dimensional $U(1)$ factor for the Nekrasov partition function has been proposed in [25],[6], where it is associated with pairs of parallel external lines in the fivebrane web diagram. However, this $U(1)$ factor does not recover that of [2] in the four dimensional limit. In this sense the relation of the $U(1)$ factor in [25],[6] and our $U(1)$ factor derived above is not clear at the moment. To simplify the computation with the quantum Seiberg-Witten curve, let us make the change of parameters and variables; $\displaystyle\Lambda^{\prime}$ $\displaystyle=tT_{1}T_{3}\Lambda,\qquad x^{\prime}=T_{1}q^{-1/2}t^{1/2}x,$ (4.6) $\displaystyle d_{1}=T_{1}^{-1}q^{1/2}t^{-1/2}Q^{-1},\quad d_{2}$ $\displaystyle=T_{2}q^{1/2}t^{-1/2},\quad d_{3}=T_{3}^{-1}q^{1/2}t^{-1/2},\quad d_{4}=T_{4}q^{1/2}t^{1/2}Q,$ (4.7) so that $\tilde{H}$ can be written as $\tilde{H}=\SS T_{qtQ,x}^{-1}T_{t,\Lambda}^{-1}$ with $\displaystyle\SS=$ $\displaystyle\frac{1}{\varphi(qx)\varphi(\Lambda/x)}\cdot\mathcal{B}\cdot\frac{\varphi(\Lambda)\varphi(q^{-1}d_{1}d_{2}d_{3}d_{4}\Lambda)}{\varphi(-d_{1}x)\varphi(-d_{2}x)\varphi(-d_{3}\Lambda/x)\varphi(-d_{4}\Lambda/x)}$ $\displaystyle~{}~{}\cdot\mathcal{B}\cdot\frac{1}{\varphi(q^{-1}d_{1}d_{2}x)\varphi(d_{3}d_{4}\Lambda/x)}.$ (4.8) where we delete ′ of $x^{\prime}$ and $\Lambda^{\prime}$. Note that by the change of variables (4.6) and (4.7), we have eliminated the parameter $t$ from (4.1). In fact we can transform $\SS$ to ${\bf H}$ defined by (3.1) by the multiplication of $\varphi(\Lambda)\varphi(q^{-1}d_{1}d_{2}d_{3}d_{4}\Lambda)$, which commutes with $\mathcal{B}$ and the adjoint action of $\varphi(q^{-1}d_{1}d_{2}x)\varphi(d_{3}d_{4}\Lambda/x)$. It is this form of $\tilde{H}$ that naturally leads to the solutions by the instanton counting of the affine Laumon space with the following property. The instanton partition function is a formal double power series in $x$ and $\Lambda/x$; $\mathcal{Z}(\Lambda,x)=\sum_{m,n\geq 0}c_{m,n}x^{m}(\Lambda/x)^{n},$ (4.9) where the coefficients $c_{m,n}$ are functions of $Q,T_{i},q$ and $t$. One of the characterestic features is that the “boundary” coefficients $c_{0,n}$ and $c_{m,0}$ factorize, which is a consequence of the fact that there is a unique fixed point of the torus action on the instanton moduli space, which corresponds to the topological number $n=0$ or $m=0$ This means $\SS$ should be regarded as the Hamiltonian on the gauge theory side. ### 4.2. Five dimensional quantum Seiberg-Witten curve When $t=1$, the $qq$-Painlevé VI equation becomes an autonomous system which admits a conserved curve. The curve is identified with a quantization of the Seiberg-Witten curve for the corresponding gauge theory. Here we work out such a quantum curve based on the commutativity with the operator $\SS$. ###### Proposition 4.1. Let $\mathcal{D}=\mathcal{D}(x,p)$ be an operator of the form $\mathcal{D}=\sum_{i,j}c_{i,j}x^{i}p^{j},$ (4.10) with nonzero coefficients $\left[\begin{array}[]{ccc}c_{-1,1}&c_{0,1}&c_{1,1}\\\ c_{-1,0}&c_{0,0}&c_{1,0}\\\ c_{-1,-1}&c_{0,-1}&c_{1,-1}\\\ \end{array}\right]=\left[\begin{array}[]{ccc}\Lambda\mu&-\frac{\mu q^{2}+\Lambda d_{1}d_{2}}{q}&d_{1}d_{2}\\\ -\Lambda\mu\left(d_{3}+d_{4}\right)&u&-d_{1}-d_{2}\\\ \Lambda\mu d_{3}d_{4}&-\Lambda\mu d_{3}d_{4}-1&1\\\ \end{array}\right].$ (4.11) Then it satisfies the relation $\SS^{-1}\mathcal{D}(x,p)\SS=\mathcal{D}(\mu x,p),$ (4.12) where $u,\mu\in\mathbb{C}$ are free parameters. ###### Proof. We compute the successive transformations of the operators $\mathcal{D}\mapsto\mathcal{D}_{1}\mapsto\cdots\mapsto\mathcal{D}_{6},$ (4.13) under the adjoint actions ${\rm Ad}(X):\mathcal{D}\mapsto X\mathcal{D}X^{-1}$ for factors $X$ in $\SS^{-1}$. In the following we will display the operators $\mathcal{D}_{i}$ by their coefficient matrices $(c_{i,j})$101010 Initialy the range (support) of the indices $(i,j)$ is $-1\leq i,j\leq 1$. However, by the adjoint action the range of the power of $x$ is extended to $-2\leq i\leq 2$. See also the Nowton polygons displayed below.. First, under the gauge transformation ${\rm Ad}\big{(}{\varphi(qx)\varphi(\frac{\Lambda}{x})}\big{)}:p\mapsto{\displaystyle\frac{1-qx}{1-\frac{\Lambda}{qx}}}p,$ we have $\mathcal{D}\mapsto\mathcal{D}_{1}=\left[\begin{array}[]{ccccc}0&0&-q\mu&\mu q^{2}+d_{1}d_{2}&-qd_{1}d_{2}\\\ 0&-\Lambda\mu\left(d_{3}+d_{4}\right)&u&-d_{1}-d_{2}&0\\\ -\Lambda^{2}\mu d_{3}d_{4}&\Lambda\left(\mu d_{3}d_{4}+1\right)&-1&0&0\end{array}\right].$ (4.14) Next, by the action of ${\rm Ad}(\mathcal{B}^{-1}):x^{i}\mapsto q^{-i(i+1)/2}x^{i}p^{-i}$, we obtain $\mathcal{D}_{1}\mapsto\mathcal{D}_{2}=\left[\begin{array}[]{ccccc}-\frac{\Lambda^{2}\mu d_{3}d_{4}}{q}&-\Lambda\mu\left(d_{3}+d_{4}\right)&-q\mu&0&0\\\ 0&\Lambda\left(\mu d_{3}d_{4}+1\right)&u&\frac{\mu q^{2}+d_{1}d_{2}}{q}&0\\\ 0&0&-1&-\frac{d_{1}+d_{2}}{q}&-\frac{d_{1}d_{2}}{q^{2}}\end{array}\right].$ (4.15) Then by the gauge transformation ${\rm Ad}\big{(}\varphi(-d_{1}x)\varphi(-d_{3}\frac{\Lambda}{x})\big{)}:p\mapsto{\displaystyle\frac{(1+d_{1}x)}{(1+d_{3}\frac{\Lambda}{qx})}}p,$ we have $\mathcal{D}_{2}\mapsto\mathcal{D}_{3}=\left[\begin{array}[]{ccc}-\Lambda\mu d_{4}&-\mu\left(q+\Lambda d_{1}d_{4}\right)&-q\mu d_{1}\\\ \Lambda\left(\mu d_{3}d_{4}+1\right)&u&\frac{\mu q^{2}+d_{1}d_{2}}{q}\\\ -\Lambda d_{3}&-\frac{q+\Lambda d_{2}d_{3}}{q}&-\frac{d_{2}}{q}\end{array}\right].$ (4.16) And by further gauge transformation ${\rm Ad}\big{(}\varphi(-d_{2}x)\varphi(-d_{4}\frac{\Lambda}{x})\big{)}:p\mapsto{\displaystyle\frac{(1+d_{2}x)}{(1+d_{4}\frac{\Lambda}{qx})}}p,$ we have $\mathcal{D}_{3}\mapsto\mathcal{D}_{4}=\left[\begin{array}[]{ccccc}0&0&-q\mu&-q\mu\left(d_{1}+d_{2}\right)&-q\mu d_{1}d_{2}\\\ 0&\Lambda\left(\mu d_{3}d_{4}+1\right)&u&\frac{\mu q^{2}+d_{1}d_{2}}{q}&0\\\ -\Lambda^{2}d_{3}d_{4}&-\Lambda\left(d_{3}+d_{4}\right)&-1&0&0\\\ \end{array}\right].$ (4.17) Then the action of ${\rm Ad}(\mathcal{B}^{-1})$ gives $\mathcal{D}_{4}\mapsto\mathcal{D}_{5}=\left[\begin{array}[]{ccccc}-\frac{\Lambda^{2}d_{3}d_{4}}{q}&\Lambda\left(\mu d_{3}d_{4}+1\right)&-q\mu&0&0\\\ 0&-\Lambda\left(d_{3}+d_{4}\right)&u&-\mu\left(d_{1}+d_{2}\right)&0\\\ 0&0&-1&\frac{\mu q^{2}+d_{1}d_{2}}{q^{2}}&-\frac{\mu d_{1}d_{2}}{q^{2}}\end{array}\right].$ (4.18) Finally, by the gauge transformation ${\rm Ad}\big{(}\varphi(\frac{d_{1}d_{2}x}{q})\varphi(d_{3}d_{4}\frac{\Lambda}{x})\big{)}:p\mapsto{\displaystyle\frac{1-\frac{d_{1}d_{2}}{q}x}{1-\frac{\Lambda d_{3}d_{4}}{qx}}}p,$ we have $\mathcal{D}_{5}\mapsto\mathcal{D}_{6}=\left[\begin{array}[]{ccc}\Lambda&-\frac{\mu q^{2}+\Lambda d_{1}d_{2}}{q}&\mu d_{1}d_{2}\\\ -\Lambda\left(d_{3}+d_{4}\right)&u&-\mu\left(d_{1}+d_{2}\right)\\\ \Lambda d_{3}d_{4}&-\Lambda\mu d_{3}d_{4}-1&\mu\end{array}\right].$ (4.19) In total, we obtained the relation $\SS^{-1}\mathcal{D}\SS=\mathcal{D}_{6}=\mathcal{D}|_{x\to\mu x},$ (4.20) as desired. ∎ ###### Corollary 4.2. If we set $\mu=(tqQ)^{-1}$, then we have ${\tilde{H}}^{-1}\mathcal{D}{\tilde{H}}=T_{t,\Lambda}(\mathcal{D}),$ (4.21) where ${\tilde{H}}=\SS T_{tqQ,x}^{-1}T_{t,\Lambda}^{-1}$. Namely, the operator $\mathcal{D}$ with $\mu=(tqQ)^{-1}$ is conserved under the evolution by $\tilde{H}$ in autonomous case: $t=1$. ###### Remark 4.3. The converse of the Proposition 4.1 is also true. Namely, for a general operator $\mathcal{D}$ of the form (4.10) with the initial support $i,j\in\\{-1,0,1\\}$, the condition (4.12) with fixed $\mu$ determines the coefficients $c_{i,j}$ up to two free parameters. In fact, in order to keep the form of the operator $\mathcal{D}$ under the successive transformations as above, one has six linear constraints on the nine coefficients $c_{i,j}$. Then the condition (4.12) gives one more constraint. Hence two coefficients (the constant term $u=c_{0,0}$ and overall normalization) remain free. ###### Remark 4.4. The position of the nonzero coefficients $c_{i,j}$ shows the Newton polygon of the operators $\mathcal{D}_{k}$. Their transitions are as follows: $\mathcal{D}$$\mapsto$$\mathcal{D}_{1}$$\mapsto$${\rm Ad}\mathcal{B}^{-1}$$\mathcal{D}_{2}$$\mapsto$$\mathcal{D}_{3}$$\mapsto$$\mathcal{D}_{4}$$\mapsto$${\rm Ad}\mathcal{B}^{-1}$$\mathcal{D}_{5}$$\mapsto$$\mathcal{D}_{6}$. We see the Newton polygon of $\mathcal{D}_{3}$ has also a rectangular shape. More precisely we have $\mathcal{D}_{3}=\mathcal{D}\Big{|}_{\\{d_{2}\to\frac{\mu q^{2}}{d_{2}},\ d_{4}\to\frac{1}{d_{4}\mu},\ \Lambda\to\frac{d_{2}d_{4}\Lambda}{q},\ x\to-\frac{d_{2}x}{q}\\}}.$ (4.22) This corresponds to the factorization property Proposition 5.1 which will be discussed in the next section. Since the parameter $u$ is stable under the adjoint action, we can regard it as a trivial free parameter. Setting $u=0$ the quantum Seiberg-Witten curve is explicitly $\displaystyle\mathcal{D}_{\rm SW}(\Lambda,x)=(d_{1}d_{2}x-\mu q)(1-\frac{\Lambda}{qx})p-(d_{1}+d_{2})x-\frac{(d_{3}+d_{4})\Lambda\mu}{x}+(x-1)(1-\frac{d_{3}d_{4}\Lambda\mu}{x})\frac{1}{p}$ $\displaystyle~{}~{}~{}=x\frac{(1-d_{1}p)(1-d_{2}p)}{p}-(1+d_{3}d_{4}\Lambda\mu)\frac{1}{p}-(\mu q+\frac{d_{1}d_{2}\Lambda}{q})p+\frac{\Lambda\mu}{x}\frac{(p-d_{3})(p-d_{4})}{p}.$ (4.23) This is a non-commutative Laurent polynomial in $(x,p)$: $px=qxp$. Note that the highest and lowest terms in $x$ and $p$ are all factorized. When $(x,p)$ are commutative, the curve (4.2) reduces to the $M$-theoretic curve, which is obtained from the toric diagram, or the five-brane web [11],[15],[28],[9]. ### 4.3. Four dimensional limit We study the $q\to 1$ limit of the operator (4.1). Let us use the short hand notation such as $F^{a,b,b}_{c}=T_{q,a}T_{q,b}^{2}T_{q,c}^{-1}F=F(qa,q^{2}b,q^{-1}c,\ldots),$ (4.24) to represent the $q$-shifts of a function $F=F(a,b,c,\ldots)$. ###### Proposition 4.5. We have $\begin{array}[]{ll}(1-\Lambda)(\SS)^{\Lambda}_{d_{1}}=(1-x)(\SS)_{x}+d_{2}(x-\Lambda)\SS p.&\qquad(d_{1}\leftrightarrow d_{2}),\\\ (1-\Lambda)(\SS)^{\Lambda}_{d_{3}}=(1-{\displaystyle\frac{\Lambda}{x}})\SS+d_{4}({\displaystyle\frac{\Lambda}{x}}-\Lambda)p^{-1}\SS,&\qquad(d_{3}\leftrightarrow d_{4}),\end{array}$ (4.25) and $\SS x=q^{2}x(\SS)^{d_{1},d_{2}}_{d_{3},d_{4}}p^{2}.$ (4.26) ###### Proof. We write $\SS$ as $\SS=A_{1}\mathcal{B}A_{2}\mathcal{B}A_{3}$, where $\displaystyle A_{1}$ $\displaystyle=\frac{1}{\varphi(qx)\varphi(\Lambda/x)}\quad A_{2}=\frac{\varphi(\Lambda)\varphi(q^{-1}d_{1}d_{2}d_{3}d_{4}\Lambda)}{\varphi(-d_{1}x)\varphi(-d_{2}x)\varphi(-d_{3}\Lambda/x)\varphi(-d_{4}\Lambda/x)}$ $\displaystyle A_{3}$ $\displaystyle=\frac{1}{\varphi(q^{-1}d_{1}d_{2}x)\varphi(d_{3}d_{4}\Lambda/x)}.$ (4.27) Then we have ${A_{1}}^{\Lambda,x}_{d_{1}}=(1-qx)A_{1},\quad{A_{2}}^{\Lambda,x}_{d_{1}}={\displaystyle\frac{1+d_{2}x}{1-\Lambda}}A_{2},\quad{A_{3}}^{\Lambda,x}_{d_{1}}=A_{3}.$ (4.28) Using these relations, we have $\displaystyle(\SS)^{\Lambda,x}_{d_{1}}\SS^{-1}$ $\displaystyle={A_{1}}^{\Lambda,x}_{d_{1}}\mathcal{B}{A_{2}}^{\Lambda,x}_{d_{1}}{A_{2}}^{-1}\mathcal{B}^{-1}A_{1}^{-1}={\displaystyle\frac{1-qx}{1-\Lambda}}A_{1}(1+d_{2}px)A_{1}^{-1}$ $\displaystyle={\displaystyle\frac{1-qx}{1-\Lambda}}\Bigl{(}1+qd_{2}x{\displaystyle\frac{1-{\displaystyle\frac{\Lambda}{qx}}}{1-qx}}p\Bigr{)},$ hence $(1-\Lambda)(\SS)^{\Lambda,x}_{d_{1}}=(1-qx)\SS+qd_{2}x(1-{\displaystyle\frac{\Lambda}{qx}})(\SS)^{x}p.$ (4.29) Putting $x\to x/q$, we obtain the first relation of (4.25). Similarly, the second relation of (4.25) follows from ${A_{1}}^{\Lambda}_{d_{3}}=(1-{\displaystyle\frac{\Lambda}{x}})A_{1},\quad{A_{2}}^{\Lambda}_{d_{3}}={\displaystyle\frac{1+d_{4}{\displaystyle\frac{\Lambda}{x}}}{1-\Lambda}}A_{2},\quad{A_{3}}^{\Lambda}_{d_{3}}=A_{3}.$ (4.30) The relation (4.26) follows easily as $\begin{array}[]{l}\SS x=A_{1}\mathcal{B}A_{2}\mathcal{B}A_{3}x=A_{1}\mathcal{B}A_{2}px\mathcal{B}A_{3}=qA_{1}\mathcal{B}xA_{2}\mathcal{B}{A_{3}}^{x}p\\\ \quad=qA_{1}px\mathcal{B}A_{2}\mathcal{B}{A_{3}}^{x}p=q^{2}x{A_{1}}\mathcal{B}{A_{2}}^{x}\mathcal{B}{A_{3}}^{x,x}p^{2}\\\ \quad=q^{2}x(\SS)^{d_{1},d_{2}}_{d_{3},d_{4}}p^{2}.\end{array}$ (4.31) ∎ ###### Theorem 4.6. We put $q=e^{h}$, $d_{i}=e^{hm_{i}}$, then we have $\displaystyle\SS$ $\displaystyle\to 1+hH_{\rm 4d}+O(h^{2}),\quad(h\to 0)$ $\displaystyle H_{\rm 4d}$ $\displaystyle=\vartheta_{x}(\vartheta_{x}+1)+{\displaystyle\frac{\Lambda-x}{1-\Lambda}}(\vartheta_{x}+m_{1})(\vartheta_{x}+m_{2})+{\displaystyle\frac{\Lambda}{x}}{\displaystyle\frac{x-1}{1-\Lambda}}(\vartheta_{x}-m_{3})(\vartheta_{x}-m_{4}).$ (4.32) ###### Proof. First we consider the expansion $v\mathbin{:=}\SS.1=v_{0}+hv_{1}+O(h^{2})$ (4.33) Obviously $v_{0}=1$. For the first order term $v_{1}$, we have from (4.25) $\begin{array}[]{ll}(1-\Lambda)\\{({v_{1}}|_{m_{1}\to m_{1}-1})-v_{1}\\}=m_{2}(x-\Lambda)v_{0},&\qquad(m_{1}\leftrightarrow m_{2}),\\\ (1-\Lambda)\\{({v_{1}}|_{m_{3}\to m_{3}-1})-v_{1}\\}=\frac{\Lambda}{x}m_{4}(1-x)v_{0},&\qquad(m_{3}\leftrightarrow m_{4}),\end{array}$ (4.34) hence we have $v_{1}={\displaystyle\frac{\Lambda-x}{1-\Lambda}}m_{1}m_{2}+{\displaystyle\frac{\Lambda}{x}}{\displaystyle\frac{x-1}{1-\Lambda}}m_{3}m_{4}.$ (4.35) Then from this and iterative use of (4.26), we obtain $\begin{array}[]{l}\SS x^{n}=q^{n(n+1)}x^{n}\Big{\\{}(T_{q,d_{1}}T_{q,d_{2}}T^{-1}_{q,d_{3}}T^{-1}_{q,d_{4}})^{n}\SS\Big{\\}}p^{n}.1=q^{n(n+1)}x^{n}(T_{q,d_{1}}T_{q,d_{2}}T^{-1}_{q,d_{3}}T^{-1}_{q,d_{4}})^{n}v\\\ =1+hn(n+1)+h\left({\displaystyle\frac{\Lambda-x}{1-\Lambda}}(n+m_{1})(n+m_{2})+{\displaystyle\frac{\Lambda}{x}}{\displaystyle\frac{x-1}{1-\Lambda}}(n-m_{3})(n-m_{4})\right)+O(h^{2}).\end{array}$ (4.36) This is the desired result. ∎ In a similar way, we can compute higher order corrections for $v=\SS.1$ as $v=\exp\left(\frac{h}{1-\Lambda}C_{1}+\frac{h^{2}}{(1-\Lambda)^{2}}C_{2}+\frac{h^{3}}{(1-\Lambda)^{3}}C_{3}+O(h^{4})\right),\\\ $ (4.37) where $\begin{array}[]{l}C_{1}=m_{1}m_{2}(\Lambda-x)+\frac{\Lambda}{x}m_{3}m_{4}(x-1),\\\\[11.38109pt] C_{2}=\frac{(1-x)}{2x}\Big{[}\frac{\Lambda}{x}m_{3}m_{4}\left(x+\Lambda-(m_{3}+m_{4})(x-\Lambda)\right)-m_{1}m_{2}x\left(x+\Lambda+(m_{1}+m_{2})(x-\Lambda)\right)\Big{]},\\\\[11.38109pt] C_{3}=(x-1)(x-\Lambda)\Big{[}m_{1}m_{2}\left(-\frac{\Lambda^{2}+\Lambda+8x^{2}+3\Lambda x-x}{12(x-\Lambda)}+\frac{(m_{1}+m_{2})(\Lambda^{2}+\Lambda+\Lambda x+x-4x^{2})}{4(x-\Lambda)}\right.\\\ \left.+\frac{(m_{1}+m_{2}){}^{2}(1+\Lambda-2x)}{6}-\frac{m_{1}m_{2}(1+\Lambda+4x)}{12}\right)+m_{3}m_{4}\left(\frac{\Lambda(\Lambda x^{2}+x^{2}+8\Lambda^{2}+3\Lambda x-\Lambda^{2}x)}{12x^{3}(x-\Lambda)}\right.\\\ \left.-\frac{\Lambda(m_{3}+m_{4})(\Lambda x^{2}+x^{2}+\Lambda^{2}x+\Lambda x-4\Lambda^{2})}{4x^{3}(x-\Lambda)}+\frac{\Lambda(m_{3}+m_{4}){}^{2}(1+\Lambda-2\frac{\Lambda}{x})}{6x^{4}}-\frac{\Lambda m_{3}m_{4}(1+\Lambda+4\frac{\Lambda}{x})}{12x^{4}}\right)+\frac{\Lambda m_{1}m_{2}m_{3}m_{4}}{x}\Big{]}.\end{array}$ (4.38) ###### Remark 4.7. The quantum Seiberg-Witten curve for four dimensional $\mathcal{N}=2$ $SU(2)$ gauge theory with $N_{f}=4$ is equivalent to the quantum $P_{\rm VI}$ equation $\mathcal{D}^{4d}_{\rm SW}\Psi(\Lambda,x)=0$, where $\mathcal{D}^{4d}_{\rm SW}$ can be written as [4] $\mathcal{D}^{4d}_{\rm SW}=(1-\Lambda)\Lambda\partial_{\Lambda}+D_{\rm Heun},$ (4.39) and the Heun operator is $\displaystyle D_{\rm Heun}$ $\displaystyle=(1-\Lambda)(v-a_{1})(v-a_{2})-\frac{\Lambda}{z}(1-z)(v-\mu_{1})(v-\mu_{2})-(z-\Lambda)(v-\mu_{3})(v-\mu_{4}),$ where $v=bz\partial_{z}$ and $b^{2}=\epsilon_{1}/\epsilon_{2}$. The result of Theorem 4.6 is consistent with the correspondence of the quantum $P_{\rm VI}$ equation and $SU(2)$ Seiberg-Witten theory with $N_{f}=4$ in four dimensions. In five dimensions there are two mutually commuting Hamiltonians, one of which requires the infinite product to generate the discrete time evolution, the other is related to the conserved quantities and takes a simple expression (like the relativistic affine Toda Hamiltonian). We have explicitly seen this is the case in the decoupling limit of the hypermultiplets (the pure Yang- Mills case). In four dimensional limit these two Hamiltonians degenerate to a single Hamiltonian, which is obtained from the Seiberg-Witten curve. ## 5\. Relation to affine Laumon space ### 5.1. Factorization as a coupled system We may use the same gauge transformation as (4.1) but exchanging $T_{1}$ and $T_{2}$ in the gauge factor. By the dictionary in section 2.2, we can see this is nothing but the action of the Weyl reflection $r_{1}$ (See Appendix A). By this $r_{1}$-reflected gauge transformation, we obtain the following Hamiltonian from (4.1) by exchanging $T_{1}$ and $T_{2}$. $\displaystyle\widetilde{\mathcal{H}}:={1\over\varphi(T_{2}q^{1/2}t^{1/2}x)\varphi(T_{3}q^{1/2}t^{1/2}\Lambda x^{-1})}\cdot\mathcal{B}$ $\displaystyle\qquad\cdot{\varphi(tT_{2}T_{3}\Lambda)\varphi(qT_{1}T_{4}\Lambda)\over\varphi(-T_{1}T_{2}x)\varphi(-Q^{-1}x)\varphi(-T_{3}T_{4}Qqt\Lambda x^{-1})\varphi(-q\Lambda x^{-1})}\cdot\mathcal{B}$ $\displaystyle\qquad\cdot T_{qtQ,x}^{-1}T_{t,\Lambda}^{-1}\cdot{1\over\varphi(T_{1}q^{1/2}t^{1/2}x)\varphi(T_{4}q^{1/2}t^{1/2}\Lambda x^{-1})}.$ (5.1) It turns out that this exchange of $T_{1}$ and $T_{2}$ is better for the purpose of factorizing the original non-stationary difference equation111111The virtue of the gauge transformation (4.1) is that it cancels $\varphi(q^{1/2}t^{1/2}T_{2}x)$ appearing in (2.18) so that the total gauge factor is written in terms the double infinite product $\Phi(x)$ only, as we have seen in the beginning of Section 4.. Our main point is that one can transform the non-stationary difference equation $\widetilde{\mathcal{H}}\mathsf{V}^{(1)}=\mathsf{V}^{(1)}$ to the following coupled system; $\displaystyle\mathsf{V}^{(1)}={\Phi(qt^{-1}b_{2}/b_{4})\Phi(b_{1}/b_{3})\over\Phi(tb_{6}/b_{8})\Phi(qb_{5}/b_{7})}{1\over\varphi(-qb_{6}/G)\varphi(-G/b_{8})}\cdot\widetilde{\mathcal{B}}$ $\displaystyle\qquad\cdot{1\over\varphi(\mathsf{p}^{-2}qb_{2}/G)\varphi(\mathsf{p}^{-2}G/b_{4})}(T_{t^{1/2},x}^{-1}T_{t,\Lambda}^{-1})\mathsf{V}^{(2)},$ (5.2) $\displaystyle\mathsf{V}^{(2)}={\Phi(tb_{6}/b_{8})\Phi(qb_{5}/b_{7})\over\Phi(qb_{2}/b_{4})\Phi(tb_{1}/b_{3})}{1\over\varphi(qb_{1}/G)\varphi(G/b_{3})}\cdot\widetilde{\mathcal{B}}$ $\displaystyle\qquad\cdot{1\over\varphi(-qb_{5}/G)\varphi(-G/b_{7})}\mathsf{V}^{(1)}.$ (5.3) where $\widetilde{\mathcal{B}}:=T_{(qt^{1/2}Q)^{1/2},x}^{-1}\mathcal{B}$. Note that by making use of the equality $\displaystyle{\Phi(qt^{-1}b_{2}/b_{4})\Phi(b_{1}/b_{3})\over\Phi(tb_{6}/b_{8})\Phi(qb_{5}/b_{7})}T_{t,\Lambda}^{-1}{\Phi(tb_{6}/b_{8})\Phi(qb_{5}/b_{7})\over\Phi(qb_{2}/b_{4})\Phi(tb_{1}/b_{3})}$ $\displaystyle=\varphi(b_{6}/b_{8})\varphi(qt^{-1}b_{5}/b_{7})T_{t,\Lambda}^{-1}=\varphi(tT_{2}T_{3}\Lambda)\varphi(qT_{1}T_{4}\Lambda)T_{t,\Lambda}^{-1}.$ (5.4) we have called back the double infinite product $\Phi(z)$, which we once eliminated to reveal the relation to the $qq$-Painlevé VI equation, to factorize the original equation as a coupled system. The possibility of such a factorization was already suggested in the proof of Conjecture 2.4 in the special case of the “Macdonald” limit [44]. We believe this factorization is a significant step towards a general proof of Conjecture 2.4. By the dictionary in section 2.2 and Definition A.5 of the variables $b_{i}$, it is straightforward to check the matching of parameters $qt^{-1}b_{2}/b_{4},b_{1}/b_{3},tb_{6}/b_{8}$ and $qb_{5}/b_{7}$. Note that the parameters $b_{i}$ involve neither $x$ nor $\Lambda$. On the other hand, $G$ is a monomial in $x$ and $\Lambda$. But it is also easy to see that matcing for $-qb_{6}/G,-G/b_{8},-qb_{5}/G$ and $-G/b_{7}$. For remaining parameters we have to take the commutation with the shift operators into account. Namely the parameters $\mathsf{p}^{-2}qb_{2}/G$ and $\mathsf{p}^{-2}G/b_{4}$ are affected by $T_{(qt^{1/2}Q)^{1/2},x}^{-1}$ and for $qb_{1}/G$ and $G/b_{3}$ the action of $T_{t^{1/2},x}^{-1}T_{t,\Lambda}^{-1}$ is also involved. To rewrite the coupled system (5.1) and (5.1) to more symmetric one in $\mathsf{V}^{(1)}$ and $\mathsf{V}^{(2)}$, let us introduce the following shift operator $\widetilde{T}_{\mathsf{p},b}$ which commutes with $\widetilde{\mathcal{B}}$. $\widetilde{T}_{\mathsf{p},b}:=T_{-t^{-1/4},x}T_{t^{-1/2},\Lambda}:b\mapsto(-\mathsf{p}^{-1}b_{1},-\mathsf{p}^{-1}b_{2},-\mathsf{p}b_{3},-\mathsf{p}b_{4},-\mathsf{p}^{-1}b_{5},-\mathsf{p}^{-1}b_{6},-\mathsf{p}b_{7},-\mathsf{p}b_{8}),$ (5.5) where $\mathsf{p}=e^{\delta}=a_{0}a_{1}a_{2}^{2}a_{3}^{2}a_{4}a_{5}=t^{1/4}$. Note that $G$ is invariant under $\widetilde{T}_{\mathsf{p},b}$. For the transformation (5.5) of the parameters $b_{i}$, the shift operator $T_{t^{-1/2},\Lambda}$ is enough, but to make $G$ invariant up to sign we need the combination with $T_{-t^{-1/4},x}$. We also note that $(\widetilde{T}_{\mathsf{p},b})^{2}=T~{}(\hbox{the discrete time evolution})$ as far as the $b$ variables are concerned (see Lemma A.8). Then we have ###### Proposition 5.1. We can rewrite (5.1) and (5.1) as follows; $\displaystyle\mathsf{V}^{(1)}$ $\displaystyle={\Phi(qt^{-1}b_{2}/b_{4})\Phi(b_{1}/b_{3})\over\Phi(tb_{6}/b_{8})\Phi(qb_{5}/b_{7})}{1\over\varphi(-qb_{6}/G)\varphi(-G/b_{8})}$ $\displaystyle\qquad\cdot(\widetilde{\mathcal{B}}\cdot\widetilde{T}_{\mathsf{p},b}){1\over\varphi(-\mathsf{p}^{-1}qb_{2}/G)\varphi(-\mathsf{p}^{-1}G/b_{4})}\widetilde{T}_{\mathsf{p},b}\cdot\mathsf{V}^{(2)},$ (5.6) $\displaystyle\widetilde{T}_{\mathsf{p},b}\cdot\mathsf{V}^{(2)}$ $\displaystyle={\Phi(\mathsf{p}^{-2}tb_{6}/b_{8})\Phi(\mathsf{p}^{-2}qb_{5}/b_{7})\over\Phi(\mathsf{p}^{-2}qb_{2}/b_{4})\Phi(\mathsf{p}^{-2}tb_{1}/b_{3})}{1\over\varphi(-\mathsf{p}^{-1}qb_{1}/G)\varphi(-\mathsf{p}^{-1}G/b_{3})}$ $\displaystyle\cdot(\widetilde{\mathcal{B}}\cdot\widetilde{T}_{\mathsf{p},b})\cdot{1\over\varphi(-qb_{5}/G)\varphi(-G/b_{7})}\mathsf{V}^{(1)}.$ (5.7) The coupled system is gauge equivalent to the non-stationary difference equation (2.14). Recall that the discrete time evolution $T$ of Painlevé VI equation is a translation element in the extended affine Weyl group of $D_{5}^{(1)}$. It is remarkable that $T$ allows a square root; $T:=r_{2}r_{1}r_{0}r_{2}\sigma_{01}r_{3}r_{4}r_{5}r_{3}\sigma_{45}=(r_{2}r_{1}r_{0}r_{2}\sigma_{01}\tau)(r_{2}r_{1}r_{0}r_{2}\sigma_{01}\tau).$ (5.8) In fact the factorization into (5.1) and (5.1) is not unrelated to the existence of the square root $T^{1/2}:=(r_{2}r_{1}r_{0}r_{2}\sigma_{01}\tau)$, which acts on the $b$ variables as follows; $T^{1/2}=(r_{2}r_{1}r_{0}r_{2}\sigma_{01}\tau):b\mapsto(b_{6},b_{5},b_{8},b_{7},\mathsf{p}^{-2}b_{2},\mathsf{p}^{-2}b_{1},\mathsf{p}^{2}b_{4},\mathsf{p}^{2}b_{3}).$ (5.9) We define an operator $X:b\mapsto(\mathsf{p}b_{6},\mathsf{p}b_{5},\mathsf{p}^{-1}b_{8},\mathsf{p}^{-1}b_{7},\mathsf{p}^{-1}b_{2},\mathsf{p}^{-1}b_{1},\mathsf{p}b_{4},\mathsf{p}b_{3}),$ (5.10) and assume that $X$ does not act on $(F,G)$. Then we have ###### Lemma 5.2. The action (5.9) on the parameters $b_{i}$ is represented by $T^{1/2}=(-1)\cdot X\cdot\widetilde{T}_{\mathsf{p},b},$ (5.11) where $(-1)$ is the overall sign flip of $b_{i}$. Two functions $\mathsf{V}^{(1)}$ and $\mathsf{V}^{(2)}$ are related by $X\mathsf{V}^{(1)}=\widetilde{T}_{\mathsf{p},b}\cdot\mathsf{V}^{(2)}.$ (5.12) Hence, the coupled system in the form of Prop.5.1 is quite natural from the viewpoint of the $qq$-Painlevé VI equations, because the second equation (5.1) is obtained by applying $X$ to (5.1). ### 5.2. Instanton counting with a surface defect Now we want to point out that solutions to the coupled system (5.1) and (5.1) are given by the instanton partition function of the affine Laumon space. In fact we have already mentioned that the gauge transformation introduced in the beginning of Section 4 is a five dimensional uplift of the gauge transformation from the Higgsed quiver gauge theory to the gauge theory with a surface defect. A torus action on the affine Laumon space is induced by the standard torus action on $\mathbb{P}^{1}\times\mathbb{P}^{1}$. The fixed points of the torus action on the affine Laumon space of type $A_{r}^{(1)}$ are labelled by $(r+1)$-tuples of partitions [18]. ###### Definition 5.3. Set $\displaystyle[u;q]_{n}=u^{-n/2}q^{-n(n-1)/4}(u;q)_{n}$ $\displaystyle=(u^{-1/2}-u^{1/2})(q^{-1/2}u^{-1/2}-q^{1/2}u^{1/2})\cdots(q^{-(n-1)/2}u^{-1/2}-q^{(n-1)/2}u^{1/2}).$ For a pair $(\lambda,\mu)$ of partitions, the $\mathbb{Z}_{N}$ orbifolded Nekrasov factor with color $k$ is121212We associate a $\sinh$ factor with each monomial term in the equivariant character. $\displaystyle{\mathsf{N}}^{(k|N)}_{\lambda,\mu}(u|q,\kappa)={\mathsf{N}}^{(k)}_{\lambda,\mu}(u|q,\kappa)$ (5.13) $\displaystyle=$ $\displaystyle\prod_{j\geq i\geq 1\atop j-i\equiv k\,\,({\rm mod}\,N)}[uq^{-\mu_{i}+\lambda_{j+1}}\kappa^{-i+j};q]_{\lambda_{j}-\lambda_{j+1}}\cdot\prod_{\beta\geq\alpha\geq 1\atop\beta-\alpha\equiv-k-1\,\,({\rm mod}\,N)}[uq^{\lambda_{\alpha}-\mu_{\beta}}\kappa^{\alpha-\beta-1};q]_{\mu_{\beta}-\mu_{\beta+1}}.$ Note that the equivariant parameters of the torus action on $\mathbb{P}^{1}\times\mathbb{P}^{1}$ are not $(q,t)$, but $(q,\kappa)$. We will substitute $\kappa=t^{-\frac{1}{2}}$ later131313The square root comes from the $\mathbb{Z}_{2}$ orbifolding.. From the equivariant character evaluated at each fixed point of the affine Laumon space of type $A_{1}^{(1)}$ [18], we obtain $\displaystyle f(u_{1},u_{2};v_{1},v_{2};w_{1},w_{2}|x_{1},x_{2}|q,\kappa)=f\left(\left.\left.\begin{array}[]{c}u_{1},u_{2}\\\ v_{1},v_{2}\\\ w_{1},w_{2}\end{array}\right|x_{1},x_{2}\right|q,\kappa\right)$ (5.17) $\displaystyle=\sum_{\lambda^{(1)},\lambda^{(2)}\in\mathsf{P}}\prod_{i,j=1}^{2}{\mathsf{N}^{(j-i|2)}_{\emptyset,\lambda^{(j)}}(u_{i}/v_{j}|q,\kappa)\mathsf{N}^{(j-i|2)}_{\lambda^{(i)},\emptyset}(v_{i}/w_{j}|q,\kappa)\over\mathsf{N}^{(j-i|2)}_{\lambda^{(i)},\lambda^{(j)}}(v_{i}/v_{j}|q,\kappa)}\cdot x_{1}^{|\lambda^{(1)}|_{o}+|\lambda^{(2)}|_{e}}x_{2}^{|\lambda^{(1)}|_{e}+|\lambda^{(2)}|_{o}},$ (5.18) where $\emptyset$ denotes the empty partition and for a partition $\lambda=(\lambda_{1}\geq\lambda_{2}\geq\cdots)$, we set $|\lambda|_{o}\mathbin{:=}\sum_{k\geq 1}\lambda_{2k-1},\qquad|\lambda|_{e}\mathbin{:=}\sum_{k\geq 1}\lambda_{2k}.$ (5.19) Note that the function $f(u_{1},u_{2};v_{1},v_{2};w_{1},w_{2}|x_{1},x_{2}|q,\kappa)$ is invariant under the overall scaling of the equivariant parameters $(u_{1},u_{2};v_{1},v_{2};w_{1},w_{2})$. The partition function (5.17) is a five dimensional uplift of the instanton partition function which is given, for example, in [4]. The parameters $(v_{1},v_{2})$ are the Coulomb moduli of $U(2)$ gauge theory, or the equivariant parameters of the Cartan subgroup $U(1)\times U(1)\subset U(2)$. The parameters $(u_{1},u_{2})$ and $(w_{1},w_{2})$ are exponentiated mass parameters of the hypermultiplets in the fundamental and the anti-fundamental representations. They are also regarded as equivariant parameters for the flavor symmetry. The expansion parameters are parametrized as $x_{1}=x,x_{2}=\Lambda/x$, where $x$ counts the monopole number (the first Chern number of the $U(1)$ connection on the defect), while $\Lambda$ counts the instanton number (the second Chern number). When $\Lambda=0$, the terms with $|\lambda^{(1)}|_{e}+|\lambda^{(2)}|_{o}\neq 0$ do not contribute to the partition function. This means the sum in (5.17) is restricted to $\lambda^{(1)}=(m)$ (a partition with a single row) and $\lambda^{(2)}=\emptyset$. Physically this corresponds to the topological sector with instanton number zero. As we see in Appendix B, the partition function is given by the Heine’s $q$-hypergeometric series. The six parameters $(u_{1},u_{2}),(w_{1},w_{2})$ and $(x_{1},x_{2})$ are “external” spectral parameters which correspond to the (independent) dynamical variables on Painlevé side. On the other hand the parameters $(v_{1},v_{2})$ are “internal” parameters or the loop parameters. The function (5.17) should be compared with the non-stationary Ruijsenaars function [45]; $\displaystyle f^{\widehat{\mathfrak{gl}}_{N}}(x,p|s,\kappa|q,t)=\sum_{\lambda^{(1)},\ldots,\lambda^{(N)}\in{\mathsf{P}}}\prod_{i,j=1}^{N}{{\mathsf{N}}^{(j-i|N)}_{\lambda^{(i)},\lambda^{(j)}}(ts_{j}/s_{i}|q,\kappa)\over{\mathsf{N}}^{(j-i|N)}_{\lambda^{(i)},\lambda^{(j)}}(s_{j}/s_{i}|q,\kappa)}\cdot\prod_{\beta=1}^{N}\prod_{\alpha\geq 1}(px_{\alpha+\beta}/x_{\alpha+\beta-1})^{\lambda^{(\beta)}_{\alpha}},$ (5.20) for $N=2$. Both functions come from the affine Laumon space of type $A_{1}^{(1)}$. The non-stationary Ruijsenaars function (5.20) corresponds to the theory with an adjoint matter, or the tangent bundle overt the affine Laumon space, while the partition function (5.17) is for the theory with four matter hypermultiplets in the (anti-)fundamental representation, or the tautological bundle. In the AGT correspondence, the former is identified with the conformal block on a punctured torus and the latter on $\mathbb{P}^{1}$ with four punctures. In the mass decoupling limit, both functions are conjectured to give solutions to the non-stationary affine Toda equation [45]. ###### Conjecture 5.4. The partition function (5.17) gives a solution to the coupled system (5.1) and (5.1) by the following specialization of parameters; $\displaystyle\mathcal{F}^{(1)}=f\left(\left.\left.\begin{array}[]{ll}q^{1/2}b_{4}/b_{8},&q^{1/2}b_{6}/b_{2}\\\ (\mathsf{p}^{2}Q)^{-1/2},&(\mathsf{p}^{2}Q)^{1/2}\\\ q^{-1/2}b_{2}/b_{5},&q^{-1/2}b_{7}/b_{4}\end{array}\right|q^{1/2}\mathsf{p}^{-1}\mathsf{t}G^{-1},q^{-1/2}\mathsf{p}^{-1}\mathsf{t}G\right|q,t^{-1/2}\right),$ (5.24) $\displaystyle\mathcal{F}^{(2)}=f\left(\left.\left.\begin{array}[]{ll}q^{1/2}b_{4}/b_{8},&q^{1/2}b_{6}/b_{2}\\\ (\mathsf{p}^{2}Q)^{-1/2},&(\mathsf{p}^{2}Q)^{1/2}\\\ q^{-1/2}b_{1}/b_{6},&q^{-1/2}b_{8}/b_{3}\end{array}\right|-q^{1/2}\mathsf{t}G^{-1},-q^{-1/2}\mathsf{t}G\right|q,t^{-1/2}\right).$ (5.28) It is remarkable that the only difference between $\mathcal{F}^{(1)}$ and $\mathcal{F}^{(2)}$ is the exchanges of $(b_{1},b_{2}),(b_{3},b_{4}),(b_{5},b_{6})$ and $(b_{7},b_{8})$ (see Figure 2) in the specialization of $w_{i}$ and the scaling $-\mathsf{p}$ of $\mathsf{t}$. Note that these exchanges of four pairs of $b_{i}$ are nothing but the action of the Weyl reflections $r_{4},r_{5},r_{1}$ and $r_{0}$, respectively. We can see that the specializations (5.24) and (5.28) are consistent with the relation (5.12). In fact under the action of $X$ $\displaystyle\frac{b_{4}}{b_{8}}\mapsto\mathsf{p}^{-2}\frac{b_{7}}{b_{3}}=\frac{b_{4}}{b_{8}},\qquad\frac{b_{2}}{b_{6}}\mapsto\mathsf{p}^{-2}\frac{b_{1}}{b_{5}}=\frac{b_{2}}{b_{6}},$ $\displaystyle\frac{b_{2}}{b_{5}}\mapsto\mathsf{p}^{2}\frac{b_{5}}{b_{2}}=\frac{b_{1}}{b_{6}},\qquad\frac{b_{7}}{b_{4}}\mapsto\mathsf{p}^{2}\frac{b_{4}}{b_{7}}=\frac{b_{8}}{b_{3}},$ where we have used the constraints (A.1). On the other hand, these ratios are invariant by the action of $\widetilde{T}_{\mathsf{p},b}$. Finally $\widetilde{T}_{\mathsf{p},b}$ generates the square root of the discrete time shift $\mathsf{t}\mapsto-\mathsf{p}^{-1}\mathsf{t}$. In terms of the variables on the gauge theory side the specialization of parameters is given as follows; $\displaystyle\mathcal{F}^{(1)}=f\left(\left.\left.\begin{array}[]{ll}q^{1/2}t^{1/2}Q^{1/2}T_{3},&q^{1/2}Q^{-1/2}T_{1}^{-1}\\\ Q^{-1/2}&t^{1/2}Q^{1/2}\\\ q^{-1/2}t^{1/2}Q^{1/2}T_{2},&q^{-1/2}Q^{-1/2}T_{4}^{-1}\end{array}\right|-t^{1/2}T_{1}^{1/2}T_{2}^{1/2}x,-t^{1/2}T_{3}^{1/2}T_{4}^{1/2}\Lambda/x\right|q,t^{-1/2}\right),$ (5.32) $\displaystyle\mathcal{F}^{(2)}=f\left(\left.\left.\begin{array}[]{ll}q^{1/2}t^{1/2}Q^{1/2}T_{3},&q^{1/2}Q^{-1/2}T_{1}^{-1}\\\ Q^{-1/2},&t^{1/2}Q^{1/2}\\\ q^{-1/2}t^{1/2}Q^{-1/2}T_{2}^{-1},&q^{-1/2}tQ^{1/2}T_{4}\end{array}\right|t^{3/4}T_{1}^{1/2}T_{2}^{1/2}x,t^{3/4}T_{3}^{1/2}T_{4}^{1/2}\Lambda/x\right|q,t^{-1/2}\right),$ (5.36) where by making use of the scaling symmetry of the partition function (5.17), we have made the overall scaling of parameters by $\mathsf{p}$. One can check that $\mathcal{F}^{(2)}$ is invariant under the exchange of $T_{1}$ and $T_{2}$. We have examined our conjecture in several cases. The results are summarized in Appendix B. Thus we can formulate Conjecture 2.4 in terms of the instanton partition functions from the affine Laumon space. According to the (four dimensional) AGT correspondence, the partition functions coming from the affine Laumon space should be identified with the conformal blocks of the current algebra. In the present case it is the affine algebra $A_{1}^{(1)}=\widehat{\mathfrak{sl}}_{2}$. In the original formulation in [44], the five dimensional Nekrasov partition function are regarded as a conformal block of the deformed Virasoro algebra. In our formulation it is natural to expect that the coupled system of non-stationary difference equations defines a conformal block of the quantum deformation of $A_{1}^{(1)}=\widehat{\mathfrak{sl}}_{2}$, namely the quantum affine algebra $U_{q}(\widehat{\mathfrak{sl}}_{2})$. The advantage of $U_{q}(\widehat{\mathfrak{sl}}_{2})$ to the deformed Virasoro algebra is that it is a quantum group (in particular we have a coproduct), while the latter is not. ###### Acknowledgements. We would like to thank H.Hayashi, A.N.Kirillov, G.Kuroki and H.Nakajima for useful discussions. Our work is supported in part by Grants-in-Aid for Scientific Research (Kakenhi); 18K03274 (H.K.), 21K03180 (R.O.), 19K03512 (J.S.), 19K03530 (J.S.) and 22H01116 (Y.Y.). The work of R.O. was partly supported by Osaka Central Advanced Mathematical Institute: MEXT Joint Usage/Research Center on Mathematics and Theoretical Physics JPMXP0619217849, and the Research Institute for Mathematical Sciences, an International Joint Usage/Research Center located in Kyoto University. ## Appendix A Difference analogue of Painlevé VI equation ### A.1. Affine Weyl group and Bäcklund transformation Let $A=A(D^{(1)}_{5})=(a_{ij})_{i,j=0}^{5}$ be the generalized Cartan matrix of type $D^{(1)}_{5}$ associated with the Dynkin diagram in Fig. 1. $\bullet$$\bullet$$\bullet$$\bullet$$\bullet$$\bullet$$\scriptstyle 0$$\scriptstyle 1$$\scriptstyle 2$$\scriptstyle 3$$\scriptstyle 5$$\scriptstyle 4$$\sigma_{01}\,\,\,\updownarrow$$\updownarrow\,\,\,\sigma_{45}$$\longleftrightarrow$$\tau$ Figure 1. Dynkin diagram of $D^{(1)}_{5}$ and its automorphism Fix a realization $(\mathfrak{h},\Pi,\Pi^{\vee})$ of $A$, where $\Pi=\\{\alpha_{0},\ldots,\alpha_{5}\\}\subset\mathfrak{h}^{*}$ denotes the set of simple roots, $\Pi^{\vee}=\\{\alpha_{0}^{\vee},\ldots,\alpha_{5}^{\vee}\\}\subset\mathfrak{h}$ the set of simple coroots, with $\Pi$ and $\Pi^{\vee}$ being linearly independent. We have $\langle\alpha_{i}^{\vee},\alpha_{j}\rangle=a_{ij}$ $(i,j=0,\ldots,5)$. Let $Q=\sum_{i=0}^{5}\mathbb{Z}\alpha_{i}$ be the root lattice. Denote by $\Delta,\Delta_{+}$ and $\Delta_{-}$ the sets of all roots, positive and negative roots respectively. We have $\Delta=\Delta_{+}\cup\Delta_{-}$ (a disjoint union). The center of $\mathfrak{g}=\mathfrak{g}(D^{(1)}_{5})$ is 1-dimensional and is spanned by the canonical central element $K=\alpha_{0}^{\vee}+\alpha_{1}^{\vee}+2\alpha_{2}^{\vee}+2\alpha_{3}^{\vee}+\alpha_{4}^{\vee}+\alpha_{5}^{\vee}.$ Denote by $\delta\in Q$ the null root $\delta=\alpha_{0}+\alpha_{1}+2\alpha_{2}+2\alpha_{3}+\alpha_{4}+\alpha_{5}$. Let $d\in\mathfrak{h}$ be the scaling element (defined up to a summand proportional to $K$) satisfying $\langle\alpha_{i},d\rangle=0$ for $i=1,\ldots,5$ and $\langle\alpha_{0},d\rangle=1$. The elements $\alpha_{0}^{\vee},\ldots,\alpha_{5}^{\vee},d$ form a basis of $\mathfrak{h}$. We have $\mathfrak{g}=[\mathfrak{g},\mathfrak{g}]+\mathbb{C}d$, $\mathfrak{h}=\sum_{i=0}^{5}\mathbb{C}\alpha_{i}^{\vee}+\mathbb{C}d$. Define the element $\Lambda_{0}\in\mathfrak{h}^{*}$ by $\langle\Lambda_{0},\alpha_{i}^{\vee}\rangle=\delta_{0i}$ for $i=0,\ldots,5$ and $\langle\Lambda_{0},d\rangle=0$. The elements $\alpha_{0},\ldots,\alpha_{5},\Lambda_{0}$ form a basis of $\mathfrak{h}^{*}$ and we have $\mathfrak{h}^{*}=\sum_{i=0}^{5}\mathbb{C}\alpha_{i}+\mathbb{C}\Lambda_{0}$. The affine Weyl group $W=W(D^{(1)}_{5})$ is defined to be the group generated by the fundamental reflections $r_{0},r_{1},\ldots,r_{5}$ which act on $\mathfrak{h}^{*}$ by $\displaystyle r_{i}(\lambda)=\lambda-\langle\lambda,\alpha_{i}^{\vee}\rangle\alpha_{i}\qquad(\lambda\in\mathfrak{h}^{*}).$ Let $a_{i}$ $(i=0,\ldots,5)$ be the formal exponentials $a_{i}=e^{\alpha_{i}}$. Write $a=(a_{0},\ldots,a_{5})$ for short. Denote by $\mathbb{K}=\mathbb{C}(a)$ the field of rational functions in $a$. Define the actions of the generators $r_{i}\in W$ on $\mathbb{K}$ by setting the rules $\displaystyle r_{i}\cdot a_{j}=a_{j}a_{i}^{-a_{ij}}=e^{\alpha_{j}-a_{ij}\alpha_{i}}\qquad(0\leq i,j\leq 5),$ and extending them as ring homomorphisms. It is clear that these actions are compatible with the group structure, namely they satisfy the Coxeter relations: $r_{i}^{2}={\rm id}$ $(a_{ij}=2)$, $r_{i}r_{j}r_{i}=r_{j}r_{i}r_{j}$ $(a_{ij}=-1)$, and $r_{i}r_{j}=r_{j}r_{i}$ $(a_{ij}=0)$. We regard $W$ as a group of birational isomorphisms of $\mathbb{K}$. Let $\sigma_{01},\sigma_{45}$ and $\tau$ be the automorphisms shown in Fig. 1, namely $\displaystyle\sigma_{01}:(0,1,2,3,4,5)\mapsto(1,0,2,3,4,5),$ $\displaystyle\sigma_{45}:(0,1,2,3,4,5)\mapsto(0,1,2,3,5,4),$ $\displaystyle\tau:(0,1,2,3,4,5)\mapsto(5,4,3,2,1,0).$ Let $\widetilde{W}$ denotes the extended affine Weyl group generated by $W$ together with $\sigma_{01},\sigma_{45}$ and $\tau$. Let $f$ and $g$ be independent indeterminates, and consider the rational function field $\mathbb{K}(f,g)=\mathbb{C}(a)(f,g)$. ###### Definition A.1. Define the actions of the generators $r_{0},\ldots,r_{5},\sigma_{01},\sigma_{45},\tau\in\widetilde{W}$ on $\mathbb{K}(f,g)$ by setting the rules $\displaystyle r_{i}\cdot a_{j}=a_{j}a_{i}^{-a_{ij}}\qquad(0\leq i,j\leq 5),$ $\displaystyle\sigma_{01}\cdot a_{i}=(a_{\sigma_{01}(i)})^{-1},\quad\sigma_{45}\cdot a_{i}=(a_{\sigma_{45}(i)})^{-1},\quad\tau\cdot a_{i}=(a_{\tau(i)})^{-1}\qquad(0\leq i\leq 5),$ $\displaystyle r_{i}\cdot f=f,\qquad r_{i}\cdot g=g\qquad(i\neq 2,3),$ $\displaystyle r_{2}\cdot f=f{a_{0}a_{1}^{-1}g+a_{2}^{2}\over a_{0}a_{1}^{-1}a_{2}^{2}g+1},\qquad r_{2}\cdot g=g,\qquad r_{3}\cdot f=f,\qquad r_{3}\cdot g={a_{3}^{2}a_{4}a_{5}^{-1}f+1\over a_{4}a_{5}^{-1}f+a_{3}^{2}}g,$ $\displaystyle\sigma_{01}\cdot f=f^{-1},\qquad\sigma_{01}\cdot g=g,\qquad\sigma_{45}\cdot f=f,\qquad\sigma_{45}\cdot g=g^{-1},$ $\displaystyle\tau\cdot f=g,\qquad\tau\cdot g=f,$ and extending them as ring homomorphisms. ###### Proposition A.2. The actions are compatible with the group structure of the extended affine Weyl group $\widetilde{W}$. Namely they satisfy the Coxeter relations: $\displaystyle r_{i}^{2}={\rm id}\qquad(\mbox{if }a_{ij}=2),$ $\displaystyle r_{i}r_{j}r_{i}=r_{j}r_{i}r_{j}\qquad(\mbox{if }a_{ij}=-1),\qquad r_{i}r_{j}=r_{j}r_{i}\qquad(\mbox{if }a_{ij}=0),$ $\displaystyle\sigma_{01}^{2}=\sigma_{45}^{2}=\tau^{2}={\rm id},\qquad\sigma_{01}\sigma_{45}=\sigma_{45}\sigma_{01},\qquad\sigma_{01}\tau=\tau\sigma_{45},$ $\displaystyle\sigma_{01}r_{0}=r_{1}\sigma_{01},\qquad\sigma_{01}r_{i}=r_{i}\sigma_{01}\qquad(i\neq 0,1),$ $\displaystyle\sigma_{45}r_{4}=r_{5}\sigma_{45},\qquad\sigma_{45}r_{i}=r_{i}\sigma_{45}\qquad(i\neq 4,5),$ $\displaystyle\tau r_{i}=r_{5-i}\tau.$ ### A.2. Difference analogue of Painlevé VI equation Set for simplicity of display that $\displaystyle\mathsf{p}=a_{0}a_{1}a_{2}^{2}a_{3}^{2}a_{4}a_{5}=e^{\delta},\qquad\mathsf{t}=a_{3}^{2}a_{4}a_{5}.$ Recall that the element $\displaystyle T=r_{2}r_{1}r_{0}r_{2}\sigma_{01}r_{3}r_{4}r_{5}r_{3}\sigma_{45}\in\widetilde{W},$ acts on $\mathfrak{h}^{*}$ as a translation. Namely, we have $\displaystyle T:(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5})\mapsto(a_{0},a_{1},\mathsf{p}a_{2},\mathsf{p}^{-1}a_{3},a_{4},a_{5}).$ Write $\overline{f}=T\cdot f$ and $\underline{g}=T^{-1}\cdot g$ for short. ###### Proposition A.3. We have the difference analogue of the Painlevé VI equation; $\displaystyle f\overline{f}$ $\displaystyle=\mathsf{p}^{2}\mathsf{t}^{-2}{g+\mathsf{t}\mathsf{p}^{-1}a_{1}^{2}\over g+\mathsf{t}^{-1}\mathsf{p}a_{0}^{2}}{g+\mathsf{t}\mathsf{p}^{-1}a_{1}^{-2}\over g+\mathsf{t}^{-1}\mathsf{p}a_{0}^{-2}},$ $\displaystyle g\underline{g}$ $\displaystyle=\mathsf{t}^{-2}{f+\mathsf{t}a_{4}^{2}\over f+\mathsf{t}^{-1}a_{5}^{2}}{f+\mathsf{t}a_{4}^{-2}\over f+\mathsf{t}^{-1}a_{5}^{-2}}.$ ###### Proposition A.4. The translation $T$ admits the symmetry of the type $D^{(1)}_{4}$ in the following sense. (1) The translation $T$ commutes with the action of the subgroup $\langle r_{0},r_{1},r_{2}r_{3}r_{2}=r_{3}r_{2}r_{3},r_{4},r_{5}\rangle\simeq W(D^{(1)}_{4})$ of $W(D^{(1)}_{5})$, (2) We have $\sigma_{01}T=T\sigma_{01}$, $\sigma_{45}T=T\sigma_{45}$ and $\tau T=T^{-1}\tau$. $\mathbb{P}^{1}\times\\{0\\}$$\mathbb{P}^{1}\times\\{\infty\\}$$\\{0\\}\times\mathbb{P}^{1}$$\\{\infty\\}\times\mathbb{P}^{1}$$\times$$\times$$\times$$\times$$\times$$\times$$\times$$\times$$-b_{1}$$-b_{2}$$-b_{3}$$-b_{4}$$-b_{5}$$-b_{6}$$-b_{7}$$-b_{8}$ Figure 2. Space of initial data for Painlevé VI (8 points blow up of $\mathbb{P}^{1}\times\mathbb{P}^{1}$) ###### Definition A.5. Introduce the following variables: $\displaystyle b_{1}=\mathsf{t}a_{4}^{-2}={a_{3}^{2}a_{5}\over a_{4}},\quad b_{2}=\mathsf{t}a_{4}^{2}=a_{3}^{2}a_{4}^{3}a_{5},\quad b_{3}=\mathsf{t}^{-1}a_{5}^{2}={a_{5}\over a_{3}^{2}a_{4}},\quad b_{4}=\mathsf{t}^{-1}a_{5}^{-2}={1\over a_{3}^{2}a_{4}a_{5}^{3}},$ $\displaystyle b_{5}=\mathsf{t}\mathsf{p}^{-1}a_{1}^{2}={a_{1}\over a_{0}a_{2}^{2}},\quad b_{6}=\mathsf{t}\mathsf{p}^{-1}a_{1}^{-2}={1\over a_{0}a_{1}^{3}a_{2}^{2}},\quad b_{7}=\mathsf{t}^{-1}\mathsf{p}a_{0}^{-2}={a_{1}a_{2}^{2}\over a_{0}},\quad b_{8}=\mathsf{t}^{-1}\mathsf{p}a_{0}^{2}=a_{0}^{3}a_{1}a_{2}^{2}.$ The parameters $b_{i}$ are not independent and they satisfy the constraints; $b_{1}b_{2}b_{3}b_{4}=1,\qquad b_{5}b_{6}b_{7}b_{8}=1,\qquad b_{1}b_{2}b_{7}b_{8}=\mathsf{p}^{2},\qquad b_{3}b_{4}b_{5}b_{6}=\mathsf{p}^{-2}.$ (A.1) Regarding $\mathsf{p}$ as a new parameter, we have six independent parameters. The root variables are expressed as the ratios of $b_{i}$ (see Figure 1); $\displaystyle a_{0}^{4}=\frac{b_{8}}{b_{7}},\qquad a_{1}^{4}=\frac{b_{5}}{b_{6}},\qquad a_{2}^{4}=\frac{b_{7}}{b_{5}},\qquad a_{3}^{4}=\frac{b_{1}}{b_{3}},\qquad a_{4}^{4}=\frac{b_{2}}{b_{1}},\qquad a_{5}^{4}=\frac{b_{3}}{b_{4}}.$ ###### Corollary A.6. We have $\displaystyle f\overline{f}=b_{7}b_{8}{g+b_{5}\over g+b_{7}}{g+b_{6}\over g+b_{8}},\qquad g\underline{g}=b_{3}b_{4}{f+b_{1}\over f+b_{3}}{f+b_{2}\over f+b_{4}}.$ ### A.3. Tables of the action of $\widetilde{W}$ on $\mathbb{K}(f,g)$ ###### Lemma A.7. Write $a=(a_{0},a_{1},a_{2},a_{3},a_{4},a_{5})$ for short. We have $\displaystyle r_{0}$ $\displaystyle:(a,f,g)\mapsto(a_{0}^{-1},a_{1},a_{0}a_{2},a_{3},a_{4},a_{5},f,g),$ $\displaystyle r_{1}$ $\displaystyle:(a,f,g)\mapsto(a_{0},a_{1}^{-1},a_{1}a_{2},a_{3},a_{4},a_{5},f,g),$ $\displaystyle r_{2}$ $\displaystyle:(a,f,g)\mapsto(a_{0}a_{2},a_{1}a_{2},a_{2}^{-1},a_{2}a_{3},a_{4},a_{5},f{a_{0}a_{1}^{-1}g+a_{2}^{2}\over a_{0}a_{1}^{-1}a_{2}^{2}g+1},g),$ $\displaystyle r_{3}$ $\displaystyle:(a,f,g)\mapsto(a_{0},a_{1},a_{2}a_{3},a_{3}^{-1},a_{3}a_{4},a_{3}a_{5},f,{a_{3}^{2}a_{4}a_{5}^{-1}f+1\over a_{4}a_{5}^{-1}f+a_{3}^{2}}g),$ $\displaystyle r_{4}$ $\displaystyle:(a,f,g)\mapsto(a_{0},a_{1},a_{2},a_{3}a_{4},a_{4}^{-1},a_{5},f,g),$ $\displaystyle r_{5}$ $\displaystyle:(a,f,g)\mapsto(a_{0},a_{1},a_{2},a_{3}a_{5},a_{4},a_{5}^{-1},f,g),$ $\displaystyle\sigma_{01}$ $\displaystyle:(a,f,g)\mapsto(a_{1}^{-1},a_{0}^{-1},a_{2}^{-1},a_{3}^{-1},a_{4}^{-1},a_{5}^{-1},f^{-1},g),$ $\displaystyle\sigma_{45}$ $\displaystyle:(a,f,g)\mapsto(a_{0}^{-1},a_{1}^{-1},a_{2}^{-1},a_{3}^{-1},a_{5}^{-1},a_{4}^{-1},f,g^{-1}),$ $\displaystyle\tau$ $\displaystyle:(a,f,g)\mapsto(a_{5}^{-1},a_{4}^{-1},a_{3}^{-1},a_{2}^{-1},a_{1}^{-1},a_{0}^{-1},g,f),$ $\displaystyle T$ $\displaystyle:a\mapsto(a_{0},a_{1},\mathsf{p}a_{2},\mathsf{p}^{-1}a_{3},a_{4},a_{5}).$ ###### Lemma A.8. Write $b=(b_{1},b_{2},b_{3},b_{4},b_{5},b_{6},b_{7},b_{8})$ for short. We have $\displaystyle r_{0}$ $\displaystyle:(b,f,g)\mapsto(b_{1},b_{2},b_{3},b_{4},b_{5},b_{6},b_{8},b_{7},f,g),$ $\displaystyle r_{1}$ $\displaystyle:(b,f,g)\mapsto(b_{1},b_{2},b_{3},b_{4},b_{6},b_{5},b_{7},b_{8},f,g),$ $\displaystyle r_{2}$ $\displaystyle:(b,f,g)\mapsto(b_{1}\sqrt{b_{7}\over b_{5}},b_{2}\sqrt{b_{7}\over b_{5}},b_{3}\sqrt{b_{5}\over b_{7}},b_{4}\sqrt{b_{5}\over b_{7}},b_{7},b_{6},b_{5},b_{8},f\sqrt{b_{5}\over b_{7}}\frac{g+b_{7}}{g+b_{5}},g),$ $\displaystyle r_{3}$ $\displaystyle:(b,f,g)\mapsto(b_{3},b_{2},b_{1},b_{4},b_{5}\sqrt{b_{3}\over b_{1}},b_{6}\sqrt{b_{3}\over b_{1}},b_{7}\sqrt{b_{1}\over b_{3}},b_{8}\sqrt{b_{1}\over b_{3}},f,\sqrt{b_{1}\over b_{3}}\frac{f+b_{3}}{f+b_{1}}g),$ $\displaystyle r_{4}$ $\displaystyle:(b,f,g)\mapsto(b_{2},b_{1},b_{3},b_{4},b_{5},b_{6},b_{7},b_{8},f,g),$ $\displaystyle r_{5}$ $\displaystyle:(b,f,g)\mapsto(b_{1},b_{2},b_{4},b_{3},b_{5},b_{6},b_{7},b_{8},f,g),$ $\displaystyle\sigma_{01}$ $\displaystyle:(b,f,g)\mapsto(b_{1}^{-1},b_{2}^{-1},b_{3}^{-1},b_{4}^{-1},b_{7},b_{8},b_{5},b_{6},f^{-1},g),$ $\displaystyle\sigma_{45}$ $\displaystyle:(b,f,g)\mapsto(b_{3},b_{4},b_{1},b_{2},b_{5}^{-1},b_{6}^{-1},b_{7}^{-1},b_{8}^{-1},f,g^{-1}),$ $\displaystyle\tau$ $\displaystyle:(b,f,g)\mapsto(b_{5},b_{6},b_{7},b_{8},b_{1},b_{2},b_{3},b_{4},g,f),$ $\displaystyle T$ $\displaystyle:b\mapsto(\mathsf{p}^{-2}b_{1},\mathsf{p}^{-2}b_{2},\mathsf{p}^{2}b_{3},\mathsf{p}^{2}b_{4},\mathsf{p}^{-2}b_{5},\mathsf{p}^{-2}b_{6},\mathsf{p}^{2}b_{7},\mathsf{p}^{2}b_{8}).$ The action of $T$ on $(f,g)$ is read from the Painlevé VI equation. ## Appendix B Evidences for the Conjecture ### B.1. Heine limit (the sector with vanishing instanton number) The power of the parameter $\Lambda$ counts the instanton number. In the limit $\Lambda\to 0$ only the fixed points with instanton number zero survive and the partition function degenerates to the $q$-hypergeometric series. Note that in this limit the non-stationary shift $T_{t,\Lambda}$ becomes trivial. ###### Lemma B.1. The equation $\displaystyle Sy(x)=y(x),\quad S:=\frac{1}{\varphi(x)}\cdot\mathcal{B}\cdot\frac{1}{\varphi(-\frac{a}{q}x)\varphi(-\frac{b}{q}x)}\cdot\mathcal{B}\cdot\frac{1}{\varphi(\frac{ab}{q^{2}}x)}T_{\frac{c}{q^{2}},x}$ (B.1) has a solution given by the Heine’s $q$-hypergeometric series: $y(x)={}_{2}\phi_{1}(x)={}_{2}\phi_{1}\left[\left.{a,b\atop c}\right|q;x\right]:=\sum_{n=0}^{\infty}\frac{(a;q)_{n}(b;q)_{n}}{(q,q)_{n}(c;q)_{n}}x^{n}.$ (B.2) ###### Proof. Recall that the Heine’s series ${}_{2}\varphi_{1}(x)$ satisfies the $q$-difference equation $\displaystyle D_{\rm Heine}$ $\displaystyle\ {}_{2}\phi_{1}(x)=0,$ $\displaystyle D_{\rm Heine}$ $\displaystyle:=x(1-ap)(1-bp)p^{-1}-(1-p)(q-cp)p^{-1},\quad p=T_{q,x}.$ (B.3) In a similar way to the proof of Prop. 4.1, we see the adjoint actions on $D_{\rm Heine}$ as $\mapsto$${\rm Ad}\varphi(x)$$\mapsto$${\rm Ad}\mathcal{B}^{-1}$$\mapsto$${\rm Ad}\varphi(-\frac{a}{q}x)\varphi(-\frac{b}{q}x)$$\mapsto$${\rm Ad}\mathcal{B}^{-1}$$\mapsto$${\rm Ad}\varphi(\frac{ab}{q^{2}}x)$, and we have $S^{-1}D_{\rm Heine}\ S=D_{\rm Heine}.$ (B.4) Hence we obtain $D_{\rm Heine}\,S\,{}_{2}\phi_{1}(x)=SD_{\rm Heine}\,{}_{2}\phi_{1}(x)=0.$ (B.5) Since the power series solution of the equation $D_{\rm Heine}y(x)=0$ of the form $y(x)=1+O(x)$ is unique, we have the conclusion $S\,{}_{2}\phi_{1}(x)={}_{2}\phi_{1}(x)$. ∎ ###### Remark B.2. Lemma B.1 shows the factorization of the coefficients $c_{m,0}$ of the “boundary” terms of the solution (4.9). The other cases $c_{0,n}$ are similar. ### B.2. Macdonald limit #### B.2.1. Macdonald function of types $A_{1}$ and $A_{2}$ [41] The asymptotically free Macdonald function $f^{\mathfrak{gl}_{2}}(x|s|q,t)$ of type $A_{1}$ is written as $\displaystyle f^{\mathfrak{gl}_{2}}(x|s|q,t)=\sum_{\theta\geq 0}{(t;q)_{\theta}(ts_{2}/s_{1};q)_{\theta}\over(q;q)_{\theta}(qs_{2}/s_{1};q)_{\theta}}(qx_{2}/tx_{1})^{\theta}.$ (B.6) We have the eigenvalue equation $\displaystyle\mathcal{D}^{\mathfrak{gl}_{2}}(s|q,t)=s_{1}{1-x_{2}/tx_{1}\over 1-x_{2}/x_{1}}T_{q,x_{1}}+s_{2}{1-tx_{2}/x_{1}\over 1-x_{2}/x_{1}}T_{q,x_{2}},$ $\displaystyle\mathcal{D}^{\mathfrak{gl}_{2}}(s|q,t)f^{\mathfrak{gl}_{2}}(x|s|q,t)=(s_{1}+s_{2})f^{\mathfrak{gl}_{2}}(x|s|q,t).$ The asymptotically free Macdonald function $f^{\mathfrak{gl}_{3}}(x|s|q,t)$ of type $A_{2}$ is written as $\displaystyle f^{\mathfrak{gl}_{3}}(x|s|q,t)$ $\displaystyle=\sum_{\theta_{12},\theta_{13},\theta_{23}\geq 0}{(t;q)_{\theta_{12}}(q^{-\theta_{23}+\theta_{13}}ts_{2}/s_{1};q)_{\theta_{12}}\over(q;q)_{\theta_{12}}(q^{-\theta_{23}+\theta_{13}}qs_{2}/s_{1};q)_{\theta_{12}}}(qx_{2}/tx_{1})^{\theta_{12}}$ $\displaystyle\cdot{(t;q)_{\theta_{13}}(ts_{3}/s_{1};q)_{\theta_{13}}\over(q;q)_{\theta_{13}}(qs_{3}/s_{1};q)_{\theta_{13}}}{(ts_{2}/s_{1};q)_{\theta_{13}}(q^{\theta_{23}-\theta_{13}}ts_{1}/s_{2};q)_{\theta_{13}}\over(qs_{2}/s_{1};q)_{\theta_{13}}(q^{\theta_{23}-\theta_{13}}qs_{1}/s_{2};q)_{\theta_{13}}}(q^{2}x_{3}/t^{2}x_{1})^{\theta_{13}}$ $\displaystyle\cdot{(t;q)_{\theta_{23}}(ts_{3}/s_{2};q)_{\theta_{23}}\over(q;q)_{\theta_{23}}(qs_{3}/s_{2};q)_{\theta_{23}}}(qx_{3}/tx_{2})^{\theta_{23}}.$ (B.7) The eigenvalue equation is $\displaystyle\mathcal{D}^{\mathfrak{gl}_{3}}(s|q,t)=s_{1}{1-x_{2}/tx_{1}\over 1-x_{2}/x_{1}}{1-x_{3}/tx_{1}\over 1-x_{3}/x_{1}}T_{q,x_{1}}$ $\displaystyle\qquad\qquad\quad+s_{2}{1-tx_{2}/x_{1}\over 1-x_{2}/x_{1}}{1-x_{3}/tx_{2}\over 1-x_{3}/x_{2}}T_{q,x_{2}}+s_{3}{1-tx_{3}/x_{1}\over 1-x_{3}/x_{1}}{1-tx_{3}/x_{2}\over 1-x_{3}/x_{2}}T_{q,x_{3}},$ $\displaystyle\mathcal{D}^{\mathfrak{gl}_{3}}(s|q,t)f^{\mathfrak{gl}_{3}}(x|s|q,t)=(s_{1}+s_{2}+s_{3})f^{\mathfrak{gl}_{3}}(x|s|q,t).$ #### B.2.2. Macdonald limit by tuning mass parameters Consider the particular choice of the mass parameters; $\displaystyle T_{1}=v^{-1},\qquad T_{2}=v,\qquad T_{3}=v^{-1},\qquad T_{4}=vt^{-1}.$ (B.8) ###### Proposition B.3. Let (B.8) be satisfied. Then the Nekrasov partition function $\Psi(t\Lambda,x)$ degenerates to the asymptotically free Macdonald function of type $\mathfrak{gl}_{3}$ with $(x_{1},x_{2},x_{3})=(1/\Lambda,1/x,1)$ and $(s_{1},s_{2},s_{3})=(1/Q,1,0)$; $\displaystyle\Psi(t\Lambda,x)=f^{\mathfrak{gl}_{3}}(1/\Lambda,1/x,1|1/Q,1,0|q,t).$ (B.9) Therefore, we have the eigenvalue equation $\displaystyle\mathcal{D}^{\mathfrak{gl}_{3}}(1/Q,1,0|q,q/t)\Psi(t\Lambda,x)=(1/Q+1)\Psi(t\Lambda,x),$ where $\displaystyle\mathcal{D}^{\mathfrak{gl}_{3}}(1/Q,1,0|q,q/t)=Q^{-1}{1-t\Lambda/qx\over 1-\Lambda/x}{1-t\Lambda/q\over 1-\Lambda}T_{q,\Lambda}^{-1}+{1-q\Lambda/tx\over 1-\Lambda/x}{1-tx/q\over 1-x}T_{q,x}^{-1}.$ We need the following lemma which is easily confirmed by the explicit formula of the Nekrasov factor. ###### Lemma B.4. (I) $\mathsf{N}_{\lambda,\mu}(1|q,\kappa)\neq 0$ if and only if $\lambda_{i}\leq\mu_{i}$ $(i\geq 1)$. (II) $\mathsf{N}_{\lambda,\mu}(\kappa^{-1}|q,\kappa)\neq 0$ if and only if $\lambda_{i+1}\leq\mu_{i}$ $(i\geq 1)$. Then the proof of Proposition B.3 is given as follows; ###### Proof. Under the condition (B.8), we have (see (2.11)) $\displaystyle v\mathfrak{f}^{+}_{2}/\mathfrak{n}_{1}=1,\qquad v\mathfrak{f}^{+}_{1}/\mathfrak{n}_{2}=1,$ $\displaystyle w\mathfrak{n}_{1}/\mathfrak{m}_{1}=t,\qquad w\mathfrak{n}_{2}/\mathfrak{m}_{2}=1,$ $\displaystyle v\mathfrak{m}_{1}/\mathfrak{f}^{-}_{1}=1,\qquad v\mathfrak{m}_{2}/\mathfrak{f}^{-}_{2}=t.$ Hence, in view of the Lemma above, we have the following parametrization of the quadruples of partitions which give rise to the non vanishing contributions in the partition function (2.10): $\displaystyle\nu_{1}=(\theta_{23}),\qquad\nu_{2}=(\theta_{13}),$ $\displaystyle\mu_{1}=\emptyset,\qquad\mu_{2}=(\theta_{12}+\theta_{13}),$ where $\theta_{12},\theta_{13},\theta_{23}\geq 0$. Then by making straightforward calculation and comparison with (B.7) under the identification $(x_{1},x_{2},x_{3})=(1/\Lambda,1/x,1)$, and $(s_{1},s_{2},s_{3})=(1/Q,1,0)$, we obtain (B.9). ∎ ###### Proposition B.5. Let (B.8) be satisfied. Then we have $\displaystyle\mathcal{F}^{(1)}=\sum_{m\geq 0}{(q/t;q)_{m}(qQ/t;q)_{m}\over(q;q)_{m}(qQ;q)_{m}}t^{m}(\Lambda/x)^{m}=f^{\mathfrak{gl}_{2}}(1/\Lambda,1/x|1/Q,1|q,t),$ (B.10) $\displaystyle\mathcal{F}^{(2)}=\sum_{m,n\geq 0}{(t;q)_{m}(q^{n}q/t;q)_{m}\over(q;q)_{m}(q^{n}qQ;q)_{m}}\left(-(qt^{1/2}Q)^{1/2}\right)^{m}(\Lambda/x)^{m}{(q/t;q)_{n}(qQ/t;q)_{n}\over(q;q)_{n}(qQ;q)_{n}}t^{2n}\Lambda^{n}.$ (B.11) Hence with the identification $(x_{1},x_{2})=(1/\Lambda,1/x)$, we have the eigenvalue equation $\displaystyle\mathcal{D}^{\mathfrak{gl}_{2}}(1/Q,1|q,q/t)\mathcal{F}^{(1)}=(1/Q+1)\mathcal{F}^{(1)},$ where $\displaystyle\mathcal{D}^{\mathfrak{gl}_{2}}(1/Q,1|q,q/t)=Q^{-1}{1-t\Lambda/qx\over 1-\Lambda/x}T_{q,\Lambda}^{-1}+{1-q\Lambda/tx\over 1-\Lambda/x}T_{q,x}^{-1}.$ ###### Proof. Under the condition (B.8), we have $\displaystyle\mathcal{F}^{(1)}=f\left(\left.\left.\begin{array}[]{ll}tQ^{1/2},&qt^{-1/2}Q^{-1/2}\\\ Q^{-1/2},&t^{1/2}Q^{1/2}\\\ Q^{1/2},&q^{-1}t^{3/2}Q^{-1/2}\end{array}\right|-t^{1/2}x,-\Lambda/x\right|q,t^{-1/2}\right),$ and $\displaystyle\mathcal{F}^{(2)}=f\left(\left.\left.\begin{array}[]{ll}tQ^{1/2},&qt^{-1/2}Q^{-1/2}\\\ Q^{-1/2},&t^{1/2}Q^{1/2}\\\ q^{-1}tQ^{-1/2},&t^{-1/2}Q^{1/2}\end{array}\right|t^{3/4}x,t^{1/4}\Lambda/x\right|q,t^{-1/2}\right).$ One finds that $\mathcal{F}^{(1)}$ has the factor $\displaystyle\mathsf{N}^{(1|2)}_{\lambda^{(2)},\emptyset}(t^{1/2}|q,t^{-1/2}),$ which is non vanishing, if and only if $\ell(\lambda^{(2)})\leq 1$. The $\mathcal{F}^{(1)}$ also has the factor $\displaystyle\mathsf{N}^{(1|2)}_{\emptyset,\lambda^{(1)}}(qt^{-1/2}|q,t^{-1/2}),$ which is non vanishing, if and only if $\lambda^{(1)}=\emptyset$. Hence the non vanishing contributions for $\mathcal{F}^{(1)}$ arise if and only if we have $(\lambda^{(1)},\lambda^{(2)})=(\emptyset,(m))$ $(m\geq 0)$. Making explicit calculation, we have (B.10). The $\mathcal{F}^{(2)}$ has the factor $\displaystyle\mathsf{N}^{(0|2)}_{\lambda^{(2)},\emptyset}(t|q,t^{-1/2}),$ which is non vanishing, if and only if $\ell(\lambda^{(2)})\leq 2$. The $\mathcal{F}^{(2)}$ also has the factor $\displaystyle\mathsf{N}^{(1|2)}_{\emptyset,\lambda^{(1)}}(qt^{-1/2}|q,t^{-1/2}),$ which is non vanishing, if and only if $\lambda^{(1)}=\emptyset$. Hence the non vanishing contributions for $\mathcal{F}^{(2)}$ arise if and only if we have $(\lambda^{(1)},\lambda^{(2)})=(\emptyset,(n+m,n))$ $(m,n\geq 0)$. Making explicit calculation, we have (B.11). ∎ We have seen how both partition functions $\Psi(t\Lambda,x)$ and $\mathcal{F}^{(1)}$ are simplified to the asymptotically free Macdonald functions in the limit (B.8). If Conjectures 2.4 and 5.4 are true, they are related by the following gauge transformation. ###### Conjecture B.6. $\displaystyle\mathcal{F}^{(1)}$ $\displaystyle={1\over\Phi(t^{2}T_{2}T_{3}\Lambda)\Phi(qtT_{1}T_{4}\Lambda)}\varphi(q^{1/2}t^{1/2}T_{1}x)\varphi(q^{1/2}t^{1/2}T_{4}\Lambda/x)\psi(\Lambda,x)$ $\displaystyle={\Phi(qtT_{2}T_{3}\Lambda)\Phi(t^{2}T_{1}T_{4}\Lambda)\over\Phi(t^{2}T_{2}T_{3}\Lambda)\Phi(qtT_{1}T_{4}\Lambda)}\varphi(q^{1/2}t^{1/2}T_{1}x)\varphi(q^{1/2}t^{1/2}T_{4}\Lambda/x)\mathcal{A}_{3}(t\Lambda,tqQx)\Psi(t\Lambda,x).$ We can prove this is indeed the case. ###### Lemma B.7. We have $\displaystyle f^{\mathfrak{gl}_{2}}(1/\Lambda,1/x|1/Q,1|q,t)={(t\Lambda;q)_{\infty}\over(q\Lambda;q)_{\infty}}{(tx;q)_{\infty}\over(qx;q)_{\infty}}f^{\mathfrak{gl}_{3}}(1/\Lambda,1/x,1|1/Q,1,0|q,t).$ (B.12) ###### Proof. We find that the Macdonald operators $\mathcal{D}^{\mathfrak{gl}_{3}}(1/Q,1,0|q,q/t)$ and $\mathcal{D}^{\mathfrak{gl}_{2}}(1/Q,1|q,q/t)$ are gauge equivalent $\displaystyle{(t\Lambda;q)_{\infty}\over(q\Lambda;q)_{\infty}}{(tx;q)_{\infty}\over(qx;q)_{\infty}}\mathcal{D}^{\mathfrak{gl}_{3}}(1/Q,1,0|q,q/t){(q\Lambda;q)_{\infty}\over(t\Lambda;q)_{\infty}}{(qx;q)_{\infty}\over(tx;q)_{\infty}}=\mathcal{D}^{\mathfrak{gl}_{2}}(1/Q,1|q,q/t).$ Then the equality (B.12) follows from the uniqueness of the normalized asymptotic eigenfunctions of $\mathcal{D}^{\mathfrak{gl}_{3}}(1/Q,1,0|q,q/t)$ and $\mathcal{D}^{\mathfrak{gl}_{2}}(1/Q,1|q,q/t)$. ∎ ###### Proposition B.8. Conjecture B.6 holds in the Macdonald limit (B.8). ###### Proof. Under the condition (B.8), we have $\displaystyle{\Phi(qtT_{2}T_{3}\Lambda)\Phi(t^{2}T_{1}T_{4}\Lambda)\over\Phi(t^{2}T_{2}T_{3}\Lambda)\Phi(qtT_{1}T_{4}\Lambda)}\varphi(q^{1/2}t^{1/2}T_{1}x)\varphi(q^{1/2}t^{1/2}T_{4}\Lambda/x)\mathcal{A}_{3}(t\Lambda,tqQx)={(t\Lambda;q)_{\infty}\over(q\Lambda;q)_{\infty}}{(tx;q)_{\infty}\over(qx;q)_{\infty}}.$ ∎ Now we are ready to prove Conjecture 5.4 in the Macdonald limit. We need the following formulas for the proof. ###### Lemma B.9 ([44]). For $n\in\mathbb{Z}$, we have $\displaystyle\mathcal{B}\cdot{1\over\varphi(\alpha x)\varphi(\beta\Lambda/x)}x^{n}={\varphi(-q^{1+n}\alpha x)\varphi(-q^{-n}\beta\Lambda/x)\over\varphi(\alpha\beta\Lambda)}q^{n(n+1)/2}x^{n},$ (B.13) $\displaystyle\mathcal{B}^{-1}\cdot{\varphi(\alpha x)\varphi(\beta\Lambda/x)}x^{n}={\varphi(q^{-1}\alpha\beta\Lambda)\over\varphi(-q^{-1-n}\alpha x)\varphi(-q^{n}\beta\Lambda/x)}q^{-n(n+1)/2}x^{n}.$ (B.14) ###### Proposition B.10 ($q$-Chu-Vandermonde sums [22]). For $n\in\mathbb{Z}_{\geq 0}$, we have $\displaystyle{}_{2}\phi_{1}\left[\left.{a,q^{-n}\atop c}\right|q;q\right]={(c/a;q)_{n}\over(c;q)_{n}}a^{n},$ (B.15) $\displaystyle{}_{2}\phi_{1}\left[\left.{a,q^{-n}\atop c}\right|q;{cq^{n}\over a}\right]={(c/a;q)_{n}\over(c;q)_{n}}.$ (B.16) ###### Proposition B.11. Conjecture 5.4 holds in the Macdonald limit (B.8). $\displaystyle T_{(qt^{1/2}Q)^{1/2},x}\mathcal{F}^{(2)}=\varphi(q\Lambda){1\over\varphi(-qtx)\varphi(-\Lambda/x)}\cdot\mathcal{B}\cdot{1\over\varphi(tx)\varphi(q\Lambda/tx)}\mathcal{F}^{(1)},$ (B.17) $\displaystyle\mathcal{F}^{(1)}=\varphi(t\Lambda){1\over\varphi(qx)\varphi(t\Lambda/x)}\cdot\mathcal{B}\cdot{1\over\varphi(-x)\varphi(-q\Lambda/x)}(T_{t^{1/2},x}^{-1}T_{t,\Lambda}^{-1})T_{(qt^{1/2}Q)^{-1/2},x}\mathcal{F}^{(2)}.$ (B.18) ###### Proof. First, we show (B.17). By using (B.13) and the $q$-binomial formula, we have $\displaystyle\varphi(q\Lambda){1\over\varphi(-qtx)\varphi(-\Lambda/x)}\cdot\mathcal{B}\cdot{1\over\varphi(tx)\varphi(q\Lambda/tx)}\mathcal{F}^{(1)}$ $\displaystyle={}$ $\displaystyle\varphi(q\Lambda){1\over\varphi(-qtx)\varphi(-\Lambda/x)}\cdot\mathcal{B}\cdot{1\over\varphi(tx)\varphi(q\Lambda/tx)}\sum_{n\geq 0}{(q/t;q)_{n}(qQ/t;q)_{n}\over(q;q)_{n}(qQ;q)_{n}}t^{n}(\Lambda/x)^{n}$ $\displaystyle={}$ $\displaystyle\sum_{n\geq 0}{\varphi(-q^{1-n}tx)\varphi(-q^{1+n}\Lambda/tx)\over\varphi(-qtx)\varphi(-\Lambda/x)}{(q/t;q)_{n}(qQ/t;q)_{n}\over(q;q)_{n}(qQ;q)_{n}}t^{n}q^{n(n-1)/2}(\Lambda/x)^{n}$ $\displaystyle={}$ $\displaystyle\sum_{k,l,n\geq 0}{(q^{-n};q)_{k}\over(q;q)_{k}}(-qtx)^{k}{(q^{n+1}/t;q)_{l}\over(q;q)_{l}}(-\Lambda/x)^{l}{(q/t;q)_{n}(qQ/t;q)_{n}\over(q;q)_{n}(qQ;q)_{n}}t^{n}q^{n(n-1)/2}(\Lambda/x)^{n}.$ Note that we have the truncation of the summation as $\sum_{k,l,n\geq 0}=\sum_{k,l\geq 0}\sum_{n=k}^{\infty}$ because of the factor $(q^{-n};q)_{k}$. Shifting the running index $n$ by $k$ as $n\rightarrow n+k$ (hence changing the summation range as $\sum_{k,l\geq 0}\sum_{n=k}^{\infty}\rightarrow\sum_{k,l\geq 0}\sum_{n=0}^{\infty}$ accordingly), we have $\displaystyle={}$ $\displaystyle\sum_{k,l,n\geq 0}{(q^{-k-n};q)_{k}\over(q;q)_{k}}(-qt)^{k}{(q^{k+n+1}/t;q)_{l}\over(q;q)_{l}}(-1)^{l}$ $\displaystyle\cdot{(q/t;q)_{k+n}(qQ/t;q)_{k+n}\over(q;q)_{k+n}(qQ;q)_{k+n}}t^{k+n}q^{(k+n)(k+n-1)/2}\cdot(\Lambda/x)^{l+n}\Lambda^{k}.$ Then we change the running index $l$ to $m$ (where $m=l+n$) as $\sum_{l,n\geq 0}=\sum_{m\geq 0}\sum_{n=0}^{m}$. Simplifying the factors, we have $\displaystyle={}$ $\displaystyle\sum_{k\geq 0}\sum_{m\geq 0}{(q^{k+1}/t;q)_{m}\over(q;q)_{m}}(-\Lambda/x)^{m}{(q/t;q)_{k}\over(q;q)_{k}}{(qQ/t;q)_{k}\over(qQ;q)_{k}}(t^{2}\Lambda)^{k}\sum_{n=0}^{m}{(q^{-m};q)_{n}\over(q;q)_{n}}{(q^{k+1}Q/t;q)_{n}\over(q^{k+1}Q;q)_{n}}q^{mn}t^{n}$ $\displaystyle={}$ $\displaystyle\sum_{k\geq 0}\sum_{m\geq 0}{(q^{k+1}/t;q)_{m}\over(q;q)_{m}}{(t;q)_{m}\over(q^{k+1}Q;q)_{m}}(-\Lambda/x)^{m}{(q/t;q)_{k}\over(q;q)_{k}}{(qQ/t;q)_{k}\over(qQ;q)_{k}}(t^{2}\Lambda)^{k}$ $\displaystyle={}$ $\displaystyle T_{(qt^{1/2}Q)^{1/2},x}\mathcal{F}^{(2)}.$ Here, we have used the $q$-Chu-Vandermonde summation formula (B.16) in the last step. The second equation (B.18) can be shown in exactly the same manner as above. Note that (B.18) is equivalent to $\displaystyle(T_{t^{1/2},x}^{-1}T_{t,\Lambda}^{-1})T_{(qt^{1/2}Q)^{-1/2},x}\mathcal{F}^{(2)}=$ $\displaystyle{}\,{1\over\varphi(t\Lambda)}{\varphi(-x)\varphi(-q\Lambda/x)}\cdot\mathcal{B}^{-1}\cdot{\varphi(qx)\varphi(t\Lambda/x)}\mathcal{F}^{(1)}.$ By using (B.14) and the $q$-binomial formula, we have $\displaystyle{1\over\varphi(t\Lambda)}{\varphi(-x)\varphi(-q\Lambda/x)}\cdot\mathcal{B}^{-1}\cdot{\varphi(qx)\varphi(t\Lambda/x)}\mathcal{F}^{(1)}$ $\displaystyle={}$ $\displaystyle{1\over\varphi(t\Lambda)}{\varphi(-x)\varphi(-q\Lambda/x)}\cdot\mathcal{B}^{-1}\cdot{\varphi(qx)\varphi(t\Lambda/x)}\sum_{n\geq 0}{(q/t;q)_{n}(qQ/t;q)_{n}\over(q;q)_{n}(qQ;q)_{n}}t^{n}(\Lambda/x)^{n}$ $\displaystyle={}$ $\displaystyle\sum_{n\geq 0}{\varphi(-x)\varphi(-q\Lambda/x)\over\varphi(-q^{n}x)\varphi(-q^{-n}t\Lambda/x)}{(q/t;q)_{n}(qQ/t;q)_{n}\over(q;q)_{n}(qQ;q)_{n}}t^{n}q^{-n(n-1)/2}(\Lambda/x)^{n}$ $\displaystyle={}$ $\displaystyle\sum_{k,l,n\geq 0}{(q^{-n};q)_{k}\over(q;q)_{k}}(-q^{n}x)^{k}{(q^{n+1}/t;q)_{l}\over(q;q)_{l}}(-q^{-n}t\Lambda/x)^{l}{(q/t;q)_{n}(qQ/t;q)_{n}\over(q;q)_{n}(qQ;q)_{n}}t^{n}q^{-n(n-1)/2}(\Lambda/x)^{n}.$ Note that we have $\sum_{k,l,n\geq 0}=\sum_{k,l\geq 0}\sum_{n=k}^{\infty}$ because of the factor $(q^{-n};q)_{k}$. Shifting the running index $n$ by $k$ as $n\rightarrow n+k$ (and changing the summation range as $\sum_{k,l\geq 0}\sum_{n=k}^{\infty}\rightarrow\sum_{k,l\geq 0}\sum_{n=0}^{\infty}$ accordingly), we have $\displaystyle={}$ $\displaystyle\sum_{k,l,n\geq 0}{(q^{-k-n};q)_{k}\over(q;q)_{k}}(-q^{n})^{k}{(q^{k+n+1}/t;q)_{l}\over(q;q)_{l}}(-q^{-n}t)^{l}$ $\displaystyle\cdot{(q/t;q)_{k+n}(qQ/t;q)_{k+n}\over(q;q)_{k+n}(qQ;q)_{k+n}}t^{k+n}q^{-(k+n)(k+n-1)/2}\cdot(\Lambda/x)^{l+n}\Lambda^{k}.$ Changing the running index $l$ to $m$ (where $m=l+n$) as $\sum_{l,n\geq 0}=\sum_{m\geq 0}\sum_{n=0}^{m}$, and simplifying the factors, we have $\displaystyle={}$ $\displaystyle\sum_{k\geq 0}\sum_{m\geq 0}{(q^{k+1}/t;q)_{m}\over(q;q)_{m}}(-\Lambda/x)^{m}{(q/t;q)_{k}\over(q;q)_{k}}{(qQ/t;q)_{k}\over(qQ;q)_{k}}\Lambda^{k}q^{-km}t^{k+m}\sum_{n=0}^{m}{(q^{-m};q)_{n}\over(q;q)_{n}}{(q^{k+1}Q/t;q)_{n}\over(q^{k+1}Q;q)_{n}}q^{n}$ $\displaystyle={}$ $\displaystyle\sum_{k\geq 0}\sum_{m\geq 0}{(q^{k+1}/t;q)_{m}\over(q;q)_{m}}{(t;q)_{m}\over(q^{k+1}Q;q)_{m}}(-qQ\Lambda/x)^{m}{(q/t;q)_{k}\over(q;q)_{k}}{(qQ/t;q)_{k}\over(qQ;q)_{k}}(t\Lambda)^{k}$ $\displaystyle={}$ $\displaystyle(T_{t^{1/2},x}^{-1}T_{t,\Lambda}^{-1})T_{(qt^{1/2}Q)^{-1/2},x}\mathcal{F}^{(2)}.$ Here, we have used the $q$-Chu-Vandermonde summation formula (B.15) in the last step. ∎ #### B.2.3. Macdonald limit in [44] For the sake of readers’ convenience, we recollect the facts concerning the choice of the parameters investigated in [44]; $\displaystyle T_{1}=vt^{-1},\qquad T_{2}=v^{-1},\qquad T_{3}=v^{-1},\qquad T_{4}=vt^{-1}.$ (B.19) Note that in the limit (B.19) we have (See Conjecture 2.4); $\displaystyle\mathcal{A}_{1}(\Lambda,x)={1\over\varphi(qt^{-1}x)\varphi(t\Lambda/x)},$ $\displaystyle\mathcal{A}_{2}(\Lambda,x)={\varphi(t\Lambda)\varphi(qt^{-2}\Lambda)\over\varphi(-t^{-1}x)\varphi(-Q^{-1}x)\varphi(-qQ\Lambda/x)\varphi(-q\Lambda/x)},$ $\displaystyle\mathcal{A}_{3}(\Lambda,x)={1\over\varphi(q^{-1}Q^{-1}x)\varphi(q^{2}t^{-1}Q\Lambda/x)}.$ Set $\displaystyle U(\Lambda,x)=\sum_{k,l\geq 0}{(q/t;q)_{k}(q^{l}qQ/t;q)_{k}\over(q;q)_{k}(q^{l}qQ;q)_{k}}(\Lambda/x)^{k}{(t;q)_{l}(tQ;q)_{l}\over(q;q)_{l}(qQ;q)_{l}}(q\Lambda/t^{2})^{l},$ $\displaystyle V(\Lambda,x)=\sum_{k,l\geq 0}{(q/t;q)_{k}(t;q)_{k}\over(q;q)_{k}(qQ;q)_{k}}(-qQ\Lambda/x)^{k}{(q^{k}t^{2};q)_{l}(tQ;q)_{l}\over(q;q)_{l}(q^{k}qQ;q)_{l}}(q\Lambda/t^{2})^{l}.$ ###### Proposition B.12 ([44]). When (B.19) is satisfied, we have $\displaystyle\Psi(\Lambda,x)=U(\Lambda,x),$ and $\displaystyle V(\Lambda,x)={\varphi(t\Lambda)\over\varphi(-Q^{-1}x)\varphi(-qQ\Lambda/x)}\cdot\mathcal{B}\cdot{1\over\varphi(q^{-1}Q^{-1}x)\varphi(q^{2}t^{-1}Q\Lambda/x)}U\left(\Lambda,{x\over tqQ}\right),$ $\displaystyle U(t\Lambda,x)={\varphi(qt^{-2}\Lambda)\over\varphi(qt^{-1}x)\varphi(t\Lambda/x)}\cdot\mathcal{B}\cdot{1\over\varphi(-t^{-1}x)\varphi(-q\Lambda/x)}V\left(\Lambda,x\right).$ ###### Proposition B.13. In the limit (B.19), we have $\displaystyle{\varphi(t\Lambda/x)\over\varphi(q\Lambda/tx)}\Psi(t\Lambda,x)$ $\displaystyle=\sum_{k,l\geq 0}{(t;q)_{k}(q^{l}tQ;q)_{k}\over(q;q)_{k}(q^{l}qQ;q)_{k}}(q\Lambda/tx)^{k}{(t;q)_{l}(tQ;q)_{l}\over(q;q)_{l}(qQ;q)_{l}}(q\Lambda/t)^{l}$ $\displaystyle=f^{\mathfrak{gl}_{3}}(1/\Lambda,1/x,1|tQ^{-1},t,1|q,t),$ $\displaystyle\mathcal{D}^{\mathfrak{gl}_{3}}(tQ^{-1},t,1|q,t){\varphi(t\Lambda/x)\over\varphi(q\Lambda/tx)}\Psi(t\Lambda,x)=(tQ^{-1}+t+1){\varphi(t\Lambda/x)\over\varphi(q\Lambda/tx)}\Psi(t\Lambda,x),$ where $\displaystyle\mathcal{D}^{\mathfrak{gl}_{3}}(tQ^{-1},t,1|q,t)$ $\displaystyle=tQ^{-1}{1-\Lambda/tx\over 1-\Lambda/x}{1-\Lambda/t\over 1-\Lambda}T_{q,\Lambda}^{-1}$ $\displaystyle+t{1-t\Lambda/x\over 1-\Lambda/x}{1-x/t\over 1-x}T_{q,x}^{-1}+{1-t\Lambda/x\over 1-\Lambda/x}{1-tx\over 1-x}T_{q,\Lambda}T_{q,x}.$ ###### Remark B.14. In the limit (B.19), the set of partitions giving rise to the non vanishing contributions for $\mathcal{F}^{(1)}$ remains rather big. This seems to be consistent with the appearance of double infinite products in the gauge transformation; $\displaystyle{\Phi(q^{-1}t^{3}\Lambda)\Phi(q^{2}t^{-2}\Lambda)\over\Phi(qt^{-1}\Lambda)\Phi(t^{2}\Lambda)}{\varphi(tx)\over\varphi(qt^{-1}x)}\mathcal{F}^{(1)}=\Psi(t\Lambda,x).$ In four dimensional theory, the set of partitions giving rise to the non vanishing contributions is the same. The four dimensional limit of the ratio of double infinite products is $(1-\Lambda)^{-\frac{2(\epsilon_{1}+\epsilon_{2})^{2}}{\epsilon_{1}\epsilon_{2}}}$. On the other hand the corresponding set of partitions for $\mathcal{F}^{(2)}$ becomes small and gives us $T_{(qt^{1/2}Q)^{1/2},x}\mathcal{F}^{(2)}=\sum_{m,n\geq 0}{(t;1)_{m}(q^{n}q/t;1)_{m}\over(q;q)_{m}(q^{n}qQ;q)_{m}}(-\Lambda/x)^{m}{(q/t;1)_{n}(qQ/t;1)_{n}\over(q;q)_{n}(qQ;q)_{n}}(t^{2}\Lambda)^{n}.$ ## Appendix C Four dimensional limit in a factorized form In section 4, we computed the four dimensional limit of the operator $\SS$. Here we will consider the limit of each factor $\begin{array}[]{l}\displaystyle K_{1}=\frac{1}{\varphi(\Lambda)}\varphi(-d_{1}x)\varphi(-d_{3}\frac{\Lambda}{x})\mathcal{B}^{-1}\varphi(qx)\varphi(\frac{\Lambda}{x}),\\\ \displaystyle K_{2}=\frac{1}{\varphi(q^{-1}d_{2}d_{4}\Lambda)}\varphi(q^{-1}d_{1}d_{2}x)\varphi(d_{3}d_{4}\frac{\Lambda}{x})\mathcal{B}^{-1}\varphi(-d_{2}x)\varphi(-d_{4}\frac{\Lambda}{x}),\\\ \displaystyle N=\frac{\varphi(q^{-1}d_{2}d_{4}\Lambda)}{\varphi(q^{-1}d_{1}d_{2}d_{3}d_{4}\Lambda)},\end{array}$ (C.1) in the decomposition $\SS^{-1}=NK_{2}K_{1},$ (C.2) separately. We show that each factor $K_{i}$ has also a well defined four dimensional limit, though the result is not usual differencial operator any more. To describe the results in a simple form, it is convenient to use the normal ordering $:\ :$ which is a $\mathbb{C}$-linear map from a commutative ring $\mathbb{C}(x,\vartheta_{x})_{/\vartheta_{x}x=x\vartheta_{x}}$ to the corresponding non-commutative ring $\mathbb{C}(x,\vartheta_{x})_{/\vartheta_{x}x=x(\vartheta_{x}+1)}$ defined by $:x^{n}F(x,\vartheta_{x}):\ =x^{n}:F(x,\vartheta_{x}):,\quad:F(x,\vartheta_{x})\vartheta_{x}^{n}:\ =\ :F(x,\vartheta_{x}):{\vartheta_{x}}^{n},\quad:1:=1.$ (C.3) ###### Theorem C.1. For $q=e^{h},d_{i}=q^{m_{i}}$, we have $\displaystyle K_{1}$ $\displaystyle=~{}:(1+x)^{-m_{1}-\vartheta_{x}}(1+\frac{\Lambda}{x})^{-m_{3}+\vartheta_{x}}\left(1+\frac{h}{2}A_{1}\right)+{\mathcal{O}}(h^{2}):,$ $\displaystyle K_{2}$ $\displaystyle=~{}:(1-x)^{-m_{1}-\vartheta_{x}}(1-\frac{\Lambda}{x})^{-m_{3}+\vartheta_{x}}\left(1+\frac{h}{2}A_{2}\right)+{\mathcal{O}}(h^{2}):,$ $\displaystyle N$ $\displaystyle=(1-\Lambda)^{-(m_{1}+m_{3})}\left(1+\frac{h}{2}\frac{(m_{1}+m_{3})(m_{1}+2m_{2}+m_{3}+2m_{4}-3)}{(1-\Lambda)}\right)+{\mathcal{O}}(h^{2}),$ where $\displaystyle A_{1}$ $\displaystyle=\frac{x(\vartheta_{x}-m_{1}+1)(\vartheta_{x}+m_{1})}{1+x}+\frac{\frac{\Lambda}{x}(\vartheta_{x}+m_{3}-1)(\vartheta_{x}-m_{3})}{1+\frac{\Lambda}{x}}-\vartheta_{x}(\vartheta_{x}+1),$ $\displaystyle A_{2}$ $\displaystyle=-\frac{x(\vartheta_{x}-m_{1}-2m_{2}+3)(\vartheta_{x}+m_{1})}{1-x}-\frac{\frac{\Lambda}{x}(\vartheta_{x}+m_{3}+2m_{4}-1)(\vartheta_{x}-m_{3})}{1-\frac{\Lambda}{x}}-\vartheta_{x}(\vartheta_{x}+1).$ ###### Proof. From (B.14) in Appendix B, we have $\mathcal{B}^{-1}\varphi(qx)\varphi(\frac{\Lambda}{x})x^{n}=\frac{\varphi(\Lambda)}{\varphi(-q^{-n}x)\varphi(-q^{n}\frac{\Lambda}{x})}q^{-n(n+1)}x^{n},$ (C.4) and hence $K_{1}x^{n}=\frac{\varphi(-d_{1}x)\varphi(-d_{3}\frac{\Lambda}{x})}{\varphi(-q^{-n}x)\varphi(-q^{n}\frac{\Lambda}{x})}q^{-n(n+1)}x^{n}.$ (C.5) Then, using the limiting formula of the $q$-binomial theorem $\frac{\varphi(q^{j}x)}{\varphi(x)}=\sum_{k=1}^{\infty}\frac{(q^{j})_{k}}{(q)_{k}}x^{k}=(1-x)^{-j}\big{\\{}1+\frac{hj(j-1)}{2}\frac{x}{1-x}\big{\\}}+{\mathcal{O}}(h^{2}),$ (C.6) we obtain $\begin{array}[]{l}\displaystyle K_{1}x^{n}=(1+x)^{-m_{1}-n}(1+\frac{\Lambda}{x})^{-m_{3}+n}(1+\frac{h}{2}A_{1})+{\mathcal{O}}(h^{2})\\\ \displaystyle\qquad A_{1}=\frac{x(n-m_{1}+1)(n+m_{1})}{1+x}+\frac{\frac{\Lambda}{x}(n+m_{3}-1)(n-m_{3})}{1+\frac{\Lambda}{x}}-n(n+1),\end{array}$ (C.7) as desired. The expression for $K_{2}$ follows from that for $K_{1}$ by the substitution $x\to-q^{-1}d_{2}x,\quad\Lambda\to q^{-1}d_{2}d_{4}\Lambda.$ (C.8) The limit of $N$ follows directly as $\begin{array}[]{l}\displaystyle N=\exp\sum_{n=1}^{\infty}\frac{(q^{-1}d_{2}d_{4}\Lambda)^{n}-(q^{-1}d_{1}d_{2}d_{3}d_{4}\Lambda)^{n}}{n(1-q^{n})}=\exp\sum_{n=1}^{\infty}\frac{(q^{m_{2}+m_{4}-1}\Lambda)^{n}(1-q^{m_{1}+m_{3}+n})}{n(1-q^{n})}\\\ \displaystyle=\exp\\{\sum_{n=1}^{\infty}\frac{1}{n}(1+nh(m_{2}+m_{4}-1))\Lambda^{n}(m_{1}+m_{3})(1+\frac{nh}{2}(m_{1}+m_{3}-1))\\}+{\mathcal{O}}(h^{2}),\\\ \displaystyle=\exp\\{(m_{1}+m_{3})\sum_{n=1}^{\infty}\frac{\Lambda^{n}}{n}(1+\frac{nh(m_{1}+2m_{2}+m_{3}+2m_{4}-3)}{2})\\}+{\mathcal{O}}(h^{2}),\\\ \displaystyle=(1+\frac{h(m_{1}+2m_{2}+m_{3}+2m_{4}-3)}{2}\vartheta_{\Lambda})\exp\\{(m_{1}+m_{3})\sum_{n=1}^{\infty}\frac{\Lambda^{n}}{n}\\}+{\mathcal{O}}(h^{2}),\\\ \displaystyle=(1-\Lambda)^{-(m_{1}+m_{3})}(1+\frac{h}{2}\frac{(m_{1}+m_{3})(m_{1}+2m_{2}+m_{3}+2m_{4}-3)}{(1-\Lambda)})+{\mathcal{O}}(h^{2}).\end{array}$ (C.9) ∎ ###### Remark C.2. The result of Theorem C.1 is consistent with the result in section 4. For instance, the consistency in the leading order is given by the identity $:(1-x)^{-m_{1}-\vartheta_{x}}(1-\frac{\Lambda}{x})^{-m_{3}+\vartheta_{x}}::(1+x)^{-m_{1}-\vartheta_{x}}(1+\frac{\Lambda}{x})^{-m_{3}+\vartheta_{x}}:=(1-\Lambda)^{m_{1}+m_{3}},$ (C.10) which follows from formal computations such as $:(1+x)^{-m_{1}-\vartheta_{x}}(1+\frac{\Lambda}{x})^{-m_{3}+\vartheta_{x}}:f(x)=(1+x)^{-m_{1}}(1+\frac{\Lambda}{x})^{-m_{3}}f(x\frac{1+\frac{\Lambda}{x}}{1+x}).$ (C.11) In relation to (C.11), the formula $:(1-x)^{-b-\vartheta_{x}}:F(a,b,c;-x)=(1-x)^{-b}F(a,b,c,\frac{x}{x-1})=F(c-a,b,c;x).$ (C.12) for the Gauss Hypergeometric series $F(a,b,c;x)$ will be instructive. ## References * [1] L. F. Alday, D. Gaiotto, S. Gukov, Y. Tachikawa and H. Verlinde, “Loop and surface operators in N=2 gauge theory and Liouville modular geometry,” JHEP 01 (2010), 113 [arXiv:0909.0945 [hep-th]]. * [2] L. F. Alday, D. Gaiotto and Y. Tachikawa, “Liouville Correlation Functions from Four-dimensional Gauge Theories,” Lett. Math. Phys. 91 (2010), 167-197 [arXiv:0906.3219 [hep-th]]. * [3] L. F. Alday and Y. Tachikawa, “Affine SL(2) conformal blocks from 4d gauge theories,” Lett. Math. Phys. 94 (2010), 87-114 [arXiv:1005.4469 [hep-th]]. * [4] H. Awata, H. Fuji, H. Kanno, M. Manabe and Y. Yamada, “Localization with a Surface Operator, Irregular Conformal Blocks and Open Topological String,” Adv. Theor. Math. Phys. 16 (2012) no.3, 725-804 [arXiv:1008.0574 [hep-th]]. * [5] H. Awata and Y. Yamada, “Five-dimensional AGT Conjecture and the Deformed Virasoro Algebra,” JHEP 01 (2010), 125 [arXiv:0910.4431 [hep-th]]. * [6] L. Bao, V. Mitev, E. Pomoni, M. Taki and F. Yagi, “Non-Lagrangian Theories from Brane Junctions,” JHEP 01 (2014), 175 [arXiv:1310.3841 [hep-th]]. * [7] A. A. Belavin, A. M. Polyakov and A. B. Zamolodchikov, “Infinite Conformal Symmetry in Two-Dimensional Quantum Field Theory,” Nucl. Phys. B 241 (1984), 333-380 * [8] M. Bershtein, P. Gavrylenko and A. Marshakov, “Cluster integrable systems, $q$-Painlevé equations and their quantization,” JHEP 02 (2018), 077 [arXiv:1711.02063 [math-ph]]. * [9] G. Bonelli, A. Grassi and A. Tanzini, “Quantum curves and $q$-deformed Painlevé equations,” Lett. Math. Phys. 109 (2019) no.9, 1961-2001 [arXiv:1710.11603 [hep-th]]. * [10] G. Bonelli, F. Del Monte and A. Tanzini, “BPS Quivers of Five-Dimensional SCFTs, Topological Strings and q-Painlevé Equations,” Annales Henri Poincare 22 (2021) no.8, 2721-2773 [arXiv:2007.11596 [hep-th]]. * [11] A. Brandhuber, N. Itzhaki, J. Sonnenschein, S. Theisen and S. Yankielowicz, “On the M theory approach to (compactified) 5-D field theories,” Phys. Lett. B 415 (1997), 127-134 [arXiv:hep-th/9709010 [hep-th]]. * [12] A. Braverman, “Instanton counting via affine Lie algebras. 1. Equivariant J functions of (affine) flag manifolds and Whittaker vectors,” In Workshop on algebraic structures and moduli spaces: CRM Workshop, AMS (2003) [arXiv:math/0401409 [math.AG]]. * [13] A. Braverman and P. Etingof, “Instanton counting via affine Lie algebras II: From Whittaker vectors to the Seiberg-Witten prepotential,” In Studies in Lie Theory: dedicated to A. Joseph on his 60th birthday Birkhäuser (2006) [arXiv:math/0409441 [math.AG]]. * [14] M. Bullimore, H. C. Kim and P. Koroteev, “Defects and Quantum Seiberg-Witten Geometry,” JHEP 05 (2015), 095 [arXiv:1412.6081 [hep-th]]. * [15] T. Eguchi and H. Kanno, “Five-dimensional gauge theories and local mirror symmetry,” Nucl. Phys. B 586 (2000), 331-345 [arXiv:hep-th/0005008 [hep-th]]. * [16] L. D. Faddeev and A. Y. Volkov, “Abelian current algebra and the Virasoro algebra on the lattice,” Phys. Lett. B 315 (1993), 311-318 [arXiv:hep-th/9307048 [hep-th]]. * [17] V. A. Fateev and A. B. Zamolodchikov, ”Self-dual solutions of the star-triangle relations in ZN-models.” Phys. Lett. 92A (1982), 37-39. * [18] B. Feigin, M. Finkelberg, A. Negut and R. Rybnikov, “Yangians and cohomology rings of Laumon spaces,” Sel. Math. New Ser. 17 (2011) 573-607, [arXiv:0812.4656 [math.AG]]. * [19] M. Finkelberg and R. Rybnikov, “Quantization of Drinfeld Zastava,” [arXiv:1009.0676 [math.AG]]. * [20] E. Frenkel, S. Gukov and J. Teschner, “Surface Operators and Separation of Variables,” JHEP 01 (2016), 179 [arXiv:1506.07508 [hep-th]]. * [21] S. Garoufalidis and C. Wheeler, “Modular $q$-holonomic modules,” [arXiv:2203.17029 [math.GT]]. * [22] R. Gasper and M. Rahman, “Basic hypergeometric series, 2nd ed., ”Encyclopedia of Mathematics and its Applications, vol 96, Cambridge University Press, (2004). * [23] K. Hasegawa, “Quantizing the Bäcklund transformations of Painlevé equations and the quantum discrete Painlevé VI equation,” Adv. Stud. Pure Math. 61 “Exploring New Structures and Natural Constructions in Mathematical Physics”, K. Hasegawa, et. al, eds. (2011), 275-288, [arXiv:math/0703036 [math.QA]]. * [24] K. Hasegawa, “Quantizing the Discrete Painleve VI Equation : The Lax Formalism,” Lett. Math. Phys. 103 (2013) 865-879, [arXiv:1210.0915 [math.QA]]. * [25] H. Hayashi, H. C. Kim and T. Nishinaka, “Topological strings and 5d $T_{N}$ partition functions,” JHEP 06 (2014), 014 [arXiv:1310.3854 [hep-th]]. * [26] M. Jimbo and H. Sakai, A $q$-analog of the sixth Painlevé equation, Lett. Math. Phys. 38(1996) 145-154. * [27] H. Kanno and Y. Tachikawa, “Instanton counting with a surface operator and the chain-saw quiver,” JHEP 1106, 119 (2011) [arXiv:1105.0357 [hep-th]]. * [28] S. S. Kim and F. Yagi, “5d En Seiberg-Witten curve via toric-like diagram,” JHEP 06 (2015), 082 [arXiv:1411.7903 [hep-th]]. * [29] A. N. Kirillov,“Dilogarithm Identities,” Prog. Theor. Phys. Supple. 118 (1995), 61-142. * [30] C. Kozcaz, S. Pasquetti, F. Passerini and N. Wyllard, “Affine sl(N) conformal blocks from N=2 SU(N) gauge theories,” JHEP 01 (2011), 045 [arXiv:1008.1412 [hep-th]]. * [31] G. Kuroki, “Quantum groups and quantization of Weyl group symmetries of Painlevé systems,” Adv. Stud. Pure Math. 61 “Exploring New Structures and Natural Constructions in Mathematical Physics”, K. Hasegawa, et. al, eds. (2011) 289-325, [arXiv:0808.2604 [math.QA]]. * [32] E. Langmann, M. Noumi and J. Shiraishi, “Basic properties of non-stationary Ruijsenaars functions,” SIGMA 16 (2020), 105 [arXiv:2006.07171 [math-ph]]. * [33] A. Marshakov, A. Mironov and A. Morozov, “On AGT Relations with Surface Operator Insertion and Stationary Limit of Beta-Ensembles,” J. Geom. Phys. 61 (2011), 1203-1222 [arXiv:1011.4491 [hep-th]]. * [34] S. Moriyama and Y. Yamada, “Quantum Representation of Affine Weyl Groups and Associated Quantum Curves,” SIGMA 17 (2021), 076 [arXiv:2104.06661 [math.QA]]. * [35] A. Negut, “Affine Laumon spaces and integrable systems,” [arXiv:1112.1756 [math.AG]]. * [36] N. Nekrasov, Seiberg-Witten prepotential from instanton counting, Advances in Theoretical and Mathematical Physics 7, no. 5 (2003): 831-864. [arXiv:hep-th/0206161]. * [37] N. Nekrasov, “BPS/CFT correspondence: non-perturbative Dyson-Schwinger equations and qq-characters,” JHEP 03 (2016), 181 [arXiv:1512.05388 [hep-th]]. * [38] N. Nekrasov, “BPS/CFT correspondence V: BPZ and KZ equations from qq-characters,” [arXiv:1711.11582 [hep-th]]. * [39] N. A. Nekrasov and S. L. Shatashvili, “Quantization of Integrable Systems and Four Dimensional Gauge Theories,” [arXiv:0908.4052 [hep-th]]. * [40] N. Nekrasov and A. Tsymbaliuk, “Surface defects in gauge theory and KZ equation,” [arXiv:2103.12611 [hep-th]]. * [41] M. Noumi and J. Shiraishi, “A direct approach to the bispectral problem for the Ruijsenaars-Macdonald $q$-difference operators,” [arXiv:1206.5364 [math.QA]]. * [42] R. Poghossian, “Deformed SW curve and the null vector decoupling equation in Toda field theory,” JHEP 04 (2016), 070 [arXiv:1601.05096 [hep-th]]. * [43] H. Sakai, Rational surfaces associated with affine root systems and geometry of the Painlevé equations, Comm. Math. Phys. 220 (2001) 165–229. * [44] S. Shakirov, “Non-stationary difference equation for $q$-Virasoro conformal blocks,” [arXiv:2111.07939 [math.RT]]. * [45] J. Shiraishi, “Affine screening operators, affine Laumon spaces and conjectures concerning non-stationary Ruijsenaars functions,” J.Integrable.Syst. 4 (2019) xyz010 [arXiv:1903.07495 [math.QA]]. * [46] J. Shiraishi, H. Kubo, H. Awata and S. Odake, “A Quantum deformation of the Virasoro algebra and the Macdonald symmetric functions,” Lett. Math. Phys. 38 (1996), 33-51 [arXiv:q-alg/9507034 [math.QA]]. * [47] M. Taki, “Surface Operator, Bubbling Calabi-Yau and AGT Relation,” JHEP 07 (2011), 047 [arXiv:1007.2524 [hep-th]].
# Stellar s-process neutron capture cross sections of 69,71Ga M. Tessler<EMAIL_ADDRESS>Soreq Nuclear Research Center, Yavne, Israel 81800 Racah Institute of Physics, Hebrew University, Jerusalem, Israel 91904 M. Paul Racah Institute of Physics, Hebrew University, Jerusalem, Israel 91904 S. Halfon Soreq Nuclear Research Center, Yavne, Israel 81800 Y. Kashiv University of Notre Dame, Notre Dame, IN 46556, USA D. Kijel Soreq Nuclear Research Center, Yavne, Israel 81800 A. Kreisel Soreq Nuclear Research Center, Yavne, Israel 81800 A. Shor Soreq Nuclear Research Center, Yavne, Israel 81800 L. Weissman Soreq Nuclear Research Center, Yavne, Israel 81800 ###### Abstract The stable gallium isotopes, 69,71Ga, are mostly produced by the weak slow ($s$) process in massive stars. We report here on measurements of astrophysically-relevant neutron capture cross sections of the 69,71Ga$(n,\gamma)$ reactions. The experiments were performed by the activation technique using a high-intensity ($3-5\times 10^{10}$ n/s), quasi- Maxwellian neutron beam that closely mimics conditions of stellar $s$-process nucleosynthesis at $kT\approx$ 40 keV. The neutron field was produced by a mA proton beam at $E_{p}=1925$ keV (beam power of 2-3 kW) from the Soreq Applied Research Accelerator Facility (SARAF), bombarding the Liquid-Lithium Target (LiLiT). A 473 mg sample of Ga2O3 of natural isotopic composition was activated in the LiLiT neutron field and the activities of 70,72Ga were measured by decay counting via $\gamma$-spectrometry with a high-purity germanium detector. The Maxwellian-averaged cross sections at $kT$ = 30 keV of 69Ga and 71Ga determined in this work are 136(8) mb and 105(7) mb, respectively, in good agreement with previous experimental values. Astrophysical implications of the measurements are discussed. ## I Introduction The majority of elements beyond iron are produced by neutron capture reactions in stars. The slow neutron capture process ($s$ process) is composed of the weak and main components [1]. The main component takes place during recurrent thermal pulses in the He shell of low-mass AGB stars ($M\lesssim 4M_{\odot}$). The weak component is produced during core He and shell C burning in massive stars ($M\gtrsim 8M_{\odot}$). Neutron densities during the $s$ process are between $10^{6}$ and $10^{12}$ $n/cm^{3}$ [2]. The weak $s$ process produces most of the $s$-process isotopes between iron and strontium ($60<A<90$). Figure 1 shows the weak $s$-process flow in the Ga region. Figure 1: The path of the $s$ process between zinc and arsenic. The numerical values are the terrestrial isotopic abundances in % (half-life) for stable (unstable) nuclides. When 69Ga captures a neutron, the product 70Ga either decays to 70Ge (99.59%) or to 70Zn (0.41%) with a half-life of 21.14 minutes. Following the neutron capture reaction on 71Ga, the product 72Ga decays to 72Ge with a half-life of 14.1 hours. Recent simulations show that gallium is the most abundant $s$-element at the end of shell carbon burning in a model of $25M_{\odot}$ star [3]. Until recently, there was only a Time-of-Flight (TOF) measurement of the 69Ga$(n,\gamma)$ cross section, with a sample of natural gallium [4] (see also [5]). Other nuclei measured within the same experimental campaign (74Ge, 75As, 81Br) show large deviations from more recent results [5]. For 71Ga, two previous results from activation measurements are in marginal agreement [6, 7]. Very recently, values were published for the neutron capture cross sections of 69,71Ga [8] which are stated to disagree with available evaluated data from KADoNiS v0.3 [5]. We report here on new measurements of the 69,71Ga$(n,\gamma)$ cross sections, which were initiated prior to the publication of the latter results. This work took advantage of the Liquid- Lithium Target (LiLiT) at the Soreq Applied Research Accelerator Facility (SARAF) intense quasi-Maxwellian neutron source ([9], see Sec. II), especially valuable for the activation measurement of short-lived nuclides like 70Ga ($t_{1/2}$= 21.1 min). Section II describes the SARAF-LiLiT neutron source, Sec. III describes the samples used and the irradiation details, and Sec. IV describes the activity measurements. In Sec. V we report the activation results, in Sec. VI we calculate the experimental cross sections, Sec. VII describes how the MACS were calculated, and in Sec. VIII we discuss the results. ## II SARAF-LiLiT Figure 2: Diagram of the Liquid-Lithium Target (LiLiT) and activation target assembly. The 1-2 mA ($\approx$9 mm full width) proton beam (dashed red arrow) impinges directly on the windowless liquid-lithium film. The light blue shows the liquid-lithium circulating flow (see [10] for details). The activation samples are mounted at the center of a ring target holder made of Al and positioned in the outgoing neutron cone (green dashed lines) at a distance of $\approx$6 mm from the liquid-lithium film surface. The targets are in a vacuum chamber separated from the LiLiT chamber by a 0.5 mm stainless steel concave vacuum wall. An intense 7Li$(p,n)^{7}$Be neutron source, in the form of a Liquid-Lithium Target (LiLiT) [10, 11], bombarded by a mA proton beam from the Soreq Applied Research Accelerator Facility (SARAF), was developed and is used for Maxwellian Average Cross Section (MACS) measurements. The SARAF accelerator is based on a continuous wave (CW), proton/deuteron RF superconducting linear accelerator. SARAF Phase I, presently undergoing a major upgrade to its Phase II, consisted of a 20 keV/u Electron Cyclotron Resonance (ECR) ion source injector, a Low-Energy Beam Transport section (LEBT), a four-rod Radio Frequency Quadrupole (RFQ, 1.5 MeV/u), a Medium Energy Beam Transport section (MEBT), a Prototype Superconducting Module (PSM) housing six half-wave resonators and three superconducting solenoids, and a diagnostic plate (D-plate). The beamline downstream of the accelerator transports the high intensity beam to the target. SARAF Phase I delivered for experiments currents of up to 2 mA of protons or deuterons, with energies of up to $\approx$ 4 and 5 MeV, respectively [12, 13, 14]. The SARAF high-intensity beam requires a Li target that can withstand its power, which is incompatible with solid Li and Li compounds. The Liquid- Lithium Target (LiLiT) [10, 11] consists of a liquid-lithium film (temperature $\approx$200 °C, above the lithium melting temperature of 180.5 °C) circulated at high velocity (3-7 $m/s$) onto a thin convex stainless-steel support wall. The target is bombarded with a high-intensity proton beam impinging directly on the Li-vacuum (windowless) interface at an energy above and close to the 7Li$(p,n)$ reaction threshold, $E_{th}=1.88$ MeV (Fig. 2). A rectangular- shaped nozzle just upstream of the curved support wall determines the film width and thickness to be 18 mm and 1.5 mm, respectively (see [10] for details). The first few microns at the surface of the liquid-lithium film serve as the neutron-producing thick target. The deeper Li film layers act as a beam dump, from which the power is transported by the flow to a heat exchanger. A spherical cap made of stainless steel foil, 0.5 mm thick and 19 cm in diameter, is located $\approx$1 mm beyond the nozzle and seals the LiLiT vacuum chamber neutron exit port (Fig. 2). The vacuum wall curvature (convex toward the Li flow with a curvature radius of 300 mm) allows us to locate a secondary activation target very close (see below) to the neutron source at the Li-vacuum interface. ## III Sample characteristics and irradiation details In the experiment described here, a 13 mm diameter natGa2O3 (99.99+% purity, [15]) pellet target was activated. It was sandwiched between two Au foils (Table 1) which were used as monitors of the neutron fluence. The distance of the 13 mm diameter Ga target from the Li neutron source was $6\pm 1$ mm, intercepting $>75\%$ of the outgoing neutrons. The characteristics of the samples used are summarized in Table 1. The Ga and Au targets were inserted into the LiLiT activation chamber (Fig. 2) and held in place by the target holder. The 9 mm full width proton beam impinged on the free-surface lithium film, resulting in an outgoing neutron cone due to the 7Li$(p,n)$ reaction, which irradiated the target. The setup is shown in Fig. 2 and explained in the caption. The number of stable nuclei $A$ per cm2, $n_{t}(A)$, for a target element with an atomic or molecular mass $M_{A}$, target area $S$ and mass $m$, is given by Eq. (1). $n_{t}(A)=s_{A}\cdot a(A)\frac{m\cdot N_{A}}{S\cdot M_{A}}.$ (1) The symbol $N_{A}$ denotes Avogadro’s number and $a(A)$ the isotopic abundance of $A$ (Table 2). The stoichiometry of element $A$ in the target is denoted by $s_{A}$ ($s_{A}$=2 or 1 for Ga or Au, respectively). Table 1: Characteristics of the samples used in this work. They are listed in the order that they were placed downstream from the Li target. The first gold foil (Au #33) was necessary to monitor the beam offset (see below). The Ga oxide targets #2 (#1) were used for neutron irradiations above (below) the Li neutron emission threshold. Sample | Diam. | Mass | Nucleus | $n_{t}$ ---|---|---|---|--- | (mm) | (mg) | | ($10^{19}$cm-2) Au #33 | 25 | 109.8(1) | 197Au | 6.839(4) Au #14 | 13 | 31.6(1) | 197Au | 7.28(1) natGa2O3 #2 | 13 | 472.5(1) | 69Ga | 137.5(1) | | | 71Ga | 91.25(1) Au #15 | 13 | 32.5(1) | 197Au | 7.49(1) natGa2O3 #1 | 13 | 443.0(1) | 69Ga | 128.9(1) | | | 71Ga | 85.55(1) The proton beam energy was measured by a TOF pick-up and Rutherford backscattering off a Au target located in the diagnostic D-plate. The beam energy was 1925 keV, with an energy spread of $\approx$15 keV. The energy spread was estimated from beam dynamics calculations and was verified experimentally under similar conditions [16]. To determine the position of the proton beam relative to the activation targets, the 25 mm diameter gold foil (Au #33) was auto-radiographically scanned after the samples irradiation (Fig. 3). An offset of 2.5 mm in the vertical direction was found and accounted for in our detailed simulations. Figure 3: An auto-radiographic scan of the 25 mm diameter gold foil (Au #33). Red indicates the area with the highest neutron irradiation followed by yellow, green and blue. A $\approx$2.5 mm offset was observed in the neutron irradiation and is attributed to the vertical steering of the proton beam. This offset was taken into account in the neutron irradiation simulations. Throughout the irradiation, the neutron yield was continuously monitored and recorded with a fission-product ionization chamber detector (PFC16A, Centronics Ltd.), counting neutron-induced fission events from a thin 235U internal foil (1 mg/cm2, 12.5 cm2 active area). The fission chamber was located at 0$\degree$ to the incident proton beam, at a distance of $\approx$80 cm downstream from the target. The fission chamber was covered with a 1 mm thick Cd sheet to absorb scattered thermal neutrons. The count rate of the fission chamber was calibrated to the beam current at low intensity (10% duty cycle, using a slow chopper), with the Faraday cup located $\approx$1 m upstream of the Li target. After the SARAF was tuned, the beam duty cycle was ramped up to 99%. Normally, the ramp up is performed rather slowly while monitoring the temperature and radiation along the beamline and LiLiT. If necessary, fine-tuning of the beamline ion-optical magnetic elements (bending magnets, steerers) is performed based on temperature reading of sensors located on the lithium nozzle (see [10, 11] for details). It is important to know the time dependence of the neutron yield in the case of a short-lived activation product (e.g., 70Ga, $t_{1/2}$= 21.1 min). In such a case, one needs to account for fluctuations of the neutron yield when evaluating the fraction of the reaction product that decayed during the irradiation. The time record of the proton beam current is presented in figure 4. The total integrated current in the irradiation was $\sim$0.96 mA$\times$h. Figure 4: Time record of the fission chamber count rate (left axis) and corresponding intensity of the proton beam current (right axis) during the irradiation. The two low-intensity intervals and gaps at beginning and end of the irradiation correspond to the calibration of the fission count rate against beam current measured at low intensity with a Faraday cup (see text). All intensity variations are taken into account by the $1/f_{b}$ correction factor (Eq. (2)). The total integrated current was 0.96 mA$\times$h. ## IV Activity measurement After the irradiation, the induced activities were measured separately for each sample with a shielded HPGe detector (ORTEC GMX 25-83). The distance of the sample to the HPGe detector was 20 cm (2 cm) for the measurement above (below) the neutron threshold. The detector efficiency was determined by standard calibrated radioactive sources: 22Na, 60Co, 88Y, 133Ba, 137Cs, 241Am, 152Eu and 155Eu. The measured efficiency curve at 20 cm is presented in Fig. 5. Figure 5: Efficiency of the HPGe detector used at a distance of 20 cm (black dots with 1$\sigma$ uncertainties) determined by standard calibrated radioactive sources: 22Na, 60Co, 88Y, 133Ba, 137Cs, 241Am, 152Eu and 155Eu. The red curve is a fit to the data with the expression: $\epsilon=a\times E_{\gamma}^{-b}$ and the dotted lines are the uncertainty (1$\sigma$). The main source of the efficiency uncertainty is the radioactive source activity; these uncertainties are given by the source manufacturers. The small value of the reduced chi square suggests that these uncertainties are overestimated. The $\gamma$-ray spectrum for the neutron activated Ga sample (natGa2O3 #2) is presented in Fig. 6, where the full-energy peaks of 70Ga and 72Ga are labeled. Figure 6: $\gamma$-ray spectrum of the activated natGa2O3 #2 sample. The spectrum was accumulated for 3000 s, starting 2464 s after the end of the neutron activation. The sample was located 20 cm from the HPGe detector. The main Ga isotope full-energy peaks, with energies in keV, are labeled. Figure 7 presents the decay curves of the 176.3 and 1039.5 keV $\gamma$ lines of 70Ga, showing an excellent agreement with the adopted 70Ga half-life of 21.14(5) min. [17]. Figure 7: Decay curves for 70Ga. The uncertainties of each data point in this plot are only the statistical uncertainties of the activity measurement. $N_{\gamma}=N_{act}{\cdot}I_{\gamma}$ (see Eq. 2) is the photon intensity ($\gamma$s per second) of the $\gamma$ transitions at the 176.3 and 1039.5 keV lines of the 70Ga decay, determined by $\gamma$ spectrometry. Excellent agreement is observed with the adopted half-life of 70Ga from the literature (21.14(5) min. [17]), and between the number of activated 70Ga nuclei (see Tables 3 and 4) derived independently from each transition, using the respective adopted $\gamma$ intensities [17]; see text. The numbers of activated 70Ga nuclei derived from each of the two transitions, using the adopted $\gamma$ intensities [17], are in excellent agreement as well. The ratio of $\gamma$ intensities of the 1039 and 176 keV lines determined here is 2.53(6) in agreement and better precision than the ratio of adopted intensities 2.24(19) [17]. The reported $\gamma$ intensity ratio in [18] is 2.30(6). The precise counting of activated 70Ga nuclei, in spite of its short half-life relative to the irradiation time and the very low intensity of the $\gamma$ transitions (Table 2), is credited to the high intensity of the LiLiT neutron source. Figure 8: Secondary $\gamma$-ray spectrum of the natGa2O3 #1 sample after irradiation with primary $\gamma$-rays. The spectrum was accumulated for 636 s, starting 1275 s after the end of the irradiation. The proton energy was below the neutron production threshold, resulting in the production of only $\gamma$ rays. The sample was located 2 cm from the HPGe detector. No Ga lines are observed. The SARAF-LiLiT setup was shown to produce high-energy $\gamma$-rays ($\approx$108 14.6 and 17.6 MeV $\gamma$-rays/(mA$\times$s)) via the 7Li$(p,\gamma)^{8}$Be reaction [19, 9]. In cases where there are a stable A-1 isotope, an unstable A isotope and a stable A+1 isotope, the nuclide A can be produced through the A-1$(n,\gamma)$ or A+1$(\gamma,n)$ reactions. To test the possible contribution of the 71Ga$(\gamma,n)$ reaction to the production of 70Ga, an irradiation of protons on LiLiT was conducted below the neutron production threshold. This resulted in the irradiation of the Ga sample (natGa2O3 #1) with $\gamma$-rays, but without neutrons. The proton energy was measured to be 1800 keV. The total integrated current in this irradiation was $\approx$0.46 mA$\times$h. The $\gamma$-ray spectrum for the activated Ga sample (natGa2O3 #1) is presented in Fig. 8. As can be seen, no Ga isotopes were created through $(\gamma,n)$ reactions. ## V Activation results Table 2: Properties of the relevant target and product nuclei studied in this work. The half-life data were taken from [20, 17, 21]. Target | Isotopic | Product | Half-life, ---|---|---|--- nucleus | abundance | nucleus | $t_{\frac{1}{2}}$ 197Au | 1 | 198Au | $2.6947(3)$ d 69Ga | 0.60108(9) | 70Ga | $21.14(5)$ m 71Ga | 0.39892(9) | 72Ga | $14.10(1)$ h The number of activated nuclei created during the irradiation, $N_{act}$, was obtained from the $\gamma$-ray spectra using Eq. (2), $N_{act}=\frac{C}{\epsilon_{\gamma}I_{\gamma}K_{\gamma}}\frac{e^{\lambda t_{cool}}}{1-e^{-\lambda t_{real}}}\frac{t_{real}}{t_{live}}\frac{1}{f_{b}},$ (2) where $C$ is the number of counts in a full-energy peak, $\epsilon_{\gamma}$ is the detector energy-dependent full-energy efficiency for the relevant target-detector geometry and $I_{\gamma}$ is the $\gamma$-intensity per decay. The $I_{\gamma}$ used in this work was taken from [20, 17, 21]. The correction due to $\gamma$-ray self absorption in the sample is $K_{\gamma}$. In the case of a disk sample of thickness $x$, $K_{\gamma}\approx\frac{1-e^{-\mu x}}{\mu x}$, where $\mu$ is the $\gamma$-ray absorption coefficient. The $\gamma$-ray absorption coefficients, $\mu$, were taken from [22]. The decay constant of the activated nucleus is $\lambda=\frac{ln(2)}{t_{1/2}}$. The cooling time between the end of the irradiation and the start of activity measurement is $t_{cool}$, and $t_{real}$ ($t_{live}$) is the real (live) measurement time. The decay of activated nuclei during the irradiation is accounted for in $f_{b}$. It is calculated using the time dependence of the neutron yield $\Phi(t)$, obtained from the fission chamber (see Fig. 4), by $f_{b}=\frac{\intop_{0}^{t_{a}}\Phi(t)e^{-\lambda(t_{a}-t)}dt}{\intop_{0}^{t_{a}}\Phi(t)dt}$. $t_{a}$ is the time of the end of irradiation. The decay parameters and correction factors used in this analysis are listed in Tables 2 and 3. Table 3: Analysis of one of the $\gamma$-ray spectra from the natGa2O3 #2 sample, measured 5180 s after the end of the neutron irradiation, with 656.8 s real time and 600 s live time; see Eq. (2) for the definition of notations. The data for $E_{\gamma}$ and $I_{\gamma}$ were taken from [17, 21]. Nucleus | $E_{\gamma}$ (keV) | counts | $I_{\gamma}$ (%) | $\epsilon_{\gamma}$ ($10^{-4}$) | $K_{\gamma}$ | $f_{b}$ | $N_{act}$ ($10^{9}$) ---|---|---|---|---|---|---|--- 70Ga | 176.115(13) | 1264(93) | 0.29(1) | 30.4(8) | 0.970(3) | 0.550 | 16.4(14) | 1039.513(10) | 602(43) | 0.65(5) | 6.50(8) | 0.990(3) | 0.550 | 16.1(17) 72Ga | 600.912(15) | 3578(99) | 5.822(19) | 10.5(1) | 0.987(3) | 0.984 | 7.96(24) | 629.967(19) | 15484(157) | 26.13(4) | 10.0(1) | 0.987(3) | 0.984 | 8.00(12) | 834.13(4) | 45243(224) | 95.45(8) | 7.87(9) | 0.989(3) | 0.984 | 8.15(10) | 894.327(18) | 4419(80) | 10.136(15) | 7.41(9) | 0.989(3) | 0.984 | 7.96(17) | 1050.794(17) | 2643(64) | 6.991(11) | 6.44(8) | 0.990(3) | 0.984 | 7.93(22) | 1861.996(18) | 1248(46) | 5.41(3) | 3.92(8) | 0.992(3) | 0.984 | 7.93(17) | 2201.586(17) | 5479(79) | 26.87(12) | 3.39(7) | 0.993(3) | 0.984 | 8.10(21) | 2491.026(17) | 1437(42) | 7.73(3) | 3.05(6) | 0.993(3) | 0.984 | 8.22(29) | 2507.718(17) | 2454(52) | 13.33(6) | 3.03(6) | 0.993(3) | 0.984 | 8.19(25) The numbers of activated nuclei at the end of the irradiation, $N_{act}$, calculated with Eq. (2), are summarized in table 4. Table 4: The number of activated nuclei at the end of the irradiation, $N_{act}$, and comparison with simulated $N_{act}$. Sample | Nucleus | $N_{act}$ ($10^{9}$) | simulated $N_{act}$ ($10^{9}$) ---|---|---|--- Au-14 | 198Au | 4.07(8) | 4.09 natGa2O3 #2 | 70Ga | 16.3(7) | | 72Ga | 8.1(1) | Au-15 | 198Au | 3.65(6) | 3.63 ## VI Experimental cross section Since the sample cross section is measured relative to the Au cross section (which is considered known), the cross section of the sample, averaged over the experimental neutron spectrum, can be expressed as $\sigma_{exp}(i)=\sigma_{ENDF}(\textrm{Au})\frac{N_{act}(i)}{N_{act}(\textrm{Au})}\frac{n_{t}(\textrm{Au})}{n_{t}(i)},$ (3) where $i$ denotes the Ga stable isotope (69 or 71) and $\sigma_{ENDF}(\textrm{Au})$ is the reference Au cross section from the ENDF/B-VIII.0 library [23] averaged over the experimental neutron spectrum. It is defined as $\sigma_{ENDF}(\textrm{Au})=\frac{\int\sigma_{ENDF}(E_{n};\textrm{Au})\frac{dn}{dE_{n}}dE_{n}}{\int\frac{dn}{dE_{n}}dE_{n}}.$ (4) The energy-dependent 197Au$(n,\gamma)^{198}$Au cross section $\sigma_{ENDF}(E_{n};\textrm{Au})$ was taken from ENDF/B-VIII.0 [23], in agreement with high-precision experimental data [24, 25]. The neutron spectrum, $\frac{dn}{dE_{n}}$, is obtained from our simulation code (Fig. 9), developed and benchmarked by experiment [26, 27, 19, 9]. The simulated neutron spectrum impinging on the Ga target, along with a $kT$ = 41.8 keV fit to a Maxwellian neutron flux ($\propto E\cdot e^{-\frac{E}{kT}}$), is presented in Fig. 9. This spectrum is generated by a GEANT4 [28] simulation, using the SimLiT code [27] output as the neutron source. The SimLiT calculation uses 7Li$(p,n)$ differential cross sections taken from [29] and takes into account the proton mean beam energy and energy spread, proton energy loss in the liquid Li using differential $dE/dx$ values taken from SRIM [30] and a Gaussian proton beam profile consistent with the Au monitor auto-radiography (Fig. 3). The detailed GEANT4 simulation takes into account the LiLiT geometry setup including the off-center position of the neutron beam relative to the Au-Ga-Au target and the surrounding materials. The simulation explicitly calculates (see [19, 9] for details) the number of activated 198Au nuclei, based on the $\sigma_{ENDF}(E_{n};Au)$ cross sections and the measured proton charge during irradiation and reproduces the experimental 198Au activity of the Au monitors within 0.5% (Table 4). Figure 9: The simulated neutron spectrum impinging on the Ga target (black). Also shown is a fit (between 0 and 90 keV) to a Maxwell-Boltzmann (MB) flux (red), where the best fit is for $kT$ = 41.8 keV. The results of the experimental cross sections, $\sigma_{exp}$, of 69Ga and 71Ga (Eq. 3) and of the $\sigma_{ENDF}(\textrm{Au})$ values (Eq. 4) are presented in Table 5. Table 5: The experimental cross sections, $\sigma_{exp}$ (Eq. 3), measured in this work for the 69,71Ga$(n,\gamma)^{70,72}$Ga reactions, together with the $\sigma_{ENDF}$ cross section averaged over the neutron energy distribution (Fig. 9). See Table 7 for explanation of $\sigma_{ENDF}(^{197}$Au) uncertainty. Isotope | $\sigma_{exp}$ (mb) | $\sigma_{ENDF}$ (mb) ---|---|--- 197Au | | $524(10)$ 69Ga | $119(5)$ | 103 71Ga | $89(2)$ | 107 ## VII Maxwellian Averaged Cross Section The Maxwellian Averaged Cross Section (MACS) at a given thermal energy, $kT$, is defined as: $\Braket{\sigma}_{kT}=\frac{\Braket{\sigma v}}{v_{T}}=\frac{2}{\sqrt{\pi}}\frac{\intop_{0}^{\infty}\sigma(E_{n})E_{n}e^{-\frac{E_{n}}{kT}}dE_{n}}{\intop_{0}^{\infty}E_{n}e^{-\frac{E_{n}}{kT}}dE_{n}},$ (5) where $\sigma(E_{n})$ is the differential $(n,\gamma)$ reaction cross section at neutron energy $E_{n}$. In this work, the MACS at a given thermal energy $kT$ is calculated with the procedure developed in [19, 31, 9] and using Eq. (6), $\textrm{MACS}_{exp}(kT)=\frac{2}{\sqrt{\pi}}\cdot C_{lib}(kT)\cdot\sigma_{exp},$ (6) where the correction factor $C_{lib}(kT)$ is given by Eq. (7): $C_{lib}(kT)=\frac{\frac{\int_{0}^{\infty}\sigma_{lib}(E_{n})E_{n}e^{-\frac{E_{n}}{kT}}dE_{n}}{\int_{0}^{\infty}E_{n}e^{-\frac{E_{n}}{kT}}dE_{n}}}{\frac{\int_{0}^{\infty}\sigma_{lib}(E_{n})\frac{dn}{dE_{n}}dE_{n}}{\int_{0}^{\infty}\frac{dn}{dE_{n}}dE_{n}}}.$ (7) In Eq. (7), $\frac{dn}{dE_{n}}$ is the simulated experimental neutron spectrum (Fig. 9) and $\sigma_{lib}(E_{n})$ is the energy-dependent neutron capture cross section taken from an evaluation library. In Table 6 we present the correction factors, $C_{lib}$, and the MACSexp at $kT=30$ keV, calculated with Eq. (6), for the Ga isotopes. The MACSexp at $kT=30$ keV derived in this work are obtained by using in Eq. (6) the average $C_{lib}$ calculated using the various cross section libraries: ENDF/B-VIII.0 [23], JENDL-4.0 [32], JEFF-3.3 [33], CENDL-3.2 [34] and TENDL-2019 [35] (see Table 6). Table 6: Comparison of correction factors, $C_{lib}$, and MACS calculated for 69Ga and 71Ga at 30 keV using the different libraries [23, 32, 33, 34, 35]. See text for explanation. | $C_{lib}$ | | MACSexp (mb) ---|---|---|--- Library | 69Ga | 71Ga | | 69Ga | 71Ga ENDF/B-VIII.0 [23] | 1.03 | 1.02 | | 138 | 103 JENDL-4.0 [32] | 1.03 | 1.01 | | 138 | 102 JEFF-3.3 [33] | 1.0 | 0.97 | | 134 | 98 CENDL-3.2 [34] | 0.99 | 1.08 | | 133 | 108 TENDL-2019 [35] | 1.04 | 1.15 | | 139 | 115 average | 1.018 | 1.046 | | 136.4 | 105.0 standard deviation | 0.02 | 0.07 | | 2.8 | 6.7 The experimental uncertainties, as discussed in detail in [19, 9], are summarized in Table 7. The uncertainties for the average $C_{lib}$ were obtained by taking the standard deviation of the $C_{lib}$ calculated for the various cross section libraries [23, 32, 34, 33, 35] (see Table 6). Table 7: Random (rand) and systematic (sys) uncertainties in the results presented in this work. | Uncertainty (%) ---|--- | 69Ga | | 71Ga Source of uncertainty | rand | sys | | rand | sys target thickness | 0.5 | | | 0.5 | activity measurement | 4.0 | | | 0.5 | full-energy eff. rel. to Au | | 0.5a | | | 0.5a intensity per decay | | 3.4 | | | 0.1 $\sigma_{ENDF}$(Au) | | 1.9b | | | 1.9b $C_{lib}$ | | 2.0 | | | 6.4 Total random uncertainty | 4.0 | | | 0.7 | Total systematic uncertainty | | 4.4 | | | 6.7 Total uncertainty | 6.0 | | 6.7 * a This contribution to the uncertainty of the 69,71Ga MACS includes only the ratio of the full-energy efficiencies of the 69,71Ga $\gamma$ lines to that of the 198Au $\gamma$ line. In Tables 3 and 4 the overall uncertainty (including the systematic uncertainty of the calibration sources) is quoted. * b This value includes the uncertainty of beam parameters (proton beam energy, energy spread, and distance of sample from Li) of 0.6%, the uncertainty of the simulations of 1.5% and the uncertainty of the ENDF cross section for Au of 1.0%, $\sqrt{0.6^{2}+1.5^{2}+1.0^{2}}=1.9\%$. See [9] for more details of the uncertainties. Table 8: Comparison of the results of this work with previous experiments and compilations. The MACS listed in the Table correspond to $kT$= 30 keV. The previously measured MACS were renormalized as specified in [36] using the recent measured 197Au$(n,\gamma)^{198}$Au cross section data [25, 24, 37] used as standard. The KADoNiS v0.3 [5] recommended MACS (30 keV) is the average of the experimental data renormalized to the 197Au$(n,\gamma)^{198}$Au MACS data measured in [38]. The KADoNiS v1.0 [36] recommended MACS(30 keV) values are an average from evaluated libraries ENDF/B-VII.1 [39], JENDL-4.0 [32] and TENDL-2015 [40]. | KADoNiS | KADoNiS | Walter, | Anand | Walter | Göbel | This work ---|---|---|---|---|---|---|--- | v0.3 [5] | v1.0 [36] | 1984 [4] | et al., 1979 [6] | et al., 1986 [7] | et al., 2021 [8] | MACS(69Ga) (mb) | 139(6) | 123(9) | 149(6) | | | | 136(8) $\sigma_{exp}$(69Ga)/$\sigma_{exp}$(197Au) 1 | | | | | | 0.286(19) | 0.227(12) MACS(71Ga) (mb) | 123(8) | 103(14) | | 79(23) | 130(8) | | 105(7) $\sigma_{exp}$(71Ga)/$\sigma_{exp}$(197Au) 1 | | | | | | 0.173(11) | 0.170(5) * 1 Note that Göbel et al. [8] denote $\sigma_{exp}$ as defined in Eq. (3) by SACS (spectrum averaged cross section) for their experimental neutron spectrum. Figure 10: Comparison of this work’s results (red) with previously measured MACS at $kT=30$ keV (black) for 69Ga [4] (left) and 71Ga [6, 7] (right). The previously measured MACS were renormalized as specified in the KADoNiS v1.0 website [36]. The TOF measurement is displayed as a circle and the activation measurements as squares. ## VIII Discussion This work’s results are compared in Table 8 and Fig. 10 with previously measured MACS at $kT$ = 30 keV for 69Ga [4] and 71Ga [6, 7], and with compilations of experimental values (KADoNiS v0.3, [5]) and of evaluated values (KADoNiS v1.0, [36]). We compare also the ratios $\sigma_{exp}$(AGa)/$\sigma_{exp}$(197Au) with the corresponding values from Göbel et al. [8], denoted there as SACS(AGa)/SACS(197Au), calculated for the relevant experimental spectrum. Our result for the 69Ga MACS (30 keV) are in good agreement within uncertainties with the previous experimental value of [4] and those recommended by Bao et al. [41] and KADoNiS v0.3 [5]). Our MACS value for 71Ga is larger than the experimental value of [6], but smaller than the experimental value of [7] recommended by Bao et al. [41] and KADoNiS v0.3 [5]), though marginally consistent within the quoted uncertainties. This smaller value is significant in view of the fact that the previous value of KADoNiS v0.3 [5] was used in extensive network calculations [3]; see below. Both 69,71Ga MACS values extracted in this work are in reasonable agreement with the evaluated values of KADoNiS v1.0 [36]. We note also the good agreement of our $\sigma_{exp}$(71Ga)/$\sigma_{exp}$(197Au) value with that of Göbel at al. [8], but a striking discrepancy for $\sigma_{exp}$(69Ga)/$\sigma_{exp}$(197Au), whose origin is not understood. Gallium, like the other elements between iron and strontium (26 $<$ Z $<$ 38, 60 $\lesssim$ A $\lesssim$ 90), is produced primarily by the weak component of the $s$ process in massive stars (M${}_{initial}>$ 8M⊙) ([42] and references therein). Pignatari et al. [3] studied nucleosynthesis, including the weak $s$ process, in a model of a Population I (solar metallicity) 25M⊙ star. For the 69,71Ga$(n,\gamma)$ MACS, the authors used the values recommended by Bao et al. [41] (KADoNiS 0.3 [5]). As shown in Table 8 and discussed above, the new 69Ga MACS measurement presented here supports the Bao et al. [41] value, but the 71Ga MACS is 15% lower. This means that less 71Ga is consumed by the $(n,\gamma)$ reaction and hence its isotopic fraction is expected to increase compared to that calculated by Pignatari et al. [3]. In addition, since the $s$ process flow goes from Ga to Ge, 69Ga$(n,\gamma)^{70}$Ga$(\beta^{-})^{70}$Ge and 71Ga$(n,\gamma)^{72}$Ga$(\beta^{-})^{72}$Ge (see Fig. 1), the lower 71Ga MACS is expected to result in a lower production of 72,73,74Ge and reduce their isotopic fraction. The lower 71Ga MACS (105(7) mb) may have as well a wider effect, as Pignatari et al. [3] concluded that the effects of MACS $\lesssim 150$ mb tend to propagate to heavier isotopes. However, we should caution that the above potential effects are tentative. The only way to check for the effect of the new 71Ga MACS is to incorporate it in a network calculation like the one performed by Pignatari et al. [3]. Such nucleosynthesis models need to be tested against isotopic compositions in the relevant star types. This is done for many elements by isotopic analysis of chemical elements in presolar grains. However, to date, there are no measurements of Ga in presolar grains ([43, 44, 45]). The most studied family of presolar grains are the carbides, mainly SiC and graphite. At the same time, Ga is thought to usually not form carbides [46], so it is unlikely that there is enough Ga in presolar carbides for isotopic analysis with the current capabilities of the experimental techniques used. On the other hand, Lodders [47] calculated that Ga condensed in the early Solar system as a trace element into the mineral feldspar, by substituting for the major element Al in the crystal lattice, and into Fe metal to a lesser extent. While presolar feldspar has not been found to date, most of the presolar oxides and silicates studied to date contain Al as a major element. This makes them more likely hosts for stellar Ga to be analyzed. ## IX Summary The neutron capture cross sections of 69,71Ga were measured by the activation technique in the intense $kT\approx$ 40 keV quasi-Maxwellian neutron field of the SARAF-LiLiT facility. The reaction products were measured by $\gamma$ spectrometry with a HPGe detector. The experimental cross sections were converted to Maxwell-averaged cross sections at $kT$ = 30 keV using the energy dependence of various neutron cross section library data. The MACS values obtained in this work are MACS (69Ga) = 136(8) mb and MACS (69Ga) = 105(7) mb. The 69Ga MACS value is in good agreement with previous experimental values and their recommended values while that of 71Ga is smaller and with reduced uncertainty than the experimental recommended value. This smaller value may have implications in network calculations such as those of Pignatari et al. [3] which used so far the recommended values. Potential natural samples to measure stellar Ga composition are presolar silicates and oxides, in which Ga may replace Al in the grains’ crystal lattice. We note a significant and not understood discrepancy between our experimental cross-section value of 69Ga with that recently published by Göbel et al. [8], beyond that expected from the different experimental conditions. ###### Acknowledgements. We would like to thank the SARAF and LiLiT (Soreq NRC) staffs for their dedicated help during the experiments. This work was supported in part by Israel Science Foundation (Grant Nr. 1387/15) and the Pazy Foundation (Israel). M.P. acknowledges support by the European Union (ChETEC-INFRA, project no. 101008324). ## References * Käppeler _et al._ [2011] F. Käppeler, R. Gallino, S. Bisterzo, and W. Aoki, Rev. Mod. Phys. 83, 157 (2011). * Reifarth _et al._ [2014] R. Reifarth, C. Lederer, and F. Käppeler, J. Phys. G 41 (2014). * Pignatari _et al._ [2010] M. Pignatari, R. Gallino, M. Heil, M. Wiescher, F. Käppeler, F. Herwig, and S. Bisterzo, Astrophys. J. 710, 1557 (2010). * Walter [1984] G. Walter, Technical Report KfK-3706, Forschungszentrum Karlsruhe (1984). * Dillmann _et al._ [2009] I. Dillmann, R. Plag, F. Käppeler, and T. Rauscher, in _EFNUDAT Fast Neutrons \- scientific workshop on neutron measurements, theory & applications_ (JRC-IRMM, Geel, Belgium, 2009) available online: http://www.kadonis.org/. * Anand _et al._ [1979] R. P. Anand, M. L. Jhingan, D. Bhattacharya, and E. Kondaiah, Nuovo Cim. A 50, 247 (1979). * Walter _et al._ [1986] G. Walter, H. Beer, F. Käppeler, G. Reffo, and F. Fabbri, Astron. Astrophys. 167, 186 (1986). * Göbel _et al._ [2021] K. Göbel, C. Beinrucker, B. Brückner, P. Erbacher, S. Fiebiger, M. Fonseca, M. Heftrich, T. Heftrich, F. Käppeler, A. Krása, D. Kurtulgil, C. Lederer-Woods, R. Plag, A. Plompen, R. Reifarth, S. Schmidt, K. Sonnabend, and M. Weigand, Phys. Rev. C 103, 025802 (2021). * Paul _et al._ [2019] M. Paul, M. Tessler, M. Friedman, S. Halfon, T. Palchan, L. Weissman, A. Arenshtam, D. Berkovits, Y. Eisen, I. Eliahu, G. Feinberg, D. Kijel, A. Kreisel, I. Mardor, G. Shimel, A. Shor, and I. Silverman, Eur. Phys. J. A 55, 44 (2019). * Halfon _et al._ [2013] S. Halfon, A. Arenshtam, D. Kijel, M. Paul, D. Berkovits, I. Eliyahu, G. Feinberg, M. Friedman, N. Hazenshprung, I. Mardor, A. Nagler, G. Shimel, M. Tessler, and I. Silverman, Rev. Sci. Instrum. 84, 123507 (2013). * Halfon _et al._ [2014] S. Halfon, A. Arenshtam, D. Kijel, M. Paul, L. Weissman, O. Aviv, D. Berkovits, O. Dudovitch, Y. Eisen, I. Eliyahu, G. Feinberg, G. Haquin, N. Hazenshprung, A. Kreisel, I. Mardor, G. Shimel, A. Shor, I. Silverman, M. Tessler, and Z. Yungrais, Rev. Sci. Instrum. 85, 056105 (2014). * Kreisel _et al._ [2014] A. Kreisel, L. Weissman, A. Arenshtam, Y. B. Aliz, D. Berkovits, Y. Buzaglo, O. Dudovich, Y. Eisen, I. Eliyahu, G. Feinberg, I. Fishman, I. Gertz, A. Grin, S. Halfon, Y. Haruvy, T. Hirsh, D. Hirschmann, Z. Horvitz, B. Kaizer, D. Kijel, Y. Luner, I. Mor, J. Rodnizki, G. Shimel, A. Shor, I. Silverman, D. Vartsky, and E. Zemach, in _proceedings of linac 2014_, 770 (Geneva, Switzerland, 2014) p. WEIOB02. * Mardor _et al._ [2016] I. Mardor, D. Berkovits, S. Halfon, T. Hirsh, Y. Mishnayot, I. Silverman, S. Vaintraub, L. Weissman, M. Hass, I. Mukul, B. Ohayon, M. Paul, G. Ron, M. Tessler, and T. Dickel, in _proceedings of The 26th International Nuclear Physics Conference_, 109 (Adelaide, Australia, 2016) p. 1. * Mardor _et al._ [2018] I. Mardor, O. Aviv, M. Avrigeanu, D. Berkovits, A. Dahan, T. Dickel, I. Eliyahu, M. Gai, I. Gavish-Segev, S. Halfon, M. Hass, T. Hirsh, B. Kaiser, D. Kijel, A. Kreisel, Y. Mishnayot, I. Mukul, B. Ohayon, M. Paul, A. Perry, H. Rahangdale, J. Rodnizki, G. Ron, R. Sasson-Zukran, A. Shor, I. Silverman, M. Tessler, S. Vaintraub, and L. Weissman, Eur. Phys. J. A 54, 91 (2018). * Ga [2] Sigma-Aldrich Co., https://www.sigmaaldrich.com/catalog/product/aldrich/215066. * Feinberg [2014] G. Feinberg, _Study of the 7Li(p,n) Reaction Towards Measurements of Neutron-Capture Cross Sections in the Astrophysical s-process With the SARAF Accelerator and a Liquid-Lithium Target_ , Ph.D. thesis, The Hebrew University of Jerusalem (2014), available online: http://arad.mscc.huji.ac.il/dissertations/W/JSL/001975970.pdf. * Gürdal and McCutchan [2016] G. Gürdal and E. McCutchan, Nucl. Data Sheets 136, 1 (2016). * Schrewe and Schmidt-Ott [1977] U. J. Schrewe and W. D. Schmidt-Ott, Z. Phys. A 128, 125 (1977). * Tessler _et al._ [2015] M. Tessler, M. Paul, A. Arenshtam, G. Feinberg, M. Friedman, S. Halfon, D. Kijel, L. Weissman, O. Aviv, D. Berkovits, Y. Eisen, I. Eliyahu, G. Haquin, A. Kreisel, I. Mardor, G. Shimel, A. Shor, I. Silverman, and Z. Yungrais, Phys. Lett. B 751, 418 (2015). * Xiaolong [2009] H. Xiaolong, Nucl. Data Sheets 110, 2533 (2009). * Abriola and Sonzogni [2010] D. Abriola and A. Sonzogni, Nucl. Data Sheets 111, 1 (2010). * Hubbell and Seltzer [2004] J. Hubbell and S. Seltzer, _Tables of X-Ray Mass Attenuation Coefficients and Mass Energy-Absorption Coefficients (version 1.4)_, technical report (National Institute of Standards and Technology, Gaithersburg, MD, 2004). * Brown _et al._ [2018] D. Brown, M. Chadwick, R. Capote, A. Kahler, A. Trkov, M. Herman, A. Sonzogni, Y. Danon, A. Carlson, M. Dunn, D. Smith, G. Hale, G. Arbanas, R. Arcilla, C. Bates, B. Beck, B. Becker, F. Brown, R. Casperson, J. Conlin, D. Cullen, M.-A. Descalle, R. Firestone, T. Gaines, K. Guber, A. Hawari, J. Holmes, T. Johnson, T. Kawano, B. Kiedrowski, A. Koning, S. Kopecky, L. Leal, J. Lestone, C. Lubitz, J. M. Damián, C. Mattoon, E. McCutchan, S. Mughabghab, P. Navratil, D. Neudecker, G. Nobre, G. Noguere, M. Paris, M. Pigni, A. Plompen, B. Pritychenko, V. Pronyaev, D. Roubtsov, D. Rochman, P. Romano, P. Schillebeeckx, S. Simakov, M. Sin, I. Sirakov, B. Sleaford, V. Sobes, E. Soukhovitskii, I. Stetcu, P. Talou, I. Thompson, S. van der Marck, L. Welser-Sherrill, D. Wiarda, M. White, J. Wormald, R. Wright, M. Zerkle, G. Zerovnik, and Y. Zhu, Nucl. Data Sheets 148, 1 (2018). * Lederer _et al._ [2011] C. Lederer, N. Colonna, C. Domingo-Pardo, F. Gunsing, F. Käppeler, C. Massimi, A. Mengoni, A. Wallner, U. Abbondanno, G. Aerts, H. Álvarez, F. Álvarez-Velarde, S. Andriamonje, J. Andrzejewski, P. Assimakopoulos, L. Audouin, G. Badurek, M. Barbagallo, P. Baumann, F. Becvar, F. Belloni, E. Berthoumieux, M. Calviani, F. Calviño, D. Cano-Ott, R. Capote, C. Carrapico, A. Carrillo de Albornoz, P. Cennini, V. Chepel, E. Chiaveri, G. Cortes, A. Couture, J. Cox, M. Dahlfors, S. David, I. Dillmann, R. Dolfini, W. Dridi, I. Duran, C. Eleftheriadis, M. Embid-Segura, L. Ferrant, A. Ferrari, R. Ferreira-Marques, L. Fitzpatrick, H. Frais-Koelbl, K. Fujii, W. Furman, I. Goncalves, E. González-Romero, A. Goverdovski, F. Gramegna, E. Griesmayer, C. Guerrero, B. Haas, R. Haight, M. Heil, A. Herrera-Martinez, M. Igashira, S. Isaev, E. Jericha, Y. Kadi, D. Karadimos, D. Karamanis, M. Kerveno, V. Ketlerov, P. Koehler, V. Konovalov, E. Kossionides, M. Krticka, C. Lampoudis, H. Leeb, A. Lindote, I. Lopes, R. Losito, M. Lozano, S. Lukic, J. Marganiec, L. Marques, S. Marrone, T. Martínez, P. Mastinu, E. Mendoza, P. M. Milazzo, C. Moreau, M. Mosconi, F. Neves, H. Oberhummer, S. O’Brien, M. Oshima, J. Pancin, C. Papachristodoulou, C. Papadopoulos, C. Paradela, N. Patronis, A. Pavlik, P. Pavlopoulos, L. Perrot, M. T. Pigni, R. Plag, A. Plompen, A. Plukis, A. Poch, J. Praena, C. Pretel, J. Quesada, T. Rauscher, R. Reifarth, M. Rosetti, C. Rubbia, G. Rudolf, P. Rullhusen, J. Salgado, C. Santos, L. Sarchiapone, R. Sarmento, I. Savvidis, C. Stephan, G. Tagliente, J. L. Tain, D. Tarrío, L. Tassan-Got, L. Tavora, R. Terlizzi, G. Vannini, P. Vaz, A. Ventura, D. Villamarin, V. Vlachoudis, R. Vlastou, F. Voss, S. Walter, H. Wendler, M. Wiescher, and K. Wisshak (n_TOF Collaboration), Phys. Rev. C 83, 034608 (2011). * Massimi _et al._ [2014] C. Massimi, B. Becker, E. Dupont, S. Kopecky, C. Lampoudis, R. Massarczyk, M. Moxon, V. Pronyaev, P. Schillebeeckx, I. Sirakov, and R. Wynants, Eur. Phys. J. A 50, 124 (2014). * Feinberg _et al._ [2012] G. Feinberg, M. Friedman, A. Krása, A. Shor, Y. Eisen, D. Berkovits, D. Cohen, G. Giorginis, T. Hirsh, M. Paul, A. J. M. Plompen, and E. Tsuk, Phys. Rev. C 85, 055810 (2012). * Friedman _et al._ [2013] M. Friedman, D. Cohen, M. Paul, D. Berkovits, Y. Eisen, G. Feinberg, G. Giorginis, S. Halfon, A. Krása, A. Plompen, and A. Shor, Nucl. Instrum. Methods Phys. Res. A 698, 117 (2013). * Agostinelli _et al._ [2003] S. Agostinelli, J. Allison, K. Amako, J. Apostolakis, H. Araujo, P. Arce, M. Asai, D. Axen, S. Banerjee, G. Barrand, F. Behner, L. Bellagamba, J. Boudreau, L. Broglia, A. Brunengo, H. Burkhardt, S. Chauvie, J. Chuma, R. Chytracek, G. Cooperman, G. Cosmo, P. Degtyarenko, A. Dell’Acqua, G. Depaola, D. Dietrich, R. Enami, A. Feliciello, C. Ferguson, H. Fesefeldt, G. Folger, F. Foppiano, A. Forti, S. Garelli, S. Giani, R. Giannitrapani, D. Gibin, J. G. Cadenas, I. González, G. G. Abril, G. Greeniaus, W. Greiner, V. Grichine, A. Grossheim, S. Guatelli, P. Gumplinger, R. Hamatsu, K. Hashimoto, H. Hasui, A. Heikkinen, A. Howard, V. Ivanchenko, A. Johnson, F. Jones, J. Kallenbach, N. Kanaya, M. Kawabata, Y. Kawabata, M. Kawaguti, S. Kelner, P. Kent, A. Kimura, T. Kodama, R. Kokoulin, M. Kossov, H. Kurashige, E. Lamanna, T. Lampén, V. Lara, V. Lefebure, F. Lei, M. Liendl, W. Lockman, F. Longo, S. Magni, M. Maire, E. Medernach, K. Minamimoto, P. M. de Freitas, Y. Morita, K. Murakami, M. Nagamatu, R. Nartallo, P. Nieminen, T. Nishimura, K. Ohtsubo, M. Okamura, S. O’Neale, Y. Oohata, K. Paech, J. Perl, A. Pfeiffer, M. Pia, F. Ranjard, A. Rybin, S. Sadilov, E. D. Salvo, G. Santin, T. Sasaki, N. Savvas, Y. Sawada, S. Scherer, S. Sei, V. Sirotenko, D. Smith, N. Starkov, H. Stoecker, J. Sulkimo, M. Takahata, S. Tanaka, E. Tcherniaev, E. S. Tehrani, M. Tropeano, P. Truscott, H. Uno, L. Urban, P. Urban, M. Verderi, A. Walkden, W. Wander, H. Weber, J. Wellisch, T. Wenaus, D. Williams, D. Wright, T. Yamada, H. Yoshida, and D. Zschiesche, Nucl. Instrum. Methods Phys. Res. A 506, 250 (2003). * Liskien and Paulsen [1975] H. Liskien and A. Paulsen, At. Data Nucl. Data Tables 15, 57 (1975). * Ziegler _et al._ [2010] J. Ziegler, M. Ziegler, and J. Biersack, Nucl. Instrum. and Methods in Phys. Res. B 268, 1818 (2010), 19th International Conference on Ion Beam Analysis. * Tessler _et al._ [2018] M. Tessler, M. Paul, S. Halfon, B. S. Meyer, R. Pardo, R. Purtschert, K. E. Rehm, R. Scott, M. Weigand, L. Weissman, S. Almaraz-Calderon, M. L. Avila, D. Baggenstos, P. Collon, N. Hazenshprung, Y. Kashiv, D. Kijel, A. Kreisel, R. Reifarth, D. Santiago-Gonzalez, A. Shor, I. Silverman, R. Talwar, D. Veltum, and R. Vondrasek, Phys. Rev. Lett. 121, 112701 (2018). * Shibata _et al._ [2011] K. Shibata, O. Iwamoto, T. Nakagawa, N. Iwamoto, A. Ichihara, S. Kunieda, S. Chiba, K. Furutaka, N. Otuka, T. Ohasawa, T. Murata, H. Matsunobu, A. Zukeran, S. Kamada, and J. Katakura, J. Nucl. Sci. Technol. 48, 1 (2011). * Plompen _et al._ [2020] A. J. M. Plompen, O. Cabellos, C. De Saint Jean, M. Fleming, A. Algora, M. Angelone, P. Archier, E. Bauge, O. Bersillon, A. Blokhin, F. Cantargi, A. Chebboubi, C. Diez, H. Duarte, E. Dupont, J. Dyrda, B. Erasmus, L. Fiorito, U. Fischer, D. Flammini, D. Foligno, M. R. Gilbert, J. R. Granada, W. Haeck, F.-J. Hambsch, P. Helgesson, S. Hilaire, I. Hill, M. Hursin, R. Ichou, R. Jacqmin, B. Jansky, C. Jouanne, M. A. Kellett, D. H. Kim, H. I. Kim, I. Kodeli, A. J. Koning, A. Y. Konobeyev, S. Kopecky, B. Kos, A. Krása, L. C. Leal, N. Leclaire, P. Leconte, Y. O. Lee, H. Leeb, O. Litaize, M. Majerle, J. I. Márquez Damián, F. Michel-Sendis, R. W. Mills, B. Morillon, G. Noguère, M. Pecchia, S. Pelloni, P. Pereslavtsev, R. J. Perry, D. Rochman, A. Röhrmoser, P. Romain, P. Romojaro, D. Roubtsov, P. Sauvan, P. Schillebeeckx, K. H. Schmidt, O. Serot, S. Simakov, I. Sirakov, H. Sjöstrand, A. Stankovskiy, J. C. Sublet, P. Tamagno, A. Trkov, S. van der Marck, F. Álvarez-Velarde, R. Villari, T. C. Ware, K. Yokoyama, and G. Žerovnik, Eur. Phys. J. A 56, 181 (2020), data available online at: https://www.oecd-nea.org/dbdata/jeff/jeff33/index.html. * Ge, Zhigang _et al._ [2020] Ge, Zhigang, Xu, Ruirui, Wu, Haicheng, Zhang, Yue, Chen, Guochang, Jin, Yongli, Shu, Nengchuan, Chen, Yongjing, Tao, Xi, Tian, Yuan, Liu, Ping, Qian, Jing, Wang, Jimin, Zhang, Huanyu, Liu, Lile, and Huang, Xiaolong, Eur. Phys. J. Web Conf. 239, 09001 (2020). * Koning _et al._ [2019] A. Koning, D. Rochman, J.-Ch.Sublet, N.Dzysiuk, M.Fleming, and S. der Marck, Nucl. Data Sheets 155, 1 (2019). * [36] Kadonis v1.0, available online: https://exp-astro.de/kadonis1.0/. * Massimi _et al._ [2010] C. Massimi, C. Domingo-Pardo, G. Vannini, L. Audouin, C. Guerrero, U. Abbondanno, G. Aerts, H. Álvarez, F. Álvarez-Velarde, S. Andriamonje, J. Andrzejewski, P. Assimakopoulos, G. Badurek, P. Baumann, F. Becvar, F. Belloni, E. Berthoumieux, F. Calviño, M. Calviani, D. Cano-Ott, R. Capote, C. Carrapico, P. Cennini, V. Chepel, E. Chiaveri, N. Colonna, G. Cortes, A. Couture, J. Cox, M. Dahlfors, S. David, I. Dillmann, W. Dridi, I. Duran, C. Eleftheriadis, L. Ferrant, A. Ferrari, R. Ferreira-Marques, K. Fujii, W. Furman, S. Galanopoulos, I. F. Goncalves, E. González-Romero, F. Gramegna, F. Gunsing, B. Haas, R. Haight, M. Heil, A. Herrera-Martinez, M. Igashira, E. Jericha, F. Käppeler, Y. Kadi, D. Karadimos, D. Karamanis, M. Kerveno, P. Koehler, E. Kossionides, M. Krticka, C. Lampoudis, C. Lederer, H. Leeb, A. Lindote, I. Lopes, M. Lozano, S. Lukic, J. Marganiec, S. Marrone, T. Martinez, P. Mastinu, E. Mendoza, A. Mengoni, P. M. Milazzo, C. Moreau, M. Mosconi, F. Neves, H. Oberhummer, S. O’Brien, J. Pancin, C. Papadopoulos, C. Paradela, A. Pavlik, P. Pavlopoulos, G. Perdikakis, L. Perrot, M. T. Pigni, R. Plag, A. Plompen, A. Plukis, A. Poch, J. Praena, C. Pretel, J. Quesada, T. Rauscher, R. Reifarth, M. Rosetti, C. Rubbia, G. Rudolf, P. Rullhusen, L. Sarchiapone, R. Sarmento, I. Savvidis, C. Stephan, G. Tagliente, J. L. Tain, L. Tassan-Got, L. Tavora, R. Terlizzi, P. Vaz, A. Ventura, D. Villamarin, V. Vlachoudis, R. Vlastou, F. Voss, S. Walter, M. Wiescher, and K. Wisshak (n_TOF Collaboration), Phys. Rev. C 81, 044616 (2010). * Ratynski and Käppeler [1988] W. Ratynski and F. Käppeler, Phys. Rev. C 37, 595 (1988). * Chadwick _et al._ [2011] M. B. Chadwick, M. Herman, P. Obložinský, M. Dunn, Y. Danon, A. Kahler, D. Smith, B. Pritychenko, G. Arbanas, R. Arcilla, R. Brewer, D. Brown, R. Capote, A. Carlson, Y. Cho, H. Derrien, K. Guber, G. Hale, S. Hoblit, S. Holloway, T. Johnson, T. Kawano, B. Kiedrowski, H. Kim, S. Kunieda, N. Larson, L. Leal, J. Lestone, R. Little, E. McCutchan, R. MacFarlane, M. MacInnes, C. Mattoon, R. McKnight, S. Mughabghab, G. Nobre, G. Palmiotti, A. Palumbo, M. Pigni, V. Pronyaev, R. Sayer, A. Sonzogni, N. Summers, P. Talou, I. Thompson, A. Trkov, R. Vogt, S. van der Marck, A. Wallner, M. White, D. Wiarda, and P. Young, Nucl. Data Sheets 112, 2887 (2011), special Issue on ENDF/B-VII.1 Library. * Koning _et al._ [2015] A. Koning, D. Rochman, J. Kopecky, J. Sublet, M. Fleming, E. Bauge, S. Hilaire, P. Romain, B. Morillon, H. Duarte, S. van der Marck, S. Pomp, H. Sjostrand, R. Forrest, H. Henriksson, O. Cabellos, J. L. S. Goriely and, A. P. H. Leeb and, and R. Mills, Tendl-2015: Talys-based evaluated nuclear data library (2015). * Bao _et al._ [2000] Z. Bao, H. Beer, F. Käppeler, F. Voss, K. Wisshak, and T. Rauscher, At. Data Nucl. Data Tables 76, 70 (2000). * Kappeler _et al._ [1989] F. Kappeler, H. Beer, and K. Wisshak, Rep. Prog. Phys. 52, 945 (1989). * Hynes and Gyngard [2009] K. M. Hynes and F. Gyngard, in _40th Lunar and Planetary Science Conference_ (Lunar and Planetary Institute, Houston, 2009) p. Abstract #1198. * Stephan _et al._ [2020] T. Stephan, M. Bose, A. Boujibar, A. M. Davis, C. J. Dory, F. Gyngard, P. Hoppe, K. M. Hynes, N. Liu, L. R. Nittler, R. C. Ogliore, and R. Trappitsch, in _51st Lunar and Planetary Science Conference_ (Lunar and Planetary Institute, Houston, 2020) p. Abstract #2140. * [45] Presolar grains database, https://presolar.physics.wustl.edu/presolar-grain-database/. * Kumar _et al._ [2017] V. B. Kumar, M. Monte, O. Mathon, S. Pascarelli, Z. Porat, and A. Gedanken, J. Am. Ceram. Soc. 100, 3305 (2017). * Lodders [2003] K. Lodders, Astrophys. J. 591, 1220 (2003).
# Reliable Quantum Certification of Bosonic Code Preparations Ya-Dong Wu QICI Quantum Information and Computation Initiative, Department of Computer Science, The University of Hong Kong, Pokfulam Road, Hong Kong ###### Abstract Bosonic fault tolerant quantum computing requires preparations of Bosonic code states like cat states and GKP states with high fidelity and reliable quantum certification of these states. Although many proposals on preparing these states have been developed, few investigation has been done on how to reliably certify these experimental states. In this paper, we develop approaches to certify whether a continuous-variable state falls inside a certain Bosonic qubit code space by detecting a witness using Gaussian measurements. Our results can be applied to certification of various cat codes including two- component cat states, four-component cat states, and squeezed two-component cat states as well as Gottesman-Kitaev-Preskill codes. Then we further apply our approach to certify resource states in continuous-variable universal fault tolerant measurement-based quantum computing and quantum outputs of instantaneous quantum polynomial circuits, which can be used to show quantum supremacy. The sample complexity of our approach is efficient in terms of number of modes, so significantly reducing overhead compared to quantum state tomography in certification of many-mode quantum states. ## I Introduction A scalable fault tolerant universal quantum computer demands quantum error correction codes to protect qubit information from decoherence noises Shor (1995); Nielsen and Chuang (2010). One promising type of quantum error correcting codes is Bosonic quantum codes that encode a logical qubit into a quantum harmonic oscillator, such that a two-dimensional code space lies in a corner of an infinite-dimensional Hilbert space. This type of quantum error correcting codes include cat codes Cochrane _et al._ (1999); Ralph _et al._ (2003); Leghtas _et al._ (2013); Mirrahimi _et al._ (2014); Bergmann and van Loock (2016), Gottesman-Kiteav-Preskill (GKP) codes Gottesman _et al._ (2001), and some other rotational symmetric codes Michael _et al._ (2016); Albert _et al._ (2018); Grimsmo _et al._ (2020). Significant experimental progress have been made on encoding qubit information into cat states Ofek _et al._ (2016) and GKP states Flühmann _et al._ (2019); Campagne-Ibarcq _et al._ (2020) on superconducting cavities as well as trapped ions, along with many theoretical proposals on state preparation and engineering Terhal and Weigand (2016); Puri _et al._ (2017); Weigand and Terhal (2018); Shi _et al._ (2019). Then a natural question is how to experimentally characterize these prepared states in harmonic oscillators. To implement fault tolerant quantum computing with Bosonic quantum error correcting codes, efficient and reliable quantum certification of scalable Bosonic code state preparations is significant. The most common approach to characterize a quantum state in experiments is quantum state tomography Lvovsky and Raymer (2009); D’Ariano _et al._ (2003), which, however, requires a huge number of measurements and long classical postprocessing time. Quantum state tomography cannot handle multi-mode entangled nonGaussian states. Quantum tomography with neural networks Tiunov _et al._ (2020); Ahmed _et al._ (2021a, b); Zhu _et al._ (2022); Wu _et al._ (2022) evidently reduces the required number of measurements and shorten the classical postprocessing time, but still cannot deal with multi-mode quantum states. In practical applications, physicists have the prior knowledge about the classical description of a target quantum state. Hence, rather than fully characterizing an experimental quantum state, in these scenarios, physicists can apply quantum certification Eisert _et al._ (2020), to efficiently determine whether a prepared quantum state is close enough to certain target state. L. Aolita et al. investigated quantum certification of photonic states Aolita _et al._ (2015), including Gaussian states, and those states prepared in Boson sampling Aaronson and Arkhipov (2013) and Knill–Laflamme–Milburn (KLM) schemes Knill _et al._ (2001). Later U. Chabaud developed another verification protocol of quantum outputs in Boson sampling Chabaud _et al._ (2021). Up to now, how to certify cat states and GKP states, without performing quantum state tomography or fidelity estimation da Silva _et al._ (2011); Chabaud _et al._ (2020a), is still open. On the other hand, although lots of work have been done on efficient verification of many-qubit states, different nature between CV states and qubit states makes certification approaches of many-qubit states Pallister _et al._ (2018) cannot be directly applied in CV quantum systems Liu _et al._ (2021). In this paper, we generalize the concept of fidelity witness Aolita _et al._ (2015); Gluza _et al._ (2018) to witness of Bosonic code space, and propose protocols to certify both two-component cat code space, four-component cat code space, and a realistic GKP code space with finite truncation in phase space. The measurements in all the protocols are experimentally friendly, being either homodyne detections or heterodyne detections. We propose certification protocols for both the resource states of CV fault-tolerant measurement-based quantum computing Menicucci (2014) and the quantum outputs of CV instantaneous quantum polynomial-time (IQP) circuits Douce _et al._ (2017) by estimating fidelity witnesses. ## II Certification of Cat State Code Space Let $\mathcal{H}$ denote the infinite-dimensional Hilbert space of a quantum harmonic oscillator. A quantum device outputs $n$ copies of quantum states $\rho^{\otimes n}$ on $\mathcal{H}^{\otimes n}$. Given a Bosonic code logical subspace $\bar{\mathcal{H}}\subset\mathcal{H}$, an experimenter, who has no knowledge of $\rho$, wants to determine whether $\operatorname{tr}_{\bar{\mathcal{H}}}\rho\geq 1-\epsilon$, by applying local measurement at each copy of $\rho$. To certify whether a single-mode state $\rho$ falls on $\bar{\mathcal{H}}$, we measure an observable $W$, which satisfies the following conditions: (i) if $\rho$ falls on $\bar{\mathcal{H}}$, i.e. $\operatorname{tr}_{\bar{\mathcal{H}}}\rho=1$, then $\braket{W}_{\rho}=1$; (ii) otherwise, $\braket{W}_{\rho}<1$. We call the observable $W$ a witness of the code space $\bar{\mathcal{H}}$. By estimating $\braket{W}_{\rho}$, we can determine whether to accept $\rho$ as a reliable Bosonic code state or not. In this subsection, we introduce code witnesses for different cat codes and explain how to estimate their mean values by Gaussian measurements. A Schrodinger cat state is a superposition of coherent states with opposite amplitudes Bužek and Knight (1995) and can be used to encode qubit information Cochrane _et al._ (1999). We call this code two-component cat code to differentiate it from the cat code with a superposition of four coherent states. A two-component cat code space $\bar{\mathcal{H}}$ is spanned by $\displaystyle\ket{\bar{0}}=\frac{1}{\sqrt{2(1+e^{-2|\alpha|^{2}})}}\left(\ket{\alpha}+\ket{-\alpha}\right)$ $\displaystyle\ket{\bar{1}}=\frac{1}{\sqrt{2(1+e^{-2|\alpha|^{2}})}}\left(\ket{\alpha}-\ket{-\alpha}\right).$ A two-component cat code witness is $W=\mathds{1}-\frac{\left(\hat{a}^{\dagger 2}-\alpha^{*\,2}\right)\left(\hat{a}^{2}-\alpha^{2}\right)}{2}.$ (1) The two-component cat code space is the degenerate ground space of the Hamiltonian $\hat{H}_{tCat}=\left(\hat{a}^{\dagger 2}-\alpha^{*\,2}\right)\left(\hat{a}^{2}-\alpha^{2}\right)$ (2) with eigenvalue zero. $\hat{H}$ can be written as $\hat{H}_{tCat}=0\left(\ket{\bar{0}}\bra{\bar{0}}_{tCat}+\ket{\bar{1}}\bra{\bar{1}}_{tCat}\right)+\lambda_{1}\ket{\psi}\bra{\psi}+\cdots$ (3) where $\ket{\psi}$ is the first excited state of $\hat{H}_{tCat}$ and $\lambda_{1}$ is the first excited energy. Then $\mathds{1}-\frac{\hat{H}_{tCat}}{2}=\ket{\bar{0}}\bra{\bar{0}}_{tCat}+\ket{\bar{1}}\bra{\bar{1}}_{tCat}+\left(1-\frac{\lambda_{1}}{2}\right)\ket{\psi}\bra{\psi}-\cdots$ (4) As $\lambda_{1}\geq 2$, we get $\mathds{1}-\frac{\hat{H}_{tCat}}{2}\leq\ket{\bar{0}}\bra{\bar{0}}_{tCat}+\ket{\bar{1}}\bra{\bar{1}}_{tCat}.$ (5) Furthermore, any state $\rho$ satisfying $\operatorname{tr}\left[\rho\left(\mathds{1}-\frac{\hat{H}_{tCat}}{2}\right)\right]=1$ if and only if $\rho$ falls on the two-component cat code space. The witness of squeezed two-component cat code space can be obtained by using the fact that squeezing operation is a Gaussian operation inducing a linear transformation on phase space. To clarify how to measure this witness, for simplicity, we rewrite it in terms of quadrature operators when $\alpha\in\mathbb{R}$, $\displaystyle W_{tCat}$ $\displaystyle=\frac{3-2\alpha^{4}}{4}\mathds{1}-\frac{1}{16}(\hat{x}^{4}+\hat{p}^{4})-\frac{1}{12}\Bigg{[}\left(\frac{\hat{x}+\hat{p}}{\sqrt{2}}\right)^{4}+$ $\displaystyle\left(\frac{\hat{x}-\hat{p}}{\sqrt{2}}\right)^{4}\Bigg{]}+\frac{1}{2}\left(1+\alpha^{2}\right)\hat{x}^{2}+\frac{1}{2}\left(1-\alpha^{2}\right)\hat{p}^{2},$ (6) where $\hat{x}=\frac{\hat{a}+\hat{a}^{\dagger}}{\sqrt{2}}$ and $\hat{p}=\frac{\hat{a}-\hat{a}^{\dagger}}{\sqrt{2}\text{i}}$ are the position and momentum operators. The expectation value of such a witness can be estimated by homodyne detections on at most four different quadrature bases, i.e. $\hat{x}$, $\hat{p}$, $\frac{\hat{x}+\hat{p}}{\sqrt{2}}$ and $\frac{\hat{x}-\hat{p}}{\sqrt{2}}$. The witness (1) can also be rewritten as $W_{tCat}=-\frac{1}{2}\hat{a}^{2}\hat{a}^{\dagger\,2}+2\hat{a}\hat{a}^{\dagger}+\frac{1}{2}(\alpha^{*2}\hat{a}^{2}+\alpha^{2}\hat{a}^{\dagger\,2})-\frac{1}{2}|\alpha|^{4}.$ (7) Each term in the above expression is a product of annihilation operators and creation operators in anti-normal order. Using optical equivalence principle Scully _et al._ (1997), i.e., $\braket{\hat{a}^{m}\hat{a}^{\dagger\,n}}=\int\text{d}^{2}\alpha Q_{\rho}(\alpha)\alpha^{m}\alpha^{*n},\quad m,n\in\mathbb{N}$ (8) $\braket{W_{tCat}}_{\rho}$ can be estimated by sampling Humusi-Q function $Q_{\rho}(\alpha):=\braket{\alpha}{\rho}{\alpha}$ using heterodyne detections. Four-component cat code, which is superposition of four coherent states, is introduced to protect qubit information from photon loss errors Mirrahimi _et al._ (2014). A four-component cat code space $\bar{\mathcal{H}}$ with even parity is spanned by $\displaystyle\bar{\ket{0}}=\frac{1}{\sqrt{2(1+e^{-2|\alpha|^{2}})}}\left(\ket{\alpha}+\ket{-\alpha}\right),$ $\displaystyle\bar{\ket{1}}=\frac{1}{\sqrt{2(1+e^{-2|\alpha|^{2}})}}\left(\ket{\text{i}\alpha}+\ket{-\text{i}\alpha}\right).$ To certify whether a single-mode state $\rho$ falls on $\bar{\mathcal{H}}$, we measure the witness $W_{fCat}=\frac{\mathds{1}+(-1)^{\hat{n}}}{2}-\frac{\hat{H}_{fCat}}{24},$ (9) where $\hat{n}=\hat{a}^{\dagger}\hat{a}$ is photon number operator, and $\hat{H}_{fCat}:=\left(\hat{a}^{\dagger 2}+\alpha^{*2}\right)\left(\hat{a}^{\dagger 2}-\alpha^{*2}\right)\left(\hat{a}^{2}-\alpha^{2}\right)\left(\hat{a}^{2}+\alpha^{2}\right)$. The first term is the parity of photon number, equal to the value of Wigner function at origin, and can be measured by a parity measurement. The second term can be written as $\displaystyle\frac{\hat{H}_{fCat}}{24}=$ $\displaystyle\hat{a}^{4}\hat{a}^{\dagger\,4}/24-2/3\hat{a}^{3}\hat{a}^{\dagger\,3}+3\hat{a}^{2}\hat{a}^{\dagger\,2}-4\hat{a}\hat{a}^{\dagger}-\alpha^{4}/24\hat{a}^{\dagger\,4}$ $\displaystyle-\alpha^{*\,4}/24\hat{a}^{4}+(|\alpha|^{8}/24+1)\mathds{1}.$ (10) Again the expectation value of each term can be estimated by heterodyne detections. A squeezed two-component cat code Schlegel _et al._ (2022) consists of superpositions of two squeezed coherent states $\displaystyle\ket{\bar{0}}\propto\left(D(-\alpha)+D(\alpha)\right)S(r)\ket{0}$ $\displaystyle\ket{\bar{1}}\propto\left(D(-\alpha)-D(\alpha)\right)S(r)\ket{0},$ where $S(r):=\text{e}^{1/2(r^{*}\hat{a}^{2}-r\hat{a}^{\dagger\,2})}$ is a squeezing operation, $r\in\mathbb{C}$ is a squeezing parameter, and $\text{arg}(r)=\text{arg}(\alpha)/2$ such that the squeezed quadrature is in the same direction as that of the displacement operation. For simplicity, we suppose $\alpha\in\mathbb{R}$. From the fact that $D(\alpha)S(r)=S(r)D(\alpha\text{e}^{r})$, we obtain a witness of the squeezed two-component cat code space $W_{stCat}=\mathds{1}-\frac{S(r)\left(\hat{a}^{\dagger 2}-\alpha^{2}\text{e}^{2r}\right)\left(\hat{a}^{2}-\alpha^{2}\text{e}^{2r}\right)S(r)^{\dagger}}{2}.$ (11) Using the fact that $S(r)\hat{x}S(r)^{\dagger}=\text{e}^{-r}\hat{x}$ and $S(r)\hat{p}S(r)^{\dagger}=\text{e}^{r}\hat{p}$, the witness (11) can be rewritten in terms of quadrature operators $\displaystyle W_{stCat}=$ $\displaystyle\frac{3-2\alpha^{4}\text{e}^{4r}}{4}\mathds{1}-\frac{1}{16}\left(\text{e}^{-4r}\hat{x}^{4}+\text{e}^{4r}\hat{p}^{4}\right)$ $\displaystyle-\frac{1}{48}\left[\left(\text{e}^{-r}\hat{x}+\text{e}^{r}\hat{p}\right)^{4}+\left(\text{e}^{-r}\hat{x}-\text{e}^{r}\hat{p}\right)^{4}\right]$ $\displaystyle+\frac{1}{2}\left(\text{e}^{-2r}+\alpha^{2}\right)\hat{x}^{2}+\frac{1}{2}\left(1-\alpha^{2}\text{e}^{2r}\right)\text{e}^{2r}\hat{p}^{2},$ which can be estimated by applying homodyne detections at four quadratures. ## III Certification of a realistic GKP state GKP states are defined to be simultaneous eigenstates of two commutative displacement operators $S_{q}=e^{2\text{i}\sqrt{\pi}\hat{x}}$ and $S_{p}=e^{-2\text{i}\sqrt{\pi}\hat{p}}$, with eigenvalue one Gottesman _et al._ (2001). A qubit is encoded into a GKP state in the following way $\displaystyle\ket{\bar{0}}_{iGKP}$ $\displaystyle:=\sum_{k=-\infty}^{\infty}\delta(x-2k\sqrt{\pi})\ket{x},$ (12) $\displaystyle\ket{\bar{1}}_{iGKP}$ $\displaystyle:=\sum_{k=-\infty}^{\infty}\delta\left(x-(2k+1)\sqrt{\pi}\right)\ket{x}.$ (13) Ideal GKP states are coherent superpositions of position eigenstates, demanding infinite squeezing, which is not physically realistic. From now on, we consider realistic GKP states with finite squeezing. A realistic GKP state replaces position eigenstates by finitely squeezed states, i.e., $\displaystyle\ket{\bar{0}}_{rGKP}$ $\displaystyle\propto\sum_{k=-\infty}^{\infty}e^{-4k^{2}\sigma^{2}\pi}D(2k\sqrt{\pi})\ket{\psi_{x}},$ (14) $\displaystyle\ket{\bar{1}}_{rGKP}$ $\displaystyle\propto\sum_{k=-\infty}^{\infty}e^{-(2k+1)^{2}\sigma^{2}\pi}D((2k+1)\sqrt{\pi})\ket{\psi_{x}},$ (15) where $D(\alpha):=\text{e}^{\text{i}\sqrt{2}(-\text{Re}(\alpha)\hat{p}+\text{Im}(\alpha)\hat{x})}$ is a displacement operator, $\ket{\psi_{x}}=\frac{1}{\sqrt{\sigma}\pi^{1/4}}\int\text{d}x\text{e}^{-\frac{x^{2}}{2\sigma^{2}}}\ket{x}$ is a position-squeezed vacuum state, and $0<\sigma<1$ is the variance of a Gaussian distribution. When $\sigma$ is small, a GKP state that is a superposition of Gaussian peaks with width $\sigma$, separated by $2\sqrt{\pi}$, with Gaussian envelop of width $1/\sigma$ in position basis is a superposition of Gaussian peaks with width $\sigma$, separated by $\sqrt{\pi}$, with Gaussian envelop of width $\frac{1}{\sigma}$ in momentum basis. A straightforward way to certify a GKP state is to detect the eigenvalue of two stabilizer operators $S_{q}$ and $S_{p}$ to check whether both eigenvalues are one. However, a realistic GKP state with finite squeezing can never pass this certification test. To circumvent this obstacle, one solution is to set a threshold on the eigenvalue in the certification test to pass those realistic GKP states which are close enough to the ideal GKP states. Nevertheless, this approach does not exclusively certify a realistic GKP state or a GKP code space with a certain target degree of squeezing. In this paper, we aim to certify a realistic GKP state with a certain target degree of squeezing. To certify a realistic GKP state, we must truncate the phase space to consider only finite number of superpositions of squeezed coherent states. Suppose we truncate at $x=\pm\sqrt{\pi}m$ in position quadrature and $p=\pm\sqrt{\pi}m$ in momentum quadrature, respectively, where $m\in\mathbb{N}^{+}$. Then we obtain a GKP code space witness $\displaystyle W_{rGKP}=$ $\displaystyle\mathds{1}-\frac{1}{(2m+1)!}\prod_{-m\leq k\leq m}\left(\cosh r\hat{a}^{\dagger}-\sinh r\hat{a}-\frac{\sqrt{\pi}k}{\sigma}\right)\cdot\text{h.c.}$ $\displaystyle-\frac{1}{(2m+1)!}\prod_{-m\leq k\leq m}\left(\cosh r\hat{a}^{\dagger}+\sinh r\hat{a}+\text{i}\frac{\sqrt{\pi}k}{\sigma}\right)\cdot\text{h.c.},$ (16) where $r=\frac{1}{2}\ln\frac{1}{\sigma}$ and h.c. denotes Hermitian conjugate. As GKP states are applied in fault-tolerant CV quantum computing, we use this witness to certify output state in fault-tolerant CV quantum computing. ## IV Application in verification of fault-tolerant quantum computing In this subsection, we apply the results in the last subsection to certify two important types of many-mode quantum states. The first are the resource states in universal fault-tolerant CV measurement-based quantum computing Menicucci (2014). These states are CV cluster states Menicucci _et al._ (2006); Gu _et al._ (2009); Larsen _et al._ (2019); Hastrup _et al._ (2021), attached with GKP states. The second are the output states of CV IQP circuits Douce _et al._ (2017). These states are prepared by applying unitary gates diagonal in position quadrature on the combination of momentum-squeezed states and GKP states. The preparations of both two types of target states are plotted in Fig. 1. We first slightly revise the code witness in Eq. (16) to obtain a fidelity witness of the GKP state $\ket{\bar{+}}_{rGKP}\propto\mathcal{N}_{0}\sum_{k=-\infty}^{\infty}e^{-\sigma^{2}\pi k^{2}}D(\sqrt{\pi}k)\ket{\psi_{x}}.$ (17) As $\ket{\bar{+}}_{rGKP}$ is a grid of Gaussian peaks separated by $\sqrt{\pi}$ in position quadrature and separated by $2\sqrt{\pi}$ in momentum quadrature. We obtain a fidelity witness $\displaystyle W_{\ket{\bar{+}}_{rGKP}}=\mathds{1}-$ $\displaystyle\frac{1}{(2m+1)!}\prod_{-m\leq k\leq m}\left(\cosh r\hat{a}^{\dagger}-\sinh r\hat{a}-\frac{\sqrt{\pi}k}{\sigma}\right)\cdot\text{h.c.}-$ $\displaystyle\frac{1}{(m+1)!}\prod_{-m/2\leq k\leq m/2}\left(\cosh r\hat{a}^{\dagger}+\sinh r\hat{a}+2\text{i}\frac{\sqrt{\pi}k}{\sigma}\right)\cdot\text{h.c.}$ (18) A CV cluster state can be considered as more generally a CV graph state, where each vertex represents a quantum mode and each edge connecting vertices $i$ and $j$ represents a quantum gate $CZ:=\text{e}^{\text{i}\hat{x}_{i}\otimes\hat{x}_{j}}$. Suppose a target cluster state has $N_{s}+N_{GKP}$ modes in total, where $N_{s}$ modes are initially momentum-squeezed vacuum states $\ket{\psi_{p}}:=\frac{1}{\sqrt{\sigma}\pi^{1/4}}\int\text{d}p\text{e}^{-\frac{p^{2}}{2\sigma^{2}}}\ket{p}$ and the other $N_{GKP}$ modes are initially $\ket{\bar{+}}_{rGKP}$. The target CV cluster state is obtained by applying a CZ gate at each pair of adjacent modes $i$ and $j$ in the graph. A fidelity witness of this resource state is $\displaystyle W=$ $\displaystyle\left(1+\frac{N_{s}}{2}\right)\mathds{1}-\sum_{i=1}^{N_{GKP}}\Bigg{[}\frac{1}{(2m+1)!}\prod_{-m\leq k\leq m}\left(\frac{\text{e}^{-r}}{\sqrt{2}}\hat{x}_{i}-\frac{\text{e}^{r}}{\sqrt{2}}\text{i}\hat{\tilde{p}}_{i}-\frac{\sqrt{\pi}k}{\sigma}\right)\cdot h.c.$ $\displaystyle+\frac{1}{(m+1)!}\prod_{-m/2\leq k\leq m/2}\left(\frac{\text{e}^{r}}{\sqrt{2}}\hat{x}_{i}-\frac{\text{e}^{-r}}{\sqrt{2}}\text{i}\hat{\tilde{p}}_{i}+2\text{i}\frac{\sqrt{\pi}k}{\sigma}\right)\cdot h.c.\Bigg{]}-\frac{1}{2}\sum_{i=1}^{N_{s}}\left(e^{2r}\hat{x}_{i}^{2}+e^{-2r}\hat{\tilde{p}}_{i}^{2}\right),$ (19) where $\hat{\tilde{p}}_{i}=\hat{p}_{i}-\sum_{j\in\mathcal{N}(i)}\hat{x}_{j}$ and $\mathcal{N}(i)$ denotes the set of modes adjacent to mode $i$ in the graph. The fidelity witness in Eq. (19) contains at most $(n+2)^{4m+2}$ products of quadrature operators with maximal order $4m+2$, where $n$ is the maximal number of neighborhood modes and is as most four in a square-lattice cluster state. To calculate the required sample complexity to estimate the mean value of fidelity witness, we denote $\sigma_{k}$ to be the uniform upper bound of the mean square of all products of $k$ quadrature operators and $\sigma_{\leq k}$ to be the maximum value of all $\sigma_{j}$ with $1\leq j\leq k$. Suppose $m$ is a constant, that is we always truncate the phase space of each mode at a certain fixed value in both position and momentum quadratures. If we use the approach of importance sampling Flammia and Liu (2011) to estimate $\braket{W}$, by Hoeffding’s inequality, we find that the minimum number of required copies of states to obtain an estimate $\omega$ such that $\text{Pr}\left(|\omega-\braket{W}_{\rho}|\geq\epsilon\right)\leq\delta$, is upper bounded by $O\left[\frac{\ln 1/\delta}{\epsilon^{2}}\left(N_{GKP}^{2}\text{e}^{r(4m+2)}\sigma_{\leq 4m+2}+N_{s}^{2}\text{e}^{2r}\sigma_{2}\right)\right].$ (20) This sample complexity scales polynomially in both $N_{s}$ and $N_{GKP}$. A CV IQP circuit Douce _et al._ (2017) is a uniformly random combination of three quantum gates in the set $\left\\{Z:=e^{\text{i}\hat{x}\sqrt{\pi}},CZ,T:=e^{\text{i}\frac{\pi}{4}[2(\hat{x}/\sqrt{\pi})^{3}+(\hat{x}/\sqrt{\pi})^{2}-2\hat{x}/\sqrt{\pi}]}\right\\},$ with the input of combination of $N_{s}$ copies of $\ket{\psi_{p}}$ and $N_{GKP}$ copies of $\ket{\bar{+}}_{rGKP}$. Denote $n_{Z}^{i}$, and $n_{T}^{i}$ as the number of Z gates and T gates applied at $i$th mode, respectively. Then the fidelity witness of the output state of a CV IQP circuit is given in Eq. (19) with $\hat{\tilde{p}}_{i}=\hat{p}_{i}-n_{T}^{i}(3\hat{x}_{i}^{2}/(2\sqrt{\pi})+\hat{x}_{i}/2-\sqrt{\pi}/2)-n_{Z}^{i}\sqrt{\pi}-\sum_{j\in\mathcal{N}(i)}\hat{x}_{j}$ There are at most $N_{GKP}(3+n_{CZ})^{4m+2}$ terms of products of quadrature operators with order at most $4m+2$, and the absolute value of coefficient is no larger than $(n_{T}\text{e}^{r})^{4m+2}$. Thus the sample complexity is bounded by $\displaystyle O$ $\displaystyle\Bigg{[}\frac{\ln 1/\delta}{\epsilon^{2}}\Big{(}N_{GKP}^{2}(n_{T}\text{e}^{r})^{4m+2}(n_{CZ}+3)^{8m+4}\sigma_{\leq 4m+2}+$ $\displaystyle N_{s}^{2}n_{T}^{2}(n_{CZ}+3)^{4}\sigma_{\leq 4}\Big{)}\Bigg{]}.$ The sample complexity scales polynomially with respect to both the number of modes and the number of gates. Figure 1: Diagrams of quantum circuits to generate the two types of target states. ## V Methods The four-component cat code space is the degenerate ground space of Hamiltonian $\displaystyle\hat{H}=$ $\displaystyle\left(\hat{a}^{\dagger 2}+\alpha^{*2}\right)\left(\hat{a}^{\dagger 2}-\alpha^{*2}\right)\left(\hat{a}^{2}-\alpha^{2}\right)\left(\hat{a}^{2}+\alpha^{2}\right)$ $\displaystyle=$ $\displaystyle\hat{a}^{\dagger 4}\hat{a}^{4}-(\alpha^{4}\hat{a}^{\dagger 4}+\alpha^{*4}\hat{a}^{4})+|\alpha^{2}|^{4}$ (21) with even photon parity. Following idea similar to above two-component cat code, and using the fact that $\frac{\mathds{1}+(-1)^{\hat{n}}}{2}$ is the projection onto the even parity subspace, we obtain the witness of four- component cat code space as shown in Eq. (9). A realistic GKP state is a superposition of squeezed coherent states. Each component $D(\sqrt{\pi}k)\ket{\psi_{x}}$ has the nullifier $S(r)\left(\hat{a}^{\dagger}-\frac{\sqrt{\pi}k}{\sigma}\right)\left(\hat{a}-\frac{\sqrt{\pi}k}{\sigma}\right)S(r)^{\dagger}$. Then the superposition of $D(\sqrt{\pi}k)\ket{\psi_{x}}$ for $k\in\mathbb{Z}$ has the nullifier $S(r)\prod_{k\in\mathbb{Z}}\left(\hat{a}^{\dagger}-\frac{\sqrt{\pi}k}{\sigma}\right)\prod_{l\in\mathbb{Z}}\left(\hat{a}-\frac{\sqrt{\pi}l}{\sigma}\right)S(r)^{\dagger}.$ (22) Similarly, any state $D(2\sqrt{\pi}k\text{i})\ket{\psi_{p}}$ has the nullifier $S(\text{i}r)\left(\hat{a}^{\dagger}+\text{i}\frac{2\sqrt{\pi}k}{\sigma}\right)\left(\hat{a}-\text{i}\frac{2\sqrt{\pi}k}{\sigma}\right)S(\text{i}r)^{\dagger}.$ (23) Thus, a realistic GKP state has another nullifier $S(\text{i}r)\prod_{k\in\mathbb{Z}}\left(\hat{a}^{\dagger}+\text{i}\frac{2\sqrt{\pi}k}{\sigma}\right)\prod_{l\in\mathbb{Z}}\left(\hat{a}-\text{i}\frac{2\sqrt{\pi}l}{\sigma}\right)S(\text{i}r)^{\dagger}.$ (24) Combining these two nullifiers together, we obtain the fidelity witness of a realistic GKP state as shown in Eq. (18). To obtain the fidelity witness of a cluster state attached with GKP states, we also need a witness of CV cluster state. A cluster state is constructed from tensor product of $n$ $\hat{p}$-squeezed vacuum states by applying CV CZ gates. A nullifier of each mode $1\leq i\leq n$ in a $n$-mode CV cluster state is , $\frac{e^{2r}\left(\hat{x}_{i}^{2}-\sum_{j\in\mathcal{N}(i)}\hat{x}_{j}^{2}\right)+e^{-2r}\hat{p}_{i}^{2}-1}{2}.$ (25) Combining all the nullifiers, we obtain a fidelity witness in Eq. (19). The fidelity witness of output states of IQP circuits can be obtained by noticing that $T\hat{p}T^{\dagger}=\hat{p}-3\hat{x}^{2}/(2\sqrt{\pi})-\hat{x}/2+\sqrt{\pi}/2,$ and following a similar strategy. The sample complexities for both the certification protocols of CV cluster states attached with GKP states and output states of IQP circuits can be calculated using importance sampling and Hoeffding’s inequality. Suppose a random variable $\bm{F}=\sum_{i=1}^{m}\lambda_{i}f_{i}$, where each $f_{i}$ is a polynomial function of quadrature operators. Then by importance sampling approach, Hoeffindg’s inequality implies $\text{Pr}(|F^{*}-\bm{F}|>\epsilon)\leq 8\text{e}^{-\frac{N\epsilon^{2}}{33\braket{\bm{F}^{2}}}}$, where $\braket{\bm{F}^{2}}\leq m^{2}\max_{i}|\lambda_{i}|\max_{i}\braket{f_{i}^{2}}$. As calculated in Ref. Liu _et al._ (2019); Farias and Aolita (2021), to make the failure probability of estimation less than $\delta$, the required sample complexity is upper bounded by $O\left(\frac{\ln(1/\delta)}{\epsilon^{2}}m^{2}\max_{i}|\lambda_{i}|\max_{i}\braket{f_{i}^{2}}\right)$. ## VI Discussion Bosonic quantum error correcting codes is a promising way to realize universal fault tolerant quantum computing is a qubit-into-qumode manner alternative to the KLM scheme. Preparation of bosonic code quantum states with high fidelity is an key issue for the implementation of Bosonic quantum error correction. Here we propose realistic protocols to certify the preparations of experimental Bosonic code sates using Gaussian measurements. Different from state tomography, this protocol is extended to certification of many-mode quantum states efficiently with respect to the number of modes. Most previous work on certification or verification of nonGaussian states are about nonGaussian states with finite stellar rank Chabaud _et al._ (2020b). In contrast to quantum states in Boson sampling and KLM schemes, cat states and GKP states have infinite stellar rank, indicating higher nonGaussianity. In this paper, we first develop an approach to certify whether a CV state falls inside a certain two-dimensional Bosonic code space. If a CV state passes this certification protocol, then we can ignore the CV nature of this state and consider only the logical code space. By combining our certification approach, together with quantum characterization approaches on many-qubit systems, including direct fidelity estimation Flammia and Liu (2011), qubit- state verification Pallister _et al._ (2018) and classical shadow estimation Huang _et al._ (2020), one can certify preparations of many-qubit Bsonic code states. Although here we assume independent and identical copies of prepared quantum states, we can extend this work into non-i.i.d scenario by using the technique developed in Ref. Wu _et al._ (2021). Hence, the protocol can be applied to verifiable blind fault-tolerant quantum computing Hayashi and Morimae (2015); Gheorghiu _et al._ (2019). A server prepares quantum states, which are claimed to be resource states for CV fault-tolerant quantum computing, and sends them to a client. After receiving these states, the client randomly chooses some of them to perform dimension test and fidelity test. If both tests are passed, then the remaining states can be used for measurement-based fault tolerant quantum computing with homodyne detections. Besides verification of resource states for universal quantum computing, we also propose protocols to certify output states of IQP circuits. It has been shown that the statstics of practical homodyne measurements on position quadratures of the quantum output of IQP circuits cannot be efficiently simulated by a classical computer Douce _et al._ (2017). Hence, a certified quantum output state of IQP circuits can be used to show quantum supremacy. ## VII Acknowleadgement This work was supported by funding from the Hong Kong Research Grant Council through grants no. 17300918 and no. 17307520. ## References * Shor (1995) Peter W. Shor, “Scheme for reducing decoherence in quantum computer memory,” Phys. Rev. A 52, R2493–R2496 (1995). * Nielsen and Chuang (2010) M. A. Nielsen and I. L. Chuang, _Quantum Computation and Quantum Information_ (Cambridge university press, 2010). * Cochrane _et al._ (1999) P. T. Cochrane, G. J. Milburn, and W. J. Munro, “Macroscopically distinct quantum-superposition states as a bosonic code for amplitude damping,” Phys. Rev. A 59, 2631–2634 (1999). * Ralph _et al._ (2003) T. C. Ralph, A. Gilchrist, G. J. Milburn, W. J. Munro, and S. Glancy, “Quantum computation with optical coherent states,” Phys. Rev. A 68, 042319 (2003). * Leghtas _et al._ (2013) Zaki Leghtas, Gerhard Kirchmair, Brian Vlastakis, Robert J. Schoelkopf, Michel H. Devoret, and Mazyar Mirrahimi, “Hardware-efficient autonomous quantum memory protection,” Phys. Rev. Lett. 111, 120501 (2013). * Mirrahimi _et al._ (2014) Mazyar Mirrahimi, Zaki Leghtas, Victor V Albert, Steven Touzard, Robert J Schoelkopf, Liang Jiang, and Michel H Devoret, “Dynamically protected cat-qubits: a new paradigm for universal quantum computation,” New J. Phys. 16, 045014 (2014). * Bergmann and van Loock (2016) Marcel Bergmann and Peter van Loock, “Quantum error correction against photon loss using multicomponent cat states,” Phys. Rev. A 94, 042332 (2016). * Gottesman _et al._ (2001) Daniel Gottesman, Alexei Kitaev, and John Preskill, “Encoding a qubit in an oscillator,” Phys. Rev. A 64, 012310 (2001). * Michael _et al._ (2016) Marios H. Michael, Matti Silveri, R. T. Brierley, Victor V. Albert, Juha Salmilehto, Liang Jiang, and S. M. Girvin, “New class of quantum error-correcting codes for a bosonic mode,” Phys. Rev. X 6, 031006 (2016). * Albert _et al._ (2018) Victor V. Albert, Kyungjoo Noh, Kasper Duivenvoorden, Dylan J. Young, R. T. Brierley, Philip Reinhold, Christophe Vuillot, Linshu Li, Chao Shen, S. M. Girvin, Barbara M. Terhal, and Liang Jiang, “Performance and structure of single-mode bosonic codes,” Phys. Rev. A 97, 032346 (2018). * Grimsmo _et al._ (2020) Arne L. Grimsmo, Joshua Combes, and Ben Q. Baragiola, “Quantum computing with rotation-symmetric bosonic codes,” Phys. Rev. X 10, 011058 (2020). * Ofek _et al._ (2016) Nissim Ofek, Andrei Petrenko, Reinier Heeres, Philip Reinhold, Zaki Leghtas, Brian Vlastakis, Yehan Liu, Luigi Frunzio, SM Girvin, Liang Jiang, _et al._ , “Extending the lifetime of a quantum bit with error correction in superconducting circuits,” Nature 536, 441–445 (2016). * Flühmann _et al._ (2019) Christa Flühmann, Thanh Long Nguyen, Matteo Marinelli, Vlad Negnevitsky, Karan Mehta, and JP Home, “Encoding a qubit in a trapped-ion mechanical oscillator,” Nature 566, 513–517 (2019). * Campagne-Ibarcq _et al._ (2020) Philippe Campagne-Ibarcq, Alec Eickbusch, Steven Touzard, Evan Zalys-Geller, Nicholas E Frattini, Volodymyr V Sivak, Philip Reinhold, Shruti Puri, Shyam Shankar, Robert J Schoelkopf, _et al._ , “Quantum error correction of a qubit encoded in grid states of an oscillator,” Nature 584, 368–372 (2020). * Terhal and Weigand (2016) B. M. Terhal and D. Weigand, “Encoding a qubit into a cavity mode in circuit qed using phase estimation,” Phys. Rev. A 93, 012315 (2016). * Puri _et al._ (2017) Shruti Puri, Samuel Boutin, and Alexandre Blais, “Engineering the quantum states of light in a kerr-nonlinear resonator by two-photon driving,” NPJ Quantum Inf. 3, 18 (2017). * Weigand and Terhal (2018) Daniel J. Weigand and Barbara M. Terhal, “Generating grid states from schrödinger-cat states without postselection,” Phys. Rev. A 97, 022341 (2018). * Shi _et al._ (2019) Yunong Shi, Christopher Chamberland, and Andrew Cross, “Fault-tolerant preparation of approximate gkp states,” New J. Phys. 21, 093007 (2019). * Lvovsky and Raymer (2009) A. I. Lvovsky and M. G. Raymer, “Continuous-variable optical quantum-state tomography,” Rev. Mod. Phys. 81, 299–332 (2009). * D’Ariano _et al._ (2003) G Mauro D’Ariano, Matteo GA Paris, and Massimiliano F Sacchi, “Quantum tomography,” Adv. Imaging Electron Phys. 128, 206–309 (2003). * Tiunov _et al._ (2020) Egor S Tiunov, VV Tiunova, Alexander E Ulanov, AI Lvovsky, and Aleksey K Fedorov, “Experimental quantum homodyne tomography via machine learning,” Optica 7, 448–454 (2020). * Ahmed _et al._ (2021a) Shahnawaz Ahmed, Carlos Sánchez Muñoz, Franco Nori, and Anton Frisk Kockum, “Quantum state tomography with conditional generative adversarial networks,” Phys. Rev. Lett. 127, 140502 (2021a). * Ahmed _et al._ (2021b) Shahnawaz Ahmed, Carlos Sánchez Muñoz, Franco Nori, and Anton Frisk Kockum, “Classification and reconstruction of optical quantum states with deep neural networks,” Phys. Rev. Res. 3, 033278 (2021b). * Zhu _et al._ (2022) Yan Zhu, Ya-Dong Wu, Ge Bai, Dong-Sheng Wang, Yuexuan Wang, and Giulio Chiribella, “Flexible learning of quantum states with generative query neural networks,” Nat. Commun. 13, 6222 (2022). * Wu _et al._ (2022) Ya-Dong Wu, Yan Zhu, Ge Bai, Yuexuan Wang, and Giulio Chiribella, “A data-driven approach to quantum cross-platform verification,” arXiv:2211.01668 (2022). * Eisert _et al._ (2020) Jens Eisert, Dominik Hangleiter, Nathan Walk, Ingo Roth, Damian Markham, Rhea Parekh, Ulysse Chabaud, and Elham Kashefi, “Quantum certification and benchmarking,” Nat. Rev. Phys. , 1–9 (2020). * Aolita _et al._ (2015) Leandro Aolita, Christian Gogolin, Martin Kliesch, and Jens Eisert, “Reliable quantum certification of photonic state preparations,” Nat. Commun. 6, 1–8 (2015). * Aaronson and Arkhipov (2013) Scott Aaronson and Alex Arkhipov, “The computational complexity of linear optics,” Theory Comput. 9, 143–252 (2013). * Knill _et al._ (2001) Emanuel Knill, Raymond Laflamme, and Gerald J Milburn, “A scheme for efficient quantum computation with linear optics,” nature 409, 46–52 (2001). * Chabaud _et al._ (2021) Ulysse Chabaud, Frédéric Grosshans, Elham Kashefi, and Damian Markham, “Efficient verification of boson sampling,” Quantum 5, 578 (2021). * da Silva _et al._ (2011) Marcus P. da Silva, Olivier Landon-Cardinal, and David Poulin, “Practical characterization of quantum devices without tomography,” Phys. Rev. Lett. 107, 210404 (2011). * Chabaud _et al._ (2020a) Ulysse Chabaud, Tom Douce, Frédéric Grosshans, Elham Kashefi, and Damian Markham, “Building Trust for Continuous Variable Quantum States,” in _15th Conference on the Theory of Quantum Computation, Communication and Cryptography (TQC 2020)_, Leibniz International Proceedings in Informatics (LIPIcs), Vol. 158, edited by Steven T. Flammia (Schloss Dagstuhl–Leibniz-Zentrum für Informatik, Dagstuhl, Germany, 2020) pp. 3:1–3:15. * Pallister _et al._ (2018) Sam Pallister, Noah Linden, and Ashley Montanaro, “Optimal verification of entangled states with local measurements,” Phys. Rev. Lett. 120, 170502 (2018). * Liu _et al._ (2021) Ye-Chao Liu, Jiangwei Shang, and Xiangdong Zhang, “Efficient verification of entangled continuous-variable quantum states with local measurements,” Phys. Rev. Research 3, L042004 (2021). * Gluza _et al._ (2018) M. Gluza, M. Kliesch, J. Eisert, and L. Aolita, “Fidelity witnesses for fermionic quantum simulations,” Phys. Rev. Lett. 120, 190501 (2018). * Menicucci (2014) Nicolas C. Menicucci, “Fault-tolerant measurement-based quantum computing with continuous-variable cluster states,” Phys. Rev. Lett. 112, 120504 (2014). * Douce _et al._ (2017) T. Douce, D. Markham, E. Kashefi, E. Diamanti, T. Coudreau, P. Milman, P. van Loock, and G. Ferrini, “Continuous-variable instantaneous quantum computing is hard to sample,” Phys. Rev. Lett. 118, 070503 (2017). * Bužek and Knight (1995) Vladimír Bužek and Peter L Knight, “I: Quantum interference, superposition states of light, and nonclassical effects,” in _Progress in optics_ , Vol. 34 (Elsevier, 1995) pp. 1–158. * Scully _et al._ (1997) Marlan O Scully, M Suhail Zubairy, _et al._ , _Quantum Optics_ (Cambridge University Press, 1997). * Schlegel _et al._ (2022) David S. Schlegel, Fabrizio Minganti, and Vincenzo Savona, “Quantum error correction using squeezed schrödinger cat states,” Phys. Rev. A 106, 022431 (2022). * Menicucci _et al._ (2006) Nicolas C Menicucci, Peter Van Loock, Mile Gu, Christian Weedbrook, Timothy C Ralph, and Michael A Nielsen, “Universal quantum computation with continuous-variable cluster states,” Phys. Rev. Lett. 97, 110501 (2006). * Gu _et al._ (2009) Mile Gu, Christian Weedbrook, Nicolas C. Menicucci, Timothy C. Ralph, and Peter van Loock, “Quantum computing with continuous-variable clusters,” Phys. Rev. A 79, 062318 (2009). * Larsen _et al._ (2019) Mikkel V Larsen, Xueshi Guo, Casper R Breum, Jonas S Neergaard-Nielsen, and Ulrik L Andersen, “Deterministic generation of a two-dimensional cluster state,” Science 366, 369–372 (2019). * Hastrup _et al._ (2021) Jacob Hastrup, Kimin Park, Jonatan Bohr Brask, Radim Filip, and Ulrik Lund Andersen, “Measurement-free preparation of grid states,” NPJ Quantum Inf. 7, 17 (2021). * Flammia and Liu (2011) Steven T. Flammia and Yi-Kai Liu, “Direct fidelity estimation from few pauli measurements,” Phys. Rev. Lett. 106, 230501 (2011). * Liu _et al._ (2019) Nana Liu, Tommaso F. Demarie, Si-Hui Tan, Leandro Aolita, and Joseph F. Fitzsimons, “Client-friendly continuous-variable blind and verifiable quantum computing,” Phys. Rev. A 100, 062309 (2019). * Farias and Aolita (2021) Renato MS Farias and Leandro Aolita, “Certification of continuous-variable gates using average channel-fidelity witnesses,” Quantum Sci. Technol. 6, 035014 (2021). * Chabaud _et al._ (2020b) Ulysse Chabaud, Damian Markham, and Frédéric Grosshans, “Stellar representation of non-gaussian quantum states,” Phys. Rev. Lett. 124, 063605 (2020b). * Huang _et al._ (2020) Hsin-Yuan Huang, Richard Kueng, and John Preskill, “Predicting many properties of a quantum system from very few measurements,” Nature Physics 16, 1050–1057 (2020). * Wu _et al._ (2021) Ya-Dong Wu, Ge Bai, Giulio Chiribella, and Nana Liu, “Efficient verification of continuous-variable quantum states and devices without assuming identical and independent operations,” Phys. Rev. Lett. 126, 240503 (2021). * Hayashi and Morimae (2015) Masahito Hayashi and Tomoyuki Morimae, “Verifiable measurement-only blind quantum computing with stabilizer testing,” Phys. Rev. Lett. 115, 220502 (2015). * Gheorghiu _et al._ (2019) Alexandru Gheorghiu, Theodoros Kapourniotis, and Elham Kashefi, “Verification of quantum computation: An overview of existing approaches,” Theory Comput. Syst. 63, 715–808 (2019). ## VIII Appendix ### VIII.1 Calculation of two-component cat code witness in Eq. (6) By using $\displaystyle\hat{a}^{\dagger\,2}\hat{a}^{2}=$ $\displaystyle\hat{n}^{2}-\hat{n}$ (26) $\displaystyle=$ $\displaystyle\frac{1}{4}\left(\hat{x}^{4}+\hat{p}^{4}+\hat{x}^{2}\hat{p}^{2}+\hat{p}^{2}\hat{x}^{2}\right)-\hat{x}^{2}-\hat{p}^{2}+\frac{3}{4}$ (27) $\displaystyle\alpha^{*2}\hat{a}^{2}+\alpha^{2}\hat{a}^{\dagger\,2}=$ $\displaystyle\operatorname{Re}(\alpha^{2})(\hat{x}^{2}-\hat{p}^{2})+\operatorname{Im}(\alpha^{2})(\hat{x}\hat{p}+\hat{p}\hat{x}),$ (28) we obtain $\displaystyle W=$ $\displaystyle\mathds{1}-\frac{\hat{H}_{tCat}}{2}$ (29) $\displaystyle=$ $\displaystyle\mathds{1}-\frac{1}{8}\left(\hat{x}^{4}+\hat{p}^{4}+\hat{x}^{2}\hat{p}^{2}+\hat{p}^{2}\hat{x}^{2}\right)+\frac{1}{2}(\hat{x}^{2}+\hat{p}^{2})-\frac{3}{8}+\frac{1}{2}\operatorname{Re}(\alpha^{2})(\hat{x}^{2}-\hat{p}^{2})$ (30) $\displaystyle+\frac{1}{2}\operatorname{Im}(\alpha^{2})(\hat{x}\hat{p}+\hat{p}\hat{x})-\frac{|\alpha|^{4}}{2}$ (31) Note that $\hat{x}\hat{p}+\hat{p}\hat{x}=(\hat{x}+\hat{p})^{2}-\hat{x}^{2}-\hat{p}^{2}$ (32) and $\hat{x}^{2}\hat{p}^{2}+\hat{p}^{2}\hat{x}^{2}=\frac{1}{6}\left[\left(\hat{x}+\hat{p}\right)^{4}+\left(\hat{x}-\hat{p}\right)^{4}\right]-\frac{1}{3}(\hat{x}^{4}+\hat{p}^{4})-1.$ (33) The second equality is because $\displaystyle(\hat{x}+\hat{p})^{4}+(\hat{x}-\hat{p})^{4}=$ $\displaystyle 2(\hat{x}^{4}+\hat{p}^{4}+\hat{x}^{2}\hat{p}^{2}+\hat{p}^{2}\hat{x}^{2})+2(\hat{x}\hat{p}\hat{x}\hat{p}+\hat{x}\hat{p}^{2}\hat{x}+\hat{p}\hat{x}\hat{p}\hat{x}+\hat{p}\hat{x}^{2}\hat{p})$ $\displaystyle=$ $\displaystyle 2(\hat{x}^{4}+\hat{p}^{4}+\hat{x}^{2}\hat{p}^{2}+\hat{p}^{2}\hat{x}^{2})+4(\hat{x}^{2}\hat{p}^{2}+\hat{p}^{2}\hat{x}^{2})+6.$ So the fidelity witness can be rewritten as $\displaystyle W=$ $\displaystyle\mathds{1}-\frac{1}{8}(\hat{x}^{4}+\hat{p}^{4})-\frac{1}{48}\left[\left(\hat{x}+\hat{p}\right)^{4}+\left(\hat{x}-\hat{p}\right)^{4}\right]+\frac{1}{24}(\hat{x}^{4}+\hat{p}^{4})$ (34) $\displaystyle+\frac{1}{8}+\frac{1}{2}(\hat{x}^{2}+\hat{p}^{2})-\frac{3}{8}+\frac{1}{2}\operatorname{Re}(\alpha^{2})(\hat{x}^{2}-\hat{p}^{2})+\frac{1}{2}\operatorname{Im}(\alpha^{2})\left(\hat{x}+\hat{p}\right)^{2}$ (35) $\displaystyle-\frac{1}{2}\operatorname{Im}(\alpha^{2})\hat{x}^{2}-\frac{1}{2}\operatorname{Im}(\alpha^{2})\hat{p}^{2}-\frac{|\alpha|^{4}}{2}$ (36) $\displaystyle=$ $\displaystyle\mathds{1}-\frac{1}{16}(\hat{x}^{4}+\hat{p}^{4})-\frac{1}{48}\left[\left(\hat{x}+\hat{p}\right)^{4}+\left(\hat{x}-\hat{p}\right)^{4}\right]$ (37) $\displaystyle+\frac{1}{2}\operatorname{Im}(\alpha^{2})\left(\hat{x}+\hat{p}\right)^{2}+\frac{1}{2}\left[1+\operatorname{Re}(\alpha^{2})-\operatorname{Im}(\alpha^{2})\right]\hat{x}^{2}$ (38) $\displaystyle+\frac{1}{2}\left[1-\operatorname{Re}(\alpha^{2})-\operatorname{Im}(\alpha^{2})\right]\hat{p}^{2}-\frac{1+2|\alpha|^{4}}{4}.$ (39) In the case that $\alpha^{2}\in\mathbb{R}$, we have $W=\mathds{1}-\frac{1}{16}(\hat{x}^{4}+\hat{p}^{4})-\frac{1}{48}\left[\left(\hat{x}+\hat{p}\right)^{4}+\left(\hat{x}-\hat{p}\right)^{4}\right]+\frac{1}{2}(1+\alpha^{2})\hat{x}^{2}+\frac{1}{2}(1-\alpha^{2})\hat{p}^{2}-\frac{1+2\alpha^{4}}{4}.$ (40) ### VIII.2 Calculation of four-component cat code witness in Eq. (10) Using the fact that for any $k\in\mathbb{N}^{+}$, $\hat{a}^{\dagger k}\hat{a}^{k}=\hat{n}(\hat{n}-1)\cdots(\hat{n}-k+1)$ (41) and $\displaystyle\hat{n}=\hat{a}\hat{a}^{\dagger}-1$ $\displaystyle\hat{n}^{2}=\hat{a}^{2}\hat{a}^{\dagger\,2}-3\hat{a}\hat{a}^{\dagger}+1$ $\displaystyle\hat{n}^{3}=\hat{a}^{3}\hat{a}^{\dagger\,3}-6\hat{a}^{2}\hat{a}^{\dagger\,2}+7\hat{a}\hat{a}^{\dagger}-1$ $\displaystyle\hat{n}^{4}=\hat{a}^{4}\hat{a}^{\dagger\,4}-10\hat{a}^{3}\hat{a}^{\dagger\,3}+25\hat{a}^{2}\hat{a}^{\dagger\,2}-15\hat{a}\hat{a}^{\dagger}+1$ we get $\displaystyle\hat{a}^{\dagger 4}\hat{a}^{4}=\hat{a}^{4}\hat{a}^{\dagger\,4}-16\hat{a}^{3}\hat{a}^{\dagger\,3}+72\hat{a}^{2}\hat{a}^{\dagger\,2}-96\hat{a}\hat{a}^{\dagger}+24,$ $\displaystyle\hat{a}^{\dagger 3}\hat{a}^{3}=\hat{a}^{3}\hat{a}^{\dagger\,3}-9\hat{a}^{2}\hat{a}^{\dagger\,2}+18\hat{a}\hat{a}^{\dagger}-6,$ $\displaystyle\hat{a}^{\dagger 2}\hat{a}^{2}=\hat{a}^{2}\hat{a}^{\dagger\,2}-4\hat{a}\hat{a}^{\dagger}+2,$ $\displaystyle\hat{a}^{\dagger}\hat{a}=\hat{a}\hat{a}^{\dagger}-1.$ Combing all the above equations, we obtain Eq. (10). ### VIII.3 Calculation of state witness in fault-tolerant quantum computing Using the fact that $CZ\hat{a}_{i}CZ^{\dagger}=\hat{a}_{i}-\frac{\text{i}}{\sqrt{2}}\hat{x}_{j}$, we have $\displaystyle\prod_{j\in\mathcal{N}(i)}CZ_{ij}\left(\cosh r\hat{a}_{i}^{\dagger}-\sinh r\hat{a}_{i}-\frac{\sqrt{\pi}k}{\sigma}\right)\prod_{j\in\mathcal{N}(i)}CZ_{ij}^{\dagger}$ $\displaystyle=$ $\displaystyle\cosh r\hat{a}_{i}^{\dagger}+\frac{\text{i}}{\sqrt{2}}\cosh r\sum_{j\in\mathcal{N}(i)}\hat{x}_{j}-\sinh r\hat{a}_{i}+\frac{\text{i}}{\sqrt{2}}\sinh r\sum_{j\in\mathcal{N}(i)}\hat{x}_{j}-\frac{\sqrt{\pi}k}{\sigma}$ $\displaystyle=$ $\displaystyle\frac{\text{e}^{-r}}{\sqrt{2}}\hat{x}_{i}-\frac{\text{e}^{r}}{\sqrt{2}}\text{i}\hat{p}_{i}+\frac{\text{i}}{\sqrt{2}}\text{e}^{r}\sum_{j\in\mathcal{N}(i)}\hat{x}_{j}-\frac{\sqrt{\pi}k}{\sigma}$ and $\displaystyle\prod_{j\in\mathcal{N}(i)}CZ_{ij}\left(\cosh r\hat{a}_{i}^{\dagger}+\sinh r\hat{a}_{i}+2\text{i}\frac{\sqrt{\pi}k}{\sigma}\right)\prod_{j\in\mathcal{N}(i)}CZ_{ij}^{\dagger}$ $\displaystyle\rightarrow$ $\displaystyle\cosh r\hat{a}_{i}^{\dagger}+\frac{\text{i}}{\sqrt{2}}\cosh r\sum_{j\in\mathcal{N}(i)}\hat{x}_{j}+\sinh r\hat{a}_{i}-\frac{\text{i}}{\sqrt{2}}\sinh r\sum_{j\in\mathcal{N}(i)}\hat{x}_{j}+2\text{i}\frac{\sqrt{\pi}k}{\sigma}$ $\displaystyle=$ $\displaystyle\frac{\text{e}^{r}}{\sqrt{2}}\hat{x}_{i}-\frac{\text{e}^{-r}}{\sqrt{2}}\text{i}\hat{p}_{i}+\frac{\text{i}}{\sqrt{2}}\text{e}^{-r}\sum_{j\in\mathcal{N}(i)}\hat{x}_{j}+2\text{i}\frac{\sqrt{\pi}k}{\sigma}.$ Plugging them into Eq. (18), we obtain the witness in Eq. (19). Since $\displaystyle Z\hat{x}Z^{\dagger}=CZ\hat{x}CZ^{\dagger}=T\hat{x}T^{\dagger}=\hat{x}$ $\displaystyle Z\hat{p}Z^{\dagger}=\hat{p}-\sqrt{\pi}$ $\displaystyle CZ\hat{p}_{1}CZ^{\dagger}=\hat{p}_{1}-\hat{x}_{2}$ $\displaystyle T\hat{p}T^{\dagger}=\hat{p}+\text{i}\pi/4(2/\sqrt{\pi^{3}}[\hat{x}^{3},\hat{p}]+1/\pi[\hat{x}^{2},\hat{p}]-2\text{i}/\sqrt{\pi})=\hat{p}-3\hat{x}^{2}/(2\sqrt{\pi})-\hat{x}/2+\sqrt{\pi}/2,$ for each mode in the QIP circuit, the annihilation operator is transformed to $\hat{a}\rightarrow\hat{a}-\text{i}\sqrt{\frac{\pi}{2}}n_{Z}-\frac{\text{i}}{\sqrt{2}}\sum_{j=1}^{n_{CZ}}\hat{x}_{j}-3\text{i}n_{T}\hat{x}^{2}/(2\sqrt{2}\pi)-\text{i}n_{T}\hat{x}/2\sqrt{2}+\sqrt{\pi}n_{T}\text{i}/2\sqrt{2}.$ (42) Hence, $\displaystyle\cosh r\hat{a}^{\dagger}-\sinh r\hat{a}$ $\displaystyle\rightarrow$ $\displaystyle\cosh r(\hat{a}^{\dagger}+\text{i}\sqrt{\frac{\pi}{2}}n_{Z}+\frac{\text{i}}{\sqrt{2}}\sum_{j=1}^{n_{CZ}}\hat{x}_{j}+3\text{i}n_{T}\hat{x}^{2}/(2\sqrt{2}\pi)+\text{i}n_{T}\hat{x}/2\sqrt{2}-\sqrt{\pi}n_{T}\text{i}/2\sqrt{2})$ $\displaystyle-\sinh r(\hat{a}-\text{i}\sqrt{\frac{\pi}{2}}n_{Z}-\frac{\text{i}}{\sqrt{2}}\sum_{j=1}^{n_{CZ}}\hat{x}_{j}-3\text{i}n_{T}\hat{x}^{2}/(2\sqrt{2}\pi)-\text{i}n_{T}\hat{x}/2\sqrt{2}+\sqrt{\pi}n_{T}\text{i}/2\sqrt{2})$ $\displaystyle=$ $\displaystyle(\frac{\text{e}^{-r}}{\sqrt{2}}+\frac{n_{T}\text{i}\text{e}^{r}}{2\sqrt{2}})\hat{x}_{i}-\frac{\text{e}^{r}}{\sqrt{2}}\text{i}\hat{p}_{i}+\frac{\text{i}}{\sqrt{2}}\text{e}^{r}\sum_{j=1}^{n_{CZ}}\hat{x}_{j}+\frac{3n_{T}\text{i}\text{e}^{r}}{2\sqrt{2}\pi}\hat{x}_{i}^{2}+\text{i}\sqrt{\frac{\pi}{2}}(n_{Z}-n_{T}/2)\text{e}^{r}.$ Plugging it into Eq. (18), we obtain the witness of quantum output of IQP circuit.
# Continual Learning with Optimal Transport based Mixture Model Quyen Tran VinAI Research <EMAIL_ADDRESS>Hoang Phan VinAI Research <EMAIL_ADDRESS>Khoat Than Hanoi University of Science and Technology <EMAIL_ADDRESS>Dinh Phung Monash University <EMAIL_ADDRESS>Trung Le Monash University <EMAIL_ADDRESS> ###### Abstract Online Class Incremental learning (CIL) is a challenging setting in Continual Learning (CL), wherein data of new tasks arrive in incoming streams and online learning models need to handle incoming data streams without revisiting previous ones. Existing works used a single centroid adapted with incoming data streams to characterize a class. This approach possibly exposes limitations when the incoming data stream of a class is naturally multimodal. To address this issue, in this work, we first propose an online mixture model learning approach based on nice properties of the mature optimal transport theory (OT-MM). Specifically, the centroids and covariance matrices of the mixture model are adapted incrementally according to incoming data streams. The advantages are two-fold: (i) we can characterize more accurately complex data streams and (ii) by using centroids for each class produced by OT-MM, we can estimate the similarity of an unseen example to each class more reasonably when doing inference. Moreover, to combat the catastrophic forgetting in the CIL scenario, we further propose Dynamic Preservation. Particularly, after performing the dynamic preservation technique across data streams, the latent representations of the classes in the old and new tasks become more condensed themselves and more separate from each other. Together with a contraction feature extractor, this technique facilitates the model in mitigating the catastrophic forgetting. The experimental results on real-world datasets show that our proposed method can significantly outperform the current state-of- the-art baselines. ## 1 Introduction Figure 1: The intuitions and motivations of our CLOT. Dynamic preservation inspires the classes in the old and new tasks to be more condensed themselves via minimizing $L_{2}(\boldsymbol{\theta})$ and more separate to each other via minimizing $L_{1}(\boldsymbol{\theta},\boldsymbol{w})$. This encourages the feature extractor $f_{\boldsymbol{\theta}}$ to have a contraction behaviour, which helps to reduce the discrepancy in the latent representations of the replay memory and the entire old-task data. Subsequently, we employ our OT-MM to characterize each class with a mixture model and multiple centroids. Artificial neural networks have created a revolution in solving many real- world problems, especially in the field of computer vision [48, 13, 30, 10]. Along with the development of advanced artificial neural networks, many intelligent systems have been widely applied to reality and in return posed many more challenging problems. Additionally, many real-world problems, typically autonomous vehicles [21, 46], sensory robot data [23, 26], and video streaming [39, 43] require the intelligent systems to continuously interact, learn, and adapt to the dynamic changes in learning environments. To further address this desideratum, Continual Learning has been introduced and investigated to tackle the dynamic changes in learning environments. In continual learning (CL), our learning system sequentially receives incoming data streams and need to adapt continuously to the dynamic changes in the learning environments. Currently, most of works in CL focuses on three main scenarios: (i) TIL where the task boundary is known during learning and task id is available during inference [33, 16, 38, 12], (ii) DIL where there is only a change in the distribution of tasks in the data stream [47, 2, 24], and (iii) CIL where the system receives infinite incoming data streams only once without revisiting previous ones [22, 44, 14]. Among these three, CIL is possibly the most realistic scenario because it does not require the task id and and allows not only data distribution but also labels being dynamically changed. Therefore, in this work, we propose a novel method to tackle the CIL setting. One of the crucial problems in continual learning is the catastrophic forgetting (CF), indicating the phenomenon where neural networks tend to forget how to perform old tasks when learning new tasks. This even becomes more serious in the CIL scenario. The reason is we use only a single classifier to predict classes seen so far and the classifier tends to bias the classes in the recent and new tasks, known as the bias phenomenon. To address the catastrophic forgetting, most of the previous works [7] only focused on avoiding forgetting for the backbone and ignored the important classification head, hence compromising the predictive performance. Additionally, some recent works [14, 44, 52] have investigated bias correction, but this is a way to solve the situation without looking at the big picture in combination with features extractor. Some others [20, 45] used tricks on the task id of the test set when predicting, which seems to violate the rule of CIL regarding no task id available. Another recent work [8] proposed CoPE to learn a single centroid for each class and used the class centroids for prediction. Although achieving impressive performance, using a single centroid for a class might not be sufficient to characterize the complexity of the incoming data stream of this class, since practical data often exhibit multimodality. Moreover, there always exists a gap between train and test latent representations, hence single class centroid tailored to the train set might highly mismatch the test set. In this work, to tackle the complexity of the data stream of a class in the OCL scenario, we propose the novel optimal transport based mixture model (OT- MM) for modelling incoming streaming data of each class. This is realized by leveraging the rich theoretical body of optimal transport or Wasserstein (WS) distance [40, 34] and mixture models [31] in the context of OCL. Interestingly, by using the entropic dual form of optimal transport [9] and Gumbel softmax distribution [15], we reach an appealing formulation in the expectation form for the WS distance of interest, making it ready for the OCL setting where the update is performed on the basis of incoming data batches. As a result, we can learn the multiple centroids and covariance matrices incrementally for each class. As shown in Figure 1, by using multiple centroids characterizing each class to do inference, we can mitigate the negative effect to the predictive performance caused by the shift between train and test latent representations for further improving the predictive performance. Another complementary component of our proposed approach is the dynamic preservation, performed dynamically across incoming data streams. This component is designed to make latent representations of data in the classes of both old and new tasks more condensed themselves and more separate to increase the class discrimination ability. More interestingly, this makes the feature extractor a contraction behavior, which helps reduce the discrepancy in the latent representations of the replay memory and the entire old data (cf. Figure 1) for mitigating the catastrophic forgetting. Finally, we name our proposed approach Class Increamental Learning with Optimal Transport based Mixture Model (CLOT) and conduct comprehensive experiments on the real-world datasets to compare our CLOT with state-of-the- art baselines. The experimental results show that CLOT significantly outperforms state-of-the-art baselines. To sum up, our work has the following contributions: * • This is the first work that leverages with the rich theoretical body of optimal transport to propose a novel OT-MM for tacking the multimodality of incoming data streams in the OCL scenario. * • Our proposed Dynamic Preservation component that strengthens the class discrimination ability and reduces the gap between train and test latent representations for relieving the catastrophic forgetting. This component is complimentary to OT-MM for devising our CLOT. * • We conduct extensive experiments on the real-world datasets to show that our proposed method performs significantly better than other existing methods. ## 2 Related work Figure 2: $t$-SNE visualization on MNIST: Motivation of OT-MM. Left: the test latent representation of CoPE [8] with one centroid (i.e., visualized by digits) per class. Right: the test latent representation of our CLOT with four centroids per class (i.e., visualized by digits). We observe that there exists a shift between the test and train representations. Therefore, centroids learned on the training set might mismatch the testing set. However, using more centroids per class as in our CLOT can relieve this mismatch. Previous works attempt to tackle the problem of catastrophic forgetting by efficiently designing the training process in a number of ways. Memory-based approaches utilize episodic memory to store past data [32, 4, 5, 33] or employ deep generative models [37, 19] to produce pseudo samples from the previous history. Architecture-based methods dynamically allocate a separate subnetwork for each task [27, 42] to maintain the knowledge of old tasks. Regularization- based approaches encourage important parameters of old tasks to lie in their close vicinity [17, 51, 49] by penalizing their changes. Among these three main lines of work, our method falls into the memory-based approach category, which utilizes the advantages of buffer memory to preserve the discriminative characteristics of data. Current CL algorithms are often assessed in an incremental classification scenario, where tasks or classes arrive at different timestamps. We here consider the online class-incremental (OCL) setting, which is more realistic but also more challenging. In particular, in comparison to task-incremental learning, the data stream is no longer divided into tasks with clear boundaries. We also have to deal with catastrophic forgetting without any information about the task id that the data come from at both training and inference time. The recent work in OCL usually focuses on: either (i) how to choose meaningful, diverse samples to store [6, 35], also how to select old samples to replay [29], or (ii) how to learn representation effectively [11, 3] , mostly inspired by contrastive learning, (iii) how to solve bias problem for single categorical head [14, 44, 52]. In this work, we present a memory- based learning strategy, also inspired by contrastive learning. However, not only that, we bring a breeze when introducing the theory of optimal transport into continuous learning for the first time. Specifically, an optimal transport-based mixing model will be applied in the online scenario to learn the multiple centroids that characterize each class. Then these centroids will be used to make more accurate predictions and improve episodic memory. ## 3 Preliminaries In what follows, we present the background of Wasserstein (WS) distance and optimal transport theory that is necessary for the development of our proposed method. ### 3.1 Optimal transport and Wasserstein distance Consider two distributions $\mathbb{P}$ and $\mathbb{Q}$ which operate on the domain $\Omega\subseteq\mathbb{R}^{d}$, let $d\left(\boldsymbol{x},\boldsymbol{y}\right)$ be a non-negative and continuous cost function or metric. Wasserstein distance [34, 40] between $\mathbb{P}$ and $\mathbb{Q}$ w.r.t the metric $d$ is defined as $\mathcal{W}_{d}\left(\mathbb{Q},\mathbb{P}\right):=\min_{\gamma\in\Gamma\left(\mathbb{Q},\mathbb{P}\right)}\mathbb{E}_{\left(\boldsymbol{x},\boldsymbol{y}\right)\sim\gamma}\left[d\left(\boldsymbol{x},\boldsymbol{y}\right)\right],$ (1) where $\gamma$ is a coupling that admits $\mathbb{Q},\mathbb{P}$ as its marginals. ### 3.2 Entropic dual-form for OT and WS distance To enable the application of optimal transport in machine learning and deep learning, [9] developed an entropic regularized dual form. First, they proposed to add an entropic regularization term to primal form (1) $\displaystyle\mathcal{W}_{d}^{\varepsilon}\left(\mathbb{Q},\mathbb{P}\right)$ $\displaystyle:=$ $\displaystyle\min_{\gamma\in\Gamma\left(\mathbb{Q},\mathbb{P}\right)}$ $\displaystyle\left\\{\mathbb{E}_{\left(\boldsymbol{x},\boldsymbol{y}\right)\sim\gamma}\left[d\left(\boldsymbol{x},\boldsymbol{y}\right)\right]+\varepsilon D_{KL}\left(\gamma\|\mathbb{Q}\otimes\mathbb{P}\right)\right\\},$ (2) where $\varepsilon$ is the regularization rate, $D_{KL}\left(\cdot\|\cdot\right)$ is the Kullback-Leibler (KL) divergence, and $\mathbb{Q}\otimes\mathbb{P}$ represents the specific coupling in which $\mathbb{Q}$ and $\mathbb{P}$ are independent. Using Fenchel-Rockafellar theorem, they obtained the following _entropic regularized dual form_ of (2) $\displaystyle\mathcal{W}_{d}^{\varepsilon}\left(\mathbb{Q},\mathbb{P}\right)$ $\displaystyle=\max_{\phi}\left\\{\int\tilde{\phi}\left(\boldsymbol{x}\right)\mathrm{d}\mathbb{Q}\left(\boldsymbol{x}\right)+\int\phi\left(\boldsymbol{y}\right)\mathrm{d}\mathbb{P}\left(\boldsymbol{y}\right)\right\\}$ $\displaystyle=\max_{\phi}\left\\{\mathbb{E}_{\mathbb{Q}}\left[\tilde{\phi}\left(\boldsymbol{x}\right)\right]+\mathbb{E}_{\mathbb{P}}\left[\phi\left(\boldsymbol{y}\right)\right]\right\\},$ (3) where $\tilde{\phi}\left(\boldsymbol{x}\right):=-\varepsilon\log\left(\mathbb{E}_{\mathbb{P}}\left[\exp\left\\{\frac{-d\left(\boldsymbol{x},\boldsymbol{y}\right)+\phi\left(\boldsymbol{y}\right)}{\varepsilon}\right\\}\right]\right)$. ## 4 Our Proposed Method In this section, we present the details of our proposed method. We start with the general framework and our motivations, followed by the technical details of online mixture model based on optimal transport theory and the simple yet but effective dynamic preservation technique that assists the model in maintaining the performance on the old tasks. ### 4.1 General Framework and Motivations In the context of the online class incremental learning, at a time step, our system receives a batch of data examples $X=[X^{c}]_{c\in\mathcal{C}_{new}}$ from a new task, where $\mathcal{C}_{new}$ represents the classes in the new task and $X^{c}$ is the batch data for the class $c$. To mitigate catastrophic forgetting, we also maintain a replay memory $\mathcal{M}$ of the old tasks. We sample a batch of data examples from the replay memory $\bar{X}=[\bar{X}^{c}]_{c\in\mathcal{C}_{old}}$, where $\mathcal{C}_{old}$ represents the classes in the old tasks and $\bar{X}^{c}$ is the batch data for the class $c$ in the replay memory. We feed $X$ and $\bar{X}$ to the feature extractor $f_{\boldsymbol{\theta}}$ to obtain the batches of latent representations $Z=f_{\boldsymbol{\theta}}(X)$ and $\bar{Z}=f_{\boldsymbol{\theta}}(\bar{X})$ for new and old tasks respectively. For the batches of $X$ and $\bar{X}$, we first perform dynamic preservation to make the classes in the old and new tasks to become more condensed themselves and more separate to each other. By doing this, we hope to strengthen the class discrimination ability for both old and new classes, hence maximally preserving the performance on the old tasks. Subsequently, given the separation of the classes in the old and new tasks, we perform OT-MM which incrementally estimates a mixture model, using optimal transport, for each class to tackle the complexity of incoming data streams. Algorithm 1 summarizes the main steps of our method which is Class Increamental Learning with Optimal Transport based Mixture Model (CLOT). Additionally, the key motivations of our approach are presented in Figure 1. In what follows, we present and discuss the technicality of dynamic preservation and OT-MM. Algorithm 1 CLOT Input: The batches $X=[X^{c}]_{c\in\mathcal{C}_{new}}$ and $\bar{X}=[\bar{X}^{c}]_{c\in\mathcal{C}_{old}}$. Output: Feature extractor $f_{\boldsymbol{\theta}}$, the centroids/ covariance matrices $[\boldsymbol{\mu}^{c}_{k},\Sigma^{c}_{k}]_{k=1}^{K}$ for each class $c$ 1: Initialize replay memory $\mathcal{M}$ 2: for each batch $(X,\bar{X})$ do 3: Step 1. Perform dynamic preservation 4: Step 2. Perform OT-MM 5: Step 3. Update the replay memory $\mathcal{M}$ 6: end for ### 4.2 Dynamic Preservation For each class $c\in\mathcal{C}=\mathcal{C}_{new}\cup\mathcal{C}_{old}$, we maintain an aligned prototype $\boldsymbol{w}^{c}$ to provide the logit to the class $c$ for $\boldsymbol{z}=f_{\boldsymbol{\theta}}(\boldsymbol{x})$ as $\langle\boldsymbol{w}^{c},\boldsymbol{z}\rangle$. To strengthen the separation of the classes in the old and new tasks, we update the feature extractor $f_{\boldsymbol{\theta}}$ and aligned prototypes $\boldsymbol{w}^{c}$ by minimizing the following cross-entropy loss over $X=[X^{c}]_{c\in\mathcal{C}_{new}}$ and $\bar{X}=[\bar{X}^{c}]_{c\in\mathcal{C}_{old}}$: $\small{}L_{1}(\boldsymbol{\theta},\boldsymbol{w})=\sum_{c\in\mathcal{C}}\mathbb{E}_{\boldsymbol{x}^{c}\sim X^{c}\text{ or}\,\bar{X}^{c}}\left[-\log\frac{\exp\left\langle\boldsymbol{w}^{c},f_{\boldsymbol{\theta}}\left(\boldsymbol{x}^{c}\right)\right\rangle}{\sum_{c^{\prime}}\exp\left\langle\boldsymbol{w}^{c^{\prime}},f_{\boldsymbol{\theta}}\left(\boldsymbol{x}^{c}\right)\right\rangle}\right].$ (4) After solving the above optimization problem (4), the representation of the classes become more separating. We then compute the mean prototypes $\boldsymbol{p}^{c}$ for each class $c$ as the average of the latent representations, i.e., $\boldsymbol{p}^{c}=\begin{cases}\frac{1}{|X^{c}|}\sum_{\boldsymbol{x}^{c}\in X^{c}}f_{\boldsymbol{\theta}}\left(\boldsymbol{x}^{c}\right)&c\in\mathcal{C}_{new}\\\ \frac{1}{|\bar{X}^{c}|}\sum_{\boldsymbol{x}^{c}\in\bar{X}^{c}}f_{\boldsymbol{\theta}}\left(\boldsymbol{x}^{c}\right)&c\in\mathcal{C}_{old}\end{cases},$ (5) where $|\cdot|$ returns the cardinality of a set. To encourage the representations of the class $c$ to be pulling around its mean prototype and the representations of other classes $c^{\prime}\neq c$ to be pushing away the mean prototype $\boldsymbol{p}^{c}$, we propose to minimize the following loss function: $\small{}L_{2}(\boldsymbol{\theta})=\sum_{c\in\mathcal{C}}\mathbb{E}_{\boldsymbol{x}^{c}\sim X^{c}\text{ or}\,\bar{X}^{c}}\left[-\log\frac{\exp\left\\{-\|f_{\boldsymbol{\theta}}\left(\boldsymbol{x}^{c}\right)-\boldsymbol{p}^{c}\|_{2}^{2}\right\\}}{\sum_{c^{\prime}}\exp\left\\{-\|f_{\boldsymbol{\theta}}\left(\boldsymbol{x}^{c}\right)-\boldsymbol{p}^{c^{\prime}}\|_{2}^{2}\right\\}}\right]$ (6) After performing the dynamic preservation, the latent representations of the classes in the old and new tasks become more condensed themselves and more separate from each other. Resultantly, we increase the class discrimination ability for those in the old and new tasks. Additionally, the feature extractor $f_{\boldsymbol{\theta}}$ has a contraction behavior which helps reduce the shift in the latent representations of the data examples in the old tasks and those in the replay memory. Summarily, we expect that the increasing class discrimination ability and the contracting feature extractor aid CLOT to mitigate the catastrophic forgetting (cf. Figure 1). The key steps of the dynamic preservation is summarized in Algorithm 2. Algorithm 2 Dynamic preservation Input:The batches $X=[X^{c}]_{c\in\mathcal{C}_{new}}$ and $\bar{X}=[\bar{X}^{c}]_{c\in\mathcal{C}_{old}}$ Output:The feature extractor $f_{\boldsymbol{\theta}}$ 1: Update $\boldsymbol{w}^{c\in\mathcal{C}},\boldsymbol{\theta}$ by minimizing (4). 2: Compute the mean prototypes using (5). 3: Update $\boldsymbol{\theta}$ by minimizing (6). We observe that though becoming more condensed and separate, the latent representations of data streams in the new task and replay memory are still sufficiently complex with multi-modality. Therefore, we propose an online OT- based mixture model (OT-MM) approach to incrementally learn the mixture model for each class. Mixture models enable our OT-MM to use multiple centroids to characterize a class in the online learning manner. Figure 2 demonstrates the advantages of using multiple centroids to characterize a class. Specifically, we use t-SNE to visualize the test latent representations of CoPE [8] using one centroid per class and our CLOT using four centroids per class. We observe that there exists a shift between test and train representations. Hence, the centroids tailored to the train set might mismatch the test set. However, it can be seen that using more centroids per class as in our CLOT can mitigate this mismatch. Together with the possibly complex distribution of incoming data streams, this is another motivation for our OT-MM. Figure 3: Average Accuracy by different number of centroids per class (CIFAR10). Figure 4: Average accuracy through tasks. ### 4.3 Optimal transport based mixture model #### 4.3.1 The Derivation of OT-MM OT-MM is one important building-block of our framework. Given a class $c\in\mathcal{C}$ of an old or a new task, we aim to use the mature theoretical body of OT to develop the online OT-MM for data in this class in the sense that the centroids and covariance matrices of the mixture model are incrementally updated according to incoming data streams. Given the latent representations or feature vectors of the class $c$: $D_{c}=\left\\{\boldsymbol{z}_{1}^{c},...,\boldsymbol{z}^{c}_{N_{c}}\right\\}$, wherein each $\boldsymbol{z}_{i}^{c}=f_{\boldsymbol{\theta}}(\boldsymbol{x}_{i}^{c})$ is the latent representation or feature vector of data example $\boldsymbol{x}_{i}^{c}$ via the feature extractor $f_{\boldsymbol{\theta}}$. We denote $\mathbb{P}_{c}:=\frac{1}{N_{c}}\sum_{i=1}^{N_{c}}\delta_{\boldsymbol{z}_{i}^{c}}$ as the empirical data distribution over all data of the class $c$ appearing in all data streams on the latent space. When data are arriving in streams, we need to learn a Gaussian mixture model (GMM) that approximates the data distribution $\mathbb{P}_{c}$. Consider the following GMM $\mathbb{Q}_{c}:=\sum_{k=1}^{K}\pi_{k,c}\mathcal{N}\left(\boldsymbol{\mu}_{k,c},\text{diag}\left(\mathbf{\sigma}_{k,c}^{2}\right)\right),$ where $\boldsymbol{\pi}^{c}=\left[\pi_{k,c}\right]_{k=1}^{K}$ represents mixing proportions and $\boldsymbol{\mu}_{k,c},\text{diag}\left(\mathbf{\sigma}_{k,c}^{2}\right)$ are the mean vector and covariance matrix of the $k$-th Gaussian. To learn this GMM, we propose minimizing a WS distance between $\mathbb{P}_{c}$ and $\mathbb{Q}_{c}$ as $\min_{\boldsymbol{\pi}^{c},\boldsymbol{\mu}^{c},\Sigma^{c}}\mathcal{W}_{d}\left(\mathbb{P}_{c},\sum_{k=1}^{K}\pi_{k,c}\mathcal{N}\left(\boldsymbol{\mu}_{k,c},\text{diag}\left(\mathbf{\sigma}_{k,c}^{2}\right)\right)\right),$ where $\boldsymbol{\pi}^{c}=\left[\pi_{k,c}\right]_{k=1}^{K}$, $\boldsymbol{\mu}^{c}=\left[\boldsymbol{\mu}_{k,c}\right]_{k=1}^{K}$, $\Sigma^{c}=\left[\text{diag}\left(\mathbf{\sigma}_{k,c}\right)\right]_{k=1}^{K}$, and $d$ is a distance on the latent space. To handle the above WS distance, we need the parameterized samples $\tilde{\boldsymbol{z}}^{c}$ from the GMM or $\tilde{\boldsymbol{z}}^{c}\sim\mathbb{Q}_{c}=\sum_{k=1}^{K}\pi_{k,c}\mathcal{N}\left(\boldsymbol{\mu}_{k,c},\text{diag}\left(\mathbf{\sigma}_{k,c}^{2}\right)\right)$. To this end, we first sample $\tilde{\boldsymbol{z}}_{k}^{c}\sim\mathcal{N}\left(\boldsymbol{\mu}_{k,c},\text{diag}\left(\mathbf{\sigma}_{k,c}^{2}\right)\right)$ by using the reparameterization trick $\tilde{\boldsymbol{z}}_{k}^{c}=\boldsymbol{\mu}_{k,c}+\boldsymbol{\epsilon}_{k}\text{diag}\left(\mathbf{\sigma}_{k,c}\right),$ where the source of randomness $\boldsymbol{\epsilon}_{k}\sim\mathcal{N}\left(\mathbf{0},\mathbb{I}\right)$. We then sample the one-hot vector $\boldsymbol{y}\sim\text{Cat}\left(\boldsymbol{\pi}^{c}\right)$ and compute $\tilde{\boldsymbol{z}}^{c}=\sum_{k=1}^{K}y_{k}\tilde{\boldsymbol{z}}_{k}^{c}=\sum_{k=1}^{K}y_{k}\left(\boldsymbol{\mu}_{k,c}+\boldsymbol{\epsilon}_{k}\text{diag}\left(\mathbf{\sigma}_{k,c}\right)\right).$ To do a continuous relaxation [15] of $\boldsymbol{y}$ for enabling learning $\boldsymbol{\pi}^{c}$, we parameterize $\boldsymbol{\pi}^{c}=\text{softmax}\left(\boldsymbol{\alpha}^{c}\right)$ and compute $\displaystyle y_{k}$ $\displaystyle=\frac{\exp\left\\{\left(\log\pi_{k,c}+G_{k}\right)/\tau\right\\}}{\sum_{j=1}^{K}\exp\left\\{\left(\log\pi_{j,c}+G_{j}\right)/\tau\right\\}},k=1,...,K$ $\displaystyle\tilde{\boldsymbol{z}}^{c}$ $\displaystyle=\sum_{k=1}^{K}y_{k}\tilde{\boldsymbol{z}}_{k}^{c}=\sum_{k=1}^{K}y_{k}\left(\boldsymbol{\mu}_{k,c}+\boldsymbol{\epsilon}_{k}\text{diag}\left(\mathbf{\sigma}_{k,c}\right)\right),$ where $\tau>0$ is a temperature parameter and random noises $G_{k}$ are i.i.d. sampled from Gumbel distribution (i.e., $G_{k}=-\log\left(-\log u_{k}\right)$ for $u_{k}\sim\text{Uniform}(0;1)$). To train the WS distance of interest, we use the $\varepsilon$-entropic dual- form [9] (cf. Section 3.2) as $\displaystyle\mathcal{W}^{\varepsilon}_{d}\left(\mathbb{P}_{c},\sum_{k=1}^{K}\pi_{k,c}\mathcal{N}\left(\boldsymbol{\mu}_{k,c},\text{diag}\left(\mathbf{\sigma}_{k,c}^{2}\right)\right)\right)$ $\displaystyle=\max_{\phi}\left\\{\mathbb{E}_{\mathbb{P}_{c}}\left[\phi\left(\boldsymbol{z}^{c}\right)\right]+\mathbb{E}_{\mathbb{Q}_{c}}\left[\tilde{\phi}\left(\tilde{\boldsymbol{z}}^{c}\right)\right]\right\\},$ (7) where $\varepsilon>0$ is a small number, $\phi$ is the Kantorovich network, $\tilde{\boldsymbol{z}}^{c}=\sum_{k=1}^{K}y_{k}\left(\boldsymbol{\mu}_{k,c}+\boldsymbol{\epsilon}_{k}\text{diag}\left(\mathbf{\sigma}_{k,c}\right)\right)$, and $\tilde{\phi}\left(\tilde{\boldsymbol{z}}^{c}\right)=-\varepsilon\log\left(\mathbb{E}_{\mathbb{P}_{c}}\left[\exp\left\\{\frac{-d\left(\boldsymbol{z}^{c},\tilde{\boldsymbol{z}}^{c}\right)+\phi\left(\boldsymbol{z}^{c}\right)}{\varepsilon}\right\\}\right]\right).$ #### 4.3.2 OT-MM in Online Continual learning Scenario We now present how to perform our OT-MM in the online continual learning scenario when we need to update the set of centroids and covariance matrices for a class $c\in\mathcal{C}$ based on the batch $X^{c}$ or $\bar{X}^{c}$. Looking into Eq. (7), this objective function is in the form of expectation, hence perfectly fitting for online learning. Specifically, we use the current batch $X^{c}$ or $\bar{X}^{c}$ for each class $c$ to solve $\displaystyle\min_{\boldsymbol{\pi}^{c},\boldsymbol{\mu}^{c},\Sigma^{c}}\max_{\phi}\left\\{\mathbb{E}_{X^{c}\,or\,\bar{X}^{c}}\left[\phi\left(f_{\boldsymbol{\theta}}(\boldsymbol{x}^{c})\right)\right]+\mathbb{E}_{\mathbb{Q}_{c}}\left[\tilde{\phi}\left(\tilde{\boldsymbol{z}}^{c}\right)\right]\right\\}.$ (8) To solve the optimization problem (8), we update the Kantorovich network $\phi$ several times and then update the mixing propotions $\boldsymbol{\pi}^{c}$, the set of centroids $\boldsymbol{\mu}^{c}$, and the set of covariance matrices $\Sigma^{c}$ for class $c$ in several times. The key steps of OT-MM is summarized in Algorithm 3. Eventually, we harvest the centroids $[\boldsymbol{\pi}^{c}]_{c\in\mathcal{C}}$ and covariance matrices $[\Sigma^{c}]_{c\in\mathcal{C}}$ to select diverge replay memory for a task and do inference in the testing phase (cf. Section 4.4). Algorithm 3 OT-GMM in the OCL scenario Input:The batches $X=[X^{c}]_{c\in\mathcal{C}_{new}}$ and $\bar{X}=[\bar{X}^{c}]_{c\in\mathcal{C}_{old}}$ Output:$[\boldsymbol{\pi}^{c},\boldsymbol{\mu}^{c},\Sigma^{c}]_{c\in\mathcal{C}}$ 1: for each $c\in\mathcal{C}$ do 2: Update $\phi$ according to (8). 3: Update $[\boldsymbol{\pi}^{c},\boldsymbol{\mu}^{c},\Sigma^{c}]$ according to (8). 4: end for Method | M = 1k | M = 3k | M = 5k | M = 0.2k | M = 0.5k | M = 1k | M = 0.1k | M = 0.5k | M = 1.5k ---|---|---|---|---|---|---|---|---|--- iid-offline | $50.28\pm 0.66$ | $50.28\pm 0.66$ | $50.28\pm 0.66$ | $83.02\pm 0.60$ | $83.02\pm 0.60$ | $83.02\pm 0.60$ | $98.44\pm 0.02$ | $98.44\pm 0.02$ | $98.44\pm 0.02$ iid-online | $20.10\pm 0.90$ | $20.10\pm 0.90$ | $20.1\pm 0.90$ | $62.31\pm 1.67$ | $62.31\pm 1.67$ | $62.31\pm 1.67$ | $96.57\pm 0.14$ | $96.57\pm 0.14$ | $96.57\pm 0.14$ DVC [11] | $\mathbf{10.93\pm 0.49}$ | $14.49\pm 0.90$ | $20.83\pm 0.89$ | $28.56\pm 2.10$ | $34.92\pm 2.27$ | $45.02\pm 2.72$ | _ | _ | _ OCD [18] | $3.71\pm 0.5$ | $9.2\pm 0.21$ | $11.56\pm 0.32$ | $15.83\pm 2.30$ | $21.02\pm 5.24$ | $26.45\pm 2.40$ | _ | _ | _ MIR [1] | $8.20\pm 0.72$ | $14.9\pm 0.40$ | $20.00\pm 0.57$ | $24.3\pm 2.01$ | $35.6\pm 2.56$ | $42.80\pm 2.22$ | $73.6\pm 0.50$ | $86.5\pm 0.22$ | $91.2\pm 0.36$ OCS [50] | $8.50\pm 0.35$ | $13.76\pm 0.70$ | $19.98\pm 0.65$ | _ | _ | _ | _ | _ | _ ASER [36] | $9.60\pm 0.88$ | $13.08\pm 0.87$ | $15.00\pm 0.64$ | $22.38\pm 2.94$ | $26.42\pm 4.80$ | $30.67\pm 3.10$ | $58.02\pm 0.07$ | $75.47\pm 0.04$ | $85.68\pm 0.03$ Reservoir [41] | $8.60\pm 1.58$ | $13.5\pm 2.61$ | $19.60\pm 0.53$ | $25.96\pm 1.41$ | $33.85\pm 1.76$ | $40.531\pm 5.30$ | $69.22\pm 1.15$ | $86.50\pm 1.04$ | $91.27\pm 0.96$ GDUMB [28] | $8.48\pm 0.26$ | $11.24\pm 0.56$ | $15.63\pm 0.60$ | $24.60\pm 1.50$ | $28.70\pm 0.63$ | $33.24\pm 2.10$ | $73.5\pm 0.12$ | $86.16\pm 0.01$ | $92.07\pm 0.01$ CoPE [8] | $9.22\pm 0.55$ | $15.15\pm 0.50$ | $19.56\pm 1.22$ | $37.20\pm 1.80$ | $42.22\pm 0.54$ | $47.98\pm 2.62$ | $72.57\pm 2.10$ | $88.92\pm 0.99$ | $92.50\pm 0.33$ CLOT (Ours) | $10.62\pm 0.45$ | $\mathbf{15.80\pm 0.55}$ | $\mathbf{22.35\pm 0.40}$ | $\mathbf{42.28\pm 0.50}$ | $\mathbf{50.08\pm 1.00}$ | $\mathbf{54.45\pm 1.20}$ | $\mathbf{77.40\pm 0.21}$ | $\mathbf{90.98\pm 0.35}$ | $\mathbf{93.71\pm 0.05}$ | a) CIFAR-100 | b) CIFAR-10 | c) MNIST Table 1: Average Accuracy (higher is better), M denotes the memory buffer size. All numbers are the average of 5 runs. The data in the table represents Average Accuracy ± standard deviation. ### 4.4 Replay memory selection and inference process We now present how to utilize the output of the OT-MM for selecting the replay memory of the new task and how to do inference. ##### Replay memory selection: After processing each batch $(X,\bar{X})$, as shown in Step 3 (Line 5) in Algorithm 1, we select some more data points to supplement the replay memory of the new task as follows: * • For a centroid of a class, we choose some closest data points of this class in the current batch to add to the replay memory. * • If the replay memory is full, we randomly pick some data points in the current replay memory to replace by the fresh-new ones. ##### Doing inference: Given an unseen data point $\boldsymbol{x}$, we compute the distance of $\boldsymbol{x}$ to each class $c$ and classify $\boldsymbol{x}$ to the closest class as follows: $\hat{y}=\arg\min_{c\in\mathcal{C}}d\left(\boldsymbol{x},c\right)\text{ where }d\left(\boldsymbol{x},c\right)=\min_{k\leq K}d(f_{\boldsymbol{\theta}}\left(\boldsymbol{x}\right),\boldsymbol{\mu}_{k,c})$ ## 5 Experiments In this section, we review the experimental settings including datasets, baselines, and metrics used to compare our method against other state-of-the- art models. We then report and analyse the results to validate the effectiveness of our approach. ### 5.1 Experimental setup Datasets: We use three benchmark datasets including: * • Split-MNIST is constructed by splitting the MNIST dataset (with 60k training samples) into 5 disjoint subsets, corresponding to 5 tasks, each of which consists of 2 classes. * • Split-CIFAR10 is constructed by splitting CIFAR10 dataset into 5 disjoint tasks, hence each task has 2 different classes. * • Split-CIFAR100 is a variant CL dataset constructed from CIFAR100 where 50k training samples of 100 classes is divided into 20 tasks of 5 classes. Architectures: For the experiments on Split-MNIST dataset, we use a simple MLP neural network with 2 hidden layers of 400 units. While a slim-version of ResNet-18 will be used to evaluate on Split-CIFAR10 and Split-CIFAR100. Baselines: We use several state-of-the-art methods in Continual learning, specifically in Online Continual Learning (OCL), to compare with our proposed method. * • iid-online and iid-offline: give the performance when ignoring the challenges of non-stationary data stream in OCL scenarios. Both of them train a model from i.i.d. sampled mini-batches (i.e. there is no change of the distribution of data in the learning process). While iid-online only allows learning once over data, iid-offline allows a model to be trained from multiple epochs. * • Regarding the recent methods for OCL, the chosen baselines can divided into 2 main groups. The first group focuses on improving the quality of replay memory: Reservoir, OCS (ICRL2022), MIR (NeurIPS2019) are well-known techniques to build memory buffer, GDumb (ECCV2020) is interested in the way to retrieve data more effectively, while ASER (AAAI2021) consider both updating and retrieving data from memory. The second group includes the methods that help improve a model mainly through more effective representation learning: DVC (CVPR2022), OCD (IJICAI2022), CoPE (ICCV2021). ### 5.2 Performance comparison Figure 5: Features on latent space of our method (a and b) and CoPE (c). It can be observed that 4 centroids is better than 1 centroid. CLOT 1-centroid is better than CoPE 1-centroid due to the effect of the dynamic preservation for learning better latent representations. Table 1 summarizes the experimental results on three datasets with various memory sizes. On MNIST, we compare our method with five baselines and achieve better results. Especially when memory sizes are small (0.1k), we exceed the strongest baseline (MIR) by roughly 4%. On CIFAR10, CLOT again outperforms the SOTA baselines. The largest gap between CLOT and the best baseline is around 8%. On CIFAR100, the memory size for each class and the batch size in each learning step both are very small, making this the most challenging dataset to learn. The results show that with a sufficient large memory size (3k, 5k), our CLOT surpasses the baselines, while with a too-small memory size, our CLOT is the runner-up method. Among three datasets, the superiority of CLOT is most apparent in CIFAR10. This because that MNIST is a fairy easy dataset, especially when the memory is large enough to get the representational diversity of the old samples. Thus, for the results on MNIST with memory size 1.5K, the methods do not show clear performance differences. As for CIFAR100, we can see that our method needs a sufficient amount of data to show its strengths in learning representation. Because with too small memory size and batch size, some samples from all classes are not always available. Figure 4 shows the visual results about the average accuracy on observed tasks of the different methods with the large enough memory size on each dataset. It is easy to see that our method outperforms the other baselines, especially as the number of tasks increases. This is most evident in CIFAR10, consistent with the reasons discussed earlier. ### 5.3 Ablation study In what follows, we present additional experimental results to analyse the effectiveness of components in CLOT. The role of using multiple centroids in general: For our OT-MM, we examine the influence of number of centroids per class. Figure 3 depicts the curves of average accuracies on CIFAR10 when Num-centroids varies. We observe that in all cases, the performance improves when we increase the number of centroids per class from 1 to a certain threshold. The bigger the memory size is, the bigger the threshold is. Moreover, if these thresholds are exceeded, the model performance depends on the support level of replay memory: the smaller memory size leads to the lower the representation learning quality, and higher prediction error. When memory size = 0.2K, if the number of centroids is bigger than 3, the model performance degrades most. On the other hand, with a larger memory size (e.g., memory size = 1K), the ideal number of centroids is 4, and when increasing the number of centroids to 5, the model quality does not degrade significantly. Therefore, it can be seen that our prediction method using many centroids would achieve the best results when there is a harmonious combination with the representation learning quality. The role of centroids in improving replay buffer: In our framework, we leverage centroids learned from OT-MM to improve the quality of the memory buffer. Table 2 is an ablation study on the effect of using or not using centroids to select samples for the replay buffer. It can be observed that in all cases, using centroids to select samples consistently offers better results than not using centroids. Regardless that the sample selection based on centroids is quite simple, the results support our intuition that centroids help effectively characterize data and improve the diversity in the episode memory. | Num of centroids ---|--- | 1 | 2 | 3 | 4 | 5 | 8 Use centroids | 50.1 | 52.5 | 53.8 | 54.5 | 54.1 | 54 w/o centroids | 50.1 | 52.11 | 53.2 | 54.0 | 53.6 | 53.3 Table 2: Effect of using centroids on replay memory selection. We compare the performance when using centroids to select samples (Use centroids) and when selecting samples at random for the replay memory (w/o centroids). The role of centroids in prediction: To verify the role of using multiple centroids when making decisions, we present $t$-SNE visualization (Figure 5) of the feature space of a setting on CIFAR10. Figures 5a and 5b illustrate our approach when using multiple centroids and only 1 centroid when predicting. We can see that, in the real scenario, the features of the classes are usually distributed with multi-modality. Therefore, using more centroids helps to characterize more accurately multi-modal distributions, thus generally giving better prediction than using only 1 centroid. Furthermore, Figure 5c illustrates the corresponding results on latent space using CoPE where the features of different classes are not well separated and each class has only one centroid for prediction. When comparing 5c with 5a and 5b, we again emphasize the effectiveness of our framework in improving the quality of representation learning via dynamic preservation, along with making decisions based on multiple centroids produced by OT-MM. ## 6 Conclusion In this work, we have proposed a novel method for Online CIL scenarios, which is for the first time that OT theory has been leveraged with Continual learning. We have introduced a new perspective that helps the network to make predictions based on many centroids for each class of data to tackle more efficiently their multi-modality. In addition, we have further proposed a strategy of remembering old knowledge flexibly, called Dynamic Preservation, which complements to the prediction mechanism above. Furthermore, the extensive experiments on three commonly used benchmark datasets demonstrate the effectiveness of our method. ## References * [1] Rahaf Aljundi, Eugene Belilovsky, Tinne Tuytelaars, Laurent Charlin, Massimo Caccia, Min Lin, and Lucas Page-Caccia. Online continual learning with maximal interfered retrieval. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 11849–11860. Curran Associates, Inc., 2019. * [2] Adrian Bulat, Jean Kossaifi, Georgios Tzimiropoulos, and Maja Pantic. Incremental multi-domain learning with network latent tensor factorization. In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, pages 10470–10477. AAAI Press, 2020. * [3] Lucas Caccia, Rahaf Aljundi, Nader Asadi, Tinne Tuytelaars, Joelle Pineau, and Eugene Belilovsky. New insights on reducing abrupt representation change in online continual learning. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, 2022. * [4] Arslan Chaudhry, Marc’Aurelio Ranzato, Marcus Rohrbach, and Mohamed Elhoseiny. Efficient lifelong learning with a-GEM. In International Conference on Learning Representations, 2019. * [5] Arslan Chaudhry, Marcus Rohrbach, Mohamed Elhoseiny, Thalaiyasingam Ajanthan, Puneet K Dokania, Philip HS Torr, and M Ranzato. Continual learning with tiny episodic memories. 2019\. * [6] Aristotelis Chrysakis and Marie-Francine Moens. Online continual learning from imbalanced data. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pages 1952–1961. PMLR, 2020. * [7] Matthias De Lange, Rahaf Aljundi, Marc Masana, Sarah Parisot, Xu Jia, Aleš Leonardis, Gregory Slabaugh, and Tinne Tuytelaars. A continual learning survey: Defying forgetting in classification tasks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(7):3366–3385, 2021. * [8] Matthias De Lange and Tinne Tuytelaars. Continual prototype evolution: Learning online from non-stationary data streams. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 8250–8259, October 2021. * [9] Aude Genevay, Marco Cuturi, Gabriel Peyré, and Francis Bach. Stochastic optimization for large-scale optimal transport. Advances in neural information processing systems, 29, 2016. * [10] Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron C. Courville, and Yoshua Bengio. Generative adversarial nets. In Zoubin Ghahramani, Max Welling, Corinna Cortes, Neil D. Lawrence, and Kilian Q. Weinberger, editors, Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada, pages 2672–2680, 2014. * [11] Yanan Gu, Xu Yang, Kun Wei, and Cheng Deng. Not just selection, but exploration: Online class-incremental continual learning via dual view consistency. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 7442–7451, June 2022. * [12] Yunhui Guo, Mingrui Liu, Tianbao Yang, and Tajana Rosing. Improved schemes for episodic memory-based lifelong learning. Advances in Neural Information Processing Systems, 33, 2020. * [13] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pages 770–778. IEEE Computer Society, 2016. * [14] Saihui Hou, Xinyu Pan, Chen Change Loy, Zilei Wang, and Dahua Lin. Learning a unified classifier incrementally via rebalancing. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019. * [15] Eric Jang, Shixiang Gu, and Ben Poole. Categorical reparameterization with gumbel-softmax. arXiv preprint arXiv:1611.01144, 2016. * [16] Zixuan Ke, Bing Liu, and Xingchang Huang. Continual learning of a mixed sequence of similar and dissimilar tasks. 33, 2020. * [17] James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, et al. Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 114(13):3521–3526, 2017. * [18] Jin Li, Zhong Ji, Gang Wang, Qiang Wang, and Feng Gao. Learning from students: Online contrastive distillation network for general continual learning. In Proceedings of the International Joint Conference on Artificial Intelligence, 2022. * [19] Zhizhong Li and Derek Hoiem. Learning without forgetting. IEEE transactions on pattern analysis and machine intelligence, 40(12):2935–2947, 2017. * [20] Zheda Mai, Ruiwen Li, Hyunwoo Kim, and Scott Sanner. Supervised contrastive replay: Revisiting the nearest class mean classifier in online class-incremental continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3589–3599, 2021. * [21] Margarita Martínez-Díaz and Francesc Soriguera. Autonomous vehicles: theoretical and practical challenges. Transportation Research Procedia, 33:275–282, 2018. * [22] Marc Masana, Xialei Liu, Bartlomiej Twardowski, Mikel Menta, Andrew D Bagdanov, and Joost van de Weijer. Class-incremental learning: survey and performance evaluation on image classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022\. * [23] Heinrich Mellmann, Marcus Scheunemann, and Oliver Stadie. Adaptive grasping for a small humanoid robot utilizing force- and electric current sensors. volume 1032, 09 2013. * [24] Muhammad Jehanzeb Mirza, Marc Masana, Horst Possegger, and Horst Bischof. An efficient domain-incremental learning approach to drive in all weather conditions. In IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2022, New Orleans, LA, USA, June 19-20, 2022, pages 3000–3010. IEEE, 2022. * [25] Seyed-Iman Mirzadeh, Mehrdad Farajtabar, Dilan Görür, Razvan Pascanu, and Hassan Ghasemzadeh. Linear mode connectivity in multitask and continual learning. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021. * [26] L. Olsson, C.L. Nehaniv, and D. Polani. Sensor adaptation and development in robots by entropy maximization of sensory data. In 2005 International Symposium on Computational Intelligence in Robotics and Automation, pages 587–592, 2005. * [27] Oleksiy Ostapenko, Mihai Puscas, Tassilo Klein, Patrick Jahnichen, and Moin Nabi. Learning to remember: A synaptic plasticity driven framework for continual learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 11321–11329, 2019. * [28] Ameya Prabhu, Philip Torr, and Puneet Dokania. Gdumb: A simple approach that questions our progress in continual learning. In The European Conference on Computer Vision (ECCV), August 2020\. * [29] Ameya Prabhu, Philip H. S. Torr, and Puneet K. Dokania. Gdumb: A simple approach that questions our progress in continual learning. In Andrea Vedaldi, Horst Bischof, Thomas Brox, and Jan-Michael Frahm, editors, Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part II, volume 12347 of Lecture Notes in Computer Science, pages 524–540. Springer, 2020\. * [30] Joseph Redmon, Santosh Kumar Divvala, Ross B. Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pages 779–788. IEEE Computer Society, 2016. * [31] Douglas A Reynolds. Gaussian mixture models. Encyclopedia of biometrics, 741(659-663), 2009. * [32] Anthony Robins. Catastrophic forgetting, rehearsal and pseudorehearsal. Connection Science, 7(2):123–146, 1995. * [33] Gobinda Saha, Isha Garg, and Kaushik Roy. Gradient projection memory for continual learning. In International Conference on Learning Representations, 2021. * [34] F. Santambrogio. Optimal transport for applied mathematicians. Birkäuser, 2015. * [35] Dongsub Shim, Zheda Mai, Jihwan Jeong, Scott Sanner, Hyunwoo Kim, and Jongseong Jang. Online class-incremental continual learning with adversarial shapley value. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, pages 9630–9638. AAAI Press, 2021. * [36] Dongsub Shim, Zheda Mai, Jihwan Jeong, Scott Sanner, Hyunwoo Kim, and Jongseong Jang. Online class-incremental continual learning with adversarial shapley value. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 9630–9638, 2021. * [37] Hanul Shin, Jung Kwon Lee, Jaehong Kim, and Jiwon Kim. Continual learning with deep generative replay. Advances in neural information processing systems, 30, 2017. * [38] Siddharth Swaroop, Cuong V. Nguyen, Thang D. Bui, and Richard E. Turner. Improving and understanding variational continual learning. arXiv:1905.02099 [cs, stat], 2019. * [39] Rahamathunnisa Usuff and Saravanan Ramakrishnan. A survey on video streaming over multimedia networks using tcp. Journal of Theoretical and Applied Information Technology, 53:205–209, 07 2013. * [40] Cédric Villani. Optimal transport: old and new, volume 338. Springer, 2009. * [41] Jeffrey Scott Vitter. Random sampling with a reservoir. ACM Trans. Math. Softw., 11(1):37–57, 1985. * [42] Johannes Von Oswald, Christian Henning, João Sacramento, and Benjamin F Grewe. Continual learning with hypernetworks. arXiv preprint arXiv:1906.00695, 2019. * [43] Jiushuang Wang, Weizhang Xu, and Jian Wang. A study of live video streaming system for mobile devices. In 2016 First IEEE International Conference on Computer Communication and the Internet (ICCCI), pages 157–160, 2016. * [44] Yue Wu, Yinpeng Chen, Lijuan Wang, Yuancheng Ye, Zicheng Liu, Yandong Guo, and Yun Fu. Large scale incremental learning. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pages 374–382. Computer Vision Foundation / IEEE, 2019. * [45] Yongqin Xian, Christoph H. Lampert, Bernt Schiele, and Zeynep Akata. Zero-shot learning - A comprehensive evaluation of the good, the bad and the ugly. IEEE Trans. Pattern Anal. Mach. Intell., 41(9):2251–2265, 2019\. * [46] Yi Xiao, Felipe Codevilla, Akhil Gurram, Onay Urfalioglu, and Antonio M. López. Multimodal end-to-end autonomous driving. IEEE Trans. Intell. Transp. Syst., 23(1):537–547, 2022. * [47] Jiangwei Xie, Shipeng Yan, and Xuming He. General incremental learning with domain-aware categorical representations. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022, pages 14331–14340. IEEE, 2022. * [48] Jiacong Xu, Zixiang Xiong, and Shankar P. Bhattacharyya. Pidnet: A real-time semantic segmentation network inspired from pid controller, 2022. * [49] Dong Yin, Mehrdad Farajtabar, and Ang Li. Sola: continual learning with second-order loss approximation. 2020\. * [50] Jaehong Yoon, Divyam Madaan, Eunho Yang, and Sung Ju Hwang. Online coreset selection for rehearsal-based continual learning. In International Conference on Learning Representations, 2022. * [51] Friedemann Zenke, Ben Poole, and Surya Ganguli. Continual learning through synaptic intelligence. In International Conference on Machine Learning, pages 3987–3995. PMLR, 2017. * [52] Bowen Zhao, Xi Xiao, Guojun Gan, Bin Zhang, and Shu-Tao Xia. Maintaining discrimination and fairness in class incremental learning. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pages 13205–13214. Computer Vision Foundation / IEEE, 2020. ## 7 Supplementary Material ### 7.1 Implementation Detail We implement our proposed method and baselines on the same code base, based on the learner-evaluator framework proposed in [8]. Our code is available at https://github.com/tranquyenbk173/Streaming_WSD_GGM #### 7.1.1 Hyperparameters configuration We shared the same replay memory size (total samples of all classes) per dataset for each method, and set the batch size for all methods to 10, the size of the retrieved batch is the same as the batch size from data stream. The optimizers are all SGD, the learning rate and other hyperparameters were trying to adjust to get the best results for each method. In addition to that, for a fair comparison, the classes in each task and the order of tasks are fixed in all experiments of the same kind. Table 3 presents the setting. Hyper-params | Split | Split | Split ---|---|---|--- | CIFAR100 | CIFAR10 | MNIST batch_size | 10 | 10 | 10 memory size | [ 1k, 3k, 5k ] | [ 0.2k, 0.5k, 1k ] | [ 0.1k, 0.5k, 1.5k ] Table 3: Shared Hyperparameter configurations among our method and baselines for three datasets. #### 7.1.2 Evaluation and metrics In all experiments, we use a single-head architecture, where the final classifier layer is shared across all the tasks (except for CLOT and CoPE, only features extractor is available), and the task identity is not provided during inference. Following the CL literature [4, 25], we used the following metrics to evaluate: * • Average accuracy ($\mathcal{A}_{T}$): Averaged test accuracy of all tasks after completing learning T task. $\mathcal{A}_{t}=\frac{1}{T}\sum_{j=1}^{T}a_{T,j},$ where $a_{i,j}$ is the performance of the model on the held-out testing set of task $j$ after the model is trained from task $1$ to $i$. * • Average forgetting ($\mathcal{F}_{T}$): The averaged gap between the highest recorded and final task accuracy at the end of continual learning on $T$ tasks. $\mathcal{F}_{T}=\frac{1}{T-1}\sum_{j=1}^{T-1}f_{T,j},$ where $f_{i,j}=\max_{l\in\\{1,...,i-1\\}}a_{l,j}-a_{i,j}$ ### 7.2 Additional Experiments #### 7.2.1 Average forgetting Table 4 shows the average forgetting at the end of data stream of methods on 3 datasets. Our CLOT is the method with the lowest average forgetting in most cases. The most enormous difference is shown on the MNIST dataset, with a gap of about 6 % from the most robust baseline (MIR). On the CIFAR10, CLOT avoids forgetting better than the best method by more than 2.6%. With the most challenging dataset - CIFAR100, our method also shows an advantage over other methods. On the other hand, Figure 6 depicts the average accuracy of methods through tasks. We can see that, in general, CLOT is the method with the lowest forgetting rate (on the MNIST and CIFAR100 datasets) / or in the group with the least forgetting rate (on the CIFAR10 set) during the continual learning process. At the end of the data stream, our proposed method shows a clear difference in avoiding catastrophic forgetting compared to the baselines on the two datasets: MNIST and CIFAR10. These evidences once again confirm the effectiveness of our framework in online continual learning scenario. Figure 6: Average forgetting through tasks. Lower is better. Method | M = 1k | M = 3k | M = 5k | M = 0.2k | M = 0.5k | M = 1k | M = 0.1k | M = 0.5k | M = 1.5k ---|---|---|---|---|---|---|---|---|--- DVC [11] | $\mathbf{52.25\pm 0.41}$ | $51.69\pm 0.34$ | $50.36\pm 0.46$ | $60.15\pm 5.5$ | $55.2\pm 5.99$ | $44.37\pm 1.21$ | _ | _ | _ OCD [18] | $68.55\pm 1.5$ | $65.33\pm 0.78$ | $64.1\pm 0.57$ | $68.48\pm 4.41$ | $62.34\pm 7.59$ | $51.09\pm 1.62$ | _ | _ | _ MiR [1] | $58.862\pm 0.97$ | $56.56\pm 0.81$ | $54.55\pm 0.72$ | $58.23\pm 4.3$ | $56.65\pm 1.6$ | $46.34\pm 3.2$ | $33.2\pm 2.59$ | $16.01\pm 1.51$ | $8.87\pm 1.8$ OCS [50] | $60.01\pm 1.81$ | $57.22\pm 0.55$ | $56.88\pm 0.6$ | _ | _ | _ | _ | _ | _ ASER [36] | $58.71\pm 0.84$ | $59.44\pm 0.92$ | $59.67\pm 0.79$ | $62.05\pm 3.2$ | $60.01\pm 1.73$ | $56.67\pm 3.27$ | $40.55\pm 3.12$ | $20.53\pm 2.52$ | $25.4\pm 0.25$ Reservoir [41] | $58.91\pm 1.7$ | $56.70\pm 0.5$ | $55.88\pm 0.87$ | $57.62\pm 2.45$ | $52.25\pm 1.41$ | $48.53\pm 2.3$ | $35.58\pm 2.5$ | $15.69\pm 2.1$ | $9.48\pm 1.13$ GDUMB [28] | $59.11\pm 1.17$ | $57.83\pm 0.82$ | $56.55\pm 0.66$ | $57.88\pm 1.89$ | $57.15\pm 2.1$ | $55.16\pm 1.55$ | $33.5\pm 2.89$ | $12.66\pm 0.2$ | $8.12\pm 0.93$ CoPE [8] | $55.17\pm 0.3$ | $54.65\pm 0.45$ | $48.47\pm 0.99$ | $52.35\pm 2.15$ | $49.76\pm 2.39$ | $42.99\pm 2.87$ | $34.49\pm 3.62$ | $12.915\pm 1.8$ | $7.53\pm 1.2$ CLOT (Ours) | $53.21\pm 0.51$ | $\mathbf{51.01\pm 0.55}$ | $\mathbf{47.59\pm 0.57}$ | $\mathbf{51.32\pm 2.3}$ | $\mathbf{47.08\pm 3.1}$ | $\mathbf{40.25\pm 1.1}$ | $\mathbf{27.46\pm 3.17}$ | $\mathbf{10.74\pm 0.21}$ | $\mathbf{6.27\pm 0.28}$ | a) CIFAR-100 | b) CIFAR-10 | c) MNIST Table 4: Average Forgetting (lower is better), M denotes the memory buffer size. All numbers are the average of 5 runs. The data in the table represents Average Forgetting ± standard deviation. #### 7.2.2 Additional technicalities Clip grad proportion of prototypes | Clip grad proportion ---|--- | 0.0 | 0.001 | 0.05 | 0.08 | 0.1 | 0.15 | 0.5 MNIST | 93.22 | 93.46 | 93.50 | 93.58 | 93.71 | 93.20 | 92.80 CIFAR10 | 53.23 | 53.55 | 54.41 | 54.45 | 52.40 | 49.12 | 46.55 Table 5: Average Accuracy by different clip grad proportion In the Dynamic Preservation phrase, when solving the problem (4): $\small{}L_{1}(\boldsymbol{\theta},\boldsymbol{w})=\sum_{c\in\mathcal{C}}\mathbb{E}_{\boldsymbol{x}^{c}\sim X^{c}\text{ or}\,\bar{X}^{c}}\left[-\log\frac{\exp\left\langle\boldsymbol{w}^{c},f_{\boldsymbol{\theta}}\left(\boldsymbol{x}^{c}\right)\right\rangle}{\sum_{c^{\prime}}\exp\left\langle\boldsymbol{w}^{c^{\prime}},f_{\boldsymbol{\theta}}\left(\boldsymbol{x}^{c}\right)\right\rangle}\right],$ (4) we use a secondary technique to control the updating of the prototypes as follows: * • Compute gradient of $L_{1}$ w.r.t prototypes: $g_{\boldsymbol{p}}=\nabla_{\boldsymbol{p}}L_{1}$ * • Adjust the gradient for prototypes: new prototypes are allowed to learn more than old prototypes. With clip grad proportion $\alpha\in\left(0,1\right)$, we update: $\boldsymbol{p^{c}}=\left\\{\begin{matrix}\boldsymbol{p}^{c}-\eta*\alpha*g_{\boldsymbol{p}}&c\in\mathcal{C}_{old},\\\ \boldsymbol{p}^{c}-\eta*g_{\boldsymbol{p}}&c\in\mathcal{C}_{new}\end{matrix}\right.$ Clip grad proportion $\alpha$ is a hyperparameter that tells us how much portability we allow for prototypes. Table 5 shows the results when changing the value of clip grad proportion. The experiments on MNIST give the best results with $alpha=0.001$, and the ideal value is $0.08$ on CIFAR10. In continual learning, the representation of features often shift too much, especially at transition times between tasks. If we learn prototypes arbitrarily without control (especially when the old data reviewed are very limited), they then move to unfavorable locations, making the features of different classes to be unseparated well, on both the training and testing sets. The table also shows that, on all 3 datasets, if the clip grad proportion is too high, the average accuracy is highly reduced. While, if it is too small, the model performance is greatly stable. That is, the features of the old classes are fixedly distributed in some places that is only good for the old tasks, and the new tasks does not have much space to place the new features. Prototype adjustment Figure 7: $t$-SNE visualization of learned features and prototypes. The features of different classes are assigned different colors. The prototypes are located in the position of the red ”X” signs. In this section, we conduct additional experiments and explanations regarding why we need to adjust the aligned prototypes to the mean prototypes (a.k.a. the prototype adjustment) when compressing the feature vectors in the dynamic preservation. Our practical observations show that the aligned prototypes (used to learn feature extractor by some objective functions like Cross Entropy loss) do not often locate at the centre of each class and also very close to each other (Figure 7). This seems to be detrimental. Additionally, in the next update step (6), we immobilize the prototypes to learn the feature extractor so that the features will move to the corresponding prototypes. Therefore, if we use the aligned prototypes directly for (6), the clusters of features (of different classes) will be overlapped. Consequently, the predictive efficiency may be degraded. Moreover, after performing (4), the features of different classes are quite separable. Hinted by this observation, we compute mean prototypes in (5) and use them to solve the optimization problem in (6) for compressing the feature vectors around the corresponding mean prototypes. Figure 8: Performance of CLOT when adjusting prototypes (Adjust proto) and not (Not adjust proto), according to (5) Figure 8 shows the effect of prototype adjustment in (5). We can see this change helps to solve the issue mentioned above and improve the performance. On all 3 datasets, when adjusting the prototypes according to (5), we obtain higher average accuracies and lower average forgetting scores.
# Dissipative reactions with intermediate-energy beams – a novel approach to populate complex-structure states in rare isotopes A. Gade Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA Department of Physics and Astronomy, Michigan State University, East Lansing, Michigan 48824, USA B. A. Brown Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA Department of Physics and Astronomy, Michigan State University, East Lansing, Michigan 48824, USA D. Weisshaar Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA D. Bazin Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA Department of Physics and Astronomy, Michigan State University, East Lansing, Michigan 48824, USA K. W. Brown Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA Department of Chemistry, Michigan State University, East Lansing, Michigan 48824, USA R. J. Charity Department of Chemistry, Washington University, St. Louis, Missouri 63130, USA P. Farris Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA Department of Physics and Astronomy, Michigan State University, East Lansing, Michigan 48824, USA A. M. Hill Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA Department of Physics and Astronomy, Michigan State University, East Lansing, Michigan 48824, USA J. Li Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA B. Longfellow Present address: Lawrence Livermore National Laboratory, Livermore, California 94550, USA Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA Department of Physics and Astronomy, Michigan State University, East Lansing, Michigan 48824, USA D. Rhodes Present address: TRIUMF, 4004 Wesbrook Mall, Vancouver, BC V6T 2A3, Canada Facility for Rare Isotope Beams, Michigan State University, East Lansing, Michigan 48824, USA Department of Physics and Astronomy, Michigan State University, East Lansing, Michigan 48824, USA W. Reviol Physics Division, Argonne National Laboratory, Argonne, Illinois 60439, USA J. A. Tostevin Department of Physics, Faculty of Engineering and Physical Sciences, University of Surrey, Guildford, Surrey GU2 7XH, United Kingdom ###### Abstract A novel pathway for the formation of multi-particle-multi-hole ($np-mh$) excited states in rare isotopes is reported from highly energy- and momentum- dissipative inelastic-scattering events measured in reactions of an intermediate-energy beam of 38Ca on a Be target. The negative-parity, complex- structure final states in 38Ca were observed following the in-beam $\gamma$-ray spectroscopy of events in the 9Be($\hbox{${}^{38}$Ca},\hbox{${}^{38}$Ca}+\gamma$)X reaction in which the scattered projectile lost longitudinal momentum of order $\Delta p_{||}=700$ MeV/c. The characteristics of the observed final states are discussed and found to be consistent with the formation of excited states involving the rearrangement of multiple nucleons in a single, highly-energetic projectile- target collision. Unlike the far-less dissipative, surface-grazing reactions usually exploited for the in-beam $\gamma$-ray spectroscopy of rare isotopes, these more energetic collisions appear to offer a practical pathway to nuclear-structure studies of more complex multi-particle configurations in rare isotopes – final states conventionally thought to be out of reach with high-luminosity fast-beam-induced reactions. ###### pacs: Beyond the proof of existence of a rare isotope and the determination of its ground-state half-life, the energies of excited states are typically the first observables that become accessible in laboratory experiments. For excited bound states, depending on their lifetime, prompt or delayed $\gamma$-ray spectroscopy is frequently used to obtain precise excitation energies from the measured transition energies [1]. In short-lived rare isotopes, excited states can be populated efficiently in (direct) nuclear reactions [2] or $\beta$ decay [3], for example, most often exploiting the unique selectivity inherent to each of these different population pathways. The selectivity of one- and two-nucleon transfer and knockout reactions, or inelastic scattering [2, 4, 5, 6, 7, 8], often enhances the population of excited states at moderate spin associated with the single-particle or collective degree of freedom. Here, we report the novel, complementary in-beam $\gamma$-ray spectroscopy of higher- spin, negative-parity states in 38Ca, observed to be populated in the 9Be($\hbox{${}^{38}$Ca},\hbox{${}^{38}$Ca}+\gamma$)X inelastic scattering at high momentum loss. From the peculiar final states observed, we argue that these complex-structure, projectile excited states are formed by the rearrangement of multiple nucleons in a single, highly-energetic projectile- target collision, giving access to multi-particle configurations not expected to be in reach of high-luminosity fast-beam reactions. The reaction channel analyzed here is populated in the same experiment as reported on in Ref. [9] where the focus was on 40Sc produced in the $pn$ pickup reaction onto the 38Ca projectile. Here, we briefly summarize the experimental scheme below and refer the reader to Ref. [9, 16] for more details. The 38Ca rare-isotope beam was produced by fragmentation of a stable 40Ca beam, accelerated to 140 MeV/nucleon by the Coupled Cyclotron Facility at NSCL [10]. The momentum width transported to the experiment was restricted to $\Delta p/p=0.25$%, resulting in 160,000 38Ca/s impinging upon a 188-mg/cm2-thick 9Be foil located at the target position of the S800 spectrograph [11]. The setting subject of this publication ran for less than 40 hours. The constituents of the incoming beam and the projectile-like reaction products were identified on an event-by-event basis using the S800 analysis beam line and focal plane with the standard detector systems [12]. As the magnetic rigidity of the S800 spectrograph was tuned for 36Ca, only part of the outermost (exponential) low-momentum tail of the reacted 38Ca distribution was transmitted to the focal plane. Specifically, the S800 momentum acceptance at this setting is $p_{0}\pm 330$ MeV/c, with $p_{0}=11.222$ GeV/c. When compared to the parallel momentum distribution of the unreacted 38Ca passing through the target, having suffered only in-target energy losses ($p_{0}=11.932$ GeV/c), the low-momentum, reacted 38Ca events detected in the reaction setting have undergone an additional longitudinal momentum loss of about 700 MeV/c (see Fig. 1). That is, approximately 18 MeV/c per nucleon in momentum or 5.4 MeV/nucleon in energy. The cross section for finding 38Ca with such large momentum loss was extracted to be $\sigma(p_{0}\pm 330~{}\mathrm{MeV/c})=3.8(4)$ mb, making these inelastic large-momentum-loss events rather rare. Figure 1: Longitudinal momentum distributions of 38Ca passing through the target and only suffering energy loss (magenta peak) and, on log scale, for the dissipative setting (inset (a)). Insets (b) and (c) confront the $\gamma$-ray spectra in coincidence with less than 100,500 38Ca at high momentum loss (black) and from nearly 179,000 38Ca in the direct setting (magenta), highlighting a stark difference in excitation probability. The mid-target energy of 38Ca in the 9Be reaction target was 60.9 MeV/nucleon. The target was surrounded by GRETINA [13, 14], an array of 48 36-fold segmented high-purity germanium crystals assembled into modules of four crystals each, used for prompt $\gamma$-ray detection to tag the final states of the reaction residues. Signal decomposition was employed to provide the $\gamma$-ray interaction points. Of these, the location of the interaction with the largest energy deposition was selected as the first hit entering the event-by-event Doppler reconstruction of the $\gamma$ rays emitted from the reaction residues in-flight at about 33% of the speed of light [14]. The event-by-event Doppler reconstructed $\gamma$-ray spectrum obtained in coincidence with the 38Ca reaction residues detected in the S800 focal plane at large momentum loss is shown in Fig. 2. Nearest neighbor addback, as detailed in [14], was used. Of the seven $\gamma$-ray transitions compiled in [15], those at 2213(5), 1489(5), 489(4) and 3684(8) keV are observed here, while the transitions at 214(4), 1048(6), 2417(7), 2537(6), 2688(7), and 2758(7) keV are reported for the first time in the present work. This letter discusses the strongly-populated states. The reader is referred to the companion paper for details on some other weakly-populated states [16]. Figure 2: Doppler-reconstructed addback $\gamma$-ray spectrum as detected in coincidence with the scattered 38Ca nuclei that underwent a large momentum loss. All $\gamma$-ray transitions are labeled by their energy. The inset magnifies the high-energy region of the spectrum. To construct the level scheme, $\gamma\gamma$ coincidences are used. Figure 3 shows the coincidence analysis of the low-energy part of the 38Ca spectrum. From Fig. 3, it is clear that the 1489-keV $\gamma$ ray feeds the 2213-keV line, the 489-keV transition feeds the level depopulated by the 1489 keV, and the 214-keV transition lies on top of the level depopulated by the 489-keV transition. There is evidence for a weak 1048-keV transition being in coincidence with the 2213 and 1489 keV $\gamma$ rays. Figure 3: Left: Doppler-corrected $\gamma\gamma$ coincidence spectra obtained from cuts on the labeled prominent transitions in the $\gamma\gamma$ coincidence matrix. Background was subtracted via a cut of equal width at slightly higher energy. Coincidence relationships are evident in the panels. Right: Resulting level scheme. The width of the arrows is proportional to the $\gamma$-ray intensity of the corresponding transition. The proton separation energy of $S_{p}=4.54727(22)$ MeV [19] places the second $3^{-}$ state above the proton separation energy. The $0^{+}_{2}$ state is shown but was not populated in the present work. Figure 3 also shows the partial level scheme with the intensities of the $\gamma$-ray transitions indicated by the arrow widths. These relative $\gamma$-ray intensities were deduced from the efficiency-corrected peak areas from the spectrum displayed in Fig. 2. Remarkably, the fourth strongest $\gamma$ ray, at 214 keV, has not been reported previously. The relative intensities and a more complete level scheme, including all $\gamma$-ray transitions observed, is provided in Ref. [16]. The 1489-keV transition in coincidence with the $2^{+}_{1}\rightarrow 0^{+}_{1}$ decay is consistent with the previously reported ($3^{-}$) state at 3702 keV. The 489-keV transition in coincidence with the 3702-keV ($3^{-}$) state suggests a level at 4191 keV, which is consistent with a previously reported state at 4194 keV. However, the $J^{\pi}$ assignment proposed in the literature of ($5^{-})$ [15] is unlikely as the 489-keV transition in our work is prompt, on the level of a few ps or faster as evident from the good resolution and absence of a low-energy tail, which – if of $E2$ character – would indicate a $B(E2;5^{-}\rightarrow 3^{-})$ strength exceeding the recommended upper limit of 100 W.u. [17]. From comparison with the mirror nucleus, 38Ar, which has a 4480-keV $4^{-}$ level with a sole transition of 670 keV connecting to the first $3^{-}$ state, resembling the situation described here, we propose $J^{\pi}=(4^{-})$ for the 4191-keV state in 38Ca. The new 214-keV transition feeding the ($4^{-}$) level establishes a state at 4405 keV which appears to correspond to the 4586-keV $5^{-}$ level in the 38Ar mirror whose far-dominant decay is a 106-keV transition to the $4^{-}$. Based on mirror symmetry, a ($5^{-}$) assignment is proposed here for the 4405-keV level in 38Ca. This establishes $(5^{-})\rightarrow(4^{-})\rightarrow(3^{-}_{1})\rightarrow 2^{+}_{1}$ as the most intense cascade seen following the 38Ca inelastic scattering populated at large momentum loss. The next strongest populated level is the $2^{+}_{2}$ state at 3684 keV for which only the transition to the ground state is observed here. A $0^{+}$ state at 4748(5) keV is claimed in 38Ca from the $(\hbox{${}^{3}$He},n)$ transfer reaction, however, with the suspicion of a doublet [15]. Due to the transition to the $(3^{-}$) state, a $0^{+}$ assignment is excluded and the level established here is tentatively assigned ($3^{-}_{2}$), consistent with the 4877-keV $3^{-}_{2}$ level in the 38Ar mirror, which also decays predominantly to the $2^{+}_{1}$ and $3^{-}_{1}$ states [15]. It is interesting to explore which low-lying levels have not been observed in the present experiment. This is, most prominently, the $0^{+}_{2}$ state reported at 3084 keV which would decay to the first $2^{+}$ state with a 871-keV transition [15]. There is no evidence for an appreciable presence of that transition in Figs. 2 and 3 (the 871-keV transition would be 13 keV above the background feature originating from neutron-induced background as indicated in Fig. 2). In the following, we discuss the configurations of the states observed. Many properties of 40Ca and the surrounding nuclei can be interpreted relative to a doubly-closed shell structure for the ground state of 40Ca with the $sd$ shell filled and the $pf$ shell empty. The first excited state of 40Ca has $J^{\pi}=0^{+}$ and is qualitatively associated with a four-particle four-hole (4p-4h) state relative to the 40Ca closed-shell ground state [18]. We will use $\Delta$, the number nucleons moved from $sd$ to $fp$ orbitals, to characterize the structure of the states. In this notation, the 4p-4h states in 40Ca have $\Delta=4$. (To remove spurious states, the $\Delta$ basis includes all components associated with the $\Delta\hbar\omega$ basis constructed in the $0s$-$0p$-$0d1s$-$0f1p$ model space). In Ref. [20], a Hamiltonian was developed for these pure $\Delta$ configurations. This Hamiltonian served as the starting point for the new Florida State University (FSU) Hamiltonian for pure $\Delta$ states [21, 22]. The $A=38$, FSU results are compared to experiment in Fig. 4, the overall agreement with experiment being good. The calculated configurations can be divided into those with $\Delta=0$ with positive parity (green), those with $\Delta=1$ with negative parity (blue) and those with $\Delta=2$ with positive parity (red). Figure 4: Comparison of the energies of the low-lying states of 38Ca, with the states observed here labeled, with shell-model calculations using the FSU $spsdfp$ interaction, and states in 38Ar [15]. In these plots, the length of the levels indicates the $J$ value and the color positive parity, $\Delta=2$ (red), negative parity, $\Delta=1$ (blue), and $sd$-shell origin, $\Delta=0$ (green). In the present 38Ca level scheme, the strongest $\gamma$ rays come from the $2^{+}_{1}$ state, which is predicted to be of $sd$-shell origin, and from states with $\Delta$=1, including the highest $J^{\pi}=5^{-}$ level possible for this $\Delta$. The $\gamma$-ray decay of the 2${}^{+}_{2}$ state is also observed. In the 36Ar$(\hbox{${}^{3}$He},n)$ reaction in [23] this state is found to have a strong $(f_{7/2})^{2}$ form factor which would come from $\Delta=2$ configurations in the FSU spectrum. However, the 0${}^{+}_{2}$ state, which also has $\Delta=2$, was not populated. In the following, we propose a view that puts the populated states within the context of the observed high-momentum-loss reaction events. From the approximately 200 MeV of energy loss in the reaction, and given that the detected 38Ca are largely within laboratory scattering angles of 3-4∘, about 150 MeV must be dissipated in the 9Be nuclei, with a total binding of 58 MeV. Thus, there must be disintegration of the target nucleus into a number of energetic fragments. The emerging picture is then one of multiple nucleons interacting in a single collision with the formation of complex multi-particle multi-hole configurations, in contrast to the situation in far-less- dissipative, surface-grazing collisions. We exclude scenarios where a 38Ca projectile undergoes multiple collisions within the target as an explanation for the observed cross sections. High-momentum loss events creating $mp$-$nh$ excitations in such a scenario would require a sequence of knockout and/or pickup processes and such pickup mechanism cross sections are small – with a typical upper limit of 2 mb at these beam energies [24]. Connecting to the shell-model picture, excitations within the FSU model space are described by many-body transition densities. In the simplest scenario, excitation of the $\Delta=1$ negative-parity states involve the $\Delta=0$ to $\Delta=1$ one-body transition densities (OBTD). The OBTD to those states observed are all large. The $\Delta=2$, 2${}^{+}_{2}$ state involves the $\Delta=0$ to $\Delta=2$ two-body transition density (TBTD). The TBTD connecting the $\Delta=0$ and $\Delta=2$ 0+ wave functions are the same ones that enter into the Hamiltonian matrix for mixing these two states. We expect that the microscopic, two-nucleon excitation mechanism should involve an operator similar to that of the two-body mixing Hamiltonian (e.g. dominated by pairing). This would explain why excitation of the 0${}^{+}_{2}$ is not observed – the mixed 0${}^{+}_{1}$ and 0${}^{+}_{2}$ eigenfunctions are orthogonal with respect to the two-nucleon excitation operator. We note that in 40Ca($p,t$) [25] the $0^{+}_{2}$ state is only very weakly populated compared to the $2^{+}_{2}$ state (see Fig. 1 in Ref. [25]). The events at momentum losses of 600-700 MeV/c, studied here, are also reminiscent of observations in the work of Podolyak et al. [30]. There, in the two-neutron knockout from 56Fe to 54Fe at 500 MeV/nucleon, the population of a 10+ isomer of complex structure was observed in the low-momentum tail of the parallel momentum distribution at about the same absolute momentum loss. The authors attributed this population to the excitation of the $\Delta$(1232) resonance at their relativistic beam energies. This mechanism is not available to our intermediate-energy beams of tens of MeV/nucleon. One may speculate that the population of the complex-structure state in the two-neutron knockout from 56Fe is rather due to a simultaneous multi-nucleon rearrangement as hypothesized here, without evoking quark degrees of freedom and consistent with the reduction of multi-step processes at their relativistic energies. For example, population of the $10^{+}$ state could be due to the $\Delta J=6$ excitation of a $(f_{7/2})^{2}$ $6^{+}$ configuration in 56Fe combined with the removal of two neutrons from the $1p_{3/2}$ and $0f_{7/2}$ orbitals having $\Delta J\geq 4$. In Ref. [16], from the high-spin spectroscopy of states up to $J=15/2$ in 39Ca, we argue that such simultaneous multi-nucleon rearrangement is also at play in intermediate-energy nucleon transfer reactions, such as 9Be($\hbox{${}^{38}$Ca}^{*},\hbox{${}^{39}$Ca}+\gamma$)X. Once again, these excitations are seen in events in the tail of the longitudinal momentum distribution at a momentum loss of 600-700 MeV/c. In the present work, the specific reaction dynamics at play in the observed large momentum loss collisions are unclear and remain a challenge for future, more complete and exclusive measurements. Specifically, it would be critical to detect the dissociation of the 9Be target nuclei in the large-momentum-loss events and clarify the kinematics of the residues. While there is much to be discovered about this type of reaction, it is evident that this presents a new opportunity in the fast-beam regime which uniquely complements classic low-energy reactions, such as multi-step Coulomb excitation and multi-nucleon transfer. Fast beams allow for the use thick targets and capitalize on an increase in $\gamma$-ray yield by a factor of about 4300 for the specific example of a 188-mg/cm2 9Be target used here vs. a 1-mg/cm2 Pb target often employed for multi-step Coulomb excitation, for example. Also, strong forward focusing enhances the collection efficiency as compared to low-energy reactions that fill a larger phase space. Multi-step Coulomb excitation studies with low-energy rare-isotope beams have been performed at beam intensities similar to those used here, but have been limited to a complementary level scheme selectively comprising cascades connected by strong $E2$ transitions, with at most the first $3^{-}$ state [26, 7]. We illustrate this with the example of the state-of-the art low- energy Coulomb excitation of the neighboring Ca isotope 42Ca on Pb [27]. The measurement was performed at 1 pnA stable-beam intensity for 5 days (resulting in more than 110,000 times the number of Ca projectiles on target as in the present measurement) – excited states up to the $4^{+}_{1,2}$ states were reported with no evidence for any of the negative-parity cross-shell excitations observed here. Multi-nucleon transfer, largely limited to stable beams at pnA beam intensities, is known to populate complex-structure states, however, without efficiently reaching 38Ca in spite of 40Ca being an often-used beam (see [28] and references within). When low-energy neutron-rich beams become available at near stable-beam intensities, multi-nucleon transfer may become an alternative to access such states in selected neutron-rich nuclei [29]. While it is interesting to also extend our approach to collective nuclei, it already promises to be a unique method to probe cross-shell excitations near magic numbers, elucidating shell evolution in rare isotopes and exploring the necessary model spaces for a region’s description on the quest for a predictive model of nuclei. In conclusion, the in-beam $\gamma$-ray spectroscopy is reported of higher- spin, complex-structure negative-parity states in 38Ca populated in highly- dissipative processes induced by a fast 38Ca projectile beam reacting with a 9Be target. This work constitutes the first high-resolution $\gamma$-ray spectroscopy of 38Ca with a modern HPGe $\gamma$-ray tracking array. The final states observed in the inelastic scattering, 9Be($\hbox{${}^{38}$Ca},\hbox{${}^{38}$Ca}+\gamma$)X, at large momentum loss are characterized through their particle-hole character relative to the 40Ca closed-shell ground state. Excellent agreement is obtained with shell-model calculations using the FSU cross-shell effective interaction. Based on the strongly populated negative-parity states and the non-observation of the first excited $0^{+}_{2}$ state, we propose a consistent picture in which these multi-particle multi-hole states are formed by simultaneous rearrangement of multiple nucleons in a single, highly-dissipative collision. These reaction processes, seen here in the extreme low-momentum tail of 38Ca+9Be inelastic scattering, identify a new pathway to gain access to excited states not usually observed in fast-beam induced reactions and likely out of reach for low-energy reactions. This work was supported by the U.S. National Science Foundation (NSF) under Grants No. PHY-1565546 and PHY-2110365, by the DOE National Nuclear Security Administration through the Nuclear Science and Security Consortium, under Award No. DE-NA0003180, and by the U.S. Department of Energy, Office of Science, Office of Nuclear Physics, under Grants No. DE-SC0020451 (MSU) and DE-FG02-87ER-40316 (WashU) and under Contract No. DE-AC02-06CH11357 (ANL). GRETINA was funded by the DOE, Office of Science. Operation of the array at NSCL was supported by the DOE under Grant No. DE-SC0019034. J.A.T. acknowledges support from the Science and Technology Facilities Council (U.K.) Grant No. ST/V001108/1. ## References * [1] A. Gade, Eur. Phys. J. A 51, 118 (2015). * [2] A. Obertelli, Eur. Phys. J. Plus 131, 319 (2016). * [3] B. Rubio, W. Gelletly, in The Euroschool Lectures on Physics with Exotic Beams, Vol. III, edited by J. Al-Khalili, E. Roeckl, (Springer, Berlin, Heidelberg, 2009), in Lecture Notes in Physics, Vol. 764, 99 (2009). * [4] T. Glasmacher, Annu. Rev. Nucl. Part. Sci. 48, 1 (1998). * [5] P. G. Hansen and J. A. Tostevin, Annu. Rev. Nucl. Part. Sci. 53, 219 (2003). * [6] A. Gade and T. Glasmacher, Prog. in Part. and Nucl. Phys. 60, 161 (2008). * [7] A. Görgen and W. Korten, J. Phys. G Nucl. Part. Phys. 43, 024002 (2010). * [8] K. Wimmer, J. Phys. G: Nucl. Part. Phys. 45 033002 (2018). * [9] A. Gade, D. Weisshaar, B. A. Brown, J. A. Tostevin, D. Bazin, K. Brown, R. J. Charity, P. J. Farris, A. M. Hill, J. Li, B. Longfellow, W. Reviol, D. Rhodes, Phys. Lett. B 808, 135637 (2020). * [10] A. Gade and B. M. Sherrill, Phys. Scr. 91, 053003 (2016). * [11] D. Bazin, J. A. Caggiano, B. M. Sherrill, J. Yurkon, and A. Zeller, Nucl. Instrum. Methods in Phys. Res. B 204, 629 (2003). * [12] J. Yurkon, D. Bazin, W. Benenson, D. J. Morrissey, B. M. Sherrill, D. Swan, R. Swanson, Nucl. Instrum. Methods in Phys. Res. A 422, 291 (1999). * [13] S. Paschalis et al., Nucl. Instrum. Methods Phys. Res., Sect. A 709, 44 (2013). * [14] D. Weisshaar, D. Bazin, P. C. Bender, C. M. Campbell, F. Recchia, V. Bader, T. Baugher, J. Belarge, M. P. Carpenter, H. L. Crawford, M. Cromaz, B. Elman, P. Fallon, A. Forney, A. Gade, J. Harker, N. Kobayashi, C. Langer, T. Lauritsen, I. Y. Lee, A. Lemasson, B. Longfellow, E. Lunderberg, A. O. Macchiavelli, K. Miki, S. Momiyama, S. Noji, D. C. Radford, M. Scott, J. Sethi, S. R. Stroberg, C. Sullivan, R. Titus, A. Wiens, S. Williams, K. Wimmer, and S. Zhu, Nuclear Instrum. Methods Phys. Res. Sect. A 847, 187 (2017). * [15] J. Chen, Nuclear Data Sheets 152, 1 (2018). * [16] A. Gade et al., Phys. Rev. C (submitted). * [17] P. M. Endt, At. Data. Nucl. Data Tables 55, 171 (1993). * [18] W. J. Gerace and A. M. Green, Nucl. Phys. A93, 110 (1967); A123, 241 (1969). * [19] M. Wang, W. J. Huang, F. G. Kondev, G. Audi, and S. Naimi, Chin. Phys. C 45, 030003 (2021). * [20] E. K. Warburton, J. A. Becker and B. A. Brown, Phys. Rev. C 41, 1147 (1990). * [21] R. S. Lubna, K. Kravvaris, S. L. Tabor, Vandana Tripathi, A. Volya, E. Rubino, J. M. Allmond, B. Abromeit, L. T. Baby, and T. C. Hensley, Phys. Rev. C 100, 034308 (2019). * [22] R. S. Lubna, K. Kravvaris, S. L. Tabor, Vandana Tripathi, E. Rubino, and A. Volya, Phys. Rev. Research 2, 043342 (2020). * [23] W. P. Alford, P. Craig, D. A. Lind, R. S. Raymond, J. Ullman, C. D. Zafiratos, B. H. Wildenthal, Nucl. Phys. A 457, 317 (1986). * [24] A. Gade, J. A. Tostevin, V. Bader, T. Baugher, D. Bazin, J. S. Berryman, B. A. Brown, D. J. Hartley, E. Lunderberg, F. Recchia, S. R. Stroberg, Y. Utsuno, D. Weisshaar, and K. Wimmer, Phys. Rev. C 93, 031601(R) (2016). * [25] S. Kubono, S. Kato, M. Yasue, H. Ohnuma, M. Sasao, K. Tsukamoto, R. Kuramasu, Nucl. Phys. A 276, 201 (1977). * [26] M. Rocchini and M. Zielinska, Physics 3, 1237 (2021). * [27] K. Hadynska-Klek, P. J. Napiorkowski, M. Zielinska, J. Srebrny, A. Maj et al., Phys. Rev. C 97, 024326 (2018). * [28] L. Corradi, S. Szilner, G. Pollarolo, D. Montanari, E. Fioretto, A. M. Stefanini, J. J. Valiente-Dobon, E. Farnea, C. Michelagnoli, G. Montagnoli, F. Scarlassara, C. A. Ur, T. Mijatovic, D. Jelavic Malenica, N. Soic F. Haas, Nucl. Instrum. Methods in Phys. Res. B 317, 743 (2013). * [29] P. Colovic, S. Szilner, A. Illana, J. J. Valiente-Dobon, L. Corradi et al., Phys. Rev. C 102, 054609 (2020). * [30] Zs. Podolyak, C. M. Shand, N. Lalovic, J. Gerl, D. Rudolph, T. Alexander, P. Boutachkov, M. L. Cortes, M. Gorska, I. Kojouharov, N. Kurz, C. Louchart, E. Merchan, C. Michelagnoli, R. M. Perez-Vidal, S. Pietri, D. Ralet, M. Reese, H. Schaffner, Ch. Stahl, H. Weick, F. Ameil, G. de Angelis, T. Arici, R. Carroll, Zs. Dombradi, A. Gadea, P. Golubev, M. Lettmann, C. Lizarazo, D. Mahboub, H. Pai, Z. Patel, N. Pietralla, P. H. Regan, L. G. Sarmiento, O. Wieland, E. Wilson, B. Birkenbach, B. Bruyneel, I. Burrows, L. Charles, E. Clement, F. C. L. Crespi, D. M. Cullen, P. Desesquelles, J. Eberth, V. Gonzalez, T. Habermann, L. Harkness-Brennan, H. Hess, D. S. Judson, A. Jungclaus, W. Korten, M. Labiche, A. Maj, D. Mengoni, D. R. Napoli, A. Pullia, B. Quintana, G. Rainovski, P. Reiter, M. D. Salsac, E. Sanchis, and J. J. Valiente Dobon, Phys. Rev. Lett. 117, 222302 (2016).
# Modelling of Spintronic Terahertz Emitters as a function of spin generation and diffusion geometry Yingshu Yang School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore, Singapore Stefano Dal Forno School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore, Singapore Marco Battiato<EMAIL_ADDRESS>School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore, Singapore ###### Abstract Spintronic THz emitters (STE) are efficient THz sources constructed using thin heavy-metal (HM) and ferromagnetic-metal (FM) layers. To improve the performance of the STE, different structuring methods (trilayers, stacked bilayers) have been experimentally applied. A theoretical description of the overall THz emission process is necessary to optimize the efficiency of STE. In particular, geometry, composition, pump laser frequency, and spin diffusion will be significant in understanding the pathways for further research developments. This work will apply a generalized model based on a modified Transfer Matrix Method (TMM). We will consider the spin generation and diffusion in the FM and HM layers and explain the spintronic THz emission process. This model is suitable for calculating emitted THz signal as a function of FM and HM thicknesses for different geometrical configurations. We will investigate a bilayer geometry as a test case, but the extension to a multi-layer configuration is straightforward. We will show how the different configurations of the sample will influence the THz emission amplitude. ††preprint: AIP/123-QED THz (0.1 -30 THz) is a frequency range that is gaining popularity because of its enormous potential in basic process studies of materials as it is able to resonantly couple to conduction-electron transport, plasmons, excitons, phonons, or magnons Kampfrath, Tanaka, and Nelson (2013); Seifert _et al._ (2022). It is also an effective tool in imaging, sensing for biomedical purposes, and security applications Seifert _et al._ (2022); Feng _et al._ (2021); Agarwal _et al._ (2022a). THz radiation can be generated using various methods, including photoconductive antennas and nonlinear optical crystals Seifert _et al._ (2022). Spintronics THz emitters (STE), a new type of THz source, based on spin-to-charge conversion, have gained popularity in recent years due to their high efficiency, broad bandwidth, and ease of manufacture. STE is usually constructed using two simple layers of ferromagnetic-metal(FM), and heavy-metal(HM)Kampfrath _et al._ (2013); Seifert _et al._ (2016); Wu _et al._ (2017). The mechanism at the core of THz emission first involves the photoexcitation of the FM layer. This induces a spin population that diffuses to the HM layer in a process called superdiffusive spin transport Battiato, Carva, and Oppeneer (2010, 2012); Battiato, Maldonado, and Oppeneer (2014); Battiato and Held (2016); Battiato (2017); Carva, Battiato, and Oppeneer (2011); Carva _et al._ (2013). Here, due to the high spin-orbit-coupling (SOC) of the HM, inverse Spin Hall effect (ISHE) Seifert _et al._ (2016) will take place and produce a transverse charge current responsible for the THz emission (as shown in Fig. 1(a)). Figure 1: (a) Schematic of a spintronic THz emitter and the THz emission process. (b) and (c) Schematics of a bilayer and trilayer systems with the spin diffusion profile in the FM layer (purple line), the spin current diffusion and reflection in the HM layer (blue line), and the energy distribution profile of the laser pump along the system (red line and shaded area). The study of STE optimization has recently received much attention in order to broaden their applications. As a result, there have been different experimental studies carried out to test the THz generation efficiency based on different materials of the structureSeifert _et al._ (2016, 2018); Qiu _et al._ (2018); Liu _et al._ (2022), different thicknesses of the layersZhou _et al._ (2018); Torosyan _et al._ (2018); Qiu _et al._ (2018); Wu _et al._ (2017), and even different stacks of layersFeng _et al._ (2018); Cheng _et al._ (2021); Seifert _et al._ (2016); Cheng _et al._ (2019); Herapath _et al._ (2019). To analyze the performance of the STE based on the above approaches, a theoretical model to describe the STE geometry and THz emission process is needed. Models based on basic THz emission processes involving the pump-pulse absorption, spin generation, and spin to charge conversion have been developed and used in different analysis Seifert _et al._ (2016); Torosyan _et al._ (2018). These models usually calculate the THz emission amplitude assuming the spin current amplitude to be proportional to the excitation energy deposition Seifert _et al._ (2022) and at the same time assuming the FM/HM interface is transparent to spins and electrons Seifert _et al._ (2022, 2016). However, no general model has considered the detailed geometry of the excitation laser and spin current with specific STE structures in the THz emission process. In a typical experimental setup, STEs are constructed by combining substrate, FM, and HM layers in a number of different arrangements. The ordering of the layers, their thicknesses, and the side illuminated by the laser pump will impact the geometry of the spin generation, the spin diffusion, and the energy profile of the laser pump. The model we applied here addresses all of the above properties on the same footing. Specifically, we want to describe the three fundamental processes that characterize any STE: the propagation of the laser excitation pump through the system, the generation and diffusion of spins in the HM and FM layers, and the generation and propagation of the THz pulse throughout the system. These processes are controlled by a number of parameters, namely, the type of materials, their thicknesses, the arrangement of the layers, the frequency of the pump, the spin diffusion length, and the degree of spin reflections between interfaces. Fig.1(b) and (c) show schematics of the processes mentioned above in both bilayer and trilayer systems. The dependence of the outgoing THz pulse strongly depends on the choice of these parameters. Our findings show that properly fitting the material properties in optical and THz regions is necessary to achieve quantitative predictions. ## I Methods To properly model STEs, we have to describe the three sub-processes that occur in the THz emission: (1) Propagation and absorption of the optical pump into the STE’s layers; (2) Production of the spin current, its propagation through the STE’s layers and the eventual conversion into a transversal charge current; and (3) Production, propagation and extraction from the STE of the THz electromagnetic radiation. The following sections will address each of these tasks individually. ### I.1 Pump laser absorption profile We describe the optical EM waves propagation and absorption using a Transfer Matrix Method (TMM) Coutaz, Garet, and Wallace (2018); Born and Wolf (2013).The system is composed by three layers: a Substrate, a FM layer and a HM layer, in different stacking orders. According to the standard Transfer Matrix Method, the transmission and reflection of the waves throughout an layer sample can be expressed as, $\begin{bmatrix}f_{\infty}^{>}\\\ f_{\infty}^{<}\end{bmatrix}=\bar{\bar{T}}_{[0,\infty]}\begin{bmatrix}f_{0}^{>}\\\ f_{0}^{<}\end{bmatrix},$ (1) where $\bar{\bar{T}}_{[0,\infty]}$ is the frequency-dependent $2\times 2$ transfer matrix that propagates the fields from the beginning to the end of the multilayer, and $f_{0}$, $f_{\infty}$ represent the field amplitudes at the beginning and the end of the multilayer. The superscripts > and < represent the right and left propagating waves respectively. Assuming the pump pulse is impinging on the sample from the left, $f_{0}^{>}$ is the time profile of the pump (which we assume known). Generally, there will be no second pump incoming from the right, so $f_{\infty}^{<}=0$. The two remaining field amplitudes $f_{\infty}^{>}$ and $f_{0}^{<}$, which represent the transmitted and reflected waves respectively, are the unknowns in the system in Eq. 1. We remind that the system of two equations in Eq. 1 is to be solved for every frequency independently. The generation of the spin currents and the subsequent diffusion of the spins in the FM and HM layers strongly depends on how the energy deposited by the laser pump is partitioned through the system Zhou _et al._ (2018); Torosyan _et al._ (2018). According to the Poynting theorem, the total energy loss due to Joule effects can be expressed as, $Q_{loss}=-\int_{-\infty}^{+\infty}\mathrm{d}t\oint_{S}(\mathbf{E}\times\mathbf{H})\cdot\mathrm{d}\bm{S}.$ (2) Here, $Q_{loss}$ is the total energy dissipated by a system enclosed by the closed surface $S$. Because of the planar symmetry of STE, the surface $S$ can be chosen to be a parallelepiped enclosing the layer. At normal incidence, only faces of $S$ parallel to the interfaces between layers contribute to the integral in Eq. 2. Hence, we can define the energy per unit area that crossed a surface $S_{z}$ at position $z$ as, $\Phi(z)=-\int_{-\infty}^{+\infty}E[t,z]H[t,z]\mathrm{d}t,$ (3) where $z$ is the position of the surface $S_{z}$. Thus, if we assume the two interface position of a layer with thickness $d$ in a multilayer system as $z_{0}$ and $z_{0}+d$, and assuming the initial input of the pump laser as $Q_{in}$, the net absorption for this layer will be: $A_{layer}=\frac{Q_{loss}}{Q_{in}}=\frac{\Phi(z_{0})-\Phi(z_{0}+d)}{Q_{in}}.$ (4) where $Q_{in}$ can also be calculated using Eq. 3 with fields at the initial surface of the system by ignoring the reflected fields ($f_{0}^{<}$ in Eq. 1). This net absorption can be used when simulating STEs with fixed FM thicknesses. Similar ideas can be used to obtain the energy distribution profile. If we assume a local axis within one single layer, the energy distribution profile can be expressed as, $D(z)=-\frac{d(\Phi(z)/Q_{in})}{dz}$ (5) The energy distribution profile of Eq.5 is schematically displayed at the bottom of Figs. 1(b) and (c) with a red solid line. ### I.2 Spin generation in the FM layer We now address how the absorbed energy in the FM layer produces a spin current profile over the whole sample. We focus first on how energy deposited between $z$ and $z+dz$ within the FM layer propagates through the whole sample. We assume that the spin current spatial distribution generated by an infinitesimally thin layer and in the absence of interfaces is in the form of an exponential decay, with an effective spin diffusion length $\lambda$, both towards the left and the right of the emission plane (See Fig. 1). However due to the finite thickness of the FM layer, as well as the presence of the other layers, we must explicitly consider the reflections and the transmissions of the spins at the layers’ boundaries. We call $\alpha$ and $\beta$ the probabilities for a spin to be transmitted over interfaces to the right and the left of the considered spin current emission point $z$. Such probabilities depend on the materials and the quality of the interfaces. We assume that the probability of transmission at the interface with air or with the substrate (assumed insulating, with sufficiently large bandgap to prevent injection of the excited spin currents) to be 0. The rate at which the spin population is transmitted to the left or to the right can be calculated by exploiting the properties of geometric series (see Suppl. Info. for the detailed derivation). By summing up the contributions of all the multiple reflections at the boundaries, we obtain the following expressions $\displaystyle\zeta_{R}(z)$ $\displaystyle=\frac{\alpha}{N}\left[e^{-\frac{d-z}{\lambda}}+\bar{\beta}e^{-\frac{d+z}{\lambda}}\right],$ (6) $\displaystyle\zeta_{L}(z)$ $\displaystyle=\frac{\beta}{N}\left[e^{-\frac{z}{\lambda}}+\bar{\alpha}e^{-\frac{2d-z}{\lambda}}\right],$ (7) where $N={1-\bar{\alpha}\bar{\beta}e^{-\frac{2d}{\lambda}}}$, $\bar{\alpha}=1-\alpha$, $\bar{\beta}=1-\beta$. Eqs. 6 and 7 measure the number of spins (generated at position $z$) that have crossed (transmitted) the right or left interfaces of the FM layer, respectively. We call $\zeta(z)$ the spin generation efficiency profile function (purple solid lines in Fig. 1). Finally, to compute the effective number of spins that is injected from FM to HM, we must consider the effects of the laser pump. We assume that the spin density is proportional to the laser fluence at the point $z$ times the left or right spin generation efficiency profile. The total spin population becomes $S(d)=\int_{z_{i}}^{z_{i}+d}D(z)\zeta(z)dz,$ (8) where $D(z)$ is the energy (per unit area) deposited by the laser pump (Eq. 5), $z_{i}$ is the position of the FM layer and $\zeta(z)$ must be chosen either to be right or left, depending on the geometrical configuration of the system. For example, if we consider the bilayer THz emitter shown in Fig.1(b), we need to calculate only the transmittance to the right interface. However, if we want to model the trilayer case shown in Fig.1(c), then the transmittance for both the left and right interfaces have to be considered. ### I.3 Spin current profile in the HM layer After the laser pump has excited the spin population in the FM layer, spins start diffusing and are eventually injected at the edge of the HM layer. We again assume an exponentially decaying spatial distribution with an effective spin diffusion length $\lambda$. We call $\gamma$ and $\mu$ the probabilities for a spin to be transmitted over the right and the left interfaces, respectively. We want to calculate the average number of spins at a given position $z$ inside the HM layer. Again, by exploiting the properties of geometric series (see Supp. Info.)Agarwal _et al._ (2022b), we can sum up the contributions of the multiple reflections and obtain the following $\sigma(z,d)=\frac{1}{N}\left[e^{-\frac{z}{\lambda}}+\bar{\gamma}e^{-\frac{2d-z}{\lambda}}\right],$ (9) where $N=1-\bar{\gamma}\bar{\mu}e^{-\frac{2d}{\lambda}}$, $d$ is the thickness of the HM layer and $\bar{\gamma}=1-\gamma$, $\bar{\mu}=1-\mu$. The above equation measures the spin current density at position $z$ for a given HM layer of thickness $d$ (see Fig. 1 light blue curves). The transversal charge current density is finally obtained by multiplying by the inverse spin Hall coefficient of the HM, similarly to our previous work. Yang, Dal Forno, and Battiato (2021a, b) Two other mechanisms of generation of transversal charge curret are also possible. The first one is caused by spin-to-charge current conversion in the FM layer. This contribution is usually much smaller than the HM layer contribution due to the low spin to charge conversion efficiency of FM layerSeifert _et al._ (2022) and will be neglected in this work. A second mechanisms involves the creation of hot electrons in the HM layer. These hot electron will diffuse back to the FM layer and act as a secondary excitation of the FM layer. This enhancement will become large when HM layer is thick Agarwal _et al._ (2022b). We stress that the above contributions, even if ignored in most cases, can be described by our model (but it is not included in this work for simplicity). The FM contribution can be included by adding a source layer to the modified TMM model developed in Sec.I.4. The secondary enhancement from the HM layer can also be added by specifically calculating the absorption of the pump-pulse in the HM layer, which can be calculated using Sec.I.1. As we are more interested in the performance of the main contribution, we show results of the HM layer emission only in the following section. We stress that the full model of including all three contributions is straightforward. However, specifying the percentage of each contribution in experiment requires detailed material data fitting (both THz frequencies and pump laser frequencies) to increase the accuracy. ### I.4 THz radiation production and propagation The final step is computing the THz radiation extracted from the STE. This requires the computation of the production of the THz within the HM layer as well as its propagation though the multilayer. In this case, standard TMM cannot be used. In the case one of the layers act as a source of electromagnetic radiation (by mean of a time and position dependent volume current) we use the modified TMM, which we call TMM-with-source, that we developed in Ref. Yang, Dal Forno, and Battiato, 2021b. The expression that we obtained maintains the structure of the TMM, but includes a source term $\begin{bmatrix}f_{\infty}^{>}\\\ f_{\infty}^{<}\end{bmatrix}=\bar{\bar{T}}_{[0,\infty]}\begin{bmatrix}f_{0}^{>}\\\ f_{0}^{<}\end{bmatrix}+\begin{bmatrix}J^{>}\\\ J^{<}\end{bmatrix}.$ (10) where $J^{>}$ and $J^{<}$ are the amplitudes of the right and left propagating fields generated in the source layer and account for time and position dependent charge currents within the multilayer. In this work, the position dependent profile used for the construction of the source term is taken as Eq. 9 and the time dependent profile is taken as a similar shape as used in Ref. Yang, Dal Forno, and Battiato, 2021b. It should be noticed that, while in the case of TMM applied to the absorption of the optical pump laser, in the case of TMM-with-source for the emitted THz radiation the constraints on the field amplitudes on the right and left of the sample are different. In this case, no external THz pulse is sent to the multilayer, and therefore $f_{\infty}^{<}$ and $f_{0}^{>}$ are to be set to zero. This means that in a THz emission process the only source is the THz generated from the HM layer and Eq. 10 can be solved for the $f_{\infty}^{>}$ and $f_{0}^{<}$ amplitudes. To be noticed that if we want to describe a trilayer system with two HM layers acting as sources (case shown in Fig.1(c)), the calculation will be straightforward with two additional terms describing fields coming from two different layers. Figure 2: The schematic of the four different configurations of STE and its corresponding emission map. (a-1)–(d-1) Different configurations of STE structure and a schematic show of the energy distribution profile (red shaded, Eq.5), overall spin generation efficiency profile (blue shaded, Eq.7), and spin current diffusion profile (green shaded, Eq.9) for each configuration. (a-2)–(d-2)The THz emission map with changing Pt(HM) (0-15nm) and Co(FM)(0-15nm) thicknesses for four different configurations.The yellow dashed line represents the peak position of the THz emission profile as a function of HM thickness, while the blue dashed line represents the peak position of the THz emission profile as a function of HM thickness. Figure 3: The linear fit of THz emission profile peak positions with changing theoretical Spin diffusion length of FM and HM under different conditions. The first row, (a)-(c), shows the peak position change of the THz emission profile as a function of HM thickness when the theoretical spin diffusion length for the HM layer changes from 1-6nm with the plasma frequency of HM layer for THz range material properties is changing from 3-7eV, and the extinction coefficient for optical range material properties is changing from 4-12. The second row, (d)-(f), shows the peak position change of the THz emission profile as a function of FM thickness when the theoretical spin diffusion length for the FM layer is changing from 1-6nm with the plasma frequency of FM layer for THz range material properties is changing from 3-7eV, and the extinction coefficient for optical range material properties is changing from 4-12. ## II Results In the following, we present the calculated THz amplitudes as a function of geometry, layer arrangements, materials, and laser pump frequency. In Fig. 2(a-1)–(d-1), we show the four different layer arrangements we considered in this work and label them C1, C2, C3, and C4, respectively. We start with a simple STE structure constructed with quartz (substrate, 1mm), Co (FM), and Pt (HM, 0–15nm). The material properties of Co and Pt at THz frequencies are taken from Ref. Ordal _et al._ , 1985, while at optical frequencies taken from Ref. Werner, Glantschnig, and Ambrosch-Draxl, 2009. The optical properties for quartz are taken from Ref. Gao, Lemarchand, and Lequime, 2012 and at the THz range are taken as the experimentally measured dielectric constants in Ref. Agarwal _et al._ , 2022b. We take the Pt and Co spin diffusion length to be 1.1 nm and 1nm, respectively,Seifert _et al._ (2016); Torosyan _et al._ (2018); Zhou _et al._ (2018); Zhang _et al._ (2013) as a test case. Figs. 2(a-2)-(d-2) show the calculated THz emission peak amplitude for different HM and FM thicknesses for the C1, C2, C3, and C4 configurations. One can notice that at parity of FM thickness, the emitted THz intensity increases at first, as more HM thickness allows for more efficient spin-to-charge conversion. However, eventually, further increases do not provide further gain in the THz intensity but become detrimental as larger metallic regions lead to absorption of the generated THz radiation within the sample itself. This reproduces the known fact that the THz intensity peaks for relatively thin layers for both changing HM thickness and FM thickness.Seifert _et al._ (2016, 2022); Cheng _et al._ (2021); Torosyan _et al._ (2018) The THz emission amplitude depends on the arrangement of the layers. We observe that exciting from the substrate side produces stronger THz emission compared to exiting from the active bilayer side (C1 and C2 $>$ C3 and C4). The reason is that in the second cases, the produces THz radiation has to traverse the quartz substrate, which absorbs in that frequency range. This reproduces experimental findings on quartz substrates Agarwal _et al._ (2022b). The situation is reversed in the case of a sapphire substrate (material properties taken from Ref. Sanjuan and Tocho, 2012 for THz range and Ref. Kelly, 1972 for optical range), where instead absorption of the pumping radiation in the substrate becomes more important. Experiments confirm this scenario.Liu _et al._ (2021) Apart from intensity, we can also extract the behaviour of the emitted THz with changing layers’ thicknesses. Although configurations C1 and C2 display similar maximum THz intensities, their behaviour with layers’ thicknesses is different. We can observe that the most relevant characteristic controlling these dependences is which one of the two active layers (HM or FM) faces the pumping laser (see Fig. 2). Each layer has three key effects in the THz production process. The FM acts as the generator of spin current, as absorber of optical photons, and absorber of THz radiation, while the HM as spin-to- charge converter and, again, as absorber of both optical and THz frequencies. The thickness dependence of all this processes depends on the material properties. Yet the relative weights of each effect in each layer depends on the relative positions. For instance the role of optical photons absorber is higher in the first layer traversed by the pump pulse. One further interesting finding is that the peak positions of the THz emission with changing FM (or HM) layer thickness are not fixed (yellow dashed line for HM peak position, blue dashed line for FM peak position) and they depend on the thickness of the adjacent layer. However, the optimal HM thickness at which the THz emission is maximal is generally used in experiments as a quick and quantitative estimation of the spin diffusion length and vice versaPapaioannou and Beigang (2021); Seifert _et al._ (2022); Zhou _et al._ (2018); Torosyan _et al._ (2018); Seifert _et al._ (2016). For that to be a meaningful estimation, the peak position should only depend on the spin diffusion length and not be affected by other characteristics of the sample. However, this contradicts our findings. To understand this better, we performed a larger set of calculations where we compare how those peak positions compare with the actual spin diffusion length when other parameters are changed. In Figs. 3(a)-(c) we show on the y axis the thickness of the HM layer for which we obtain the strongest THz emission, while on the x coordinates the spin diffusion length in the HM used in the calculations. To really claim that the peak HM thickness can be used as an estimation of the spin diffusion length, one should require the points to be over, or close enough to, the $y=x$ line (blue dashed line in Figs), or more generally have a stable functional dependence unaffected by other layer’s properties. However, while a sufficiently linear relationship can be found between the two quantities, important deviations can be observed from the desired correlation. Yet even more crucially, the relationship between the two quantities is very strongly dependent on the other parameters of the system. In Fig. 3(a) the thickness of the FM layer is shown to impact the peak position. For instance if the measured HM peak thickness were $4$nm, the extrapolated spin diffusion should be $3$nm if the FM is $10$nm thick, or twice as large for a sample with a $1$nm thick FM. One could still argue that the FM thickness is generally known and one could couple experimental results to theory to do more reliable estimations of the spin diffusion length. However Figs. 3(b) and (c) show that that will require a very careful characterisation of the sample. A change in the dielectric properties of the HM (THz plasma frequency $\omega_{p}$ and optical extinction coefficient $k$) can, in fact, strongly impact the relationship (in particular Fig. 3(c) shows the huge change in the dependence, with increasing extinction coefficient). Similarly, we compute the FM layer thickness at which one obtains the highest THz emission at parity of HM thickness. It might be tempting to use that to estimate the spin current diffusion length in the FM. However Figs. 3(d)-(f) again show that the correlation between the two quantities is far too strongly dependent on other properties of the multilayer. Figure 4: Three different tests of THz emission. (a) THz emission profile for the same Sub/HM/FM configuration with different FM material. (b) THz emission profile for the same Sub/HM/FM configuration with different HM material. (c) THz emission profile for the same Sub/Pt/Fe configuration with different excitation lasers. We now move onto analysing the dependence of the THz emission on the used materials and pump laser frequency. We choose configuration C1 as the test case and describe three material combinations. We set the HM material as Pt and Pd and the FM material as Co and Fe (material properties for pump laser frequencies taken from Ref. Werner, Glantschnig, and Ambrosch-Draxl, 2009 and for THz frequencies from Ref. Ordal _et al._ , 1985.) We set the FM thicknesses to 3nm, the spin diffusion length $\lambda=1.1$nm for the HM, and $\lambda=1$nm for the FM cases. Fig. 4(a) and (b) show the calculated THz amplitudes. Changes are direct consequence of the different optical properties of the materials. Fig. 4(c) shows the dependence of the THz pulse for two different pump laser wavelengths (400nm and 800nm). We see that, for the given thicknesses and $\lambda$, the THz emission profile changes. Specifically, the 400nm emitted THz amplitude is higher than 800nm one. This is in qualitative agreement with experiments Adam _et al._ (2019). Figure 5: The comparison of THz emission profile with changing thicknesses of HM between experimental data and theoretical calculation when spin reflection at FM/HM interface is 20% and 100%. (a) Comparison for sapphire/Co/Pt structure with data taken from Zhou _et al._ (2018). (b) Comparison for MgO/Fe/Pt structure with data taken from Torosyan _et al._ (2018). Finally, we show another interesting finding for the THz emission profile at low HM layer thicknesses. From the generalized spin diffusion model (Eq. 9) in the HM layer, we know that the spin reflections at the interfaces will play a more significant role at lower compared to higher thicknesses. To see the influence of the spin reflections, we calculated two sets of THz emission profiles as a function of HM layer thicknesses. Then, we compared them to two different sets of normalized experimental data for Sapphire(1mm)/Co(3nm)/Pt(xnm) and MgO(0.5mm)/Fe(12nm)/Pt(xnm) samples taken from Refs. Torosyan _et al._ , 2018; Zhou _et al._ , 2018. In these two sets of calculations, we took the spin reflection percentage as 20% and 100% for the FM/HM interface and compared them to the experimental data ($\bar{\mu}=20\%$ and $\bar{\mu}=100\%$ in Eq. 9). We can see that when 20% of spins are reflected (80% spins transmitted), the profile shows a much better fit to the experimental data at lower thicknesses. This can be evidence to show that one possible reason for the appearance of a positive second derivative at low thicknesses in the experiment comes from the spin reflections at the interfaces in the HM layer. Hence, it is crucial to consider the spin reflections when dealing with low HM thicknesses. ## III Conclusion We built a theoretical model that included a spin diffusion profile in HM, a spin generation efficiency profile in FM, and an excitation laser energy profile in FM with the modified TMM to describe the spintronic THz emission. The thicknesses of the layers and the substrate, the material choice, the layer arrangement, and the pump laser frequency all affect the THz emission profile. Using this model, we showed that the peak position of the THz pulse as a function of the HM thickness is dominated by the spin diffusion length. However, other contributions, such as laser absorption and layer arrangements, play a non-negligible role. We also showed that an accurate description of the spin current reflection at interfaces is important to achieve accurate predictions at low thicknesses ($d_{HM}<4nm$). This is important because the THz signal usually peaks within this thickness range . Finally, we showed that a proper experimental fitting needs to be done to extract reliable information about the STE, such as spin diffusion length for both HM and FM, spin transmission and reflection percentage at each interface, and THz emission efficiency. ## References * Kampfrath, Tanaka, and Nelson (2013) T. Kampfrath, K. Tanaka, and K. A. Nelson, “Resonant and nonresonant control over matter and light by intense terahertz transients,” Nature Photonics 7, 680–690 (2013). * Seifert _et al._ (2022) T. S. Seifert, L. Cheng, Z. Wei, T. Kampfrath, and J. Qi, “Spintronic sources of ultrashort terahertz electromagnetic pulses,” Applied Physics Letters 120, 180401 (2022). * Feng _et al._ (2021) Z. Feng, H. Qiu, D. Wang, C. Zhang, S. Sun, B. Jin, and W. Tan, “Spintronic terahertz emitter,” Journal of Applied Physics 129, 010901 (2021), publisher: American Institute of Physics. * Agarwal _et al._ (2022a) P. Agarwal, Y. Yang, J. Lourembam, R. Medwal, M. Battiato, and R. Singh, “Terahertz spintronic magnetometer (tsm),” Applied Physics Letters 120, 161104 (2022a). * Kampfrath _et al._ (2013) T. Kampfrath, M. Battiato, P. Maldonado, G. Eilers, J. Nötzold, S. Mährlein, V. Zbarsky, F. Freimuth, Y. Mokrousov, S. Blügel, M. Wolf, I. Radu, P. M. Oppeneer, and M. Münzenberg, “Terahertz spin current pulses controlled by magnetic heterostructures,” Nature Nanotechnology 8, 256–260 (2013), number: 4 Publisher: Nature Publishing Group. * Seifert _et al._ (2016) T. Seifert, S. Jaiswal, U. Martens, J. Hannegan, L. Braun, P. Maldonado, F. Freimuth, A. Kronenberg, J. Henrizi, I. Radu, E. Beaurepaire, Y. Mokrousov, P. M. Oppeneer, M. Jourdan, G. Jakob, D. Turchinovich, L. M. Hayden, M. Wolf, M. Münzenberg, M. Kläui, and T. Kampfrath, “Efficient metallic spintronic emitters of ultrabroadband terahertz radiation,” Nature Photonics 10, 483–488 (2016), number: 7 Publisher: Nature Publishing Group. * Wu _et al._ (2017) Y. Wu, M. Elyasi, X. Qiu, M. Chen, Y. Liu, L. Ke, and H. Yang, “High-Performance THz Emitters Based on Ferromagnetic/Nonmagnetic Heterostructures,” Advanced Materials 29, 1603031 (2017). * Battiato, Carva, and Oppeneer (2010) M. Battiato, K. Carva, and P. M. Oppeneer, “Superdiffusive Spin Transport as a Mechanism of Ultrafast Demagnetization,” Physical Review Letters 105, 027203 (2010), publisher: American Physical Society. * Battiato, Carva, and Oppeneer (2012) M. Battiato, K. Carva, and P. M. Oppeneer, “Theory of laser-induced ultrafast superdiffusive spin transport in layered heterostructures,” Physical Review B 86, 024404 (2012), publisher: American Physical Society. * Battiato, Maldonado, and Oppeneer (2014) M. Battiato, P. Maldonado, and P. M. Oppeneer, “Treating the effect of interface reflections on superdiffusive spin transport in multilayer samples (invited),” Journal of Applied Physics 115, 172611 (2014), publisher: American Institute of Physics. * Battiato and Held (2016) M. Battiato and K. Held, “Ultrafast and gigantic spin injection in semiconductors,” Phys. Rev. Lett. 116, 196601 (2016). * Battiato (2017) M. Battiato, “Spin polarisation of ultrashort spin current pulses injected in semiconductors,” J. Phys. Condens. Matter 29, 174001 (2017). * Carva, Battiato, and Oppeneer (2011) K. Carva, M. Battiato, and P. M. Oppeneer, “Ab Initio Investigation of the Elliott-Yafet Electron-Phonon Mechanism in Laser-Induced Ultrafast Demagnetization,” Physical Review Letters 107, 207201 (2011), publisher: American Physical Society. * Carva _et al._ (2013) K. Carva, M. Battiato, D. Legut, and P. M. Oppeneer, “Ab initio theory of electron-phonon mediated ultrafast spin relaxation of laser-excited hot electrons in transition-metal ferromagnets,” Physical Review B 87, 184425 (2013), publisher: American Physical Society. * Seifert _et al._ (2018) T. S. Seifert, N. M. Tran, O. Gueckstock, S. M. Rouzegar, L. Nadvornik, S. Jaiswal, G. Jakob, V. V. Temnov, M. Münzenberg, M. Wolf, M. Kläui, and T. Kampfrath, “Terahertz spectroscopy for all-optical spintronic characterization of the spin-Hall-effect metals Pt, W and Cu ${}_{\textrm{80}}$ Ir ${}_{\textrm{20}}$,” Journal of Physics D: Applied Physics 51, 364003 (2018). * Qiu _et al._ (2018) H. Qiu, K. Kato, K. Hirota, N. Sarukura, M. Yoshimura, and M. Nakajima, “Layer thickness dependence of the terahertz emission based on spin current in ferromagnetic heterostructures,” Optics Express 26, 15247–15254 (2018). * Liu _et al._ (2022) J. Liu, K. Lee, Y. Yang, Z. Li, R. Sharma, L. Xi, T. Salim, C. Boothroyd, Y. M. Lam, H. Yang, _et al._ , “Spintronic terahertz emitters in silicon-based heterostructures,” Physical Review Applied 18, 034056 (2022). * Zhou _et al._ (2018) C. Zhou, Y. Liu, Z. Wang, S. Ma, M. Jia, R. Wu, L. Zhou, W. Zhang, M. Liu, Y. Wu, and J. Qi, “Broadband Terahertz Generation via the Interface Inverse Rashba-Edelstein Effect,” Physical Review Letters 121, 086801 (2018), publisher: American Physical Society. * Torosyan _et al._ (2018) G. Torosyan, S. Keller, L. Scheuer, R. Beigang, and E. T. Papaioannou, “Optimized Spintronic Terahertz Emitters Based on Epitaxial Grown Fe/Pt Layer Structures,” Scientific Reports 8, 1–9 (2018), number: 1 Publisher: Nature Publishing Group. * Feng _et al._ (2018) Z. Feng, R. Yu, Y. Zhou, H. Lu, W. Tan, H. Deng, Q. Liu, Z. Zhai, L. Zhu, J. Cai, B. Miao, and H. Ding, “Highly Efficient Spintronic Terahertz Emitter Enabled by Metal–Dielectric Photonic Crystal,” Advanced Optical Materials 6, 1800965 (2018). * Cheng _et al._ (2021) L. Cheng, Z. Li, D. Zhao, and E. E. M. Chia, “Studying spin–charge conversion using terahertz pulses,” APL Materials 9, 070902 (2021), publisher: American Institute of Physics. * Cheng _et al._ (2019) L. Cheng, X. Wang, W. Yang, J. Chai, M. Yang, M. Chen, Y. Wu, X. Chen, D. Chi, K. E. J. Goh, J.-X. Zhu, H. Sun, S. Wang, J. C. W. Song, M. Battiato, H. Yang, and E. E. M. Chia, “Far out-of-equilibrium spin populations trigger giant spin injection into atomically thin MoS 2,” Nature Physics 15, 347–351 (2019), number: 4 Publisher: Nature Publishing Group. * Herapath _et al._ (2019) R. I. Herapath, S. M. Hornett, T. S. Seifert, G. Jakob, M. Kläui, J. Bertolotti, T. Kampfrath, and E. Hendry, “Impact of pump wavelength on terahertz emission of a cavity-enhanced spintronic trilayer,” Applied Physics Letters 114, 041107 (2019), publisher: American Institute of Physics. * Coutaz, Garet, and Wallace (2018) J.-L. Coutaz, F. Garet, and V. Wallace, _Principles of Terahertz time-domain spectroscopy_ (CRC Press, 2018). * Born and Wolf (2013) M. Born and E. Wolf, _Principles of optics: electromagnetic theory of propagation, interference and diffraction of light_ (Elsevier, 2013). * Agarwal _et al._ (2022b) P. Agarwal, Y. Yang, R. Medwal, H. Asada, Y. Fukuma, M. Battiato, and R. Singh, “Secondary spin current driven efficient thz spintronic emitters,” (2022b). * Yang, Dal Forno, and Battiato (2021a) Y. Yang, S. Dal Forno, and M. Battiato, “Removal of Spectral Distortion Due to Echo for Ultrashort THz Pulses Propagating Through Multilayer Structures with Thick Substrate,” Journal of Infrared, Millimeter, and Terahertz Waves , 1–11 (2021a). * Yang, Dal Forno, and Battiato (2021b) Y. Yang, S. Dal Forno, and M. Battiato, “Transfer-matrix description of heterostructured spintronics terahertz emitters,” Physical Review B 104, 155437 (2021b), publisher: American Physical Society. * Ordal _et al._ (1985) M. A. Ordal, R. J. Bell, R. W. Alexander, L. L. Long, and M. R. Querry, “Optical properties of fourteen metals in the infrared and far infrared: Al, Co, Cu, Au, Fe, Pb, Mo, Ni, Pd, Pt, Ag, Ti, V, and W,” Applied Optics 24, 4493 (1985). * Werner, Glantschnig, and Ambrosch-Draxl (2009) W. S. M. Werner, K. Glantschnig, and C. Ambrosch-Draxl, “Optical Constants and Inelastic Electron-Scattering Data for 17 Elemental Metals,” Journal of Physical and Chemical Reference Data 38, 1013–1092 (2009), publisher: American Institute of Physics. * Gao, Lemarchand, and Lequime (2012) L. Gao, F. Lemarchand, and M. Lequime, “Exploitation of multiple incidences spectrometric measurements for thin film reverse engineering,” Opt. Express 20, 15734–15751 (2012). * Zhang _et al._ (2013) W. Zhang, V. Vlaminck, J. E. Pearson, R. Divan, S. D. Bader, and A. Hoffmann, “Determination of the Pt spin diffusion length by spin-pumping and spin Hall effect,” Applied Physics Letters 103, 242414 (2013), publisher: American Institute of Physics. * Sanjuan and Tocho (2012) F. Sanjuan and J. O. Tocho, “Optical properties of silicon, sapphire, silica and glass in the terahertz range,” in _Latin America Optics and Photonics Conference_ (Optica Publishing Group, 2012) pp. LT4C–1. * Kelly (1972) R. L. Kelly, “Program of the 1972 annual meeting of the optical society of america,” J. Opt. Soc. Am. 62, 1336–1336 (1972). * Liu _et al._ (2021) Y. Liu, H. Cheng, Y. Xu, P. Vallobra, S. Eimer, X. Zhang, X. Wu, T. Nie, and W. Zhao, “Separation of emission mechanisms in spintronic terahertz emitters,” Physical Review B 104, 064419 (2021). * Papaioannou and Beigang (2021) E. T. Papaioannou and R. Beigang, “Thz spintronic emitters: a review on achievements and future challenges,” Nanophotonics 10, 1243–1257 (2021). * Adam _et al._ (2019) R. Adam, G. Chen, D. E. Bürgler, T. Shou, I. Komissarov, S. Heidtfeld, H. Hardtdegen, M. Mikulics, C. M. Schneider, and R. Sobolewski, “Magnetically and optically tunable terahertz radiation from ta/nife/pt spintronic nanolayers generated by femtosecond laser pulses,” Applied physics letters 114, 212405 (2019).
# Solution to the Number Rotation Puzzle Thomas Lam Carnegie Mellon University<EMAIL_ADDRESS> ###### Abstract. The Number Rotation Puzzle (NRP) is a combination puzzle in which the goal is to rearrange a scrambled rectangular grid of numbers back into order via moves that consist of rotating square blocks of numbers of fixed size. Over all possible boards and rotating block sizes, we find all solvable initial configurations and provide algorithms to solve such configurations. For sufficiently large board and rotating block sizes, solvability conditions depend only on parity restrictions, with special additional conditions for smaller sizes. One special case leads to a novel construction of the exotic outer automorphism on $S_{6}$. 2010 Mathematics Subject Classification. Primary 05E99; Secondary 00A08 Key Words. combination puzzle, three-cycle, exotic automorphism ## 1\. Introduction The Number Rotation Puzzle (NRP) is a combination puzzle in which the goal is to rearrange a scrambled square grid of integers back into ascending order. Numbers can be moved by rotating square blocks of numbers by some multiple of $90^{\circ}$. The game is most commonly played on a $3\times 3$ board of the integers $1$ to $9$, with $2\times 2$ rotating blocks. We call this the standard NRP. See Figure 1 for an example of a game played on the standard NRP. Figure 1. A two move solution to a sample standard NRP. In addition to the standard NRP, other variations exist, such as those with large board size and rotating block size, e.g. a $5\times 5$ grid with $3\times 3$ rotating blocks, and those with rectangular boards. The NRP has been implemented on Nokia phones and a few puzzle game websites, most notably Simon Tatham’s Twiddle (see [4]). There has been very little research done on the NRP. In fact, it has no “official" name. Most researchers focus on computational methods for solving the standard NRP. For example, Lee programmed an AI using decision matrices to solve the standard NRP [2], and Wang and Song developed an efficient brute-force algorithm to find solutions to the standard NRP [5]. However, one of the more outstanding feats in the research of the NRP is Fernando’s solution to the $n\times n$ NRP with $(n-1)\times(n-1)$ rotating blocks, which was in turn used to bound God’s Number — the least number of moves required to solve a combination puzzle — of the NRP [1]. The focus of this paper will be solving the NRP on rectangular boards, and thus generalizing the work of Fernando. By solve, it is meant that all possible solvable initial configurations are identified, and a logical solving algorithm is developed that will solve all such configurations to prove that they are indeed solvable. Particularly, our main result is as follows: ###### Theorem 1.1. Consider the $n\times m$ board with $b\times b$ rotating blocks, where $n$ is the number of rows and $m$ is the number of columns. * • If $b=n=m$, then there are exactly four solvable initial configurations, each a rotation of the solved board. Assume in the following criteria that this case does not occur. * • If $b=1$, then the only solvable initial configuration is the solved board. * • If $b=2$, then there are two further cases. * – If $(m,n)=(2,3),(3,2)$, then there are only 120 solvable configurations. * – Otherwise, all initial configurations are solvable. * • If $b=3$, then there are two further cases. * – If $(m,n)=(3,4),(4,3)$, then there are only $6!$ solvable initial configurations. In fact, solving one parity will solve the other. * – Otherwise, all initial configurations whose numbers lie in their correct parity set are solvable. * • If $b\geq 4$, then case on the residue of $b$ modulo 8. * – If $b\equiv 2,6\pmod{8}$, all initial configurations are solvable. * – If $b\equiv 0,4\pmod{8}$, the solvable boards are those that are even permutations of the solved boards. * – If $b\equiv 3,\ 5\pmod{8}$, the solvable boards are those that are even permutations of the solved boards such that each number lies in its correct parity. * – If $b\equiv 1,\ 7\pmod{8}$, the solvable boards are such that each parity is an even permutation of their respective solved states. We will also reveal a surprising construction of the exotic outer automorphism on $S_{6}$ in the special case $(n,m,b)=(3,4,3)$. Through this research, it is hoped that this puzzle be popularized in light of its intuitive mechanics. ## 2\. Preliminaries We denote by $(n,m,b)$ the NRP on an $n\times m$ board with $b\times b$ rotating blocks, where $n$ denotes the number of rows and $m$ denotes the number of columns. We denote moves by capital letters, and typically such a move is a $90^{\circ}$ counter-clockwise rotation of a block. If $X$ is such a move, then $X^{-1}$ is the $90^{\circ}$ clockwise rotation of said block, and $X^{2}$ is the $180^{\circ}$ rotation of said block. We read algorithms from left to right, so that e.g. $XYZ$ denotes executing $X$, $Y$, then $Z$. For an algorithm $A$ we write $A^{n}$ to denote $A$ repeated $n$ times, and write $A^{-1}$ to denote the inverse algorithm of $A$. We also impose a coordinate system with $(i,j)$ denoting the square in the $i$th row from the top and the $j$th column from the left. ## 3\. Parity Restrictions We let the case in which $m,n,b\geq 5$ with $m\neq n$ be the general case. We start by proving that our claimed conditions for solvability are necessary in the general case. These conditions come from three parity restrictions (PRs): 1. (1) Parity of a square 2. (2) Even permutations 3. (3) Even permutations within a parity of squares The “parity of a square" PR simply refers to the invariance of the color of a square in which a number lies if the board is colored like a checkerboard. Clearly this invariance occurs exactly when $b$ is odd, hence for odd $b$ we have that each of the numbers in the board must lie in the correct square parity in order to have solvability. This gives us the first PR. For some values of $b$, moves will execute even permutations, which indicates that solvable boards must be even permutations of the solved state. To identify such $b$, we take cases: * • If $b$ is even, then a move will move $b^{2}$ numbers in $b^{2}/4$ 4-cycles. Since 4-cycles are odd, we have that the permutation introduced by a move is even iff $b^{2}/4$ is even, and this occurs exactly when $b\equiv 0\pmod{4}$. * • If $b$ is odd, then each move will move $b^{2}-1$ numbers because the center of rotation stays in place. Again, it follows that moves execute even permutations iff $\frac{b^{2}-1}{4}$ is even, and some simple casework shows that this occurs for all odd $b$. Hence for $b\equiv 0,1,3\pmod{4}$, solvable boards must be even permutations, giving the second PR. The third PR is a combination of the first two. For odd $b$, each move executes a permutation that can be decomposed into a permutation of one parity and a permutation of the other parity. If these permutations are even, then solvable boards must have each of their parities be even permutations of their respective solved states. Since a $b\times b$ rotating block moves $\frac{b^{2}-1}{2}$ numbers within each parity, we see as before that a move is an even permutation within a parity iff $\frac{b^{2}-1}{8}$ is even. It is not hard to show that this occurs exactly when $b\equiv 1,7\pmod{8}$, giving the third PR and completing the proof of the necessary direction in the general case. ## 4\. General Solving Algorithm We now show that these conditions are sufficient for solvability in the general case by constructing a solving algorithm. We claim that it is sufficient to prove that * • for even $n$, we may 3-cycle any three numbers, and * • for odd $n$, we may 3-cycle any three numbers lying in a common parity. To see this, let us consider each of the cases described in the PRs. * • If $n\equiv 0\pmod{4}$, then the board must be an even permutation, and since we can execute any 3-cycle, we clearly can solve the board. * • For $n\equiv 2\pmod{4}$, the board may be an odd permutation. In this case, simply rotating any block $90^{\circ}$ will bring the board to an even permutation, in which case we can solve the board since we can execute any 3-cycle. * • For $n\equiv 1,7\pmod{8}$, each parity must be an even permutation of their respective solved states. Since we can compute any 3-cycle of numbers of the same number, it follows that we can execute any even permutation of a parity, hence the solvability. * • For $n\equiv 3,5\pmod{8}$, the board must be an even permutation. Looking at each parity, they are either both even permutations or both odd permutations of their respective solved states. If the latter, then rotating any block $90^{\circ}$ must bring both parities to even permutations, from which we deduce solvability as in the previous case. In practice, a solving algorithm would involve first executing a move if needed to bring the board to an even permutation, and then using 3-cycles to solve one number at a time until the entire board is inevitably solved. Next, we claim that it is sufficient to solve the case in which $m=n+1$ and $b=n$, i.e. the $(n,n+1,n)$ NRP for $n\geq 5$. One way to argue this is as follows: For the $(M,N,n)$ NRP with $M\geq n$ and $N\geq n+1$, we may 3-cycle any three numbers for $n$ odd, and 3-cycle any three numbers on the same parity for $n$ even, by executing 3-cycles within the $n\times(n+1)$ sub- boards in which each of the three numbers lie until they lie in a common $n\times(n+1)$ sub-board. Then we may 3-cycle them and undo the previous intermediate moves, resulting in the desired 3-cycle. For the $(n,n+1,b)$ NRP, there are only two rotating blocks that may be used. We denote rotation of the left and right rotating blocks counter-clockwise $90^{\circ}$ by $X$ and $Y$, respectively. To prove that any three numbers may be 3-cycled, subject to parity restrictions, we will first prove that there exists an algorithm cycling some three numbers. ### 4.1. Existence of a 3-cycle Algorithm We define an extremely useful algorithm, called cycle and denoted by $C$, as $C:=(YX)^{4}=YXYXYXYX.$ The importance of $C$ is that it does not move many numbers. Imagine that the numbers along the leftmost column, the bottom-most row, and the rightmost column form a sort of looped “conveyor belt", with $(1,n+1)$ looping back to $(1,1)$. Then $C$ will “cycle" the numbers on the conveyor belt counter- clockwise a total of $n-3$ squares along the belt (See Figure 2). Figure 2. The algorithm cycle in action on the $6\times 7$ board. Left: Initial board. Right: The board after execution of $(YX)^{4}$. Since $n=6$, all numbers in the belt (outlined in bold) are cycled $6-3=3$ squares counter- clockwise. To see that $C$ indeed executes this sort of behavior, first note that the numbers not in the belt form a square which is rotated $90^{\circ}$ with every execution of $YX$, hence these numbers’ positions are unaffected after cycle. Then, observe that the cyclic order of the numbers on the belt is invariant. After a $YX$, the number at $(1,1)$ will move to $(n,1)$. Thus, the belt is advanced $n-1$ squares counter-clockwise after each $YX$. It follows that $(YX)^{4}$ advances the belt a total of $4n-4$ squares counter-clockwise. Since there are only $3n-1$ squares on the belt, this is a net advancement of $n-3$ squares counter-clockwise. We are now ready to define an algorithm $\varphi$ that executes a 3-cycle. If we let $A=XYYXY^{-1}X^{-1}$, then we take $\varphi:=C(ACA^{-1})C^{-1}(AC^{-1}A^{-1}).$ The motivation is that the sequence $ACA^{-1}$ essentially “applies cycle to a different set of squares", so in some sense $\varphi$ is simply commuting two different instances of cycle. If the squares moved by $C$ and $ACA^{-1}$ intersect at exactly one square, then it is not hard to see that $\varphi$ must be a 3-cycle. To prove this, it is sufficient to show that after $A$, only one number that was in the belt will remain in the belt. This can be visually seen in Figure 3. See Section A.1 for a complete proof. Figure 3. The cycle “belt", after executing $A$. ### 4.2. The Spiral Algorithm To build up our plan for sending the three numbers that we wish to 3-cycle to the squares cycled by $\varphi$, we first develop a methodology for sending one number to any desired square (while obeying PRs in the sense that the number’s square and the goal square lie in the same parity when $n$ is odd). To do this, it is sufficient to be able to send any number to a “center square" of the board. In the discussion that follows, we will assume that if $n$ is odd then the numbers we are maneuvering lie in the “even parity" in the sense that their coordinates have even sum. The case where $n$ is even and the numbers in question have odd parity is handled by reflecting our arguments. Our specific goal is to send a number, which we will call $a$, to the square $u=(n/2,n/2+1)$ if $n$ is even (the “upper center square"), and $u=\left(\frac{n+1}{2},\frac{n+1}{2}\right)$ if $n$ is odd (the center of the $X$-rotating block). Note that these squares are of minimum positive distance to the center of the $Y$-rotating block in either case. The procedure for accomplishing this is the spiral algorithm, described as follows. 1. (1) If $a$ is located in the leftmost column, execute $X^{2}$ so that it is not. 2. (2) If either $a$ is at $u$, or is at any of the other three squares of minimum positive distance to the center of rotation of $Y$, then execute $Y$ as needed to move $a$ to $u$ and terminate. 3. (3) Execute $Y$ until the following inequalities hold, where $(i,j)$ are the coordinates of $a$: $\frac{n+1}{2}\leq i\leq n,\qquad 2\leq j\leq\frac{n+3}{2}.$ Visually, this condition entails that $a$ is located in the lower-left quadrant of the square rotated by $Y$, so this step is necessarily possible. 4. (4) Execute $X$. 5. (5) Loop back to Step 2. Figure 4. Geometrical interpretation of the spiral distance monovariant. $c_{x}$ is the center of rotation of the move $X$, and similarly for $c_{y}$. We claim that this algorithm must terminate. Let $c_{y}$ be the point at the center of rotation of $Y$, which may or may not be the center of a square. It suffices to show that the straightline distance from $a$ to $c$ is a strictly decreasing monovariant over each loop of the algorithm, since there are only a finite number of possible distances between the centers of two squares. Note that this distance is invariant as $Y$ is executed, so it suffices to observe a decrease in the distance after $X$ is executed. Right before $X$ is executed, let the distance from $a$ to $c_{y}$ be $d$. Then, since the coordinates of $a$ will satisfy the inequalities specified in step 3, $a$ will lie on the lower-left quarter arc of the circle centered at $c$ with radius $d$ (See Figure 4). Now, let us calculate a lower bound for $d$. If $n$ odd, we see that the coordinates of $c_{y}$ sum to $n+1$, which is odd, so the premise that $a$ is at $c_{y}$ contradicts the assumption that $a$ lies on an even-parity square. Moreover if $d=1$ then $a$ is already as close as it can be to $c_{y}$, which would have terminated the algorithm. Hence we may assume $d>1$ for odd $n$. For even $n$, we see that the closest possible straightline distance to $c_{y}$ is $\frac{\sqrt{2}}{2}$, and if this distance is obtained then the algorithm would have terminated. Otherwise, the next closest distance to $c_{y}$ would be $\frac{\sqrt{10}}{2}$, so we may assume that $d\geq\frac{\sqrt{10}}{2}$ for even $n$. In either case, we may assume that $d>1$. Now, executing $X$ will rotate the aforementioned arc $90^{\circ}$ counter- clockwise about the center of rotation of $x$, denoted by $c_{x}$. This will bring the arc closer to $c_{y}$. Since only the distances between this arc and $c_{y}$ are of concern here, it is equivalent to rotate $c_{y}$ $90^{\circ}$ clockwise about $c_{x}$ to $c^{\prime}_{y}$. The new distance between $a$ and $c_{y}$ after $X$ will simply be the distance between $a$ and $c_{y}^{\prime}$. Let this new distance be $d^{\prime}$. Now let the angle formed by $a$, $c_{y}$, $c^{\prime}_{y}$ be $\theta$. Since $a$ must lie on the lower-left quarter arc, we see that $0^{\circ}\leq\theta\leq 45^{\circ}$. The distance from $a$ to $c_{y}$ is the radius $d$, and since the distances between $c_{x}$ to $c_{y}$ and $c^{\prime}_{y}$ are both 1, the distance from $c_{y}$ to $c_{y}^{\prime}$ is $\sqrt{2}$. By Law of Cosines we have that $d^{\prime}=\sqrt{d^{2}+2-2\sqrt{2}d\cos{\theta}}.$ We want to show that $d^{\prime}<d$, or that $\sqrt{d^{2}+2-2\sqrt{2}d\cos{\theta}}<d.$ After algebraic manipulations, we see that this holds if and only if $d\cos{\theta}>\frac{1}{\sqrt{2}}$. This follows because we may multiply the assumption $d>1$ with the inequality $\cos\theta\geq\frac{1}{\sqrt{2}}$ which holds for $0\leq\theta\leq 45$. This completes the proof that spiral terminates. We conclude that any number $a$ may be sent to any target square, provided that the square in which $a$ lies shares the same parity as the target square in the case that $n$ is odd. We note that in spiral, there was nothing special about using the lower-left quadrant of the $Y$-rotating block and then executing $X$. It would have been just as effective to instead move $a$ to the upper-left quadrant of the $Y$-rotating block and then executing $X^{-1}$. Denote this variant as spiral*. ### 4.3. The Spiral-Cycle Algorithm We are now ready to describe the procedure for sending any three numbers to any three desired squares. As before, if $n$ is odd then we will require that the squares occupied by the three numbers, as well as the three goal squares, all lie on the same parity. In this case we will assume without loss of generality that these six squares lie on the even parity, as before. It suffices to choose three fixed squares $u_{1},u_{2},u_{3}$ and show that we can move any three numbers $a_{1},a_{2},a_{3}$ to these squares, since we can find the moves $B_{1}$ that send the three numbers we need to cycle to $u_{1},u_{2},u_{3}$, find the moves $B_{2}$ that send the goal squares to $u_{1},u_{2},u_{3}$, and then execute $B_{1}B_{2}^{-1}$. Our choice for $u_{1},u_{2},u_{3}$ is as follows: * • If $n$ is even, we choose $u_{1}=(1,1)$, $u_{2}=(2,1)$, and $u_{3}=(\frac{n}{2},\frac{n}{2}+1)$. * • If $n$ is odd, we choose $u_{1}=(1,1)$, $u_{2}=(3,1)$, and $u_{3}=(\frac{n+1}{2},\frac{n+1}{2})$. Note that these squares are chosen to lie on the even parity. The procedure for sending $a_{1},a_{2},a_{3}$ to $u_{1},u_{2},u_{3}$ is the spiral-cycle algorithm, described as follows for all $n\geq 6$: 1. (1) Use spiral to move $a_{1}$ to $u_{1}$. 2. (2) If $a_{2}$ is now in the first column, we move it out via $XYX^{-1}$. 3. (3) Use a modified spiral* to bring $a_{2}$ to $(n,n+1)$ for even $n$ or $(n-1,n+1)$ for odd $n$, with the following caveats: * • If we require usage of the $X$ rotating block in the spiral* algorithm, but rotating the $X$ block will displace the current position of $a_{1}$ (i.e. $a_{1}$ lies in the first column), then we execute the inverse cycle $C^{-1}$. This “hides" $a_{1}$ in the last column. Note that $C^{-1}$ cannot displace $a_{2}$, which would have to lie in the upper-left quadrant of the $Y$ rotating block when spiral* requires the use of the $X$ block, so $a_{2}$ cannot lie in the belt at this time. * • If we require usage of the $Y$ rotating block in the spiral* algorithm, but rotating the $Y$ block will displace the current position of $a_{1}$ (i.e. $a_{1}$ lies in the last column), then we execute $C$. This cannot displace $a_{2}$ either, since at this stage it must be the case that at least one $X^{-1}$ move of the spiral* algorithm was executed, so $a_{2}$ is too close to the center of the $Y$ rotating block to lie in the belt at this time. * • Once spiral* is terminated in this way, we execute $C$ is needed to restore $a_{1}$ to its position at $u_{1}=(1,1)$. 4. (4) Execute $XY^{-1}X^{-1}$. This will bring $a_{2}$ to $(2,1)$ for even $n$, or $(3,1)$ for odd $n$. Now $a_{1},a_{2}$ are at $u_{1},u_{2}$. 5. (5) If $a_{3}$ is now in the first column, then move it out using $XY^{2}XYXY^{-1}X^{2}Y^{-1}XY^{-1}X^{-1}$ for odd $n$, or $XY^{2}X^{-1}Y^{-1}XY^{-1}X^{-1}$ for even $n$. See Section A.2 for a proof that this works for $n\geq 5$. 6. (6) Mimic Step 3 in order to get $a_{3}$ to $u_{3}$. This will still work because $C^{-1}$ always moves the numbers on $(1,1),(2,1),$ and $(3,1)$ in the last column. If $n\geq 5$, replace each instance of $C$ with $C^{2}$. To see that the application of cycle always “hides" the numbers on $(1,1),(2,1),$ and $(3,1)$ as described, recall that $C^{-1}$ moves the numbers along the belt $n-3$ squares clockwise. For $n\geq 6$ we see that $3\leq n-3\leq n$, so it must move the first three numbers on the belt into the last $n$, i.e. the last column. For $n=5$, we use two applications of cycle instead so that the numbers move $2(5)-6=4$ squares clockwise along the belt. Since $3\leq 4\leq 5$, this gives the same result of moving $(1,1),(2,1),$ and $(3,1)$ into the last column. This justifies the correctness of spiral-cycle. We conclude that * • for $n$ even, we may send any three numbers to any three squares, and * • for $n$ odd, we may send any three numbers lying in a common parity to any three squares of that parity. Consequently, * • for $n$ even, we may 3-cycle any three numbers and * • for $n$ odd, we may 3-cycle any three numbers lying in a common parity, which is what we wanted to prove. This completes the proof of the theorem in the general case. ## 5\. Special Cases Our general solution fails for some small cases. This is due to the fact that the general 3-cycle algorithms described do not have enough space to execute the permutation that contains the 3-cycle. As a result, we have some special cases for smaller values of $m$, $n$, and $b$, that will introduce further solvability conditions. ### 5.1. $(2,n,2)$ $n$ is not large enough for any of the general algorithms to work. In fact, for $n=3$, general 3-cycle algorithms do not exist. It was proven by Jaap Scherphuis that out of all $6!$ theoretically reachable permutations in the $(2,3,2)$ variation, only $5!$ are achievable [3]. However, when $n\geq 4$, there exist algorithms that switch two numbers. Take any $4\times 2$ sub-board, and label the possible moves from left to right $X$, $Y$, and $Z$. Then the following algorithm will switch two numbers: $XYZ^{-1}Y^{2}X^{-1}Z^{-1}YZ^{2}Y^{-1}$ ### 5.2. $(m,n,2)$ where $m,n\geq 3$ Although this is classified as a special case, this is the most common version of the NRP, especially $(3,3,2)$, which is the standard NRP. Thus, many different resolutions to this variation are well-known. There are no restrictions on solvability. Consider a $3\times 3$ sub-board, and let $X$ rotate the lower-left $2\times 2$ block, $Y$ rotate the upper- right block, and $Z$ rotate the lower-right block. The following algorithm switches two numbers: $XY^{-1}X^{-1}YZ$ From here, it is easy to prove that all initial configurations are solvable. ### 5.3. $(3,4,3)$ We claim that out of all $12!$ theoretically reachable permutations, only $6!$ are achievable. Moreover, we claim that solving one parity will solve the other. Figure 5. The $3\times 4$ NRP with $3\times 3$ rotating blocks. The two possible moves are notated as shown. To show this, first observe that since $b$ is odd, the parity in which each number lies is preserved by each move. Hence we may view the puzzle as two separate but superimposed puzzles formed by by the 6 numbers in each parity. Now, let the two counter-clockwise moves be $X$ and $Y$, as in Figure 5. We let $P_{0}$ and $P_{1}$ be the set of coordinates with even coordinate sum and odd coordinate sum, respectively. To prove our claim, we need to show that, if the initial configuration is solvable, then 1. (1) we can obtain any permutation of $P_{0}$ (without any regard as to what happens to $P_{1}$), and 2. (2) $P_{0}$ is solved if and only if $P_{1}$ is solved. To prove (1), note that $XYX^{-1}Y^{-1}XY^{-1}XYXYX$ swaps $(2,4)$ and $(3,3)$, and fixes all other elements of $P_{0}$. It follows easily that any two numbers in $P_{0}$ may be swapped. As for (2), it suffices to prove that $P_{1}$ being solved implies that $P_{0}$ is solved by symmetry. In particular, it is sufficient to prove that any sequence of moves that fixes all elements of $P_{1}$ must also fix all elements of $P_{0}$. This is because if for some solvable board we have that all elements of $P_{1}$ are solved but $P_{0}$ is not, then by virtue of solvability there exists an algorithm $A$ that solves the $P_{0}$ parity but fixes $P_{1}$. What we wish to prove will contradict this. We define a pair to be a set of two elements. Let $S_{0}$ be the set of all pairs with elements in $P_{0}$, and let $S_{1}$ be the set of partitions of $P_{1}$ into three pairs. For a sequence of moves $A$, we let $A$ “act" on $S_{0}$ and $S_{1}$ in an element-wise sense. For example: * • We have the pair $\\{(1,1),(2,4)\\}\in S_{0}$, and after executing the $Y^{2}$ this pair becomes $\\{(1,1),(2,2)\\}$. * • We have the partition $\\{\\{(1,2),(3,4)\\},\\{(1,4),(3,2)\\},\\{(2,1),(2,3)\\}\\}\in S_{1}$, and after executing $X$ this becomes the partition $\\{\\{(2,1),(3,4)\\},\\{(1,4),(2,3)\\},\\{(3,2),(1,2)\\}\\}$. We shall denote such actions by $\cdot$, so e.g. $Y^{2}\cdot\\{(1,1),(2,4)\\}=\\{(1,1),(2,2)\\}$. The remainder of the proof relies in the following remarkably strong property. ###### Lemma 5.4. There exists a bijection $\phi:S_{0}\to S_{1}$ that preserves structure in the following sense: For all pairs $p\in S_{0}$, we have that $\phi(A\cdot p)=A\cdot\phi(p)$ for all $A$. Figure 6. Movement graphs on the sets $S_{1}$ and $S_{2}$. Top: Movement graph on the set of pairs on the parity set $P_{0}$. Bottom: Movement graph on the set of partitions into pairs on the parity set $P_{1}$. Each edge represents a move. It can be seen that the structures of these graphs are identical. ###### Proof. We list out all 15 elements of $S_{0}$ in a graph with two types of directed edges: an “$X$ edge" and a “$Y$ edge", where each edge represents the actions of $X$ and $Y$. That is, we place an $X$ edge between the pairs $p_{1}$ and $p_{2}$ if and only if $X\cdot p_{1}=p_{2}$, and similarly for the $Y$ edges. We may construct the same sort of graph for $S_{1}$. In doing so, we see that the two graphs are isomorphic! See Figure 6 for a visual of these graphs. The proof of our claim follows from this isomorphic structure by taking $\phi$ to be the natural correspondence that arises. ∎ To finish, consider a sequence of moves $A$ that fixes the elements of $P_{1}$. Take any pair $\\{a,b\\}\in S_{0}$. We claim that $A$ fixes this pair, i.e. $A\cdot\\{a,b\\}=\\{a,b\\}$. Indeed, using the lemma and the fact that $A$ fixes the elements of $P_{1}$, we have that $A\cdot\\{a,b\\}=\phi^{-1}(A\cdot\phi(\\{a,b\\}))=\phi^{-1}(\phi(\\{a,b\\}))=\\{a,b\\}.$ Since $A$ fixes all pairs, it must fix all elements of $P_{0}$. This is because for any $a,b,c\in P_{0}$, we have $A\cdot\\{a,b\\}=\\{a,b\\}$ and $A\cdot\\{a,c\\}=\\{a,c\\}$, so $A\cdot a\in\\{a,b\\}\cap\\{a,c\\}=\\{a\\}$, forcing $A\cdot a=a$ for all $a\in P_{0}$, as needed. As a remark, we note the following fascinating connection to group theory: Let $\sigma_{P_{0}}$ and $\sigma_{P_{1}}$ denote that permutation groups on $P_{0}$ and $P_{1}$, respectively. Construct a map $\psi:\sigma_{P_{0}}\to\sigma_{P_{1}}$ as follows: If a sequence of moves executes a permutation $\phi_{0}\in\sigma_{P_{0}}$ on $P_{0}$ and a permutation $\phi_{1}\in\sigma_{P_{1}}$ on $P_{1}$, then $\psi:\phi_{0}\mapsto\phi_{1}$. Then by the work we have done, $\psi$ is a well-defined isomorphism, and since $\sigma_{P_{0}}\cong\sigma_{p_{1}}\cong S_{6}$, we may view $\psi$ as an automorphism on $S_{6}$. In fact, a quick check reveals that $\psi$ is the exotic outer automorphism on $S_{6}$! ### 5.5. $(3,n,3)$ for $n\geq 5$ Consider some $3\times 5$ sub-board. Let $P_{0}$ and $P_{1}$ denote the squares in this sub-board of even and odd parity respectively, so that $P_{0}$ contains the upper-left corner. Label, from left to right, the three possible moves, $X$, $Y$, and $Z$. Then the algorithm $YZY^{-1}Z^{-1}YZ^{2}Y^{-1}Z^{-1}$ switches two numbers in $P_{0}$ (with other side effects in $P_{1}$), hence we can solve all numbers in $P_{0}$. Since each move is a rotation of a $3\times 3$ block, each move enacts an even permutation, so the resulting configuration of the numbers in the other parity, $P_{1}$, must be an even permutation of the solved order. Thus, it suffices to find an arbitrary 3-cycle algorithm on numbers in $P_{1}$ that does not displace the numbers in $P_{0}$. Consider $(XZX^{-1}Z^{-1})^{2}$, which is a 3-cycle. Then, one can demonstrate that any three numbers can be sent anywhere by adapting the transposition algorithm on $P_{0}$ to work on $P_{1}$ instead by "translating" it as such: $XYX^{-1}Y^{-1}XY^{2}X^{-1}Y^{-1}$. This will act as a transposition on $P_{1}$, and applying this repeatedly to move the three desired numbers to be cycled to the correct squares will serve as an intermediary algorithm. ### 5.6. $(m,n,3)$ where $m,n\geq 4$ The solvability condition here is the same as in the general case: the initial configuration must be an even permutation of the solved board, and every number must lie in its correct parity. Fernando’s general 3-cycle algorithm on square boards is enough to resolve this case [1]. Nevertheless, we present a short 3-cycle algorithm. Let $A$ rotate the upper-left $3\times 3$ block, $B$ rotate the upper-right block, $C$ rotate the lower-left block, and $D$ rotate the lower-right block. Then $ADA^{-1}D^{-1}C^{-1}DA^{-1}D^{-1}AC$ is a 3-cycle. ### 5.7. $(m,n,4)$ where $m,n\geq 4$ with $m\neq n$ The solvability condition here is the same as in the general case: the initial configuration must be an even permutation of the solved board. As in the proof of the general case, it suffices to consider the case $(4,5,4)$ and find a 3-cycle. An “easy" 3-cycle is given by $(XY^{2}X^{-1}YX^{-1}Y^{2})^{35}.$ All remaining special cases are completely trivial. ## 6\. Future Research There are still some open problems concerning the NRP. For example, all rotations were assumed to preserve the orientations of the numbers, so they always stay upright. It would be interesting to determine the additional solvability conditions if orientation of numbers are not preserved with rotation. Additionally, all moves were assumed to be rotations of square rotating blocks. Thus, a variation of the NRP can be made by letting the rotating blocks be rectangular, with moves consisting of rotating said blocks $180^{\circ}$. This would add an exciting dimension to an already convoluted puzzle. ## 7\. Acknowledgments I would like to thank Ravi Fernando for revealing to me the connection between my result on the $(3,4,3)$ NRP and the exotic automorphism on $S_{6}$. ## References * [1] R. Fernando. Nxn corner rotation puzzle. https://www.speedsolving.com/forum/threads/nxn- corner-rotation-puzzle.15472/, 2009. * [2] Y. Lee. Solving the “rotation” puzzle in stages. https://codemyroad.wordpress.com/2015/04/13 /solving-the-rotation-puzzle-in-stages/, 2015. * [3] J. Scherphuis. Two-generator corners group. URL:https://www.jaapsch.net/puzzles/pgl25.htm. * [4] S. Tatham. Chapter 7: Twiddle, 2018. * [5] G. Wang and J. Song. Bbfs-stt: An efficient algorithm for number rotation puzzle. Entertainment Computing, 12:1–7, November 2015. ## 8\. Appendix Here we include the unsightly proofs necessary to justify some of the claims we have made. We recall the notations of $X$ and $Y$ denoting the $90^{\circ}$ counter- clockwise rotations of the left and right blocks respectively on the $(n,n+1,n)$ NRP. We may view $X$ and $Y$ as functions on the board $\\{1,\cdots,n\\}\times\\{1,\cdots,n+1\\}$ defined via: $\displaystyle X(x,y)$ $\displaystyle=\begin{cases}(n+1-y,x),&1\leq y\leq n\\\ (x,y),&\text{otherwise}\end{cases}$ $\displaystyle Y(x,y)$ $\displaystyle=\begin{cases}(n+2-y,x+1),&2\leq y\leq n+1\\\ (x,y),&\text{otherwise}\end{cases}$ It will also be useful to state the behavior of $X^{-1},Y^{-1},X^{2},$ and $Y^{2}$: $\displaystyle X^{-1}(x,y)$ $\displaystyle=\begin{cases}(y,n+1-x),&1\leq y\leq n\\\ (x,y),&\text{otherwise}\end{cases}$ $\displaystyle Y^{-1}(x,y)$ $\displaystyle=\begin{cases}(y-1,n+2-x),&2\leq y\leq n+1\\\ (x,y),&\text{otherwise}\end{cases}$ $\displaystyle X^{2}(x,y)$ $\displaystyle=\begin{cases}(n+1-x,n+1-y),&1\leq y\leq n\\\ (x,y),&\text{otherwise}\end{cases}$ $\displaystyle Y^{2}(x,y)$ $\displaystyle=\begin{cases}(n+1-x,n+3-y),&2\leq y\leq n+1\\\ (x,y),&\text{otherwise}\end{cases}$ For ease, let us define the abuse of notation $([a,b],c):=\\{(a,c),(a+1,c),\cdots,(b,c)\\}$, and likewise for $(a,[b,c])$. For further ease and abuse, we view $\\{(a,b)\\}$ and $(a,b)$ as the same. ### 8.1. Image of the belt under the algorithm $A$ Let $A=XYYXY^{-1}X^{-1}$ be an algorithm, and let $E=([1,n],1)\cup(n,[2,n])\cup([1,n],n+1)$ denote the “belt". We show that the image of $E$ under the algorithm $A$ intersects with $E$ at exactly one square, provided that $n\geq 5$. We recall that our algorithms are read left-to-right, so eg. $(XY)(E)=Y(X(E))$. Our proof strategy simply involves computing the image of $E$ after every move, and splitting said image to prepare for the next move. * • After $X$: $\displaystyle X(E)$ $\displaystyle=X([1,n],1)\cup X(n,[2,n])\cup([1,n],n+1)$ $\displaystyle=(n,[1,n])\cup([1,n-1],n)\cup([1,n],n+1)$ $\displaystyle=(n,1)\cup(n,[2,n])\cup([1,n-1],n)\cup([1,n],n+1)$ * • After $Y^{2}$: $\displaystyle(XY^{2})(E)$ $\displaystyle=(n,1)\cup Y^{2}(n,[2,n])\cup Y^{2}([1,n-1],n)\cup Y^{2}([1,n],n+1)$ $\displaystyle=(n,1)\cup(1,[3,n+1])\cup([2,n],3)\cup([1,n],2)$ $\displaystyle=(n,1)\cup(1,n+1)\cup(1,[3,n])\cup([2,n],3)\cup([1,n],2)$ * • After $X$: $\displaystyle(XY^{2}X)(E)$ $\displaystyle=X(n,1)\cup(1,n+1)\cup X(1,[3,n])\cup X([2,n],3)\cup X([1,n],2)$ $\displaystyle=(n,n)\cup(1,n+1)\cup([1,n-2],1)\cup(n-2,[2,n])\cup(n-1,[1,n])$ $\displaystyle=(n,n)\cup(1,n+1)\cup([1,n-1],1)\cup(n-2,[2,n])\cup(n-1,[2,n])$ * • After $Y^{-1}$: $\displaystyle\phantom{{}={}}(XY^{2}XY^{-1})(E)$ $\displaystyle=Y^{-1}(n,n)\cup Y^{-1}(1,n+1)\cup([1,n-1],1)\cup Y^{-1}(n-2,[2,n])\cup Y^{-1}(n-1,[2,n])$ $\displaystyle=(n-1,2)\cup(n,n+1)\cup([1,n-1],1)\cup([1,n-1],4)\cup([1,n-1],3)$ * • After $X^{-1}$: $\displaystyle\phantom{{}={}}(XY^{2}XY^{-1}X^{-1})(E)$ $\displaystyle=X^{-1}(n-1,2)\cup X^{-1}([1,n-1],1)\cup X^{-1}([1,n-1],3)\cup X([1,n-1],4)\cup(n,n+1)$ $\displaystyle=(2,2)\cup(1,[2,n])\cup(3,[2,n])\cup(4,[2,n])\cup(n,n+1)$ From this we see that, under the key assumption that $n\geq 4$, we have that $(n,n+1)$ is the sole element in the intersection of $E$ and $A(E)$. ### 8.2. Moving $a_{3}$ out of the first column After executing the first four steps of the Spiral-Cycle algorithm, it may be the case that $a_{3}$ lies in the first column, in which case we cannot proceed. If so, then it must be moved out of this column without disturbing the positions of $a_{1}$ and $a_{2}$. Let us first consider the case in which $n$ is even. We claim that after $XY^{2}X^{-1}Y^{-1}XY^{-1}X^{-1}$, all numbers in the first column are moved out of it, without displacing $a_{1}$ and $a_{2}$, which are are $(1,1)$ and $(2,1)$ respectively. It is easy to verify that $a_{1}$ and $a_{2}$ are fixed under this algorithm. $\displaystyle(1,1)$ $\displaystyle\stackrel{{\scriptstyle X}}{{\mapsto}}(n,1)\stackrel{{\scriptstyle Y^{2}}}{{\mapsto}}(n,1)\stackrel{{\scriptstyle X^{-1}}}{{\mapsto}}(1,1)$ $\displaystyle\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(1,1)\stackrel{{\scriptstyle X}}{{\mapsto}}(n,1)\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(n,1)\stackrel{{\scriptstyle X^{-1}}}{{\mapsto}}(1,1)$ $\displaystyle(2,1)$ $\displaystyle\stackrel{{\scriptstyle X}}{{\mapsto}}(n,2)\stackrel{{\scriptstyle Y^{2}}}{{\mapsto}}(1,n+1)\stackrel{{\scriptstyle X^{-1}}}{{\mapsto}}(1,n+1)$ $\displaystyle\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(n,n+1)\stackrel{{\scriptstyle X}}{{\mapsto}}(n,n+1)\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(n,2)\stackrel{{\scriptstyle X^{-1}}}{{\mapsto}}(2,1)$ Now let $E=([3,n],1)$ be the rest of the first column. We show that the image of $E$ under the algorithm is disjoint from $E$, which suffices. Here we must assume $n\geq 4$. * • After $X$: $X(E)=(n,[3,n])$ * • After $Y^{2}$: $(XY^{2})(E)=Y^{2}(n,[3,n])=(1,[3,n])$ * • After $X^{-1}$: $(XY^{2}X^{-1})(E)=X^{-1}(1,[3,n])=([3,n],n)$ * • After $Y^{-1}$: $(XY^{2}X^{-1}Y^{-1})(E)=Y^{-1}([3,n],n)=(n-1,[2,n-1])$ * • After $X$: $(XY^{2}X^{-1}Y^{-1}X)(E)=X(n-1,[2,n-1])=([2,n-1],n-1)$ * • After $Y^{-1}$: $(XY^{2}X^{-1}Y^{-1}XY^{-1})(E)=Y^{-1}([2,n-1],n-1)=(n-2,[3,n])$ * • After $X^{-1}$: $(XY^{2}X^{-1}Y^{-1}XY^{-1}X^{-1})(E)=X^{-1}(n-2,[3,n])=([3,n],3)$ Evidently $([3,n],3)$ is disjoint from $([3,n],1)$, as needed. Now we consider the case in which $n$ is odd. Then we claim that $XY^{2}XYXY^{-1}X^{2}Y^{-1}XY^{-1}X^{-1}$ moves out all numbers in the first column without displacing $a_{1}$ and $a_{2}$, which are at $(1,1)$ and $(3,1)$ respectively. This algorithm is more difficult to motivate, but nevertheless we may manually verify that $a_{1}$ and $a_{2}$ are fixed under this algorithm. $\displaystyle(1,1)$ $\displaystyle\stackrel{{\scriptstyle X}}{{\mapsto}}(n,1)\stackrel{{\scriptstyle Y^{2}}}{{\mapsto}}(n,1)\stackrel{{\scriptstyle X}}{{\mapsto}}(n,n)\stackrel{{\scriptstyle Y}}{{\mapsto}}(2,n+1)\stackrel{{\scriptstyle X}}{{\mapsto}}(2,n+1)$ $\displaystyle\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(n,n)\stackrel{{\scriptstyle X^{2}}}{{\mapsto}}(1,1)\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(1,1)\stackrel{{\scriptstyle X}}{{\mapsto}}(n,1)\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(n,1)\stackrel{{\scriptstyle X^{-1}}}{{\mapsto}}(1,1)$ $\displaystyle(3,1)$ $\displaystyle\stackrel{{\scriptstyle X}}{{\mapsto}}(n,3)\stackrel{{\scriptstyle Y^{2}}}{{\mapsto}}(1,n)\stackrel{{\scriptstyle X}}{{\mapsto}}(1,1)\stackrel{{\scriptstyle Y}}{{\mapsto}}(1,1)\stackrel{{\scriptstyle X}}{{\mapsto}}(n,1)$ $\displaystyle\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(n,1)\stackrel{{\scriptstyle X^{2}}}{{\mapsto}}(1,n)\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(n-1,n+1)\stackrel{{\scriptstyle X}}{{\mapsto}}(n-1,n+1)\stackrel{{\scriptstyle Y^{-1}}}{{\mapsto}}(n,3)\stackrel{{\scriptstyle X^{-1}}}{{\mapsto}}(3,1)$ The difficulty is showing that the rest of the column, $E=(2,1)\cup([4,n],1)$, is moved out. We will need the assumpton that $n\geq 5$. * • After $X$: $X(E)=X(2,1)\cup X([4,n],1)=(n,2)\cup(n,[4,n])$ * • After $Y^{2}$: $(XY^{2})(E)=Y^{2}(n,2)\cup Y^{2}(n,[4,n])=(1,n+1)\cup(1,[3,n-1])$ * • After $X$: $(XY^{2}X)(E)=X(1,n+1)\cup X(1,[3,n-1])=(1,n+1)\cup([2,n-2],1)$ * • After $Y$: $(XY^{2}XY)(E)=Y(1,n+1)\cup Y([2,n-2],1)=(1,2)\cup([2,n-2],1)$ * • After $X$: $(XY^{2}XYX)(E)=X(1,2)\cup X([2,n-2],1)=(n-1,1)\cup(n,[2,n-2])$ * • After $Y^{-1}$: $(XY^{2}XYXY^{-1})(E)=Y^{-1}(n-1,1)\cup Y^{-1}(n,[2,n-2])=(n-1,1)\cup([1,n-3],2)$ * • After $X^{2}$: $\displaystyle(XY^{2}XYXY^{-1}X^{2})(E)$ $\displaystyle=X^{2}(n-1,1)\cup X^{2}([1,n-3],2)$ $\displaystyle=(2,n)\cup([4,n],n-1)$ * • After $Y^{-1}$: $\displaystyle(XY^{2}XYXY^{-1}X^{2}Y^{-1})(E)$ $\displaystyle=Y^{-1}(2,n)\cup Y^{-1}([4,n],n-1)$ $\displaystyle=(n-1,n)\cup(n-2,[2,n-2])$ * • After $X$: $\displaystyle(XY^{2}XYXY^{-1}X^{2}Y^{-1}X)(E)$ $\displaystyle=X(n-1,n)\cup X(n-2,[2,n-2])$ $\displaystyle=(1,n-1)\cup([3,n-1],n-2)$ * • After $Y^{-1}$: $\displaystyle(XY^{2}XYXY^{-1}X^{2}Y^{-1}XY^{-1})(E)$ $\displaystyle=Y^{-1}(1,n-1)\cup Y^{-1}([3,n-1],n-2)$ $\displaystyle=(n-2,n+1)\cup(n-3,[3,n-1])$ * • After $X^{-1}$: $\displaystyle(XY^{2}XYXY^{-1}X^{2}Y^{-1}XY^{-1}X^{-1})(E)$ $\displaystyle=X^{-1}(n-2,n+1)\cup X^{-1}(n-3,[3,n-1])$ $\displaystyle=(n-2,n+1)\cup([3,n-1],4)$ We see that $(n-2,n+1)\cup([3,n-1],4)$ does not intersect the first column, as needed.
# Anomalous relaxation of density waves in a ring-exchange system Pranay Patil<EMAIL_ADDRESS>Max-Planck-Institut für Physik komplexer Systeme, Nöthnitzer Straße 38, Dresden 01187, Germany Laboratoire de Physique Théorique, Université de Toulouse, CNRS, UPS, France Markus Heyl Max-Planck- Institut für Physik komplexer Systeme, Nöthnitzer Straße 38, Dresden 01187, Germany Theoretical Physics III, Center for Electronic Correlations and Magnetism, Institute of Physics, University of Augsburg, 86135 Augsburg, Germany Fabien Alet Laboratoire de Physique Théorique, Université de Toulouse, CNRS, UPS, France ###### Abstract We present the analysis of the slowing down exhibited by stochastic dynamics of a ring-exchange model on a square lattice, by means of numerical simulations. We find the preservation of coarse-grained memory of initial state of density-wave types for unexpectedly long times. This behavior is inconsistent with the prediction from a low frequency continuum theory developed by assuming a mean field solution. Through a detailed analysis of correlation functions of the dynamically active regions, we exhibit an unconventional transient long ranged structure formation in a direction which is featureless for the initial condition, and argue that its slow melting plays a crucial role in the slowing-down mechanism. We expect our results to be relevant also for the dynamics of quantum ring-exchange dynamics of hard- core bosons. ## I Introduction The field of dynamics in isolated quantum systems has recently received an increasing amount of attention thanks to the discoveries of a plethora of interesting non-equilibrium behaviors Calabrese (2018); Garrahan (2018); Gogolin and Eisert (2016), and of versatile experimental platforms to realize the same Browaeys and Lahaye (2020); Gross and Bloch (2017). These studies have been partially motivated by the desire to achieve the protection of quantum information from scrambling caused by Hamiltonian dynamics or environmental noise. This has lead to the rapid development of the field of many body localization Abanin _et al._ (2019); Alet and Laflorencie (2018); Abanin and Papić (2017); Luitz _et al._ (2015), which relies on strong disorder to provide a safeguard against scrambling many body dynamics, and the nascent field of Hilbert space fragmentation Sala _et al._ (2020); Khemani _et al._ (2020); Moudgalya _et al._ (2021); Yang _et al._ (2020); Mukherjee _et al._ (2021); Karpov _et al._ (2021), which results from the highly constrained configuration space of Hamiltonians with a large number of strong local constraints and/or highly frustrated interactions Zhang and Røising (2022); Sikora _et al._ (2011). A milder version of the total arresting of dynamics generated by the phenomena mentioned above is realized by systems which approach equilibrium in a manner which is qualitatively different from standard diffusion. An example of this has recently been explored for the spin-1/2 Heisenberg chain as well as its classical version at intermediate energies where evidence of super-diffusion has recently been seen both theoretically Ilievski _et al._ (2018); McRoberts _et al._ (2022) and experimentally Scheie _et al._ (2021), leading to connections to surface growth dynamics studied by Kardar-Parisi-Zhang Weiner _et al._ (2020); Kardar _et al._ (1986). An approach to equilibrium which is slower than that expected from diffusion has also been realized in systems which conserve higher moments (such as dipolar and octupolar) of the spin configuration Iaconis _et al._ (2019); Feldmeier _et al._ (2020); Morningstar _et al._ (2020); Iaconis _et al._ (2021); Richter and Pal (2022). For two- and higher-dimensional systems, this has also lead to the realization of exotic fractonic phases Gromov _et al._ (2020); Nandkishore and Hermele (2019); Pretko _et al._ (2020). Remarkable advances have also been made on probing experimental realizations of Hilbert space fragmentation and/or higher-moment conservation and associated sub-diffusion Guardado- Sanchez _et al._ (2020); Scherg _et al._ (2021); Kohlert _et al._ (2021). While progress has been made on the analytical description and experimental detection of the phenomena mentioned above, there has been a dire need for numerical simulations on microscopic models to lend support to many of the predictions. This is in general a difficult task as powerful methods to simulate large-scale quantum dynamics on the needed long-time scales are relatively few and capable of handling only specific regimes Baiardi and Reiher (2019); Schmitt and Heyl (2020). It was found insightful Iaconis _et al._ (2019); Iadecola and Vijay (2020) to adopt methods from stochastic classical mechanics in the framework of cellular automata Iaconis _et al._ (2019); Gopalakrishnan and Zakirov (2018); Iaconis (2021), which ignore part of the quantum phase fluctuations and have been able to successfully describe the long-term dynamical behavior of strongly interacting quantum systems. This intuition arises from the expectation that for generic systems with sufficiently large Hilbert spaces and for times long compared to the microscopic energy scales of the Hamiltonian, the dynamics does not show quantum coherence, thus reducing to a classical dynamics problem. As mentioned above, well-known exceptions to this exist, but as simulation of exact quantum dynamics is out of reach using current methods, a study of the classical equivalent becomes of particular interest. This can also serve as a natural starting point to understand the complete quantum dynamics. Following up on the studies of constrained systems, we consider in this work the case of a simple hard-core bosonic model living on a square lattice, undergoing ring-exchange dynamics. This model has already been studied from the perspective of cuprates, as they serve as promising candidates to realize high-Tc superconductivity Paramekanti _et al._ (2002); Rousseau _et al._ (2005); Schaffer _et al._ (2009). Although traditionally most studies have focused on the possible exotic ground state features of this model Tay and Motrunich (2010); Tay _et al._ (2011), some recent works Iaconis _et al._ (2019); Khudorozhkov _et al._ (2021) have considered the constrained dynamics generated by the ring-exchange mechanism, including starting from random configurations Iaconis _et al._ (2019). However, the relevance of fragmentation to generic low-momentum states which are only described by macroscopic patterns has not been investigated. In this work we address this question using a classical approach based on stochastic dynamics. We find that structured initial configurations in the form of a boson density wave retain their coarse grained structure for a time which grows as a tunable power of the wavelength, with an eventual melting which is approximately described by a continuum model derived from a simple Taylor expansion. We study the detailed structure of the melting process via spatial correlation functions and find that the dynamics proceed through the development of strongly correlated large active regions which merge and destroy the initial modulated pattern. The detailed plan of the paper is as follows. In Sec.II, we present the model, discuss the quantities conserved under the dynamics, and elucidate the general profile of stripe-like configurations which are perfectly frozen under the dynamics. The bulk of this section discusses the effects of small perturbations on these exactly frozen patterns, and the preservation of the memory of the initial state to infinite times as illustrated by simulation of exact quantum and stochastic classical dynamics on a small system size. The close agreement between exact quantum dynamics and its stochastic classical equivalent seen in this section motivates our approximation and we focus purely on the classical system for the following sections. Sec.III recalls the expected continuum field theory based on Taylor expansions by first considering the simpler case of correlated random walkers, along with numerical checks of the equations developed. This is followed by a treatment of the hard-core model using a mean-field assumption. For Sec.IV, we move to more general configurations which take the form of boson density waves, and show the persistence of the memory of the initial state for unexpectedly long periods of time. We also compare the prediction of the continuum field theory developed in Sec.III with the initial dynamics. We follow up in Sec.V with a detailed analysis of the evolution of the dynamical active regions, and present a phenomenological picture of the mechanism leading up to the melting of the initial density wave configuration. We summarize our results in Sec.VI and present possibilities for future follow ups via direct simulation of the quantum many body dynamics. In a following appendix A, we briefly discuss the long-time momentum space profile of correlation functions, show that it is consistent with a mean-field treatment obtained in a previous work, and that it cannot be a basis of the anomalous scaling observed in the present work (as could have been deduced from a recent analysis Sala _et al._ (2021) of a related model). ## II Ring-exchange boson dynamics, conserved quantities and frozen patterns We consider a system of hard-core bosons living on a square lattice, which evolve stochastically in time using only ring-exchange dynamics where bosons hop by pairs around a 4-sites plaquette of the lattice if and only if a single diagonal of the plaquette is occupied by two bosons (“flippable” plaquettes). This only allowed dynamics is shown in Fig. 1. This dynamic rule trivially conserves the total particle number as well as the number of particles in each individual column and in each individual row Tay and Motrunich (2010); Tay _et al._ (2011); Iaconis _et al._ (2019); Khudorozhkov _et al._ (2021). For the rest of this work, we restrict ourselves to half-filling, i.e. $L^{2}/2$ sites occupied by bosons, on a periodic $L\times L$ lattice, and to the sector where each column and each row has exactly $L/2$ occupied sites. One expects it to be the sector with the largest number of configurations, as it is maximally symmetric. The total number of configurations in this sector can be computed on large lattices using combinatorial techniques Canfield and McKay (2005). We find that the conserved quantities discussed above do not in themselves completely describe the dynamically connected sets of configurations. This can be seen by considering a “perfect” stripe configuration as shown in Fig. 1, where an alternating pattern of filled and empty sites does not leave a single flippable plaquette, making the configuration frozen under ring-exchange dynamics. By varying the widths and locations of the filled and empty stripes, we can create many similar frozen configurations. One may also naively expect that this restriction on the number of accessible configurations extends to the case where we do not have perfect stripes. To see this, we consider a configuration generated by interchanging the two neighboring diagonals on the edge between a filled and an empty region. This creates a diagonal made of flippable plaquettes, where the influence of this active region may be expected to only extend to a few lattice spacings around this diagonal. Figure 1: (a) The ring-exchange rule that is the only allowed dynamics in our system. (b) A perfect stripe configuration which is frozen under the dynamic rule above. (c) Initial configuration with mixed filled-staggered pattern on a $L=16$ sample used to seed Hilbert space fragment. The dynamics of the (d) overlap $O(t)$ and (e) Fourier ratio $R(t)$ with this initial state for both exact dynamics (ED) in the quantum case and stochastic classical automaton (SCA) dynamics, which quickly saturate to the exact values calculated from enumerating all connected configurations. To test this intuition and illustrate this effect, we exactly enumerate all the configurations connected to an initial condition of the form described above for an $16\times 16$ lattice (shown in Fig. 1). We find that the total number of configurations, which we call $N_{c}$ which belong to this “fragment” is $27,990$. To show that they retain some structure of the representative which we have chosen, we compute an overlap as defined below, $\bar{O}=\frac{4}{N_{c}L^{2}}\sum_{c}\sum_{x,y}n^{\rm seed}_{x,y}n^{c}_{x,y}.$ (1) where $n_{x,y}^{c/{\rm seed}}=0,1$ is the number of bosons at site of coordinates $(x,y)$ in the ($c$-th/initial) configuration. If we use all possible configurations (in which case $N_{c}$ would be the total Hilbert space size), we would expect $\bar{O}=0$ due to the symmetry under $n\to 1-n$. For the restricted Hilbert space belonging to the fragment being considered we find $\bar{0}=0.844003...$, showing that for a dynamic simulation restricted to this sector, the initial and late time states would still retain significant overlap. To confirm this, we also run a stochastic classical automaton (SCA) simulation, where at each time step, we propose $L^{2}$ random plaquette flips, and if the chosen plaquette is flippable, we flip it with probability $1/2$. The resulting overlap with the initial state is shown in Fig. 1, and we see that it quickly approaches the value expected from exact enumeration, and retains it indefinitely. To study the accuracy with which the SCA reproduces the exact quantum dynamics, we also perform an exact diagonalization for this Hilbert space fragment, and compute $\braket{\psi(t)}{O}{\psi(t)}$ for $\ket{\psi(t)}=e^{-iHt}\ket{\psi_{0}}$ using $H=\sum_{x,y}b^{\dagger}_{x,y}b_{x+1,y}b_{x,y+1}b^{\dagger}_{x+1,y+1}+h.c.$, where the initial condition is the same occupation-basis state as we initialized the SCA with and $O$ now denotes an operator which is diagonal in the occupation-basis and measures the overlap with the initial state. This operator can be generated directly from the expression above for the classical case by promoting $n^{c}_{x,y}$ to the number operator, while retaining the integer status of $n^{\rm seed}_{x,y}$. As shown in Fig. 1, we find that the overlap in the quantum dynamics closely traces the SCA during the initial decay away from an overlap of $1$. In the following sections, we quantify the relaxation to equilibrium using the Fourier transform $n_{k}(t)$ with frequency $k$ along the $\hat{x}+\hat{y}$ direction, at time $t$. To this end, we measure the Fourier ratio $R(t)=n_{k}(t)/n_{k}(0)$, and make a similar comparison as done for the spatial overlap above. This is presented in Fig. 1 as well, and once again we find a close agreement between the exact quantum dynamics and it’s classical equivalent in the way they approach the equilibrium value of $R(t)$ within the sector of interest. This result suggests that the quantum dynamics within sectors matches the classical automaton upon coarse graining past time scales of $O$(1), suggesting that quantum phases do not play a substantial role in the aspects of dynamics which we want to study, and that a classical automaton approach could be sufficient to study the effect of the kinetic constraints on large scale features. Following the arguments above and noting that exact quantum dynamics for larger sizes is not possible with current computational capabilities, we directly study the long time behavior of our stochastic dynamics simulations to gather information about large system sizes in the following sections. ## III Expected hydrodynamic description Here we discuss the coarse grained description in continuous space, of the microscopic dynamics which we have introduced in the previous section. We begin by relaxing the hard-core constraint and considering the limit of average number of particles per site being $\gg 1$. This allows us to reduce the problem to non-interacting correlated random walkers, and leads to an analogue of the diffusion equation which encodes vanilla sub-diffusion. We show numerical evidence for the validity of the same. The expected continuum dynamical behavior of this type of ring-exchange in 2d was first presented in Ref. Iaconis _et al._ (2019), but we recall it for completeness as well as to understand how it should be perturbed to take into account the hard-core nature. Next we move to an equivalent model of a real-valued field on the lattice which makes the connection to the hydrodynamic limit more transparent, and we again provide support with numerical simulations. Lastly, we return to the hard-core model presented in the previous section, and show that the continuum theory describing the model must include non- linear terms in addition to the sub-diffusive term, and we show the possibility for a quantitative change in the behavior of the system due to the non-linearities. ### III.1 Large particle number limit We relax the hard-core constraint of the stripe configurations described in the previous section and assume the pattern to exist over a featureless background of an average density of $n_{d}\gg 1$ particles per site. The dynamics can now be understood in terms of correlated random walkers in the following way. First, we label each particle in the system as an independent walker. A move is defined now as first picking a walker (called walker $a$) at random, picking one of its four next nearest neighboring sites with probability $1/4$, and moving one of the walkers on the chosen site (called walker $b$) in tandem with walker $a$ in a ring exchange type manner. Due to the large number of walkers per site, we expect to always be able to find walker $b$. To write down the number of particles $n$ at a particular position $(x,y)$ at time $t+1$ as a function of the values at time $t$, we must consider all processes which can change $n_{x,y}$. These processes are (1) choosing a walker at $(x,y)$ and moving it away ($\propto n_{x,y}$); (2) choosing a walker at one of the four nearest neighbor sites and moving it or its partner to $(x,y)$ ($\propto 1/2(n_{x\pm 1,y})$ or $1/2(n_{x,y\pm 1})$); (3) choosing one of the walkers at one of the next nearest neighbors and moving it in tandem with a walker at $(x,y)$, thus reducing the number of walkers at $(x,y)$ by one ($\propto 1/4(n_{x\pm 1,y\pm 1})$). Putting these terms together, we can write the change in $n_{x,y}$, which is an whole number, from time $t$ to time $t+1$ as (note that all terms in the expression below are at time $t$, and we do not explicitly mention it for ease of representation) $\displaystyle\Delta_{t}$ $\displaystyle n_{x,y}\propto-n_{x,y}$ $\displaystyle+\frac{1}{2}[n_{x+1,y}+n_{x-1,y}+n_{x,y+1}+n_{x,y-1}]$ $\displaystyle-\frac{1}{4}[n_{x+1,y+1}+n_{x+1,y-1}+n_{x-1,y+1}+n_{x-1,y-1}].$ (2) This equation was obtained using similar arguments in Ref. Iaconis _et al._ (2019). If the system is initialized over a background of $n_{d}$ particles per site using the stripe configuration described in the previous section (shown in Fig. 1), $n_{(x,y)}$ can be seen as a step function switching periodically between $n_{d}$ and $n_{d}+1$. We would naively expect that the correlated dynamics discussed above would quickly eliminate the sharp boundaries of the $n_{(x,y)}$ texture and lead to a smooth function once we average over many realizations of the stochastic dynamics. For a function which varies slowly as a function of $(x,y)$ (note that this implies that the stripes in the initial condition should be wide compared to lattice spacing), we can perform a Taylor expansion of the expression above. We find that all terms to fourth order cancel, and the only term at fourth order yields ${\partial_{t}}n_{(x,y)}=-c{\partial_{x}^{2}}{\partial_{y}^{2}}n_{(x,y)}$ (3) where $c=1$ if following the treatment above. For further convenient reporting of the wave-vector $k$ in units of $\pi$, and taking into account in addition a factor of $4$ coming from the acceptance probabilities in our numerical implementation, we consider instead a different normalization with $c=\pi^{4}/4$. Rescaling the x-axis as in Fig. 2, this allows to recover a match to $e^{-x}$ for the fit in Fig. 2. For the rest of this manuscript, we maintain this convention for $c$. As the stripe initial condition is a periodic square wave in $(x+y)$, it is convenient to rewrite the above equation in the Fourier basis, and consider only the lowest harmonic (largest wavelength) of the square wave transform. This reduces the dynamical equation to ${\partial_{t}}n_{k_{x},k_{y}}=-ck_{x}^{2}k_{y}^{2}n_{k_{x},k_{y}}$. For diagonal stripes, $k_{x}=k_{y}=k$, and $n_{k}$ can be exactly reduced to $n_{k}(0)\exp{(-ck^{4}t)}$, where $n_{k}(0)$ is the value at $t=0$. We can now numerically verify this behavior by calculating the Fourier ratio $R_{k}(t)=n_{k}(t)/n_{k}(0)$, and looking for a data collapse onto a single exponential for various values of $k$. We find that $n_{d}=3$ provides a sufficiently large background for a good data collapse, and show the same for a $128\times 128$ lattice for various stripe widths (encoded in $k$) in Fig. 2 and averaged over $80$ realizations of the stochastic dynamics for each $k$. For smaller values of $n_{d}$, we find an increasing discrepancy between different values of $k$, with the divergence growing with decreasing $n_{d}$. Figure 2: Dynamics of the ratio $R_{k}(t)=n_{k}(t)/n_{k}(0)$, against $k^{4}t$ for various values of $k$ and on a log-linear scale, for a $128\times 128$ system of correlated walkers, with a background base walker density per site $n_{d}=3$. The solid line is a fit to a single exponential. ### III.2 Discretization of continuum equation Here we study a discretization of Eq. (3) on the lattice by defining the real valued field $n_{x,y}$, and attempt to recover the limit of the hard-core model. From the analysis of the correlated random walkers above, we expect that a ring exchange dynamic on a plaquette should lead to the ${\partial_{x}^{2}}{\partial_{y}^{2}}$ form. To this end, we define the “activity” of a plaquette whose left bottom site is $(x,y)$ to be $a=[n_{x,y+1}-n_{x,y}]-[n_{x+1,y+1}+n_{x+1,y}]$. This is evaluated at time $t$, and the fields living on the plaquette are updated at time $t+1$ by adding $\epsilon a$ to sites $(x,y)$ and $(x+1,y+1)$ and subtracting $\epsilon a$ from the other two sites. The constant $\epsilon$ is included to ensure that the field remains $\in[-1,1]$, and plays only a quantitative role in the scaling study. To deduce the dynamic rule for the activity at a single lattice site, we must consider the four neighboring plaquettes around it which can affect the field on the chosen site via plaquette updates. By summing $a$ for the four plaquettes with equal weight, and carrying out a careful grouping of terms, we see that the equation reduces exactly to the dynamical equation discussed in the previous subsection. Using a Taylor series expansion once again leads to a dynamical equation of the form given in Eq. (3). We attempt a data collapse for the decay of $R(t)$ as defined in the previous subsection for stripe configurations using the dynamic rule mentioned above. As shown in Fig. 3, we once again find a satisfactory data collapse to a single exponential for a large range of $k$ values. Before we turn to the case of the hard-core model, we must note that the dynamics described in this subsection differ in one crucial way from the hard- core model. We can see this by considering all the plaquette configurations which yield a non-zero value of $a$ (these are listed in Fig. 4), and observing that only the two completely “flippable” (or equivalently with the largest magnitude of $a$) contribute to dynamics in the hard-core case. This is not the case if one considers the dynamical rule used here and in the previous subsection (all plaquettes with all values of $a$ are updated). We argue in the next subsection that this leads to a strong violation of Eq. (3). Figure 3: Dynamics of $R(t)$ for a $128\times 128$ system with continuum fields at each lattice site, against $k^{4}t$ for various values of $k$ and on a log-linear scale. The solid line is a fit to a single exponential. ### III.3 Continuum theory under the hard-core constraint Now we turn to the model described in the previous section, and restrict particle number to at most one per site, and dynamics only to proceed via exactly flippable plaquettes. Note that now we cannot define an updated state using just the definition of $a$ as in the previous subsection. We require a function which evaluates to unity only for the two flippable plaquette configurations and zero for all others. For the purpose of this subsection, it is more convenient to set $n_{(x,y)}=1$ for an occupied site and $-1$ for an unoccupied site. It is not apparent if there is a unique function which achieves this, and one of the simplest functions which we were able to find on the plaquette to satisfy these constraints is $\displaystyle h=\frac{1}{4}\big{[}$ $\displaystyle(n_{x,y+1}-n_{x,y})-(n_{x+1,y+1}-n_{x+1,y})$ $\displaystyle-\frac{1}{4}(\Sigma_{p}n)(n_{x,y}n_{x+1,y}-n_{x+1,y+1}n_{x,y+1})$ $\displaystyle\ \ \ \ \ \ \ \times(n_{x,y}n_{x,y+1}-n_{x+1,y}n_{x+1,y+1})\big{]},$ (4) where $\Sigma_{p}n=n_{x,y}+n_{x,y+1}+n_{x+1,y+1}+n_{x+1,y}$ is the sum of all $n$ belonging to the plaquette. A careful consideration of the expression above reveals that it generates a value of $\pm 1$ for the flippable plaquettes shown in Fig. 4a, while returning zero for all other configurations (including those in Fig. 4b), thus satisfying our requirements for a hard-core ring exchange. The second term in the above expression is formed by noticing that the only configurations which violated the assignment of values we desired had a difference in the types of pair arrangements on opposite edges. Before performing approximations on this expression to derive a continuum theory, we find that it is convenient to expand the product of the last two bracketed terms discussed above as $(2n_{x+1,y}n_{x,y+1}-2n_{x,y}n_{x+1,y+1})$, where we have used $n_{x,y}^{2}=1$ for all $(x,y)$. Figure 4: Plaquettes which have a non-zero contribution to dynamics in Sec. III.2. Plaquettes (a) are flippable, and (b) do not contribute under hard-core discrete dynamics of Sec. III.3. For (b), we did not include particle-hole symmetric partners (with black and white dots interchanged) for simplicity. For the evolution $\Delta_{t}n_{x,y}$ of density of a single site, we must once again consider the four plaquettes in which it participates. The resulting expression has a term identical to the dynamic equation in the previous subsections, but has an important addition in the form of the sum of $\frac{1}{2}(\Sigma_{p}n)(n_{{x^{\prime}},y}n_{x,{y^{\prime}}}-n_{x,y}n_{{x^{\prime}},{y^{\prime}}})$ over the four plaquettes. To get the continuum limit as done previously, we assume initial conditions with several flippable plaquettes and a smooth density profile which varies slowly spatially. An average over stochastic dynamics and initial conditions thus allows us to replace the terms linear in $n$ with the differential form $-{\partial_{x}^{2}}{\partial_{y}^{2}}g$, where $g$ is a real valued field living in continuous space-time aiming at replacing $n$ (we explicitly distinguish it from $n$ for this subsection due to the correlations possibly generated by the products of $n$). To understand the behavior of the averaged value of terms such as $n_{x+1,y}n_{x,y+1}$, an important assumption about the correlations has to be made. In the following of this subsection, we work in a mean-field picture ignoring correlation effects and assume that such terms can be rewritten as the product of $g$ at the two points. In doing so, we will be able to derive the complete dynamics only in terms of the field $g$. This mean-field like assumption is true for the initial condition we work with, as it is drawn from an uncorrelated ensemble, but it is not a priori evident if the dynamics maintains this uncorrelated nature or rapidly builds up correlations. The non- linear term $f$ is expressed as $\displaystyle(\sum_{x,y}n)(n_{x+1,y}n_{x,y+1}-n_{x,y}n_{x+1,y+1})$ $\displaystyle+$ $\displaystyle(\sum_{x-1,y-1}n)(n_{x,y-1}n_{x-1,y}-n_{x-1,y-1}n_{x,y})$ $\displaystyle-$ $\displaystyle(\sum_{x-1,y}n)(n_{x,y}n_{x-1,y+1}-n_{x-1,y}n_{x,y+1})$ $\displaystyle-$ $\displaystyle(\sum_{x,y-1}n)(n_{x+1,y-1}n_{x,y}-n_{x,y-1}n_{x+1,y}),$ where the subscript below the sum indicates the plaquette over which the sum is performed indexed by its left bottom site. Once again we expand the products and ensure that we replace all occurrences of $n_{x,y}^{2}$ with unity for all $(x,y)$. The equation above is thus reduced to a linear combination of single body and three body terms. This can be further reduced by performing the mean field decoupling $\braket{n_{x,y}n_{{x^{\prime}},{y^{\prime}}}n_{{x^{\prime\prime}},{y^{\prime\prime}}}}=\braket{n_{x,y}}\braket{n_{{x^{\prime}},{y^{\prime}}}}\braket{n_{{x^{\prime\prime}},{y^{\prime\prime}}}}=g_{x,y}g_{{x^{\prime}},{y^{\prime}}}g_{{x^{\prime\prime}},{y^{\prime\prime}}}$, followed by a Taylor expansion around the relevant site in the derivatives of $g$. As we began our analysis by considering a sum over the four plaquettes surrounding the site $(x,y)$, symmetry restrictions apply to the terms generated by the Taylor series, which require that only terms with non-zero coefficients have an even number of derivatives in $x$ and $y$, and are symmetric with respect to the $x\to y$ transformation. Using this constraint and after some algebra, we find that the only surviving terms arise at fourth order in the derivatives and that the complete dynamical equation reduces to the following expression (ignoring a global factor of $1/4$), given by $\displaystyle\partial_{t}g=-\Big{(}\frac{1}{2}-g^{2}\Big{)}$ $\displaystyle{\partial_{x}^{2}}{\partial_{y}^{2}}g-{\partial_{y}^{2}}g({\partial_{x}}g)^{2}-{\partial_{x}^{2}}g({\partial_{y}}g)^{2}$ $\displaystyle-2({\partial_{x}}g)({\partial_{y}}g)({\partial_{x}}{\partial_{y}}g)-g{\partial_{x}^{2}}g{\partial_{y}^{2}}g.$ (5) The presence of non-linearities, derived even under a crude mean-field approximation, suggests that at leading order in the dynamics, the hard core constraint indeed plays an important role, and may invalidate the expectation that the coarse grained dynamics are equivalent to those of vanilla sub- diffusion. ## IV Long time persistence and eventual melt of approximate stripe configurations We have seen in the Sec. II that configurations which can be viewed as small perturbations around a perfect stripe configuration may maintain the memory of the initial state indefinitely. As these configurations are highly specific, it would seem unrealistic to choose one of these as the initial state for the dynamics of large systems. This motivated us to study “approximate” stripe patterns, which are chosen to be boson density waves with a wave-vector $\vec{k}=(k_{x},k_{y})$. Figure 5: A sample boson configuration created by annealing to the target function $f_{\vec{k}}$ Eq. 6 with $k_{x}^{t}=k_{y}^{t}=0.1875$ and $A=1.0$ for a $96\times 96$ lattice, along with Fourier components of random initial states for larger sizes generated using the annealing method for $k_{x}^{t}=k_{y}^{t}=0.1875$, plotted against the nearest distance $k_{d}$, shows a clear peak for $k_{d}=0$. ### IV.1 Initial state preparation To prepare such configurations, we first generate a target distribution using the function $f_{\vec{k^{t}}}(\vec{r})=A\sin(\vec{k^{t}}\cdot\vec{r}),$ (6) where $A$ takes continuous values between $0$ and $1$. We cannot generate a configuration which has exactly this pattern as a boson configuration can take only $\pm 1$ (filled/empty) at each site. Thus we perform a Monte Carlo simulated annealing starting from a charge-density wave state with $\vec{k^{t}}=k^{t}\hat{e}_{x}+k^{t}\hat{e}_{y}$ using an energy defined as $E=\sum_{l}(D_{l}-Lf_{\vec{k^{t}}}(\vec{r}))^{2}$, where the sum is over all diagonals and $D_{l}=\sum_{(i,j)\in l}\sigma_{i,j}$, is the occupancy in diagonal number $l$. This favors an exact match between the current and target configurations. By tuning the inverse temperature for the annealing from $\beta=0.01$ to $\beta=20.48$ in a geometric progression consisting of multiplying by $2$ every $10L^{2}$ steps, we achieve a stripe configuration which has a single Fourier component at the target $\vec{k^{t}}$. We ensure that the proposed configuration changes are long-ranged and respect the conservation laws discussed in the previous section, thus staying in the sector where every column and every row has exactly $L/2$ bosons. An example of a configuration created by this procedure is shown in Fig. 5. To check the effectiveness of the annealing procedure, we record the final energy and find it to be $O(L)$, implying that each $D_{l}$ is within an $O(1)$ value of its target value, which is expected as $D_{l}$ is an integer and the target value is in general a real number, and not necessarily close to an integer value. We find that this procedure generates a sharp peak for the desired $\vec{k^{t}}$ amid a weak background which decays with increasing system size. This can be seen in Fig. 5, where we plot the Fourier component as a function of the distance from the target peak location in Fourier space $k_{d}^{2}=(k_{x}-k^{t})^{2}+(k_{y}-k^{t})^{2}$ for various sizes. Due to the periodic boundary conditions in Fourier space, given by $(k_{x},k_{y})\leftrightarrow(1-k_{x},1-k_{y})$, we consider the shortest distance to the expected peak over the naive distance for open boundary conditions. Note that such a configuration still contains a large number of flippable plaquettes due to the smooth nature of the target pattern. Figure 6: Plots for $A=0.7$ : (a) Comparison of average scaled profile $B_{x}(t)$ from microscopic hard-core dynamics compared to continuum expression (Eq. (III.3)) for wavelength $\lambda=16$ for $t_{0}=16$. (b) $R(t)=n_{k}(t)/n_{k}(0)$ against $k^{4}t$ on log-linear scale, with long times fit to a single exponential. ### IV.2 Comparing with continuum field theory In the context of the continuum field theory discussed in Sec.IIIC, the field $g$ at time zero would now simply be equal to $A\sin(\vec{k}\cdot\vec{r})$ with $\vec{k}=\vec{k^{t}}$. It is essential to numerically verify the validity of the assumption used in developing our continuum theory, namely that spatial correlations do not play a role in the initial dynamics. Using this as the initial source field in Eq. (III.3), we find that ${\partial_{t}}g$ reduces to $Ak^{4}\sin(\vec{k}\cdot\vec{r})\Big{[}-\frac{1}{2}+4A^{2}\cos^{2}(\vec{k}\cdot\vec{r})-\frac{A^{2}}{2}\sin^{2}(\vec{k}\cdot\vec{r})\Big{]}.$ (7) This immediately indicates that we have lost linearity (which would imply no dependence on amplitude other than a global proportionality), and that the theory is no longer exactly separable in Fourier space. Note here that the terms proportional to $A^{2}$ within the brackets oppose the decay generated by the simple sub-diffusion, leading to a possibility of further slowing down the dynamics. The above equation also suggests that for $A\ll 1$, we should expect to recover vanilla sub-diffusive dynamics. For $\vec{k}\cdot\vec{r}\in(0,\pi)$, the sign is controlled by the expression within the brackets, and we can see that a growth of the function can be achieved for $\sin(\vec{k}\cdot\vec{r})<((8A^{2}-1)/(9A^{2}))^{1/2}$, provided $8A^{2}>1$ (that is large enough amplitude). This condition is satisfied for $\vec{k}\cdot\vec{r}$ in the vicinity of $0$ and $\pi$. This effect is rather non-intuitive as it implies that the local density tends away (towards $\pm 1$) from its equilibrium value (0) under non-equilibrium dynamics for a tunable range of $\vec{r}$, leading to a local reduction of entropy as the number of states available locally reduces if we require their average to be closer to the most extreme values which it can take. To ascertain the extent to which our microscopic dynamics is consistent with the continuum field theory developed under the mean field assumption, we can now compare the averaged value of $n_{x,y}$ against a numerical evolution of Eq. (III.3). Looking specifically for the feature described above, we plot $B_{x}(t)=\langle n_{x,x}(t)\rangle/\langle n_{x,x}(0)\rangle$ in Fig. 6a as a function of $x$ for the minimal period. Our initial condition already sets $n_{x,x}(0)=A\sin(2kx)$. We see a qualitative match to the prediction from the continuum equation (in the sense of the sign of the difference to unity of the ratio is well captured), but a quantitative disagreement (in the amplitude of this difference) potentially due to the build up of correlation effects beyond mean-field. Before we move to the study of the microscopic dynamics, it is worth considering the time scales over which the initial pattern melts under the continuum dynamics, as done in Sec. II. We show in Fig. 6b that the dynamics is still consistent with a single scaling variable, given by $k^{4}t$, although the relaxation deviates from the single exponential seen in the simpler cases considered in Sec. II. This implies that the additional non- linear terms in the continuum dynamics do not alter the scaling of space-time, and this can be intuitively understood by observing that all non-linear terms have the same order in derivatives. Figure 7: For “weak” amplitude $A=0.2$. (a) Dynamics of $R(t)$ as a function of $k^{4}t$, in a log-linear scale. (b) Zoom on initial dynamics, that shows a consistent scaling with $k^{4}t$ for sample initial conditions. We do not present an average over samples as it leads to large fluctuations over the time axis. Figure 8: For “large” values of amplitude $A$ : (a) Slow initial behavior of $R(t)$ as a function of time, for $A=0.5$, $k=0.125$ and a $L=256$ sample. To detect the beginning of the melting process, we define a threshold of 0.99 (shown by dotted line), and define $t_{0}$ as the intercept. (b) Average melting starting time $t_{0}$ as a function of $1/k$, for different values of $A$. The log-log scale emphasizes the power-law dependence. The black solid line corresponds to a $1/k^{4}$ power-law. ### IV.3 Exact numerical evolution: Beginning of melting We now study the dynamical behavior starting from a configuration generated by the method discussed above. To carry out an analysis of the evolution of the coarse-grained structure of the initial configuration, we choose to study the decay of the dominant Fourier component via the already defined ratio $R(t)=\frac{n_{\vec{k}}(t)}{n_{\vec{k}}(0)}$. To get an intuition about the effect of lattice spacing, we first consider the decay of $R(t)$ over a range of decreasing values of $k$, starting from $k=1/2$. As shown in Fig. 7a, for $A=0.2$, there are clear deviations away from a single exponential for large values of $k$, whereas intermediate values of $k$ appear to agree partially, and for small values of $k$, we once again deviate from the expected $k^{-4}$ scaling. To study the initial stages of the decay of the wave pattern, we can look at $R(t)$ in the regime where it is close to unity. Here we find a collapse consistent with a time scaling with $k^{-4}$ for a large range of $k$, including large $k$ values (as shown in Fig. 7b). This is quite surprising, because the regime of short-time scales and large-wave vectors is not the one where we expect the hydrodynamic prediction of Sec. III to hold. We have no simple explanation for this observation. For larger values of $A$, we observe a completely different behavior. We find that $R(t)\approx 1$ for a non-trivial amount of time after the initialization of the dynamics, an illustration of this behavior of $R(t)$ is given in Fig. 8 for $A=0.5$. We define the beginning of the melting process by the first time $t_{0}$ at which $R(t)$ crosses $0.99$. This threshold is chosen arbitrarily and a different threshold does not change the result qualitatively (as shown using a threshold of $0.9$ in the inset of Fig. 8b). We study this for a few values of $k$, and for $20$ realizations of the initial condition for each $k$. In addition to this, for each realization, we run sufficient number of realization of the stochastic dynamics to ensure that we get a good estimate for $t_{0}$. We have taken systems of linear size in the range of $150-250$, as the property of self-averaging allows us to narrow the spread in the values of $t_{0}$. We find a strong dependence of the averaged $t_{0}$ on the inverse wave-vector $1/k$. A linear fit on a log-log scale of $t_{0}$ vs $1/k$ reveals various power law regimes (see Fig. 8b) that depend on the amplitude $A$, the most extreme of which is achieved for $A=1$, where the dependence is $t_{0}\sim(1/k)^{18.1(3)}$. Such a strong dependence on the initial pattern suggests that the mechanism for melting is initiated by a coordinated rearrangement of bosons which is favorable for dynamics. Note that for all values of $A$, there exists a window of $1/2>k>k^{0}(A)$, where we find good agreement with $(1/k)^{4}$, as predicted from the continuum theory, with $k^{0}(A)$ reducing with reducing $A$. ### IV.4 Exact numerical evolution: Post-melt scaling The continuum theory developed in Sec. III, which can be valid only for small wave-vectors and long times, cannot describe the results in Fig. 8b and the period for which we observed $R(t)\approx 1$, and this suggests that it is not the appropriate theory to understand this “prethermal” behavior. We can check however if the continuum prediction is upheld after the melting process with a simple scaling collapse. We first define a time ${t^{\prime}}=0$, which is taken to be the well after the beginning of the melting process, as the time at which $R(t)$ drops below a threshold of $0.01$ (chosen just for the convenience of analysis, once again we checked that this value only plays a quantitative role). We can now attempt a parameter free scaling collapse of $R({t^{\prime}})$ and would expect to get a linear profile on a log-linear plot for a range of $k$ by scaling the ${t^{\prime}}\to k^{4}{t^{\prime}}$. We present this analysis in Fig. 9 for $A=1.0$ and find a reasonable agreement with the expectation developed above, even if not entirely satisfactory. To further check the applicability of the continuum theory, we also consider ring exchange dynamics over $2\times 1$ plaquettes, which we expect to the same Eq. (3), but have much short pre-thermal times (as naturally be expected from a longer-range type of exchange). This allows to probe a larger range of wave-vectors $k$ values, as presented in Fig. 9 where the adherence to a $k^{4}{t^{\prime}}$ is clearly improved. Figure 9: $R(t)$ scaled by the value closest to $0.01$ ($R(t_{b})$) in our dataset vs $k^{4}t^{\prime}$ where $t^{\prime}=t-t_{b}$ is the time shifted by the reference time $t_{b}$,for different values of $k$. A log-linear scale is chosen to emphasize the expected scaling in $k$ for different values of $k$. (a) : For the main ring-exchange model discussed in the manuscript. (b) Similar scaling for an extended model where we allow ring-exchanges models around $2\times 1$ and $1\times 2$ plaquettes. ## V Pattern of dynamical activity in prethermal and melting process Figure 10: (a) A sample initial configuration for an $80\times 80$ lattice with $k=0.05$ and $A=0.6$. Intersection of axes indicates origin. (b-d) Flippability correlator $C(x,y,t)$ for the same system size, averaged over $80$ initial conditions and $128$ realizations of the dynamics for each, at times (left to right) $t=2^{17},2^{19},2^{21}$. The previous section shows that a mean field treatment is unable to recreate the slow dynamics seen numerically, thus hinting at the presence of correlations which are neglected at a mean field level. To understand the large-scale mechanisms involved in the prolongation of the prethermal lifetime and the onset of melting, we consider the build up and correlations of flippable plaquettes. This serves as a proxy for identifying dynamically active regions and their evolution. We study real-space flippability correlations through the normalized connected correlator $C(x,y,t)=\frac{\braket{P(0,0,t)P(x,y,t)}-\braket{P(0,0,t)}\braket{P(x,y,t)}}{\braket{P(t)}^{2}},$ where $P(x,y,t)=1$ if the plaquette whose left bottom site is $(x,y)$ is flippable and zero otherwise, and $P(t)$ is the spatially averaged density of flippable plaquettes. Large (low) values of $C(x,y)$ at a given time will thus indicate strong (weak) correlations of flippability with the initial point. Recall that we have chosen starting configurations from annealing to a potential which is only a function of the tilted coordinate $x+y$. This means that we have a freedom of choosing the origin for our correlation function at any $x-y$ for fixed $x+y$. We take advantage of this symmetry by averaging over all equivalent positions of the origin. For the $x+y$ position of the origin we use our potential function as defined in Eq. (6) and set our origin to be the point satisfying $\vec{k}\cdot\vec{r}=\pi/4$. We make this choice as it lies at the threshold between highly active and inactive regions, defined by $\braket{n}=1/2$ (medium boson density, high density of flippable plaquettes) and $\braket{n}=1$ (high boson density, low density of flippable plaquettes). Figure 11: (a) A sample filled-staggered-unfilled-staggered type initial configuration, as described in the main text, with $W=13$ consecutive diagonals filled. (b) $R(t)$ at the dominant wave-vector for various values of $W$, against rescaled time $k^{4}t$. (c) Flippability correlator $C(x,y,t)$ for configuration in (1) at $t=2^{23}$ (d) $C(x,y,t)$ for $W=27$ at $t=2^{21}$ with the same conditions. (e) Same as plot for $W=27$, scaled instead by $\langle P_{d_{i}}\rangle\langle P_{d_{j}}\rangle$. ### V.1 Density wave patterns We begin by considering the initial conditions studied in the previous sections. To ensure that we are able to observe large scale (slowly varying spatial) features in the correlation function, we choose a system size of $80\times 80$ and $k=0.05$ at $A=0.6$ (a particular realization of this is shown in Fig. 10). To gather high quality statistics we average over 80 realizations for the initial condition and for each realization we run 128 realizations of the stochastic dynamics. The time $t_{0}$, which denotes the beginning of the melting process, is $\approx 10^{5}$ or $2^{17}$ for $k=0.05$, as seen in Fig. 8. We study $C(x,y,t)$ up to time scales of $2^{21}$ and find that strong anti-correlations develop at short range for the density of flippable plaquettes at early times, and sustain until the intermediate stages of melting. This is shown in Fig. 10, where snapshots of $C(x,y,t)$ at $t=2^{17}\to 2^{21}$ are presented. This profile shows the development of active regions (given in red) surrounded by inactive regions (blue), and suggests that a mechanism of excluded dynamical regions may be the source of dynamical slowing down. Note also that Fig. 10 suggests a bias of the dynamics towards the $x$ and $y$ axes, which may be expected from the continuum theory as well, due to the lack of radial symmetry. We see that even before the melting time, weak correlations are already built up along the $x$ and $y$ axes, but in a manner which is anti- correlated and modulated with the approximate stripe width. We find that this pattern begins to appear at times as short as $t=2^{4}=16$, but we present data at the last possible time value before melting which we have recorded as the pattern is visible with significantly more clarity. Another important aspect of the correlation pattern to note here is the dependence on the $x-y$ coordinate (“perpendicular” to the initial wave- pattern), which is not built in in the initial condition as the energy used in the annealing process used to generate the configuration has no $x-y$ dependence. This behavior cannot also be expected from the continuum theory since we start of with a single wave like configuration with $k_{x}=k_{y}$. We find that an important condition for the existence of such correlation patterns is that they are specific values of $A$ where the dynamics is not described by simple sub-diffusion. For example, we studied the case of $k=0.05$ at $A=0.2$, which shows conventional dynamics (as far as we can see from Fig. 8), and did not find any non-trivial correlations down to a precision of $10^{-4}$ in $C(x,y,t)$. This is as expected from a mean-field treatment, as it forbids correlations by definition. The evolution of the correlation landscape for the duration of the melting process presented in Fig. 10 suggests that it proceeds through a merging of dynamically active regions. This will be discussed in more detail for step- like initial conditions, where a complete melting process can be observed with higher clarity and on longer times. ### V.2 Maximally active square wave initial conditions To gain a better understanding of the melting process, we begin with a more artificial initial condition chosen to have regions of flippable plaquettes with maximal and minimal density. We consider alternating patterns such as the one shown in Fig. 11, with $W$ (necessarily odd) consecutive diagonals filled, followed by $W+1$ in a perfect staggered (“Néel”) pattern, followed by $W$ empty diagonals, and finally by another $W+1$ in a Néel pattern to close out one period. These constraints are chosen to ensure that a filled region is bounded by empty diagonals and vice versa. The lattice size is then $N(2W+2(W+1))$, where $N$ is the number of periods. Although these configurations show melting times (quantified by the decay of the appropriate Fourier component) which are faster than those studied above, they show a scaling fore pre-melting which is slower than the conventional $k^{-4}$, as seen in Fig.11. We find that the flippability correlator $C(x,y,t)$ for these unconventional initial conditions shows an extension of the anti-correlated pattern in the $x-y$ direction with a periodic modulation across distances large compared to lattice spacing. This is seen clearly for $W=13$ with $N=2$ at time of $t=2^{23}=8\times 10^{6}$ in Fig. 11. Figure 12: (a) $R(t)$ for $W=7$ and 2 stripes with vertical lines marking the times (b-f) $[4,6,8,11,14]\times 2^{20}$ for which $C(x,y,t)$ profiles are collected. An important characteristic of the pattern described above is the value of the wave-vector associated to this pattern. We find that this is determined by the approximate stripe width, i.e. the wavelength $1/k$. Our data suggest that this occurs due to a development of the correlation along the $x$ and $y$ axes, which is limited by the size of active regions (high density of flippable plaquettes). Growth of correlation within an active region is mediated by alternating patterns of correlation and anti-correlation, which can be thought of as being generated by reflections off of the boundary between active and inactive regions, as shown in Fig. 11. Repeated reflections create the observed periodic pattern, thus linking the periodicity and the width of the initial stripes. This is clearly seen in the approximate periodicities of the pattern for $W=13$ and $27$ in Fig. 11 where the profiles look similar even though the stripe width is changed by a factor of 2. Strong signatures of the periodic pattern for stripe widths (or alternatively wavelengths) of $13$ and $27$ at approximately the same time indicates the emergence of this pattern at a time scale which only weakly depends on the wave-vector $k$, and that it develops at a time scale parametrically much smaller than the melting time. Note that an extrapolation based on Fig.11b indicates the beginning of melting to occur earliest at $t\approx 2^{27}$, assuming a lower bound of the scaling behavior as $k^{-4}$ (which is clearly slower than the scaling our numerics show in Fig.11b). To develop an estimate of the relevance of this patterning to the process of melting, we first remove the normalization of $\braket{P(t)}$ which we have absorbed into the definition and instead look at $C(x,y,t)$ scaled by $\langle P_{d_{i}}(t)\rangle\langle P_{d_{j}}(t)\rangle$, where $\braket{P_{d_{i}}(t)}$ is the average number of flippable plaquettes in diagonal number $i$ at time $t$. This makes the correlation function effectively lie between $-1$ and $1$, with either limit describing a saturation to the largest (smallest) values. In particular, a value of $-1$ implies that the density of flippable plaquettes is zero, leading to a complete arrest of the dynamics. This version of $C(x,y,t)$ is plotted in Fig. 11e and we see clearly that close to the frontier of dynamical activity, we do get a complete anti-correlation around the reference region. This suggests that the dynamical activity is likely strongly controlled by the unconventional pattern in the $x-y$ direction. Now we turn to an investigation of the role of the periodic patterns observed above in the melting process. We can study this by choosing a small enough $W$ so that the melting process is captured within the time scale of $10^{9}$ which we can simulate. We find that for $W=7$, the ratio $R(t)$ begins to deviate from unity at a time $t_{0}\sim 2^{21}=2\times 10^{6}$, and reaches a value of 0.01 by $t_{b}\sim 2^{24}=1.6\times 10^{7}$ (data plotted in Fig. 12). In this interval, we find at time $t=2^{20}$ a formation of stable periodic structures in the $x-y$ direction, and see a connection of the dynamically active regions for $t=2^{22}$. Finally by $t=2^{24}$, the correlation pattern has relaxed into a single wave in the $x+y$ direction, restoring the symmetry in the $x-y$ direction expected from the continuum prediction. The role of correlations is expected to be insignificant past this time point. ## VI Summary and outlook We studied a system of hard-core bosons on a square lattice evolving under classical stochastic dynamics using ring exchanges. We find that boson density waves motivated by the patterns present in the Hilbert space fragmentation of this model approach equilibrium on an extremely long time scale, which diverges with the inverse wave-vector as $k^{-\alpha}$ for small $k$, where $\alpha$ depends on the amplitude $A$ of the initial density wave configuration. For decreasing values of $A$, we find increasingly larger windows in the large $k$ regime where we observe a scaling of $k^{-4}$. Our numerical studies and mean-field treatments of simplified models show the emergence of sub-diffusion on the other-hand in the hydrodynamic low $k$ limit. For the hard-core case, we derive a dynamical equation using a mean- field like assumption which is able to capture partially the unconventional feature of an increase in slope around the nodes of the pattern driven by the non-linear terms in the dynamical equation but which also predicts a time scaling of $k^{-4}$. This leads us to the conclusion that the microscopic dynamics are controlled by strong correlations which are built up in the prethermal regime, which cannot be accounted for by this mean-field treatment. To understand the mechanism driving the long prethermal regime, we study the correlation function of flippable plaquette density for different instances in time. We find that a strong anti-correlation is built up in the direction orthogonal to the modulation direction for the initial condition, even extending to the entire length of the diagonals for intermediate system sizes. We thus observe the formation of isolated dynamical puddles and find that the mechanism leading to the melting is characterized by a merger of these puddles, which relax into a profile which is consistent with what we expect from the continuum theory. Thus, we have found an example where correlations control crucial aspects of the dynamical behavior, and a simple mean-field like hydrodynamic description is insufficient. An important question remains: how generic are the long melting processes that we observe, in particular to which additional classes of initial conditions do they apply? We have shown that the decay of the overlap with the initial state with time shows a similar behavior upon comparing the exact quantum dynamics and our classical automaton for a small system size. This leads us to speculate that the behavior studied in the automaton language might inherit properties which can also be relevant for the exact quantum evolution for long times, as already seen for the study of sub-diffusive behavior in 1D Feldmeier _et al._ (2020). First of all, we expect that our hydrodynamic description might be equally valid also for the quantum dynamics, in case the quantum system exhibits a hydrodynamic long-time behavior. This would imply sub-diffusion also in the quantum case. Further, it might be possible that the onset of hydrodynamic behavior also in the quantum model might be delayed to very long times. This would have the consequence that a large intermediate time window exists with unconventionally slow dynamics. As this model is related to the quantum link model Karpov _et al._ (2021) via a relaxation of charge conservationKhudorozhkov _et al._ (2021), similar physics may be expected to emerge there in a more restricted setting. The crossover between Hilbert space fragmentation, which dictates preservation of the memory of the initial state to arbitrarily long times, and the long prethermal plateaus we see here, is another promising direction in which investigations can be carried out to better our understanding of the processes involved. Finally, the rapid increase of the lifetime of the prethermal behavior with inverse wave-vector suggests that rare events, which create configurations allowing the applicability of the hydrodynamic theory, play a key role. An improved understanding of the potentially large deviations which lead to the above mentioned phenomenon would help greatly in developing a coarse-grained description of the dynamics at the threshold between the prethermal and equilibrium regimes. ###### Acknowledgements. We would like to thank Clément Sire, Kedar Damle and Frank Pollmann for fruitful discussions. Computational resources for this project were provided by LPT Toulouse and MPIPKS. This project has received funding from the European Research Council (ERC) under the European Unions Horizon 2020 research and innovation program (grant agreement No. 853443). ## Appendix A Low energy behavior of random configurations Figure 13: $C(\vec{k},t)$ normalized by $C(\vec{0},t)$ for times $t=2^{12}$ (a) and $2^{16}$ (b). Here we follow the analysis developed in Ref. Sala _et al._ (2021) for a similar stochastic model. The authors consider a classical automaton on a square lattice which also conserves the dipole moment similar to the model we have considered. The microscopic dynamics is considered from the perspective of the long time correlation function, defined as $C(\vec{r},t)=\braket{n_{\vec{r}}(t)n_{0}(0)}$, where the expectation value is over random initial conditions and realizations of stochastic dynamics. The authors of Ref. Sala _et al._ (2021) estimate the correlation function in momentum space ($C(\vec{k},t)$) for a time much larger than the scale of the microscopic dynamics (using both automaton dynamics and an effective analytical ansatz) and find sub-diffusive features with “hidden” modulated symmetries corresponding to certain patterns in the Brillouin zone (see Fig.2b of Ref. Sala _et al._ (2021)). In this appendix, we would like to check if this analysis can help identify the slow dynamics we observe for large scale modulated patterns. The energy spectrum as a function of $\vec{k}$ has already been identified in a mean- field framework by Paramekanti et al. in Ref. Paramekanti _et al._ (2002) for the form of ring-exchange used in our work, to be given by $E_{\vec{k}}\propto|\sin(k_{x}/2)\sin(k_{y}/2)|$. This trivially implies that the lines $k_{x}=0$ and $k_{y}=0$ host zero modes, and that this should be visible in $C(\vec{k},t)$ in the long time limit. Note that this profile does not explain the slow dynamics which we have explored in this work, as that would at the minimum require the presence of slow modes along tilted directions such as $k_{x}=k_{y}$. To check if this expectation is borne out by ring-exchange dynamics, we perform numerical evolutions of random configurations of a $200\times 200$ lattice for $t=2^{12}$ and $2^{16}$, shown in Fig. 13. We can clearly see that the profile is consistent with non-zero values being present only along the lines $k_{x}=0$ and $k_{y}=0$ at long times. This aspect of the dynamics is consistent with the mean-field treatment developed in the body of the manuscript. This is to be expected since random configurations correspond to the $A\to 0$ limit of our study of modulated patterns, which was shown to be well-described by the “vanilla” mean-field dynamics. ## References * Calabrese (2018) P. Calabrese, Physica A: Statistical Mechanics and its Applications 504, 31 (2018). * Garrahan (2018) J. P. Garrahan, Physica A: Statistical Mechanics and its Applications 504, 130 (2018). * Gogolin and Eisert (2016) C. Gogolin and J. Eisert, Reports on Progress in Physics 79, 056001 (2016). * Browaeys and Lahaye (2020) A. Browaeys and T. Lahaye, Nature Physics 16, 132 (2020). * Gross and Bloch (2017) C. Gross and I. Bloch, Science 357, 995 (2017). * Abanin _et al._ (2019) D. A. Abanin, E. Altman, I. Bloch, and M. Serbyn, Rev. Mod. Phys. 91, 021001 (2019). * Alet and Laflorencie (2018) F. Alet and N. Laflorencie, Comptes Rendus Physique 19, 498 (2018). * Abanin and Papić (2017) D. A. Abanin and Z. Papić, Annalen der Physik 529, 1700169 (2017). * Luitz _et al._ (2015) D. J. Luitz, N. Laflorencie, and F. Alet, Physical Review B 91, 081103 (2015). * Sala _et al._ (2020) P. Sala, T. Rakovszky, R. Verresen, M. Knap, and F. Pollmann, Physical Review X 10, 011047 (2020). * Khemani _et al._ (2020) V. Khemani, M. Hermele, and R. Nandkishore, Phys. Rev. B 101, 174204 (2020). * Moudgalya _et al._ (2021) S. Moudgalya, B. A. Bernevig, and N. Regnault, arXiv preprint arXiv:2109.00548 (2021). * Yang _et al._ (2020) Z.-C. Yang, F. Liu, A. V. Gorshkov, and T. Iadecola, Physical review letters 124, 207602 (2020). * Mukherjee _et al._ (2021) B. Mukherjee, D. Banerjee, K. Sengupta, and A. Sen, Physical Review B 104, 155117 (2021). * Karpov _et al._ (2021) P. Karpov, R. Verdel, Y.-P. Huang, M. Schmitt, and M. Heyl, Physical Review Letters 126, 130401 (2021). * Zhang and Røising (2022) Z. Zhang and H. S. Røising, arXiv preprint arXiv:2206.01758 (2022). * Sikora _et al._ (2011) O. Sikora, N. Shannon, F. Pollmann, K. Penc, and P. Fulde, Physical Review B 84, 115129 (2011). * Ilievski _et al._ (2018) E. Ilievski, J. De Nardis, M. Medenjak, and T. Prosen, Physical review letters 121, 230602 (2018). * McRoberts _et al._ (2022) A. J. McRoberts, T. Bilitewski, M. Haque, and R. Moessner, Physical Review B 105, L100403 (2022). * Scheie _et al._ (2021) A. Scheie, N. Sherman, M. Dupont, S. Nagler, M. Stone, G. Granroth, J. Moore, and D. Tennant, Nature Physics 17, 726 (2021). * Weiner _et al._ (2020) F. Weiner, P. Schmitteckert, S. Bera, and F. Evers, Physical Review B 101, 045115 (2020). * Kardar _et al._ (1986) M. Kardar, G. Parisi, and Y.-C. Zhang, Physical Review Letters 56, 889 (1986). * Iaconis _et al._ (2019) J. Iaconis, S. Vijay, and R. Nandkishore, Physical Review B 100, 214301 (2019). * Feldmeier _et al._ (2020) J. Feldmeier, P. Sala, G. De Tomasi, F. Pollmann, and M. Knap, Physical Review Letters 125, 245303 (2020). * Morningstar _et al._ (2020) A. Morningstar, V. Khemani, and D. A. Huse, Physical Review B 101, 214205 (2020). * Iaconis _et al._ (2021) J. Iaconis, A. Lucas, and R. Nandkishore, Physical Review E 103, 022142 (2021). * Richter and Pal (2022) J. Richter and A. Pal, Physical Review Research 4, L012003 (2022). * Gromov _et al._ (2020) A. Gromov, A. Lucas, and R. M. Nandkishore, Physical Review Research 2, 033124 (2020). * Nandkishore and Hermele (2019) R. M. Nandkishore and M. Hermele, Annual Review of Condensed Matter Physics 10, 295 (2019). * Pretko _et al._ (2020) M. Pretko, X. Chen, and Y. You, International Journal of Modern Physics A 35, 2030003 (2020). * Guardado-Sanchez _et al._ (2020) E. Guardado-Sanchez, A. Morningstar, B. M. Spar, P. T. Brown, D. A. Huse, and W. S. Bakr, Phys. Rev. X 10, 011042 (2020). * Scherg _et al._ (2021) S. Scherg, T. Kohlert, P. Sala, F. Pollmann, B. Hebbe Madhusudhana, I. Bloch, and M. Aidelsburger, Nature Communications 12, 4490 (2021). * Kohlert _et al._ (2021) T. Kohlert, S. Scherg, P. Sala, F. Pollmann, B. Hebbe Madhusudhana, I. Bloch, and M. Aidelsburger, arXiv e-prints , arXiv:2106.15586 (2021), arXiv:2106.15586 . * Baiardi and Reiher (2019) A. Baiardi and M. Reiher, Journal of chemical theory and computation 15, 3481 (2019). * Schmitt and Heyl (2020) M. Schmitt and M. Heyl, Physical Review Letters 125, 100503 (2020). * Iadecola and Vijay (2020) T. Iadecola and S. Vijay, Physical Review B 102, 180302 (2020). * Gopalakrishnan and Zakirov (2018) S. Gopalakrishnan and B. Zakirov, Quantum Science and Technology 3, 044004 (2018). * Iaconis (2021) J. Iaconis, PRX Quantum 2, 010329 (2021). * Paramekanti _et al._ (2002) A. Paramekanti, L. Balents, and M. P. Fisher, Physical Review B 66, 054526 (2002). * Rousseau _et al._ (2005) V. Rousseau, R. Scalettar, and G. Batrouni, Physical Review B 72, 054524 (2005). * Schaffer _et al._ (2009) R. Schaffer, A. A. Burkov, and R. G. Melko, Physical Review B 80, 014503 (2009). * Tay and Motrunich (2010) T. Tay and O. I. Motrunich, Physical review letters 105, 187202 (2010). * Tay _et al._ (2011) T. Tay, O. I. Motrunich, _et al._ , Physical Review B 83, 205107 (2011). * Khudorozhkov _et al._ (2021) A. Khudorozhkov, A. Tiwari, C. Chamon, and T. Neupert, arXiv preprint arXiv:2107.09690 (2021). * Sala _et al._ (2021) P. Sala, J. Lehmann, T. Rakovszky, and F. Pollmann, arXiv preprint arXiv:2110.08302 (2021). * Canfield and McKay (2005) E. Canfield and B. McKay, Electronic Journal Of Combinatorics 12 (2005).
# NOPE-SAC: Neural One-Plane RANSAC for Sparse-View Planar 3D Reconstruction Bin Tan, Nan Xue, Tianfu Wu, Gui-Song Xia B. Tan, N. Xue and G.-S. Xia are with the School of Computer Science, Wuhan University, Wuhan, China, 430072. T. Wu is with the Department of Electrical and Computer Engineering, North Carolina State University, Raleigh, NC, USA, 27606.(Corresponding author: Nan Xue.) ###### Abstract This paper studies the challenging two-view 3D reconstruction in a rigorous sparse-view configuration, which is suffering from insufficient correspondences in the input image pairs for camera pose estimation. We present a novel Neural One-PlanE RANSAC framework (termed NOPE-SAC in short) that exerts excellent capability to learn one-plane pose hypotheses from 3D plane correspondences. Building on the top of a siamese plane detection network, our NOPE-SAC first generates putative plane correspondences with a coarse initial pose. It then feeds the learned 3D plane parameters of correspondences into shared MLPs to estimate the one-plane camera pose hypotheses, which are subsequently reweighed in a RANSAC manner to obtain the final camera pose. Because the neural one-plane pose minimizes the number of plane correspondences for adaptive pose hypotheses generation, it enables stable pose voting and reliable pose refinement in a few plane correspondences for the sparse-view inputs. In the experiments, we demonstrate that our NOPE- SAC significantly improves the camera pose estimation for the two-view inputs with severe viewpoint changes, setting several new state-of-the-art performances on two challenging benchmarks, i.e., MatterPort3D and ScanNet, for sparse-view 3D reconstruction. The source code is released at https://github.com/IceTTTb/NopeSAC for reproducible research. ###### Index Terms: Planar 3D Reconstruction, Two-view Camera Pose Estimation, Sparse-view 3D Reconstruction, Deep Learning. ## 1 Introduction Two-view 3D reconstruction is a fundamental and longstanding problem in computer vision, which is usually formulated to recover the relative camera pose and the scene geometry by feature correspondences [1]. Although such a classic formulation has been well established with keypoints in conventional Structure-from-Motion systems [2, 3, 4, 5], they are challenged by severe viewpoint changes and low-texture appearance of the captured images in the indoor scenes as shown in Fig. 1(a). In this paper, we are interested in addressing such a challenging configuration for two-view pose estimation and scene reconstruction, which is named as _sparse-view_ 3D reconstruction in literature [6, 7, 8]. (a) SuperGlue Keypoint Correspondences[9] (b) Plane Correspondences (c) Camera Poses and Planar Reconstruction Figure 1: An illustration of the indoor sparse-view 3D reconstruction. (a) shows the keypoint correspondences of SuperGlue [9] including only 4 inliers (Green lines) and 8 outliers (Blue lines). (b) and (c) present the plane correspondences and reconstructed 3D planar scene of our NOPE-SAC from two sparse-view images. The Black frustum shows the ground truth camera of the second image. The Pink, Gray, and Blue frustums show the initial, the one- plane hypothetical, and the final cameras of the first image estimated by our NOPE-SAC. The _sparse-view images_ are often resulted by large viewpoint changes up to scene scale in image capturing, which manifests as a low overlapping rate between the images. For instance, as shown in Fig. 1(a), even though adopting the state-of-the-art SuperGlue [9] for keypoint correspondences in the sparse- view image pair, it is infeasible for the well-known 5-point algorithm [10] because there are only $4$ inliers among the $12$ putative correspondences. However, when we carefully check the input sparse-view images from pixel level to object level with a focus on the wall, floor, and door, in our mind, these two images can be aligned in a unified 3D space. The pioneering works [7, 8] justified such an observation in learning-based reconstruction systems by relying on 3D plane correspondences, in which they formulate the challenging relative pose estimation problem in a voting form in accordance with the geometric correctness of plane correspondences. Voting the pose hypotheses by putative 3D plane correspondences in [7, 8] shares a similar spirit with the keypoint-based RANSAC framework [11] for pose estimation, but they use precomputed pose hypotheses instead of computing them from the inlier correspondences. To ensure that the precomputed pose hypotheses can cope with complicated pose distribution, the number of hypotheses would be large (_e.g., 1024 in SparsePlane[7]_), which induced a contradiction between a large number of hypotheses and the small number of plane correspondences. For example, there are only 4 plane correspondences in Fig. 1(b) for hypotheses voting, thus being unstable and unreliable to distinguish the best hypothesis from the 1024 ones. To this end, a time- consuming continuous optimization with the assistance of keypoint correspondences is required [7] for accurate sparse-view 3D reconstruction. Although they achieved impressive reconstruction results, from the perspective of RANSAC [11], they ignored a key problem of pose hypotheses generation and are thus inflexible and computationally expensive for sparse-view 3D reconstruction. In this paper, we pursue to formulate the camera pose estimation with pure plane correspondences in RANSAC to get rid of the fixed pose hypotheses. Considering the fact that explicitly modeling the relationship between the plane correspondences and the relative camera pose in the sparse-view setting would lead to ill-posed minimal problems [12], we leverage neural networks to encode the parameters of plane correspondences into a pose embedding space and learn to generate the pose hypotheses. We minimize the number of plane correspondences to 1 for the pose hypothesis generation, namely one-plane pose, and thus propose our Neural One-PlanE RANSAC (NOPE-SAC) framework. With the one-plane pose generation, the number of pose hypotheses is significantly reduced to be the same as the number of plane correspondences. Consequently, the contradiction between the number of hypotheses and those of the supported inlier plane correspondences is largely alleviated, which makes it possible to fully utilize the information of plane correspondences for recovering accurate camera pose from sparse-view images without using post-processing steps such as the continuous optimization in [7]. We build our NOPE-SAC together with the plane detection and matching modules to establish a complete planar reconstruction system. More precisely, given two sparse-view images, our NOPE-SAC first detects 3D planes on each image and makes the plane correspondences by solving an optimal transport problem as in [9]. To start the pose estimation in RANSAC with a coarse initial camera pose, we leverage a convolutional neural network (ConvNet) for the learning of pose embedding which is then used to decode the initial pose. Then, the parameters of each plane correspondence are encoded together with the initial pose embedding to form a new pose embedding for the generation of one-plane pose hypotheses. Due to the flexibility issue of pose initialization from ConvNets, we present an arbitrary pose initialization module (AIM) that follows the auto-encoder structure to encode poses from other methods into embeddings. Next, with a coarse initial camera pose and $N$ putative plane correspondences, our NOPE-SAC generates $N$ one-plane pose hypotheses, which are then voted by the geometric cost of the plane correspondences. Finally, the refined pose is estimated by fusing all hypotheses together with voting scores, and the holistic planar reconstruction is approached as the final output. Fig. 1 presents an illustrative example of our NOPE-SAC for the sparse-view input images. For the four plane correspondences predicted by the siamese network, our NOPE-SAC yields an initial pose (with the pink frustum) and four one-plane pose hypotheses (with the gray frustums), and then use the plane correspondences to vote and fuse the hypotheses (including the initial pose) for achieving a holistic 3D reconstruction. In the experiments, we evaluate our NOPE-SAC on two indoor benchmark datasets, i.e., Matterport3D [13] with sparse-view splits created by SparsePlanes [7] and ScanNet [14] with a more challenging split created by ourselves (see Sec. 5.1 for details). On both benchmarks, our NOPE-SAC achieves state-of-the-art performance in terms of pose estimation accuracy and planar 3D reconstruction precision. Compared to the prior arts (_e.g.,_ PlaneFormers [8]), our NOPE-SAC pushed the accuracy of pose estimation on the Matterport3D dataset to $73.2\%$ and $89.0\%$ for translation and rotation ($6.4\%$ and $5.2\%$ absolute improvements), the accuracy of pose estimation on the ScanNet dataset to $82.0\%$ and $82.6\%$ for translation and rotation ($6.7\%$ and $9.4\%$ absolute improvements), and the Average Precision (AP) of 3D plane reconstruction to $43.29\%$ on the Matterport3D and $39.39\%$ on the ScanNet ($5.76\%$ and $4.75\%$ absolute improvements). The comprehensive ablation studies further verified the design rationales of the proposed NOPE-SAC. In summary, in this article, we present a novel approach, i.e., NOPE-SAC, to address the challenging problem of sparse-view planar 3D reconstruction in a RANSAC framework, which fully takes the advantage of end-to-end learning of deep neural networks. By our NOPE-SAC, the neural one-plane pose hypotheses alleviate the contradiction between the number of the pose hypotheses and the 3D plane correspondences, thus improving the accuracy of camera pose estimation without incurring any offline optimization procedures. In performance, our method sets several new state-of-the-art performances on both the Matterport3D [13] and the ScanNet [14] datasets for pose estimation and holistic planar reconstruction. The comprehensive experimental analyzes also aligned benchmarking protocols for sparse-view 3D reconstruction. ## 2 Related Work ### 2.1 Single-View 3D Reconstruction One relevant task to indoor 3D reconstruction is to recover 3D scenes from single images. As one of the most widely used solutions, single-view depth estimation has been extensively studied [15, 16, 17, 18, 19] to pixel-wisely predict the depth values of the input images. Benefitting from the advances of deep learning and the richness of the training data for depth estimation [20, 21], we have witnessed their significant improvements in estimation accuracy and generalization ability. Because the single-view depth estimation can only yield unstructured 3D point clouds, they would bring structural distortions into predictions for structured scenes, _e.g.,_ the indoor scenes. To this end, some researchers propose to directly predict 3D planes from a single image to reconstruct the geometry of structured scene [22, 23, 24, 25]. For example, Liu et al. [23] applied a two-stage instance segmentation framework to jointly detect plane instance masks and estimate 3D plane parameters for the single view planar reconstruction. Although these insightful methods work well for single-view indoor plane reconstruction, they can not recover the holistic scene because of the limited field of view in every single-view image. In this paper, we take the merits of single-view planar 3D reconstruction and go further to the challenging sparse two-view configurations. We show that the estimated 3D planes from single-view images are favorable for both camera pose estimation and planar 3D reconstruction from the low-overlapping two-view images. ### 2.2 Two-View Camera Pose Estimation Two-view 3D reconstruction is the most fundamental task in computer vision, which was formulated to solve the relative camera poses between the input images and estimate the scene geometry from the camera motions. The problem of camera pose estimation is the core of this task. A common solution is to estimate camera poses from keypoint correspondences [26, 27, 28, 29, 30] relying on a typical 5-point solver [10] with a RANSAC [11] framework. Following this paradigm, there have been tremendous efforts on improving the performance of keypoint detection and matching by neural networks [31, 32, 33, 34, 9, 35]. As our study mainly focuses on the sparse- view configuration for two-view 3D reconstruction, such a common solution would be infeasible due to the fact of low-overlapping rate between the sparse-view inputs for feature correspondences. Recently, some approaches leverage neural networks to directly estimate camera poses from feature cost volumes built up on dense pixel correspondences [36, 37, 38, 39]. Besides, Wei et al. [36] combined the traditional and learning- based methods by estimating dense pixel correspondences with a neural network and retrieving camera poses with the 5-point solver [10]. Although these learning-based methods have shown their excellent performances, they largely depend on sufficient image overlap to achieve motion cues from correspondences and thus are challenged by sparse-view indoor images with small overlapping rates. Most recently, the problem of pose estimation from sparse-view indoor images has been extensively studied [6, 7, 8]. Due to the challenges of obtaining a sufficient number of inlier correspondences, these approaches use a large number of a-priori poses (by clustering the ground-truth camera poses) as the label space and formulate the problem of pose estimation to label scoring with respect to classification likelihoods and plane correspondences. Because a large number of a-priori poses would affect the flexibility and their estimation accuracy, these approaches have to adopt an optimization scheme for estimation refinement. Compared to those approaches, we show that although the number of 3D plane correspondences is very limited in sparse-view configuration, it works very well in a consensus sampling pipeline by generating the one-plane pose hypotheses and voting to obtain the accurate estimation of camera poses in end-to-end learning. Figure 2: Overview of the proposed NOPE-SAC. Our network first detects 3D planes (Sec. 3.1) and estimates plane correspondences (Sec. 3.2) from the input sparse-view images. Then, the final relative camera pose of the first image (Blue frustum) is voted from a coarse initial pose (Pink frustum) and a few one-plane poses (Gray frustum) according to the geometric cost of plane correspondences (Sec. 4). At last, the planar 3D reconstruction is achieved as the final output. The Black frustum shows the camera of the second image. ## 3 Planar Correspondences Preparation The overview of our proposed Nope-SAC for planar 3D reconstruction is illustrated in Fig. 2, which detects 3D planes from the two-view images and establishes the plane correspondences between views for camera pose estimation and planar 3D reconstruction. We focus on 3D plane detection and matching in this section and leave the key components for pose estimation in Sec. 4. With the estimated relative camera pose, the scene reconstruction is finally achieved by aligning the 3D plane correspondences between viewpoints. ### 3.1 3D Plane Instance Detection Similar to [7], we define a plane instance as $\pi=\left\\{\mathbf{n},d,\mathcal{M},\text{emb}\right\\}$, where $\mathbf{n}\in\mathbb{R}^{3}$ and $d$ are the plane normal and the offset from the plane to the camera center, $\mathcal{M}\in\mathbb{R}^{H\times W}$ is the plane segmentation mask ($H$ and $W$ are the image height and width, respectively.) and $\text{emb}\in\mathbb{R}^{256}$ is a plane appearance embedding. We use the recent PlaneTR [25] as the 3D plane detection module for each input view with two main modifications: (1) the line segmentation branch in PlaneTR [25] is excluded to keep the simplicity; (2) the backbone network of PlaneTR [25] is replaced to ResNet-50 [40] for efficiency. In computation, we use the output feature of the Transformer decoder in [25] as the plane appearance embedding and predict the normal and offset of each plane on top of the plane appearance embedding with a linear layer. The 3D plane detection module is supervised as: $\mathcal{L}_{\text{plane}}=\mathcal{L}_{\text{cls}}+20\mathcal{L}_{\text{mask}}+\mathcal{L}_{\text{dice}}+\mathcal{L}_{\text{parm}}+0.5\mathcal{L}_{\text{center}},$ (1) where $\mathcal{L}_{cls}$, $\mathcal{L}_{mask}$, $\mathcal{L}_{dice}$, $\mathcal{L}_{parm}$ and $\mathcal{L}_{center}$ are the classification loss, the mask loss, the dice loss [41], the plane parameter loss and the plane center loss as used in [25, 42]. We refer readers to [25, 42] for more details. ### 3.2 Plane Matching Denoted by $\Pi_{1}={\\{\pi_{i}^{(1)}\\}}_{i=1}^{K_{1}}$ and $\Pi_{2}={\\{\pi_{j}^{(2)}\\}}_{j=1}^{K_{2}}$ the plane sets from two view images, we aim to find an optimal partial assignment $A\in[0,1]^{K_{1}\times K_{2}}$ from $\Pi_{1}$ to $\Pi_{2}$ for plane matching, where $K_{1}$ and $K_{2}$ are the number of plane instances in two images. The optimal partial assignment $A$ is solved as an Optimal Transport (OT) problem based on the scoring matrix $S\in\mathbb{R}^{K_{1}\times K_{2}}$. #### Scoring Matrix. Given the plane sets $\Pi_{1}$ and $\Pi_{2}$, we compute the affinity matrices in terms of appearance and geometry and linearly add them as the final scoring matrix. For the affinity of plane appearance, we apply an Attentional Graph Neural Network used in [9] on the plane appearance embeddings to obtain two encoded sequence ${E_{1}}\in\mathbb{R}^{K_{1}\times 256}$ and ${E_{2}}\in\mathbb{R}^{K_{2}\times 256}$. Then, the appearance score matrix is computed by $S_{e}={E_{1}}{E_{2}}^{T}.$ (2) The geometric affinity is defined by the plane parameters between two 3D planes. Given an initial relative camera pose $R,\mathbf{t}$ (see Sec. 4.1 for details), we first convert the 3D plane parameters of $\Pi_{1}$ to ${\widetilde{\Pi}}_{1}$, the warped counterpart of $\Pi_{1}$ under the coordinate frame of $\Pi_{2}$, and then calculate the geometric affinity matrix by: $S_{g}(i,j)=-\lambda_{1}\text{acos}(\tilde{n}_{i}^{(1)},{n}_{j}^{(2)})-\lambda_{2}|\tilde{d}_{i}^{(1)}-d_{j}^{(2)}|,$ (3) where $\tilde{n}_{i}^{(1)}$ and $\tilde{d}_{i}^{(1)}$ are the $i^{th}$ plane normal and offset in ${\widetilde{\Pi}}_{1}$. ${n}_{j}^{(2)}$ and $d_{j}^{(2)}$ are the $j^{th}$ plane parameters in $\Pi_{2}$. $\lambda_{1}$ and $\lambda_{2}$ are set to 0.125 and 0.25 to balance the magnitude of the two terms. Finally, the scoring matrix $S$ can be calculated as: $S=S_{e}+S_{g}$ (4) #### Optimal Matching. Given a scoring matrix $S$, we apply a differentiable Sinkhorn Algorithm [43, 44] to output a soft assignment matrix $A$. Then, for a pair of planes $\\{\pi_{i}^{(1)},\pi_{j}^{(2)}\\}$, they can be regarded as matched planes if $A(i,j)$ is the maximal score both in the $i^{th}$ row and the $j^{th}$ column of $A$, and $A(i,j)$ is larger than a fixed matching threshold (0.2 in this paper). We supervise the plane matching module with the loss defined in Eqn. (5). More precisely, let $\bar{A}\in\mathbb{R}^{(K_{1}+1)\times(K_{2}+1)}$ to be the soft assignment matrix augmented with dustbins and the expected soft assignment matrix $A=\bar{A}_{1:K_{1},1:K_{2}}$, the plane matching loss can be calculated as: $\displaystyle\mathcal{L}_{\text{match}}=$ $\displaystyle-\sum_{(i,j)\in\mathcal{A}}\text{log}\bar{A}_{i,j}-\sum_{i\in\mathcal{I}}\text{log}\bar{A}_{i,K_{2}+1}$ (5) $\displaystyle-\sum_{j\in\mathcal{J}}\text{log}\bar{A}_{K_{1}+1,j},$ where, $\mathcal{A}=\\{(i,j)\\}$ are the indices of ground truth matches. $\mathcal{I}$ and $\mathcal{J}$ are the indices of unmatched planes in two images respectively. #### Correspondences Preparation for Pose Estimation. After the computation of optimal matching by the Sinkhorn layer, we extract a total number of $M$ plane correspondences, denoted in the set $\mathbb{P}=\left\\{(\mathcal{P}_{m}^{(1)},\mathcal{P}_{m}^{(2)})\right\\}_{m=1}^{M}$, where the superscript (i) indicates the viewpoint index. For each plane $\mathcal{P}_{m}^{(i)}$, we denote the normal and offset in $\mathbf{n}_{m}^{(i)}\in\mathbb{R}^{3}$ and $d_{m}^{(i)}\in\mathbb{R}_{+}$, respectively. With the plane correspondence set $\mathbb{P}$, we aim at estimating the relative camera pose in Sec. 4. ## 4 NOPE-SAC Pose Estimation Given the initial camera pose $\xi_{0}=(R_{0},\mathbf{t}_{0})$ and a pair of matched planes $\mathcal{P}_{m}^{(1)}$ in the first-view image $\mathcal{I}_{1}$ and $\mathcal{P}_{m}^{(2)}$ in the second-view image $\mathcal{I}_{2}$, the proposed NOPE-SAC learns the one plane pose in the embedding space of the initial camera pose $\xi_{0}$ and the plane correspondence $(\mathcal{P}_{m}^{(1)},\mathcal{P}_{m}^{(2)})$. For a total of $M$ putative plane matches, there will be $M$ one-plane poses and we then learn to vote those one-plane pose hypotheses according to the geometric cost in the planar consensus set. Thanks to the pose hypotheses by NOPE-SAC only requiring one plane correspondence to yield a pose hypothesis, the imbalance issue between the number of hypotheses and the cardinal of the consensus set is largely alleviated for accurate pose estimation. ### 4.1 Camera Pose Initialization In the consensus sampling pipelines for camera pose estimation, a coarse estimation is required to kick off the estimation of pose parameters. Different from the keypoint-based RANSAC (and its variants) that did the coarse estimation by randomly sampling a small subset of keypoint correspondences, it would be more challenging to obtain the coarse estimation from plane correspondences by an explicit mathematical model. To this end, we resort to the convolutional neural network to learn the initial pose $\xi_{0}$ from the feature volume of image $\mathcal{I}_{1}$ and $\mathcal{I}_{2}$. #### Pose Regression Network. We take ResNet-50 [40] as our backbone network for the initial pose regression. Given two input images $\mathcal{I}_{1}$ and $\mathcal{I}_{2}$ of which they are with the spatial size of $H\times W$, the output backbone features $\mathcal{F}_{1}$ and $\mathcal{F}_{2}$ are taken to build a 4D affinity volume $V\in\mathbb{R}^{(h\times w)\times(h\times w)}$ by $V(\mathbf{p},\mathbf{p}^{\prime})=\frac{\exp(\mathcal{F}_{2}^{T}(\mathbf{p})\mathcal{F}_{1}^{T}(\mathbf{p}^{\prime}))}{\sum_{\mathbf{p}}\exp(\mathcal{F}_{2}^{T}(\mathbf{p})\mathcal{F}_{1}^{T}(\mathbf{p}^{\prime}))},$ (6) for any pair of pixels $\mathbf{p}$ and $\mathbf{p}^{\prime}$ in the coordinates of $\mathcal{F}_{1}$ and $\mathcal{F}_{2}$, where $h=H/s$, $w=W/s$, and $s=32$ is the stride of backbone network. The computation of $V$ is similar to [7], however, we choose to directly predict the pose parameters instead of learning the classification likelihood with respect to a set of discrete pose templates as the formulation of classification would hurt the end-to-end learning for camera pose estimation. To predict the initial camera pose $\xi_{0}=(R_{0},\mathbf{t}_{0})\in\text{SE}(3)$, the affinity volume is rearranged to $\mathcal{V}\in\mathbb{R}^{hw\times h\times w}$, and is then transformed respectively by 6 convolutional layers (including BatchNorm and LeakyReLU) and a linear layer to yield two feature embeddings, $f_{r}^{(0)},f_{t}^{(0)}\in\mathbb{R}^{256}$ for the rotation $R_{0}$ and translation $\mathbf{t}_{0}$. Here, the stride of each even-numbered convolution operation is set to 2 to reduce the feature size. With $f_{r}^{(0)},f_{t}^{(0)}$, it is straightforward to predict the rotation and translation by linear layers $\text{LinearRot}(\cdot)$ and $\text{LinearTrans}(\cdot)$ with $\mathbf{q}_{0}=\text{LinearRot}(f_{r}^{(0)}),\mathbf{t}_{0}=\text{LinearTrans}(f_{t}^{(0)}),$ (7) where $\mathbf{q}_{0}\in\mathbb{R}^{4}$ is the unit quaternion of the rotation matrix that satisfies $\left\|\mathbf{q}_{0}\right\|_{2}=1$. The weights of $\text{LinearRot}(\cdot)$ and $\text{LinearTrans}(\cdot)$ are optimized by the MSE (Mean-Square Error) loss $\mathcal{L}_{\text{cam}}^{\text{init}}=\|\mathbf{q}_{0}-\mathbf{q}^{gt}\|_{2}^{2}+\|\mathbf{t}_{0}-\mathbf{t}^{gt}\|_{2}^{2},$ (8) where $\mathbf{q}^{gt}$ and $\mathbf{t}^{gt}$ are the ground truths of the rotation (in unit quaternions) and translation between the input images $(\mathcal{I}_{1},\mathcal{I}_{2})$. Besides predicting the initial camera poses, the embedding $f_{r}^{(0)}$ and $f_{t}^{(0)}$ plays a vital role in our proposed NOPE-SAC because it generates the pose hypotheses by the embedding vectors of initial poses and plane correspondences. Thus, there would be an issue of inflexibility for the initial poses that do not come from the convolutional pose estimation network. We present an Arbitrary Initialization Module (AIM) that reconstructs a given camera pose via auto-encoding to cope with this situation. #### Arbitrary Initialization Module. We featurize any given quaternion $\mathbf{q}^{*}$ (of rotation $R^{*}\in\text{SO}(3)$) and translation $\mathbf{t}^{*}\in\mathbb{R}^{3}$ in an auto-encoding manner by: $\begin{split}f_{r}^{*}=\text{RotEnc}(\mathbf{q}^{*}),\,f_{t}^{*}=\text{TransEnc}(\mathbf{t}^{*}),\\\ \tilde{\mathbf{q}}^{*}=\text{LinearRot}(f_{r}^{*}),\,\tilde{\mathbf{t}}^{*}=\text{LinearTrans}(f_{t}^{*}),\end{split}$ (9) where $\text{RotEnc}(\cdot)$ and $\text{TransEnc}(\cdot)$ are the encoders for rotation and translation by MLPs, $\text{LinearRot}(\cdot)$ and $\text{LinearTrans}(\cdot)$ are the shared layers defined in Eqn. (7) to reconstruct the camera poses by $\mathcal{L}_{\text{rec}}=\left\|\tilde{\mathbf{q}}^{*}-\mathbf{q}^{*}\right\|_{2}^{2}+\left\|\tilde{\mathbf{t}}^{*}-\mathbf{t}^{*}\right\|_{2}^{2}.$ (10) For the training of AIM, we randomly sample the rotations and translations from the uniform distribution, please move to Section 5.3 for the details. TABLE I: Detailed architecture of MLPs used in NOPE-SAC Pose Estimation Module. For each MLP, the corresponding equation, notation, the number of linear layers, the channel dimensions of inputs and outputs, and the activation functions are listed. Equation | Notation | $\\#$ Layers | $\\#$ Channels | Activation ---|---|---|---|--- In | Out Eqn. (9) | RotEnc / TransEnc | 6 | 256 | 256 | ReLU Eqn. (11) | $\mathcal{G}$ | 1 | 8 | 1024 | ReLU 4 | 1024 | 1024 | ReLU 1 | 1024 | 1024 | - 2 | 1024 | 1024 | ReLU 1 | 1024 | 1024 | - Eqn. (12) | $\mathcal{E}_{r}$ | 1 | 1024 | 512 | ReLU 4 | 512 | 512 | ReLU 1 | 512 | 256 | - Eqn. (12) | $\mathcal{E}_{t}$ | 1 | 1280 | 1024 | ReLU 1 | 1024 | 1024 | ReLU 1 | 1024 | 1024 | - 1 | 1024 | 512 | ReLU 4 | 512 | 512 | ReLU 1 | 512 | 256 | - Eqn. (13) | $\mathcal{D}_{r}$ / $\mathcal{D}_{t}$ | 2 | 512 | 512 | ReLU 1 | 512 | 256 | ReLU ### 4.2 Camera Pose Refinement With the initial camera pose $\xi_{0}=(R_{0},\mathbf{t}_{0})$ and its corresponding feature embedding $f_{r}^{(0)}$ and $f_{t}^{(0)}$, we present the core of NOPE-SAC that consists of the following steps: (1) One-plane Pose Hypotheses Generation, (2) Hypotheses Scoring and (3) Pose Refinement based on the $M$ 3D plane correspondences $\mathbb{P}=\\{\mathcal{P}_{m}\\}_{m=1}^{M}$. For each plane correspondence $\mathcal{P}_{m}$, the normal vector (3D real vectors) and the offset (real scalars) from the plane to the camera center in the input images $\mathcal{I}_{1}$ and $\mathcal{I}_{2}$ are denoted by $(\mathbf{n}_{m}^{(1)},d_{m}^{(1)})$ and $(\mathbf{n}_{m}^{(2)},d_{m}^{(2)})$, respectively. #### One-plane Pose Hypotheses Generation. For the $m$-th 3D plane correspondence $\mathcal{P}_{m}^{(1)}=(\mathbf{n}_{m}^{(1)},{d}_{m}^{(1)})$ in $\mathcal{I}_{1}$ and $\mathcal{P}_{m}^{(2)}=(\mathbf{n}_{m}^{(2)},{d}_{m}^{(2)})$ in $\mathcal{I}_{2}$, we leverage an MLP layer $\mathcal{G}(\cdot):\mathbb{R}^{8}\mapsto\mathbb{R}^{1024}$ to embed the correspondence by $g_{m}=\mathcal{G}(\tilde{\mathbf{n}}_{m}^{(1)}\oplus{\tilde{\mathbf{d}}}_{m}^{(1)}\oplus\mathbf{n}_{m}^{(2)}\oplus\mathbf{d}_{m}^{(2)}),$ (11) where $\tilde{\mathbf{n}}_{m}^{(1)}$ and $\tilde{d}_{m}^{(1)}$ are the warped normal of $\mathbf{n}_{m}^{(1)}$ and offset of $d_{m}^{(1)}$ by the relative camera pose $(R_{0},\mathbf{t}_{0})$, $\oplus$ is the concatenation operation. Here, the warped plane parameters by the initial camera pose could be regarded as a kind of normalization to facilitate the learning of plane correspondence embedding in neural networks. With the correspondence embedding $g_{m}$, we further transform it to the rotation feature $e_{m}^{r}\in\mathbb{R}^{256}$ and the translation feature $e_{m}^{t}\in\mathbb{R}^{256}$ by two MLPs $\mathcal{E}_{r}$ and $\mathcal{E}_{t}$ as $\begin{split}e_{m}^{r}&=\mathcal{E}_{r}(g_{m}),\\\ e_{m}^{t}&=\mathcal{E}_{t}(g_{m}\oplus e_{m}^{r}).\end{split}$ (12) Then, we concatenate $e_{m}^{r}$ and $e_{m}^{t}$ with the embedding $f_{r}^{(0)}$ and $f_{t}^{(0)}$ of the initial camera pose, and transform the concatenated embedding by two MLP layers $\mathcal{D}_{r}(\cdot):\mathbb{R}^{512}\mapsto\mathbb{R}^{256}$ and $\mathcal{D}_{t}(\cdot):\mathbb{R}^{512}\mapsto\mathbb{R}^{256}$ to yield the one-plane pose embedding $f_{m}^{r}$ and $f_{m}^{t}$ as $\begin{split}f_{m}^{r}&=\mathcal{D}_{r}(f_{r}^{(0)}\oplus e_{m}^{r}),\\\ f_{m}^{t}&=\mathcal{D}_{t}(f_{t}^{(0)}\oplus e_{m}^{t}).\end{split}$ (13) Finally, to obtain the one-plane rotation and translation for the $m$-th plane correspondence, we leverage the linear layers $\text{LinearRot}(\cdot)$ and $\text{LinearTrans}(\cdot)$ defined in Eqn. (9) as the shared headnets by $\begin{split}\mathbf{q}^{(m)}&=\text{LinearRot}(f_{m}^{r}),\\\ \mathbf{t}^{(m)}&=\text{LinearTrans}(f_{m}^{r}),\end{split}$ (14) where $\mathbf{q}^{(m)}$ and $\mathbf{t}^{(m)}$ are the predicted quaternion of rotation and the translation. #### Hypotheses Scoring Layer. Next to the pose hypotheses generation, we compute the geometric cost for each predicted rotation and translation. For each pose hypothesis $R^{(m)}$ and $\mathbf{t}^{(m)}$, the rotation cost $c_{r}^{(m)}\in\mathbb{R}^{M}$ and translation cost $c_{t}^{(m)}\in\mathbb{R}^{M}$ throughout all $M$ plane correspondences are calculated by $\begin{split}c_{r}^{(m)}&=\left(\|\mathbf{n}_{i,m}^{(1)}-\mathbf{n}_{i}^{(2)}\|_{2}\right)_{i=1}^{M}\\\ c_{t}^{(m)}&=\left(\|d_{i,m}^{(1)}\mathbf{n}_{i,m}^{(1)}-d_{i}^{(2)}\mathbf{n}_{i}^{(2)}\|_{2}\right)_{i=1}^{M},\end{split}$ (15) where $\mathbf{n}_{i,m}^{(1)}$ and $d_{i,m}^{(1)}$ are the warped normal and offset of the plane $\mathcal{P}_{i}^{(1)}$ by the $m$-th pose hypothesis. By including the costs of initial pose $(R_{0},\mathbf{t}_{0})$, there will be two cost matrices $C_{r}\in\mathbb{R}^{(M+1)\times M}$ and $C_{t}\in\mathbb{R}^{(M+1)\times M}$, which are then fed into two parallel Hypotheses Scoring Layers to yield the rotation scores $\mathbf{w}_{r}=(w_{r}^{(0)},\ldots,w_{r}^{(M)})$ and the translation scores $\mathbf{w}_{t}=(w_{t}^{(0)},\ldots,w_{t}^{(M)})$, respectively. Here, the sum of scores is equal to $1$ and we use the Softmax operation in our implementation. As shown in Fig. 3, each Hypotheses Scoring Layer consists of an MLP layer (including 3 linear layers with ReLU activations and 64 channels per layer) and one linear layer for score prediction. To train the scoring layer, we design a scoring loss to supervise the learning of pose scores dynamically. Let $i$ and $j$ be the indices of rotation and translation in all pose hypotheses $\\{R^{(m)}\\}_{m=0}^{M}$ and $\\{t^{(m)}\\}_{m=0}^{M}$ that are closest to the ground truth camera poses. Then, the heuristic scoring loss can be calculated as: $\displaystyle\mathcal{L}_{\text{score}}=\|1-w_{r}^{i}\|_{1}+2\|1-w_{t}^{j}\|_{1}+\frac{10}{M}\sum_{m=1}^{M}c_{t}^{(m)}(m),$ (16) where $w_{r}^{i}$ and $w_{t}^{j}$ are the predicted rotation and translation scores, and $c_{t}^{(m)}(m)$ means the translation cost of plane correspondence $\mathcal{P}_{m}$ with the $m$-th pose hypothesis. Figure 3: An illustration of the Hypotheses Scoring Layer. The input cost matrix is first passed through an MLP layer and then converted to the scores with a linear layer and Softmax operation. The top line shows the output dimensions of each step. #### Final Pose Estimation. Following the hypotheses scoring layer, we present a _Soft_ fusion strategy by default to fuse all the pose hypotheses to obtain the final pose. In detail, we use the predicted scores $\mathbf{w}_{r}$, $\mathbf{w}_{t}$ to obtain the embedding of the final pose by $f_{r}=\sum_{i=0}^{M}w_{r}^{i}f_{r}^{(i)},\,f_{t}=\sum_{i=0}^{M}w_{t}^{i}f_{t}^{(i)},$ (17) and leverages the layers $\text{LinearRot}(\cdot)$ and $\text{LinearTrans}(\cdot)$ again as the shared pose predictors to obtain the final pose $\xi_{\rm ref}=(\mathbf{q}_{\rm ref},\mathbf{t}_{\rm ref})$. We denote the computation of the camera pose refinement module by $\xi_{\rm ref}={\rm PoseRef}(\xi_{0},\mathbb{P};\Theta),$ (18) where $\xi_{0}$ and $\mathbb{P}$ are the required input of the initial pose and the plane correspondences, and $\Theta$ indicates the parameters of the neural networks. Apart from the _Soft_ fusion scheme, there are three representative alternatives to get the final refined pose from pose hypotheses, which are summarized as follows: 1. - Avg: It treats all pose hypotheses equally and sets all scores used in Eqn. (17) to be $\frac{1}{M+1}$. 2. - Min-Cost: It selects the minimal-cost rotation/translation hypothesis according to Eqn. (15) and discard all the remaining hypotheses. 3. - Max-Score: It takes the maximal-score rotation/translation hypothesis according to the hypotheses scoring layer as the final pose prediction. If not explicitly stated, we use the Soft fusion strategy in all experiments, and we analyze these fusion strategies in the ablation study. We supervise the final pose prediction with the MSE loss. Because there are two initial pose predictions $\xi_{0}^{\rm reg}$ (from the regression branch) and $\xi_{0}^{\rm rec}$ (from the Arbitrary Initialization Module), by taking the predicted plane correspondences $\mathbb{P}_{\rm pred}$ and the ground- truth $\mathbb{P}_{gt}$ into account during training, there will be four predictions of the refined camera pose for the _Soft_ fusion scheme. We compute the loss functions $\mathcal{L}_{\rm cam}^{\rm soft}$ and $\mathcal{L}_{\rm score}$ for the refinement module similar to Eqn. (8) and Eqn. (16), but the four predictions are all taken in the computation. To avoid instability during training, the Avg fusion strategy is also involved to compute a loss function $\mathcal{L}_{\rm cam}^{\rm avg}$. Finally, the total loss function of the refinement module is denoted by $\mathcal{L}_{\rm cam}^{\rm ref}=\mathcal{L}_{\rm cam}^{\rm soft}+\mathcal{L}_{\rm cam}^{\rm avg}+0.01\mathcal{L}_{\rm score}.$ (19) To train the NOPE-SAC for pose estimation, we linearly add the loss functions $\mathcal{L}_{\text{cam}}^{\text{init}}$, $\mathcal{L}_{\text{rec}}$ and $\mathcal{L}_{\text{cam}}^{\text{ref}}$ by $\mathcal{L}_{\text{pose}}=\mathcal{L}_{\text{cam}}^{\text{init}}+\mathcal{L}_{\text{rec}}+\mathcal{L}_{\text{cam}}^{\text{ref}}$ (20) as the total loss for the camera pose estimation. ## 5 Experiments and Analysis This section presents the experiments of our NOPE-SAC. (a) Matterport3D [13] train (b) Matterport3D [13] test (c) ScanNet [14] train (d) ScanNet [14] test (e) Rotation train (f) Rotation test (g) Translation train (h) Translation test Figure 4: Dataset analysis of the sparse-view split on the Matterport3D [13] and the ScanNet [14] datasets. The top row shows the image overlap ratio on the training and test set of two datasets (from left to right, the average percentages of image overlap are $20.9\%$, $21.0\%$, $20.6\%$ and $18.6\%$). At the bottom, we visualize the rotation and translation distributions on the ScanNet (Blue dot) and the Matterport3D (Red dot) datasets. ### 5.1 Datasets #### Matterport3D Dataset. We use the sparse view dataset based on Matterport3D [13] with ground truth camera poses and plane annotations generated by [7], which contains 31932, 4707, and 7996 image pairs for training, validation, and testing. The size of each image is $480\times 640$. #### ScanNet Dataset. We create a new sparse view split on the indoor ScanNet [14] video dataset with plane annotations generated by [23]. The image size is $480\times 640$. We randomly sample 17237/4051 image pairs from 1210/303 non-overlapping scenes for training/testing. The frame interval within a sampled image pair is at least 20 and 40 frames in the training and test sets, respectively. #### Dataset Analysis. We first analyze the image overlap on the Matterport3D [13] and the ScanNet [14] datasets. As shown in the top row of Fig. 4, our split on the ScanNet dataset contains more low-overlap image pairs than the split on the Matterport3D dataset created by [7]. Further, we visualize the distributions of rotations (represented as rotation vectors) and translations on the training and test set of these two datasets. As shown in the bottom row of Fig. 4, although the translation range on the ScanNet dataset is smaller than the Matterport3D dataset, the rotation distribution on the ScanNet dataset is much more complex than the Matterport3D dataset, which makes our split on the ScanNet dataset more challenging. ### 5.2 Evaluation Metrics #### Metrics for Camera Pose Estimation. We evaluate camera poses with the rotation angle error and the translation distance error as used in [7, 37, 36]. The reported metrics include the mean and median errors, and the percentage of errors lower than a threshold. #### Metrics for 3D Plane Reconstruction. According to SparsePlanes [7], the matched planes are merged together and all 3D planes from two views are converted to the same coordinate system. Then, the reconstructed full scene is evaluated like a detection problem with the metric of average precision (AP). A reconstructed 3D plane is regarded as a true positive if the following conditions of (1) its mask IoU $\geq$ 0.5, (2) its plane normal angle error in degree is less than $\alpha^{\circ}$, and (3) its plane offset error in meters is less than $\beta$ are all satisfied. TABLE II: Comparison of camera poses on the Matterport3D [13] dataset and the ScanNet [14] dataset. Method | Translation | Rotation ---|---|--- Med. $\downarrow$ | Mean $\downarrow$ | ($\leq$1m) $\uparrow$ | ($\leq$0.5m) $\uparrow$ | ($\leq$0.2m) $\uparrow$ | Med. $\downarrow$ | Mean $\downarrow$ | ($\leq 30^{\circ}$) $\uparrow$ | ($\leq 15^{\circ}$) $\uparrow$ | ($\leq 10^{\circ}$) $\uparrow$ Matterport3D dataset | SuperGlue [9] | - | - | - | - | - | 3.88 | 24.17 | 77.8% | 71.0% | 65.7% SparsePlanes [7] | 0.63 | 1.15 | 66.6% | 40.4% | 11.9% | 7.33 | 22.78 | 83.4% | 72.9% | 61.2% PlaneFormers [8] | 0.66 | 1.19 | 66.8% | 36.7% | 8.7% | 5.96 | 22.20 | 83.8% | 77.6% | 68.0% SparsePlanes-TR [7, 25] | 0.61 | 1.13 | 67.3% | 41.7% | 12.2% | 6.87 | 22.17 | 83.8% | 74.5% | 63.3% PlaneFormers-TR [8, 25] | 0.64 | 1.17 | 67.9% | 38.7% | 8.9% | 5.28 | 21.90 | 83.9% | 79.0% | 70.8% NOPE-SAC-Cls (ours) | 0.66 | 1.20 | 65.6% | 37.9% | 9.8% | 2.98 | 19.68 | 84.9% | 83.1% | 80.2% NOPE-SAC-Reg (ours) | 0.52 | 0.94 | 73.2% | 48.3% | 16.2% | 2.77 | 14.37 | 89.0% | 86.9% | 84.0% ScanNet dataset | SuperGlue [9] | - | - | - | - | - | 10.90 | 31.00 | 67.8% | 56.0% | 48.4% SparsePlanes [7] | 0.56 | 0.81 | 73.7% | 44.6% | 10.7% | 15.46 | 33.38 | 70.5% | 48.7% | 28.0% PlaneFormers [8] | 0.55 | 0.81 | 75.3% | 45.5% | 11.3% | 14.34 | 32.08 | 73.2% | 52.1% | 32.3% SparsePlanes-TR [7, 25] | 0.57 | 0.82 | 73.4% | 43.6% | 10.1% | 14.57 | 32.36 | 72.8% | 51.2% | 30.1% PlaneFormers-TR [8, 25] | 0.53 | 0.79 | 76.2% | 47.0% | 11.4% | 13.81 | 31.58 | 74.5% | 54.1% | 33.6% NOPE-SAC-Cls (ours) | 0.49 | 0.76 | 77.5% | 50.9% | 14.1% | 9.01 | 27.84 | 77.9% | 69.5% | 55.1% NOPE-SAC-Reg (ours) | 0.41 | 0.65 | 82.0% | 59.1% | 21.2% | 8.27 | 22.12 | 82.6% | 73.2% | 59.5% ### 5.3 Implementation Details Our NOPE-SAC is implemented with Detectron2 [45] and the AdamW optimizer [46] is used for training with a batch size of 16. Because the sparse-view planar 3D reconstruction consists of multiple neural modules for the 3D plane detection and matching, the camera pose initialization (including the initial pose regression and the arbitrary pose initialization), and the pose refinement with one-plane pose hypotheses, we train the whole network in three stages on the Matterport3D [13] training set. In the first stage, only the 3D plane detection module is optimized for 12k iterations with the learning rate of $10^{-4}$. Then, we add the plane matching module, the initial pose regression network, and the arbitrary initialization module for the second stage of training in 50K iterations. The initial learning rate in this training stage is also $10^{-4}$, and we use the multi-step learning rate schedule to adjust the learning rate at the milestones of 34k and 44k iterations with the decay factor of 0.1. In the last stage, we add the camera pose refinement module in training for a total of 14k iterations. The learning rate in this stage is initialized to $10^{-4}$ for the first 6k iterations and is decayed to $10^{-5}$ for the last 8k iterations. On the ScanNet [14] dataset, we finetune the model trained on the Matterport3D dataset with two additional stages. In the first stage, we train the 3D plane detection module, the plane matching module, the initial pose regression network, and the arbitrary initialization module for 20K iterations. The initial learning rate is set to $10^{-4}$ and divided by 10 after 4.4K iterations. In the second stage, we add the camera pose refinement module and train the whole network for 15K iterations. The initial learning rate is set to $10^{-4}$ and divided by 10 after 2.2K iterations. When training the arbitrary initialization module (AIM), we first represent the input rotation as a rotation vector $\textbf{v}^{*}\in\mathbb{R}^{3}$. Then, we randomly sample each axis of $\textbf{v}^{*}$ and the input translation $\textbf{t}^{*}\in\mathbb{R}^{3}$ from the uniform distribution $U(-2.5,2.5)$. At last, the sampled rotation vector $\textbf{v}^{*}$ is converted to a unit quaternion $\textbf{q}^{*}$ as used in Eqn. 9. TABLE III: Detailed camera pose comparison with SparsePlanes-TR [7, 25] on the Matterport3D [13] dataset and the ScanNet [14] dataset. ‘Cls. Top-1’ means the Top-1 classification pose of SparsePlanes [7]. ‘Con.’ means the continuous optimization proposed by SparsePlanes. Method | Translation | Rotation ---|---|--- Med. $\downarrow$ | Mean $\downarrow$ | ($\leq$0.5m) $\uparrow$ | Med. $\downarrow$ | Mean $\downarrow$ | ($\leq 15^{\circ}$) $\uparrow$ Matterport3D dataset Cls. Top-1 | 0.90 | 1.40 | 21.1% | 7.65 | 24.57 | 71.7% [7, 25] w/o Con. | 0.88 | 1.35 | 21.9% | 7.17 | 22.36 | 74.6% [7, 25] | 0.61 | 1.13 | 41.7% | 6.87 | 22.17 | 74.5% NOPE-SAC-Cls | 0.66 | 1.20 | 37.9% | 2.98 | 19.68 | 83.1% ScanNet dataset Cls. Top-1 | 0.56 | 0.83 | 43.9% | 15.21 | 33.16 | 49.5% [7, 25] w/o Con. | 0.55 | 0.83 | 44.2% | 14.55 | 32.36 | 51.3% [7, 25] | 0.57 | 0.82 | 43.6% | 14.57 | 32.36 | 51.2% NOPE-SAC-Cls | 0.49 | 0.76 | 50.9% | 9.01 | 27.84 | 69.5% ### 5.4 Baseline Configurations We compare our method with the state-of-the-art learning solutions including SparsePlanes [7] and PlaneFormers [8] for pose estimation and planar 3D reconstruction, as well as a keypoint-based solution built on the top- performing keypoint matcher, SuperGlue [9]. For the SparsePlanes [7] and PlaneFormers [8], we additionally set two variant baselines to compare the performance gain with the same 3D plane detection module. #### SparsePlanes and SparsePlanes-TR . SparsePlanes [7] is the first sparse-view planar 3D reconstruction approach. Given a set of precompute pose hypotheses, it detects 3D planes in each view and predicts the likelihood of the precomputed set of camera poses by deep neural networks. Then, a discrete optimization problem is solved to obtain the optimal plane correspondences and the camera pose by taking the geometric correctness and embedding consistency between 3D planes across views, as well as the classification likelihood into account. Following the discrete optimization, a continuous optimization problem is formulated to refine the camera pose with the assistance of SIFT [26] keypoints. Since the 3D plane detection modules between SparsePlanes [7] and NOPE-SAC are different, we build a variant baseline SparsePlanes-TR to ensure the fairness of comparisons, in which we replace the plane detection module to PlaneTR [25] as our network used. Because the official implementation of SparsePlanes [7] did not provide results on the ScanNet [14] dataset, we train these baselines on the ScanNet [14] dataset by ourselves. #### PlaneFormers and PlaneFormers-TR . PlaneFormers [8] is the prior art developed based on SparsePlanes [7], which selects Top-9 classification poses from [7] as hypotheses. For each pose hypothesis, PlaneFormers jointly estimates the pose score and plane matching cost with a neural network. Finally, the pose with the best score is selected and refined with an extra estimated pose residual, and the plane correspondences are achieved by conducting an offline Hungarian algorithm on the matching cost. Like SparsePlane [7], we implemented the results of PlaneFormers on the ScanNet [14] dataset by ourselves. In PlaneFormers-TR, we use the planes detected by PlaneTR [25] as our network used. #### SuperGlue . It is a strong baseline for keypoints detection and matching with neural networks. The camera poses are then calculated from matched keypoints with ‘5-point solver [10] \+ RANSAC [11]’. Since SuperGlue [9] lacks scale information in translations, we do not report its translation errors. In case SuperGlue fails to estimate camera poses, we set its rotation results as identity matrices like [7]. For the results of plane reconstruction, we use the ground truth scales in translations and match planes detected by PlaneTR [25] with our plane matching module. (a) Image 1 (b) Image 2 (c) Plane Correspondences (d) Poses (view 1) (e) Poses (view 2) Figure 5: Qualitative results of the refined poses and plane correspondences on the Matterport3D [13] dataset (first three rows) and the ScanNet [14] dataset (last three rows). The Pink, Blue and Red frustums show the initial, the refined, and the ground truth cameras of the first image respectively. Gray frustums show the one-plane pose hypotheses of the first image generated from plane correspondences. Black frustums show the camera of the second image. ### 5.5 Comparison of Camera Pose Estimation We evaluate our NOPE-SAC with the regressed initial poses, namely NOPE-SAC- Reg, which is the main version of our proposed method. We also evaluate a variant of our NOPE-SAC which directly uses the Top-1 a-pripori pose with a classification module of SparsePlanes [7] as initial poses, namely NOPE-SAC- Cls. #### Quantitative Results. We first compare the performance of rotation estimation. As shown in Tab. II, both our NOPE-SAC-Reg and Nope-SAC-Cls significantly outperform all baselines on both two datasets especially when the rotation threshold is small, _e.g.,_ $84.0\%$ (NOPE-SAC-Reg) v.s. $63.3\%$ (SparsePlanes-TR) with the threshold of $10^{\circ}$ on the Matterport3D [13] dataset. Then, we further evaluate the translation results. When compared with SparsePlanes-TR [7, 25] and PlaneFormers-TR [8, 25] which use the a-pripori classification poses, our NOPE-SAC-Cls performs slightly worse than them on the Matterport3D [13] dataset but outperforms them on the more challenging ScanNet [14] dataset. With a further comparison to SparsePlanes-TR [7, 25] as shown in Tab. III, we find that the improvement of translation estimation for SparsePlanes-TR [7, 25] largely depends on its keypoint-based continuous optimization. Although they are well-performing on the MatterPort3D [13] dataset, the performance of continuous optimization dramatically degenerates on the ScanNet [14] dataset because it is much more difficult to achieve good keypoint matches on the ScanNet dataset (as shown in Fig. 1(a)). By contrast, our NOPE-SAC-Cls consistently improves the initial poses with only the plane correspondences on both datasets. #### Qualitative Results. Fig. 5 shows the pose estimation results of our NOPE-SAC from two different viewpoints (last two columns) on the Matterport3D [13] and the ScanNet [14] datasets. As described in Sec. 4, we compute one-plane pose hypotheses (Gray frustums) from estimated plane correspondences (the third column). Despite the outlier poses caused by incorrect correspondences and inaccurate plane parameters, NOPE-SAC effectively achieved the final refined pose (Blue frustum) from all one-plane pose hypotheses (Gray frustums) and the initial pose (Pink frustum). TABLE IV: Average Precision (AP) of 3D plane reconstruction conditioned with mask IoU, normal angle error, and offset distance error. The threshold of mask IoU is fixed to 0.5. ‘All’ means we consider all three conditions. ‘-Offset’ and ‘-Normal’ mean we ignore the offset and the normal conditions respectively. ‘Con.’ means the continuous optimization proposed by SparsePlanes [7]. Method | Offset$\leq$1m, Normal$\leq 30^{\circ}$ | Offset$\leq$0.5m, Normal$\leq 15^{\circ}$ | Offset$\leq$0.2m, Normal$\leq 5^{\circ}$ ---|---|---|--- All | -Offset | -Normal | All | -Offset | -Normal | All | -Offset | -Normal Matterport3D dataset SuperGlue-TR [9, 25] | 39.51 | 44.11 | 44.08 | 28.98 | 37.53 | 34.40 | 11.29 | 21.87 | 17.27 SparsePlanes [7] w/o Con. | 35.87 | 42.13 | 38.80 | 23.36 | 35.34 | 27.48 | 8.07 | 17.28 | 12.99 SparsePlanes [7] | 36.02 | 42.01 | 39.04 | 23.53 | 35.25 | 27.64 | 6.76 | 17.18 | 11.52 PlaneFormers [8] | 37.62 | 43.19 | 40.36 | 26.10 | 36.88 | 29.99 | 9.44 | 18.82 | 14.78 SparsePlanes-TR [7, 25] w/o Con. | 39.91 | 46.50 | 42.53 | 27.37 | 40.79 | 31.03 | 9.99 | 22.80 | 14.64 SparsePlanes-TR [7, 25] | 40.35 | 46.39 | 43.03 | 27.81 | 40.65 | 31.38 | 9.02 | 22.80 | 13.66 PlaneFormers-TR [8, 25] | 41.87 | 47.50 | 44.43 | 30.78 | 42.82 | 34.03 | 12.45 | 25.98 | 17.34 NOPE-SAC-Cls Init. (ours) | 38.94 | 46.60 | 41.96 | 26.17 | 40.48 | 29.89 | 9.89 | 22.55 | 14.29 NOPE-SAC-Cls Ref. (ours) | 41.92 | 48.18 | 44.01 | 31.36 | 44.24 | 34.01 | 13.59 | 30.05 | 17.45 NOPE-SAC-Reg Init. (ours) | 40.07 | 46.03 | 43.59 | 26.78 | 36.76 | 31.95 | 10.09 | 19.09 | 15.55 NOPE-SAC-Reg Ref. (ours) | 43.29 | 49.00 | 45.32 | 32.61 | 44.94 | 35.36 | 14.25 | 30.39 | 18.37 ScanNet dataset SuperGlue-TR [9, 25] | 33.20 | 33.77 | 43.40 | 22.78 | 24.94 | 36.72 | 4.35 | 6.19 | 19.33 SparsePlanes [7] w/o Con. | 33.20 | 34.12 | 40.74 | 22.89 | 25.62 | 33.67 | 3.03 | 4.52 | 17.17 SparsePlanes [7] | 33.08 | 34.12 | 40.51 | 21.69 | 25.59 | 32.20 | 2.52 | 4.50 | 14.85 PlaneFormers [8] | 34.64 | 35.47 | 41.37 | 24.48 | 27.19 | 34.69 | 3.93 | 5.52 | 18.58 SparsePlanes-TR [7, 25] w/o Con. | 35.56 | 36.51 | 42.14 | 26.01 | 29.61 | 35.12 | 3.96 | 6.10 | 18.59 SparsePlanes-TR [7, 25] | 35.32 | 36.50 | 41.92 | 24.71 | 29.55 | 33.50 | 3.21 | 6.07 | 15.32 PlaneFormers-TR [8, 25] | 36.82 | 37.87 | 43.01 | 27.41 | 30.72 | 36.31 | 4.83 | 7.02 | 19.94 NOPE-SAC-Cls Init. (ours) | 35.41 | 36.68 | 42.44 | 25.21 | 28.78 | 34.96 | 3.84 | 5.74 | 18.66 NOPE-SAC-Cls Ref. (ours) | 38.23 | 39.36 | 43.27 | 30.25 | 34.15 | 36.93 | 6.23 | 9.57 | 20.56 NOPE-SAC-Reg Init. (ours) | 36.39 | 37.35 | 43.15 | 25.59 | 28.54 | 36.06 | 4.59 | 6.41 | 19.92 NOPE-SAC-Reg Ref. (ours) | 39.39 | 40.30 | 43.88 | 31.21 | 34.89 | 37.88 | 6.74 | 10.10 | 21.41 (a) Image 1 (b) Image 2 (c) SparsePlanes-TR (d) PlaneFormers-TR (e) Ours (f) Ground Truth Figure 6: Comparison of 3D plane reconstruction results on the Matterport3D [13] dataset (first four rows) and the ScanNet [14] dataset (last three rows). Blue and Black frustums show cameras of the first and the second images respectively. TABLE V: Ablation studies for NOPE-SAC camera pose refinement. Method | Translation | Rotation ---|---|--- Med. $\downarrow$ | Mean $\downarrow$ | ($\leq$1m) $\uparrow$ | ($\leq$0.5m) $\uparrow$ | ($\leq$0.2m) $\uparrow$ | Med. $\downarrow$ | Mean $\downarrow$ | ($\leq 30^{\circ}$) $\uparrow$ | ($\leq 15^{\circ}$) $\uparrow$ | ($\leq 10^{\circ}$) $\uparrow$ Matterport3D dataset | Initial Pose | 0.69 | 1.08 | 65.0% | 37.0% | 10.1% | 11.16 | 21.49 | 81.3% | 60.5% | 46.5% Nume-Ref-I | 0.52 | 1.06 | 69.2% | 48.5% | 19.1% | 7.17 | 21.96 | 79.8% | 67.4% | 58.3% Nume-Ref-II | 0.68 | 1.07 | 65.5% | 38.2% | 10.7% | 8.38 | 19.97 | 82.2% | 65.7% | 54.9% Homo-Ref | 0.61 | 1.05 | 68.5% | 42.9% | 14.9% | 5.03 | 17.47 | 85.1% | 74.5% | 66.3% PlaneFormers-TR [8, 25] | 0.66 | 1.06 | 66.8% | 37.5% | 9.8% | 8.96 | 20.01 | 83.0% | 65.9% | 53.5% NOPE-SAC (ours) | 0.52 | 0.94 | 73.2% | 48.3% | 16.2% | 2.77 | 14.37 | 89.0% | 86.9% | 84.0% ScanNet dataset | Initial Pose | 0.48 | 0.72 | 77.7% | 51.9% | 16.5% | 14.68 | 26.75 | 73.7% | 51.0% | 34.4% Nume-Ref-I | 0.54 | 0.80 | 73.2% | 47.2% | 13.7% | 16.57 | 29.47 | 68.7% | 47.0% | 32.3% Nume-Ref-II | 0.48 | 0.72 | 77.7% | 51.8% | 16.5% | 14.64 | 26.72 | 73.7% | 51.1% | 34.6% Homo-Ref | 0.49 | 0.73 | 77.5% | 51.3% | 16.5% | 14.03 | 26.27 | 74.2% | 52.7% | 36.8% PlaneFormers-TR [8, 25] | 0.48 | 0.72 | 78.3% | 52.2% | 14.2% | 14.30 | 26.51 | 74.5% | 52.2% | 33.7% NOPE-SAC (ours) | 0.41 | 0.65 | 82.0% | 59.1% | 21.2% | 8.27 | 22.12 | 82.6% | 73.2% | 59.5% ### 5.6 Comparison of 3D Planar Reconstruction We evaluate two versions of our method named NOPE-SAC-Reg and NOPE-SAC-Cls as described in Sec. 5.5 on the Matterport3D [13] and ScanNet [14] datasets. #### Quantitative Results. We evaluate our method with initial (Init.) and refined (Ref.) poses, and compare with baselines under various plane offset and normal error thresholds from loose to strict. As shown in Tab. IV, Our NOPE-SAC-Reg achieves state-of- the-art performance especially under the strictest settings with ‘Offset$\leq$0.2m’ and ‘Normal$\leq 5^{\circ}$’ on both Matterport3D [13] and ScanNet[14] datasets. When compared with SparsePlanes-TR [7, 25] which also uses classification poses, our NOPE-SAC-Cls effectively improves the reconstruction performance (_e.g.,_ from 25.21 to 30.25 with offset$\leq$0.5m and normal$\leq 10^{\circ}$ on the ScanNet dataset). Those performance gains confirmed the superiority of our proposed NOPE-SAC. #### Qualitative Results. Fig. 6 visualizes the 3D plane reconstruction results by different approaches on the Matterport3D [13] and the ScanNet [14] datasets. As it is shown, our method successfully reconstructs the scenes from sparse views even when the image overlap is very small (_e.g.,_ the third row in Fig. 6) and the viewpoint change is very large (_e.g.,_ the fourth row in Fig. 6). ### 5.7 Ablation for NOPE-SAC Pose Estimation This section presents a series of ablation study for our NOPE-SAC pose estimation as it plays the most important role for the end task. We use the initial pose achieved by our pose regression network by default. #### NOPE-SAC VS. Other Pose Refinement Methods. Here, we first compare our NOPE-SAC with three traditional methods and one learning based method for pose refinement, including: 1. (1) Nume-Ref-I: We use the numerical optimization like SparsePlanes [7] but only optimizes the initial camera poses $R^{(0)}$, $t^{(0)}$ as follows: $\begin{split}\mathop{\min}\limits_{R,t}\sum_{\mathcal{P}\in\mathbb{P}}&L(d_{\text{par}}(\mathcal{P},R,t))+d_{\text{pix}}(\mathcal{P},R,t)\\\ &+d_{\text{cam}}(R,R^{(0)}),\end{split}$ (21) where $\mathcal{P}$ is a pair of matched planes in the predicted plane correspondence set $\mathbb{P}$, $L(\cdot)$ is the Huber loss, $d_{\text{par}}$ calculates the euclidean distance between plane parameters, $d_{\text{pix}}$ is the reprojection error of matched SIFT [26] points on plane regions, and $d_{\text{cam}}$ is a regularization term which restricts the geodesic distance of rotations. 2. (2) Nume-Ref-II: A variant of Nume-Ref-I which excludes $d_{\text{pix}}$ in Eq. 21. 3. (3) Homo-Ref: We use homographies to estimate refined poses from predicted plane correspondences $\mathbb{P}$. Specifically, to each predicted plane correspondence $\mathcal{P}\in\mathbb{P}$, matched SIFT points are found from plane regions and then used to calculate the homography matrix $H$. Then a refined pose hypothesis can be decomposed from $H$. We use the scale of the initial translation and select the refined pose from all hypotheses which minimizes $\sum_{\mathcal{P}\in\mathbb{P}}d_{\text{par}}(\mathcal{P},R,t)$. 4. (4) PlaneFormers-TR [8, 25]: We estimate the refined pose by giving planes detected by PlaneTR [25] and our regressed initial pose as inputs to the PlaneFormers [8]. As shown in Tab. V, compared to the learning based PlaneFormers-TR [8, 25] which directly estimates a pose residual from features of all plane correspondences, our method improves the initial poses more effectively than PlaneFormers-TR [8, 25] on both two datasets with our accurate one-plane pose hypotheses. Benefiting from matched keypoints, the traditional Nume-Ref-I achieves the results which are most close to our NOPE-SAC on the Matterport3D [13] dataset. However, due to the difficulty to find sufficient good keypoint matches, Nume-Ref-I performs even worse than the initial pose on the ScanNet [14] dataset. Similarly, Homo-Ref also suffers from unsatisfied keypoint matches on the ScanNet dataset. In contrast, benefiting from directly learning pose refinement with plane parameters in embedding space, our NOPE-SAC avoids the problem of keypoints detection and matching and achieves state-of-the-art performance. TABLE VI: Ablation studies of the arbitrary initialization module and plane warping in NOPE-SAC pose estimation on the Matterport3D [13] and the ScanNet [14] datasets. Settings | Trans. | Rot. ---|---|--- AIM | Warp Plane | Med. $\downarrow$ | Mean $\downarrow$ | ($\leq$0.5m) $\uparrow$ | Med. $\downarrow$ | Mean $\downarrow$ | ($\leq 15^{\circ}$) $\uparrow$ Matterport3D dataset $\checkmark$ | | 0.61 | 1.00 | 42.3% | 3.19 | 14.10 | 85.7% | $\checkmark$ | 0.51 | 0.92 | 49.3% | 2.97 | 14.34 | 86.8% $\checkmark$ | $\checkmark$ | 0.52 | 0.94 | 48.3% | 2.77 | 14.37 | 86.9% ScanNet dataset $\checkmark$ | | 0.42 | 0.66 | 57.4% | 8.17 | 22.05 | 73.2% | $\checkmark$ | 0.40 | 0.64 | 59.9% | 8.21 | 21.95 | 73.0% $\checkmark$ | $\checkmark$ | 0.41 | 0.65 | 59.1% | 8.27 | 22.12 | 73.2% #### Arbitrary Initialization Module. In this part, we analyze the necessity of our arbitrary initialization module (AIM). As introduced in Sec. 4.1, the initial pose embeddings can be achieved from (1) the pose regression network or (2) the AIM. Thus, we first compare the influence of these two methods to pose refinement. As shown in Tab. VI, using initial pose embeddings from AIM achieves similar results to embeddings from the regression network, which demonstrates the effectiveness of our AIM. Furthermore, with the usage of AIM, our NOPE-SAC is flexible to refine initial poses that do not come from the convolutional pose regression network. #### Plane Warping for Correspondence Embedding. Here, we discuss the influence of warping plane parameters before calculating correspondence embeddings as described in Eqn. 11. As shown in Tab. VI, warping plane parameters effectively improves the translation results, especially on the Matterport3D [13] dataset (_e.g.,_ from $42.3\%$ to $48.3\%$). #### Pose Hypotheses Fusion. We then discuss the fusion strategies of one-plane pose hypotheses for pose refinement in Sec. 4.2, including (1)Soft, (2)Avg, (3)Min-Cost, (4)Max-Score. As shown in Tab. VII, all strategies, even selecting only one pose hypothesis (Min-Cost and Max-Score), can improve the initial pose, which demonstrates the effectiveness and the reasonability of our one-plane RANSAC. Specifically, when evaluating translations, Avg performs closely to Soft, while Min-Cost and Max-Score degenerate significantly on both two datasets. It indicates that it is necessary to leverage more than one hypothesis to get better translation refinement results. When evaluating rotations, both Min-Cost and Max-Score perform closely to or slightly better than Soft, while Avg degenerates significantly. It indicates that rotations can be refined more easily than translations with one rotation hypothesis. However, because of the influence of matching outliers and the plane parameter errors, the strategies of Avg, Min-Cost, and Max-Score are not stable for camera pose estimation. By contrast, Soft achieves the best overall performance benefiting from the learned pose scores. TABLE VII: Ablation studies of NOPE-SAC pose estimation with different one plane pose fusion strategies on the Matterport3D [13] and the ScanNet [14] datasets. Method | Trans. | Rot. ---|---|--- Med. $\downarrow$ | Mean $\downarrow$ | ($\leq$0.5m) $\uparrow$ | Med. $\downarrow$ | Mean $\downarrow$ | ($\leq 15^{\circ}$) $\uparrow$ Matterport3D dataset Initial Pose | 0.69 | 1.08 | 37.0% | 11.16 | 21.49 | 60.5% Avg | 0.56 | 0.96 | 45.3% | 5.03 | 17.21 | 77.1% Min-Cost | 0.56 | 1.00 | 45.4% | 2.85 | 14.47 | 86.8% Max-Score | 0.59 | 1.03 | 43.1% | 2.84 | 14.38 | 87.0% Soft | 0.52 | 0.94 | 48.3% | 2.77 | 14.37 | 86.9% ScanNet dataset Initial Pose | 0.48 | 0.72 | 51.9% | 14.68 | 26.75 | 51.0% Avg | 0.42 | 0.65 | 57.8% | 10.28 | 23.33 | 63.2% Min-Cost | 0.46 | 0.73 | 53.6% | 8.75 | 22.68 | 72.1% Max-Score | 0.46 | 0.72 | 53.0% | 8.77 | 22.91 | 72.4% Soft | 0.41 | 0.65 | 59.1% | 8.27 | 22.12 | 73.2% TABLE VIII: Influence of plane parameter accuracy to NOPE-SAC pose estimation on the Matterport3D [13] and the ScanNet [14] datasets. Setting | Trans. | Rot. ---|---|--- Med. $\downarrow$ | Mean $\downarrow$ | ($\leq$0.5m) $\uparrow$ | Med. $\downarrow$ | Mean $\downarrow$ | ($\leq 15^{\circ}$) $\uparrow$ Matterport3D dataset Initial Pose | 0.69 | 1.08 | 37.0% | 11.16 | 21.49 | 60.5% w/ GT | 0.32 | 0.70 | 63.5% | 0.34 | 3.85 | 96.2% (0.1m, $5^{\circ}$) | 0.39 | 0.76 | 58.6% | 3.27 | 6.85 | 95.4% (0.2m, $10^{\circ}$) | 0.48 | 0.83 | 51.5% | 6.75 | 10.76 | 83.6% (0.3m, $15^{\circ}$) | 0.56 | 0.89 | 45.3% | 10.40 | 15.17 | 66.0% ScanNet dataset Initial Pose | 0.48 | 0.72 | 51.9% | 14.68 | 26.75 | 51.0% w/ GT | 0.26 | 0.47 | 72.1% | 3.48 | 8.42 | 90.7% (0.1m, $5^{\circ}$) | 0.29 | 0.49 | 70.4% | 7.22 | 11.63 | 88.2% (0.2m, $10^{\circ}$) | 0.35 | 0.55 | 65.4% | 11.95 | 16.35 | 65.4% (0.3m, $15^{\circ}$) | 0.40 | 0.60 | 59.9% | 16.35 | 21.11 | 44.3% TABLE IX: Influence of the plane matching precision (P) to NOPE-SAC pose estimation on the Matterport3D [13] and the ScanNet [14] datasets. Threshold | P | Translation | Rotation ---|---|---|--- Med. $\downarrow$ | Mean $\downarrow$ | ($\leq$0.5m) $\uparrow$ | Med. $\downarrow$ | Mean $\downarrow$ | ($\leq 15^{\circ}$) $\uparrow$ Matterport3D dataset | 0.2 | 49.9 | 0.52 | 0.94 | 48.3% | 2.77 | 14.37 | 86.9% 0.1 | 48.8 | 0.52 | 0.94 | 48.4% | 2.75 | 14.40 | 86.9% 0.01 | 46.4 | 0.53 | 0.94 | 47.9% | 2.76 | 14.63 | 86.7% 0.001 | 44.7 | 0.53 | 0.94 | 47.7% | 2.77 | 14.96 | 86.3% ScanNet dataset | 0.2 | 44.3 | 0.41 | 0.66 | 59.1% | 8.27 | 22.12 | 73.2% 0.1 | 43.0 | 0.41 | 0.65 | 58.9% | 8.26 | 22.11 | 73.5% 0.01 | 40.8 | 0.41 | 0.65 | 58.6% | 8.21 | 22.36 | 73.2% 0.001 | 39.8 | 0.41 | 0.65 | 58.5% | 8.22 | 22.56 | 73.2% #### Influence of Plane Parameters. In this part, we study the influence of plane parameter accuracy on our NOPE- SAC pose estimation on the Matterport3D [13] and the ScanNet [14] datasets. The upper bound of our method is achieved by using ground truth plane correspondences and ground truth plane parameters in our pose refinement, defined as ‘w/ GT’ in Tab. VIII. Then, we add various Gaussian noises to the offset and normal of ground truth plane parameters in both two image views. The mean of the Gaussian noises is set to zero. The standard deviation of plane offset increases from 0.1m to 0.3m, and the standard deviation of plane normal increases from $5^{\circ}$ to $15^{\circ}$. As shown in Tab. VIII, our method effectively improves the initial pose even in the challenging setting of (0.2m, $10^{\circ}$), but fails to improve rotations in the setting of (0.3m, $15^{\circ}$) on the ScanNet dataset because of too large noises on plane parameters. It demonstrates that our method is robust to the accuracy of plane parameters. #### Influence of Matching Precision. Besides plane parameters, we also evaluate the influence of plane matching precision to our NOPE-SAC on the Matterport3D [13] and ScanNet [14] datasets. Here, a predicted plane correspondence is regarded as a true positive if it can be matched to a ground truth plane correspondence with mask IoU $\geq$ 0.5 in both two images. We conduct the experiments by gradually reducing the threshold of plane matching from 0.2 to 0.001. As shown in Tab. IX, with the reduction of the matching threshold, the plane matching precision reduces from 49.9 to 44.7 while the pose metrics only change slightly on the Matterport3D dataset. Similar results can be found on the ScanNet dataset. It demonstrates that our NOPE-SAC is able to cope with incorrect plane correspondences. ## 6 Conclusion This paper studies the challenging two-view 3D reconstruction in a rigorous sparse-view configuration. At the core of estimating the camera poses from one-plane pose hypotheses generated from plane correspondences with neural networks, our proposed NOPE-SAC formulates the problem in a consensus sampling paradigm while enjoying end-to-end learning without incurring any offline optimization schemes. In the experiments, our NOPE-SAC achieves new state-of- the-art performances for sparse-view camera pose estimation and planar 3D reconstruction in indoor scenes on the challenging Matterport3D [13] and ScanNet [14] datasets. The comprehensive analyzes also align benchmarking protocols for sparse-view 3D reconstruction. ## References * [1] A. Harltey and A. Zisserman, _Multiple View Geometry in Computer Vision (2. ed.)_. Cambridge University Press, 2006. * [2] D. J. Crandall, A. Owens, N. Snavely, and D. P. Huttenlocher, “Sfm with mrfs: Discrete-continuous optimization for large-scale structure from motion,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 35, no. 12, pp. 2841–2853, 2013. * [3] A. J. Davison, I. D. Reid, N. Molton, and O. Stasse, “Monoslam: Real-time single camera SLAM,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 29, no. 6, pp. 1052–1067, 2007. * [4] R. Mur-Artal, J. M. M. Montiel, and J. D. Tardós, “ORB-SLAM: A versatile and accurate monocular SLAM system,” _IEEE Trans. Robotics_ , vol. 31, no. 5, 2015. * [5] J. L. Schönberger and J. Frahm, “Structure-from-motion revisited,” in _IEEE Conf. Comput. Vis. Pattern Recog._ , 2016, pp. 4104–4113. * [6] S. Qian, L. Jin, and D. F. Fouhey, “Associative3d: Volumetric reconstruction from sparse views,” in _Eur. Conf. Comput. Vis._ , vol. 12360, 2020, pp. 140–157. * [7] L. Jin, S. Qian, A. Owens, and D. F. Fouhey, “Planar surface reconstruction from sparse views,” in _IEEE Int. Conf. Comput. Vis._ , 2021. * [8] S. Agarwala, L. Jin, C. Rockwell, and D. F. Fouhey, “Planeformers: From sparse view planes to 3d reconstruction,” in _Eur. Conf. Comput. Vis._ , vol. 13663, 2022, pp. 192–209. * [9] P. Sarlin, D. DeTone, T. Malisiewicz, and A. Rabinovich, “Superglue: Learning feature matching with graph neural networks,” in _IEEE Conf. Comput. Vis. Pattern Recog._ , 2020, pp. 4937–4946. * [10] D. Nistér, “An efficient solution to the five-point relative pose problem,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 26, no. 6, pp. 756–777, 2004. * [11] M. A. Fischler and R. C. Bolles, “Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography,” _Commun. ACM_ , vol. 24, no. 6, pp. 381–395, 1981. * [12] C. Raposo, M. Lourenço, M. Antunes, and J. P. Barreto, “Plane-based odometry using an RGB-D camera,” in _Brit. Mach. Vis. Conf._ , 2013. * [13] A. X. Chang, A. Dai, T. A. Funkhouser, M. Halber, M. Nießner, M. Savva, S. Song, A. Zeng, and Y. Zhang, “Matterport3d: Learning from RGB-D data in indoor environments,” in _Int. Conf. 3D Vis._ , 2017, pp. 667–676. * [14] A. Dai, A. X. Chang, M. Savva, M. Halber, T. A. Funkhouser, and M. Nießner, “Scannet: Richly-annotated 3d reconstructions of indoor scenes,” in _IEEE Conf. Comput. Vis. Pattern Recog._ , 2017, pp. 2432–2443. * [15] C. Godard, O. M. Aodha, M. Firman, and G. J. Brostow, “Digging into self-supervised monocular depth estimation,” in _IEEE Int. Conf. Comput. Vis._ , 2019, pp. 3827–3837. * [16] K. Lasinger, R. Ranftl, K. Schindler, and V. Koltun, “Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer,” _IEEE Trans. Pattern Anal. Mach. Intell._ , 2019. * [17] T. Chen, S. An, Y. Zhang, C. Ma, H. Wang, X. Guo, and W. Zheng, “Improving monocular depth estimation by leveraging structural awareness and complementary datasets,” in _Eur. Conf. Comput. Vis._ , vol. 12359, 2020, pp. 90–108. * [18] W. Yin, Y. Liu, C. Shen, and Y. Yan, “Enforcing geometric constraints of virtual normal for depth prediction,” in _IEEE Int. Conf. Comput. Vis._ , 2019, pp. 5683–5692. * [19] Z. Li and N. Snavely, “Megadepth: Learning single-view depth prediction from internet photos,” in _IEEE Conf. Comput. Vis. Pattern Recog._ , 2018, pp. 2041–2050. * [20] A. Eftekhar, A. Sax, J. Malik, and A. Zamir, “Omnidata: A scalable pipeline for making multi-task mid-level vision datasets from 3d scans,” in _IEEE Int. Conf. Comput. Vis._ , 2021, pp. 10 766–10 776. * [21] R. Ranftl, K. Lasinger, D. Hafner, K. Schindler, and V. Koltun, “Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer,” _IEEE Trans. Pattern Anal. Mach. Intell._ , vol. 44, no. 3, pp. 1623–1637, 2022. * [22] Z. Yu, J. Zheng, D. Lian, Z. Zhou, and S. Gao, “Single-image piece-wise planar 3d reconstruction via associative embedding,” in _IEEE Conf. Comput. Vis. Pattern Recog._ , 2019, pp. 1029–1037. * [23] C. Liu, K. Kim, J. Gu, Y. Furukawa, and J. Kautz, “Planercnn: 3d plane detection and reconstruction from a single image,” in _IEEE Conf. Comput. Vis. Pattern Recog._ , 2019, pp. 4450–4459. * [24] C. Liu, J. Yang, D. Ceylan, E. Yumer, and Y. Furukawa, “Planenet: Piece-wise planar reconstruction from a single RGB image,” in _IEEE Conf. Comput. Vis. Pattern Recog._ , 2018. * [25] B. Tan, N. Xue, S. Bai, T. Wu, and G. Xia, “Planetr: Structure-guided transformers for 3d plane recovery,” in _IEEE Int. Conf. Comput. Vis._ , 2021, pp. 4166–4175. * [26] D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” _Int. J. Comput. Vis._ , vol. 60, no. 2, pp. 91–110, 2004. * [27] E. Rublee, V. Rabaud, K. Konolige, and G. R. Bradski, “ORB: an efficient alternative to SIFT or SURF,” in _IEEE Int. Conf. Comput. Vis._ , 2011, pp. 2564–2571. * [28] H. Bay, A. Ess, T. Tuytelaars, and L. V. Gool, “Speeded-up robust features (SURF),” _Comput. Vis. Image Underst._ , vol. 110, no. 3, pp. 346–359, 2008. * [29] S. Leutenegger, M. Chli, and R. Siegwart, “BRISK: binary robust invariant scalable keypoints,” in _IEEE Int. Conf. Comput. Vis._ , 2011, pp. 2548–2555. * [30] J. Morel and G. Yu, “ASIFT: A new framework for fully affine invariant image comparison,” _SIAM J. Imaging Sci._ , vol. 2, no. 2, pp. 438–469, 2009. * [31] J. Revaud, C. R. de Souza, M. Humenberger, and P. Weinzaepfel, “R2D2: reliable and repeatable detector and descriptor,” in _Adv. Neural Inform. Process. Syst._ , 2019, pp. 12 405–12 415. * [32] D. DeTone, T. Malisiewicz, and A. Rabinovich, “Superpoint: Self-supervised interest point detection and description,” in _IEEE Conf. Comput. Vis. Pattern Recog. Worksh._ , 2018, pp. 224–236. * [33] Z. Luo, T. Shen, L. Zhou, J. Zhang, Y. Yao, S. Li, T. Fang, and L. Quan, “Contextdesc: Local descriptor augmentation with cross-modality context,” in _IEEE Conf. Comput. Vis. Pattern Recog._ , 2019, pp. 2527–2536. * [34] P. Lindenberger, P. Sarlin, V. Larsson, and M. Pollefeys, “Pixel-perfect structure-from-motion with featuremetric refinement,” in _IEEE Int. Conf. Comput. Vis._ , 2021, pp. 5967–5977. * [35] W. Jiang, E. Trulls, J. Hosang, A. Tagliasacchi, and K. M. Yi, “COTR: correspondence transformer for matching across images,” in _IEEE Int. Conf. Comput. Vis._ , 2021, pp. 6187–6197. * [36] X. Wei, Y. Zhang, Z. Li, Y. Fu, and X. Xue, “Deepsfm: Structure from motion via deep bundle adjustment,” in _Eur. Conf. Comput. Vis._ , vol. 12346, 2020, pp. 230–247. * [37] A. Kendall, M. Grimes, and R. Cipolla, “Posenet: A convolutional network for real-time 6-dof camera relocalization,” in _IEEE Int. Conf. Comput. Vis._ , 2015, pp. 2938–2946. * [38] P. Ji, R. Li, B. Bhanu, and Y. Xu, “Monoindoor: Towards good practice of self-supervised monocular depth estimation for indoor environments,” in _IEEE Int. Conf. Comput. Vis._ , 2021, pp. 12 767–12 776. * [39] S. En, A. Lechervy, and F. Jurie, “Rpnet: An end-to-end network for relative camera pose estimation,” in _Eur. Conf. Comput. Vis._ , vol. 11129, 2018, pp. 738–745. * [40] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in _IEEE Conf. Comput. Vis. Pattern Recog._ , 2016, pp. 770–778. * [41] F. Milletari, N. Navab, and S. Ahmadi, “V-net: Fully convolutional neural networks for volumetric medical image segmentation,” in _Int. Conf. 3D Vis._ , 2016, pp. 565–571. * [42] B. Cheng, A. G. Schwing, and A. Kirillov, “Per-pixel classification is not all you need for semantic segmentation,” in _Adv. Neural Inform. Process. Syst._ , 2021, pp. 17 864–17 875. * [43] R. Sinkhorn and P. Knopp, “Concerning nonnegative matrices and doubly stochastic matrices,” _Pac. J. Math._ , vol. 21, no. 2, pp. 343–348, 1967\. * [44] G. Peyré and M. Cuturi, “Computational optimal transport,” _Found. Trends Mach. Learn._ , vol. 11, no. 5-6, pp. 355–607, 2019. * [45] Y. Wu, A. Kirillov, F. Massa, W.-Y. Lo, and R. Girshick, “Detectron2,” https://github.com/facebookresearch/detectron2, 2019. * [46] I. Loshchilov and F. Hutter, “Decoupled weight decay regularization,” in _Int. Conf. Learn. Represent._ , 2019.
Probing New Physics in the Vector-like Lepton Model by Lepton Electric Dipole Moments Koichi Hamaguchia,b***<EMAIL_ADDRESS>Natsumi Nagataa††† <EMAIL_ADDRESS>Genta Osakia‡‡‡ osaki@hep- th.phys.s.u-tokyo.ac.jp , and Shih-Yen Tsenga§§§ shihyen@hep- th.phys.s.u-tokyo.ac.jp aDepartment of Physics, University of Tokyo, Bunkyo-ku, Tokyo 113–0033, Japan bKavli IPMU (WPI), University of Tokyo, Kashiwa, Chiba 277–8583, Japan We examine the lepton dipole moments in an extension of the Standard Model (SM), which contains vector-like leptons that couple only to the second- generation SM leptons. The model naturally leads to sizable contributions to the muon $g-2$ and the muon electric dipole moment (EDM). One feature of this model is that a sizable electron EDM is also induced at the two-loop level due to the existence of new vector-like leptons in the loops. We find parameter regions that can explain the muon $g-2$ anomaly and are also consistent with the experimental constraints coming from the electron EDM and the Higgs decay $h\rightarrow\mu^{+}\mu^{-}$. The generated EDMs can be as large as $\mathcal{O}(10^{-22})~{}e\cdot\mathrm{cm}$ for the muon and $\mathcal{O}(10^{-30})~{}e\cdot\mathrm{cm}$ for the electron, respectively, which can be probed in future experiments for the EDM measurements. ## 1 Introduction Last year, Fermilab published their first result [1] on the measurement of the muon anomalous magnetic moment $a_{\mu}\equiv\frac{g_{\mu}-2}{2}~{},$ (1) which gives a value of $a_{\mu}(\mathrm{exp})=116592061(41)\times 10^{-11}~{},$ (2) while the Standard Model (SM) prediction is [2] $a_{\mu}(\mathrm{SM})=116591810(43)\times 10^{-11}~{}.$ (3) There is a $4.2\sigma$ tension between the experiment and theory, $\Delta a_{\mu}=a_{\mu}(\mathrm{exp})-a_{\mu}(\mathrm{SM})=251(59)\times 10^{-11}~{},$ (4) which may indicate the existence of physics beyond the Standard Model (BSM)111The hadronic vacuum polarization (HVP) contribution to the muon $g-2$ has been a challenge for theoretical calculations. The value obtained in the data-driven method, which is adopted in Ref. [2], is $a^{\mathrm{HVP}}_{\mu}=6845(40)\times 10^{-11}$. A discrepancy exists between this value and the lattice QCD calculations performed by the Budapest- Marseille-Wuppertal (BMW) group [3], which gives $a^{\mathrm{HVP,LO}}_{\mu,\mathrm{BMW}}=7075(55)\times 10^{-11}$. This is $2.1\sigma$ larger than the recommended data-driven result. Naturally, the two values should be compatible with each other because they correspond to the same physical processes in the SM. However, the current situation is that there is a significant difference between the two approaches, and the reason is not yet clear. More intriguingly, recent results from other lattice QCD groups [4, 5] support the result obtained by the BMW group. For the time being, we simply fix on the value given in Eq. (3).. Various new physics models have been proposed to explain this deviation. In general, these models contain hypothetical new particles and couplings, whose corresponding parameters are complex, and thus contain complex phases that break the $CP$ symmetry. It is well-known that the flavor-conserving $CP$ violation in the SM is very small, such that the induced particle electric dipole moments (EDMs) are vanishingly small. The non-zero SM contributions to lepton EDMs appear at the four-loop level and are thus strongly suppressed. For example, the electron EDM is estimated to be $\left|d^{\mathrm{SM}}_{e}\right|\leq 10^{-38}~{}e\cdot\mathrm{cm}$ [6]. Since it is far below the sensitivity of the current experimental techniques, any observation of a particle EDM will be an unambiguous sign of the new physics beyond the SM. Currently, the upper bound on the muon EDM is $\left|d_{\mu}\right|<1.8\times 10^{-19}~{}e\cdot\mathrm{cm}~{}\left(95\%~{}\mathrm{C.L.}\right)$ (5) set by the Muon $\left(g-2\right)$ Collaboration at Brookhaven National Laboratory [7], which is about ten orders of magnitude weaker than the one on the electron EDM, $\left|d_{e}\right|<1.1\times 10^{-29}~{}e\cdot\mathrm{cm}~{}\left(90\%~{}\mathrm{C.L.}\right)$ (6) set by the ACME Collaboration [8]. In order to improve the sensitivity on the muon EDM, there are several future experiments proposed to measure the muon EDM. For example, J-PARC Muon $g-2$/EDM experiment [9] and the one using the frozen-spin technique at the Paul Scherrer Institute (PSI) [10] will have sensitivities of $\sigma\left(d_{\mu}\right)\leq 1.5\times 10^{-21}~{}e\cdot\mathrm{cm}$ and $\sigma\left(d_{\mu}\right)\leq 6\times 10^{-23}~{}e\cdot\mathrm{cm}$, respectively. In this paper, we consider a model with extra vector-like leptons (VLLs) as a possible explanation of the muon $g-2$ deviation, and investigate the EDMs of the muon and electron in the model. Models with VLLs have been discussed previously as solutions to the muon $g-2$ anomaly [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28]. See also Refs. [29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46] for the works that studied the muon EDMs in the models motivated by the muon $g-2$ anomaly. We consider a simple extension to the SM with two vector-like leptons, one SU(2)L doublet and one SU(2)L singlet. We show that the model can naturally induce sizable EDMs of the muon and the electron at the one-loop and two-loop levels, respectively, in the parameter regions motivated by the muon $g-2$; the predicted EDMs can be as large as $|d_{\mu}|\sim 10^{-22}~{}e\cdot\mathrm{cm}$ and $|d_{e}|\sim 10^{-30}~{}e\cdot\mathrm{cm}$, which are within the reach of the proposed EDM experiments. There are also researches discussing the indirect constraints on the muon EDM extracted from the EDM measurements of heavy atoms and molecules; for example, see Ref. [47]. This paper is structured as follows. In section 2, we describe the model used in the analysis. In section 3, we summarize the calculation of the one-loop contributions to the muon dipole moments. In section 4, the induced electron EDM in this model is discussed. The experimental constraints on this model are presented in section 5, and the results are given in section 6. Finally, we summarize the study in section 7. ## 2 Model | $\ell_{L}$ | $\mu_{R}$ | $H$ | $L_{L,R}$ | $E_{L,R}$ ---|---|---|---|---|--- SU(3)C | 1 | 1 | 1 | 1 | 1 SU(2)L | 2 | 1 | 2 | 2 | 1 U(1)Y | $-\frac{1}{2}$ | $-1$ | $\frac{1}{2}$ | $-\frac{1}{2}$ | $-1$ Table 1: The quantum numbers of the SM and extra vector-like leptons in our model. We consider an extension of the SM with one SU(2)L doublet ($L$) and one SU(2)L singlet ($E$) vector-like leptons. For simplicity, we consider the minimal scenario where the vector-like leptons couple only to the second- generation lepton, not to the first- and third-generation leptons.222Such kind of structure may be realized by imposing flavor symmetries on the model. We discuss a specific example in appendix B. Therefore, the electron and tau do not mix with extra vector-like leptons, and their masses totally originate from the Higgs Yukawa couplings. The quantum numbers of the fields necessary for the analysis are listed in Table 1, where $\ell_{L}$ and $\mu_{R}$ are the second-generation leptons in the SM and $L_{L,R}$ and $E_{L,R}$ denote the vector-like leptons, which respectively have the same quantum numbers as can be seen from the table. In the discussion below, our notation basically follows the one used in Ref. [13]. The components of the doublet fields are labeled as $\displaystyle\ell_{L}=\begin{pmatrix}\nu_{\mu}\\\ \mu_{L}\end{pmatrix},~{}L_{L,R}=\begin{pmatrix}L^{0}_{L,R}\\\ L^{-}_{L,R}\end{pmatrix},~{}H=\frac{1}{\sqrt{2}}\begin{pmatrix}0\\\ v+h\end{pmatrix},$ (7) where $v=246.22$ GeV is the vacuum expectation value of the Higgs field. In the rest of this paper, we denote the indices of muon and two vector-like leptons in the mass basis as $2,4,5$, respectively, and for the muon, we use the symbol $\mu$ and $2$ interchangeably. Without loss of generality, we work in the basis where the Yukawa matrix of the leptons in the SM sector is already diagonal. The most relevant part of the Lagrangian is the Yukawa interactions among the muon and the vector-like leptons and the mass term of the vector-like leptons, which are given by333In general, the other mass terms $M^{\prime}_{L}\bar{\ell}_{L}L_{R}$ and $M^{\prime}_{R}\bar{E}_{L}\mu_{R}$ are also allowed, but they can be removed by field redefinitions. $\displaystyle\begin{split}\mathcal{L}\supset&-y_{\mu}\bar{\ell}_{L}\mu_{R}H-\lambda_{E}\bar{\ell}_{L}E_{R}H-\lambda_{L}\bar{L}_{L}\mu_{R}H-\lambda\bar{L}_{L}E_{R}H-\bar{\lambda}H^{{\dagger}}\bar{E}_{L}L_{R}\\\ &-M_{L}\bar{L}_{L}L_{R}-M_{E}\bar{E}_{L}E_{R}+\mathrm{h.c.}.\end{split}$ (8) The parameters $y_{\mu},\lambda_{E},\lambda_{L},\lambda,\bar{\lambda},M_{L}$ and $M_{E}$ are in general complex. However, most of the complex phases are not physical since they can be removed via field redefinitions. It turns out that in the model there are two independent complex phases $\displaystyle\phi_{\lambda}$ $\displaystyle=\arg\left(y_{\mu}\lambda^{\ast}_{L}\lambda^{\ast}_{E}\lambda\right)~{},$ (9) $\displaystyle\phi_{\bar{\lambda}}$ $\displaystyle=\arg\left(y_{\mu}\lambda_{L}^{*}\lambda_{E}^{*}M_{L}M_{E}\bar{\lambda}^{*}\right)~{},$ (10) which are invariant under the phase rotation of the fields and therefore can serve as the sources of new $CP$ violation. In this work, we take the two $CP$ phases to be the phases of $\lambda$ and $\bar{\lambda}$ and set the other parameters to be real. After the Higgs field develops a vacuum expectation value, the leptons acquire masses and the mass matrix of charged leptons is given by $\displaystyle\bar{f}_{L}Mf_{R}=(\bar{\mu}_{L},\bar{L}^{-}_{L},\bar{E}_{L})\begin{pmatrix}y_{\mu}v/\sqrt{2}&0&\lambda_{E}v/\sqrt{2}\\\ \lambda_{L}v/\sqrt{2}&M_{L}&\lambda v/\sqrt{2}\\\ 0&\bar{\lambda}v/\sqrt{2}&M_{E}\end{pmatrix}\begin{pmatrix}\mu_{R}\\\ L^{-}_{R}\\\ E_{R}\end{pmatrix},$ (11) where the leptons in the flavor eigenbasis are denoted collectively as $f_{L}=(\mu_{L},L^{-}_{L},E_{L})^{T}$ and $f_{R}=(\mu_{R},L^{-}_{R},E_{R})^{T}$. We can diagonalize the mass matrix in Eq. (11) by a bi-unitary transformation with two unitary matrices $U_{L}$ and $U_{R}$: $\displaystyle U^{\dagger}_{L}\begin{pmatrix}y_{\mu}v&0&\lambda_{E}v\\\ \lambda_{L}v&M_{L}&\lambda v\\\ 0&\bar{\lambda}v&M_{E}\end{pmatrix}U_{R}=\begin{pmatrix}m_{2}&0&0\\\ 0&m_{4}&0\\\ 0&0&m_{5}\\\ \end{pmatrix},$ (12) where the eigenmass $m_{2}$ is set to be the mass of muon, $m_{2}=m_{\mu}$, and the mass ordering is fixed as $m_{2}<m_{4}<m_{5}$. We also note that the mass of the neutral component of the doublet $L$, $L^{0}\equiv\nu_{4}$, solely originates from the mass term $-M_{L}\bar{L}_{L}L_{R}$ and hence its mass is determined by the value of $M_{L}$. ## 3 Muon dipole moments In this section, we summarize the one-loop contributions to the dipole moments of the muon, which are induced by the diagrams of the Higgs, $Z$ boson and $W$ boson mediations, as shown in Fig. 1. For the relevant interactions, see appendix A. Figure 1: One-loop contributions to the muon dipole moments. $\ell_{2,4,5}$ are the muon and two vector-like leptons in the mass basis, while $\nu_{2,4}$ are the muon neutrino and the heavy neutrino in the vector-like doublet, respectively. ### 3.1 Higgs boson mediation The contributions to the moun dipole moments from the Higgs boson mediation are given by $\displaystyle\Delta a_{\mu}^{h}$ $\displaystyle=\frac{m_{\mu}}{8\pi^{2}m_{h}^{2}}\sum_{i=2,4,5}\left[(|\lambda_{i2}|^{2}+|\lambda_{2i}|^{2})\,m_{\mu}f_{h}(r_{i})+{\rm Re}\,(\lambda_{i2}\lambda_{2i})\,m_{i}g_{h}(r_{i})\right]-a^{h,\mathrm{SM}}_{\mu},$ (13) $\displaystyle d_{\mu}^{h}$ $\displaystyle=-\frac{e}{16\pi^{2}m_{h}^{2}}\sum_{i=2,4,5}{\rm Im}(\lambda_{i2}\lambda_{2i})m_{i}g_{h}(r_{i}),$ (14) where $a^{h,\mathrm{SM}}_{\mu}$ is the SM contribution from the diagram with muons in the loop. The loop functions are $\displaystyle g_{h}(r_{i})$ $\displaystyle=-\frac{r_{i}^{2}-4r_{i}+3+2\,\mathrm{ln}(r_{i})}{2(1-r_{i})^{3}},$ (15) $\displaystyle f_{h}(r_{i})$ $\displaystyle=\frac{r_{i}^{3}-6r_{i}^{2}+3r_{i}+2+6r_{i}\,\mathrm{ln}(r_{i})}{12(1-r_{i})^{4}},$ (16) with $r_{i}=m^{2}_{i}/m^{2}_{h}$ and $i=2,4,5$. ### 3.2 $Z$ boson mediation The contributions to the moun dipole moments from the $Z$ boson mediation are given by $\displaystyle\Delta a_{\mu}^{Z}$ $\displaystyle=\frac{m_{\mu}}{8\pi^{2}m_{Z}^{2}}\sum_{i=2,4,5}\big{\\{}\left[|(g^{Z}_{L})_{i2}|^{2}+|(g^{Z}_{R})_{i2}|^{2}\right]\,m_{\mu}f_{Z}(r_{i})$ $\displaystyle\qquad\qquad\qquad\qquad+{\rm Re}\,\left[(g^{Z}_{L})_{i2}(g^{Z\ast}_{R})_{i2}\right]\,m_{i}g_{Z}(r_{i})\big{\\}}-a^{Z,\mathrm{SM}}_{\mu},$ (17) $\displaystyle d_{\mu}^{Z}$ $\displaystyle=\frac{e}{16\pi^{2}m^{2}_{Z}}\sum_{i=2,4,5}{\rm Im}\,\left[(g^{Z}_{L})_{i2}(g^{Z\ast}_{R})_{i2}\right]\,m_{i}g_{Z}(r_{i}),$ (18) where $a^{Z,\mathrm{SM}}_{\mu}$ is the SM contribution from the diagram with muons in the loop. The loop functions are then given by $\displaystyle g_{Z}(r_{i})$ $\displaystyle=-\frac{r_{i}^{3}+3r_{i}-4-6r_{i}\,\mathrm{ln}(r_{i})}{2(1-r_{i})^{3}},$ (19) $\displaystyle f_{Z}(r_{i})$ $\displaystyle=-\frac{5r_{i}^{4}-14r_{i}^{3}+39r^{2}_{i}-38r_{i}+8-18r^{2}_{i}\,\mathrm{ln}(r_{i})}{12(1-r_{i})^{4}},$ (20) with $r_{i}=m^{2}_{i}/m^{2}_{Z}$ and $i=2,4,5$. ### 3.3 $W$ boson mediation The contributions to the muon dipole moments from the $W$ boson mediation are given by $\displaystyle\Delta a^{W}_{\mu}$ $\displaystyle=\frac{m_{\mu}}{8\pi^{2}m_{W}^{2}}\big{\\{}\sum_{i=2,4}\left[|(g^{W}_{L})_{i2}|^{2}+|(g^{W}_{R})_{i2}|^{2}\right]m_{\mu}f_{W}(r_{i})$ $\displaystyle\quad\qquad\qquad\qquad+{\rm Re}\,\left[(g^{W}_{L})_{42}(g^{W\ast}_{R})_{42}\right]\,M_{L}g_{W}(r_{4})\big{\\}}-a^{W,\mathrm{SM}}_{\mu},$ (21) $\displaystyle d^{W}_{\mu}$ $\displaystyle=\frac{e}{16\pi^{2}m^{2}_{W}}{\rm Im}\,\left[(g^{W}_{L})_{42}(g^{W\ast}_{R})_{42}\right]\,M_{L}g_{W}(r_{4}),$ (22) where $a^{W,\mathrm{SM}}_{\mu}$ is the SM contribution from the diagram with muon neutrino in the loop. The loop functions are given by $\displaystyle g_{W}(r_{i})$ $\displaystyle=\frac{r^{3}_{i}-12r^{2}_{i}+15r_{i}-4+6r^{2}_{i}\,\mathrm{ln}(r_{i})}{2(1-r_{i})^{3}},$ (23) $\displaystyle f_{W}(r_{i})$ $\displaystyle=\frac{4r^{4}_{i}-49r^{3}_{i}+78r^{2}_{i}-43r_{i}+10+18r^{3}_{i}\,\mathrm{ln}(r_{i})}{12(1-r_{i})^{4}},$ (24) where $r_{2}=m^{2}_{\nu_{\mu}}/m^{2}_{W}$ assuming $m_{\nu_{\mu}}\simeq 0$, and $r_{4}=M^{2}_{L}/m^{2}_{W}$. ## 4 Electron EDM In this model, the mixings among the muon and the extra vector-like leptons also contribute to the electron EDM. These contributions appear in the internal lepton loop of the two-loop Barr-Zee diagrams [48] shown in Fig. 2, where all possible combinations of muon and vector-like leptons contribute to the inner loop, as shown in Fig. 3. Figure 2: Contributions to the electron EDM from the Barr-Zee diagram induced by the vector-like leptons. Figure 3: The inner loop inside the blob of Fig. 2. To obtain the value of the electron EDM, let us first discuss the left diagram in Fig. 2. The first step is to extract the effective vertex $\Gamma^{\mu\nu}_{hV\gamma}$ from the inner loop. By the gauge invariance of the on-shell photon, $q_{\mu}\Gamma^{\mu\nu}=0$, the effective vertex must have the form of [49] $\displaystyle\Gamma^{\mu\nu,ij}_{hV\gamma}(q,k)=\int^{1}_{0}dx\frac{eQ}{4\pi^{2}D^{ij}}\left[c_{E}^{Vij}(q^{\nu}k^{\mu}-q\cdot k\,g^{\mu\nu})+ic_{O}^{Vij}\epsilon^{\mu\nu\alpha\beta}q_{\alpha}k_{\beta}\right],$ (25) where $V=\gamma$, $Z$, and $Q$ is the charge of the lepton coupled to the external photon which is $-1$ for all cases considered here. The coefficients $c_{E}^{Vij}$, $c_{O}^{Vij}$, and the $\Delta^{ij}$ in the denominator are given by $\displaystyle c^{Vij}_{E}$ $\displaystyle=m_{i}x^{2}(1-x)(g^{ij\ast}_{s}g^{Vij}_{v}+g^{ij\ast}_{p}g^{Vij}_{a})+m_{j}(1-x)^{3}(g^{ij\ast}_{s}g^{Vij}_{v}-g^{ij\ast}_{p}g^{Vij}_{a}),$ (26) $\displaystyle c^{Vij}_{O}$ $\displaystyle=-m_{i}x(1-x)(g^{ij\ast}_{s}g^{Vij}_{a}+g^{ij\ast}_{p}g^{Vij}_{v})+m_{j}(1-x)^{2}(g^{ij\ast}_{s}g^{Vij}_{a}-g^{ij\ast}_{p}g^{Vij}_{v}),$ (27) $\displaystyle D^{ij}$ $\displaystyle=x(1-x)k^{2}-xm^{2}_{i}-(1-x)m^{2}_{j}.$ (28) Here, $i,j=2,4,5$, and $g_{s},g_{p},g_{v}^{V},g_{a}^{V}$ are the couplings of scalar, pseudoscalar, vector, and axial-vector types, respectively.444 The details of interaction couplings can be found in appendix A. The particle $j$ is defined as the one who interacts with the external photon, accompanied by particle $i$ in the inner loop of the Barr-Zee diagram. The expression of the electron EDM from the $hV\gamma$ diagrams is given by $\displaystyle d^{hV\gamma}_{e}=\frac{eg^{Vee}_{v}g^{ee}_{s}}{32\pi^{4}}\sum_{i,j=2,4,5}\int^{1}_{0}dx\,\mathrm{Im}(c^{Vij}_{O})\,I^{ij}_{hV},$ (29) where $g^{Vee}_{v}$ and $g^{ee}_{s}$ are the SM couplings of the electron to vector bosons, $V=\gamma$, $Z$, and Higgs boson. The momentum integration over $k$ in the outer loop, $I^{ij}_{hV}$, is given by $\displaystyle I^{ij}_{hV}=\frac{1}{m^{2}_{h}}\left[F\left(x,\frac{m^{2}_{V}}{m^{2}_{h}},\frac{\Delta_{ij}}{m^{2}_{h}}\right)-F\left(x,\frac{m^{2}_{V}}{m^{2}_{h}},\frac{\Delta_{ij}}{m^{2}_{V}}\right)\right],$ (30) where $\displaystyle F(x,y,z)$ $\displaystyle=\frac{1}{(1-y)[z-x(1-x)]}\mathrm{ln}\frac{z}{x(1-x)},$ (31) $\displaystyle\Delta_{ij}$ $\displaystyle=xm^{2}_{i}+(1-x)m^{2}_{j}.$ (32) Another class of diagrams is the ones with two $W$ bosons as shown in the right diagram of Fig. 2. In this case, the contribution to the electron EDM originates from the interaction of the heavy neutrino $\nu_{4}$ with the charged leptons, and is given by $\displaystyle d^{WW\gamma}_{e}=\frac{eg^{2}}{256\pi^{4}}\sum_{i=2,4,5}\frac{m_{e}m_{i}M_{L}}{m^{2}_{W}}\mathrm{Im}\left[(g^{W}_{L})_{4i}(g^{W\ast}_{R})_{4i}\right]\int^{1}_{0}dxI^{4i}_{WW},$ (33) where $\displaystyle I^{4i}_{WW}=\frac{(1-x)}{-x(1-x)m^{2}_{W}+(1-x)m^{2}_{i}+xM^{2}_{L}}\mathrm{ln}\frac{(1-x)m^{2}_{i}+xM^{2}_{L}}{x(1-x)m^{2}_{W}}.$ (34) ## 5 Constraints In this section, we discuss the constraints on the model of vector-like leptons coming from the precision electroweak measurements, the electron EDM, the Higgs decay $h\rightarrow\mu^{+}\mu^{-}$, and the collider constraints on the heavy charged leptons. ### 5.1 Precision electroweak measurements Since the muon and the muon neutrino mix with extra vector-like leptons in this model, the gauge couplings in the mass basis are modified (for details, see appendix A). Therefore, several electroweak observables are affected correspondingly, such as the muon lifetime, decay asymmetries involving muons, partial widths of $W$ and $Z$ bosons, etc. These constraints have been considered in the previous analysis [26, 11] and can be translated into upper bounds of the couplings $\lambda_{L}$ and $\lambda_{E}$ given by $\displaystyle\frac{\lambda_{L}v}{\sqrt{2}M_{L}}\lesssim 0.04~{},~{}\frac{\lambda_{E}v}{\sqrt{2}M_{E}}\lesssim 0.03.$ (35) ### 5.2 $h\rightarrow\mu^{+}\mu^{-}$ Since the Yukawa coupling of the muon is also modified from its SM value, the deviation from the SM prediction is expected in the decay channel of the Higgs boson to a muon pair, $h\rightarrow\mu^{+}\mu^{-}$. In the latest search of this Higgs decay channel, the CMS group found the first evidence of the Higgs- to-dimuon decay channel [50] with a significance of three standard deviations. Currently, the branching fraction of $h\rightarrow\mu^{+}\mu^{-}$ is constrained to be within the range $0.8\times 10^{-4}<\mathcal{B}(h\rightarrow\mu^{+}\mu^{-})<4.5\times 10^{-4}$ at 95$\%$ confidence level. This can be transformed into $\displaystyle 0.37<R(h\rightarrow\mu^{+}\mu^{-})\equiv\frac{\Gamma(h\rightarrow\mu^{+}\mu^{-})}{\Gamma(h\rightarrow\mu^{+}\mu^{-})_{\mathrm{SM}}}<2.1,$ (36) where we have used $\Gamma(h\rightarrow\mu^{+}\mu^{-})_{\mathrm{SM}}=2.16\times 10^{-4}$ [51] for $m_{H}=125.25$ GeV [52]. In our setup, the ratio is given by $R(h\rightarrow\mu^{+}\mu^{-})=|\lambda_{22}|^{2}/(m_{\mu}/v)^{2}$. ### 5.3 Electron EDM The latest measurement of the electron EDM is performed by the ACME collaboration [8], which measured the precession of the electron spin in a superposition of the quantum states of an electron inside a strong intramolecular electric field. The group obtained an upper limit on the value of the electron EDM, $\displaystyle|d_{e}|<1.1\times 10^{-29}~{}e\cdot\mathrm{cm}$ (37) at 90$\%$ confidence level. We note that in obtaining the above limit, possible contributions to the spin precession frequency from the $CP$-odd electron-nucleon scalar coupling are set to zero. In the case of the model of vector-like leptons, such kinds of coupling exist as higher-order quantum effects and, therefore, can be safely ignored in our analysis.555The size of these higher-order effects can be estimated from the results in Ref. [47], where the contribution of the muon EDM to the electron EDM $d_{e}$ and the CP- odd electron-nucleon coupling $C_{S}$ is evaluated. This is expressed in terms of the equivalent electron EDM, $d_{e}^{\rm equiv}$, a linear combination of $d_{e}$ and $C_{S}$ that is constrained by experiments: $d_{e}^{\rm equiv}\simeq 5.8\times 10^{-10}d_{\mu}$. As we see in the subsequent section, in our model we have $|d_{\mu}|\lesssim 1.5\times 10^{-22}~{}e\cdot\mathrm{cm}$, and thus the size of the equivalent electron EDM induced radiatively by the muon EDM is $\lesssim 8.7\times 10^{-32}~{}e\cdot\mathrm{cm}$, which is smaller than the Barr-Zee contribution for most of the parameter points shown in Fig. 5. ### 5.4 Direct search of heavy leptons The extra vector-like leptons decay via charged or neutral weak currents through the mixing with the SM second generation leptons. The tree-level decay modes include $\ell_{i}\to\ell_{j}+Z/h$, and $\ell_{i}\to\nu_{j}+W$, with $i>j$. However, there are not so many experimental searches concentrating on vector-like leptons compared to the searches on vector-like quarks. The LEP experiment searched for such kinds of decays and set a lower bound on the mass of these new leptons [53]. The lower bound was set to be around 100 GeV. Recently, the CMS group reported the lower bound on the mass of vector- like leptons coupled to the third generation of SM lepton, $\tau$ [54]. The doublet type is constrained to be heavier than 1045 GeV, and for the singlet type, the mass range 125–150 GeV is excluded. Expecting that the constraints on the vector-like leptons coupled to the muons are comparable, in this study, we assume that the vector-like leptons have masses of $\mathcal{O}(1)$ TeV. ## 6 Results In this section, we show the results of our analysis on the model of vector- like leptons. We randomly choose the sampling points and scan over the parameters in the model with the ranges of parameters listed in Table 2. We note that the masses of the vector-like leptons are chosen to be at the TeV scale in this analysis so that the constraints from the direct search of heavy leptons are avoided. The muon Yukawa coupling $y_{\mu}$ is solved for the correct muon mass with all other 8 parameters fixed randomly in the range indicated by Table 2. We note that in solving the Yukawa coupling, it can be written in the form of a quadratic equation, therefore there are in general two solutions of Yukawa coupling for each randomly chosen parameter set. We include both of them in the result. Parameter | Value ---|--- $M_{L},M_{E}$ | $1-5$ TeV $|\lambda_{L}|,|\lambda_{E}|$ | $\leq$ EW constraints in Eq. (35) $|\lambda|,|\bar{\lambda}|$ | $0-1$ $\phi_{\lambda},\phi_{\bar{\lambda}}$ | $0-2\pi$ $y_{\mu}$ | solved for the correct $m_{\mu}$ Table 2: Ranges of parameters randomly chosen for the sampling points. Around $4\times 10^{5}$ sampling points for the results are shown in the scattering plots. In the left plot of Fig. 4, the correlation between the muon $g-2$ and the muon EDM is shown. The horizontal red line and the (light) red bands correspond to the central value and (2$\sigma$) 1$\sigma$ regions of the observed muon $g-2$ in the experiment, respectively. The gray dots are the ones excluded by the $h\rightarrow\mu^{+}\mu^{-}$ constraint. After this selection, we further exclude the sampling points that predict the electron EDM to be too large and exceed the current upper bound given in Eq. (37). These are represented by the blue dots. The black dots are the sampling points that satisfy both the constraints from $h\rightarrow\mu^{+}\mu^{-}$ and electron EDM. We can see that there are predictions consistent with the observed muon $g-2$, and the largest muon EDM is around $|d_{\mu}|\simeq 1.5\times 10^{-22}~{}e\cdot\mathrm{cm}$. The PSI experiment [10] can examine some regions of the parameters, and its sensitivity to the muon EDM is shown by the vertical dotted line. In this result, we confirm the elliptical correlation between muon $g-2$ and muon EDM obtained from the EFT approach [46]. In the right plot of Fig. 4, the correlation between the muon $g-2$ and the electron EDM is shown. The gray dots show the sampling points excluded by the $h\rightarrow\mu^{+}\mu^{-}$ constraint, and the region with the blue band is excluded by the current limit on the electron EDM from the ACME experiment. From the plot, we see that most of the sampling points that are consistent with both the $h\rightarrow\mu^{+}\mu^{-}$ and muon $g-2$ predict $|d_{e}|\gtrsim 10^{-32}~{}e\cdot\mathrm{cm}$. The vertical dashed line, $|d_{e}|=10^{-30}~{}e\cdot\mathrm{cm}$, indicates a representative sensitivity of near future experiments. The next-generation ACME experiment (ACME III) is expected to improve the current sensitivity by an order of magnitude [55], and there are also proposals to improve the sensitivity by several orders of magnitude in the next 10-20 years [56]. Figure 4: The correlations of muon $g-2$ with muon EDM (left plot) and electron EDM (right plot). The horizontal red line and (light) red bands correspond to the central value and (2$\sigma$) 1$\sigma$ regions of the observed muon $g-2$, respectively. The $h\rightarrow\mu^{+}\mu^{-}$ constraint is first applied in both plots and the sampling points which do not satisfy it are shown as the gray dots. In the left plot, we also show the sampling points further excluded by the electron EDM constraint in blue. The black dots are the points that satisfy both the constraints from $h\rightarrow\mu^{+}\mu^{-}$ and electron EDM. The vertical dotted line represents the PSI experiment. In the right plot, the blue band shows the constraint from electron EDM measurement and the vertical dashed line indicates the prospect of the near- future electron EDM experiments. In Fig. 5, we show the correlation between the muon and the electron EDMs. Again, we first excluded the sampling points that do not satisfy the $h\rightarrow\mu^{+}\mu^{-}$ constraint. These are the gray dots in the figure. The remaining sampling points are categorized into two classes, with the (dark) red dots representing the points satisfying the (1$\sigma$) 2$\sigma$ region of the muon $g-2$, respectively. The horizontal blue band is the constraint from the ACME experiment for the electron EDM measurement, and the prospects of the future muon/electron EDM experiments are also given in the plot. We find that these proposed experiments can investigate a significant fraction of the parameter space of this model, as they can exclude most of the available sampling points in the figure if no signals of EDMs are discovered. Furthermore, all of the allowed parameter space in Fig. 5 can be covered by the electron EDM experiments offering sensitivities better than the current value by several orders of magnitude [56]. Figure 5: The correlations between the muon and electron EDMs. The gray dots show the points excluded by the $h\rightarrow\mu^{+}\mu^{-}$ constraint. The (dark) red dots represent the points satisfying the (1$\sigma$) 2$\sigma$ region of the muon $g-2$. The blue band is the constraint from the ACME experiment for the electron EDM measurement. The prospect of the PSI experiment for the muon EDM measurement and the near-future electron EDM experiments are shown as the vertical dotted line and horizontal dashed line, respectively. Finally, we show the dependence of the muon and electron EDMs on the two $CP$-violating phases, $\phi_{\bar{\lambda}}$ and $\phi_{\lambda}$, in Fig. 6. Here, we fix the other parameters to be $|\lambda|=0.2,|\bar{\lambda}|=0.6,M_{E}=2.5~{}\mathrm{TeV},M_{L}=5~{}\mathrm{TeV}$, $\lambda_{L}=-0.04\times\sqrt{2}M_{L}/v\sim-0.57$, and $\lambda_{E}=0.03\times\sqrt{2}M_{E}/v\sim 0.86$. The magnitude of the EDMs is indicated by the black contours. The gray bands are the regions where no solution exists for the $y_{\mu}$ to give the correct muon mass after the mass diagonalization. The (2$\sigma$) 1$\sigma$ level of muon $g-2$ are also given in the left figure as the (light) red bands. As we can see from the plots, the muon EDM only depends on one of the two $CP$-violating phases, $\phi_{\bar{\lambda}}$, while in the case of the electron EDM, it depends on both of the phases. If the muon EDM is indeed nonzero and the vector-like leptons exist, we may be able to determine the masses and couplings of the vector-like leptons from the measurement of their decays, and the value of $\phi_{\bar{\lambda}}$ can then be fixed by the result of the muon EDM measurement. We can input these parameters to the electron EDM and find the value of the remaining phase $\phi_{\lambda}$ according to the result of the electron EDM measurement. Figure 6: The dependence of muon EDM (left) and electron EDM (right) on the two $CP$-violating phases in the model. The other parameters are chosen to be $|\lambda|=0.2,|\bar{\lambda}|=0.6,M_{E}=2.5~{}\mathrm{TeV},M_{L}=5~{}\mathrm{TeV}$, $\lambda_{L}=-0.04\times\sqrt{2}M_{L}/v\sim-0.57$, and $\lambda_{E}=0.03\times\sqrt{2}M_{E}/v\sim 0.86$. The size of the EDMs are given by the black contours. The gray bands represent the regions where no solution exists for the $y_{\mu}$ to give the correct muon mass after the mass diagonalization. The region consistent with muon $g-2$ at (2$\sigma$) 1$\sigma$ level is also indicated in the left figure by the (light) red bands. ## 7 Summary In this work, we have investigated a simple extension of the Standard Model, with the addition of one SU(2)L doublet and one SU(2)L singlet vector-like leptons, which are coupled to the second-generation SM leptons only. In this framework, the muon dipole moments are generated through the mediation of the Higgs, $W$, and $Z$ bosons. Furthermore, one interesting feature of this model is that a sizable electron EDM can also be induced at the two-loop level. Besides the latest value of the muon $g-2$ published by the Muon $g-2$ collaboration at the Fermilab, we have also considered the recent constraint on the Higgs decay channel $h\rightarrow\mu^{+}\mu^{-}$ reported by the CMS group and the constraint on the electron EDM from the ACME experiment. We have found that there are parameter regions consistent with all these constraints. Because of the chirality flip from the heavy vector-like leptons, the sizes of muon dipole moments are enhanced. The muon EDM can be as large as $10^{-22}~{}e\cdot\mathrm{cm}$, which can be probed in the future muon EDM measurement such as the PSI experiment, whose sensitivity is estimated to be around $6\times 10^{-23}~{}e\cdot\mathrm{cm}$ [10]. An electron EDM of $\mathcal{O}(10^{-30})~{}e\cdot\mathrm{cm}$ can be generated in this model, and future electron EDM measurement like the ACME III experiment, whose sensitivity is expected to be one order of magnitude better than the previous ACME II [55], is able to examine the prediction. ## Acknowledgments This work is supported in part by the Grant-in-Aid for Innovative Areas (No.19H05810 [KH], No.19H05802 [KH], No.18H05542 [NN]), Scientific Research B (No.20H01897 [KH and NN]), Young Scientists (No.21K13916 [NN]), and JSPS KAKENHI Grant (No.20J22214 [SYT]). ## Appendix A Interactions In this appendix, we briefly summarize the interactions in the model we used in the analysis. Concerning the index notation, the quantities related to the flavor basis are indexed by Greek alphabets ($\alpha,\beta,\gamma,\cdots$), while the quantities related to the mass basis are indexed by Latin alphabets ($i,j,k,\cdots$). For the flavor indices, we have $2=\mu,4=L$ and $5=E$. ### A.1 Higgs boson couplings Since the couplings related to electron and tau are not modified by the vector-like leptons, their couplings with the Higgs boson are the SM values $\lambda_{e,\tau}=-m_{e,\tau}/v$. Other couplings of charged leptons to the Higgs boson are modified by the mixings. The Yukawa interaction in the flavor basis is given by $\displaystyle{\cal L}_{Y}\supset-\frac{1}{\sqrt{2}}\,\bar{f}_{L\rho}\,Y_{\rho\sigma}\,f_{R\sigma}\,h+\mathrm{h.c.}.$ (38) This can be transformed to the the mass basis given by $\displaystyle{\cal L}_{Y}\supset-\frac{1}{\sqrt{2}}\sum_{\rho,\sigma=2,4,5}\bar{f}_{Li}(U^{{\dagger}}_{L})_{i\rho}\,Y_{\rho\sigma}\,(U_{R})_{\sigma j}\,f_{Rj}\,h+\mathrm{h.c.}\equiv\bar{f}_{Li}\,\lambda_{ij}\,f_{Rj}\,h+\mathrm{h.c.},$ (39) where the Yukawa matrix is written as $\displaystyle Y=\begin{pmatrix}y_{\mu}&0&\lambda_{E}\\\ \lambda_{L}&0&\lambda\\\ 0&\bar{\lambda}&0\end{pmatrix},$ (40) and the Yukawa couplings in the mass basis is $\displaystyle\lambda_{ij}=-\frac{1}{\sqrt{2}}\sum_{\rho,\sigma=2,4,5}(U^{\dagger}_{L})_{i\rho}Y_{\rho\sigma}\left(U_{R}\right)_{\sigma j}.$ (41) We notice that the Yukawa coupling in the flavor basis is similar to the mass matrix of the charged leptons with the absence of the masses of vector-like leptons, that is, $Yv/\sqrt{2}=M-\mathrm{diag}(0,M_{L},M_{E})$. With this relation, the Higgs boson couplings in the mass basis can be written as: $\displaystyle-\lambda v$ $\displaystyle=U^{\dagger}_{L}\begin{pmatrix}y_{\mu}v/\sqrt{2}&0&\lambda_{E}v/\sqrt{2}\\\ \lambda_{L}v/\sqrt{2}&M_{L}&\lambda v/\sqrt{2}\\\ 0&\bar{\lambda}v/\sqrt{2}&M_{E}\end{pmatrix}U_{R}-U^{\dagger}_{L}\begin{pmatrix}0&0&0\\\ 0&M_{L}&0\\\ 0&0&M_{E}\end{pmatrix}U_{R}$ (42) $\displaystyle=\begin{pmatrix}m_{\mu}&0&0\\\ 0&m_{4}&0\\\ 0&0&m_{5}\end{pmatrix}-U^{\dagger}_{L}\begin{pmatrix}0&0&0\\\ 0&M_{L}&0\\\ 0&0&M_{E}\end{pmatrix}U_{R}.$ (43) In this expression, we can see clearly that the first term has a simple form which is the same as the Yukawa couplings in the SM. The second term corresponds solely to the contributions from the mass term of the vector-like leptons. ### A.2 $Z$ boson couplings The couplings of leptons to the $Z$ boson come from the kinetic term of the leptons. Due to the mixing among muon and vector-like leptons, the kinetic term is modified to $\displaystyle{\cal L}_{\mathrm{kin}}\supset\bar{f}_{L\sigma}i\not{D}_{\sigma}f_{L\sigma}+\bar{f}_{R\sigma}i\not{D}_{\sigma}f_{R\sigma}=\bar{f}_{Li}(U^{{\dagger}}_{L})_{i\sigma}i\not{D}_{\sigma}(U_{L})_{\sigma j}f_{Lj}+\bar{f}_{Ri}(U^{{\dagger}}_{R})_{i\sigma}i\not{D}_{\sigma}(U_{R})_{\sigma j}f_{Rj}.$ (44) The covariant derivative $D_{\mu\sigma}$ is defined as $\displaystyle D_{\mu\sigma}=\partial_{\mu}+i\,\frac{g}{\mathrm{cos}\,\theta_{W}}(T^{3}_{\sigma}-Q_{\sigma}\mathrm{sin}^{2}\theta_{W})Z_{\mu}+ieQ_{\sigma}A_{\mu},$ (45) where $e$ and $g$ are the electromagnetic and the SU(2) couplings of the Standard Model, and $T^{3}$ and $Q$ are the weak isospin and electric charge of leptons obtained from the quantum numbers listed in Table 1. The couplings of the $Z$ boson to charged leptons $\ell_{i}$ and $\ell_{j}$ are defined in the Lagrangian $\displaystyle{\cal L}_{Z}\supset\left[\bar{f}_{Li}\gamma^{\mu}(g^{Z}_{L})_{ij}f_{Lj}+\bar{f}_{Ri}\gamma^{\mu}(g^{Z}_{R})_{ij}f_{Rj}\right]Z_{\mu}.$ (46) The couplings of left-handed and right-handed charged leptons with $Z$ boson are given by $\displaystyle(g^{Z}_{L})_{ij}=-\frac{g}{\cos\theta_{W}}\displaystyle\sum\limits_{\sigma=2,4,5}(T^{3}_{L\sigma}-\sin^{2}\theta_{W}Q_{\sigma})(U^{\dagger}_{L})_{i\sigma}(U_{L})_{\sigma j},$ (47) $\displaystyle(g^{Z}_{R})_{ij}=-\frac{g}{\cos\theta_{W}}\displaystyle\sum\limits_{\sigma=2,4,5}(T^{3}_{R\sigma}-\sin^{2}\theta_{W}Q_{\sigma})(U^{\dagger}_{R})_{i\sigma}(U_{R})_{\sigma j},$ (48) where the electric charge $Q_{\sigma}$ and the third component of weak isospins $T^{3}_{L\sigma}$ and $T^{3}_{R\sigma}$ for each flavor are summarized in Table 3. Flavor | 2 | 4 | 5 ---|---|---|--- $Q_{\sigma}$ | $-$1 | $-$1 | $-$1 $T^{3}_{L\sigma}$ | $-$1/2 | $-$1/2 | 0 $T^{3}_{R\sigma}$ | 0 | $-$1/2 | 0 Table 3: This table shows the electric charge $Q_{\sigma}$ and the third component of weak isospins $T^{3}_{L\sigma}$ and $T^{3}_{R\sigma}$ for each flavor considered in the mixing of charged leptons. Recall that in the flavor basis, we have $2=\mu,4=L$, and $5=E$. Since the electric charge $Q_{\sigma}$ is the same for all the charged leptons considered, no modification is made on the couplings of photon with charged leptons. On the other hand, because of the different weak isospins of the charged leptons in the mixing, the couplings of the $Z$ boson in Eq. (47) and Eq. (48) are modified from their SM values. In $(g^{Z}_{L})_{ij}$, we have $\displaystyle-\frac{g}{\mathrm{cos}\,\theta_{W}}\left\\{\left(-\frac{1}{2}+\mathrm{sin}^{2}\theta_{W}\right)[(U^{\dagger}_{L})_{i2}(U_{L})_{2j}+(U^{\dagger}_{L})_{i4}(U_{L})_{4j}]+\mathrm{sin}^{2}\theta_{W}(U^{\dagger}_{L})_{i5}(U_{L})_{5j}\right\\}.$ (49) To see the modification to the SM coupling, we can arrange and separate Eq. (47) in two parts. The first one is $\displaystyle-\frac{g}{\mathrm{cos}\,\theta_{W}}\left\\{\left(-\frac{1}{2}+\mathrm{sin}^{2}\theta_{W}\right)[(U^{\dagger}_{L})_{i2}(U_{L})_{2j}+(U^{\dagger}_{L})_{i4}(U_{L})_{4j}+(U^{\dagger}_{L})_{i5}(U_{L})_{5j}]\right\\}.$ (50) By the unitarity of the matrix $U_{L}$, we have $\sum_{\sigma=2,4,5}(U^{\dagger}_{L})_{i\sigma}(U_{L})_{\sigma j}=\delta_{ij}$. This gives the form of the SM coupling of $Z$ boson with the left-handed leptons $\displaystyle(g^{Z}_{L})^{\mathrm{SM}}_{ij}=-\frac{g}{\mathrm{cos}\,\theta_{W}}\left(-\frac{1}{2}+\mathrm{sin}^{2}\theta_{W}\right)\delta_{ij}.$ (51) The second one corresponds to the modification of the SM coupling, $\displaystyle(\delta g^{Z}_{L})_{ij}=-\frac{g}{2\,\mathrm{cos}\,\theta_{W}}(U^{\dagger}_{L})_{i5}(U_{L})_{5j}.$ (52) Similarly, we can separate Eq. (48) in parts of the SM coupling $\displaystyle(g^{Z}_{R})^{\mathrm{SM}}_{ij}=-\frac{g}{\mathrm{cos}\,\theta_{W}}\sin^{2}\theta_{W}\,\delta_{ij}$ (53) and the modification $\displaystyle(\delta g^{Z}_{R})_{ij}=\frac{g}{2\mathrm{cos}\,\theta_{W}}(U^{\dagger}_{R})_{i4}(U_{R})_{4j}.$ (54) ### A.3 $W$ boson couplings The couplings of $W$ with leptons are also derived from the kinetic term. Since the charged lepton $E$ is an SU(2)L singlet, it decouples from the interaction with $W$ boson. The kinetic term is given by $\displaystyle{\cal L}_{\mathrm{kin}}$ $\displaystyle\supset-\frac{g}{\sqrt{2}}\left(\bar{\nu}_{\mu}\gamma^{\mu}\mu_{L}+\bar{L}_{L}^{0}\gamma^{\mu}L_{L}^{-}+\bar{L}_{R}^{0}\gamma^{\mu}L_{R}^{-}\right)W^{+}_{\mu}+h.c.$ (55) $\displaystyle=-\frac{g}{\sqrt{2}}\left(\bar{\nu}_{2}\gamma^{\mu}(U_{L})_{2j}f_{Lj}+\bar{\nu}_{L4}\gamma^{\mu}(U_{L})_{4j}f_{Lj}+\bar{\nu}_{R4}\gamma^{\mu}(U_{R})_{4j}f_{Rj}\right)W^{+}_{\mu}+h.c..$ (56) This can be written in a more compact form $\displaystyle{\cal L}_{W}\supset\left[\bar{\nu}_{Li}\gamma^{\mu}(g^{W}_{L})_{ij}f_{Lj}+\bar{\nu}_{Ri}\gamma^{\mu}(g^{W}_{R})_{ij}f_{Rj}\right]W^{+}_{\mu}+h.c.,$ (57) and the couplings of $W$ boson with leptons are defined as $\displaystyle(g^{W}_{L})_{2j}=-\frac{g}{\sqrt{2}}(U_{L})_{2j},\qquad(g^{W}_{L})_{4j}=-\frac{g}{\sqrt{2}}(U_{L})_{4j},\qquad(g^{W}_{R})_{4j}=-\frac{g}{\sqrt{2}}(U_{R})_{4j}~{},$ (58) with $(g^{W}_{R})_{2j}=(g^{W}_{L})_{5j}=(g^{W}_{R})_{5j}=0$. ### A.4 Couplings in the electron EDM formulas The interaction of the Higgs boson with the charged leptons is $\displaystyle\bar{f}_{Li}\,\lambda_{ij}\,f_{Rj}\,h+\bar{f}_{Ri}\,\lambda^{\ast}_{ji}\,f_{Lj}\,h=\bar{f}_{i}(\lambda_{ij}P_{R}+\lambda^{\ast}_{ji}P_{L})f_{j}\,h,$ (59) where $P_{L,R}$ are the projection operators for the left-handed and right- handed components of the leptons. This can be rearranged in the form of the scalar and pseudoscalar couplings $\displaystyle\bar{f}_{i}(g_{s}^{ij}+g_{p}^{ij}\gamma_{5})f_{j}\,h,$ (60) where $\displaystyle g_{s}^{ij}=\frac{1}{2}(\lambda_{ij}+\lambda^{\ast}_{ji})~{},~{}g_{p}^{ij}=\frac{1}{2}(\lambda_{ij}-\lambda^{\ast}_{ji})$ (61) are the scalar and pseudoscalar types of coupling. The interaction of charged leptons with $Z$ boson can be written as $\displaystyle\bar{f}_{i}\gamma^{\mu}\left(g^{Z,ij}_{v}+g^{Z,ij}_{a}\gamma_{5}\right)f_{j}Z_{\mu},$ (62) where $\displaystyle g^{Z,ij}_{v}=\frac{1}{2}\left[(g^{Z}_{L})_{ij}+(g^{Z}_{R})_{ij}\right]~{},~{}g^{Z,ij}_{a}=\frac{1}{2}\left[(-g^{Z}_{L})_{ij}+(g^{Z}_{R})_{ij}\right]$ (63) are the vector and axial-vector types of coupling. The electron couplings in Eq. (29) are given by $\displaystyle g^{ee}_{s}$ $\displaystyle=-m_{e}/v,$ (64) $\displaystyle g^{\gamma ee}_{v}$ $\displaystyle=e,$ (65) $\displaystyle g^{Zee}_{v}$ $\displaystyle=-\frac{g}{4\,\mathrm{cos}\,\theta_{W}}\left(-1+4\,\mathrm{sin}^{2}\theta_{W}\right),g^{Zee}_{a}=-\frac{g}{4\,\mathrm{cos}\,\theta_{W}}.$ (66) ## Appendix B An example for the muon-only coupling In the model discussed in our work, we just assume that the vector-like leptons couple only to the second-generation leptons, not to the first- and third-generation ones. In this appendix, we briefly discuss a model where this setup is realized as a consequence of U(1) gauge symmetries. As discussed in Refs. [57, 58], there are two types of lepton flavor-dependent U(1) symmetries which can be gauged and introduced to a model without anomalies. One is the linear combination of the lepton numbers U(1)${}_{\alpha_{e}L_{e}+\alpha_{\mu}L_{\mu}-(\alpha_{e}+\alpha_{\mu})L_{\tau}}$ and the other is a linear combination of baryon and lepton numbers U(1)${}_{B+\beta_{e}L_{e}+\beta_{\mu}L_{\mu}-(3+\beta_{e}+\beta_{\mu})L_{\tau}}$. As an example, we choose them to be U(1)${}_{L_{\mu}-L_{\tau}}$ and U(1)${}_{B+3L_{e}-L_{\mu}-5L_{\tau}}$. The charge assignment of the fields is summarized in Table 4. Note that we introduce three right-handed neutrinos, $N_{e},N_{\mu},N_{\tau}$. To break the new U(1) gauge symmetries, we also introduce two scalar fields that are singlet under the SM gauge interactions, $\sigma$ and $\sigma^{\prime}$. We choose the charges of the vector-like leptons to be the same as the ones of muon so that the interaction structure of coupling solely to the muon in Eq. (8) can be realized. The charges of the new scalars are chosen for the reason explained in the following. | $e$ | $\mu$ | $\tau$ | $H$ | $L_{L,R}$ | $E_{L,R}$ | $\sigma$ | $\sigma^{\prime}$ | $N_{e}$ | $N_{\mu}$ | $N_{\tau}$ ---|---|---|---|---|---|---|---|---|---|---|--- U(1)${}_{L_{\mu}-L_{\tau}}$ | 0 | 1 | $-$1 | 0 | 1 | 1 | 1 | 0 | 0 | 1 | $-$1 U(1)${}_{B+3L_{e}-L_{\mu}-5L_{\tau}}$ | 3 | $-$1 | $-$5 | 0 | $-$1 | $-$1 | 2 | 6 | 3 | $-$1 | $-$5 Table 4: Charge assignment of gauged U(1)${}_{L_{\mu}-L_{\tau}}$ and U(1)${}_{B+3L_{e}-L_{\mu}-5L_{\tau}}$ symmetries to the fields considered in the model in appendix B. After the spontaneous symmetry breaking of scalar fields, we obtain a Majorana mass matrix of the right-handed neutrinos in the form of $\displaystyle\begin{pmatrix}\lambda_{ee}\langle\sigma^{\prime\ast}\rangle&\lambda_{e\mu}\langle\sigma^{\ast}\rangle&\lambda_{e\tau}\langle\sigma\rangle\\\ \lambda_{e\mu}\langle\sigma^{\ast}\rangle&0&\lambda_{\mu\tau}\langle\sigma^{\prime}\rangle\\\ \lambda_{e\tau}\langle\sigma\rangle&\lambda_{\mu\tau}\langle\sigma^{\prime}\rangle&0\end{pmatrix},$ (67) while the neutrino Dirac mass matrix is diagonal. This implies that after applying the seesaw mechanism [59, 60, 61, 62], we obtain an active neutrino mass matrix with the so-called two-zero minor structure [63, 64]. We can then follow the same analysis performed in [65, 66, 67] to extract the predictions of the sum of three active neutrino masses and the Dirac $CP$-violating phase $\delta$. In Fig. 7, we show these predictions as a function of the neutrino mixing angle $\theta_{23}$, with other two mixing angles and the two squared mass differences fixed by the values provided in [68]. The neutrino masses are taken to be in the normal ordering, $m_{1}<m_{2}<m_{3}$, as the inverse ordering turns out to be incompatible with the observed neutrino oscillation data in this model [65]. In both plots, the boundary of $x$ axis corresponds to the 3$\sigma$ range of $\theta_{23}$, while the vertical dashed line shows the central value of $\theta_{23}$ and the 1$\sigma$ range of $\theta_{23}$ are enclosed by the vertical dotted lines. The horizon dashed line in the left plot represents the upper bound on the sum of active neutrino masses given in [69], which is $\sum m_{\nu}<0.13$ eV at 95$\%$ confidence level. The (light) green band in the right plot shows the (3) 1$\sigma$ range of the Dirac $CP$-violating phase obtained in [68]. The predictions from the two-zero minor structure are plotted as red curves, dark (light) red bands corresponding to the parameters fixed to their central value and 1 (3) sigma value, respectively, given in [68]. As we can see from the plots, although a large fraction of the parameter region are excluded by the constraint on the sum of the neutrino masses, there is still a possibility around the corner close to the 3$\sigma$ boundary of $\theta_{23}$, which can be examined in the future neutrino experiments. Figure 7: Left plot: predictions of the sum of the three active neutrino masses as a function of $\theta_{23}$. Right plot: predictions of the Dirac $CP$ phase as a function of $\theta_{23}$. The predictions are plotted as red curves, dark (light) red bands corresponding to the parameters fixed to their central value and 1 (3) sigma value given in [68]. In both plots, the boundary of $x$ axis corresponds to the 3$\sigma$ range of $\theta_{23}$, while the vertical dashed line shows the central value of $\theta_{23}$ and the 1$\sigma$ range of $\theta_{23}$ are enclosed by the vertical dotted lines. The horizon dashed line in the left plot represents the upper bound on the sum of the active neutrino masses given in [69], which is $\sum m_{\nu}<0.13$ eV at 95$\%$ confidence level. The (light) green band in the right plot shows the (3) 1$\sigma$ range of the Dirac $CP$-violating phase obtained in [68]. ## References * [1] Muon g-2 Collaboration, Measurement of the Positive Muon Anomalous Magnetic Moment to 0.46 ppm, Phys. Rev. Lett. 126 (2021) 141801 [arXiv:2104.03281]. * [2] T. Aoyama et al., The anomalous magnetic moment of the muon in the Standard Model, Phys. Rept. 887 (2020) 1–166 [arXiv:2006.04822]. * [3] S. Borsanyi et al., Leading hadronic contribution to the muon magnetic moment from lattice QCD, Nature 593 (2021) 51–55 [arXiv:2002.12347]. * [4] M. Cè et al., Window observable for the hadronic vacuum polarization contribution to the muon $g-2$ from lattice QCD, arXiv:2206.06582 (2022). * [5] C. Alexandrou et al., Lattice calculation of the short and intermediate time-distance hadronic vacuum polarization contributions to the muon magnetic moment using twisted-mass fermions, arXiv:2206.15084 (2022). * [6] M. Pospelov and A. Ritz, Electric dipole moments as probes of new physics, Annals Phys. 318 (2005) 119–169 [hep-ph/0504231]. * [7] Muon (g-2) Collaboration, An Improved Limit on the Muon Electric Dipole Moment, Phys. Rev. D 80 (2009) 052008 [arXiv:0811.1207]. * [8] ACME Collaboration, Improved limit on the electric dipole moment of the electron, Nature 562 (2018) 355–360. * [9] M. Abe et al., A New Approach for Measuring the Muon Anomalous Magnetic Moment and Electric Dipole Moment, PTEP 2019 (2019) 053C02 [arXiv:1901.03047]. * [10] A. Adelmann et al., Search for a muon EDM using the frozen-spin technique, arXiv:2102.08838 (2021). * [11] K. Kannike, M. Raidal, D. M. Straub, and A. Strumia, Anthropic solution to the magnetic muon anomaly: the charged see-saw, JHEP 02 (2012) 106 [arXiv:1111.2551]. [Erratum: JHEP 10, 136 (2012)]. * [12] A. Falkowski, D. M. Straub, and A. Vicente, Vector-like leptons: Higgs decays and collider phenomenology, JHEP 05 (2014) 092 [arXiv:1312.5329]. * [13] R. Dermisek and A. Raval, Explanation of the Muon g-2 Anomaly with Vectorlike Leptons and its Implications for Higgs Decays, Phys. Rev. D 88 (2013) 013017 [arXiv:1305.3522]. * [14] P. Arnan, L. Hofer, F. Mescia, and A. Crivellin, Loop effects of heavy new scalars and fermions in $b\to s\mu^{+}\mu^{-}$, JHEP 04 (2017) 043 [arXiv:1608.07832]. * [15] E. Megias, M. Quiros, and L. Salas, $g_{\mu}-2$ from Vector-Like Leptons in Warped Space, JHEP 05 (2017) 016 [arXiv:1701.05072]. * [16] K. Kowalska and E. M. Sessolo, Expectations for the muon g-2 in simplified models with dark matter, JHEP 09 (2017) 112 [arXiv:1707.00753]. * [17] S. Raby and A. Trautner, Vectorlike chiral fourth family to explain muon anomalies, Phys. Rev. D 97 (2018) 095006 [arXiv:1712.09360]. * [18] Z. Poh and S. Raby, Vectorlike leptons: Muon g-2 anomaly, lepton flavor violation, Higgs boson decays, and lepton nonuniversality, Phys. Rev. D 96 (2017) 015032 [arXiv:1705.07007]. * [19] J. Kawamura, S. Raby, and A. Trautner, Complete vectorlike fourth family and new U(1)’ for muon anomalies, Phys. Rev. D 100 (2019) 055030 [arXiv:1906.11297]. * [20] G. Hiller, C. Hormigos-Feliu, D. F. Litim, and T. Steudtner, Anomalous magnetic moments from asymptotic safety, Phys. Rev. D 102 (2020) 071901 [arXiv:1910.14062]. * [21] G. Hiller, C. Hormigos-Feliu, D. F. Litim, and T. Steudtner, Model Building from Asymptotic Safety with Higgs and Flavor Portals, Phys. Rev. D 102 (2020) 095023 [arXiv:2008.08606]. * [22] M. Endo and S. Mishima, Muon $g-2$ and CKM unitarity in extra lepton models, JHEP 08 (2020) 004 [arXiv:2005.03933]. * [23] M. Frank and I. Saha, Muon anomalous magnetic moment in two-Higgs-doublet models with vectorlike leptons, Phys. Rev. D 102 (2020) 115034 [arXiv:2008.11909]. * [24] E. J. Chun and T. Mondal, Explaining $g-2$ anomalies in two Higgs doublet model with vector-like leptons, JHEP 11 (2020) 077 [arXiv:2009.08314]. * [25] K. Kowalska and E. M. Sessolo, Minimal models for g-2 and dark matter confront asymptotic safety, Phys. Rev. D 103 (2021) 115032 [arXiv:2012.15200]. * [26] R. Dermisek, K. Hermanek, and N. McGinnis, Muon g-2 in two-Higgs-doublet models with vectorlike leptons, Phys. Rev. D 104 (2021) 055033 [arXiv:2103.05645]. * [27] H. M. Lee and K. Yamashita, A model of vector-like leptons for the muon $g-2$ and the W boson mass, Eur. Phys. J. C 82 (2022) 661 [arXiv:2204.05024]. * [28] A. de Giorgi, L. Merlo, and S. Pokorski, The Low-Scale Seesaw Solution to the $M_{W}$ and $(g-2)_{\mu}$ Anomalies, arXiv:2211.03797 (2022). * [29] K. S. Babu, B. Dutta, and R. N. Mohapatra, Enhanced electric dipole moment of the muon in the presence of large neutrino mixing, Phys. Rev. Lett. 85 (2000) 5064–5067 [hep-ph/0006329]. * [30] T. Ibrahim and P. Nath, Slepton flavor nonuniversality, the muon EDM and its proposed sensitive search at Brookhaven, Phys. Rev. D 64 (2001) 093002 [hep-ph/0105025]. * [31] J. L. Feng, K. T. Matchev, and Y. Shadmi, Theoretical expectations for the muon’s electric dipole moment, Nucl. Phys. B 613 (2001) 366–381 [hep-ph/0107182]. * [32] A. Romanino and A. Strumia, Electron and Muon Electric Dipoles in Supersymmetric Scenarios, Nucl. Phys. B 622 (2002) 73–94 [hep-ph/0108275]. * [33] J. R. Ellis, J. Hisano, S. Lola, and M. Raidal, CP violation in the minimal supersymmetric seesaw model, Nucl. Phys. B 621 (2002) 208–234 [hep-ph/0109125]. * [34] J. R. Ellis, J. Hisano, M. Raidal, and Y. Shimizu, Lepton electric dipole moments in nondegenerate supersymmetric seesaw models, Phys. Lett. B 528 (2002) 86–96 [hep-ph/0111324]. * [35] A. Bartl, W. Majerotto, W. Porod, and D. Wyler, Effect of supersymmetric phases on lepton dipole moments and rare lepton decays, Phys. Rev. D 68 (2003) 053005 [hep-ph/0306050]. * [36] K. Cheung, O. C. W. Kong, and J. S. Lee, Electric and anomalous magnetic dipole moments of the muon in the MSSM, JHEP 06 (2009) 020 [arXiv:0904.4352]. * [37] G. Hiller, K. Huitu, T. Ruppell, and J. Laamanen, A Large Muon Electric Dipole Moment from Flavor?, Phys. Rev. D 82 (2010) 093015 [arXiv:1008.5091]. * [38] C. Cesarotti, Q. Lu, Y. Nakai, A. Parikh, and M. Reece, Interpreting the Electron EDM Constraint, JHEP 05 (2019) 059 [arXiv:1810.07736]. * [39] W. Dekens, J. de Vries, M. Jung, and K. K. Vos, The phenomenology of electric dipole moments in models of scalar leptoquarks, JHEP 01 (2019) 069 [arXiv:1809.09114]. * [40] A. Crivellin, M. Hoferichter, and P. Schmidt-Wellenburg, Combined explanations of $(g-2)_{\mu,e}$ and implications for a large muon EDM, Phys. Rev. D 98 (2018) 113002 [arXiv:1807.11484]. * [41] W. Altmannshofer, S. Gori, H. H. Patel, S. Profumo, and D. Tuckler, Electric dipole moments in a leptoquark scenario for the $B$-physics anomalies, JHEP 05 (2020) 069 [arXiv:2002.01400]. * [42] I. Bigaran and R. R. Volkas, Reflecting on chirality: CP-violating extensions of the single scalar-leptoquark solutions for the (g-2)e,$\mu$ puzzles and their implications for lepton EDMs, Phys. Rev. D 105 (2022) 015002 [arXiv:2110.03707]. * [43] Y. Omura, E. Senaha, and K. Tobe, $\tau$\- and $\mu$-physics in a general two Higgs doublet model with $\mu-\tau$ flavor violation, Phys. Rev. D 94 (2016) 055019 [arXiv:1511.08880]. * [44] W.-S. Hou, G. Kumar, and S. Teunissen, Charged lepton EDM with extra Yukawa couplings, JHEP 01 (2022) 092 [arXiv:2109.08936]. * [45] Y. Nakai, R. Sato, and Y. Shigekami, Muon electric dipole moment as a probe of flavor-diagonal CP violation, Phys. Lett. B 831 (2022) 137194 [arXiv:2204.03183]. * [46] R. Dermisek, K. Hermanek, N. McGinnis, and S. Yoon, Ellipse of Muon Dipole Moments, Phys. Rev. Lett. 129 (2022) 221801 [arXiv:2205.14243]. * [47] Y. Ema, T. Gao, and M. Pospelov, Improved Indirect Limits on Muon Electric Dipole Moment, Phys. Rev. Lett. 128 (2022) 131803 [arXiv:2108.05398]. * [48] S. M. Barr and A. Zee, Electric Dipole Moment of the Electron and of the Neutron, Phys. Rev. Lett. 65 (1990) 21–24. [Erratum: Phys.Rev.Lett. 65, 2920 (1990)]. * [49] Y. Nakai and M. Reece, Electric Dipole Moments in Natural Supersymmetry, JHEP 08 (2017) 031 [arXiv:1612.08090]. * [50] CMS Collaboration, Evidence for Higgs boson decay to a pair of muons, JHEP 01 (2021) 148 [arXiv:2009.04363]. * [51] LHC Higgs Cross Section Working Group Collaboration, Handbook of LHC Higgs Cross Sections: 4. Deciphering the Nature of the Higgs Sector, arXiv:1610.07922 (2016). * [52] Particle Data Group Collaboration, Review of Particle Physics,. to be published (2022). * [53] L3 Collaboration, Search for heavy neutral and charged leptons in $e^{+}e^{-}$ annihilation at LEP, Phys. Lett. B 517 (2001) 75–85 [hep-ex/0107015]. * [54] CMS Collaboration, Inclusive nonresonant multilepton probes of new phenomena at $\sqrt{s}$=13 TeV, Phys. Rev. D 105 (2022) 112007 [arXiv:2202.08676]. * [55] The ACME EDM Experiment, http://doylegroup.harvard.edu/edm/. * [56] R. Alarcon et al. in 2022 Snowmass Summer Study. 2022\. arXiv:2203.08103. * [57] T. Araki, J. Heeck, and J. Kubo, Vanishing Minors in the Neutrino Mass Matrix from Abelian Gauge Symmetries, JHEP 07 (2012) 083 [arXiv:1203.4951]. * [58] K. Asai, Predictions for the neutrino parameters in the minimal model extended by linear combination of U(1)${}_{L_{e}-L_{\mu}}$, U(1)${}_{L_{\mu}-L_{\tau}}$ and U(1)B-L gauge symmetries, Eur. Phys. J. C 80 (2020) 76 [arXiv:1907.04042]. * [59] P. Minkowski, $\mu\to e\gamma$ at a Rate of One Out of $10^{9}$ Muon Decays?, Phys. Lett. B 67 (1977) 421–428. * [60] T. Yanagida, Horizontal Symmetry and Masses of Neutrinos, Prog. Theor. Phys. 64 (1980) 1103. * [61] M. Gell-Mann, P. Ramond, and R. Slansky, Complex Spinors and Unified Theories, Conf. Proc. C 790927 (1979) 315–321 [arXiv:1306.4669]. * [62] R. N. Mohapatra and G. Senjanovic, Neutrino Mass and Spontaneous Parity Nonconservation, Phys. Rev. Lett. 44 (1980) 912. * [63] L. Lavoura, Zeros of the inverted neutrino mass matrix, Phys. Lett. B 609 (2005) 317–322 [hep-ph/0411232]. * [64] E. I. Lashin and N. Chamoun, Zero minors of the neutrino mass matrix, Phys. Rev. D 78 (2008) 073002 [arXiv:0708.2423]. * [65] K. Asai, K. Hamaguchi, and N. Nagata, Predictions for the neutrino parameters in the minimal gauged U(1)${}_{L_{\mu}-L_{\tau}}$ model, Eur. Phys. J. C 77 (2017) 763 [arXiv:1705.00419]. * [66] K. Asai, K. Hamaguchi, N. Nagata, S.-Y. Tseng, and K. Tsumura, Minimal Gauged U(1)${}_{L_{\alpha}-L_{\beta}}$ Models Driven into a Corner, Phys. Rev. D 99 (2019) 055029 [arXiv:1811.07571]. * [67] K. Asai, K. Hamaguchi, N. Nagata, and S.-Y. Tseng, Leptogenesis in the minimal gauged U(1)${}_{L_{\mu}-L_{\tau}}$ model and the sign of the cosmological baryon asymmetry, JCAP 11 (2020) 013 [arXiv:2005.01039]. * [68] I. Esteban, M. C. Gonzalez-Garcia, M. Maltoni, T. Schwetz, and A. Zhou, The fate of hints: updated global analysis of three-flavor neutrino oscillations, JHEP 09 (2020) 178 [arXiv:2007.14792]. * [69] DES Collaboration, Dark Energy Survey Year 3 results: Cosmological constraints from galaxy clustering and weak lensing, Phys. Rev. D 105 (2022) 023520 [arXiv:2105.13549].
# Generalised Spherical Text Embedding Souvik Banerjee1, Bamdev Mishra2, Pratik Jawanpuria2 Manish Shrivastava1 1 International Institute of Information Technology, Hyderabad 2 Microsoft India <EMAIL_ADDRESS> <EMAIL_ADDRESS> <EMAIL_ADDRESS> <EMAIL_ADDRESS> ###### Abstract This paper aims to provide an unsupervised modelling approach that allows for a more flexible representation of text embeddings. It jointly encodes the words and the paragraphs as individual matrices of arbitrary column dimension with unit Frobenius norm. The representation is also linguistically motivated with the introduction of a novel similarity metric. The proposed modelling and the novel similarity metric exploits the matrix structure of embeddings. We then go on to show that the same matrices can be reshaped into vectors of unit norm and transform our problem into an optimization problem over the spherical manifold. We exploit manifold optimization to efficiently train the matrix embeddings. We also quantitatively verify the quality of our text embeddings by showing that they demonstrate improved results in document classification, document clustering, and semantic textual similarity benchmark tests.111https://github.com/SouvikBan/matrix_rep. ## 1 Introduction Most unsupervised text embedding models are trained by encoding the words or paragraphs acquired from the training data as a feature length vector, with the assumption that they reside in the Euclidean space. Such models are ubiquitous for good reason. Aside from their efficiency, they have also proven to be very effective providing us with state of the art results in various intrinsic and extrinsic embedding evaluation tasks. Word2vecMikolov et al. (2013b, a), and GLoVE Pennington et al. (2014) are two notable examples where word embeddings are learned in the Euclidean space and are trained to be oriented such that word vectors that appear in the same context have higher cosine similarity. Some of the most common methods of intrinsic evaluation of word embeddings include word similarity, word analogy, and compositionality. Doc2vec Le and Mikolov (2014), an unsupervised document embedding model generalises the training method introduced in Word2vec to documents and achieves improved results in various downstream tasks like sentiment analysis, information retrieval and multi-class classification. There are other document embedding models like skip-thought Kiros et al. (2015) and infersent Conneau et al. (2017); Moghadasi and Zhuang (2020). The joint spherical embedding model, JoSE as proposed in Meng et al. (2019), shows that directional similarity is often more effective in tasks such as word similarity and document clustering. They show that when embeddings are trained in the Euclidean space, there is a performance gap between the training stage and usage stage of text embeddings. To bridge that gap, they propose a model which trains both words and paragraphs on a spherical space with tools from Riemannian optimization methods. The resulting embeddings are also shown to give considerably better results in word similarity, document clustering, and document classification tasks when compared with other standard models. Such application of manifold geometry has also been explored in substantial depth in works like Batmanghelich et al. (2016); Reisinger et al. (2010); Gopal and Yang (2014). There are also other notable Riemannian optimization based embedding training models like Tifrea et al. (2018); Nickel and Kiela (2017) which train embeddings on the hyperbolic manifold space and uses its tree like property for better hierarchical representation of data. Hyperbolic word embeddings are also intrinsically linked with Gaussian word embeddings Vilnis and McCallum (2014) which gives a lot more insight into the geometry of word embeddings. However, most of these text embedding models like JoSE, Word2vec, Doc2vec, and fastText Bojanowski et al. (2017); Meng et al. (2020) are trained with the goal of getting a single dense vector representation per word or document. These models treat both polysemous and monosemous words in the same way resulting in the most frequent meaning of the word dominating the others or the meanings getting mixed in the case of former. It is especially detrimental for documents where we use a single dense vector representation to encode information which span over several sentences, often involving multiple topics. This paper aims to address this problem by using matrices as the mode of representation instead of vectors. Our model is the joint word and document training generative model proposed in JoSE where we replace the cosine similarity metric with a novel metric that exploits the matrix structure of the embeddings. This robust metric takes word or document matrices of arbitrary number of columns and calculates the similarity between them. We also show that a few reshape operations allow us to reformulate the optimization problem of our model in terms of the spherical manifold optimization problem. Thus, we offer more flexibility in the way of matrix dimensions while retaining efficiency. Our choice of metric also suggests that the word, sentence, paragraph/document embeddings do not need to have the same number of columns, which has linguistic validation. ## 2 Matrix Representation of Texts and Optimization Problem The text embeddings are represented as elements of the following set $\mathcal{S}(p,r)=\\{\textbf{X}\in\mathbb{R}^{p\times r}:||\textbf{X}||_{F}=1\\},$ where $r\leq p$ and $\|\cdot\|_{F}$ denotes the Frobenius norm. The Frobenius norm is the matrix norm of a $p\times r$ matrix X defined as the square root of the sum of the absolute squares of its elements, i.e., $||\textbf{X}||_{F}=\sqrt{\sum_{i=1}^{p}\sum_{j=1}^{r}x_{ij}^{2}}.$ Our model design is consistent with JoSE where it is assumed that text generation is a two-step process: a center word is first generated according to the semantics of the paragraph, and then the surrounding words are generated based on the center word’s semantics. Consider a positive tuple $(\mathcal{U},\mathcal{V},\mathcal{D})$ where word $\mathcal{V}$ appears in the local context window of word $\mathcal{U}$ in paragraph $\mathcal{D}$ and negative tuple $(\mathcal{V},\mathcal{U}^{\prime},\mathcal{D})$ where $\mathcal{U}^{\prime}$ is a randomly sampled word from the vocabulary serving as a negative sample. We represent words $\mathcal{V},\mathcal{U},\mathcal{U}^{\prime}$ as matrices $\textbf{V},\textbf{U},\textbf{N}$ which are elements of the set $\mathcal{S}(p,r_{1})$ and paragraph $\mathcal{D}$ as matrix D which is an element of the set $\mathcal{S}(p,r_{2})$, where $p,r_{1},r_{2}>0$. From a linguistic perspective, these matrices can be considered as a set of latent variables that govern the semantics of a word or a document. Each column is given some arbitrary unit of linguistic information to encode, a latent variable which contributes to the mathematical representation of a word or a document. For example, the columns of a matrix D that represent the document $\mathcal{D}$ might encode latent variables that contain information about some topic contained in that document. Similarly, the columns of the word matrix U might encode information about a specific context in which a polysemous word $\mathcal{U}$ appears. We also keep the number of columns for word matrices less than or equal to the number of columns for sentence/document matrices, i.e., $r_{1}\leq r_{2}$, so that the number of latent variables governing a word should not be more than the ones that govern a sentence or paragraph. Novel metric. To model the above mentioned linguistic representation mathematically, we define a novel similarity metric for the ambient space in which we train our matrix embeddings. The proposed metric function is a measure of similarity between two sets of latent variables (matrices) - a function analogous to the cosine similarity measure for vectors in the Euclidean space. Given two arbitrary matrices $\textbf{A}\in\mathcal{S}(p,r_{1}),\textbf{B}\in\mathcal{S}(p,r_{2})$, we propose the similarity metric $\textbf{g}:\mathcal{S}(p,r_{1})\times\mathcal{S}(p,r_{2})\rightarrow\mathbb{R}$ as $\textbf{g}(\textbf{A},\textbf{B})=\frac{\sum_{i=1}^{r_{1}}\sum_{j=1}^{r_{2}}a_{i}^{\top}b_{j}}{r_{1}r_{2}},$ (1) where $\textbf{A}=[a_{1}\,a_{2}\,\cdots\,a_{r_{1}}]$ , $\textbf{B}=[b_{1}\,b_{2}\,\cdots\,b_{r_{2}}]$ , $a_{i},b_{j}\in\mathbb{R}^{p}\,\forall i\,\in[1,2,\cdots r_{1}]$ , $\forall\,j\in[1,2,\cdots r_{2}]$. Motivation for our similarity metric. The metric g (1) calculates the average of all the entries in the matrix $\textbf{A}^{\top}\textbf{B}$. The linguistic intuition behind the choice of this metric is that we want to define a metric that takes the average of dot products between all possible pairs of latent variables (columns) from each matrix. In the case of $r_{1}=r_{2}=1$, $\textbf{g}(\textbf{A},\textbf{B})$ reduces to the cosine similarity metric between unit norm vectors A and B which is the metric used in the spherical space model of JoSE. However, in the case of higher values of $r_{1},r_{2}$, For example, let two words $\mathcal{V}_{1}$ and $\mathcal{V}_{2}$ be represented by proposed $p\times r_{1}$ matrix embeddings - $\textbf{V}_{1}=[\text{a}_{1},\text{a}_{2}]$ and $\textbf{V}_{2}=[\text{b}_{1},\text{b}_{2}]$, where $p=1$ and $r_{1}=2$. The proposed similarity metric $\textbf{g}(\textbf{V}_{1},\textbf{V}_{2})$ is computed as ${(a_{1}b_{1}+a_{1}b_{2}+a_{2}b_{1}+a_{2}b_{2}})/{4}$. Note that this is different from computing the cosine similarity which gives $(a_{1}b1+a_{2}b_{2})$. Moreover, the regular cosine similarity between word and paragraph embedding matrices with unequal dimensions $(p\times r_{1}$ and $p\times r_{2}$ respectively) is not defined. On the other hand, our proposed similarity metric is still applicable. Modelling. As our model has the same generative process as JoSE, we take the same max-margin loss function and substitute the cosine similarity metric with our new similarity metric g where the word matrices $\textbf{U},\textbf{V},\textbf{N}\in\mathcal{S}(p,r_{1})$ and paragraph matrix $\textbf{D}\in\mathcal{S}(p,r_{2})$ with $r_{1}\leq r_{2}$. We get the following loss, i.e., $\begin{array}[]{lll}\mathcal{L}(\textbf{V},\textbf{U},\textbf{N},\textbf{D})\\\ \quad\quad=\max\,(0,m-\textbf{g}(\textbf{V},\textbf{U})-\textbf{g}(\textbf{U},\textbf{D})\\\ \qquad\qquad\qquad\quad+\ \textbf{g}(\textbf{V},\textbf{N})+\textbf{g}(\textbf{N},\textbf{D})).\end{array}$ (2) where $m>0$ is the margin. Optimization. For the purpose of optimization, matrices of different dimensions are reshaped and embedded into Riemannian spherical manifolds of different dimensions. Overall, they are combined using the Riemannian product manifold structure. Therefore, the optimization of $\mathcal{L}$ (2) is done by performing two reshape operations per iteration while training. For example, the unit Frobenius norm matrices of dimension $\mathbb{R}^{p\times r}$ can be reshaped into vectors of dimension $\mathbb{R}^{pr}$ with the unit norm. To calculate the value of our loss function (2) at every iteration and the Euclidean gradient (partial derivatives), the vectors in question are reshaped into matrices for calculating the g values and their gradients. Subsequently, the matrices are reshaped back into vectors. We then apply the Riemannian gradient descent update rule to update the parameters Meng et al. (2019); Absil et al. (2008); Smith (2014); Edelman et al. (1998). Note that our proposed modelling and optimization are different from just training on the spherical manifold with unit vectors and using the cosine similarity metric (which is the case in JoSE). ## 3 Experiments To highlight the quality of our obtained matrix representations, we run the same set of evaluations as JoSE with a relatively lower number of columns, i.e., $1\leq r_{1}\leq r_{2}\leq 6$. We notice that for even higher values, the quality of our embeddings gradually decrease. Moreover, the word similarity experiment results are not added as words seemingly do not benefit from our representation directly. Indeed, the best word similarity score are obtained for $r_{1}=r_{2}=1$. Instead, we add semantic textual similarity benchmark tests to show that sentences can benefit from this matrix representation model. Unless otherwise stated, our model and JoSE are trained for 35 iterations on the respective corpora; the local context window size is 5; the embedding dimension is kept at 100; the number of negative samples are 2. Other hyperparameters in our model are kept the same as JoSE. ### 3.1 Document Clustering We perform document clustering on the 20 Newsgroup222http://qwone.com/~jason/20Newsgroups/ dataset using spectral clustering. Each paragraph in the dataset is separated by a new line and is considered a separate document while training. JoSE uses K-Means and SK-Means as the clustering algorithm that assume the ambient space to be the Euclidean and the spherical space, respectively. Our non-Euclidean space with its custom metric requires a clustering algorithm that allows the freedom of using custom metric, i.e., the algorithm should be space agnostic. We found spectral clustering to suit those requirements perfectly. The four external measures used for validating the results are kept unchanged from JoSE Banerjee et al. (2005); Manning et al. (2008); Steinley (2004). These measures are Mutual Information (MI), Normalized Mutual Information (NMI), Adjusted Rand Index (ARI), and Purity. We run the clustering algorithm with our custom similarity metric as written in (1) with kernel coefficient, $\gamma=0.001$. Table 1 shows quantitatively how matrix representations benefit document embeddings for clustering tasks. Keeping $r_{1}=1$ fixed, we see a steady increase in performance as $r_{2}$ is increased from 1 (the score of our baseline model - JoSE ) to 6. Table 1: Evaluation results for spectral clustering of document embeddings on the 20 Newsgroup dataset for kernel coefficient, $\gamma=0.001$ ($r_{1}=1,r_{2}=1$ is JoSE score). Document embeddings benefit from matrix representations as demonstrated by better scores for higher values of $r_{2}$. Here, $r_{1}=1$. $r_{2}$, $r_{1}=1$ | MI | NMI | ARI | Purity ---|---|---|---|--- $r_{2}=1$ | 1.73 | 0.58 | 0.45 | 0.64 $r_{2}=2$ | 1.75 | 0.59 | 0.46 | 0.63 $r_{2}=3$ | 1.75 | 0.59 | 0.46 | 0.62 $r_{2}=4$ | 1.77 | 0.60 | 0.46 | 0.63 $r_{2}=5$ | 1.84 | 0.62 | 0.49 | 0.65 $r_{2}=6$ | 1.85 | 0.62 | 0.49 | 0.67 ### 3.2 Document Classification Following Meng et al. (2019), the document classification evaluations are ran on the following two datasets: the topic classification 20 Newsgroup dataset (which we used for document clustering as well) and a binary sentiment classification dataset consisting of 1 000 positive and 1 000 negative movie reviews333http://www.cs.cornell.edu/people/pabo/movie-review-data/. The train/test split is the original split for 20 Newsgroup while for the movie review datasets, the splitting is done by randomly selecting 80% of the data as training and 20% as testing. The classification algorithm we use is K-NN with $k=3$ and a custom distance metric that is suitable for our space. The custom distance metric for two paragraph matrices $\textbf{U}=[u_{1}\,u_{2}\,\cdots\,u_{r_{2}}]$ and $\textbf{V}=[v_{1}\,v_{2}\,\cdots\,v_{r_{2}}]$ where $\textbf{U},\textbf{V}\in\mathcal{S}(p,r_{2})$ and $u_{i},v_{i}\in\mathbb{R}^{p}\,\forall i\in[1,2,\cdots,r_{2}]$ is defined as $\begin{array}[]{lll}\text{dist}^{2}(\textbf{U},\textbf{V})=\frac{{\sum_{k=1}^{r_{2}}\sum_{l=1}^{r_{2}}(u_{k}-v_{l})^{\top}(u_{k}-v_{l})}}{r_{2}^{2}}.\end{array}$ (3) The intuition for the distance metric in (3) comes from our interpretation of each individual column as encoding a latent variable governing the semantics of that specific document. A quick look at (3) tells us that the distance metric takes the square root of the average of the squared Euclidean distances between all pairs of columns formed from one matrix with another. Tables 2 and 3 list the Macro-F1 and Micro-F1 scores for 20 Newsgroup dataset and Movie Reviews dataset respectively for increasing values of $r_{1}$ and $r_{2}$. We again see an increase in scores for higher values of both $r_{2}$ and $r_{1}$ compared to JoSE ($r_{1}$=1, $r_{2}$=1). Table 2: F1-macro, F1-micro for 20 Newsgroup dataset classification using K-NN with K=3 ($r_{1}=1,r_{2}=1$ is JoSE score). Increasing the value $r_{2}$ benefits documents embeddings in classification tasks. $r_{1}$\$r_{2}$ | $r_{2}$=1 | $r_{2}$=2 | $r_{2}$=3 | $r_{2}$=4 ---|---|---|---|--- $r_{1}$=1 | 0.74, 0.74 | 0.77, 0.77 | 0.78, 0.78 | 0.78, 0.78 $r_{1}$=2 | – | 0.76, 0.76 | 0.77, 0.77 | 0.76, 0.77 $r_{1}$=3 | – | – | 0.76, 0.76 | 0.73, 0.74 $r_{1}$=4 | – | – | – | 0.72, 0.72 Table 3: F1-macro, F1-micro for movie review dataset classification using K-NN with K=3 ($r_{1}=1,r_{2}=1$ is the JoSE score). $r_{1}$\$r_{2}$ | $r_{2}$=1 | $r_{2}$=2 | $r_{2}$=3 | $r_{2}$=4 ---|---|---|---|--- $r_{1}$=1 | 0.74, 0.74 | 0.75, 0.75 | 0.76, 0.76 | 0.75, 0.76 $r_{1}$=2 | – | 0.75, 0.75 | 0.75, 0.75 | 0.74, 0.74 $r_{1}$=3 | – | – | 0.74, 0.74 | 0.76, 0.76 $r_{1}$=4 | – | – | – | 0.74, 0.74 ### 3.3 Semantic Textual Similarity Task Semantic Textual Similarity Benchmark comprises a selection of the English datasets used in the STS tasks organized in the context of SemEval Cer et al. (2017) between 2012 and 2017444http://ixa2.si.ehu.eus/stswiki/index.php/STSbenchmark. We perform semantic textual similarity tasks on the sts-benchmark dataset to show that even sentences can benefit from being represented as matrices. The benchmark comprises of 8 628 sentence pairs split into 3 partitions: train, development and test. The results are reported on both the test and dev sets. Each sentence in the dataset is treated as a separate document by our model and we use all the sentences in the train, development and test set to train. The rationale for this is that the model is completely unsupervised, i.e., it takes only the raw text and uses no supervised or annotated information, and thus there is no need to hold out the test data as it is unlabelled. We train for 1 000 iters with window size 15 and negative samples 5 while the rest of the hyperparameters were kept at their default values. To score a sentence pair representation, similarity was computed between them using our custom metric described in 1 for our model. We report the dev and test Pearson correlation score for $r_{1},r_{2}=1,2,3,4,\,r_{1}\leq r_{2}$. As Table 4 reports, higher values of $r_{2}$ give better scores compared to our baseline model JoSE ($r_{1}=r_{2}=1$). Table 4: Pearson Correlation for STS Benchmark on dev and test data ($r_{1}=1,r_{2}=1$ is the JoSE score). Even sentences can benefit from our matrix representation as demonstrated by better scores with higher values of $r_{2}$. $r_{1}$\$r_{2}$ | $r_{2}$=1 | $r_{2}$=2 | $r_{2}$=3 | $r_{2}$=4 ---|---|---|---|--- $r_{1}$=1 | 0.51, 0.40 | 0.51, 0.39 | 0.52, 0.40 | 0.53, 0.40 $r_{1}$=2 | – | 0.53, 0.40 | 0.53, 0.40 | 0.53, 0.40 $r_{1}$=3 | – | – | 0.53, 0.40 | 0.54, 0.40 $r_{1}$=4 | – | – | – | 0.53, 0.40 ## 4 Conclusion In this paper, we extend the joint modelling idea used for training text embeddings from vectors with unit norm to matrices with unit Frobenius norm. Each word/sentence/document matrix is made to encode information in a way that each column of the matrix represents some latent topic, context, or discourse. Since the standard vector dot product can no longer be applied, we introduce a novel similarity metric that allows the measurement of similarity between matrices of arbitrary number of columns. For optimization simplicity, we reshape our matrices to vectors of unit norm that allows using the Riemannian gradient descent optimization algorithm on the spherical manifold. Our theory is validated quantitatively by the results which shows that our text embeddings outperform or produce similar results when compared with JoSE in document classification, clustering, and semantic textual similarity tasks. This paper is meant to serve as a ground work for more involved research topics which integrate concepts of differential geometry and NLP. Future directions could include qualitative analysis on the columns of the matrices to see what tangible information they encode which will allow for better modelling. Another direction would be to exploit other matrix structures on the embeddings, e.g., treating each word embedding as a symmetric positive definite matrix and to study whether they can be beneficial. ## References * Absil et al. (2008) P.-A. Absil, R. Mahony, and R. Sepulchre. 2008. _Optimization Algorithms on Matrix Manifolds_. Princeton University Press, Princeton, NJ. * Banerjee et al. (2005) Arindam Banerjee, Inderjit S. Dhillon, Joydeep Ghosh, and Suvrit Sra. 2005. Clustering on the unit hypersphere using von mises-fisher distributions. _J. Mach. Learn. Res._ , 6:1345–1382. * Batmanghelich et al. (2016) Kayhan Batmanghelich, Ardavan Saeedi, Karthik Narasimhan, and Sam Gershman. 2016\. Nonparametric spherical topic modeling with word embeddings. In _Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)_ , pages 537–542, Berlin, Germany. Association for Computational Linguistics. * Bojanowski et al. (2017) Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2017. Enriching word vectors with subword information. _Transactions of the Association for Computational Linguistics_ , 5:135–146. * Cer et al. (2017) Daniel Cer, Mona Diab, Eneko Agirre, Inigo Lopez-Gazpio, and Lucia Specia. 2017\. SemEval-2017 task 1: Semantic textual similarity multilingual and crosslingual focused evaluation. In _Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017)_. Association for Computational Linguistics. * Conneau et al. (2017) Alexis Conneau, Douwe Kiela, Holger Schwenk, Loic Barrault, and Antoine Bordes. 2017\. Supervised learning of universal sentence representations from natural language inference data. * Edelman et al. (1998) Alan Edelman, Tomás A. Arias, and Steven T. Smith. 1998. The geometry of algorithms with orthogonality constraints. _SIAM Journal on Matrix Analysis and Applications_ , 20(2):303–353. * Gopal and Yang (2014) Siddharth Gopal and Yiming Yang. 2014. Von mises-fisher clustering models. In _Proceedings of the 31st International Conference on Machine Learning_ , volume 32 of _Proceedings of Machine Learning Research_ , pages 154–162, Bejing, China. PMLR. * Kiros et al. (2015) Ryan Kiros, Yukun Zhu, Ruslan Salakhutdinov, Richard S. Zemel, Antonio Torralba, Raquel Urtasun, and Sanja Fidler. 2015. Skip-thought vectors. * Le and Mikolov (2014) Quoc V. Le and Tomas Mikolov. 2014. Distributed representations of sentences and documents. * Manning et al. (2008) Christopher D. Manning, Prabhakar Raghavan, and Hinrich Schütze. 2008. _Introduction to Information Retrieval_. Cambridge University Press, USA. * Meng et al. (2019) Yu Meng, Jiaxin Huang, Guangyuan Wang, Chao Zhang, Honglei Zhuang, Lance Kaplan, and Jiawei Han. 2019. Spherical text embedding. In _Advances in neural information processing systems_. * Meng et al. (2020) Yu Meng, Yunyi Zhang, Jiaxin Huang, Yu Zhang, Chao Zhang, and Jiawei Han. 2020. Hierarchical topic mining via joint spherical tree and text embedding. KDD ’20, page 1908–1917, New York, NY, USA. Association for Computing Machinery. * Mikolov et al. (2013a) Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013a. Efficient estimation of word representations in vector space. _arXiv preprint arXiv:1301.3781_. * Mikolov et al. (2013b) Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013b. Distributed representations of words and phrases and their compositionality. In _Advances in neural information processing systems_ , pages 3111–3119. * Moghadasi and Zhuang (2020) Mahdi Naser Moghadasi and Yu Zhuang. 2020. Sent2vec: A new sentence embedding representation with sentimental semantic. In _2020 IEEE International Conference on Big Data (Big Data)_ , pages 4672–4680. * Nickel and Kiela (2017) Maximilian Nickel and Douwe Kiela. 2017. Poincaré embeddings for learning hierarchical representations. * Pennington et al. (2014) Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. In _Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)_ , pages 1532–1543. * Reisinger et al. (2010) Joseph Reisinger, Austin Waters, Bryan Silverthorn, and Raymond J. Mooney. 2010\. Spherical topic models. ICML’10, page 903–910, Madison, WI, USA. Omnipress. * Smith (2014) Steven Thomas Smith. 2014. Optimization techniques on riemannian manifolds. * Steinley (2004) Douglas L. Steinley. 2004. Properties of the hubert-arabie adjusted rand index. _Psychological methods_ , 9 3:386–96. * Tifrea et al. (2018) Alexandru Tifrea, Gary Bécigneul, and Octavian-Eugen Ganea. 2018. Poincaré glove: Hyperbolic word embeddings. * Vilnis and McCallum (2014) Luke Vilnis and Andrew McCallum. 2014. Word representations via gaussian embedding.
[title=Index of definitions and notations,intoc] # Existence obstructions for separating morphisms and totally real pencils Matilde Manzaroli ###### Abstract. It goes back to Ahlfors that a real algebraic curve $C$ admits a separating morphism $f$ to the complex projective line if and only if the real part of the curve disconnects its complex part, i.e. the curve is separating. The degree of such $f$ is bounded from below by the number $l$ of real connected components of $\mathbb{R}C$. The sharpness of this bound is not a priori clear. We prove that real algebraic separating curves, embedded in some ambient surface and with $l$ bounded in a certain way, do not admit separating morphisms of lowest possible degree. Moreover, this result of non-existence can be applied to show that certain real separating plane curves of degree $d$, do not admit totally real pencils of curves of degree $k$ such that $kd\leq l$. Finally, we construct an explicit family of real separating plane curves of increasing degree $2s+1$ such that each member of the family does not admit totally real pencils of curves of degree lower or equal to $\frac{s^{2}+1}{2s+1}$. ###### Contents 1. 1 Introduction 2. 2 Preliminaries 3. 3 Theorem 1.13 and applications ## 1\. Introduction A real algebraic variety is a compact complex algebraic variety $X$ equipped with an anti-holomorphic involution $\sigma:X\rightarrow X$, called real structure. The real part $\mathbb{R}X$ of $X$ is the set of points fixed by the involution $\sigma$. Let $C$ be any non-singular real algebraic curve. Denote by $C_{1},\dots,C_{l}$ the connected components of $\mathbb{R}C$. Harnack-Klein’s inequality [Har76, Kle73] bounds $l$ by the genus $g$ of $C$ plus one. If $C\setminus\mathbb{R}C$ is connected, we say that $C$ is of type II or non- separating, otherwise of type I or separating [Kle73]. Looking at the real part of the curve and its position with respect to its complexification gives information about $l$ and viceversa. For example, we know that if $l$ equals $g+1$, then $C$ is of type I. Or, if $C$ is of type I then $l$ has the parity of $g+1$. Rokhlin promoted and contributed to this new point of view; an example of Rokhlin’s contribution is the introduction and the study of the complex orientations of a separating real algebraic curve [Rok72]. Namely, if $C$ is of type I, the two halves of $C\setminus\mathbb{R}C$ induce two opposite orientations on $\mathbb{R}C$ called complex orientations of the curve. Looking at complex orientations of separating real curves has allowed a change of prospective and a remarkable progress in the study of real algebraic curves, their topology and a refinement of their classifications. ###### Definition 1.1. We say that a real morphism $f$ from a real algebraic curve $C$ to the complex projective line $\mathbb{C}\mathbb{P}^{1}$ is separating if $f^{-1}(\mathbb{R}\mathbb{P}^{1})=\mathbb{R}C$. According to Ahlfors [Ahl50, §4.2], there exists a separating morphism $f:C\rightarrow\mathbb{C}\mathbb{P}^{1}$ if and only if $C$ is separating. In this paper, we focus on the relation of separating curves embedded in some ambient surface and their separating morphisms; see Definition 1.1. Before stating the main result of this paper, Theorem 1.13, in its generality, we present it restricted to the case of real algebraic plane curves. ### 1.1. Real plane curves Let us consider separating real algebraic plane curves $C$. Let $l$ denote the number of the connected components of $\mathbb{R}C$. The following proposition show that there is a relation between $l$ and the existence of separating morphisms $f:C\rightarrow\mathbb{C}\mathbb{P}^{1}$ of degree $r(l)$. ###### Proposition 1.2 (Particular case of Theorem 1.13). Let $C$ be a non-singular real algebraic plane curve of type I and with $l$ real connected components. Assume one of the following: 1. (1) The degree of $C$ is $2s+1$, for some $s\in\mathbb{Z}_{>1}$ and $l+\varepsilon\leq s^{2}+s-2,$ where $\varepsilon\in\\{0,1\\}$ such that $l+\varepsilon$ is even. 2. (2) The degree of $C$ is $2s$, for some $s\in\mathbb{Z}_{\geq 4}$ and $l+\varepsilon\leq s^{2}-2s+2,$ where $\varepsilon\in\\{0,1\\}$ such that $l-s+\varepsilon$ is even. Then $C$ admits no separating morphisms of degree $l$. Proposition 1.2 falls within the line of results relating topology, complex orientations and properties of separating plane curves, such as Rokhlin’s complex orientation formulas (2.1), (2.2), and such as [Ore21, Theorem 1.1], where Orevkov shows that there are finer relations for the numbers which intervene in (2.2). As corollaries of Proposition 1.2, one can find obstructions for the existences of totally real pencils of curves of a certain degree for given separating plane curves; see Section 1.3. ### 1.2. Generalities of separating morphisms A separating morphism $f:C\rightarrow\mathbb{C}\mathbb{P}^{1}$ is always unramified once restricted to $\mathbb{R}C$; see [KS20a, Theorem 2.19]. Therefore, the restriction of $f$ to each connected component of $\mathbb{R}C$ is a covering map of $\mathbb{R}\mathbb{P}^{1}$. This implies that the degree of a separating morphism is at least as big as the number of connected components of $\mathbb{R}C$. Actually, the definition of separating morphism is more general and includes real morphisms between any real algebraic varieties of same dimension. In order to a have a general idea of the subject, we refer the interested reader to [KS20a] and [KTM22]. In the context of this paper, we only need Definition 1.1. In [Hui01], [Gab06],[CH13], [Cop13], [KS20b], [Ore18] properties of separating morphisms and their existence are treated. For example, [Gab06, Theorem 7.1] states that a genus $g$ real separating curve with $l$ real connected components admits a separating morphism of degree at most $\frac{g+l+1}{2}$. Later, Coppens, in [Cop13], constructed, for every value $h$ between $\tilde{l}$, the minimum for a separating morphism degree, and $\frac{\tilde{g}+\tilde{l}+1}{2}$, a separating curve $\tilde{C}$ of genus $\tilde{g}$ and with $\tilde{l}$ real connected components such that $h$ is the smallest possible degree of a separating morphism of $\tilde{C}$. In [KS20b], the authors fix a separating curve $C$ and study all separating morphisms of $C$ as follows. Set $d_{i}(f)\in\mathbb{N}$ the degree of the covering map $f:C_{i}\rightarrow\mathbb{R}\mathbb{P}^{1}$ and set $d(f):=(d_{1}(f),\dots,d_{l}(f))$. Let us denote by $t:=\sum_{i=1}^{l}d_{i}(f)$ the degree of $f$. The set $Sep(C)$ of all such degree partitions forms a semigroup, called separating semigroup, and for all element $d\in Sep(C)$, it is shown that $d+\mathbb{Z}^{l}_{\geq 0}$ is also contained in $Sep(C)$. So, in order to understand $Sep(C)$ for a given separating curve $C$, it is important to understand which minimal possible element $Sep(C)$ contains, where with minimal we mean that $t$ is of minimal possible value. ###### Remark 1.3. Thanks to Harnack-Klein inequality, any real curve of genus $g$ cannot have more than $g+1$ real connected components. By Riemann-Roch theorem, all non- singular real curves with $l=g+1$ real connected components admits a separating morphism of degree $l$. But, whenever $l<g+1$, it is not a priori clear whether a separating morphism of degree $l$ exists. ### 1.3. Obstruction of existence of totally real pencils Kummer and Shaw, in [KS20b], also prove that for all real separating curves embedded in the complex projective plane, there exist infinitely many totally real pencils of curves. ###### Theorem 1.4. ([KS20b, Theorem 1.6]) Let $C$ be a non-singular real algebraic plane curve of type I. Then there exists an integer $k$ such that $\forall k^{\prime}\geq k$, there are $f,g\in\mathbb{R}[x,y,z]_{k^{\prime}}$ such that $V(f,g)\cap C=\emptyset$ and $V(\lambda f+\mu g)\cap C$ consists of real points only for all $\lambda,\mu\in\mathbb{R}$ not both zero. ###### Definition 1.5. Let $C$ be a non-singular real algebraic plane curve of type I. We say that $C$ admits a totally real pencil of curves of degree $k$ if there exists an integer $k$ such that there are $f,g\in\mathbb{R}[x,y,z]_{k}$ such that $V(f,g)\cap C=\emptyset$ and $V(\lambda f+\mu g)\cap C$ consists of real points only for all $\lambda,\mu\in\mathbb{R}$ not both zero. In [KS20b, Question 1], the authors wonder which may be the minimal possible value of $k$ in Theorem 1.4. An immediate corollary of Proposition 1.2 gives a lower bound for $k$, as follows. ###### Corollary 1.6. Let $C$ be a non-singular degree $d$ real algebraic plane curve of type I and with $l$ real connected components. Assume that $C$ respects either $(1)$ or $(2)$ of Proposition 1.2. Then $C$ admits no totally real pencil of curves of degree $k$ such that $kd\leq l$. ###### Remark 1.7. As remarked in [KS20b, Remark 3.5], according to [Ahl50, §4.2], real separating plane curves admit separating morphisms (whose degree is a priori unknown). Hence, there exists an integer $k$ such that there are $f,g\in\mathbb{R}[x,y,z]_{k}$ such that $V(\lambda f+\mu g)\cap C$ consists of real points only for all $\lambda,\mu\in\mathbb{R}$ not both zero. The difference with Definition 1.5 is that $V(f,g)\cap C$ may be non-empty. Analogously, [Gab06, Theorem 1.7] implies the existence of separating morphisms of degree between $l$ and $\frac{g+l+1}{2}$ and, once again, this does not imply that a given real separating plane curve of genus $g$ and with $l$ real connected components, admits totally real pencils of curves of degree $k(l,g)$ in the sense of Definition 1.5. In order to have examples of separating real curves respecting the hypothesis of Proposition 1.2, we construct an explicit family of real separating plane curves of increasing degree $2s+1$ such that each member of the family admits no totally real pencils of curves of degree lower than $\frac{s^{2}+1}{2s+1}$; see Proposition 1.8. In particular, Proposition 1.8 presents some explicit examples of real curves in which the bound for the minimal degree of separating morphisms is higher than the Brill-Noether bound. Namely, the Brill-Noether theorem states that a generic curve of genus $g$ has gonality $\lfloor\frac{g+3}{2}\rfloor$. This implies in particular that for a generic real curve of genus $g$, there cannot be any separating morphism of degree less than $\frac{g+3}{2}$. For the real separating plane curves $C$ of odd degree $2s+1$ constructed to prove Proposition 1.8, this gonality $\displaystyle{\frac{2s^{2}-s+3}{2}}$ is strictly smaller than $s^{2}+1$, which is the number of connected components of $\mathbb{R}C$. ###### Proposition 1.8. For all $s\in\mathbb{N}_{\geq 2}$, there exists a non-singular real plane curve $B_{d}$ of degree $d=2s+1$, of type I and with $s^{2}+1$ real connected components, which admit no totally real pencils of curves of degree $k$ such that $k\leq\frac{s^{2}+1}{2s+1}$. ###### Remark 1.9. We do not know yet whether the curves $B_{2s+1}$, constructed (see Lemma 3.2) to prove Proposition 1.8, admit totally real pencils of degree $k^{\prime}=\lceil\frac{s^{2}+1}{2s+1}\rceil$. ### 1.4. Statement of the main result and strategy In this paper, we consider real separating curves embedded in some ambient real surface and focus on their separating morphisms. The key tools of our approach are the fact that all separating curves come equipped with two possible complex orientations and the use of [Ore21, Theorem 3.2] as shown in [Ore21, Example 3.3]. Before recalling [Ore21, Theorem 3.2] and state the main result of this paper (Theorem 1.13), let us introduce some notation. ###### Notation 1.10. Let $X$ be a smooth real algebraic surface and $C\subset X$ a non-singular real algebraic curve of class $d$ in $H_{2}(X;\mathbb{Z})$. Let us consider a real divisor $D$ belonging to the linear system $|C+K_{X}|$. Assume that $D$ has not $C$ as a component. This divisor can always be written as $2D_{0}+D_{1}$ with the following properties. The divisor $D_{0}$ is an effective divisor in $H_{2}(X;\mathbb{Z})$ and $D_{1}$ is a reduced curve realising some class $\beta\in H_{2}(X;\mathbb{Z})$. ###### Example 1.11. Let $C$ be a non-singular real algebraic curve of degree $2k$ in $\mathbb{C}\mathbb{P}^{2}$. Then $D_{0}$ can be chosen as the divisor realising the class of a plane curve of degree $k-2$ and $D_{1}$ as a line in $\mathbb{C}\mathbb{P}^{2}$. ###### Theorem 1.12. ([Ore21, Theorem 3.2]) Let $X$ be a smooth real algebraic surface, let $C$ be a smooth irreducible real separating curve on $X$ and $D$ be a real divisor on $X$ as in Notation 1.10. Let us fix a complex orientation on $\mathbb{R}C$ and an orientation $\mathcal{O}$ on $\mathbb{R}X\setminus(\mathbb{R}C\cup\mathbb{R}D_{1})$ which changes each time we cross $\mathbb{R}C\cup\mathbb{R}D_{1}$ at its smooth points. The latter orientation induces a boundary orientation on $\mathbb{R}C\setminus(\mathbb{R}C\cap D_{1})$. Let $f:C\rightarrow\mathbb{C}\mathbb{P}^{1}$ be a separating morphism. Then it is impossible that, for some $p\in\mathbb{R}\mathbb{P}^{1}$, the set $f^{-1}(p)\setminus supp(D)$ is non-empty and the two orientations coincide at each point of the set. In [Ore21], Orevkov shows powerful applications of Theorem 1.12, such as[Ore21, Theorem 1.1] and the construction of complex schemes (Definition 2.7) in the real projective plane which are realisable by real pseudoholomorphic plane curves of odd degree and not realisable by real algebraic plane curves of same degree. Moreover, other applications of [Ore21, Theorem 3.2] are presented via examples. Therefore, the possible directions of use of the content of [Ore21, Theorem 3.2] are still wide open. Now, let us present the main result of this paper. We refer the reader to Notation 1.10. Unless differently stated, we choose $\beta\in H_{2}(X;\mathbb{Z})$ in order to maximise $\frac{D_{0}^{2}-D_{0}K_{X}}{2}\in\mathbb{Z}_{>0}$. ###### Theorem 1.13. Let $X$ be a smooth real algebraic surface with holomorphic Euler characteristic of the trivial bundle $\chi(0)\geq 1$ and $C\subset X$ a non- singular real algebraic curve of class $d$ and of type I. Denote by $l$ the number of connected components of $\mathbb{R}C$. Assume one of the following: 1. (1) $D_{1}$ can be chosen empty and $2\lceil\frac{l}{2}\rceil\leq\frac{d^{2}-k_{X}^{2}}{4}$; 2. (2) $D_{1}$ cannot be chosen empty and $2\lceil\frac{l-m-\theta}{2}\rceil\leq D_{0}^{2}-D_{0}K_{X}-2\theta$, where $\theta+m$ is the number of connected components of $\mathbb{R}C$ which have non-empty intersection with $\mathbb{R}D_{1}$, and $m=min(\frac{\beta^{2}-\beta K_{X}}{2},l)$; Then there are no separating morphisms $f:C\rightarrow\mathbb{C}\mathbb{P}^{1}$ of degree $l$. ###### Remark 1.14. For example, Theorem 1.13 holds for separating real curves in any smooth rational surface $X$, because the holomorphic Euler characteristic of the trivial bundle equals one. In particular, Theorem 1.13 directly implies Proposition 1.2. In fact, for plane curves of odd degree $2s+1$, the curve $D_{1}$ can be chosen empty and $2\lceil\frac{l}{2}\rceil\leq\frac{(2s+1)^{2}-9}{4}=s^{2}+s-2$. On the other hand, for plane curves of even degree $2s$, the curve $D_{1}$ can be chosen to be a real line so that $m+\theta\leq s$ with $m=2$, and $D_{0}=(s-2)[D_{1}]\in H_{2}(\mathbb{C}\mathbb{P}^{2};\mathbb{Z})$. Hence $2\Bigl{\lceil}\frac{l-s}{2}\Bigl{\rceil}\leq l-2-\theta\leq(s-2)^{2}+3(s-2)-2\theta$ and one gets $l+\varepsilon\leq s^{2}-s-\theta$, where $\varepsilon=1$ if $l-s$ is odd and $0$ otherwise. At worst, the number $\theta$ equals $s-2$ and we obtain the bound given in the hypothesis of Proposition 1.2. ### 1.5. Acknowledgements I would like to thank Mario Kummer and Kris Shaw for useful discussions. A special thanks to Stepan Orevkov for addressing my attention to [Vir86, (4.8)]. Moreover I would like to thank Erwan Brugallé and Hannah Markwig for suggestions and remarks on a preliminary version of the paper. Thanks to Athene Grant. ### Organisation of the paper We start presenting some tools and terminology in Section 2. Then, in Section 3.1, we give a proof of Theorem 1.13. Finally, in Section 3.2, we present examples and applications of Theorem 1.13, and we prove Proposition 1.8. ## 2\. Preliminaries Let us consider non-singular real algebraic curves embedded in the complex projective plane. The real part of real plane curves is homeomorphic to a disjoint union of $S^{1}$ embedded in $\mathbb{R}\mathbb{P}^{2}$. Each $S^{1}$ can be embedded in $\mathbb{R}\mathbb{P}^{2}$ in two different ways: if it realises the trivial-class in $H_{1}(\mathbb{R}\mathbb{P}^{2};\mathbb{Z}/2\mathbb{Z})$, it is called oval, otherwise it is called pseudo-line and denoted by $J$. If a non-singular real plane curve has even degree then its real part consists of ovals only; otherwise of exactly one pseudo-line and ovals. An oval in $\mathbb{R}P^{2}$ separates two disjoint non-homeomorphic connected components: the connected component homeomorphic to a disk is called interior of the oval; the other one is called exterior of the oval. For each pair of ovals, if one is in the interior of the other we speak about an injective pair, otherwise a non-injective pair. For a separating real plane curve, equipped with one of its two complex orientations, an injective pair of ovals is said to be positive if the complex orientations of the ovals are induced by an orientation of the annulus bounded by the ovals; otherwise negative. An oval $\mathcal{O}$ of an odd degree separating curve is said positive if $[\mathcal{O}]=-2[J]$ in $H_{1}(M)$, where $M$ is the closure of the non- orientable component of $\mathbb{R}\mathbb{P}^{2}\setminus\mathcal{O}$; and negative otherwise. ### 2.1. Rokhlin’s complex orientation formula For any separating plane curve of even degree $2s$ and with $l$ real connected components, Rokhlin in [Rok74] proved that (2.1) $2(\Pi_{+}-\Pi_{-})=l-s^{2},$ where $\Pi_{+}$ and $\Pi_{-}$ denote the number of respectively positive and negative injective pairs of ovals of the curve. Later, for separating plane curves of odd degree $2s+1$ and with $l$ real connected components, Mishachev [Mis75] proved that (2.2) $\Lambda_{+}-\Lambda_{-}+2(\Pi_{+}-\Pi_{-})=l-1-s(s+1),$ where $\Lambda_{+}$ and $\Lambda_{-}$ denote respectively positive and negative ovals of the curve. In order to give examples of separating real curves respecting the hypothesis of Proposition 1.2 and to prove Proposition 1.8, we need the following result. ###### Proposition 2.3. [Vir86, Section (4.8)] Let $A_{1}$ and $A_{2}$ be separating real plane curves of degree $d_{1}$ and $d_{2}$, transversal to each other and intersecting in $r$ real points. Let $B$ be a real curve of degree $d=d_{1}+d_{2}$ obtained from $A_{1}\cup A_{2}$ by a small perturbation such that some complex orientations of $A_{1}$ and $A_{2}$ determine an orientation $\mathfrak{O}$ of $\mathbb{R}B$. Then $B$ is of type I if and only if the pair $(\mathbb{R}B,\mathfrak{O})$ respects Rokhlin’s complex orientation formula. ### 2.2. Encoding real schemes In this section, it is introduced some notation and definitions used Section 3.2. ###### Definition 2.4. Let $B_{1},\dots,B_{n}$ be any collection of disjoint ovals (and of a pseudo- line) in $\mathbb{R}\mathbb{P}^{2}$. We call topological type the arrangement $\mathcal{S}$ realised by the pair $(\mathbb{R}\mathbb{P}^{2},\bigsqcup_{i=1}^{n}B_{i})$. ###### Notation 2.5 (Encoding topological types). Let us consider collections of disjoint ovals in $\mathbb{R}\mathbb{P}^{2}$. An empty union of ovals is denoted by $0$. We say that a union of $l$ ovals realises $l$ if there are no injective pairs. The symbol $\langle\mathcal{S}\rangle$ denotes the disjoint union of a non-empty collection of ovals realising $\mathcal{S}$, and an oval forming an injective pair with each oval of the collection. A collection of $h$ ovals in $\mathbb{R}\mathbb{P}^{2}$ is called a nest of depth $h$ if any two ovals of the collection form an injective pair. The disjoint union of any two collections of ovals, realising respectively $\mathcal{S}^{\prime}$ and $\mathcal{S}^{\prime\prime}$ in $\mathbb{R}\mathbb{P}^{2}$, is denoted by $\mathcal{S}^{\prime}\sqcup\mathcal{S}^{\prime\prime}$ if none of the ovals of one collection forms an injective pair with the ovals of the other one and they are both non-empty collections. A disjoint union of the form $\mathcal{S}^{\prime}\sqcup 0$ is still denoted $\mathcal{S}^{\prime}$. Moreover, as previously mentioned, a pseudo-line in $\mathbb{R}\mathbb{P}^{2}$ is denoted by $J$. ###### Definition 2.6. A topological type $\mathcal{S}$, up to homeomorphism of $\mathbb{R}\mathbb{P}^{2}$, is called real scheme. Let $A\subset\mathbb{C}\mathbb{P}^{2}$ be a non-singular real curve. We say that $A$ has real scheme $\mathcal{S}$ if the pair $(\mathbb{R}\mathbb{P}^{2},\mathbb{R}A)$ realises $\mathcal{S}$. ###### Definition 2.7. A real scheme $\mathcal{S}$ in $\mathbb{R}\mathbb{P}^{2}$, where each connected component is endowed with an orientation is called complex scheme. Let $A\subset\mathbb{C}\mathbb{P}^{2}$ be a non-singular real separating curve. We say that $A$ has complex scheme $\mathcal{S}$ if the pair $(\mathbb{R}\mathbb{P}^{2},\mathbb{R}A)$ realises $\mathcal{S}$ as real scheme and one of the two complex orientations of $\mathbb{R}A$ agrees with that of $\mathcal{S}$. ## 3\. Theorem 1.13 and applications Section 3.1 is uniquely devoted to the proof of Theorem 1.13. Then Section 3.2 presents the proof of Proposition 1.8 and some examples. ### 3.1. Proof of Theorem 1.13 ###### Proof of Theorem 1.13. Denote by $C_{i}$ the connected components of $\mathbb{R}C$ where $i=1,\dots l$. Let us consider a real divisor $D$ belonging to the linear system $|C+K_{X}|$. For the sake of contradiction, assume that there exists a separating morphism $f:C\rightarrow\mathbb{P}^{1}$ of degree $l$. Then, for all $p\in\mathbb{R}\mathbb{P}^{1}$, the set $f^{-1}(p)=\\{p_{1},\dots,p_{l}\\}$ is a collection of $l$ real points such that every $C_{i}$ contains exactly one $p_{i}$. Let us choose a curve $D_{1}$ and a curve of class $D_{0}$; see Notation 1.10. For our purposes, the choice of $\beta\in H_{2}(X;\mathbb{Z})$, and consequently of $D_{1}$, is done in order to maximise $\frac{D_{0}^{2}-D_{0}K_{X}}{2}\in\mathbb{Z}_{>0}$; where, thanks to Riemann-Roch theorem and the hypothesis that the holomorphic Euler characteristic of the trivial bundle of $X$ is bigger or equal to one, the positive integer $\frac{D_{0}^{2}-D_{0}K_{X}}{2}$ is always less or equal than the dimension of the linear system of curves in $X$ of class $D_{0}$. Therefore, fixed a configuration of $\frac{D_{0}^{2}-D_{0}K_{X}}{2}$ points, there always exists at least one curve of class $D_{0}$ passing through the configuration. Once $D_{1}$ is chosen, let us fix an orientation on $\mathbb{R}X\setminus(\mathbb{R}C\cup\mathbb{R}D_{1})$ which changes each time we cross $\mathbb{R}C\cup\mathbb{R}D_{1}$ at its smooth points; see [Ore21, Proof of Theorem 3.2] for details. The latter orientation induces a boundary orientation $\mathfrak{O}$ on $\mathbb{R}C$. Fix some $p$ in $\mathbb{R}\mathbb{P}^{1}$. Before treating cases $(1)$ and $(2)$ and prove the statement. Let us give an idea of the strategy. In order to get a contradiction, we use Theorem 1.12. Namely, define $Y_{p}$ (resp. $NY_{p}$) as the set of points in $f^{-1}(p)\setminus supp(D)$ on which the complex orientation and the orientation $\mathfrak{O}$ agree (resp. do not agree). Then, Theorem 1.12 tell us that if the set $f^{-1}(p)\setminus supp(D)$ is non-empty, one has that $Y_{p}$ and $NY_{p}$ are both non-empty. First, let us treat case $(1)$ in which $D_{1}$ can be chosen empty. Fix one of the two complex orientations on $\mathbb{R}C$. Then, remark that either the number $V^{1}_{p}$ of $p_{i}$’s where the complex orientation and the orientation $\mathfrak{O}$ agree, or the number $V^{2}_{p}$ of the $p_{i}$’s where the two orientations do not agree, does not exceed $\lceil\frac{l}{2}\rceil$. Given a collection $\mathcal{P}$ of $\frac{d^{2}-K_{X}^{2}}{8}$ real points in $X$, there exists a real curve $A$ of class $D_{0}$ passing through $\mathcal{P}$. By hypothesis $2\lceil\frac{l}{2}\rceil\leq\frac{d^{2}-K_{X}^{2}}{4}$, therefore it is possible to choose such a collection $\mathcal{P}$ so that it contains at least $\lceil\frac{l}{2}\rceil$ points of $f^{-1}(p)$ belonging to $V^{j}_{p}$, for an opportune $j\in\\{1,2\\}$. It follows that the set $f^{-1}(p)\setminus supp(D)$ is non-empty and either $NY_{p}$ or $Y_{p}$ is empty, which contradicts Theorem 1.12. Now, let us consider case $(2)$ in which $D_{1}$ is not empty. Let us pick $D_{1}$ as the real curve passing through a given collection $\tilde{\mathcal{P}}$ of $\frac{\beta^{2}-\beta K_{X}}{2}$ real points contained in $m=min(\frac{\beta^{2}-\beta K_{X}}{2},l)$ connected components $C_{j_{1}},\dots,C_{j_{m}}$. Moreover, choose the collection $\tilde{\mathcal{P}}$ such that it contains the points $p_{j_{1}},\dots,p_{j_{m}}$ of $f^{-1}(p)$. Now, we proceed as in the previous case with the difference that one must take $D_{1}$ into account when looking at the orientation $\mathfrak{O}$. Let $C_{j_{m+1}},\dots,C_{j_{m+\theta}}$ be the connected components of $\mathbb{R}C\setminus\bigsqcup_{i=j_{1}}^{j_{m}}C_{i}$ intersecting $\mathbb{R}D_{1}$. Each connected components $C_{j_{m+i}}$, with $i=1,\dots,\theta$, may contain, regardless of the chosen complex orientation on $\mathbb{R}C$, a point of $f^{-1}(p)$ where the orientation $\mathfrak{O}$ and the complex orientation do not coincide. Fix one of the two complex orientations on $\mathbb{R}C$. Then, remark that either the number $V^{1}_{p}$ of $p_{i}$’s contained in $\mathbb{R}C\setminus\bigsqcup_{i=1}^{m+\theta}C_{j_{i}}$, where the complex orientation and the orientation $\mathfrak{O}$ agree, or the number $V^{2}_{p}$ of the $p_{i}$’s in $\mathbb{R}C\setminus\bigsqcup_{i=1}^{m+\theta}C_{j_{i}}$, where the two orientations do not agree, does not exceed $\lceil\frac{l-m-\theta}{2}\rceil$. Therefore, we pick a collection $\mathcal{P}$ of $\frac{D_{0}^{2}-D_{0}K_{X}}{2}$ real points such that $\theta$ of those are the points of $f^{-1}(p)$ contained in the $C_{j_{m+i}}$’s, where $i=1,\dots,\theta$; and the others include the points of $V^{j}_{p}$, for an opportune $j\in\\{1,2\\}$. Once again, the set $f^{-1}(p)\setminus supp(D)$ is non-empty and either $NY_{p}$ or $Y_{p}$ is empty, which contradicts Theorem 1.12. ∎ ### 3.2. Examples and Applications Let us consider real separating curves in the complex projective plane and focus on [KS20b, Question 1]. A first known example [KS20b, Example 2.8] concerns real separating plane curves of degree $2s$ or $2s+1$ having a a nest of maximal depth $s$. Such curves admit a totally real pencil of lines. Some more examples can be found in [Tou13], where for some pairs $(A,\mathcal{S})$, where $A$ is a real separating plane sextic curve with $9$ real connected components and $\mathcal{S}$ its real scheme in $\mathbb{R}\mathbb{P}^{2}$, the minimal value for $k$ of Theorem 1.4, is shown to be equal to $3$. $q$ Figure 1. $(\mathbb{R}\mathbb{P}^{2},\mathbb{R}A,\mathbb{R}L)$ of Example 3.1. Double arrows denote $\mathfrak{O}$, simple arrows the fixed complex orientation of $\mathbb{R}A$ and $\bullet$ the points in $f^{-1}(p)$. ###### Example 3.1 (Application of Theorem 1.12). Let $A$ be a non-singular plane quintic of type I realising the real scheme $J\quad\sqcup\quad 4$. We use Theorem 1.12 in order to show that there exists no separating morphism $f:A\rightarrow\mathbb{C}\mathbb{P}^{1}$ of degree $5$. For the sake of contradiction assume that such a $f$ exists. By Rokhlin’s complex orientation formula (2.2), among the ovals of $\mathbb{R}A$, three are negative and one is positive. Fix $p\in\mathbb{R}\mathbb{P}^{1}$. Denote $q$ the point lying on the positive oval of $\mathbb{R}A$ and contained in $f^{-1}(p)$. Let us consider Notation 1.10. Choose $D_{1}$ to be empty. Therefore $D_{0}$ realises the class of a line. Fix an orientation $\mathfrak{O}$ that changes every time time we cross $\mathbb{R}A$. Up to choose one of the two complex orientations of $\mathbb{R}A$, one gets that the fixed complex orientation and $\mathfrak{O}$ agree on the pseudo-line and the three negative ovals of $\mathbb{R}A$. Hence, if we pick any line $L$ passing through the point $q$, we get a contradiction thanks to Theorem 1.12; see Fig. 1, A key information used to deal with the examples above is the fact that their real schemes are given. In this paper, on the other hand, we focus on real separating curves $C$ about which only the number of real connected components is set. We manage to obstruct the existence of certain totally real pencils of curves with respect to $C$, but we do not know yet how to compute the minimal value for $k$ of Theorem 1.4. The rest of the section is dedicated to the proof of Proposition 1.8. We start with a lemma, where a recursive construction of curves produces odd degree plane curves respecting the hypothesis of Proposition 1.8. ###### Lemma 3.2. For all positive integer $s$ and $\varepsilon\in\\{0,1\\}$, there exist non- singular real plane projective curves of degree $2s+\varepsilon$ of type I and with $s^{2}+\varepsilon$ real connected components. ###### Proof. For small positive integers $d\leq 4$, real maximal, therefore separating, plane curves of degree $d$ have the required properties of the statement. Therefore, we present a recursive construction of real plane curves $B_{d}$ with the wanted features via a variation of Harnack construction method [Har76] combined with Proposition 2.3, which allows us to check, at every step of the construction, that the realised curves are of type I. Set $l_{d}$ as the number of connected components of $\mathbb{R}B_{d}$. Moreover if $d=2s+1$, denote $\Lambda_{\pm}$ with $\Lambda^{s}_{\pm}$ for $B_{d}$. In the following, whenever we consider a plane curve $A$, we denote its polynomial by $a(x,y,z)$. Let us fix a real line $L$. 1. (i) Step $0$ induction: In order to construct a curve with the wanted properties for any given degree, we start from the following. One can construct a non- singular real maximal cubic $B_{3}$ whose pseudo-line $J_{3}$ intersects $\mathbb{R}L$ in $3$ distinct points; see Fig 2.1. Equip $\mathbb{R}B_{3}$ with one of its two complex orientations and $\mathbb{R}L$ with the complex orientation opposite to that of $J_{3}$. Let $\mathcal{H}_{1},\mathcal{H}_{2}$ and $\mathcal{H}_{3}$ be the connected components of $\mathbb{R}L\setminus\mathbb{R}B_{3}$. Up to fix $\mathcal{H}_{1}$, label the $\mathcal{H}_{i}$’s following the order given by the fixed complex orientation on $J_{3}$. Pick four real lines $L_{1},\dots,L_{4}$ intersecting pairwise respectively $\mathcal{H}_{1},\mathcal{H}_{2}$. Now, we apply Brusotti’s perturbation method [Bru21] to $B_{3}\cup L$ in a compatible way with the fixed complex orientations of $B_{3}$ and $L$. Namely, we take a small perturbation of $L\cup B_{3}$ replacing $b_{3}(x,y,z)l(x,y,z)$ with $b_{4}(x,y,z):=b_{3}(x,y,z)l(x,y,z)+\varepsilon l_{1}(x,y,z)\dots,l_{4}(x,y,z)$, where $\varepsilon\not=0$ is a sufficient small real number. Up to a choice of the sign of $\varepsilon$, the real curve, defined by $b_{4}(x,y,z)=0$, is a non-singular real curve $B_{4}$ of degree $4$. In particular $B_{4}$ is maximal, therefore of type I, equipped with a complex orientation $\mathfrak{O}_{4}$ coming form the perturbation; see Fig. 2.2. Take for $L$ the opposite complex orientation to the previously chosen one. Set $\mathcal{H}_{1},\dots,\mathcal{H}_{4}$ for the connected components of $\mathbb{R}L\setminus\mathbb{R}B_{4}$. Label the $\mathcal{H}_{i}$’s following the order given by the fixed complex orientation on $\mathbb{R}L$. Without loss of generality, we can say that $\mathcal{H}_{1}$ and $\mathcal{H}_{3}$ are in the interior of the ovals of $\mathbb{R}B_{4}$ intersecting $\mathbb{R}L$. Pick five real lines $L_{1},\dots,L_{5}$ such that $L_{1}$ intersects $\mathcal{H}_{2}$ and the others $\mathcal{H}_{4}$. Now, we apply Brusotti’s perturbation method to $L\cup B_{4}$ with respect to $\bigcup_{i=1}^{5}L_{i}$, in a compatible way with the complex orientation $\mathfrak{O}_{4}$ and the fixed complex orientation of $L$, and obtain a non- singular real curve $B_{5}$ of degree $5$ whose real part is equipped with an orientation $\mathfrak{O}_{5}$ coming from the perturbation. In order to assure that $\mathfrak{O}_{5}$ is a complex orientation of $B_{5}$ and, hence $B_{5}$ is of type I, it is enough to check that $B_{5}$ respects (2.2); see Proposition 2.3. Because of the absence of nests in $\mathbb{R}B_{5}$ and the number of positive and negative ovals of $\mathbb{R}B_{5}$ differ by $-2$, it follows that $(B_{5},\mathfrak{O}_{5})$ respect Rokhlin’s complex orientation formula. Fig. 2.1: $(\mathbb{R}\mathbb{P}^{2},\mathbb{R}B_{3},\mathbb{R}L)$Fig. 2.3: $(\mathbb{R}\mathbb{P}^{2},\mathbb{R}B_{5},\mathbb{R}L)$Fig. 2.2: $(\mathbb{R}\mathbb{P}^{2},\mathbb{R}B_{4},\mathbb{R}L)$Fig. 2.4: $(\mathbb{R}\mathbb{P}^{2},\mathbb{R}B_{6},\mathbb{R}L)$ Figure 2. $\mathbb{R}L$ in red and $\mathbb{R}B_{d}$ in black, where, from top to bottom and from left to right, $d=3,4,5,6$. 2. (ii) Inductive Step: Assume that there exists a non-singular real plane curve $B_{2s-1}$ of degree $2s-1$ and of type I. Equip $\mathbb{R}B_{2s-1}$ with one of its two complex orientations and $\mathbb{R}L$ with the complex orientation opposite to that of $J_{2s-1}$. Moreover, assume that * • the pseudo-line $J_{2s-1}$ intersects $\mathbb{R}L$ in $2s-1$ distinct points such that, if one labels the intersections points $2s-1,\dots,1$ with respect to the complex orientation on $J_{2s-1}$, the order of the labeled points agrees with the complex orientation of $\mathbb{R}L$; * • the number $l_{2s-1}$ equals $(s-1)^{2}+1$; * • the curve $B_{2s-1}$ has no nests and $\Lambda^{(s-1)}_{+}=\frac{(s-1)(s-2)}{2}\quad\text{and}\quad\Lambda^{(s-1)}_{-}=\frac{s(s-1)}{2}.$ Let $\mathcal{H}_{1},\dots,\mathcal{H}_{2s-1}$ be the connected components of $\mathbb{R}L\setminus\mathbb{R}B_{2s-1}$. Up to fix $\mathcal{H}_{1}$, label the $\mathcal{H}_{i}$’s following the order given by the fixed complex orientation on $J_{2s-1}$. Pick $2s$ real lines $L_{1},\dots,L_{2s}$ intersecting pairwise $\mathcal{H}_{j},\mathcal{H}_{j+1},\dots,\mathcal{H}_{j+s}$, for some $j\in\\{1,\dots,2s+1\\}$. Now, we apply Brusotti’s perturbation method to $L\cup B_{2s-1}$ with respect to $\bigcup_{i=1}^{2s}L_{i}$, in a compatible way with the fixed complex orientations of $B_{2s-1}$ and $L$, and obtain a degree $2s$ non-singular real algebraic curve $B_{2s}$, equipped with an orientation $\mathfrak{O}_{2s}$. Let us check that (2.1) applies to the pair $(B_{2s},\mathfrak{O}_{2s})$. The curve $B_{2s}$ has no nests and $l_{2s}=l_{2s-1}-1+2s-1=s^{2}$. Hence, Proposition 2.3 implies that $B_{2s}$ is of type I. In addition, remark that $s$ ovals of $\mathbb{R}B_{2s}$ intersect $\mathbb{R}L$. Take for $L$ the opposite complex orientation to the previously chosen one. Set $\mathcal{H}_{1},\dots,\mathcal{H}_{2s}$ for the connected components of $\mathbb{R}L\setminus\mathbb{R}B_{2s}$. Label the $\mathcal{H}_{i}$’s following the order given by the fixed complex orientation on $\mathbb{R}L$. Without loss of generality, we can say that $\mathcal{H}_{2h+1}$, for $0\leq h\leq s$ are in the interior of the ovals of $\mathbb{R}B_{2s}$ intersecting $\mathbb{R}L$. Pick $2s+1$ real lines $L_{1},\dots,L_{2s+1}$ such that $L_{h}$ intersects $\mathcal{H}_{2h}$, for all $1\leq h\leq s-1$ and all others lines intersect $\mathcal{H}_{2s}$. Once again, apply Brusotti’s perturbation method to $L\cup B_{2s}$ with respect to $\bigcup_{i=1}^{2s+1}L_{i}$, in a compatible way with the fixed complex orientations of $B_{2s}$ and $L$. We get a degree $2s+1$ non-singular real curve $B_{2s+1}$ whose real part is equipped with an orientation $\mathfrak{O}_{2s+1}$ coming from the perturbation, and $l_{2s+1}=l_{2s}+1=s^{2}+1$. Let us check that the pair $(B_{2s+1},\mathfrak{O}_{2s+1})$ respects (2.2). First, observe that $\Lambda^{s}_{+}=\Lambda^{(s-1)}_{+}+s-1\quad\text{and}\quad\Lambda^{s}_{-}=\Lambda^{(s-1)}_{-}+s.$ It follows that, because of the absence of nests in $\mathbb{R}B_{2s+1}$ and the number of positive and negative ovals of $\mathbb{R}B_{2s+1}$ differ by $-s$, the pair $(B_{2s+1},\mathfrak{O}_{2s+1})$ satisfies Rokhlin’s complex orientation formula. In conclusion, one obtains a real curve of degree $2s+1$ which is separating and with $s^{2}+1$ real connected components. ∎ ###### Proof of Proposition 1.8. The non-singular real algebraic plane curves $B_{2s+1}$ of odd degree $2s+1\geq 5$ and of type I, constructed in Lemma 3.2, satisfy the hypothesis of Corollary 1.6. Therefore $B_{2s+1}$ admit no totally real pencils of curves of degree lower or equal to $\frac{s^{2}+1}{2s+1}$. ∎ ###### Remark 3.3. Notice that the non-singular real algebraic plane curves $B_{2s}$ of even degree $2s\geq 6$ and of type I constructed in Lemma 3.2, does not satisfy hypothesis of Corollary 1.6. We do not know yet whether there is a way to apply [Ore21, Theorem 3.2] to this particular case in order to show that $B_{2s}$ admit no totally real pencils of curves of degree lower or equal to $\frac{s}{2}$. Mimic the constructions adopted in the proof of Lemma 3.2, one may also construct an explicit family of real separating plane curves of increasing even degree, which respect the hypothesis of Proposition 1.2 or even those in Remark 1.14. Instead of presenting to the reader such a construction, that would be a modified version of proof of Lemma 3.2, we conclude with an example of separating real curves in a different ambient surface, respecting hypothesis of Theorem 1.13 and, therefore, non admitting separating morphisms of degree equal to the number of their real connected components. ###### Example 3.4. Let $Q$ be $\mathbb{C}P^{1}\times\mathbb{C}P^{1}$, equipped with the anti- holomorphic involution $\sigma:Q\rightarrow Q$ sending $(x,y)$ to $(\overline{y},\overline{x})$, where $x=[x_{0}:x_{1}]$ and $y=[y_{0}:y_{1}]$ are in $\mathbb{C}P^{1}$ and $\overline{x}=[\overline{x_{0}}:\overline{x_{1}}]$ and $\overline{y}=[\overline{y_{0}}:\overline{y_{1}}]$ are respectively the images of $x$ and $y$ via the standard complex conjugation on $\mathbb{C}P^{1}$. The real part of $Q$ is homeomorphic to a $2$-sphere $S^{2}$ and $Q$ is called quadric ellipsoid. Moreover, the Euler characteristic of the trivial bundle of $Q$ equals one; see Remark 1.14. A non-singular real algebraic curve A on $Q$ is defined by a bi-homogeneous polynomial of bidegree $(d,d)$ $P(x,y)=\sum\limits_{0\leq i,j\leq d}a_{i,j}x_{0}^{i}x_{1}^{d-i}y_{0}^{j}y_{1}^{d-j}=0,$ where $d$ is a positive integer and the coefficients satisfy $a_{i,j}=\overline{a_{j,i}}$. Real schemes realised by non-singular real algebraic curves of bidegree $(d,d)$ on $Q$ are completely classified for $d\leq 5$; see [GS80], [Mik94] and [Man21]. Moreover such classifications distinguish the cases in which a given topological type may or may not be realised by a real algebraic curve of type I. For $d\leq 5$, we report in Table 1 a list of real schemes realised by type I real algebraic curves of bidegree $(d,d)$, with $l$ real connected components, where $l$ undergoes the hypothesis of Theorem 1.13. Hence, the curves do not admit separating morphisms of degree $l$. In order to understand Table 1, let us recall some notation. Let $A$ be a non-singular bidegree $(d,d)$ real curve on $Q$. The connected components of $A$ are called ovals. An oval in $S^{2}$ bounds two disks; therefore, on $S^{2}$ interior and exterior of an oval are not well defined. It follows that the encoding of real schemes on $S^{2}$ is not well defined either and it depends on the choice of a point $p$ on $S^{2}\setminus\mathbb{R}A$. Let us take $S^{2}$ deprived of $p$, which is homeomorphic to $\mathbb{R}^{2}$ and let us call oval any circle embedded in $\mathbb{R}^{2}$. Analogously to the case of $\mathbb{R}P^{2}$, in $\mathbb{R}^{2}$ one define interior and exterior of an oval and (non-)injective pairs for each pair of ovals. Up to choose a point, we can now adopt the same notation introduced in Section 2.2. d | Real schemes realised by type I curves | $l\leq$ | $r(D_{0},D_{1})$ | $D_{0}$ | $D_{1}$ | $m$ | $\theta$ ---|---|---|---|---|---|---|--- $3$ | $\langle\langle 1\rangle\rangle$ | $3\leq$ | 3 | $-$ | $(1,1)$ | 3 | 0 $4$ | $\langle\langle\langle 1\rangle\rangle\rangle$ | $4\leq$ | 6 | $(1,1)$ | $-$ | $-$ | $-$ $4$ | $4\quad\sqcup\quad\langle 1\rangle$ | $6\leq$ | 6 | $(1,1)$ | $-$ | $-$ | $-$ $4$ | $3\quad\sqcup\quad\langle 2\rangle$ | $6\leq$ | 6 | $(1,1)$ | $-$ | $-$ | $-$ $5$ | $\langle\langle\langle\langle 1\rangle\rangle\rangle\rangle$ | $5\leq$ | 7 | $(1,1)$ | $(1,1)$ | 3 | 2 $5$ | $\alpha\quad\sqcup\quad\langle\beta\rangle\quad\sqcup\quad\langle\gamma\rangle,$ for all $\alpha,\beta,\gamma$ such | $7\leq$ | 7 | $(1,1)$ | $(1,1)$ | 3 | 2 | that $\alpha=0\pmod{2}$ and $\alpha+\beta+\gamma=5$ | | | | | | Table 1. For all $d\leq 5$, this is a list of real schemes realised by type I real algebraic curves of bidegree $(d,d)$, with $l$ real connected components, where $l$ undergoes the hypothesis of Theorem 1.13. Hence, the curves do not admit separating morphisms of degree $l$. The symbol $r(D_{0},D_{1})$ represents the bound for the number $l$, depending on choices of $D_{0},D_{1}$, for which the hypothesis of Theorem 1.13 are satisfied and such that one maximises the number $\frac{D_{0}^{2}-D_{0}K_{S}}{2}\in\mathbb{Z}_{>0}$. ###### Remark 3.5. In the context of Example 3.4, one can construct recursively, in a way similar to the proof of Lemma 3.2, for all $d\geq 6$, bidegree $(d,d)$ separating real algebraic curves on the quadric ellipsoid not admitting separating morphism of degree equal to the number of their real connected components. In fact, variations of Harnack’s construction method, Brusotti’s perturbation method and Rokhlin’s complex orientation formula [Zvo83, Ore07] are also available on the quadric ellipsoid. ## References * [Ahl50] Lars L. Ahlfors. Open Riemann surfaces and extremal problems on compact subregions. Comment. Math. Helv., 24:100–134, 1950. * [Bru21] L. Brusotti. Sulla ”piccola variazione” di una curva piana algebrica reale. Rend. della R. Acc. Nazionale dei Lincei, (5), 30, 375-379, 268(1):22–26, 1921. * [CH13] M. Coppens and J. Huisman. Pencils on real curves. Math. Nachr., 286(8-9):302–320, 2013. * [Cop13] M. Coppens. The separating gonality of a separating real curve. Monatsh. Math., 170(1):1–10, 2013. * [Gab06] A. Gabard. Sur la représentation conforme des surfaces de Riemann à bord et une caractérisation des courbes séparantes. Comment. Math. Helv., 81(4):945–964, 2006. * [GS80] D. A. Gudkov and E. I. Shustin. Classification of non-singular 8th order curves on ellipsoid. In ”Methods of Qualitative Theory”. Gorky university Press, pages 104–107 (in Russian), 1980. * [Har76] A. Harnack. Über vieltheiligkeit der ebenen algebraischen Curven. In Math. Ann., 10: 189-199,, volume 1060 of Lecture Notes in Math., pages 187–200. 1876. * [Hui01] J. Huisman. On the geometry of algebraic curves having many real components. Rev. Mat. Complut., 14(1):83–92, 2001. * [Kle73] F. Klein. Über Flächen dritter Ordnung. Math. Ann. 6, pag. 551–581, 1873. * [KS20a] M. Kummer and E. Shamovich. Real fibered morphisms and Ulrich sheaves. J. Algebraic Geom., 29(1):167 – 198, 2020. * [KS20b] M. Kummer and K. Shaw. The separating semigroup of a real curve. Annales de la Faculté des sciences de Toulouse: Mathématiques, 29(1):79–96, 2020. * [KTM22] M. Kummer, C. Le Texier, and M. Manzaroli. Real-Fibered Morphisms of del Pezzo Surfaces and Conic Bundles. Discrete & Computational Geometry, https://doi.org/10.1007/s00454-022-00427-3, 2022. * [Man21] M. Manzaroli. Real algebraic curves of bidegree (5,5) on the quadric ellipsoid. St. Petersburg Math. J., 32(2):279–306, 2021. * [Mik94] G. Mikhalkin. Congruences for real algebraic curves on an ellipsoid. In Topology of manifolds and varieties, volume 18 of Adv. Soviet Math., pages 223–233. Amer. Math. Soc., Providence, RI, 1994. * [Mis75] N. M. Mishachev. Complex orientations of plane M-curves of odd degree. Funktsional Anal. i Prilozhen., 9:77–78, 1975. * [Ore07] S. Y. Orevkov. Positions of an $M$-quintic with respect to a conic that maximally intersect the odd branch of the quintic. Algebra i Analiz, 19(4):174–242, 2007. * [Ore18] S. Y. Orevkov. On the hyperbolicity locus of a real curve. Funk. anal. i ego prilozhenia, 52(2):86–89 (Russian), 2018. * [Ore21] S. Y. Orevkov. Algebraically unrealizable complex orientations of plane real pseudoholomorphic curves. Geom. and Funct. Anal., (31):930–947, 2021. * [Rok72] V. A. Rokhlin. Congruences modulo 16 in Hilbert’s sixteenth problem. Funktsional. Anal. i Prilozhen. 6, 58–64 (Russian), 1972. * [Rok74] V. A. Rokhlin. Complex orientations of real algebraic curves. Funktsional. Anal. i Prilozhen. 8, 71–75 (Russian), 1974. * [Tou13] S. Fiedler-Le Touzé. Totally real pencils of cubics with respect to sextics. arXiv:1303.4341, 2013. * [Vir86] O. Y. Viro. Progress in the topology of real algebraic varieties over the last six years. Russian Mathematical Surveys, 41(3):55–82, 1986. * [Zvo83] V. I. Zvonilov. Complex orientations of real algebraic curves with singularities. Dokl. Akad. Nauk SSSR, 268(1):22–26, 1983. Matilde Manzaroli, Universität Tübingen, Germany E-mail address: matilde.manzaroli’at’uni-tuebingen.de
# Camelira: An Arabic Multi-Dialect Morphological Disambiguator Ossama Obeid1, Go Inoue1,2, Nizar Habash1 1Computational Approaches to Modeling Language (CAMeL) Lab New York University Abu Dhabi 2Mohamed bin Zayed University of Artificial Intelligence <EMAIL_ADDRESS> <EMAIL_ADDRESS> ###### Abstract ## 1 Introduction The last two decades have witnessed remarkable progress in Natural Language Processing (NLP) for Arabic and its dialects despite many challenges such as its diglossic nature, morphological complexity, and orthographic ambiguity Darwish et al. (2021). These efforts have led to many practical applications for various NLP tasks including tokenization, part-of-speech (POS) tagging, morphological disambiguation, named entity recognition, dialect identification (DID), and sentiment analysis (Pasha et al., 2014; Abdelali et al., 2016; Obeid et al., 2019; Abdul-Mageed et al., 2020b, inter alia). Tools for core technologies like POS tagging and morphological disambiguation are primary examples of such successful applications, e.g., MADAMIRA Pasha et al. (2014), Farasa Abdelali et al. (2016), UDPipe Straka et al. (2016), and Stanza Qi et al. (2020). However, there are still gaps to be filled in terms of coverage and usability. For example, these systems only support Modern Standard Arabic (MSA) and Egyptian Arabic, but not other widely spoken dialects such as Gulf and Levantine. In addition, these web interfaces only present the top prediction, although the alternative readings could provide valuable information for analyzing the models’ behavior. In contrast, morphological analyzers such as ElixirFM Smrž (2007), CALIMAStar Taji et al. (2018b), CALIMA Egyptian Habash et al. (2012) show all the different readings for a given word out of context but without disambiguated analyses in context. These tools assume that users already know the input DID; however, this is not necessarily the case for second language learners. To address these limitations, we present Camelira,111http://camelira.camel-lab.com,222Camelira is named after CAMeL Tools Obeid et al. (2020), and in homage to MADAMIRA Pasha et al. (2014). a web interface for Arabic multi-dialect morphological disambiguation that covers four major variants of Arabic: MSA, Egyptian, Gulf, and Levantine. Our system takes an input sentence and provides automatically disambiguated readings for each word in context, as well as its alternative out-of-context readings. We also showcase the integration of a state-of-the-art morphological disambiguator Inoue et al. (2022) with the highest performing fine-grained Arabic DID system Salameh et al. (2018) on the MADAR DID shared task Bouamor et al. (2019). Camelira provides an option to automatically choose a dialect- specific disambiguator based on the prediction of the DID component. To the best of our knowledge, our work is the first to demonstrate an integrated web application that leverages both Arabic morphological disambiguation and DID systems. Our contributions are as follows: (a) We present a user-friendly web interface that allows researchers and language learners to explore the detailed linguistic analysis of a given Arabic sentence. (b) We include three major Arabic dialects (Egyptian, Gulf, and Levantine) in addition to MSA, to make our tool more accessible to a wider audience. (c) We integrate DID to automatically select the appropriate disambiguator; a feature that helps users with limited knowledge of Arabic dialects. ## 2 Arabic Linguistic Facts The Arabic language poses a number of challenges for NLP Habash (2010). We highlight three aspects that are most relevant to multi-dialectal morphological modeling: dialectal variations, morphological richness, and orthographic ambiguity. First, Arabic is characterized with diglossia and its large number of dialects Ferguson (1959); Holes (2004). MSA is the shared standard variant used in official contexts, while the dialects are the varieties of daily use. MSA and the dialects vary among themselves in different aspects, such as lexicons, morphology, and syntax. Second, Arabic is a morphologically rich and complex language. It employs a combination of templatic, affixational, and cliticization morphological operations to represent numerous grammatical features such as gender, number, person, case, state, mood, aspect, and voice, in addition to a number of attachable pronominal, preposition, and determiner clitics. Third, Arabic is orthographically highly ambiguous. This is due to its orthographic conventions where diacritical marks are often omitted, leading to a high degree of ambiguity. For example, MSA can have 12 different morphological analyses per word on average Pasha et al. (2014). ## 3 Related Work #### Morphological Analysis and Disambiguation Morphological analysis is the task of producing a complete list of readings (analyses) for a given word out of context. Morphological analysis has a wide range of applications, including treebank annotation Maamouri et al. (2003, 2011, 2009) and improving morphological modeling Habash et al. (2005); Inoue et al. (2017); Zalmout and Habash (2017); Khalifa et al. (2020). Over the past two decades, there have been numerous efforts in building morphological analyzers for Arabic, e.g. BAMA Buckwalter (2002), MAGEAD Habash and Rambow (2006); Altantawy et al. (2010), ALMORGEANA Habash (2007), ElixirFM Smrž (2007), SAMA Graff et al. (2009), CALIMA Egyptian Habash et al. (2012), CALIMA Gulf Khalifa et al. (2017), AlKhalil Morpho Sys Boudlal et al. (2010); Boudchiche et al. (2017) and CALIMAStar Taji et al. (2018b). Among these efforts, ElixirFM333http://quest.ms.mff.cuni.cz/elixir and CALIMAStar444http://calimastar.camel-lab.com/ provide easy-to-use web interfaces, allowing the user to explore all the possible morphological analyses for a given word. In addition to these rule-based approaches, Eskander et al. (2016) used a corpus-based paradigm completion technique Eskander et al. (2013) to develop a morphological analyzer for Levantine; and Khalifa et al. (2020) used the same technique to develop a morphological analyzer for Gulf. Morphological disambiguation is the subsequent process of identifying the correct analysis in context from the list of different analyses produced by a morphological analyzer. Examples of this in Arabic start with MADA Habash et al. (2005) and many following efforts Pasha et al. (2014); Khalifa et al. (2016); Zalmout and Habash (2017, 2020); Khalifa et al. (2020); Inoue et al. (2022), where they rank the analyses based on the predictions of morphological taggers. While these models have achieved significant improvement over time, only MADAMIRA Pasha et al. (2014) offers a web interface555http://madamira.camel-lab.com/ that’s accessible to a general audience. In this work, we present a user-friendly web interface for state-of- the-art morphological disambiguation models to make these recent advances more accessible to a wider audience, such as linguists and language learners. Our interface also provides all the alternative readings of each input word with the associated prediction scores, allowing researchers to investigate the model’s behavior. Figure 1: The Camelira interface with an MSA example sentence celebrating the winning of a racehorse named “Dream.” In this example, the automatically diacritized forms of the words are presented together with their POS. The first word (on the right), which is highlighted, is selected by the user. The two lower boxes show all the possible out-of-context analyses (on the right) and the detailed features and gloss for the top in-context analysis (on the left). Figure 2: The Camelira interface presenting the same example in Figure 1 using the Arabic user interface. #### Dialect Identification Dialect identification (DID) is the task of automatically identifying the language variety of a given text. DID for Arabic and its variants has attracted increasing attention in recent years. A number of shared tasks have been organized, including VarDial Malmasi et al. (2016); Zampieri et al. (2017, 2018), MADAR Bouamor et al. (2019), and NADI Abdul-Mageed et al. (2020a, 2021, 2022), along with continuous efforts in dataset creation (Zaidan and Callison-Burch, 2011; Mubarak and Darwish, 2014; Zaghouani and Charfi, 2018; Baimukan et al., 2022, inter alia). These evaluation campaigns have led to the development of practical applications, such as ADIDA666http://adida.camel-lab.com/ Obeid et al. (2019), a web interface for fine-grained Arabic DID based on the highest performing system in the MADAR shared task Salameh et al. (2018). In this work, we employ one of the DID systems described by Salameh et al. (2018);777We use regional level classification instead of fine-grained city-level classification because the morphological analyzers are designed at the regional level. however, we differ from their work in that we combine DID with multi-dialect morphological disambiguation to allow users to easily select an appropriate dialect-specific Arabic disambiguator based on the DID prediction. ## 4 System Design and Implementation ### 4.1 Design Considerations We want an easy-to-use one-stop online-accessible user interface that supports the analysis of Arabic sentences from different dialects, and with access to under-the-hood decisions about disambiguation. To that end, we are inspired by three web interfaces: MADAMIRA Pasha et al. (2014) for in-context disambiguation, CALIMAStar Taji et al. (2018a) for out-of-context analysis, and ADIDA Obeid et al. (2019) for dialect identification. Furthermore, we would like the web interface to have a responsive design with streamlined user experiences across a range of devices from mobile to desktops. ### 4.2 Implementation #### Back-end The back-end is implemented in Python using Flask888https://flask.palletsprojects.com/ to serve a REST API. We implemented the MODEL-6 DID system described by Salameh et al. (2018) for automatic dialect identification and the morphological disambiguation system described by Inoue et al. (2022). The implementation of the morphological disambiguator was provided by the CAMeL Tools999https://github.com/CAMeL-Lab/camel_tools Python API Obeid et al. (2020). We plan to add our MODEL-6 implementation to CAMeL Tools. For morphological disambiguation, we use the unfactored model with a morphological analyzer for all variants. We chose the unfactored models because they are faster than the factored models and only slightly lower in performance. Table 1 shows the performance accuracy of Camelira’s morphological disambiguation models. We report numbers on DEV as presented in Inoue et al. (2022). For DID, we train our MODEL-6 using the TRAIN split and evaluate using the DEV and TEST splits following Salameh et al. (2018). Table 2 compares the performance of our implementation with that of Salameh et al. (2018). Our results are slightly lower due to implementation differences. | ALL TAGS | POS ---|---|--- MSA | 95.9 | 98.7 EGY | 90.5 | 94.0 GLF | 93.8 | 96.6 LEV | 85.5 | 92.7 Table 1: Accuracy of Camelira’s morphological disambiguation models based on Inoue et al. (2022)’s unfactored+Morph models. ALL TAGS is the accuracy of the combined morphosyntactic features. | DEV | TEST ---|---|--- Camelira | 92.8 | 93.5 Salameh et al. | 93.1 | 93.6 Table 2: Accuracy of Camelira’s implementation of the MODEL-6 DID model compared with Salameh et al. (2018)’s implementation of the same model. #### Front-end The front-end was implemented using Vue.js101010https://vuejs.org/ for model view control and Bulma111111https://bulma.io/ for styling and creating a responsive design that works well across devices. Figure 3: The Camelira interface with an MSA example sentence and “Tokenized” display tab. This is an exact replica of the input and output choices as in Figure 1 except that the word forms are presented in full tokenization. ### 4.3 The Camelira Interface The Camelira interface is divided into three main areas, the Input Area, Text Output Area, and Morphological Analysis Area. Figure 1 shows an example of a disambiguated MSA sentence in the Camelira web interface. We also provide the option of viewing the interface in Arabic as seen in Figure 2. #### Input Area At first, only the Input Area is displayed which provides users with an input box where they can enter the sentence they wish to disambiguate. Users are also presented with a drop-down menu where they can select whether to disambiguate the input sentence as a particular dialect (MSA, Egyptian, Gulf, or Levantine) or to have the dialect be automatically selected. #### Text Output Area Once the submit button is clicked and the sentence has been disambiguated, the Text Output Area is displayed. First, the dialect indicator displays which dialect was used to analyze the provided input. Then, an output box displays the disambiguated sentence in three different views: (a) the Diacritized/POS view which displays the diacritized text (if supported by the selected dialect’s resources) along with the POS tag of each word, (b) the Tokenized view which displays each disambiguated word in its tokenized form where tokens are delimited by a ‘+’ character, and (c) the Lemmatized view where each word is displayed in its lemmatized form. Figure 3 is the same as Figure 1 except that the text output is in Tokenized mode. #### Morphological Analysis Area Below the Text Output Area, the Morphological Analysis Area consists of the Analysis List box (on the right), which displays all analyses of a given word sorted by their disambiguation ranking order, and the Analysis Viewer box (on the left), which displays a selected analysis in an easy-to-read form with more morphological feature details. The analysis list displays the disambiguation score of each analysis as well as the values for a reduced set of features. Clicking on a word in the Text Output Area selects that word, displaying its analyses in the analysis list and analysis viewer boxes. Clicking on an analysis in the Analysis List will display its user-friendly form in the Analysis Viewer. By default, the top analysis is selected. #### Dialect Identification and Morphological Disambiguation Figures 4 and 5 present Egyptian and Gulf Arabic examples, respectively. Both are presented in a mobile setting to demonstrate our responsive design. In the case of Figure 4, the user selected Auto-Detect for dialect identification. In the Gulf example, the user selected Gulf Arabic directly. Note that the Gulf Arabic does not show diacritizations since its training data did not include diacritized forms Khalifa et al. (2020). Figure 4: The Camelira interface with an Egyptian example sentence: _"A very cool song [video clip], you’ll regret it if you don’t watch it."_ In this example, the input text is automatically correctly detected as Egyptian. Figure 5: The Camelira interface with a Gulf example sentence: _"Go up to your room, I don’t want to hear you talking about this subject again."_ In this example, the user specified the input dialect as Gulf. ## 5 Conclusion and Future Work We presented Camelira, a user-friendly web interface for Arabic multi-dialect morphological disambiguation that covers four major variants of Arabic. The system takes a sentence as input and provides an automatically disambiguated reading for each word, as well as its alternative readings, allowing users to explore various linguistic information, such as part-of-speech, morphological features, and lemmas. Camelira also provides an option to automatically choose an appropriate dialect-specific disambiguator based on the prediction of its dialect identification component. In the future, we plan to extend our disambiguation system to cover other Arabic dialects such as Maghrebi and Yemeni Arabic. We also plan to continue to update the system using future improvements in terms of efficiency and accuracy in CAMeL Tools Obeid et al. (2020). ## Limitations and Ethical Considerations We acknowledge that our system is currently limited to specific variants of Arabic and it can produce erroneous predictions especially on different dialects, genres, and styles that are not covered in the current system’s training data. We also acknowledge that our work on core and generic NLP technologies can be used as part of the pipeline of other systems with malicious intents. ## Acknowledgements Some of this work was carried out on the High Performance Computing resources at New York University Abu Dhabi. We thank Salam Khalifa and Bashar Alhafni for their insightful comments and helpful discussions. We also thank anonymous reviewers for their helpful comments. ## References * Abdelali et al. (2016) Ahmed Abdelali, Kareem Darwish, Nadir Durrani, and Hamdy Mubarak. 2016. Farasa: A fast and furious segmenter for Arabic. In _Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics (NAACL)_ , pages 11–16, San Diego, California. * Abdul-Mageed et al. (2020a) Muhammad Abdul-Mageed, Chiyu Zhang, Houda Bouamor, and Nizar Habash. 2020a. NADI 2020: The first nuanced Arabic dialect identification shared task. In _Proceedings of the Fifth Arabic Natural Language Processing Workshop_ , pages 97–110, Barcelona, Spain (Online). * Abdul-Mageed et al. (2021) Muhammad Abdul-Mageed, Chiyu Zhang, AbdelRahim Elmadany, Houda Bouamor, and Nizar Habash. 2021. NADI 2021: The second nuanced Arabic dialect identification shared task. In _Proceedings of the Sixth Arabic Natural Language Processing Workshop_ , pages 244–259, Kyiv, Ukraine (Virtual). * Abdul-Mageed et al. (2022) Muhammad Abdul-Mageed, Chiyu Zhang, AbdelRahim Elmadany, Houda Bouamor, and Nizar Habash. 2022. NADI 2022: The Third Nuanced Arabic Dialect Identification Shared Task. In _Proceedings of the Seven Arabic Natural Language Processing Workshop (WANLP 2022)_. * Abdul-Mageed et al. (2020b) Muhammad Abdul-Mageed, Chiyu Zhang, Azadeh Hashemi, and El Moatez Billah Nagoudi. 2020b. AraNet: A deep learning toolkit for Arabic social media. In _Proceedings of the 4th Workshop on Open-Source Arabic Corpora and Processing Tools, with a Shared Task on Offensive Language Detection_ , pages 16–23, Marseille, France. European Language Resource Association. * Altantawy et al. (2010) Mohamed Altantawy, Nizar Habash, Owen Rambow, and Ibrahim Saleh. 2010. Morphological Analysis and Generation of Arabic Nouns: A Morphemic Functional Approach. In _Proceedings of the Language Resources and Evaluation Conference (LREC)_ , Valletta, Malta. * Baimukan et al. (2022) Nurpeiis Baimukan, Houda Bouamor, and Nizar Habash. 2022. Hierarchical aggregation of dialectal data for Arabic dialect identification. In _Proceedings of the Language Resources and Evaluation Conference_ , pages 4586–4596, Marseille, France. European Language Resources Association. * Bouamor et al. (2019) Houda Bouamor, Sabit Hassan, and Nizar Habash. 2019. The MADAR shared task on Arabic fine-grained dialect identification. In _Proceedings of the Fourth Arabic Natural Language Processing Workshop_ , pages 199–207, Florence, Italy. * Boudchiche et al. (2017) Mohamed Boudchiche, Azzeddine Mazroui, Mohamed Ould Abdallahi Ould Bebah, Abdelhak Lakhouaja, and Abderrahim Boudlal. 2017. AlKhalil Morpho Sys 2: A robust Arabic morpho-syntactic analyzer. _Journal of King Saud University - Computer and Information Sciences_ , 29(2):141–146. * Boudlal et al. (2010) Abderrahim Boudlal, Abdelhak Lakhouaja, Azzeddine Mazroui, Abdelouafi Meziane, MOAO Bebah, and M Shoul. 2010. Alkhalil Morpho Sys1: A morphosyntactic analysis system for Arabic texts. In _Proceedings of the International Arab Conference on Information Technology_ , pages 1–6. * Buckwalter (2002) Tim Buckwalter. 2002. Buckwalter Arabic morphological analyzer version 1.0. Linguistic Data Consortium (LDC) catalog number LDC2002L49, ISBN 1-58563-257-0. * Darwish et al. (2021) Kareem Darwish, Nizar Habash, Mourad Abbas, Hend Al-Khalifa, Huseein T Al-Natsheh, Houda Bouamor, Karim Bouzoubaa, Violetta Cavalli-Sforza, Samhaa R El-Beltagy, Wassim El-Hajj, et al. 2021. A panoramic survey of natural language processing in the arab world. _Communications of the ACM_ , 64(4):72–81. * Eskander et al. (2013) Ramy Eskander, Nizar Habash, and Owen Rambow. 2013. Automatic extraction of morphological lexicons from morphologically annotated corpora. In _Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)_ , pages 1032–1043, Seattle, Washington, USA. * Eskander et al. (2016) Ramy Eskander, Nizar Habash, Owen Rambow, and Arfath Pasha. 2016. Creating resources for Dialectal Arabic from a single annotation: A case study on Egyptian and Levantine. In _Proceedings of the International Conference on Computational Linguistics (COLING)_ , pages 3455–3465, Osaka, Japan. * Ferguson (1959) Charles F Ferguson. 1959. Diglossia. _Word_ , 15(2):325–340. * Graff et al. (2009) David Graff, Mohamed Maamouri, Basma Bouziri, Sondos Krouna, Seth Kulick, and Tim Buckwalter. 2009. Standard Arabic Morphological Analyzer (SAMA) Version 3.1. Linguistic Data Consortium LDC2009E73. * Habash (2007) Nizar Habash. 2007. Arabic Morphological Representations for Machine Translation. In Antal van den Bosch and Abdelhadi Soudi, editors, _Arabic Computational Morphology: Knowledge-based and Empirical Methods_. Kluwer/Springer. * Habash et al. (2012) Nizar Habash, Ramy Eskander, and Abdelati Hawwari. 2012. A Morphological Analyzer for Egyptian Arabic. In _Proceedings of the Workshop of the Special Interest Group on Computational Morphology and Phonology (SIGMORPHON)_ , pages 1–9, Montréal, Canada. * Habash and Rambow (2006) Nizar Habash and Owen Rambow. 2006. MAGEAD: A morphological analyzer and generator for the Arabic dialects. In _Proceedings of the International Conference on Computational Linguistics and the Conference of the Association for Computational Linguistics (COLING-ACL)_ , pages 681–688, Sydney, Australia. * Habash et al. (2005) Nizar Habash, Owen Rambow, and George Kiraz. 2005. Morphological Analysis and Generation for Arabic Dialects. In _Proceedings of the Workshop on Computational Approaches to Semitic Languages (CASL)_ , pages 17–24, Ann Arbor, Michigan. * Habash (2010) Nizar Y Habash. 2010. _Introduction to Arabic natural language processing_. Morgan & Claypool Publishers. * Holes (2004) Clive Holes. 2004. _Modern Arabic: Structures, Functions, and Varieties_. Georgetown Classics in Arabic Language and Linguistics. Georgetown University Press. * Inoue et al. (2022) Go Inoue, Salam Khalifa, and Nizar Habash. 2022. Morphosyntactic tagging with pre-trained language models for Arabic and its dialects. In _Findings of the Association for Computational Linguistics: ACL 2022_ , pages 1708–1719, Dublin, Ireland. * Inoue et al. (2017) Go Inoue, Hiroyuki Shindo, and Yuji Matsumoto. 2017. Joint Prediction of Morphosyntactic Categories for Fine-Grained Arabic Part-of-Speech Tagging Exploiting Tag Dictionary Information. In _Proceedings of the Conference on Computational Natural Language Learning (CoNLL)_ , pages 421–431, Vancouver, Canada. * Khalifa et al. (2017) Salam Khalifa, Sara Hassan, and Nizar Habash. 2017. A morphological analyzer for Gulf Arabic verbs. In _Proceedings of the Workshop for Arabic Natural Language Processing (WANLP)_ , Valencia, Spain. * Khalifa et al. (2016) Salam Khalifa, Nasser Zalmout, and Nizar Habash. 2016. Yamama: Yet another multi-dialect Arabic morphological analyzer. In _Proceedings of the International Conference on Computational Linguistics (COLING)_ , pages 223–227, Osaka, Japan. * Khalifa et al. (2020) Salam Khalifa, Nasser Zalmout, and Nizar Habash. 2020. Morphological analysis and disambiguation for Gulf Arabic: The interplay between resources and methods. In _Proceedings of the 12th Language Resources and Evaluation Conference_ , pages 3895–3904, Marseille, France. European Language Resources Association. * Maamouri et al. (2003) Mohamed Maamouri, Ann Bies, Hubert Jin, and Tim Buckwalter. 2003. Arabic treebank: Part 1 v 2.0. Linguistic Data Consortium (LDC2003T06). * Maamouri et al. (2009) Mohamed Maamouri, Ann Bies, Seth Kulick, Fatma Gaddeche, Wigdan Mekki, Sondos Krouna, and Basma Bouziri. 2009. The penn Arabic treebank part 3 v 3.1. Linguistic Data Consortium (LDC2008E22). * Maamouri et al. (2011) Mohamed Maamouri, Ann Bies, Seth Kulick, Fatma Gaddeche, Wigdan Mekki, Sondos Krouna, Basma Bouziri, and Wadji Zaghouani. 2011. Arabic treebank: Part 2 v 3.1. * Malmasi et al. (2016) Shervin Malmasi, Marcos Zampieri, Nikola Ljubešić, Preslav Nakov, Ahmed Ali, and Jörg Tiedemann. 2016. Discriminating between similar languages and Arabic dialect identification: A report on the third DSL shared task. In _Proceedings of the Third Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial3)_ , pages 1–14, Osaka, Japan. The COLING 2016 Organizing Committee. * Mubarak and Darwish (2014) Hamdy Mubarak and Kareem Darwish. 2014. Using Twitter to collect a multi-dialectal corpus of Arabic. In _Proceedings of the Workshop for Arabic Natural Language Processing (WANLP)_ , Doha, Qatar. * Obeid et al. (2019) Ossama Obeid, Mohammad Salameh, Houda Bouamor, and Nizar Habash. 2019. ADIDA: Automatic dialect identification for Arabic. In _Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (Demonstrations)_ , pages 6–11, Minneapolis, Minnesota. * Obeid et al. (2020) Ossama Obeid, Nasser Zalmout, Salam Khalifa, Dima Taji, Mai Oudah, Bashar Alhafni, Go Inoue, Fadhl Eryani, Alexander Erdmann, and Nizar Habash. 2020. CAMeL tools: An open source python toolkit for Arabic natural language processing. In _Proceedings of the 12th Language Resources and Evaluation Conference_ , pages 7022–7032, Marseille, France. European Language Resources Association. * Pasha et al. (2014) Arfath Pasha, Mohamed Al-Badrashiny, Mona Diab, Ahmed El Kholy, Ramy Eskander, Nizar Habash, Manoj Pooleery, Owen Rambow, and Ryan Roth. 2014. MADAMIRA: A fast, comprehensive tool for morphological analysis and disambiguation of Arabic. In _Proceedings of the Language Resources and Evaluation Conference (LREC)_ , pages 1094–1101, Reykjavik, Iceland. * Qi et al. (2020) Peng Qi, Yuhao Zhang, Yuhui Zhang, Jason Bolton, and Christopher D. Manning. 2020\. Stanza: A python natural language processing toolkit for many human languages. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations_ , pages 101–108, Online. * Salameh et al. (2018) Mohammad Salameh, Houda Bouamor, and Nizar Habash. 2018. Fine-grained Arabic dialect identification. In _Proceedings of the International Conference on Computational Linguistics (COLING)_ , pages 1332–1344, Santa Fe, New Mexico, USA. * Smrž (2007) Otakar Smrž. 2007. ElixirFM — Implementation of Functional Arabic Morphology. In _Proceedings of the Workshop on Computational Approaches to Semitic Languages (CASL)_ , pages 1–8, Prague, Czech Republic. ACL. * Straka et al. (2016) Milan Straka, Jan Hajič, and Jana Straková. 2016. UDPipe: Trainable pipeline for processing CoNLL-U files performing tokenization, morphological analysis, POS tagging and parsing. In _Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16)_ , pages 4290–4297, Portorož, Slovenia. European Language Resources Association (ELRA). * Taji et al. (2018a) Dima Taji, Jamila El Gizuli, and Nizar Habash. 2018a. An Arabic dependency treebank in the travel domain. In _Proceedings of the Workshop on Open-Source Arabic Corpora and Processing Tools (OSACT)_ , Miyazaki, Japan. * Taji et al. (2018b) Dima Taji, Salam Khalifa, Ossama Obeid, Fadhl Eryani, and Nizar Habash. 2018b. An Arabic morphological analyzer and generator with copious features. In _Proceedings of the Fifteenth Workshop on Computational Research in Phonetics, Phonology, and Morphology_ , pages 140–150, Brussels, Belgium. * Zaghouani and Charfi (2018) Wajdi Zaghouani and Anis Charfi. 2018. ArapTweet: A Large Multi-Dialect Twitter Corpus for Gender, Age and Language Variety Identification. In _Proceedings of the Language Resources and Evaluation Conference (LREC)_ , Miyazaki, Japan. * Zaidan and Callison-Burch (2011) Omar F Zaidan and Chris Callison-Burch. 2011. The Arabic Online Commentary Dataset: an Annotated Dataset of Informal Arabic With High Dialectal Content. In _Proceedings of the Conference of the Association for Computational Linguistics (ACL)_ , pages 37–41. * Zalmout and Habash (2017) Nasser Zalmout and Nizar Habash. 2017. Don’t throw those morphological analyzers away just yet: Neural morphological disambiguation for Arabic. In _Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)_ , pages 704–713, Copenhagen, Denmark. * Zalmout and Habash (2020) Nasser Zalmout and Nizar Habash. 2020. Joint diacritization, lemmatization, normalization, and fine-grained morphological tagging. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , pages 8297–8307, Online. * Zampieri et al. (2017) Marcos Zampieri, Shervin Malmasi, Nikola Ljubešić, Preslav Nakov, Ahmed Ali, Jörg Tiedemann, Yves Scherrer, and Noëmi Aepli. 2017. Findings of the VarDial evaluation campaign 2017. In _Proceedings of the Fourth Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial)_ , pages 1–15, Valencia, Spain. * Zampieri et al. (2018) Marcos Zampieri, Shervin Malmasi, Preslav Nakov, Ahmed Ali, Suwon Shon, James Glass, Yves Scherrer, Tanja Samardžić, Nikola Ljubešić, Jörg Tiedemann, Chris van der Lee, Stefan Grondelaers, Nelleke Oostdijk, Dirk Speelman, Antal van den Bosch, Ritesh Kumar, Bornini Lahiri, and Mayank Jain. 2018. Language identification and morphosyntactic tagging: The second VarDial evaluation campaign. In _Proceedings of the Fifth Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial 2018)_ , pages 1–17, Santa Fe, New Mexico, USA.
# Efficient Adversarial Input Generation via Neural Net Patching Tooba Khan, Kumar Madhukar, Subodh Vishnu Sharma ###### Abstract The adversarial input generation problem has become central in establishing the robustness and trustworthiness of deep neural nets, especially when they are used in safety-critical application domains such as autonomous vehicles and precision medicine. This is also practically challenging for multiple reasons – scalability is a common issue owing to large-sized networks, and the generated adversarial inputs often lack important qualities such as naturalness and output-impartiality. We relate this problem to the task of patching neural nets, i.e. applying small changes in some of the network’s weights so that the modified net satisfies a given property. Intuitively, a patch can be used to produce an adversarial input because the effect of changing the weights can also be brought about by changing the inputs instead. This work presents a novel technique to patch neural networks, and an innovative approach of using it to produce perturbations of inputs which are adversarial for the original net. We note that the proposed solution is significantly more effective than the prior state-of-the-art techniques. ## Introduction Deep Neural Networks (DNNs) today are omnipresent. An important reason behind their widespread use is their ability to generalize and thereby perform well even on previously unseen inputs. While this is a great practical advantage, it may sometimes make DNNs unreliable. In safety- or business-critical applications this lack of reliability can indeed have dreadful costs. Evidently, central to a trained network’s unreliability lies the lack of robustness against input perturbations, i.e., small changes to some inputs cause a substantial change in the network’s output. This is undesirable in many application domains. For example, consider a network that has been trained to issue advisories to aircrafts to alter their paths based on approaching intruder aircrafts. It is natural to expect such a network to be robust in its decision-making, i.e. the advisory issued for two very similar situations should not be vastly different. At the same time, if that is not the case, then demonstrating the lack of robustness through _adversarial inputs_ can help not only in improving the network but also in deciding when the network should relinquish control to a more dependable entity. Given a network and an input, an adversarial input is one which is _very close_ to the given input and yet the network’s outputs for the two inputs are quite different. In the last several years, there has been much work on finding adversarial inputs (Pei et al. 2017; Guo et al. 2018; Goodfellow, Shlens, and Szegedy 2015; Kurakin, Goodfellow, and Bengio 2017; Alparslan et al. 2021). These approaches can be divided into black-box and white-box methods based on whether they consider the network’s architecture during the analysis or not. A variety of techniques have been developed in both these classes, ranging from generation of random attacks (Goodfellow, Shlens, and Szegedy 2015; Kurakin, Goodfellow, and Bengio 2017) and gradient-based methods (Alparslan et al. 2021) to symbolic execution (Wang et al. 2018; Yang et al. 2021; Li et al. 2019), fault localization (Eniser, Gerasimou, and Sen 2019), coverage-guided testing (Pei et al. 2017; Guo et al. 2018), and SMT and ILP solving (Gopinath et al. 2018). However, there are several issues that limit the practicability of these techniques – poor success rate, large distance between the adversarial and the original input (both in terms of the number of input values changed, and the degree of the change), unnatural or perceivably different inputs, and output partiality (techniques’ bias to produce adversarial examples for just one of the network’s outputs). This paper presents a useful approach to generate adversarial inputs in a way that addresses these issues. We relate the problem of finding adversarial inputs to the task of patching neural nets, i.e. applying small changes in some of the network’s weights so that the modified net satisfies a given property. Patching DNNs is a topic of general interest to the Machine Learning community because of its many applications, which include bug-fixing, watermark resilience, and fine-tuning of DNNs, among others (Goldberger et al. 2020). Intuitively, the relation between these two problems is based on the observation that a patch can be translated into an adversarial input because the effect of changing the weights may be brought about by changing the inputs instead. In fact, a patch in the very first _edge-layer_ of a network can very easily be transformed into a corresponding change in the input by just solving linear equations. While there are techniques to solve the patching problem (Goldberger et al. 2020; Refaeli and Katz 2021), it is in general a difficult task, particularly for layers close to the input layer. Informally, this is because the computation of the entire network needs to be encoded and passed to a constraint solver in order to obtain a patch. For large-sized networks, this gives rise to a big monolithic constraint leading to scalability issues for the solver. We address this by proposing an improvement in the technique of (Goldberger et al. 2020), and then using it repeatedly to find a middle-layer patch and chop off the latter half of the network, till a first-layer patch has been obtained. Our experiments on three popular image-dataset benchmarks show that our approach does significantly better than other state-of-the-art techniques, in terms of the number of pixels modified as well as the magnitude of the change. This reflects in the quality of the adversarial images, both visually and in several qualitative metrics that we present later. The core contributions of this paper are: * • A novel technique to patch neural networks, and an innovative approach of using it to produce small perturbations of inputs which are adversarial for the original net. * • An extensive experimental evaluation using CIFAR-10 (Krizhevsky and Hinton 2009), MNIST (Deng et al. 2009a), and ImageNet (Deng et al. 2009b) datasets, and a number of qualitative parameters, to show the efficacy of our approach over the state-of-the-art. ## Illustrative Example Let us consider a toy DNN, $\mathcal{N}$, shown in Fig. 1. It has two neurons in each of its four layers – the input layer first, followed by two hidden layers, and then the output layer. We assume that the hidden layers have ReLU111ReLU($x$) = $\mathit{max}(0,x)$ as the activation function. For neurons without an activation function, the value of a neuron is computed by summing up, for each incoming edge to the neuron, the product of the edge weight and the value of the neuron at the edge’s source. In presence of an activation function, the value is computed by applying the function on this sum. For example, for the input $\langle 0.5,0.5\rangle$, the values at the next three layers are $\langle 1,0.5\rangle$, $\langle 3.5,1.5\rangle$, and $\langle-5,-6.5\rangle$ respectively. 1$\rightarrow$-3/42$\rightarrow$7/8 1/12$\leftarrow$1/2 1/24$\leftarrow$1/2 1/2$\rightarrow$0 3/2$\rightarrow$1/8 -6.5$\rightarrow$-0.375 1$\rightarrow$1/8 7/2$\rightarrow$1/4 -5$\rightarrow$-0.5 $i_{1}$$i_{2}$$x_{1}$$x_{2}$$x_{3}$$x_{4}$$o_{1}$$o_{2}$$1$$-1$$2$$1$$3$$1$$2$$5$$-4$$-8$ Figure 1: Adversarial inputs from first-layer modification. The adversarial input and the corresponding values of each neuron are written in red. The modification required in first layer weights are shown in black boxes. Finding an adversarial input $\langle i_{1},i_{2}\rangle$ for the input $\langle 0.5,0.5\rangle$ amounts to finding a value for each $i_{k}$ ($k\in\\{1,2\\}$) such that $|i_{k}-0.5|\leq\delta$ and the corresponding output $o_{2}>o_{1}$, for a given small $\delta$. It is noteworthy that if we want the second output to become bigger than the first one, this can be achieved by modifying the weights instead of the inputs. For example, if the edges connecting the second input neuron to the first hidden layer had the weights $\langle-0.75,0.875\rangle$ instead of $\langle 1,2\rangle$, then the next three layers would have the values $\langle 0.125,0\rangle$, $\langle 0.25,0.125\rangle$, and $\langle-0.5,-0.375\rangle$, and our goal would be met by modifying the weights while keeping the inputs unchanged. We will come to the question of how to find the changed first-layer weights in a bit, but let us first see how the changed weights can help us obtain an adversarial input. This is a rather simple exercise. Note that with the changed weights, the values of the neurons in the first hidden layer were $\langle 0.125,0\rangle$. So, our task is simply to find inputs for which the first hidden layer values stay as $\langle 0.125,0\rangle$, but with the original weights $\langle 1,2\rangle$. This can be done by solving the following equations, where $\delta_{1}$ and $\delta_{2}$ ($\delta_{1},\delta_{2}$ $\leq\delta\leq 0.5$, say) are the changes in the two inputs respectively. $(1/2+\delta_{1})+(1/2+\delta_{2})=1/8$ (1) $-(1/2+\delta_{1})+2*(1/2+\delta_{2})=0$ (2) We get $\delta_{1}=-5/12,\delta_{2}=-11/24$ and, thus, the adversarial input as $\langle 1/12,1/24\rangle$. The dotted rectangles in Fig. 1 contain the adversarial inputs and the corresponding values at each layer. The first thing to notice here is that Eqn. 2 could have been relaxed as $-(1/2+\delta_{1})+2*(1/2+\delta_{2})\leq 0$ because of the ReLU activation function. This may give us smaller values of $\delta_{i}$’s. Moreover, along with minimizing the change in each input pixel, we can also minimize the number of pixels that are modified, as shown here. $(1/2+\delta_{1}*M_{1})+(1/2+\delta_{2}*M_{2})=1/8$ (3) $-(1/2+\delta_{1}*M_{1})+2*(1/2+\delta_{2}*M_{2})<=0$ (4) $M_{1},M_{2}\in\\{0,1\\};~{}~{}~{}\mathit{minimize}~{}~{}\sum M_{i}$ (5) In pratice, we solve these constraints in place of Eqns. 1-2; this gives us better adversarial inputs. There are a few more points to note before we proceed. First, it is only the first-layer modification that may be easily translated into an adversarial input as illustrated. Modification in deeper layers are not immediately helpful; they cannot be translated easily into an adversarial input because of the non-linear activation functions. Second, we need to find _small_ modifications in the weights, so that the corresponding $\delta_{i}$’s in the inputs fall within the allowed $\delta$. And, lastly, while there are ways to compute a first-layer change directly using an SMT or an ILP solver (Goldberger et al. 2020), this approach is not very scalable in practice as large-sized networks give rise to big monolithic formulas that may be difficult for the solver. Instead, we propose an iterative approach that finds a middle-layer modification using (Goldberger et al. 2020)222In fact, we use an improved version of this which is described in the next section. and chops off the latter half of the network, repeatedly till a first-layer patch is found. Since our network has three edge-layers, we start by finding a small modification of the weights in the second edge-layer with which we can achieve our target of making $o_{2}$ bigger than $o_{1}$ for the input $\langle 0.5,0.5\rangle$. We propose an $\epsilon_{i,j}$ change in the weight of the $j^{th}$ edge in $i^{th}$ edge-layer, and encode the constraints for $o_{2}>o_{1}$ as follows: $x_{3}=\mathit{max}(0,1*(2+\epsilon_{2,1})+1/2*(3+\epsilon_{2,2}))$ (6) $x_{4}=\mathit{max}(0,1*(1+\epsilon_{2,3})+1/2*(1+\epsilon_{2,4}))$ (7) $-4*x_{3}+5*x_{4}>2*x_{3}-8*x_{4}$ (8) The range of each $\epsilon_{i,j}$ is [$-\alpha,\alpha$] if $\alpha$ is the biggest permissible modification for an edge-weight. We minimize the magnitude of the total change using Gurobi (Gurobi Optimization, LLC 2022). For the equations above, we get $\langle\epsilon_{2,1},\epsilon_{2,2},\epsilon_{2,3},\epsilon_{2,4}\rangle=\langle-9/8,-17/4,-5/4,-1/4\rangle$. These changes indeed make the second output bigger (see Fig. 2). Extracted Network 1/2 3/2$\rightarrow$1/8 -6.5$\rightarrow$-0.375 1 7/2$\rightarrow$1/4 -5$\rightarrow$-0.5 $i_{1}$$1/2$$i_{2}$$1/2$$x_{1}$$x_{2}$$x_{3}$$x_{4}$$o_{1}$$o_{2}$$1$$2$$-1$$1$2$\rightarrow$7/81$\rightarrow$3/43$\rightarrow$-5/41$\rightarrow$-1/4$2$$5$$-4$$-8$ Figure 2: Middle-layer modification and sub-net extraction Our next step is to extract a sub-network (see Fig. 2) and look for a modification in its middle layer. Since the extracted network has only two edge-layers, this step will give us a first-layer modification. The equations, subject to the constraint that $x_{3}$ and $x_{4}$ get the values 1/4 and 1/8 respectively, are shown below. $x_{1}=\mathit{max}(0,1/2*(1+\epsilon_{1,1})+1/2*(1+\epsilon_{1,2}))$ (9) $x_{2}=\mathit{max}(0,1/2*(-1+\epsilon_{1,3})+1/2*(2+\epsilon_{1,4}))$ (10) $2*x_{1}+3*x_{2}=1/4;~{}~{}~{}1*x_{1}+1*x_{2}=1/8$ (11) Gurobi gives us the solution $\langle\epsilon_{1,1},\epsilon_{1,2},\epsilon_{1,3},\epsilon_{1,4}\rangle=\langle 0,-7/4,0,-9/8\rangle$, from which we can obtain the adversarial input $\langle 1/12,1/24\rangle$ as already shown above. ## Methodology We now describe the technical details of our approach and present our algorithm. We begin with the notation that we use. Let $\mathcal{N}$ denote the DNN that we have with $n$ inputs ($i_{1},i_{2},\ldots,i_{n}$), $m$ outputs ($o_{1},o_{2},\ldots,o_{m}$), and $k$ layers ($l_{1},l_{2},\ldots,l_{k}$). We use $v_{p,q}$ to denote the value of the $q^{th}$ neuron in $l_{p}$. We assume that $\mathcal{N}$ is feed-forward, i.e., the (weighted) edges connect neurons in adjacent layers only, and point in the direction of the output layer. We use the term _edge-layer_ to refer to all the edges between any two adjacent layers of $\mathcal{N}$, and denote the edge layers as $el_{1},el_{2},\ldots,el_{k-1}$. We also assume the hidden layers ($l_{2},l_{3},\ldots,l_{k-1}$) in $\mathcal{N}$ have ReLU activation function, and that there is no activation function on any output neuron. For simplicity, we assume that the neurons do not have any biases. This is not a limitation in any sense; our implementation handles them directly. Moreover, a DNN with biases can be converted into an equivalent one without any biases. Like in the previous section, we use $\delta_{p}$ to denote the change in the $p^{th}$ input, and $\epsilon_{q,s}$ to denote the change in the weight of the $q^{th}$ edge in $el_{s}$. The $\delta_{p}$’s are constrained to be $\leq\delta$, which is the biggest perturbation allowed in any pixel to find an adversarial input. Algorithm 1 presents a pseudocode of our algorithm. The inputs to the algorithm are: $\mathcal{N},\delta$, and the values for the input neurons $v_{1,1},v_{1,2},\ldots,v_{1,n}$, for which the corresponding output does not satisfy a given adversarial property $\phi(o_{1},o_{2},\ldots,o_{m})$ (denoted simply as $\phi$, henceforth). The aim of our algorithm is to find a new set of input values $v^{\prime}_{1,1},v^{\prime}_{1,2},\ldots,v^{\prime}_{1,n}$ such that $\mathcal{N}$’s output corresponding to these new inputs satisfies $\phi$. The algorithm works in the following two steps. First, it finds a _small_ modification in the weights of $el_{1}$ to derive $\mathcal{N}_{mod}$ (which is essentially $\mathcal{N}$ with the modified weights in $el_{1}$), such that the output of $\mathcal{N}_{mod}$ for the input $v_{1,1},v_{1,2},\ldots,v_{1,n}$ satisfies $\phi$. Then, the algorithm translates this first-layer modification into adversarial inputs $v^{\prime}_{1,1},v^{\prime}_{1,2},\ldots,v^{\prime}_{1,n}$, subject to the constraint that $|v^{\prime}_{1,p}-v_{1,p}|\leq\delta$, for every $p\in[1,n]$. This second step is shown in the algorithm as the function $\mathit{mod2adv}$, the pseudocode of which has been omitted as this is a simple call to Gurobi as illustrated in the previous section. Let us assume for the time being that we have a sub-routine $\mathit{modifyEdgeLayer}$ that takes as input a DNN $\mathcal{N}$, one of its edge-layers $el_{k}$, values of the input neurons $v_{1,1},v_{1,2},\ldots,v_{1,n}$, and a property $\phi$ on the output layer neurons, and returns a new network $\mathcal{N}_{mod}$ with the constraints that: * • $\mathcal{N}_{mod}$ is same as $\mathcal{N}$ except for the weights in $el_{k}$, and * • the output of $\mathcal{N}_{mod}$ on $v_{1,1},v_{1,2},\ldots,v_{1,n}$ satisfies $\phi$. Clearly, with such a sub-routine, the first step of our algorithm becomes trivial. We would simply call $\mathit{modifyEdgeLayer}$ with the given input, $\mathcal{N}$, $\phi$, and $el_{1}$. We refer to the work of Goldberger et al. (Goldberger et al. 2020) which gives us exactly this. However, we do not use it directly to find our first-layer modification. Informally, the technique of (Goldberger et al. 2020) uses variables $\epsilon_{q,s}$ to denote the changes in the weights (in a given edge-layer $s$) and encodes the computation of the entire network, and then adds the constraint that the output must satisfy $\phi$. It then uses Gurobi on these constraints, to solve for (and optimize) the values of $\epsilon_{q,s}$. Consider an example (reproduced from (Goldberger et al. 2020)) shown in Fig. 3 with the output property $\phi:=(v_{3,1}\geq v_{3,2})$. Let us ignore the color of the output neurons for now. If the input neurons are given values $v_{1,1}=3$ and $v_{1,2}=4$, the output neurons get the value $v_{3,1}=-2$ and $v_{3,2}=2$, which does not satisfy $\phi$. Note that the hidden layer neurons have ReLU activation function. In order to obtain a second edge-layer modification such that $\phi$ holds for the input $\langle 3,4\rangle$, the technique of (Goldberger et al. 2020) generates the following constraints. $v_{1,1}$$v_{1,2}$$v_{2,1}$$v_{2,2}$$v_{3,1}$$v_{3,2}$1-12-2 1 1 -1 -1 Figure 3: Example illustrating DNN modification, from (Goldberger et al. 2020) $\mathit{minimize}~{}~{}M$ (12) $M\geq 0$ (13) $-M\leq\epsilon_{1,2}\leq M$ (14) $-M\leq\epsilon_{2,2}\leq M$ (15) $-M\leq\epsilon_{3,2}\leq M$ (16) $-M\leq\epsilon_{4,2}\leq M$ (17) $v_{3,1}=0.(1+\epsilon_{1,2})+2.(-1+\epsilon_{2,2})$ (18) $v_{3,2}=0.(-1+\epsilon_{3,2})+2.(1+\epsilon_{4,2})$ (19) $v_{3,1}\geq v_{3,2}$ (20) Figure 4: DNN modification constraints for Fig.3 In a similar way, the constraints can be generated for modification in any layer, by propagating the input values up to that layer, encoding the computation from there onward, and adding the output property $\phi$. If a first-layer modification has to be found, this gives rise to a big monolithic constraint, particularly for large-sized networks. This does not scale in practice, and therefore we propose an iterative approach of doing this in our algorithm. The iterative approach uses the above technique to find a middle- layer modification, derive a new output property $\phi^{\prime}$ for the first half of the network, and does this repeatedly till a first-layer modification is found. This has been illustrated in the function $\mathit{findFirstLayerMod}$ in Alg. 1, which calls $\mathit{modifyEdgeLayer}$ in a loop. The last bit here is to understand how the modified output property $\phi^{\prime}$ may be derived in each iteration. The way this works is as follows. Let us say that the last call to $\mathit{modifyEdgeLayer}$ was made on edge-layer $el_{(j-1)}$, which connects the layers $l_{(j-1)}$ and $l_{j}$. We can use the modified weights to propagate the inputs all the way to layer $l_{j}$, by simply simulating the network on the inputs. This gives us values for all the neurons in layer $l_{j}$, say $c_{1},c_{2},\ldots$ and so on. Now, the modified weights are replaced by the original weights, and all the layers after $l_{j}$ are dropped off from the network. This reduced network $\mathcal{N}^{\prime}$ has exactly the layers $l_{1}$ to $l_{j}$ of $\mathcal{N}$. We denote this reduction as $\mathcal{N}\downarrow_{(l_{1}\ldots l_{j})}$. Ideally, we would want to find a middle-layer modification in $\mathcal{N}^{\prime}$ under the new output constraint as $\phi^{\prime}:=(v_{j,1}=c_{1})\wedge(v_{j,2}=c_{2})\wedge\ldots$ and so on. The updated $\phi^{\prime}$ is correct because we know that $\phi$ gets satisfied when these values are propagated further to the output layer. But, we can relax the strict equalities of $\phi^{\prime}$ into inequalities as discussed in the next subsection. Algorithm 1 Adversarial Inputs via Network Patching Input: $\mathcal{N},l,\delta,\phi,$ and input $\langle v_{1,1},\ldots,v_{1,n}\rangle$ Output: Adversarial input $\langle v^{\prime}_{1,1},\ldots,v^{\prime}_{1,n}\rangle$ findFirstLayerMod($\mathcal{N},l,\mathit{in},\phi$): 1: while true do 2: $p\leftarrow\lceil(l/2)\rceil$ 3: $\mathcal{N}_{mod}\leftarrow\mathit{modifyEdgeLayer}(\mathcal{N},\mathit{in},\phi,el_{p})$ 4: return $\mathcal{N}_{mod}$ if$(p=1)$ 5: $\langle c_{1},c_{2},\ldots\rangle\leftarrow\mathit{simulate}(\mathcal{N}_{mod},\mathit{in})$ 6: $\phi^{\prime}=(v_{(p+1),1}=c_{1})\wedge(v_{(p+1),2}=c_{2})\wedge\ldots$ 7: $\mathcal{N}^{\prime}=\mathcal{N}\downarrow_{(l_{1}\ldots l_{(p+1)})}$ 8: $\mathcal{N}\leftarrow\mathcal{N}^{\prime}$; $l\leftarrow(p+1)$; $\phi\leftarrow\phi^{\prime}$ 9: end while main(): 1: $\mathit{in}\leftarrow\langle v_{1,1},\ldots,v_{1,n}\rangle$ 2: $\mathcal{N}_{mod}\leftarrow\mathit{findFirstLayerMod}$($\mathcal{N},l,\mathit{in},\phi$) 3: $\langle\delta_{1},\delta_{2},\ldots,\delta_{n}\rangle$ = $\mathit{mod2adv}$($\mathcal{N}_{mod},\mathit{in},\delta$) 4: $\langle v^{\prime}_{1,1},\ldots,v^{\prime}_{1,n}\rangle$ = $\langle v_{1,1},\ldots,v_{1,n}\rangle$ \+ $\langle\delta_{1},\ldots,\delta_{n}\rangle$ 5: return $\langle v^{\prime}_{1,1},\ldots,v^{\prime}_{1,n}\rangle$ ### Simplifying DNN Modification Constraints Let us revisit the example of Fig. 3, and the constraints corresponding to the modification problem shown in Fig. 4. Recollect that the modification problem in this example was to find small changes in the weights of the second edge- layer, such that $v_{3,1}\geq v_{3,2}$ for the input $\langle 3,4\rangle$. This is not true for the DNN in this example as $v_{3,1}$ gets the value -2, whereas $v_{3,2}$ gets the value 2\. A possible way of satisfying $v_{3,1}\geq v_{3,2}$ is to change weights in such a way that $v_{3,2}$ _decreases_ and $v_{3,2}$ _increases_. This can give us a marking of the final layer neurons as decrement and increment, which has been indicated by the neuron colors red and green in Fig. 3. The useful thing about such a marking is that it can be propagated backward to other layers. For instance, in the same example, $v_{2,1}$ and $v_{2,2}$ can also be colored green and red, resp. This works by looking at the edge weights. Since $v_{2,1}$ is connected to $v_{3,1}$ with a positive-weight edge, an increase in $v_{3,1}$ can be brought about by _increasing_ $v_{2,1}$. If we look at $v_{2,2}$, since it connected to $v_{3,1}$ with a negative-weight edge, a _decrease_ in $v_{2,2}$ would result in an increase in $v_{3,1}$. This marking was proposed by Elboher et al. (Elboher, Gottschlich, and Katz 2020), although it was in the context of abstraction-refinement of neural networks. We refer the interested readers to (Elboher, Gottschlich, and Katz 2020) for more details about this marking scheme. In what follows, we explain how this marking can be useful in simplifying the modification constraints. Since we are interested in modifying weights in the second edge-layer ($el_{2}$), we propagate the inputs to the second layer of neurons. This gives us the values $\langle 0,2\rangle$ for $\langle v_{2,1},v_{2,2}\rangle$. Having propagated the input to the source neurons of $el_{2}$, and the increment-decrement marking at the target neurons of $el_{2}$, we claim that we can identify whether a given edge-weight should be increased, or decreased. Let us consider the edge between $v_{2,2}$, which has a value of 2, and $v_{3,1}$, which has an increment marking. We claim that the change in this edge, $\epsilon_{2,2}$ should be positive. Naturally, since the value is positive, we should multiply with a _bigger_ weight to get an increased output. Instead, if a positive value was connected to a decrement neuron, we should decrease the weight (for example, for the edge between $v_{2,2}$ and $v_{3,2}$). In case of negative values, just the opposite needs to be done. And if the value is zero, no change needs to be made at all. With this, the constraints in Fig. 4 get simplified as $\epsilon_{1,2}=\epsilon_{2,2}=0,0\leq\epsilon_{3,2}\leq M$ and $-M\leq\epsilon_{3,2}\leq 0$. We have implemented this on top of the tool corresponding to (Goldberger et al. 2020) and used it in our call to $\mathit{modifyEdgeLayer}$. We end this section with a brief note on how this increment decrement marking may help us relax the modified output property $\phi^{\prime}$. Recollect that $\phi^{\prime}$ was derived as a conjunction of equality constraints, where each conjunct was equating a last-layer neuron of the reduced network with the values obtained by simulating the input on $\mathcal{N}_{mod}$. Since we know the increment-decrement marking of last layer neurons, we can relax each conjunct into an inequality by replacing the equality sign with $\leq$ ($\geq$) for decrement (increment) neurons. This allows us to obtain, possibly better, solutions more often. ## Related Work Robustness of DNNs has gained a lot of attention in the last several years as DNNs are permeating our lives, even safety- and business-critical aspects of it. A number of techniques have been developed to establish robustness or demonstrate the lack of it through adversarial examples. These techniques may be broadly classified as black box and white box approaches. Black box methods do not consider the architecture of DNNs in trying to argue about its robustness. Attacks based on the L2 and L0 norms were introduced in (Alparslan et al. 2021). These attacks change the pixel values by some fixed amount, and measure their effectiveness by the decrease in confidence of the original label. Fast Gradient sign method (Goodfellow, Shlens, and Szegedy 2015) is a gradient based method, which uses gradient of the loss of neural network when the original and modified images are fed to it. These methods modify a large number of pixels in the original image, to induce a misclassification. Changing more pixels makes the adversarial image visibly different from the original image. White box methods, on the other hand, involve looking at the complete architecture of the DNNs to obtain adversarial inputs or argue that none exists. Verification of neural networks using Symbolic execution is one such white box approach. As discussed in (Gopinath et al. 2018), it translates the neural network into an imperative program, and uses SMT (Symbolic Modulo Theory) based solver to prove given properties. However, such techniques are not scalable due to exponential time complexity. Symbolic propagation is another white box method which converts inputs to symbolic representations and propagates them through the hidden layers and the output layer. (Li et al. 2019)] uses symbolic propagation along with abstract interpretation to compute bounds on the neurons. (Wang et al. 2018) has used integer arithmetic to find bounds on the outputs. (Yang et al. 2021) verifies local robustness properties using symbolic propagation and global robustness properties using Lipschitz constant based framework. However, these techniques often given loose bounds and lack precision. Another white box technique is to find flaws in the training phase of the neural network such as the use of a inappropriate loss function. The most popularly used loss function while training classification networks is softmax cross entropy loss. The authors of (Pang et al. 2020) claim that the presence of adversarial behavior in a deep neural network can be attributed to its loss function. They propose Max-Mahalanobis center (MMC) loss which learns from dense feature regions in the input images. (Raj Dhamija, Gunther, and Boult 2019) has proposed two loss functions: Entropic Open-Set Loss and Objectoshpere Loss. Entropic Open-Set Loss makes the softmax values of all classes uniform when the input is unknown. Objectoshpere Loss increases the boundary of feature magnitudes of known and unknown samples. Loss function replacement techniques are still vulnerable to adversarial attacks. They can not be transferred either. These methods cannot be used to test the robustness of DNNs that are already in use. DeepFault (Eniser, Gerasimou, and Sen 2019) is another white-box approach for testing DNNs. It uses fault localization, i.e, finding the areas of the network that are mainly responsible for incorrect behaviors. It analyses a trained DNN to find faulty neurons from its hit spectrum analysis using suspiciousness measures. Adversarial inputs are those inputs that achieve high activation values for suspicious neurons. Another popular category of white box methods depends on coverage based techniques. These methods use structural coverage metrics such as neuron coverage and modified condition/decision coverage (MC/DC). These metrics are claimed to be related to faults in neural networks. (Sun et al. 2019) developed a tool which uses MC/DC variants for neural networks and neuron coverage. This tool is used to verify neural networks based on the results of the specified coverage. (Xie et al. 2019) uses mutation based strategies to generate test cases which can maximize neuron coverage. Similarly, (Gao et al. 2020) uses a genetic algorithm based approach which can generate adversarial inputs to maximize neuron coverage and they retrain the DNN simultaneously while generating inputs. (Pei et al. 2017) implements a white-box approach that maximises neuron coverage and differential behaviour of a group of DNNs, implying that different DNNs will produce different output classes for the adversarial input while maximising neuron coverage. (Guo et al. 2018) also focuses on maximising neuron coverage in order to generate adversarial inputs. Even though the code coverage criteria of software engineering test methodologies corresponds to neuron coverage, it is not a useful indicator of the production of adversarial inputs. Neuron coverage statistics, according to authors in (Harel-Canada et al. 2020), lead to the detection of fewer flaws, making them inappropriate for proving the robustness of DNNs. They also present three new standards – defect detection, naturalness, and output impartiality – that can be used to gauge the quality of adversarial inputs produced, as alternatives to the L2 and L-inf norms. Their findings establish that the adversarial image set generated by using neuron coverage measures did not perform well on these three standards. Thus, neuron coverage is not a good measure for testing neural networks and generating adversarial images. We have explained these metrics in the following section. Since our technique relies on finding modifications in DNNs, we also discuss a couple of recent work that is related in this respect. (Goldberger et al. 2020) proposes a technique to find minimal modifications in a single-layer in a DNN, in order to meet a given outcome. This work has been extended further in (Refaeli and Katz 2021) to come up with multi-layer modifications by dividing the problem into multiple sub problems and applying the idea of (Goldberger et al. 2020) on each one of them. Our work proposes an improvement over their idea and uses the improved technique to find small modifications which are then translated into adversarial inputs. ## Experimental Setup We implemented our approach in a tool called Aigent (Adversarial Input Generator), using the Tensorflow and Keras libraries for working with the DNNs. Aigent uses Gurobi for constrained optimization. ### Benchmark datasets We conducted our experiments on three popular datasets: MNIST, CIFAR-10, and ImageNet. We chose these benchmarks because they are readily available and are acceptable as inputs by a number of tools, making it easier to compare different techniques in a fair way. MNIST consists of 60,000 black and white images of handwritten digits for the purpose of training and 10,000 for testing. Each image is of 28x28 size. It has 10 classes with labels corresponding to each digit. CIFAR-10 consists of 60000 32x32 colour images in 10 classes. ImageNet is a large dataset which consists of images in 1000 classes. ### Metrics of Evaluation In addition to the usual metrics like L2 and L-$\infty$ distance, and the time taken, we have used the following metrics to compare our results with the results of existing approaches. 1. 1. Defect detection: to highlight the attack success rate, or the number of benchmarks for which our tool could successfully produce an adversarial image. 2. 2. Naturalness: to score the adversarial images for being admissible, i.e. visibly not very different from the original image. We use Frechet Incéption Distance (FID) (Harel-Canada et al. 2020) to measure naturalness. Values of FID close to 0 indicate that the adversarial images are natural, and are therefore desirable. 3. 3. Output impartiality: to reflect whether the adversarial image generation is biased towards any one of the output classes or not. We measure this using the Pielou score (Harel-Canada et al. 2020), which can range from 0 (biased, undesirable) to 1 (unbiased, desirable). (Harel-Canada et al. 2020) have observed that neuron coverage was negatively correlated with defect detection, naturalness, and output impartiality. Naturalness is considered an essential metric while assessing the quality of adversarial images. Fig. 5 shows how some of the existing methods generate images that perform well on L2 and L-$\infty$ distance metric, but appear very different from the original image. Figure 5: Examples of adversarial images generated by other techniques lacking originality. ## Results Table 1 presents a comparison of Aigent with other state-of-the-art approaches of generating adversarial images, on all the three benchmarks datasets, for several important metrics including FID, Pielou score, defect detection rate, L-2 and L-$\infty$ distance, and the number of pixels modified. The results demonstrate that Aigent performs better than all the other approaches in terms of FID, which indicates that the adversarial images generated by our method are natural and visibly quite similar to the corresponding original images. This is further reinforced by the fact that Aigent modifies far fewer pixels as compared to the other approaches. S. No. Technique FID Pielou score L-2 L-$\infty$ Time (seconds) Pixels modified % of pixels modified Defect detection Benchmark dataset: MNIST 1 Aigent 0.001 0.725 1.82 0.66 1.726 24 3.06% 72.00% 2 Aigent (high defect)1 0.03 0.74 4.1 0.80 1.799 24 3.06% 91.40% 3 FGSM2 1.73 0.95 2.8 0.1 0.069 784 100.00% 99.00% 4 Black Box3 1.98 0.14 6.58 0.23 0.065 784 100.00% 88.40% 5 DeepXplore 0.02 0.47 5.16 1 11.74 60 7.65% 45.66% 6 DLFuzz4 0.17 0.88 2.29 0.39 30 586 74.74% 92.36% Benchmark dataset: CIFAR-10 1 Aigent 0.00009 0.927 1.6 0.5 12.01 12 0.39% 100.0% 2 FGSM2 0.071 0.92 5.5 0.1 0.079 3072 100.00% 100.0% 3 Black Box3 0.44 0.703 13.04 0.23 0.082 3072 100.00% 76.20% Benchmark dataset: ImageNet 1 Aigent 0.00011 0.75 6.81 0.73 35 300 0.61% 98.60% 2 FGSM2 0.43 0.87 22 0.1 0.4 16384 100.00% 97.00% 3 Black Box3 0.05 0.8 52 0.4 0.3 16384 100.00% 90.00% 4 DeepXplore 0.032 N.A 58.04 0.51 84 15658 95.57% 59.13% 5 DLFuzz4 0.11 N.A 61.1 0.6 57 16102 98.28% 92.00% Table 1: Comparison of Aigent with other state-of-the-art techniques on MNIST, CIFAR-10 and ImageNet datasets. Bold values indicate the best figure for each metric. DeepXplore and DLFuzz did not work on CIFAR-10. 1: Tuned to achieve higher defect detection. 2:Gradient Based technique. 3:(Goodfellow, Shlens, and Szegedy 2015). 4: We were getting a few compilation errors in the DLFuzz code (https://github.com/turned2670/DLFuzz) which we fixed for this comparison. Our method could achieve 72% defect detection for MNIST when constraints were stricter. When we allowed the quality of generated adversarial images to degrade slightly in order to achieve a higher defect detection, shown as Aigent (high defect) in the table, we were able to generate adversarial images for 91.4% of the original images. Our method performs comparable to white box methods. Although black box methods achieve higher defect detection, they modify 100% pixels which leads to visibly distinguishable images. Thus, our method performs well in terms of defect detection, while keeping the modification quite small. Default detection on CIFAR-10 and ImageNet benchmarks for Aigent was better than all other approaches. For measuring the Pielou score, we took 50 original images of each class and then calculated the frequency distribution on the classes of adversarial images generated. Aigent was able to achieve a good Pielou score on all the benchmark datasets. While techniques such as FGSM and DLFuzz have a better Pielou score, it comes at the expense of other metrics such as FID and the percentage of pixels modified. Our method aims at striking a good balance between these metrics as it is crucial for the quality of adversarial images. Figure 6: Adversarial images produced by Aigent (bottom row), and the corresponding original images (top row). Fig. 6 shows sample adversarial images produced for MNIST, CIFAR-10 and ImageNet datasets. The first row contains original images and the second row contains their corresponding adversarial images. Fig. 7 shows a comparison of Aigent with black-box, gradient-based, and coverage-guided approaches, on the same input image. Figure 7: Sample adversarial images from different tools. ## Conclusion Finding adversarial inputs for DNNs is not just useful for identifying situations when a network may behave unexpectedly, but also for adversarial training, which can make the network robust. We have proposed a technique to generate adversarial inputs via patching of neural networks. In our experiments over three benchmark image datasets, we observed that the proposed method is significantly more effective than the existing state-of-the-art, in terms of the naturalness of the adversarial images as well as the fraction of pixels that were modified. There are several interesting directions for future work. Since the proposed method works by finding a patch repeatedly, better algorithms for DNN patching would also make our technique more efficient. Another useful direction would be to find a _minimal_ patch in order to get the _closest_ adversarial example. And even though the technique can, in principle, be applied to DNNs with any activation function, it would be worthwhile to engineer our approach to handle activations other than ReLU efficiently. ## References * Alparslan et al. (2021) Alparslan, Y.; Alparslan, K.; Keim-Shenk, J.; Khade, S.; and Greenstadt, R. 2021\. Adversarial Attacks on Convolutional Neural Networks in Facial Recognition Domain. Technical Report arXiv:2001.11137, arXiv. ArXiv:2001.11137 [cs, eess, stat] type: article. * Deng et al. (2009a) Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; and Fei-Fei, L. 2009a. ImageNet: A large-scale hierarchical image database. In _2009 IEEE Conference on Computer Vision and Pattern Recognition_ , 248–255. * Deng et al. (2009b) Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; and Fei-Fei, L. 2009b. ImageNet: A large-scale hierarchical image database. In _2009 IEEE Conference on Computer Vision and Pattern Recognition_ , 248–255. * Elboher, Gottschlich, and Katz (2020) Elboher, Y. Y.; Gottschlich, J.; and Katz, G. 2020. An Abstraction-Based Framework for Neural Network Verification. In Lahiri, S. K.; and Wang, C., eds., _Computer Aided Verification - 32nd International Conference, CAV 2020, Los Angeles, CA, USA, July 21-24, 2020, Proceedings, Part I_ , volume 12224 of _Lecture Notes in Computer Science_ , 43–65. Springer. * Eniser, Gerasimou, and Sen (2019) Eniser, H. F.; Gerasimou, S.; and Sen, A. 2019. DeepFault: Fault Localization for Deep Neural Networks. In Hähnle, R.; and van der Aalst, W., eds., _Fundamental Approaches to Software Engineering_ , Lecture Notes in Computer Science, 171–191. Cham: Springer International Publishing. ISBN 978-3-030-16722-6. * Gao et al. (2020) Gao, X.; Saha, R. K.; Prasad, M. R.; and Roychoudhury, A. 2020. Fuzz testing based data augmentation to improve robustness of deep neural networks. In _Proceedings of the ACM/IEEE 42nd International Conference on Software Engineering_ , 1147–1158. Seoul South Korea: ACM. ISBN 978-1-4503-7121-6. * Goldberger et al. (2020) Goldberger, B.; Katz, G.; Adi, Y.; and Keshet, J. 2020. Minimal Modifications of Deep Neural Networks using Verification. In Albert, E.; and Kovacs, L., eds., _LPAR23. LPAR-23: 23rd International Conference on Logic for Programming, Artificial Intelligence and Reasoning_ , volume 73 of _EPiC Series in Computing_ , 260–278. EasyChair. * Goodfellow, Shlens, and Szegedy (2015) Goodfellow, I. J.; Shlens, J.; and Szegedy, C. 2015. Explaining and Harnessing Adversarial Examples. Technical Report arXiv:1412.6572, arXiv. ArXiv:1412.6572 [cs, stat] type: article. * Gopinath et al. (2018) Gopinath, D.; Wang, K.; Zhang, M.; Pasareanu, C. S.; and Khurshid, S. 2018. Symbolic Execution for Deep Neural Networks. Technical Report arXiv:1807.10439, arXiv. ArXiv:1807.10439 [cs] type: article. * Guo et al. (2018) Guo, J.; Jiang, Y.; Zhao, Y.; Chen, Q.; and Sun, J. 2018. DLFuzz: Differential Fuzzing Testing of Deep Learning Systems. In _Proceedings of the 2018 26th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering_ , 739–743. ArXiv:1808.09413 [cs]. * Gurobi Optimization, LLC (2022) Gurobi Optimization, LLC. 2022. Gurobi Optimizer Reference Manual. * Harel-Canada et al. (2020) Harel-Canada, F.; Wang, L.; Gulzar, M. A.; Gu, Q.; and Kim, M. 2020. _Is Neuron Coverage a Meaningful Measure for Testing Deep Neural Networks?_ , 851–862. New York, NY, USA: Association for Computing Machinery. ISBN 9781450370431. * Krizhevsky and Hinton (2009) Krizhevsky, A.; and Hinton, G. 2009. Learning multiple layers of features from tiny images. Technical Report 0, University of Toronto, Toronto, Ontario. * Kurakin, Goodfellow, and Bengio (2017) Kurakin, A.; Goodfellow, I.; and Bengio, S. 2017. Adversarial examples in the physical world. Technical Report arXiv:1607.02533, arXiv. ArXiv:1607.02533 [cs, stat] type: article. * Li et al. (2019) Li, J.; Liu, J.; Yang, P.; Chen, L.; Huang, X.; and Zhang, L. 2019. Analyzing Deep Neural Networks with Symbolic Propagation: Towards Higher Precision and Faster Verification. In Chang, B.-Y. E., ed., _Static Analysis_ , 296–319. Cham: Springer International Publishing. ISBN 978-3-030-32304-2. * Pang et al. (2020) Pang, T.; Xu, K.; Dong, Y.; Du, C.; Chen, N.; and Zhu, J. 2020. Rethinking Softmax Cross-Entropy Loss for Adversarial Robustness. Technical Report arXiv:1905.10626, arXiv. ArXiv:1905.10626 [cs, stat] type: article. * Pei et al. (2017) Pei, K.; Cao, Y.; Yang, J.; and Jana, S. 2017. DeepXplore: Automated Whitebox Testing of Deep Learning Systems. In _Proceedings of the 26th Symposium on Operating Systems Principles_ , 1–18. ArXiv:1705.06640 [cs]. * Raj Dhamija, Gunther, and Boult (2019) Raj Dhamija, A.; Gunther, M.; and Boult, T. E. 2019. Improving Deep Network Robustness to Unknown Inputs with Objectosphere. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops_. * Refaeli and Katz (2021) Refaeli, I.; and Katz, G. 2021. Minimal Multi-Layer Modifications of Deep Neural Networks. _ArXiv_ , abs/2110.09929. * Sun et al. (2019) Sun, Y.; Huang, X.; Kroening, D.; Sharp, J.; Hill, M.; and Ashmore, R. 2019. DeepConcolic: Testing and Debugging Deep Neural Networks. In _2019 IEEE/ACM 41st International Conference on Software Engineering: Companion Proceedings (ICSE-Companion)_ , 111–114. ISSN: 2574-1934. * Wang et al. (2018) Wang, S.; Pei, K.; Whitehouse, J.; Yang, J.; and Jana, S. 2018. Formal Security Analysis of Neural Networks Using Symbolic Intervals. In _Proceedings of the 27th USENIX Conference on Security Symposium_ , SEC’18, 1599–1614. USA: USENIX Association. ISBN 9781931971461. * Xie et al. (2019) Xie, X.; Ma, L.; Juefei-Xu, F.; Xue, M.; Chen, H.; Liu, Y.; Zhao, J.; Li, B.; Yin, J.; and See, S. 2019. DeepHunter: a coverage-guided fuzz testing framework for deep neural networks. In _Proceedings of the 28th ACM SIGSOFT International Symposium on Software Testing and Analysis_ , 146–157. New York, NY, USA: Association for Computing Machinery. ISBN 978-1-4503-6224-5. * Yang et al. (2021) Yang, P.; Li, J.; Liu, J.; Huang, C.-C.; Li, R.; Chen, L.; Huang, X.; and Zhang, L. 2021. Enhancing Robustness Verification for Deep Neural Networks via Symbolic Propagation. _Formal Aspects of Computing_ , 33(3): 407–435.
# A note on additive complements of the squares Yuchen Ding, Yu–Chen Sun, Li–Yuan Wang and Yutong Xia (Yuchen Ding) School of Mathematical Science, Yangzhou University, Yangzhou 225002, People’s Republic of China<EMAIL_ADDRESS>(Yu–Chen Sun) Department of Mathematics and Statistics, University of Turku, Turku 20014 , Finland<EMAIL_ADDRESS>(Li-Yuan Wang) School of Physical and Mathematical Sciences, Nanjing Tech University, Nanjing 211816, People’s Republic of China<EMAIL_ADDRESS>(Yutong Xia) School of Mathematical Science, Yangzhou University, Yangzhou 225002, People’s Republic of China<EMAIL_ADDRESS> ###### Abstract. Let $\mathcal{S}=\\{1^{2},2^{2},3^{2},...\\}$ be the set of squares and $\mathcal{W}\subset\mathbb{N}$ be an additive complement of $\mathcal{S}$ so that $\mathcal{S}+\mathcal{W}\supset\\{n\in\mathbb{N}:n\geq N_{0}\\}$ for some $N_{0}$. Let also $\mathcal{R}_{\mathcal{S},\mathcal{W}}(n)=\\#\\{(s,w):n=s+w,s\in\mathcal{S},w\in\mathcal{W}\\}$. Chen and Fang [5] proved that $\sum_{n=1}^{N}R_{\mathcal{S},\mathcal{W}}(n)-N\gg N^{1/4}\log N.$ In this note, we improve their result and obtain that $\sum_{n=1}^{N}R_{\mathcal{S},\mathcal{W}}(n)-N\gg N^{1/3}.$ 2020 Mathematics Subject Classification. Primary 11B13; Secondary 11B75. Keywords. Additive complements, Squares. ## 1\. Introduction Let $\mathbb{N}$ be the set of all non-negative integers. We say that $A,B\subset\mathbb{N}$ are additive complements if $A+B\supset\\{n\in\mathbb{N}:n\geq N_{0}\\}$ for some fixed $N_{0}$. If $A,B$ are additive complements, we say that $A$ is an additive complement of $B$. Let $\mathcal{S}=\\{1^{2},2^{2},3^{2},...\\}$ the set of squares and $\mathcal{W}\subset\mathbb{N}$ be an additive complement of $\mathcal{S}$. Let $N$ be a large integer and $\mathcal{W}(N)$ be the number of elements of $\mathcal{W}$ not exceeding $N$. Clearly, we have $\mathcal{W}(N)\sqrt{N}\geq N-N_{0}$ for some given integer $N_{0}$, from which we deduce that $\liminf_{N\rightarrow\infty}\mathcal{W}(N)/\sqrt{N}\geq 1.$ It is of great interest to ask whether $\liminf_{N\rightarrow\infty}\mathcal{W}(N)/\sqrt{N}$ is strictly larger than $1$. In 1956, Erdős [8] posed this problem, which was later settled affirmatively by Moser [10] with an accurate lower bound $1.06$. The number was then improved in a few articles [1, 2, 4, 7, 11, 12]. Up to now the best result is $\displaystyle\liminf_{N\rightarrow\infty}\mathcal{W}(N)/\sqrt{N}\geq 4/\pi,$ (1.1) given by Cilleruelo [6], Habsieger [9], Balasubramanian and Ramana [3]. Note that $\mathcal{W}(N)/\sqrt{N}>1$ implies that there are some integers $n\in\mathbb{N}$ which do not have an unique representation. Thus it is worthwhile to study the number of representations for writing $n$ as a sum $s+w$ with $s\in\mathcal{S}$ and $w\in\mathcal{W}$. For any $A,B\subset\mathbb{N}$, let $R_{A,B}(n)=\\#\\{(a,b):n=a+b,a\in A,b\in B\\}.$ Obviously, if $A,B$ are complements, then $\liminf_{N\rightarrow\infty}\frac{1}{N}\sum_{n\leq N}R_{A,B}(n)\geq 1.$ Ben Green had a nice observation that for $B=\\{b_{1},b_{2},...\\}$ with $b_{n}=\frac{\pi^{2}}{16}n^{2}+o(n^{2})$, $\lim_{N\rightarrow\infty}\frac{1}{N}\sum_{n=1}^{N}R_{S,B}(n)=1.$ Later, he asked Fang (private communication, see [5]) whether there is an additive complement of $\mathcal{S}$ has the same property as $B$. Namely, can we find an additive complement $\mathcal{W}=\\{w_{n}\\}_{n=1}^{\infty}$ of $\mathcal{S}$ satisfying the asymptotic condition $w_{n}=\frac{\pi^{2}}{16}n^{2}+o(n^{2})$? Motivated by Ben Green’s problem, Chen and Fang [5, Theorem 1.1] proved the following result. Let $N$ be a sufficiently large integer. For any additive complement $\mathcal{W}$ of $\mathcal{S}$, we have $\displaystyle\sum_{n=1}^{N}R_{\mathcal{S},\mathcal{W}}(n)-N\gg\mathcal{W}(2\sqrt{N})\log\mathcal{W}(2\sqrt{N})$ (1.2) From (1.1), one can immediately obtain the following corollary. If $\mathcal{W}$ is any complement of $\mathcal{S}$, then $\sum_{n=1}^{N}R_{\mathcal{S},\mathcal{W}}(n)-N\gg N^{1/4}\log N.$ (1.3) Chen and Fang [5, Remark 1] also gave a counterexample to show that (1.3) does not always hold, if $\mathcal{S}$ and $\mathcal{W}$ are substituted with general additive complements $A,B$. In this note, we will give an improvement of the result of Chen and Fang and get the following theorem. ###### Theorem 1.1. Let $N$ be a sufficiently large integer. For any additive complement $\mathcal{W}$ of $\mathcal{S}$, we have $\sum_{n=1}^{N}R_{\mathcal{S},\mathcal{W}}(n)-N\gg\mathcal{W}\left(\frac{1}{5}N^{2/3}\right).$ ###### Remark 1.1. The coefficient $1/5$ can be further optimized by carefully calculating. From Theorem 1.1 and (1.1), one can immediately obtain the following corollary. ###### Corollary 1.1. For any additive complement $\mathcal{W}$ of $\mathcal{S}$, we have $\sum_{n=1}^{N}R_{\mathcal{S},\mathcal{W}}(n)-N\gg N^{1/3}.$ The proof of Theorem 1.1 is based on the structure of the proof in Chen–Fang [5]. In Chen–Fang’s proof, for $d_{1},d_{2}\in\mathcal{W}$, they considered the equation $x^{2}+d_{1}=y^{2}+d_{2}<N$ and required that this equation has $\gg\log N$ solutions (see [5, Lemma 2.1]). Thus the $\log$-factor, in (1.3), comes from the number of solutions. However, in Chen-Fang’s arguments, the above equation has some solutions $(x,y)$ such that $|x-y|$ is very small, namely $O(1)$, and thus $x+y$ is very large, namely $O(|d_{2}-d_{1}|)$. Since $\max\\{x^{2},y^{2}\\}<N$, we must have $|d_{2}-d_{1}|\ll N^{1/2}$. For this reason, in their arguments, they restricted $d_{1},d_{2}\in\mathcal{W}\cap[cN^{1/2}]$ for some $c>0$. Here $[N]$ denotes the set $\\{n\in\mathbb{N}:n\leq N\\}$. Our new idea is that we exclude those solutions $(x,y)$ such that $|x-y|$ is small. Namely, we require solutions $(x,y)$ to satisfy that $|x-y|\gg N^{1/6}$, so $x+y$ is not too large, namely, $x+y\ll|d_{2}-d_{1}|/N^{1/6}$. Hence, in our case, $\max\\{x^{2},y^{2}\\}<N$ follows from $|d_{2}-d_{1}|/N^{1/6}\ll N^{1/2}$. Thus, one can allow $d_{1}$ and $d_{2}$ to come from the larger set $\mathcal{W}\cap[cN^{1/6+1/2}]$ for some $c>0$ and $c$ will be determined in our proof. Here the threshold $N^{1/6}$ is chosen to balance certain terms in the proof. Consequently, we can improve the lower bound $\mathcal{W}(2N^{1/2})$ in (1.2) to $\mathcal{W}(cN^{2/3})$ for some explicit $c>0$, but the payoff is that since we just pick solutions $(x,y)$ such that $|x-y|$ is large, we can only show that the number of available solutions in our case is $\gg 1$ instead of $\gg\log N$. ###### Acknowledgment. The authors would like to thank Kaisa Matomäki for her helpful comments. The first author was supported by National Natural Science Foundation of China (Grant No. 12201544), Natural Science Foundation of Jiangsu Province of China (Grant No. BK20210784) and China Postdoctoral Science Foundation (Grant No. 2022M710121). He was also supported by foundation numbers JSSCBS20211023 and YZLYJF2020PHD051. The second author was supported by UTUGS funding and was working in the Academy of Finland project No. $333707$. The third author was supported by the National Natural Science Foundation of China (Grant No. 12201291) and the Natural Science Foundation of the Higher Education Institutions of Jiangsu Province (21KJB110001). ## 2\. Proof of Theorem 1.1 The proof of our theorem relies on the following lemma, which is a refinement of the original argument of Chen and Fang [5, Theorem 2.1]. ###### Lemma 2.1. Let $K$ be a positive integer and $\mathcal{D}\subseteq\mathbb{N}$ be such that $4K^{2}|d-d^{\prime}$ for any $d,d^{\prime}\in\mathcal{D}$. Then for all positive numbers $x$, we have $\sum_{\begin{subarray}{c}n\leq x\\\ R_{\mathcal{S},\mathcal{D}}(n)\geq 1\end{subarray}}\left(R_{\mathcal{S},\mathcal{D}}(n)-1\right)\geq\mathcal{D}(2K\sqrt{x})-2.$ ###### Proof. The lemma is trivial if $\mathcal{D}(2K\sqrt{x})\leq 2$. So we only need to consider the case $\mathcal{D}(2K\sqrt{x})>2$. In this case, we assume $\mathcal{D}(2K\sqrt{x})=\ell$ and $\mathcal{D}\cap\left[0,2K\sqrt{x}\right]=\\{d_{1}<d_{2}<\cdot\cdot\cdot<d_{\ell}\\}.$ For any $2<s\leq\ell$, we have $4K^{2}|d_{s}-d_{1}$. So we can assume $d_{s}-d_{1}=4K^{2}n_{s}$ for some positive integer $n_{s}>1$. Then one of the solutions of the equation $x^{2}-y^{2}=d_{s}-d_{1}=4K^{2}n_{s}$ has the form $\begin{cases}x_{s}=Kn_{s}+K,\\\ y_{s}=Kn_{s}-K.\end{cases}$ (2.1) Moreover, the solution given by (2.1) satisfies $\displaystyle x_{s}^{2}+d_{1}=y_{s}^{2}+d_{s}<x,$ (2.2) because $x_{s}^{2}+d_{1}<\left(\frac{d_{s}-d_{1}}{2K}\right)^{2}+2\frac{d_{1}(d_{s}-d_{1})}{(2K)^{2}}<\left(\frac{d_{s}}{2K}\right)^{2}\leq\frac{(2K\sqrt{x})^{2}}{4K^{2}}=x.$ If an integer $n\leq x$ can be written as the sum of $d_{1}$ and a square, then $\displaystyle R_{\mathcal{S},\mathcal{D}}(n)-1\geq\sum_{\begin{subarray}{c}d_{s}>d_{1}\\\ n-d_{s}\in\mathcal{S}\end{subarray}}1,$ from which we deduce that $\sum_{\begin{subarray}{c}n\leq x\\\ R_{\mathcal{S},\mathcal{D}}(n)\geq 1\end{subarray}}\\!\\!\\!\left(R_{\mathcal{S},\mathcal{D}}(n)-1\right)\geq\\!\\!\\!\sum_{\begin{subarray}{c}n\leq x\\\ n-d_{1}\in\mathcal{S}\end{subarray}}\\!\\!\\!\left(R_{\mathcal{S},\mathcal{D}}(n)-1\right)\geq\\!\\!\\!\sum_{\begin{subarray}{c}n\leq x\\\ n-d_{1}\in\mathcal{S}\end{subarray}}\sum_{\begin{subarray}{c}d_{s}>d_{1}\\\ n-d_{s}\in\mathcal{S}\end{subarray}}1=\sum_{\begin{subarray}{c}1<s\leq\ell\end{subarray}}\sum_{\begin{subarray}{c}n\leq x\\\ n-d_{s}\in\mathcal{S}\\\ n-d_{1}\in\mathcal{S}\end{subarray}}1.$ (2.3) For any $2<s\leq\ell$, from (2.1) and (2.2) with $n=x_{s}^{2}+d_{1}$, we have $\sum_{\begin{subarray}{c}n\leq x\\\ n-d_{s}\in\mathcal{S}\\\ n-d_{1}\in\mathcal{S}\end{subarray}}1\geq 1$ It follows that by (2.3), $\sum_{\begin{subarray}{c}n\leq x\\\ R_{\mathcal{S},\mathcal{D}}(n)\geq 1\end{subarray}}\left(R_{\mathcal{S},\mathcal{D}}(n)-1\right)\geq\sum_{\begin{subarray}{c}2<s\leq\ell\end{subarray}}1=\ell-2=\mathcal{D}(2K\sqrt{x})-2.$ ∎ Now, we turn to prove Theorem 1.1. ###### Proof of Theorem 1.1. Let $N$ be a sufficiently large integer. It suffices to prove $\displaystyle\sum_{n=1}^{N}\left(R_{\mathcal{S},\mathcal{W}}(n)-1\right)\gg\mathcal{W}\left(\frac{1}{5}N^{2/3}\right).$ Let $K<N^{1/2}$ be a large number to be chosen later. We follow the proof of Chen and Fang [5]. For any $1\leq j\leq 4K^{2}$, let $\mathcal{W}_{j}=\\{w\in\mathcal{W}:w\equiv j\ ({\rm{mod}}\ 4K^{2})\\}.$ Then we have $\mathcal{W}=\bigcup_{j=1}^{4K^{2}}\mathcal{W}_{j}$ with $\mathcal{W}_{i}\bigcap\mathcal{W}_{j}=\varnothing$ for $i\neq j$. Thus, $\displaystyle\sum_{n=1}^{N}\left(R_{\mathcal{S},\mathcal{W}}(n)-1\right)=\sum_{n=1}^{N}\left(\sum_{j=1}^{4K^{2}}R_{\mathcal{S},\mathcal{W}_{j}}(n)-1\right)$ $\displaystyle\geq$ $\displaystyle\sum_{n=1}^{N}\sum_{\begin{subarray}{c}j=1\\\ R_{\mathcal{S},\mathcal{W}_{j}}(n)\geq 1\end{subarray}}^{4K^{2}}\left(R_{\mathcal{S},\mathcal{W}_{j}}(n)-1\right)=\sum_{j=1}^{4K^{2}}\sum_{\begin{subarray}{c}n=1\\\ R_{\mathcal{S},\mathcal{W}_{j}}(n)\geq 1\end{subarray}}^{N}\left(R_{\mathcal{S},\mathcal{W}_{j}}(n)-1\right).$ (2.4) By Lemma 2.1 with $\mathcal{D}=\mathcal{W}_{j}$, we have $\displaystyle\sum_{\begin{subarray}{c}n=1\\\ R_{\mathcal{S},\mathcal{W}_{j}}(n)\geq 1\end{subarray}}^{N}\left(R_{\mathcal{S},\mathcal{W}_{j}}(n)-1\right)\geq\mathcal{W}_{j}(2K\sqrt{N})-2$ (2.5) for any $1\leq j\leq 4K^{2}$. Combining (2) and (2.5), we obtain $\displaystyle\sum_{n=1}^{N}\left(R_{\mathcal{S},\mathcal{W}}(n)-1\right)\geq\sum_{j=1}^{4K^{2}}(\mathcal{W}_{j}(2K\sqrt{N})-2)=\mathcal{W}(2K\sqrt{N})-8K^{2}.$ (2.6) We choose $K=\frac{1}{10}N^{1/6}$, then our claims follows from (1.1). ∎ ## References * [1] H.L. Abbott, On the additive completion of sets of integers, J. Number Theory, 17 (1983), 135–143. * [2] R. Balasubramanian,On the additive completion of squares, J. Number Theory, 29 (1988), 10–12. * [3] R. Balasubramanian, D.S. Ramana, Additive complements of the squares, C. R. Math. Acad. Sci. Soc. R. Can., 23 (2001), 6–11. * [4] R. Balasubramanian, K. Soundararajan, On the additive completion of squares, II, J. Number Theory, 40 (1992), 127–129. * [5] Y.-G. Chen, J.-H. Fang, Additive complements of the squares, J. Number Theory 180 (2017), 410–422. * [6] J. Cilleruelo, The additive completion of $k$–th powers, J. Number Theory, 44 (1993), 237–243. * [7] R. Donagi, M. Herzog, On the additive completion of polynomial sets of integers, J. Number Theory, 3 (1971), 150–154. * [8] P. Erdős, Problems and Results in Additive Number Theory Colloque sur la Théorie des Nombres, Bruxelles (1955), 127–137 George Thone, Liège; Masson and Cie, Paris, 1956. * [9] L. Habsieger, On the additive completion of polynomial sets, J. Number Theory, 51 (1995), 130–135. * [10] L. Moser, On the additive completion of sets of integers, Proceedings of Symposia in Pure Mathematics, vol. VIII, Amer. Math. Soc., Providence, R.I. (1965), 175–180. * [11] D.S. Ramana, Some Topics in Analytic Number Theory, PhD thesis University of Madras (May 2000). * [12] D.S. Ramana, A report on additive complements of the squares, Number Theory and Discrete Mathematics, Trends Math., Chandigarh, 2000, Birkhäuser, Basel (2002), 161–167.